Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
4,330
3,012,735,799
IssuesEvent
2015-07-29 02:01:51
OData/WebApi
https://api.github.com/repos/OData/WebApi
closed
OData scaffolding in VS2015 should be with OData v4
3: Resolved (code ready)
When building a new OData controller, we will see this, which we're told is OData V3. Let’s get that renamed or removed for RTM. ![scaffolding](https://cloud.githubusercontent.com/assets/7425148/7923050/1d7f2db4-0865-11e5-8485-ee3a18f15f4d.PNG)
1.0
OData scaffolding in VS2015 should be with OData v4 - When building a new OData controller, we will see this, which we're told is OData V3. Let’s get that renamed or removed for RTM. ![scaffolding](https://cloud.githubusercontent.com/assets/7425148/7923050/1d7f2db4-0865-11e5-8485-ee3a18f15f4d.PNG)
non_defect
odata scaffolding in should be with odata when building a new odata controller we will see this which we re told is odata let’s get that renamed or removed for rtm
0
23,655
3,851,865,236
IssuesEvent
2016-04-06 05:27:46
GPF/imame4all
https://api.github.com/repos/GPF/imame4all
closed
X and Y buttons invisible on Landcape Mode on Samsung Galaxy Nexus
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Playing any game in landscape mode. 2. 3. What is the expected output? What do you see instead? X and Y (blue buttons) are supposed to appear, however they are invisible put can still be pressed. What version of the product are you using? On what operating system? MAME4droid (0.37b5) ver, 1.5.2 on Android 4.1.1 using Samsung Galaxy Nexus Please provide any additional information below. ``` Original issue reported on code.google.com by `ugoplata...@gmail.com` on 7 Jan 2013 at 5:45
1.0
X and Y buttons invisible on Landcape Mode on Samsung Galaxy Nexus - ``` What steps will reproduce the problem? 1. Playing any game in landscape mode. 2. 3. What is the expected output? What do you see instead? X and Y (blue buttons) are supposed to appear, however they are invisible put can still be pressed. What version of the product are you using? On what operating system? MAME4droid (0.37b5) ver, 1.5.2 on Android 4.1.1 using Samsung Galaxy Nexus Please provide any additional information below. ``` Original issue reported on code.google.com by `ugoplata...@gmail.com` on 7 Jan 2013 at 5:45
defect
x and y buttons invisible on landcape mode on samsung galaxy nexus what steps will reproduce the problem playing any game in landscape mode what is the expected output what do you see instead x and y blue buttons are supposed to appear however they are invisible put can still be pressed what version of the product are you using on what operating system ver on android using samsung galaxy nexus please provide any additional information below original issue reported on code google com by ugoplata gmail com on jan at
1
81,867
31,777,506,435
IssuesEvent
2023-09-12 15:13:30
idaholab/HERON
https://api.github.com/repos/idaholab/HERON
closed
[DEFECT] Error when running opt mode, templates need an update
defect priority-critical
-------- Defect Description -------- **Describe the defect** Running optimization mode causes an error due to a recent RAVEN update (https://github.com/idaholab/raven/pull/2177) ##### What did you expect to see happen? Running the `optimization_settings` integration test should have passed running HERON tests via Rook. ##### What did you see instead? ``` Traceback (most recent call last): File "C:\Users\sotogj\projects\raven\raven_framework.py", line 25, in <module> sys.exit(main(True)) File "C:\Users\sotogj\projects\raven\ravenframework\Driver.py", line 151, in main simulation.XMLread(root,runInfoSkip=set(["DefaultInputFile"]),xmlFilename=inputFile) File "C:\Users\sotogj\projects\raven\ravenframework\Simulation.py", line 381, in XMLread self.instantiateEntities(xmlNode, runInfoSkip, xmlFilename) File "C:\Users\sotogj\projects\raven\ravenframework\Simulation.py", line 494, in instantiateEntities entity.readXML(block, varGroups, globalAttributes=globalAttributes) File "C:\Users\sotogj\projects\raven\ravenframework\BaseClasses\BaseEntity.py", line 99, in readXML self._readMoreXML(xmlNode) File "C:\Users\sotogj\projects\raven\ravenframework\Steps\Step.py", line 148, in _readMoreXML self._handleInput(paramInput) File "C:\Users\sotogj\projects\raven\ravenframework\Steps\Step.py", line 195, in _handleInput self._localInputAndCheckParam(paramInput) File "C:\Users\sotogj\projects\raven\ravenframework\Steps\MultiRun.py", line 59, in _localInputAndCheckParam SingleRun._localInputAndCheckParam(self,paramInput) File "C:\Users\sotogj\projects\raven\ravenframework\Steps\SingleRun.py", line 99, in _localInputAndCheckParam Models.validate(self.parList[modelIndex][2], role, toBeTested[role]) File "C:\Users\sotogj\projects\raven\ravenframework\Models\Factory.py", line 50, in validate return factory.returnClass(className).localValidateMethod(role, what) File "C:\Users\sotogj\projects\raven\ravenframework\Models\Model.py", line 190, in localValidateMethod raise IOError('The number of times class = '+str(tester['class'])+' type= ' +str(tester['type'])+' is used as '+str(who)+' is improper. Number of allowable times is '+str(tester['multiplicity'])+'.Got '+str(tester['tempCounter'])) OSError: The number of times class = Optimizers type= ['GradientDescent', 'GeneticAlgorithm', 'SimulatedAnnealing', 'BayesianOptimizer'] is used as Optimizer is improper. Number of allowable times is 1.Got 0 ``` ##### Do you have a suggested fix for the development team? Switch the Optimizer class type from 'FiniteDifference' to 'GradientDescent' in the outer XML template. **Describe how to Reproduce** Steps to reproduce the behavior: 1. 2. 3. 4. **Screenshots and Input Files** Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. **Platform (please complete the following information):** - OS: [e.g. iOS] - Version: [e.g. 22] - Dependencies Installation: [CONDA or PIP] ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or task? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [x] 1. If the issue is a defect, is the defect fixed? - [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)? - [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
1.0
[DEFECT] Error when running opt mode, templates need an update - -------- Defect Description -------- **Describe the defect** Running optimization mode causes an error due to a recent RAVEN update (https://github.com/idaholab/raven/pull/2177) ##### What did you expect to see happen? Running the `optimization_settings` integration test should have passed running HERON tests via Rook. ##### What did you see instead? ``` Traceback (most recent call last): File "C:\Users\sotogj\projects\raven\raven_framework.py", line 25, in <module> sys.exit(main(True)) File "C:\Users\sotogj\projects\raven\ravenframework\Driver.py", line 151, in main simulation.XMLread(root,runInfoSkip=set(["DefaultInputFile"]),xmlFilename=inputFile) File "C:\Users\sotogj\projects\raven\ravenframework\Simulation.py", line 381, in XMLread self.instantiateEntities(xmlNode, runInfoSkip, xmlFilename) File "C:\Users\sotogj\projects\raven\ravenframework\Simulation.py", line 494, in instantiateEntities entity.readXML(block, varGroups, globalAttributes=globalAttributes) File "C:\Users\sotogj\projects\raven\ravenframework\BaseClasses\BaseEntity.py", line 99, in readXML self._readMoreXML(xmlNode) File "C:\Users\sotogj\projects\raven\ravenframework\Steps\Step.py", line 148, in _readMoreXML self._handleInput(paramInput) File "C:\Users\sotogj\projects\raven\ravenframework\Steps\Step.py", line 195, in _handleInput self._localInputAndCheckParam(paramInput) File "C:\Users\sotogj\projects\raven\ravenframework\Steps\MultiRun.py", line 59, in _localInputAndCheckParam SingleRun._localInputAndCheckParam(self,paramInput) File "C:\Users\sotogj\projects\raven\ravenframework\Steps\SingleRun.py", line 99, in _localInputAndCheckParam Models.validate(self.parList[modelIndex][2], role, toBeTested[role]) File "C:\Users\sotogj\projects\raven\ravenframework\Models\Factory.py", line 50, in validate return factory.returnClass(className).localValidateMethod(role, what) File "C:\Users\sotogj\projects\raven\ravenframework\Models\Model.py", line 190, in localValidateMethod raise IOError('The number of times class = '+str(tester['class'])+' type= ' +str(tester['type'])+' is used as '+str(who)+' is improper. Number of allowable times is '+str(tester['multiplicity'])+'.Got '+str(tester['tempCounter'])) OSError: The number of times class = Optimizers type= ['GradientDescent', 'GeneticAlgorithm', 'SimulatedAnnealing', 'BayesianOptimizer'] is used as Optimizer is improper. Number of allowable times is 1.Got 0 ``` ##### Do you have a suggested fix for the development team? Switch the Optimizer class type from 'FiniteDifference' to 'GradientDescent' in the outer XML template. **Describe how to Reproduce** Steps to reproduce the behavior: 1. 2. 3. 4. **Screenshots and Input Files** Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. **Platform (please complete the following information):** - OS: [e.g. iOS] - Version: [e.g. 22] - Dependencies Installation: [CONDA or PIP] ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or task? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [x] 1. If the issue is a defect, is the defect fixed? - [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)? - [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
defect
error when running opt mode templates need an update defect description describe the defect running optimization mode causes an error due to a recent raven update what did you expect to see happen running the optimization settings integration test should have passed running heron tests via rook what did you see instead traceback most recent call last file c users sotogj projects raven raven framework py line in sys exit main true file c users sotogj projects raven ravenframework driver py line in main simulation xmlread root runinfoskip set xmlfilename inputfile file c users sotogj projects raven ravenframework simulation py line in xmlread self instantiateentities xmlnode runinfoskip xmlfilename file c users sotogj projects raven ravenframework simulation py line in instantiateentities entity readxml block vargroups globalattributes globalattributes file c users sotogj projects raven ravenframework baseclasses baseentity py line in readxml self readmorexml xmlnode file c users sotogj projects raven ravenframework steps step py line in readmorexml self handleinput paraminput file c users sotogj projects raven ravenframework steps step py line in handleinput self localinputandcheckparam paraminput file c users sotogj projects raven ravenframework steps multirun py line in localinputandcheckparam singlerun localinputandcheckparam self paraminput file c users sotogj projects raven ravenframework steps singlerun py line in localinputandcheckparam models validate self parlist role tobetested file c users sotogj projects raven ravenframework models factory py line in validate return factory returnclass classname localvalidatemethod role what file c users sotogj projects raven ravenframework models model py line in localvalidatemethod raise ioerror the number of times class str tester type str tester is used as str who is improper number of allowable times is str tester got str tester oserror the number of times class optimizers type is used as optimizer is improper number of allowable times is got do you have a suggested fix for the development team switch the optimizer class type from finitedifference to gradientdescent in the outer xml template describe how to reproduce steps to reproduce the behavior screenshots and input files please attach the input file s that generate this error the simpler the input the faster we can find the issue platform please complete the following information os version dependencies installation for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided
1
61,074
17,023,595,174
IssuesEvent
2021-07-03 02:50:01
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Merkaartor TMS implementation is backwards
Component: merkaartor Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 11.29am, Saturday, 22nd May 2010]** Hi, After a bit of frustration ... it appears that Merkaartor (well, Slippy tiles really) follows Google maps in making the TMS tile grid start from 0,0 in the top-left, not the bottom left as specified by the TMS spec. Slippy tiles spec: http://wiki.openstreetmap.org/wiki/Slippy_map_tilenames#X_and_Y TMS spec: http://wiki.osgeo.org/wiki/Tile_Map_Service_Specification#Tile_Resources gdal2tiles.py code comments on the subject: https://trac.osgeo.org/gdal/browser/trunk/gdal/swig/python/scripts/gdal2tiles.py?rev=19288#L146 I'm not asking that Slippy tiles, T@H, CycleMap, etc all suddenly turn on their head and start following the official spec, but as Merkaartor calls it "TMS" it would be nice if there were at least a tickbox on the Merkaartor `TMS server editor` setup that let you choose the alternate convention for those times you would like to connect to a TMS which is following the spec. (I'm not sure which spec came first, but at this point they are both established so it doesn't really matter) Manifestation: TMS tilesets created with GDAL's gdal2tiles.py script don't show up where you expect them (they end up reversed in the wrong hemisphere). thanks, Hamish
1.0
Merkaartor TMS implementation is backwards - **[Submitted to the original trac issue database at 11.29am, Saturday, 22nd May 2010]** Hi, After a bit of frustration ... it appears that Merkaartor (well, Slippy tiles really) follows Google maps in making the TMS tile grid start from 0,0 in the top-left, not the bottom left as specified by the TMS spec. Slippy tiles spec: http://wiki.openstreetmap.org/wiki/Slippy_map_tilenames#X_and_Y TMS spec: http://wiki.osgeo.org/wiki/Tile_Map_Service_Specification#Tile_Resources gdal2tiles.py code comments on the subject: https://trac.osgeo.org/gdal/browser/trunk/gdal/swig/python/scripts/gdal2tiles.py?rev=19288#L146 I'm not asking that Slippy tiles, T@H, CycleMap, etc all suddenly turn on their head and start following the official spec, but as Merkaartor calls it "TMS" it would be nice if there were at least a tickbox on the Merkaartor `TMS server editor` setup that let you choose the alternate convention for those times you would like to connect to a TMS which is following the spec. (I'm not sure which spec came first, but at this point they are both established so it doesn't really matter) Manifestation: TMS tilesets created with GDAL's gdal2tiles.py script don't show up where you expect them (they end up reversed in the wrong hemisphere). thanks, Hamish
defect
merkaartor tms implementation is backwards hi after a bit of frustration it appears that merkaartor well slippy tiles really follows google maps in making the tms tile grid start from in the top left not the bottom left as specified by the tms spec slippy tiles spec tms spec py code comments on the subject i m not asking that slippy tiles t h cyclemap etc all suddenly turn on their head and start following the official spec but as merkaartor calls it tms it would be nice if there were at least a tickbox on the merkaartor tms server editor setup that let you choose the alternate convention for those times you would like to connect to a tms which is following the spec i m not sure which spec came first but at this point they are both established so it doesn t really matter manifestation tms tilesets created with gdal s py script don t show up where you expect them they end up reversed in the wrong hemisphere thanks hamish
1
34,414
7,451,215,926
IssuesEvent
2018-03-29 01:35:25
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
KY muutmisel nupp Tühista võiks anda tagasisidet
P: normal R: fixed T: defect
**Reported by aadikaljuvee on 13 Mar 2013 09:46 UTC** Olen KY muutmise/lisamise lehel http://rahvusarhiiv.tietotest.ee/et/description_unit/edit/110100615521/ Praegu nupp Tühista ei tee _pealtnäha_ midagi. Kasutajaile oleks selgem, kui Tühista järel jõuaks kasutaja tagasi KY muutmise/lisamise eelselt avatud olnud lehele (ehk enamasti KY andmete vaatamise lehele). (Ainult Bread-crumb linkide kaudu tagasi liikumine ei ole alati intuitiivne)
1.0
KY muutmisel nupp Tühista võiks anda tagasisidet - **Reported by aadikaljuvee on 13 Mar 2013 09:46 UTC** Olen KY muutmise/lisamise lehel http://rahvusarhiiv.tietotest.ee/et/description_unit/edit/110100615521/ Praegu nupp Tühista ei tee _pealtnäha_ midagi. Kasutajaile oleks selgem, kui Tühista järel jõuaks kasutaja tagasi KY muutmise/lisamise eelselt avatud olnud lehele (ehk enamasti KY andmete vaatamise lehele). (Ainult Bread-crumb linkide kaudu tagasi liikumine ei ole alati intuitiivne)
defect
ky muutmisel nupp tühista võiks anda tagasisidet reported by aadikaljuvee on mar utc olen ky muutmise lisamise lehel praegu nupp tühista ei tee pealtnäha midagi kasutajaile oleks selgem kui tühista järel jõuaks kasutaja tagasi ky muutmise lisamise eelselt avatud olnud lehele ehk enamasti ky andmete vaatamise lehele ainult bread crumb linkide kaudu tagasi liikumine ei ole alati intuitiivne
1
519,172
15,046,496,732
IssuesEvent
2021-02-03 07:29:21
redhat-developer/vscode-openshift-tools
https://api.github.com/repos/redhat-developer/vscode-openshift-tools
opened
Create 'Components View'
demo kind/feature priority/major
Extension reveals developer related resources like: 1. Components (s2i, devfile) 2. Versions (s2i) 3. Starter Projects (devile) only as simple odo command running in terminal or in the middle of create component workflow. Former is just a text output in terminal which has no means to start using described resources. Latter is buried in the middle of workflow and required to do several steps just to see what is in store. Extension should provide the view where all resources presented as a tree and has context menu/default commands to trigger workflow to use that item to create s2i component with specific version and devfile component based on specific starter project.
1.0
Create 'Components View' - Extension reveals developer related resources like: 1. Components (s2i, devfile) 2. Versions (s2i) 3. Starter Projects (devile) only as simple odo command running in terminal or in the middle of create component workflow. Former is just a text output in terminal which has no means to start using described resources. Latter is buried in the middle of workflow and required to do several steps just to see what is in store. Extension should provide the view where all resources presented as a tree and has context menu/default commands to trigger workflow to use that item to create s2i component with specific version and devfile component based on specific starter project.
non_defect
create components view extension reveals developer related resources like components devfile versions starter projects devile only as simple odo command running in terminal or in the middle of create component workflow former is just a text output in terminal which has no means to start using described resources latter is buried in the middle of workflow and required to do several steps just to see what is in store extension should provide the view where all resources presented as a tree and has context menu default commands to trigger workflow to use that item to create component with specific version and devfile component based on specific starter project
0
376,986
11,161,231,517
IssuesEvent
2019-12-26 12:57:22
myceworld/myce
https://api.github.com/repos/myceworld/myce
opened
[marketing] Faucet
Priority: Low Status: Help Type: Enhancement
**Describe the solution** New user flow that ends in receiving coins to showcase how fast and easy myce is. 1) explain myce 2) download myce wallet 3) receive coins **Describe reference or alternative examples** https://github.com/yasmikash/cryptomaze made in python/flask
1.0
[marketing] Faucet - **Describe the solution** New user flow that ends in receiving coins to showcase how fast and easy myce is. 1) explain myce 2) download myce wallet 3) receive coins **Describe reference or alternative examples** https://github.com/yasmikash/cryptomaze made in python/flask
non_defect
faucet describe the solution new user flow that ends in receiving coins to showcase how fast and easy myce is explain myce download myce wallet receive coins describe reference or alternative examples made in python flask
0
21,364
10,606,556,383
IssuesEvent
2019-10-10 23:51:59
isubasinghe/gatsby-contentful-portfolio
https://api.github.com/repos/isubasinghe/gatsby-contentful-portfolio
opened
CVE-2018-16487 (High) detected in lodash-4.11.1.tgz, lodash-3.10.1.tgz
security vulnerability
## CVE-2018-16487 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-4.11.1.tgz</b>, <b>lodash-3.10.1.tgz</b></p></summary> <p> <details><summary><b>lodash-4.11.1.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.11.1.tgz">https://registry.npmjs.org/lodash/-/lodash-4.11.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/gatsby-contentful-portfolio/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/gatsby-contentful-portfolio/node_modules/webpack-validator/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - gatsby-1.9.279.tgz (Root Library) - webpack-validator-2.3.0.tgz - :x: **lodash-4.11.1.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/gatsby-contentful-portfolio/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/gatsby-contentful-portfolio/node_modules/webpack-configurator/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - gatsby-1.9.279.tgz (Root Library) - webpack-configurator-0.3.1.tgz - :x: **lodash-3.10.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/isubasinghe/gatsby-contentful-portfolio/commit/f33d27b6f7e490a39f317d19ad6eecafc73ac173">f33d27b6f7e490a39f317d19ad6eecafc73ac173</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-16487>CVE-2018-16487</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-16487 (High) detected in lodash-4.11.1.tgz, lodash-3.10.1.tgz - ## CVE-2018-16487 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-4.11.1.tgz</b>, <b>lodash-3.10.1.tgz</b></p></summary> <p> <details><summary><b>lodash-4.11.1.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.11.1.tgz">https://registry.npmjs.org/lodash/-/lodash-4.11.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/gatsby-contentful-portfolio/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/gatsby-contentful-portfolio/node_modules/webpack-validator/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - gatsby-1.9.279.tgz (Root Library) - webpack-validator-2.3.0.tgz - :x: **lodash-4.11.1.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/gatsby-contentful-portfolio/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/gatsby-contentful-portfolio/node_modules/webpack-configurator/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - gatsby-1.9.279.tgz (Root Library) - webpack-configurator-0.3.1.tgz - :x: **lodash-3.10.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/isubasinghe/gatsby-contentful-portfolio/commit/f33d27b6f7e490a39f317d19ad6eecafc73ac173">f33d27b6f7e490a39f317d19ad6eecafc73ac173</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-16487>CVE-2018-16487</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in lodash tgz lodash tgz cve high severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz lodash modular utilities library home page a href path to dependency file tmp ws scm gatsby contentful portfolio package json path to vulnerable library tmp ws scm gatsby contentful portfolio node modules webpack validator node modules lodash package json dependency hierarchy gatsby tgz root library webpack validator tgz x lodash tgz vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file tmp ws scm gatsby contentful portfolio package json path to vulnerable library tmp ws scm gatsby contentful portfolio node modules webpack configurator node modules lodash package json dependency hierarchy gatsby tgz root library webpack configurator tgz x lodash tgz vulnerable library found in head commit a href vulnerability details a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
821,345
30,818,396,527
IssuesEvent
2023-08-01 14:47:40
broadinstitute/gnomad-browser
https://api.github.com/repos/broadinstitute/gnomad-browser
closed
Don't show variant cooccurrence table or constraint if not using a v2 dataset
Type: Bug Priority: High
per Anne, "The variant co-occurrence data should not be shown on gnomAD v3 or SV v3 (same as constraint data) as it was generated on v2 (but we will eventually generate it on v3)." (I assume that applies to Exac as well). Should be a simple matter of adding another `dataset-metadata` flag and checking that in the appropriate place.
1.0
Don't show variant cooccurrence table or constraint if not using a v2 dataset - per Anne, "The variant co-occurrence data should not be shown on gnomAD v3 or SV v3 (same as constraint data) as it was generated on v2 (but we will eventually generate it on v3)." (I assume that applies to Exac as well). Should be a simple matter of adding another `dataset-metadata` flag and checking that in the appropriate place.
non_defect
don t show variant cooccurrence table or constraint if not using a dataset per anne the variant co occurrence data should not be shown on gnomad or sv same as constraint data as it was generated on but we will eventually generate it on i assume that applies to exac as well should be a simple matter of adding another dataset metadata flag and checking that in the appropriate place
0
70,029
22,793,066,007
IssuesEvent
2022-07-10 09:49:24
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
0.7.13 / CentOS 7.9 - list_del corruption
Type: Defect Status: Stale Status: Triage Needed
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Distribution Name | CentOS Distribution Version | 7.9 Linux Kernel | 3.10.0-1062.18.1.el7.DPC_enabled Architecture | x86_64 ZFS Version | 0.7.13 SPL Version | 0.7.13 Additional Info | All NVMe zpool, Lustre MDT use case <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing Unknown what I/O pattern was occurring in the middle of the night but something caused ZFS to cough up a hairball, specifically "list_del corruption ffff8e1b9ca82b78->next is LIST_POISON1 (dead000000000100)" Note on kernel version, regular 3.10.0-1062.18.1.el7, recompiled to enable DPC (PCIe downstream port containment) ### Describe how to reproduce the problem I don't know what I/O was occurring at the time so reproduction isn't straightforward. System required a reboot, zfs was permanently ticked off until cleared by reboot. Traces kept looping incessantly. I have seen other reports similar so I am adding this to the mix. https://github.com/openzfs/zfs/issues/9068 https://github.com/openzfs/zfs/issues/10707 ### Include any warning/errors/backtraces from the system logs ``` [Wed Mar 31 06:57:46 2021] WARNING: CPU: 48 PID: 10073 at lib/list_debug.c:53 __list_del_entry+0x63/0xd0 [Wed Mar 31 06:57:46 2021] list_del corruption, ffff8e1b9ca82b78->next is LIST_POISON1 (dead000000000100) [Wed Mar 31 06:57:46 2021] Modules linked in: osp(OE) mdd(OE) lod(OE) mdt(OE) lfsck(OE) mgs(OE) mgc(OE) osd_zfs(OE) lquota(OE) nfsv3 nfs_acl rpcsec_gss_krb5 auth_rpcgss nfsv4 dns_resolver nfs lockd grace fscache lustre(OE) lmv(OE) mdc(OE) osc(OE) lov(OE) fid(OE) fld(OE) ko2iblnd(OE) ptlrpc(OE) obdclass(OE) lnet(OE) libcfs(OE) ntb_perf 8021q garp mrp stp llc ntb_transport ntb rdma_ucm(OE) ib_ucm(OE) rdma_cm(OE) iw_cm(OE) ib_ipoib(OE) ib_cm(OE) ib_umad(OE) mlx5_fpga_tools(OE) mlx4_en(OE) mlx4_ib(OE) mlx4_core(OE) ip6_tables iptable_filter sunrpc vfat fat amd64_edac_mod edac_mce_amd kvm_amd kvm irqbypass crc32 _pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr cls_switchtec(OE) sg k10temp i2c_piix4 ipmi_si ipmi_devintf ipmi_msghandler i2c_designware_platform pinctrl_amd i2c_designware_core [Wed Mar 31 06:57:46 2021] pcc_cpufreq acpi_cpufreq binfmt_misc knem(OE) ip_tables ext4 mbcache jbd2 raid1 sd_mod crc_t10dif crct10dif_generic zfs(POE) zunicode(POE) zavl(POE) icp(POE) zcommon(POE) znvpair(POE) spl(OE) mlx5_ib(OE) ib_uverbs(OE) ib_core(OE) mlx5_core(OE) ast mlxfw(OE) drm_kms_helper vfio_m dev(OE) vfio_iommu_type1 syscopyarea sysfillrect vfio sysimgblt mdev(OE) fb_sys_fops crct10dif_pclmul crct10dif_common devlink ttm crc32c_intel mlx_compat(OE) ahci drm igb libahci nvme ptp dca libata nvme_core drm_panel_orientation_quirks i2c_algo_bit pps_core nfit libnvdimm [last unloaded: libcfs] [Wed Mar 31 06:57:46 2021] CPU: 48 PID: 10073 Comm: dp_sync_taskq Kdump: loaded Tainted: P W OE ------------ 3.10.0-1062.18.1.el7.DPC_enabled.x86_64 #1 [Wed Mar 31 06:57:46 2021] Hardware name: Aeon Computing R182-Z90-00/MZ92-FS0-00, BIOS R20g 01/27/2021 [Wed Mar 31 06:57:46 2021] Call Trace: [Wed Mar 31 06:57:46 2021] [<ffffffff9517c416>] dump_stack+0x19/0x1b [Wed Mar 31 06:57:46 2021] [<ffffffff94a9bab8>] __warn+0xd8/0x100 [Wed Mar 31 06:57:46 2021] [<ffffffff94a9bb3f>] warn_slowpath_fmt+0x5f/0x80 [Wed Mar 31 06:57:46 2021] [<ffffffffc0f01dae>] ? dnode_rele_and_unlock+0x5e/0xc0 [zfs] [Wed Mar 31 06:57:46 2021] [<ffffffff94da1483>] __list_del_entry+0x63/0xd0 [Wed Mar 31 06:57:46 2021] [<ffffffff94da14fd>] list_del+0xd/0x30 [Wed Mar 31 06:57:46 2021] [<ffffffffc0f2a585>] multilist_sublist_remove+0x15/0x20 [zfs] [Wed Mar 31 06:57:46 2021] [<ffffffffc0ef30bf>] userquota_updates_task+0xff/0x5b0 [zfs] [Wed Mar 31 06:57:46 2021] [<ffffffffc0ef03f0>] ? dmu_objset_userobjspace_upgradable+0x60/0x60 [zfs] [Wed Mar 31 06:57:46 2021] [<ffffffffc0ef03f0>] ? dmu_objset_userobjspace_upgradable+0x60/0x60 [zfs] [Wed Mar 31 06:57:46 2021] [<ffffffff94da14fd>] ? list_del+0xd/0x30 [Wed Mar 31 06:57:46 2021] [<ffffffffc02c6d7c>] taskq_thread+0x2ac/0x4f0 [spl] [Wed Mar 31 06:57:46 2021] [<ffffffff94adb600>] ? wake_up_state+0x20/0x20 [Wed Mar 31 06:57:46 2021] [<ffffffffc02c6ad0>] ? taskq_thread_spawn+0x60/0x60 [spl] [Wed Mar 31 06:57:46 2021] [<ffffffff94ac6321>] kthread+0xd1/0xe0 [Wed Mar 31 06:57:46 2021] [<ffffffff94ac6250>] ? insert_kthread_work+0x40/0x40 [Wed Mar 31 06:57:46 2021] [<ffffffff9518ed24>] ret_from_fork_nospec_begin+0xe/0x21 [Wed Mar 31 06:57:46 2021] [<ffffffff94ac6250>] ? insert_kthread_work+0x40/0x40 [Wed Mar 31 06:57:46 2021] ---[ end trace 7eb94146e5b7009d ]--- ```
1.0
0.7.13 / CentOS 7.9 - list_del corruption - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Distribution Name | CentOS Distribution Version | 7.9 Linux Kernel | 3.10.0-1062.18.1.el7.DPC_enabled Architecture | x86_64 ZFS Version | 0.7.13 SPL Version | 0.7.13 Additional Info | All NVMe zpool, Lustre MDT use case <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing Unknown what I/O pattern was occurring in the middle of the night but something caused ZFS to cough up a hairball, specifically "list_del corruption ffff8e1b9ca82b78->next is LIST_POISON1 (dead000000000100)" Note on kernel version, regular 3.10.0-1062.18.1.el7, recompiled to enable DPC (PCIe downstream port containment) ### Describe how to reproduce the problem I don't know what I/O was occurring at the time so reproduction isn't straightforward. System required a reboot, zfs was permanently ticked off until cleared by reboot. Traces kept looping incessantly. I have seen other reports similar so I am adding this to the mix. https://github.com/openzfs/zfs/issues/9068 https://github.com/openzfs/zfs/issues/10707 ### Include any warning/errors/backtraces from the system logs ``` [Wed Mar 31 06:57:46 2021] WARNING: CPU: 48 PID: 10073 at lib/list_debug.c:53 __list_del_entry+0x63/0xd0 [Wed Mar 31 06:57:46 2021] list_del corruption, ffff8e1b9ca82b78->next is LIST_POISON1 (dead000000000100) [Wed Mar 31 06:57:46 2021] Modules linked in: osp(OE) mdd(OE) lod(OE) mdt(OE) lfsck(OE) mgs(OE) mgc(OE) osd_zfs(OE) lquota(OE) nfsv3 nfs_acl rpcsec_gss_krb5 auth_rpcgss nfsv4 dns_resolver nfs lockd grace fscache lustre(OE) lmv(OE) mdc(OE) osc(OE) lov(OE) fid(OE) fld(OE) ko2iblnd(OE) ptlrpc(OE) obdclass(OE) lnet(OE) libcfs(OE) ntb_perf 8021q garp mrp stp llc ntb_transport ntb rdma_ucm(OE) ib_ucm(OE) rdma_cm(OE) iw_cm(OE) ib_ipoib(OE) ib_cm(OE) ib_umad(OE) mlx5_fpga_tools(OE) mlx4_en(OE) mlx4_ib(OE) mlx4_core(OE) ip6_tables iptable_filter sunrpc vfat fat amd64_edac_mod edac_mce_amd kvm_amd kvm irqbypass crc32 _pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr cls_switchtec(OE) sg k10temp i2c_piix4 ipmi_si ipmi_devintf ipmi_msghandler i2c_designware_platform pinctrl_amd i2c_designware_core [Wed Mar 31 06:57:46 2021] pcc_cpufreq acpi_cpufreq binfmt_misc knem(OE) ip_tables ext4 mbcache jbd2 raid1 sd_mod crc_t10dif crct10dif_generic zfs(POE) zunicode(POE) zavl(POE) icp(POE) zcommon(POE) znvpair(POE) spl(OE) mlx5_ib(OE) ib_uverbs(OE) ib_core(OE) mlx5_core(OE) ast mlxfw(OE) drm_kms_helper vfio_m dev(OE) vfio_iommu_type1 syscopyarea sysfillrect vfio sysimgblt mdev(OE) fb_sys_fops crct10dif_pclmul crct10dif_common devlink ttm crc32c_intel mlx_compat(OE) ahci drm igb libahci nvme ptp dca libata nvme_core drm_panel_orientation_quirks i2c_algo_bit pps_core nfit libnvdimm [last unloaded: libcfs] [Wed Mar 31 06:57:46 2021] CPU: 48 PID: 10073 Comm: dp_sync_taskq Kdump: loaded Tainted: P W OE ------------ 3.10.0-1062.18.1.el7.DPC_enabled.x86_64 #1 [Wed Mar 31 06:57:46 2021] Hardware name: Aeon Computing R182-Z90-00/MZ92-FS0-00, BIOS R20g 01/27/2021 [Wed Mar 31 06:57:46 2021] Call Trace: [Wed Mar 31 06:57:46 2021] [<ffffffff9517c416>] dump_stack+0x19/0x1b [Wed Mar 31 06:57:46 2021] [<ffffffff94a9bab8>] __warn+0xd8/0x100 [Wed Mar 31 06:57:46 2021] [<ffffffff94a9bb3f>] warn_slowpath_fmt+0x5f/0x80 [Wed Mar 31 06:57:46 2021] [<ffffffffc0f01dae>] ? dnode_rele_and_unlock+0x5e/0xc0 [zfs] [Wed Mar 31 06:57:46 2021] [<ffffffff94da1483>] __list_del_entry+0x63/0xd0 [Wed Mar 31 06:57:46 2021] [<ffffffff94da14fd>] list_del+0xd/0x30 [Wed Mar 31 06:57:46 2021] [<ffffffffc0f2a585>] multilist_sublist_remove+0x15/0x20 [zfs] [Wed Mar 31 06:57:46 2021] [<ffffffffc0ef30bf>] userquota_updates_task+0xff/0x5b0 [zfs] [Wed Mar 31 06:57:46 2021] [<ffffffffc0ef03f0>] ? dmu_objset_userobjspace_upgradable+0x60/0x60 [zfs] [Wed Mar 31 06:57:46 2021] [<ffffffffc0ef03f0>] ? dmu_objset_userobjspace_upgradable+0x60/0x60 [zfs] [Wed Mar 31 06:57:46 2021] [<ffffffff94da14fd>] ? list_del+0xd/0x30 [Wed Mar 31 06:57:46 2021] [<ffffffffc02c6d7c>] taskq_thread+0x2ac/0x4f0 [spl] [Wed Mar 31 06:57:46 2021] [<ffffffff94adb600>] ? wake_up_state+0x20/0x20 [Wed Mar 31 06:57:46 2021] [<ffffffffc02c6ad0>] ? taskq_thread_spawn+0x60/0x60 [spl] [Wed Mar 31 06:57:46 2021] [<ffffffff94ac6321>] kthread+0xd1/0xe0 [Wed Mar 31 06:57:46 2021] [<ffffffff94ac6250>] ? insert_kthread_work+0x40/0x40 [Wed Mar 31 06:57:46 2021] [<ffffffff9518ed24>] ret_from_fork_nospec_begin+0xe/0x21 [Wed Mar 31 06:57:46 2021] [<ffffffff94ac6250>] ? insert_kthread_work+0x40/0x40 [Wed Mar 31 06:57:46 2021] ---[ end trace 7eb94146e5b7009d ]--- ```
defect
centos list del corruption thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information distribution name centos distribution version linux kernel dpc enabled architecture zfs version spl version additional info all nvme zpool lustre mdt use case commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing unknown what i o pattern was occurring in the middle of the night but something caused zfs to cough up a hairball specifically list del corruption next is list note on kernel version regular recompiled to enable dpc pcie downstream port containment describe how to reproduce the problem i don t know what i o was occurring at the time so reproduction isn t straightforward system required a reboot zfs was permanently ticked off until cleared by reboot traces kept looping incessantly i have seen other reports similar so i am adding this to the mix include any warning errors backtraces from the system logs warning cpu pid at lib list debug c list del entry list del corruption next is list modules linked in osp oe mdd oe lod oe mdt oe lfsck oe mgs oe mgc oe osd zfs oe lquota oe nfs acl rpcsec gss auth rpcgss dns resolver nfs lockd grace fscache lustre oe lmv oe mdc oe osc oe lov oe fid oe fld oe oe ptlrpc oe obdclass oe lnet oe libcfs oe ntb perf garp mrp stp llc ntb transport ntb rdma ucm oe ib ucm oe rdma cm oe iw cm oe ib ipoib oe ib cm oe ib umad oe fpga tools oe en oe ib oe core oe tables iptable filter sunrpc vfat fat edac mod edac mce amd kvm amd kvm irqbypass pclmul ghash clmulni intel aesni intel lrw glue helper ablk helper cryptd pcspkr cls switchtec oe sg ipmi si ipmi devintf ipmi msghandler designware platform pinctrl amd designware core pcc cpufreq acpi cpufreq binfmt misc knem oe ip tables mbcache sd mod crc generic zfs poe zunicode poe zavl poe icp poe zcommon poe znvpair poe spl oe ib oe ib uverbs oe ib core oe core oe ast mlxfw oe drm kms helper vfio m dev oe vfio iommu syscopyarea sysfillrect vfio sysimgblt mdev oe fb sys fops pclmul common devlink ttm intel mlx compat oe ahci drm igb libahci nvme ptp dca libata nvme core drm panel orientation quirks algo bit pps core nfit libnvdimm cpu pid comm dp sync taskq kdump loaded tainted p w oe dpc enabled hardware name aeon computing bios call trace dump stack warn warn slowpath fmt dnode rele and unlock list del entry list del multilist sublist remove userquota updates task dmu objset userobjspace upgradable dmu objset userobjspace upgradable list del taskq thread wake up state taskq thread spawn kthread insert kthread work ret from fork nospec begin insert kthread work
1
72,858
24,335,001,480
IssuesEvent
2022-10-01 01:23:55
scipy/scipy
https://api.github.com/repos/scipy/scipy
opened
BUG: osx-64 scipy 1.9.1 test_bad_geneig numerical error
defect
### Describe your issue. The test test_bad_geneig is failing on scipy 1.9.1 osx-64 with openblas 0.3.21 for py38, py39 and py310. It looks like a permutation issue. ### Reproducing Code Example ```python This is reproducible on osx-64 with openblas 0.3.21 for py38, py39 and py310. python -c 'import scipy, sys; scipy.test(verbose=2, label='\''full'\'', tests=None, extra_argv=['\''-k'\'', '\''(test_bad_geneig)'\'', '\''-n'\'', '\''3'\'', '\''--timeout=1800'\'', '\''--durations=50'\'']); sys.exit(0)' ``` ### Error message ```shell INFO [gw0] darwin -- Python 3.10.4 $PREFIX/bin/python INFO $PREFIX/lib/python3.10/site-packages/scipy/linalg/tests/test_decomp.py:337: in test_bad_geneig INFO self._check_gen_eig(A, B) INFO A = [[1, 0, 0, 0], [0, 1, 0, 0], [0, 0, -8.9375, 0], [0, 0, 0, -8.9375]] INFO B = [[0, 0, 1, 0], [0, 0, 0, 1], [1, 0, 0, -0.5], [0, 1, 0.5, 0]] INFO k = 5 INFO matrices = <function TestEig.test_bad_geneig.<locals>.matrices at 0x7fc4988a2560> INFO self = <scipy.linalg.tests.test_decomp.TestEig object at 0x7fc49970d510> INFO $PREFIX/lib/python3.10/site-packages/scipy/linalg/tests/test_decomp.py:274: in _check_gen_eig INFO assert_allclose(w[perm], wt[permt], INFO E AssertionError: INFO E Not equal to tolerance rtol=1e-07, atol=1e-07 INFO E INFO E array([[ 1. , 0. , 0. , 0. ], INFO E [ 0. , 1. , 0. , 0. ], INFO E [ 0. , 0. , -8.9375, 0. ], INFO E [ 0. , 0. , 0. , -8.9375]]) INFO E array([[ 0. , 0. , 1. , 0. ], INFO E [ 0. , 0. , 0. , 1. ], INFO E [ 1. , 0. , 0. , -0.5], INFO E [ 0. , 1. , 0.5, 0. ]]) INFO E Mismatched elements: 3 / 4 (75%) INFO E Max absolute difference: 6. INFO E Max relative difference: 2.18181818 INFO E x: array([-3.866788e-17-2.75j, -3.866788e-17+2.75j, 0.000000e+00-3.25j, INFO E 0.000000e+00+3.25j]) INFO E y: array([0.-3.25j, 0.-2.75j, 0.+2.75j, 0.+3.25j]) INFO A = array([[ 1. , 0. , 0. , 0. ], INFO [ 0. , 1. , 0. , 0. ], INFO [ 0. , 0. , -8.9375, 0. ], INFO [ 0. , 0. , 0. , -8.9375]]) INFO B = array([[ 0. , 0. , 1. , 0. ], INFO [ 0. , 0. , 0. , 1. ], INFO [ 1. , 0. , 0. , -0.5], INFO [ 0. , 1. , 0.5, 0. ]]) INFO B0 = array([[ 0. , 0. , 1. , 0. ], INFO [ 0. , 0. , 0. , 1. ], INFO [ 1. , 0. , 0. , -0.5], INFO [ 0. , 1. , 0.5, 0. ]]) INFO beta_nonzero = array([ True, True, True, True]) INFO i = 3 INFO length = array([1., 1., 1., 1.]) INFO msg = '\narray([[ 1. , 0. , 0. , 0. ],\n [ 0. , 1. , 0. , 0. ],\n [ 0. , 0. ... , 1. , 0. ],\n [ 0. , 0. , 0. , 1. ],\n [ 1. , 0. , 0. , -0.5],\n [ 0. , 1. , 0.5, 0. ]])' INFO perm = array([3, 2, 1, 0]) INFO permt = array([1, 3, 2, 0]) INFO res = array([[ 5.55111512e-16+0.00000000e+00j, 5.55111512e-16+0.00000000e+00j, INFO 7.98297505e-17-4.44089210e-16j, 7....0000e+00j, 8.88178420e-16+0.00000000e+00j, INFO -3.89981295e-17+0.00000000e+00j, -3.89981295e-17+0.00000000e+00j]]) INFO self = <scipy.linalg.tests.test_decomp.TestEig object at 0x7fc49970d510> INFO val1 = array([[-6.75837818e-01+0.00000000e+00j, -6.75837818e-01+0.00000000e+00j, INFO 9.65826743e-17+6.64534303e-01j, 9....0000e+00j, 1.85855400e+00+0.00000000e+00j, INFO -5.64792532e-16+2.15973648e+00j, -5.64792532e-16-2.15973648e+00j]]) INFO val2 = array([[-6.75837818e-01+0.00000000e+00j, -6.75837818e-01-0.00000000e+00j, INFO 1.67529238e-17+6.64534303e-01j, 1....0000e+00j, 1.85855400e+00+0.00000000e+00j, INFO -5.25794402e-16+2.15973648e+00j, -5.25794402e-16-2.15973648e+00j]]) INFO vr = array([[-6.75837818e-01+0.00000000e+00j, -6.75837818e-01+0.00000000e+00j, INFO 9.65826743e-17+6.64534303e-01j, 9....0000e+00j, -2.07950098e-01+0.00000000e+00j, INFO 6.31935700e-17-2.41648837e-01j, 6.31935700e-17+2.41648837e-01j]]) INFO w = array([ 0.00000000e+00+3.25j, 0.00000000e+00-3.25j, INFO -3.86678798e-17+2.75j, -3.86678798e-17-2.75j]) INFO w_fin = array([ 0.00000000e+00+3.25j, 0.00000000e+00-3.25j, INFO -3.86678798e-17+2.75j, -3.86678798e-17-2.75j]) INFO wh = array([ 0.00000000e+00+3.25j, 0.00000000e+00-3.25j, INFO -3.86678798e-17+2.75j, -3.86678798e-17-2.75j]) INFO wt = array([0.+3.25j, 0.-3.25j, 0.+2.75j, 0.-2.75j]) INFO wt_fin = array([0.+3.25j, 0.-3.25j, 0.+2.75j, 0.-2.75j]) ``` ### SciPy/NumPy/Python version information 1.9.1 1.23.1 sys.version_info(major=3, minor=10, micro=4, releaselevel='final', serial=0)
1.0
BUG: osx-64 scipy 1.9.1 test_bad_geneig numerical error - ### Describe your issue. The test test_bad_geneig is failing on scipy 1.9.1 osx-64 with openblas 0.3.21 for py38, py39 and py310. It looks like a permutation issue. ### Reproducing Code Example ```python This is reproducible on osx-64 with openblas 0.3.21 for py38, py39 and py310. python -c 'import scipy, sys; scipy.test(verbose=2, label='\''full'\'', tests=None, extra_argv=['\''-k'\'', '\''(test_bad_geneig)'\'', '\''-n'\'', '\''3'\'', '\''--timeout=1800'\'', '\''--durations=50'\'']); sys.exit(0)' ``` ### Error message ```shell INFO [gw0] darwin -- Python 3.10.4 $PREFIX/bin/python INFO $PREFIX/lib/python3.10/site-packages/scipy/linalg/tests/test_decomp.py:337: in test_bad_geneig INFO self._check_gen_eig(A, B) INFO A = [[1, 0, 0, 0], [0, 1, 0, 0], [0, 0, -8.9375, 0], [0, 0, 0, -8.9375]] INFO B = [[0, 0, 1, 0], [0, 0, 0, 1], [1, 0, 0, -0.5], [0, 1, 0.5, 0]] INFO k = 5 INFO matrices = <function TestEig.test_bad_geneig.<locals>.matrices at 0x7fc4988a2560> INFO self = <scipy.linalg.tests.test_decomp.TestEig object at 0x7fc49970d510> INFO $PREFIX/lib/python3.10/site-packages/scipy/linalg/tests/test_decomp.py:274: in _check_gen_eig INFO assert_allclose(w[perm], wt[permt], INFO E AssertionError: INFO E Not equal to tolerance rtol=1e-07, atol=1e-07 INFO E INFO E array([[ 1. , 0. , 0. , 0. ], INFO E [ 0. , 1. , 0. , 0. ], INFO E [ 0. , 0. , -8.9375, 0. ], INFO E [ 0. , 0. , 0. , -8.9375]]) INFO E array([[ 0. , 0. , 1. , 0. ], INFO E [ 0. , 0. , 0. , 1. ], INFO E [ 1. , 0. , 0. , -0.5], INFO E [ 0. , 1. , 0.5, 0. ]]) INFO E Mismatched elements: 3 / 4 (75%) INFO E Max absolute difference: 6. INFO E Max relative difference: 2.18181818 INFO E x: array([-3.866788e-17-2.75j, -3.866788e-17+2.75j, 0.000000e+00-3.25j, INFO E 0.000000e+00+3.25j]) INFO E y: array([0.-3.25j, 0.-2.75j, 0.+2.75j, 0.+3.25j]) INFO A = array([[ 1. , 0. , 0. , 0. ], INFO [ 0. , 1. , 0. , 0. ], INFO [ 0. , 0. , -8.9375, 0. ], INFO [ 0. , 0. , 0. , -8.9375]]) INFO B = array([[ 0. , 0. , 1. , 0. ], INFO [ 0. , 0. , 0. , 1. ], INFO [ 1. , 0. , 0. , -0.5], INFO [ 0. , 1. , 0.5, 0. ]]) INFO B0 = array([[ 0. , 0. , 1. , 0. ], INFO [ 0. , 0. , 0. , 1. ], INFO [ 1. , 0. , 0. , -0.5], INFO [ 0. , 1. , 0.5, 0. ]]) INFO beta_nonzero = array([ True, True, True, True]) INFO i = 3 INFO length = array([1., 1., 1., 1.]) INFO msg = '\narray([[ 1. , 0. , 0. , 0. ],\n [ 0. , 1. , 0. , 0. ],\n [ 0. , 0. ... , 1. , 0. ],\n [ 0. , 0. , 0. , 1. ],\n [ 1. , 0. , 0. , -0.5],\n [ 0. , 1. , 0.5, 0. ]])' INFO perm = array([3, 2, 1, 0]) INFO permt = array([1, 3, 2, 0]) INFO res = array([[ 5.55111512e-16+0.00000000e+00j, 5.55111512e-16+0.00000000e+00j, INFO 7.98297505e-17-4.44089210e-16j, 7....0000e+00j, 8.88178420e-16+0.00000000e+00j, INFO -3.89981295e-17+0.00000000e+00j, -3.89981295e-17+0.00000000e+00j]]) INFO self = <scipy.linalg.tests.test_decomp.TestEig object at 0x7fc49970d510> INFO val1 = array([[-6.75837818e-01+0.00000000e+00j, -6.75837818e-01+0.00000000e+00j, INFO 9.65826743e-17+6.64534303e-01j, 9....0000e+00j, 1.85855400e+00+0.00000000e+00j, INFO -5.64792532e-16+2.15973648e+00j, -5.64792532e-16-2.15973648e+00j]]) INFO val2 = array([[-6.75837818e-01+0.00000000e+00j, -6.75837818e-01-0.00000000e+00j, INFO 1.67529238e-17+6.64534303e-01j, 1....0000e+00j, 1.85855400e+00+0.00000000e+00j, INFO -5.25794402e-16+2.15973648e+00j, -5.25794402e-16-2.15973648e+00j]]) INFO vr = array([[-6.75837818e-01+0.00000000e+00j, -6.75837818e-01+0.00000000e+00j, INFO 9.65826743e-17+6.64534303e-01j, 9....0000e+00j, -2.07950098e-01+0.00000000e+00j, INFO 6.31935700e-17-2.41648837e-01j, 6.31935700e-17+2.41648837e-01j]]) INFO w = array([ 0.00000000e+00+3.25j, 0.00000000e+00-3.25j, INFO -3.86678798e-17+2.75j, -3.86678798e-17-2.75j]) INFO w_fin = array([ 0.00000000e+00+3.25j, 0.00000000e+00-3.25j, INFO -3.86678798e-17+2.75j, -3.86678798e-17-2.75j]) INFO wh = array([ 0.00000000e+00+3.25j, 0.00000000e+00-3.25j, INFO -3.86678798e-17+2.75j, -3.86678798e-17-2.75j]) INFO wt = array([0.+3.25j, 0.-3.25j, 0.+2.75j, 0.-2.75j]) INFO wt_fin = array([0.+3.25j, 0.-3.25j, 0.+2.75j, 0.-2.75j]) ``` ### SciPy/NumPy/Python version information 1.9.1 1.23.1 sys.version_info(major=3, minor=10, micro=4, releaselevel='final', serial=0)
defect
bug osx scipy test bad geneig numerical error describe your issue the test test bad geneig is failing on scipy osx with openblas for and it looks like a permutation issue reproducing code example python this is reproducible on osx with openblas for and python c import scipy sys scipy test verbose label full tests none extra argv sys exit error message shell info darwin python prefix bin python info prefix lib site packages scipy linalg tests test decomp py in test bad geneig info self check gen eig a b info a info b info k info matrices matrices at info self info prefix lib site packages scipy linalg tests test decomp py in check gen eig info assert allclose w wt info e assertionerror info e not equal to tolerance rtol atol info e info e array info e info e info e info e array info e info e info e info e mismatched elements info e max absolute difference info e max relative difference info e x array info e info e y array info a array info info info info b array info info info info array info info info info beta nonzero array info i info length array info msg narray n n n n n info perm array info permt array info res array info info info self info array info info info array info info info vr array info info info w array info info w fin array info info wh array info info wt array info wt fin array scipy numpy python version information sys version info major minor micro releaselevel final serial
1
184,990
14,292,030,365
IssuesEvent
2020-11-24 00:02:54
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
zvol_ENOSPC test suite failure
Component: Test Suite Status: Stale
<!-- Thank you for reporting an issue. *IMPORTANT* - Please search our issue tracker *before* making a new issue. If you cannot find a similar issue, then create a new issue. https://github.com/zfsonlinux/zfs/issues *IMPORTANT* - This issue tracker is for *bugs* and *issues* only. Please search the wiki and the mailing list archives before asking questions on the mailing list. https://github.com/zfsonlinux/zfs/wiki/Mailing-Lists Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | TOSS/RHEL Distribution Version | 6.7 Linux Kernel | 2.6.32-573.18.1.1chaos.ch5.4.x86_64 Architecture | x86_64 ZFS Version | zfsonlinux/zfs@66eead5 SPL Version | zfsonlinux/spl@9704820 <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing The zvol_ENOSPC test suite is repeatedly failing in the automated tests performed on PRs by the buildbot. Specifically, zvol_ENOSPC_001_pos times out and is killed, and the cleanup step fails. Subsequent tests often fail as well. The failure does not appear to depend on the nature of the patch. ``` Test: /usr/share/zfs/zfs-tests/tests/functional/zvol/zvol_ENOSPC/setup (run as root) [00:02] [PASS] Test: /usr/share/zfs/zfs-tests/tests/functional/zvol/zvol_ENOSPC/zvol_ENOSPC_001_pos (run as root) [10:00] [KILLED] Test: /usr/share/zfs/zfs-tests/tests/functional/zvol/zvol_ENOSPC/cleanup (run as root) [00:00] [FAIL] ``` I've given my VM's specific distro/kernel/etc above but do not know if this issue is specific to this environment. ### Describe how to reproduce the problem I believe I've reproduced the problem in a VM on my desktop: * truncate -s 64M /tmp/file.1 * sudo cmd/zpool/zpool create frogger /tmp/file.1 * sudo cmd/zfs/zfs create -V 8M frogger/myvol * sudo strace -T dd if=/dev/urandom of=/dev/zd0 The dd does eventually get ENOSPC in response to a write() call. However, the close() call for output file (the zvol's device) takes a very long time to return - 100s in the example I captured. ``` read(0, "\203\242\247\30\216Y\355Tp\323\371\17\300\30&q'\210\16\314\301\360i0\242\356*O\305\rT5"..., 512) = 512 <0.000140> write(1, "\203\242\247\30\216Y\355Tp\323\371\17\300\30&q'\210\16\314\301\360i0\242\356*O\305\rT5"..., 512) = -1 ENOSPC (No space left on device) <0.000008> ``` and then ``` close(0) = 0 <0.000027> close(1) = 0 <100.729140> ``` ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` --> I saw two stacks while waiting on the close() call to return. I don't know if either is a useful clue: ``` [faaland1@locks branch:b_refcount_t zfs] $sudo cat /proc/23312/stack [<ffffffff81127acd>] sync_page+0x3d/0x50 [<ffffffff81127a67>] __lock_page+0x67/0x70 [<ffffffff8113cebd>] write_cache_pages+0x38d/0x4c0 [<ffffffff8113d014>] generic_writepages+0x24/0x30 [<ffffffff8113d041>] do_writepages+0x21/0x40 [<ffffffff8112828b>] __filemap_fdatawrite_range+0x5b/0x60 [<ffffffff811287ef>] filemap_fdatawrite+0x1f/0x30 [<ffffffff81128835>] filemap_write_and_wait+0x35/0x60 [<ffffffff811d0f94>] __sync_blockdev+0x24/0x50 [<ffffffff811d0fd3>] sync_blockdev+0x13/0x20 [<ffffffff811d1168>] __blkdev_put+0x188/0x1c0 [<ffffffff811d11b0>] blkdev_put+0x10/0x20 [<ffffffff811d11f3>] blkdev_close+0x33/0x60 [<ffffffff81194858>] __fput+0x108/0x230 [<ffffffff811949a5>] fput+0x25/0x30 [<ffffffff8118fa6d>] filp_close+0x5d/0x90 [<ffffffff8118fb45>] sys_close+0xa5/0x100 [<ffffffff8100b328>] tracesys+0xd9/0xde [<ffffffffffffffff>] 0xffffffffffffffff ``` and ``` [faaland1@locks branch:b_refcount_t zfs] $sudo cat /proc/23312/stack [<ffffffffa028523d>] cv_wait_common+0x15d/0x2b0 [spl] [<ffffffffa02853e5>] __cv_wait+0x15/0x20 [spl] [<ffffffffa045a063>] txg_wait_open+0xf3/0x1d0 [zfs] [<ffffffffa03f9b3c>] dmu_tx_wait+0x4ac/0x510 [zfs] [<ffffffffa03f9c8c>] dmu_tx_assign+0xec/0x7a0 [zfs] [<ffffffffa04d972c>] zvol_request+0x4ec/0xa20 [zfs] [<ffffffff81277062>] generic_make_request+0x2b2/0x660 [<ffffffff81277480>] submit_bio+0x70/0x120 [<ffffffff811c927d>] submit_bh+0x11d/0x1f0 [<ffffffff811cbb38>] __block_write_full_page+0x1d8/0x340 [<ffffffff811cbd80>] block_write_full_page_endio+0xe0/0x120 [<ffffffff811cbdd5>] block_write_full_page+0x15/0x20 [<ffffffff811d0c08>] blkdev_writepage+0x18/0x20 [<ffffffff8113ba27>] __writepage+0x17/0x40 [<ffffffff8113cd48>] write_cache_pages+0x218/0x4c0 [<ffffffff8113d014>] generic_writepages+0x24/0x30 [<ffffffff8113d041>] do_writepages+0x21/0x40 [<ffffffff8112828b>] __filemap_fdatawrite_range+0x5b/0x60 [<ffffffff811287ef>] filemap_fdatawrite+0x1f/0x30 [<ffffffff81128835>] filemap_write_and_wait+0x35/0x60 [<ffffffff811d0f94>] __sync_blockdev+0x24/0x50 [<ffffffff811d0fd3>] sync_blockdev+0x13/0x20 [<ffffffff811d1168>] __blkdev_put+0x188/0x1c0 [<ffffffff811d11b0>] blkdev_put+0x10/0x20 [<ffffffff811d11f3>] blkdev_close+0x33/0x60 [<ffffffff81194858>] __fput+0x108/0x230 [<ffffffff811949a5>] fput+0x25/0x30 [<ffffffff8118fa6d>] filp_close+0x5d/0x90 [<ffffffff8118fb45>] sys_close+0xa5/0x100 [<ffffffff8100b328>] tracesys+0xd9/0xde [<ffffffffffffffff>] 0xffffffffffffffff ```
1.0
zvol_ENOSPC test suite failure - <!-- Thank you for reporting an issue. *IMPORTANT* - Please search our issue tracker *before* making a new issue. If you cannot find a similar issue, then create a new issue. https://github.com/zfsonlinux/zfs/issues *IMPORTANT* - This issue tracker is for *bugs* and *issues* only. Please search the wiki and the mailing list archives before asking questions on the mailing list. https://github.com/zfsonlinux/zfs/wiki/Mailing-Lists Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | TOSS/RHEL Distribution Version | 6.7 Linux Kernel | 2.6.32-573.18.1.1chaos.ch5.4.x86_64 Architecture | x86_64 ZFS Version | zfsonlinux/zfs@66eead5 SPL Version | zfsonlinux/spl@9704820 <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing The zvol_ENOSPC test suite is repeatedly failing in the automated tests performed on PRs by the buildbot. Specifically, zvol_ENOSPC_001_pos times out and is killed, and the cleanup step fails. Subsequent tests often fail as well. The failure does not appear to depend on the nature of the patch. ``` Test: /usr/share/zfs/zfs-tests/tests/functional/zvol/zvol_ENOSPC/setup (run as root) [00:02] [PASS] Test: /usr/share/zfs/zfs-tests/tests/functional/zvol/zvol_ENOSPC/zvol_ENOSPC_001_pos (run as root) [10:00] [KILLED] Test: /usr/share/zfs/zfs-tests/tests/functional/zvol/zvol_ENOSPC/cleanup (run as root) [00:00] [FAIL] ``` I've given my VM's specific distro/kernel/etc above but do not know if this issue is specific to this environment. ### Describe how to reproduce the problem I believe I've reproduced the problem in a VM on my desktop: * truncate -s 64M /tmp/file.1 * sudo cmd/zpool/zpool create frogger /tmp/file.1 * sudo cmd/zfs/zfs create -V 8M frogger/myvol * sudo strace -T dd if=/dev/urandom of=/dev/zd0 The dd does eventually get ENOSPC in response to a write() call. However, the close() call for output file (the zvol's device) takes a very long time to return - 100s in the example I captured. ``` read(0, "\203\242\247\30\216Y\355Tp\323\371\17\300\30&q'\210\16\314\301\360i0\242\356*O\305\rT5"..., 512) = 512 <0.000140> write(1, "\203\242\247\30\216Y\355Tp\323\371\17\300\30&q'\210\16\314\301\360i0\242\356*O\305\rT5"..., 512) = -1 ENOSPC (No space left on device) <0.000008> ``` and then ``` close(0) = 0 <0.000027> close(1) = 0 <100.729140> ``` ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` --> I saw two stacks while waiting on the close() call to return. I don't know if either is a useful clue: ``` [faaland1@locks branch:b_refcount_t zfs] $sudo cat /proc/23312/stack [<ffffffff81127acd>] sync_page+0x3d/0x50 [<ffffffff81127a67>] __lock_page+0x67/0x70 [<ffffffff8113cebd>] write_cache_pages+0x38d/0x4c0 [<ffffffff8113d014>] generic_writepages+0x24/0x30 [<ffffffff8113d041>] do_writepages+0x21/0x40 [<ffffffff8112828b>] __filemap_fdatawrite_range+0x5b/0x60 [<ffffffff811287ef>] filemap_fdatawrite+0x1f/0x30 [<ffffffff81128835>] filemap_write_and_wait+0x35/0x60 [<ffffffff811d0f94>] __sync_blockdev+0x24/0x50 [<ffffffff811d0fd3>] sync_blockdev+0x13/0x20 [<ffffffff811d1168>] __blkdev_put+0x188/0x1c0 [<ffffffff811d11b0>] blkdev_put+0x10/0x20 [<ffffffff811d11f3>] blkdev_close+0x33/0x60 [<ffffffff81194858>] __fput+0x108/0x230 [<ffffffff811949a5>] fput+0x25/0x30 [<ffffffff8118fa6d>] filp_close+0x5d/0x90 [<ffffffff8118fb45>] sys_close+0xa5/0x100 [<ffffffff8100b328>] tracesys+0xd9/0xde [<ffffffffffffffff>] 0xffffffffffffffff ``` and ``` [faaland1@locks branch:b_refcount_t zfs] $sudo cat /proc/23312/stack [<ffffffffa028523d>] cv_wait_common+0x15d/0x2b0 [spl] [<ffffffffa02853e5>] __cv_wait+0x15/0x20 [spl] [<ffffffffa045a063>] txg_wait_open+0xf3/0x1d0 [zfs] [<ffffffffa03f9b3c>] dmu_tx_wait+0x4ac/0x510 [zfs] [<ffffffffa03f9c8c>] dmu_tx_assign+0xec/0x7a0 [zfs] [<ffffffffa04d972c>] zvol_request+0x4ec/0xa20 [zfs] [<ffffffff81277062>] generic_make_request+0x2b2/0x660 [<ffffffff81277480>] submit_bio+0x70/0x120 [<ffffffff811c927d>] submit_bh+0x11d/0x1f0 [<ffffffff811cbb38>] __block_write_full_page+0x1d8/0x340 [<ffffffff811cbd80>] block_write_full_page_endio+0xe0/0x120 [<ffffffff811cbdd5>] block_write_full_page+0x15/0x20 [<ffffffff811d0c08>] blkdev_writepage+0x18/0x20 [<ffffffff8113ba27>] __writepage+0x17/0x40 [<ffffffff8113cd48>] write_cache_pages+0x218/0x4c0 [<ffffffff8113d014>] generic_writepages+0x24/0x30 [<ffffffff8113d041>] do_writepages+0x21/0x40 [<ffffffff8112828b>] __filemap_fdatawrite_range+0x5b/0x60 [<ffffffff811287ef>] filemap_fdatawrite+0x1f/0x30 [<ffffffff81128835>] filemap_write_and_wait+0x35/0x60 [<ffffffff811d0f94>] __sync_blockdev+0x24/0x50 [<ffffffff811d0fd3>] sync_blockdev+0x13/0x20 [<ffffffff811d1168>] __blkdev_put+0x188/0x1c0 [<ffffffff811d11b0>] blkdev_put+0x10/0x20 [<ffffffff811d11f3>] blkdev_close+0x33/0x60 [<ffffffff81194858>] __fput+0x108/0x230 [<ffffffff811949a5>] fput+0x25/0x30 [<ffffffff8118fa6d>] filp_close+0x5d/0x90 [<ffffffff8118fb45>] sys_close+0xa5/0x100 [<ffffffff8100b328>] tracesys+0xd9/0xde [<ffffffffffffffff>] 0xffffffffffffffff ```
non_defect
zvol enospc test suite failure thank you for reporting an issue important please search our issue tracker before making a new issue if you cannot find a similar issue then create a new issue important this issue tracker is for bugs and issues only please search the wiki and the mailing list archives before asking questions on the mailing list please fill in as much of the template as possible system information type version name distribution name toss rhel distribution version linux kernel architecture zfs version zfsonlinux zfs spl version zfsonlinux spl commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing the zvol enospc test suite is repeatedly failing in the automated tests performed on prs by the buildbot specifically zvol enospc pos times out and is killed and the cleanup step fails subsequent tests often fail as well the failure does not appear to depend on the nature of the patch test usr share zfs zfs tests tests functional zvol zvol enospc setup run as root test usr share zfs zfs tests tests functional zvol zvol enospc zvol enospc pos run as root test usr share zfs zfs tests tests functional zvol zvol enospc cleanup run as root i ve given my vm s specific distro kernel etc above but do not know if this issue is specific to this environment describe how to reproduce the problem i believe i ve reproduced the problem in a vm on my desktop truncate s tmp file sudo cmd zpool zpool create frogger tmp file sudo cmd zfs zfs create v frogger myvol sudo strace t dd if dev urandom of dev the dd does eventually get enospc in response to a write call however the close call for output file the zvol s device takes a very long time to return in the example i captured read q o write q o enospc no space left on device and then close close include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with i saw two stacks while waiting on the close call to return i don t know if either is a useful clue sudo cat proc stack sync page lock page write cache pages generic writepages do writepages filemap fdatawrite range filemap fdatawrite filemap write and wait sync blockdev sync blockdev blkdev put blkdev put blkdev close fput fput filp close sys close tracesys and sudo cat proc stack cv wait common cv wait txg wait open dmu tx wait dmu tx assign zvol request generic make request submit bio submit bh block write full page block write full page endio block write full page blkdev writepage writepage write cache pages generic writepages do writepages filemap fdatawrite range filemap fdatawrite filemap write and wait sync blockdev sync blockdev blkdev put blkdev put blkdev close fput fput filp close sys close tracesys
0
41,664
10,562,908,787
IssuesEvent
2019-10-04 19:32:31
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
InvocationFuture race: race on exception
Internal breaking change Module: Invocation System Source: Internal Team: Core To Triage Type: Defect
When an InvocationFuture processes a thrown exception e.g one from the operation.run, it will rewrite the stacktrace so that both local and remote stacktraces are integrated. ``` private Object resolveApplicationResponse(Object unresolvedResponse) { ... if (response instanceof Throwable) { Throwable throwable = ((Throwable) response); if (invocation.remote) { fixRemoteStackTrace((Throwable) response, currentThread().getStackTrace()); } return throwable; } return response; } ``` The problem is that this exception-instance can be shared between multiple threads since the future can be shared between multiple threads. So you have multiple threads modifying a stacktrace of the same exception instance. Solution: clone the exception first. Edit: for a remote thrown exception this probably isn't an issue since each thread will get a Data and deserialize that data. Not completely sure about this. But for local exception is could certainly be an issue.
1.0
InvocationFuture race: race on exception - When an InvocationFuture processes a thrown exception e.g one from the operation.run, it will rewrite the stacktrace so that both local and remote stacktraces are integrated. ``` private Object resolveApplicationResponse(Object unresolvedResponse) { ... if (response instanceof Throwable) { Throwable throwable = ((Throwable) response); if (invocation.remote) { fixRemoteStackTrace((Throwable) response, currentThread().getStackTrace()); } return throwable; } return response; } ``` The problem is that this exception-instance can be shared between multiple threads since the future can be shared between multiple threads. So you have multiple threads modifying a stacktrace of the same exception instance. Solution: clone the exception first. Edit: for a remote thrown exception this probably isn't an issue since each thread will get a Data and deserialize that data. Not completely sure about this. But for local exception is could certainly be an issue.
defect
invocationfuture race race on exception when an invocationfuture processes a thrown exception e g one from the operation run it will rewrite the stacktrace so that both local and remote stacktraces are integrated private object resolveapplicationresponse object unresolvedresponse if response instanceof throwable throwable throwable throwable response if invocation remote fixremotestacktrace throwable response currentthread getstacktrace return throwable return response the problem is that this exception instance can be shared between multiple threads since the future can be shared between multiple threads so you have multiple threads modifying a stacktrace of the same exception instance solution clone the exception first edit for a remote thrown exception this probably isn t an issue since each thread will get a data and deserialize that data not completely sure about this but for local exception is could certainly be an issue
1
397,105
27,149,091,134
IssuesEvent
2023-02-16 22:50:03
TravelTimN/oregon-trail-python
https://api.github.com/repos/TravelTimN/oregon-trail-python
opened
River Crossing anomalies
bug documentation
Fording: if D < 2.5: if BT = 2 (rocky): 16% of overturning This isn't possible. Only the Green and Snake rivers have BT = 2 (rocky). The Green has a min_depth of 20ft. The Snake has a min_depth of 6ft. --------------------------------------- Fording: NOTE: Most of these risks are reduced by 80% if you hire a Native American guide. You can only hire help at the Snake River Crossing. However, you cannot select "hire help" and "ford" simultaneously.
1.0
River Crossing anomalies - Fording: if D < 2.5: if BT = 2 (rocky): 16% of overturning This isn't possible. Only the Green and Snake rivers have BT = 2 (rocky). The Green has a min_depth of 20ft. The Snake has a min_depth of 6ft. --------------------------------------- Fording: NOTE: Most of these risks are reduced by 80% if you hire a Native American guide. You can only hire help at the Snake River Crossing. However, you cannot select "hire help" and "ford" simultaneously.
non_defect
river crossing anomalies fording if d if bt rocky of overturning this isn t possible only the green and snake rivers have bt rocky the green has a min depth of the snake has a min depth of fording note most of these risks are reduced by if you hire a native american guide you can only hire help at the snake river crossing however you cannot select hire help and ford simultaneously
0
2,224
2,603,991,349
IssuesEvent
2015-02-24 19:06:42
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳哪里看生殖疱疹好
auto-migrated Priority-Medium Type-Defect
``` 沈阳哪里看生殖疱疹好〓沈陽軍區政治部醫院性病〓TEL:024-3 1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。� ��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:31
1.0
沈阳哪里看生殖疱疹好 - ``` 沈阳哪里看生殖疱疹好〓沈陽軍區政治部醫院性病〓TEL:024-3 1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。� ��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:31
defect
沈阳哪里看生殖疱疹好 沈阳哪里看生殖疱疹好〓沈陽軍區政治部醫院性病〓tel: 〓 , 。� �� 。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 original issue reported on code google com by gmail com on jun at
1
60,300
17,023,390,940
IssuesEvent
2021-07-03 01:47:05
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
osmosis does not implement maxDownloadCount in 0.6 code
Component: osmosis Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 8.58am, Saturday, 25th April 2009]** It looks like the 0.6 changeset download code does not implement the maxDownloadCount feature. The 0.6 code was branched in r7941, but the download limit was added to 0.5 until r7949. I'll attach a patch which looks like it will fix it.
1.0
osmosis does not implement maxDownloadCount in 0.6 code - **[Submitted to the original trac issue database at 8.58am, Saturday, 25th April 2009]** It looks like the 0.6 changeset download code does not implement the maxDownloadCount feature. The 0.6 code was branched in r7941, but the download limit was added to 0.5 until r7949. I'll attach a patch which looks like it will fix it.
defect
osmosis does not implement maxdownloadcount in code it looks like the changeset download code does not implement the maxdownloadcount feature the code was branched in but the download limit was added to until i ll attach a patch which looks like it will fix it
1
692,816
23,749,979,006
IssuesEvent
2022-08-31 19:37:49
harvard-lil/perma
https://api.github.com/repos/harvard-lil/perma
opened
Case-insensitive email addresses
bug priority-high
As reported by @clare-stanton: Some of our users are running into issues because email addresses are considered case sensitive during sign-up / sign-in. - [ ] Figure out if the entire email address, or just the part before `@` should be case insensitive - [ ] See how many existing users would be impacted by proposed fix - [ ] Apply the fix and remediate edge cases --- **Resources:** - @lizadaly's work and thoughts on fixing a similar problem on H2O: https://github.com/harvard-lil/h2o/pull/1592
1.0
Case-insensitive email addresses - As reported by @clare-stanton: Some of our users are running into issues because email addresses are considered case sensitive during sign-up / sign-in. - [ ] Figure out if the entire email address, or just the part before `@` should be case insensitive - [ ] See how many existing users would be impacted by proposed fix - [ ] Apply the fix and remediate edge cases --- **Resources:** - @lizadaly's work and thoughts on fixing a similar problem on H2O: https://github.com/harvard-lil/h2o/pull/1592
non_defect
case insensitive email addresses as reported by clare stanton some of our users are running into issues because email addresses are considered case sensitive during sign up sign in figure out if the entire email address or just the part before should be case insensitive see how many existing users would be impacted by proposed fix apply the fix and remediate edge cases resources lizadaly s work and thoughts on fixing a similar problem on
0
649,288
21,262,969,572
IssuesEvent
2022-04-13 07:09:00
bounswe/bounswe2022group8
https://api.github.com/repos/bounswe/bounswe2022group8
opened
Milestone I Report
Effort: High Priority: High Status: in progress milestone
**To Do**: As discussed in [Meeting-10](https://github.com/bounswe/bounswe2022group8/wiki/Week6---Meeting-%2310-(06.04.2022)-(PS)), it's time to prepare our first milestone report that gathers the major deliverables related to the requirements, design and plan of our project. According to our meeting, task allocation is as follows: * Executive Summary (Project Description and Project Status): @serdarakol * List and Status of Deliverables & Evaluation of the status of deliverables: @KarahanS * A summary of work done by each member: **everyone** (refer to related [wikipage](https://github.com/bounswe/bounswe2022group8/wiki/Work-Summary-by-Each-Member)) * Evaluation of tools and processes that you have used to manage your team project: @kostanya * RAM (responsibility assignment matrix): @simsekhebun, @ooodogodogodogo, @serdarakol, @mumcusena * Project Plan: @sinemKocoglu, @kostanya, @KarahanS, @BElifb * Combining deliverables (project repository, requirements, software design documents, project plan, RAM, communication plan etc.): @mustafa-cihan We will be working in Google Docs to prepare the necessary items. Refer to our Discord channel for the link. Please refer to this issue in the related sub-issues. Deadline: 15.04.2022 @19.00
1.0
Milestone I Report - **To Do**: As discussed in [Meeting-10](https://github.com/bounswe/bounswe2022group8/wiki/Week6---Meeting-%2310-(06.04.2022)-(PS)), it's time to prepare our first milestone report that gathers the major deliverables related to the requirements, design and plan of our project. According to our meeting, task allocation is as follows: * Executive Summary (Project Description and Project Status): @serdarakol * List and Status of Deliverables & Evaluation of the status of deliverables: @KarahanS * A summary of work done by each member: **everyone** (refer to related [wikipage](https://github.com/bounswe/bounswe2022group8/wiki/Work-Summary-by-Each-Member)) * Evaluation of tools and processes that you have used to manage your team project: @kostanya * RAM (responsibility assignment matrix): @simsekhebun, @ooodogodogodogo, @serdarakol, @mumcusena * Project Plan: @sinemKocoglu, @kostanya, @KarahanS, @BElifb * Combining deliverables (project repository, requirements, software design documents, project plan, RAM, communication plan etc.): @mustafa-cihan We will be working in Google Docs to prepare the necessary items. Refer to our Discord channel for the link. Please refer to this issue in the related sub-issues. Deadline: 15.04.2022 @19.00
non_defect
milestone i report to do as discussed in it s time to prepare our first milestone report that gathers the major deliverables related to the requirements design and plan of our project according to our meeting task allocation is as follows executive summary project description and project status serdarakol list and status of deliverables evaluation of the status of deliverables karahans a summary of work done by each member everyone refer to related evaluation of tools and processes that you have used to manage your team project kostanya ram responsibility assignment matrix simsekhebun ooodogodogodogo serdarakol mumcusena project plan sinemkocoglu kostanya karahans belifb combining deliverables project repository requirements software design documents project plan ram communication plan etc mustafa cihan we will be working in google docs to prepare the necessary items refer to our discord channel for the link please refer to this issue in the related sub issues deadline
0
77,230
26,865,354,822
IssuesEvent
2023-02-03 22:52:32
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
closed
CCE1.1 - total en USD y tipo de cambio dentro del complemento incorrectos
bug defect
Si la moneda del CFDI es distinto a USD, entonces estos atributos deben calcularse, con base a lo capturado en el catálogo de monedas.. Este atributo dentro del complemento, siempre será en USD. Si la moneda original es pesos, entonces hay que hacer la conversión. Si la moneda no es ni pesos, ni USD... entonces hay que convertir a pesos, y luego a USD (supongo este último escenario sería poco común)...
1.0
CCE1.1 - total en USD y tipo de cambio dentro del complemento incorrectos - Si la moneda del CFDI es distinto a USD, entonces estos atributos deben calcularse, con base a lo capturado en el catálogo de monedas.. Este atributo dentro del complemento, siempre será en USD. Si la moneda original es pesos, entonces hay que hacer la conversión. Si la moneda no es ni pesos, ni USD... entonces hay que convertir a pesos, y luego a USD (supongo este último escenario sería poco común)...
defect
total en usd y tipo de cambio dentro del complemento incorrectos si la moneda del cfdi es distinto a usd entonces estos atributos deben calcularse con base a lo capturado en el catálogo de monedas este atributo dentro del complemento siempre será en usd si la moneda original es pesos entonces hay que hacer la conversión si la moneda no es ni pesos ni usd entonces hay que convertir a pesos y luego a usd supongo este último escenario sería poco común
1
163,843
20,364,218,036
IssuesEvent
2022-02-21 02:21:10
turkdevops/framevuerk
https://api.github.com/repos/turkdevops/framevuerk
opened
html-webpack-plugin-3.2.0.tgz: 4 vulnerabilities (highest severity is: 9.1)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>html-webpack-plugin-3.2.0.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2019-10744](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | lodash-4.17.11.tgz | Transitive | 4.0.0-alpha | ❌ | | [CVE-2020-8203](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.4 | lodash-4.17.11.tgz | Transitive | 4.0.0-alpha | ❌ | | [CVE-2021-23337](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.2 | lodash-4.17.11.tgz | Transitive | 4.0.0-alpha | ❌ | | [CVE-2020-28500](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | lodash-4.17.11.tgz | Transitive | 4.0.0-alpha | ❌ | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-10744</summary> ### Vulnerable Library - <b>lodash-4.17.11.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - html-webpack-plugin-3.2.0.tgz (Root Library) - :x: **lodash-4.17.11.tgz** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> <p></p> ### Vulnerability Details <p> Versions of lodash lower than 4.17.12 are vulnerable to Prototype Pollution. The function defaultsDeep could be tricked into adding or modifying properties of Object.prototype using a constructor payload. <p>Publish Date: 2019-07-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744>CVE-2019-10744</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-jf85-cpcp-j695">https://github.com/advisories/GHSA-jf85-cpcp-j695</a></p> <p>Release Date: 2019-07-26</p> <p>Fix Resolution (lodash): 4.17.12</p> <p>Direct dependency fix Resolution (html-webpack-plugin): 4.0.0-alpha</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-8203</summary> ### Vulnerable Library - <b>lodash-4.17.11.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - html-webpack-plugin-3.2.0.tgz (Root Library) - :x: **lodash-4.17.11.tgz** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> <p></p> ### Vulnerability Details <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.4</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution (lodash): 4.17.19</p> <p>Direct dependency fix Resolution (html-webpack-plugin): 4.0.0-alpha</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23337</summary> ### Vulnerable Library - <b>lodash-4.17.11.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - html-webpack-plugin-3.2.0.tgz (Root Library) - :x: **lodash-4.17.11.tgz** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> <p></p> ### Vulnerability Details <p> Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.2</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution (lodash): 4.17.21</p> <p>Direct dependency fix Resolution (html-webpack-plugin): 4.0.0-alpha</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-28500</summary> ### Vulnerable Library - <b>lodash-4.17.11.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - html-webpack-plugin-3.2.0.tgz (Root Library) - :x: **lodash-4.17.11.tgz** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> <p></p> ### Vulnerability Details <p> Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution (lodash): 4.17.21</p> <p>Direct dependency fix Resolution (html-webpack-plugin): 4.0.0-alpha</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"html-webpack-plugin","packageVersion":"3.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"html-webpack-plugin:3.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0.0-alpha","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2019-10744","vulnerabilityDetails":"Versions of lodash lower than 4.17.12 are vulnerable to Prototype Pollution. The function defaultsDeep could be tricked into adding or modifying properties of Object.prototype using a constructor payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744","cvss3Severity":"high","cvss3Score":"9.1","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"html-webpack-plugin","packageVersion":"3.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"html-webpack-plugin:3.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0.0-alpha","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-8203","vulnerabilityDetails":"Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"html-webpack-plugin","packageVersion":"3.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"html-webpack-plugin:3.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0.0-alpha","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2021-23337","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337","cvss3Severity":"high","cvss3Score":"7.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"html-webpack-plugin","packageVersion":"3.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"html-webpack-plugin:3.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0.0-alpha","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-28500","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.\n WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> -->
True
html-webpack-plugin-3.2.0.tgz: 4 vulnerabilities (highest severity is: 9.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>html-webpack-plugin-3.2.0.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2019-10744](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | lodash-4.17.11.tgz | Transitive | 4.0.0-alpha | ❌ | | [CVE-2020-8203](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.4 | lodash-4.17.11.tgz | Transitive | 4.0.0-alpha | ❌ | | [CVE-2021-23337](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.2 | lodash-4.17.11.tgz | Transitive | 4.0.0-alpha | ❌ | | [CVE-2020-28500](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | lodash-4.17.11.tgz | Transitive | 4.0.0-alpha | ❌ | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-10744</summary> ### Vulnerable Library - <b>lodash-4.17.11.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - html-webpack-plugin-3.2.0.tgz (Root Library) - :x: **lodash-4.17.11.tgz** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> <p></p> ### Vulnerability Details <p> Versions of lodash lower than 4.17.12 are vulnerable to Prototype Pollution. The function defaultsDeep could be tricked into adding or modifying properties of Object.prototype using a constructor payload. <p>Publish Date: 2019-07-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744>CVE-2019-10744</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-jf85-cpcp-j695">https://github.com/advisories/GHSA-jf85-cpcp-j695</a></p> <p>Release Date: 2019-07-26</p> <p>Fix Resolution (lodash): 4.17.12</p> <p>Direct dependency fix Resolution (html-webpack-plugin): 4.0.0-alpha</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-8203</summary> ### Vulnerable Library - <b>lodash-4.17.11.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - html-webpack-plugin-3.2.0.tgz (Root Library) - :x: **lodash-4.17.11.tgz** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> <p></p> ### Vulnerability Details <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.4</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution (lodash): 4.17.19</p> <p>Direct dependency fix Resolution (html-webpack-plugin): 4.0.0-alpha</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23337</summary> ### Vulnerable Library - <b>lodash-4.17.11.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - html-webpack-plugin-3.2.0.tgz (Root Library) - :x: **lodash-4.17.11.tgz** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> <p></p> ### Vulnerability Details <p> Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.2</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution (lodash): 4.17.21</p> <p>Direct dependency fix Resolution (html-webpack-plugin): 4.0.0-alpha</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-28500</summary> ### Vulnerable Library - <b>lodash-4.17.11.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - html-webpack-plugin-3.2.0.tgz (Root Library) - :x: **lodash-4.17.11.tgz** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> <p></p> ### Vulnerability Details <p> Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution (lodash): 4.17.21</p> <p>Direct dependency fix Resolution (html-webpack-plugin): 4.0.0-alpha</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"html-webpack-plugin","packageVersion":"3.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"html-webpack-plugin:3.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0.0-alpha","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2019-10744","vulnerabilityDetails":"Versions of lodash lower than 4.17.12 are vulnerable to Prototype Pollution. The function defaultsDeep could be tricked into adding or modifying properties of Object.prototype using a constructor payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744","cvss3Severity":"high","cvss3Score":"9.1","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"html-webpack-plugin","packageVersion":"3.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"html-webpack-plugin:3.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0.0-alpha","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-8203","vulnerabilityDetails":"Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"html-webpack-plugin","packageVersion":"3.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"html-webpack-plugin:3.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0.0-alpha","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2021-23337","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337","cvss3Severity":"high","cvss3Score":"7.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"html-webpack-plugin","packageVersion":"3.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"html-webpack-plugin:3.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0.0-alpha","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-28500","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.\n WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> -->
non_defect
html webpack plugin tgz vulnerabilities highest severity is vulnerable library html webpack plugin tgz path to dependency file package json path to vulnerable library node modules lodash package json vulnerabilities cve severity cvss dependency type fixed in remediation available high lodash tgz transitive alpha ❌ high lodash tgz transitive alpha ❌ high lodash tgz transitive alpha ❌ medium lodash tgz transitive alpha ❌ details cve vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy html webpack plugin tgz root library x lodash tgz vulnerable library found in base branch dev vulnerability details versions of lodash lower than are vulnerable to prototype pollution the function defaultsdeep could be tricked into adding or modifying properties of object prototype using a constructor payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution html webpack plugin alpha step up your open source security game with whitesource cve vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy html webpack plugin tgz root library x lodash tgz vulnerable library found in base branch dev vulnerability details prototype pollution attack when using zipobjectdeep in lodash before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution html webpack plugin alpha step up your open source security game with whitesource cve vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy html webpack plugin tgz root library x lodash tgz vulnerable library found in base branch dev vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution html webpack plugin alpha step up your open source security game with whitesource cve vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy html webpack plugin tgz root library x lodash tgz vulnerable library found in base branch dev vulnerability details lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions whitesource note after conducting further research whitesource has determined that cve only affects environments with versions to of lodash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution html webpack plugin alpha step up your open source security game with whitesource istransitivedependency false dependencytree html webpack plugin isminimumfixversionavailable true minimumfixversion alpha isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails versions of lodash lower than are vulnerable to prototype pollution the function defaultsdeep could be tricked into adding or modifying properties of object prototype using a constructor payload vulnerabilityurl istransitivedependency false dependencytree html webpack plugin isminimumfixversionavailable true minimumfixversion alpha isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails prototype pollution attack when using zipobjectdeep in lodash before vulnerabilityurl istransitivedependency false dependencytree html webpack plugin isminimumfixversionavailable true minimumfixversion alpha isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails lodash versions prior to are vulnerable to command injection via the template function vulnerabilityurl istransitivedependency false dependencytree html webpack plugin isminimumfixversionavailable true minimumfixversion alpha isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions n whitesource note after conducting further research whitesource has determined that cve only affects environments with versions to of lodash vulnerabilityurl
0
129,775
18,109,740,410
IssuesEvent
2021-09-23 01:02:09
maorkuriel/ksa
https://api.github.com/repos/maorkuriel/ksa
opened
CVE-2019-10086 (High) detected in commons-beanutils-1.8.3.jar
security vulnerability
## CVE-2019-10086 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-beanutils-1.8.3.jar</b></p></summary> <p>BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p> <p> Dependency Hierarchy: - shiro-core-1.2.0.jar (Root Library) - :x: **commons-beanutils-1.8.3.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean. <p>Publish Date: 2019-08-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10086>CVE-2019-10086</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/victims/victims-cve-db/commit/16a669c84d95bbbd4294f30e609049a36700847f">https://github.com/victims/victims-cve-db/commit/16a669c84d95bbbd4294f30e609049a36700847f</a></p> <p>Release Date: 2019-08-20</p> <p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-beanutils","packageName":"commons-beanutils","packageVersion":"1.8.3","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"org.apache.shiro:shiro-core:1.2.0;commons-beanutils:commons-beanutils:1.8.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-beanutils:commons-beanutils:1.9.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-10086","vulnerabilityDetails":"In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10086","cvss3Severity":"high","cvss3Score":"7.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-10086 (High) detected in commons-beanutils-1.8.3.jar - ## CVE-2019-10086 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-beanutils-1.8.3.jar</b></p></summary> <p>BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p> <p> Dependency Hierarchy: - shiro-core-1.2.0.jar (Root Library) - :x: **commons-beanutils-1.8.3.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean. <p>Publish Date: 2019-08-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10086>CVE-2019-10086</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/victims/victims-cve-db/commit/16a669c84d95bbbd4294f30e609049a36700847f">https://github.com/victims/victims-cve-db/commit/16a669c84d95bbbd4294f30e609049a36700847f</a></p> <p>Release Date: 2019-08-20</p> <p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-beanutils","packageName":"commons-beanutils","packageVersion":"1.8.3","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"org.apache.shiro:shiro-core:1.2.0;commons-beanutils:commons-beanutils:1.8.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-beanutils:commons-beanutils:1.9.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-10086","vulnerabilityDetails":"In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10086","cvss3Severity":"high","cvss3Score":"7.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in commons beanutils jar cve high severity vulnerability vulnerable library commons beanutils jar beanutils provides an easy to use but flexible wrapper around reflection and introspection dependency hierarchy shiro core jar root library x commons beanutils jar vulnerable library found in base branch master vulnerability details in apache commons beanutils a special beanintrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all java objects we however were not using this by default characteristic of the propertyutilsbean publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution commons beanutils commons beanutils isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache shiro shiro core commons beanutils commons beanutils isminimumfixversionavailable true minimumfixversion commons beanutils commons beanutils basebranches vulnerabilityidentifier cve vulnerabilitydetails in apache commons beanutils a special beanintrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all java objects we however were not using this by default characteristic of the propertyutilsbean vulnerabilityurl
0
816,562
30,602,821,531
IssuesEvent
2023-07-22 16:03:41
Willings-Inc/SCG
https://api.github.com/repos/Willings-Inc/SCG
closed
Extend the expiration date of points for user id 95.
Highest Priority(1)
https://takashiichima-af82034.slack.com/archives/C030P8L1AER/p1689659734972169 `②吉田響子様(ID:95)のポイントを延長 ポイント有効期限8/6→9/7に変更` Please extend the current expiration date of points for user id 95, Kyoko Yoshida. From August 6th to Sep 7th.
1.0
Extend the expiration date of points for user id 95. - https://takashiichima-af82034.slack.com/archives/C030P8L1AER/p1689659734972169 `②吉田響子様(ID:95)のポイントを延長 ポイント有効期限8/6→9/7に変更` Please extend the current expiration date of points for user id 95, Kyoko Yoshida. From August 6th to Sep 7th.
non_defect
extend the expiration date of points for user id ②吉田響子様 id のポイントを延長 → please extend the current expiration date of points for user id kyoko yoshida from august to sep
0
20,182
3,309,984,717
IssuesEvent
2015-11-05 05:34:40
navjagpal/browser-timetracker
https://api.github.com/repos/navjagpal/browser-timetracker
closed
Incompatibility with IE tab
auto-migrated Priority-Medium Type-Defect
``` The extension does not track any pages loaded through the IE tab extension. Pages opened with ie tab will have an address starting with chrome-extension://hehijbfgiekmjfkfjpbkbammjbdenadd/ ``` Original issue reported on code.google.com by `datatecr...@aol.com` on 29 Apr 2011 at 4:29
1.0
Incompatibility with IE tab - ``` The extension does not track any pages loaded through the IE tab extension. Pages opened with ie tab will have an address starting with chrome-extension://hehijbfgiekmjfkfjpbkbammjbdenadd/ ``` Original issue reported on code.google.com by `datatecr...@aol.com` on 29 Apr 2011 at 4:29
defect
incompatibility with ie tab the extension does not track any pages loaded through the ie tab extension pages opened with ie tab will have an address starting with chrome extension hehijbfgiekmjfkfjpbkbammjbdenadd original issue reported on code google com by datatecr aol com on apr at
1
738,237
25,550,078,955
IssuesEvent
2022-11-29 22:38:29
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
opened
Disabling already disabled storage-provisioner-gluster addon errors
kind/bug area/addons priority/backlog
**How to reproduce:** ``` # start minikube $ minikube start # disable the already disabled storage-provisioner-gluster addon $ minikube addons disable storage-provisioner-gluster ❌ Exiting due to MK_ADDON_DISABLE: run callbacks: running callbacks: [Error disabling glusterfile as the default storage class: Error getting storage class glusterfile: storageclasses.storage.k8s.io "glusterfile" not found] ╭─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮ │ │ │ 😿 If the above advice does not help, please let us know: │ │ 👉 https://github.com/kubernetes/minikube/issues/new/choose │ │ │ │ Please run `minikube logs --file=logs.txt` and attach logs.txt to the GitHub issue. │ │ Please also attach the following file to the GitHub issue: │ │ - /var/folders/9l/6wpxv6wd1b901m1146r579wc00rqw3/T/minikube_addons_a1222cee5cdad746bbb0c5aa40a377a17b232811_0.log │ │ │ ╰─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯ ``` **Expected:** ``` 🌑 "The 'storage-provisioner-gluster' addon is disabled ```
1.0
Disabling already disabled storage-provisioner-gluster addon errors - **How to reproduce:** ``` # start minikube $ minikube start # disable the already disabled storage-provisioner-gluster addon $ minikube addons disable storage-provisioner-gluster ❌ Exiting due to MK_ADDON_DISABLE: run callbacks: running callbacks: [Error disabling glusterfile as the default storage class: Error getting storage class glusterfile: storageclasses.storage.k8s.io "glusterfile" not found] ╭─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮ │ │ │ 😿 If the above advice does not help, please let us know: │ │ 👉 https://github.com/kubernetes/minikube/issues/new/choose │ │ │ │ Please run `minikube logs --file=logs.txt` and attach logs.txt to the GitHub issue. │ │ Please also attach the following file to the GitHub issue: │ │ - /var/folders/9l/6wpxv6wd1b901m1146r579wc00rqw3/T/minikube_addons_a1222cee5cdad746bbb0c5aa40a377a17b232811_0.log │ │ │ ╰─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯ ``` **Expected:** ``` 🌑 "The 'storage-provisioner-gluster' addon is disabled ```
non_defect
disabling already disabled storage provisioner gluster addon errors how to reproduce start minikube minikube start disable the already disabled storage provisioner gluster addon minikube addons disable storage provisioner gluster ❌ exiting due to mk addon disable run callbacks running callbacks ╭─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮ │ │ │ 😿 if the above advice does not help please let us know │ │ 👉 │ │ │ │ please run minikube logs file logs txt and attach logs txt to the github issue │ │ please also attach the following file to the github issue │ │ var folders t minikube addons log │ │ │ ╰─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯ expected 🌑 the storage provisioner gluster addon is disabled
0
40,309
9,942,254,937
IssuesEvent
2019-07-03 13:31:46
OpenMS/OpenMS
https://api.github.com/repos/OpenMS/OpenMS
closed
TOPPView documentation is still a buildtarget with WITH_GUI=off
Documentation defect minor wontfix
eventually leading to a build error.
1.0
TOPPView documentation is still a buildtarget with WITH_GUI=off - eventually leading to a build error.
defect
toppview documentation is still a buildtarget with with gui off eventually leading to a build error
1
397,583
27,170,425,864
IssuesEvent
2023-02-17 18:53:13
aws/aws-cdk
https://api.github.com/repos/aws/aws-cdk
closed
SSM: Invalid link on SSM Document "Target Type" Property
@aws-cdk/aws-ssm needs-cfn documentation
### Describe the issue Minor link issue on the ssm documentation page for "CfnDocument". The "targetType" has a link with the text "AWS resource and property types reference" which directs to "https://awscloudformation/latest/UserGuide/aws-template-resource-type-ref.html". This looks to be down to the documentation line having a double slash between `docs.aws.amazon.com` and `AWSCloudFormation`. ``` "TargetType": "Specify a target type to define the kinds of resources the document can run on. For example, to run a document on EC2 instances, specify the following value: `/AWS::EC2::Instance` . If you specify a value of '/' the document can run on all types of resources. If you don't specify a value, the document can't run on any resources. For a list of valid resource types, see [AWS resource and property types reference](https://docs.aws.amazon.com//AWSCloudFormation/latest/UserGuide/aws-template-resource-type-ref.html) in the *AWS CloudFormation User Guide* .", ``` (See packages/@aws-cdk/cfnspec/spec-source/cfn-docs/cfn-docs.json#L44463) ([Locked line number by making fork](https://raw.githubusercontent.com/ryanolee/aws-cdk/main/packages/%40aws-cdk/cfnspec/spec-source/cfn-docs/cfn-docs.json)) ### Links https://docs.aws.amazon.com/cdk/api/v2/docs/aws-cdk-lib.aws_ssm.CfnDocument.html https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-resource-ssm-document.html
1.0
SSM: Invalid link on SSM Document "Target Type" Property - ### Describe the issue Minor link issue on the ssm documentation page for "CfnDocument". The "targetType" has a link with the text "AWS resource and property types reference" which directs to "https://awscloudformation/latest/UserGuide/aws-template-resource-type-ref.html". This looks to be down to the documentation line having a double slash between `docs.aws.amazon.com` and `AWSCloudFormation`. ``` "TargetType": "Specify a target type to define the kinds of resources the document can run on. For example, to run a document on EC2 instances, specify the following value: `/AWS::EC2::Instance` . If you specify a value of '/' the document can run on all types of resources. If you don't specify a value, the document can't run on any resources. For a list of valid resource types, see [AWS resource and property types reference](https://docs.aws.amazon.com//AWSCloudFormation/latest/UserGuide/aws-template-resource-type-ref.html) in the *AWS CloudFormation User Guide* .", ``` (See packages/@aws-cdk/cfnspec/spec-source/cfn-docs/cfn-docs.json#L44463) ([Locked line number by making fork](https://raw.githubusercontent.com/ryanolee/aws-cdk/main/packages/%40aws-cdk/cfnspec/spec-source/cfn-docs/cfn-docs.json)) ### Links https://docs.aws.amazon.com/cdk/api/v2/docs/aws-cdk-lib.aws_ssm.CfnDocument.html https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-resource-ssm-document.html
non_defect
ssm invalid link on ssm document target type property describe the issue minor link issue on the ssm documentation page for cfndocument the targettype has a link with the text aws resource and property types reference which directs to this looks to be down to the documentation line having a double slash between docs aws amazon com and awscloudformation targettype specify a target type to define the kinds of resources the document can run on for example to run a document on instances specify the following value aws instance if you specify a value of the document can run on all types of resources if you don t specify a value the document can t run on any resources for a list of valid resource types see in the aws cloudformation user guide see packages aws cdk cfnspec spec source cfn docs cfn docs json links
0
259,885
22,574,932,991
IssuesEvent
2022-06-28 06:18:56
eclipse/tycho
https://api.github.com/repos/eclipse/tycho
closed
BundleException: Bundle bundle cannot be resolved with tycho 2.3.0
test case required
I'm getting the following BundleException with tycho 2.3.0, but tycho 2.2.0 works: ``` Caused by: org.osgi.framework.BundleException: Bundle bundle cannot be resolved:bundle [100] Unresolved requirement: Require-Bundle: com.collabnet.subversion.merge -> Bundle-SymbolicName: com.collabnet.subversion.merge; bundle-version="4.2.0.1"; singleton:="true" com.collabnet.subversion.merge [1] Unresolved requirement: Require-Bundle: org.eclipse.jdt.core at org.eclipse.tycho.core.osgitools.EquinoxResolver.assertResolved (EquinoxResolver.java:423) at org.eclipse.tycho.core.osgitools.EquinoxResolver.newResolvedState (EquinoxResolver.java:127) at org.eclipse.tycho.core.osgitools.OsgiBundleProject.getResolverState (OsgiBundleProject.java:281) at org.eclipse.tycho.core.osgitools.OsgiBundleProject.resolveClassPath (OsgiBundleProject.java:177) at org.eclipse.tycho.core.resolver.DefaultTychoResolver.resolveProject (DefaultTychoResolver.java:142) ``` I've published a reproducer here: https://github.com/jhonnen/tychorepro
1.0
BundleException: Bundle bundle cannot be resolved with tycho 2.3.0 - I'm getting the following BundleException with tycho 2.3.0, but tycho 2.2.0 works: ``` Caused by: org.osgi.framework.BundleException: Bundle bundle cannot be resolved:bundle [100] Unresolved requirement: Require-Bundle: com.collabnet.subversion.merge -> Bundle-SymbolicName: com.collabnet.subversion.merge; bundle-version="4.2.0.1"; singleton:="true" com.collabnet.subversion.merge [1] Unresolved requirement: Require-Bundle: org.eclipse.jdt.core at org.eclipse.tycho.core.osgitools.EquinoxResolver.assertResolved (EquinoxResolver.java:423) at org.eclipse.tycho.core.osgitools.EquinoxResolver.newResolvedState (EquinoxResolver.java:127) at org.eclipse.tycho.core.osgitools.OsgiBundleProject.getResolverState (OsgiBundleProject.java:281) at org.eclipse.tycho.core.osgitools.OsgiBundleProject.resolveClassPath (OsgiBundleProject.java:177) at org.eclipse.tycho.core.resolver.DefaultTychoResolver.resolveProject (DefaultTychoResolver.java:142) ``` I've published a reproducer here: https://github.com/jhonnen/tychorepro
non_defect
bundleexception bundle bundle cannot be resolved with tycho i m getting the following bundleexception with tycho but tycho works caused by org osgi framework bundleexception bundle bundle cannot be resolved bundle unresolved requirement require bundle com collabnet subversion merge bundle symbolicname com collabnet subversion merge bundle version singleton true com collabnet subversion merge unresolved requirement require bundle org eclipse jdt core at org eclipse tycho core osgitools equinoxresolver assertresolved equinoxresolver java at org eclipse tycho core osgitools equinoxresolver newresolvedstate equinoxresolver java at org eclipse tycho core osgitools osgibundleproject getresolverstate osgibundleproject java at org eclipse tycho core osgitools osgibundleproject resolveclasspath osgibundleproject java at org eclipse tycho core resolver defaulttychoresolver resolveproject defaulttychoresolver java i ve published a reproducer here
0
46,748
13,055,968,873
IssuesEvent
2020-07-30 03:15:55
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
[steamshovel] ShovelLogger does not honor logging unit filter levels (Trac #1805)
Incomplete Migration Migrated from Trac cmake defect
Migrated from https://code.icecube.wisc.edu/ticket/1805 ```json { "status": "closed", "changetime": "2016-08-16T08:38:32", "description": "Steamshovel's ShovelLogger does not check whether a log messages should be displayed or not, resulting in a cluttered log if e.g. a lot of TRACE messages are logged.\n\n'''Example code:'''\n{{{\n#!python\nfrom icecube.icetray import logging\n\nlogging.set_level_for_unit(\"UnitName\", logging.I3LogLevel.LOG_WARN)\nlogging.log_info(\"Foo bar\", \"UnitName\")\n}}}\n\n'''Behaviour:'''\nThere should be no log entry with \"''INFO (UnitName): ...''\" since the level for UnitName was set to `WARN`. However, no filtering is applied to log statements in the log viewer, resulting in messages down to the lowest level.", "reporter": "sander.vanheule", "cc": "hdembinski", "resolution": "fixed", "_ts": "1471336712035911", "component": "cmake", "summary": "[steamshovel] ShovelLogger does not honor logging unit filter levels", "priority": "normal", "keywords": "steamshovel", "time": "2016-07-27T15:33:43", "milestone": "", "owner": "", "type": "defect" } ```
1.0
[steamshovel] ShovelLogger does not honor logging unit filter levels (Trac #1805) - Migrated from https://code.icecube.wisc.edu/ticket/1805 ```json { "status": "closed", "changetime": "2016-08-16T08:38:32", "description": "Steamshovel's ShovelLogger does not check whether a log messages should be displayed or not, resulting in a cluttered log if e.g. a lot of TRACE messages are logged.\n\n'''Example code:'''\n{{{\n#!python\nfrom icecube.icetray import logging\n\nlogging.set_level_for_unit(\"UnitName\", logging.I3LogLevel.LOG_WARN)\nlogging.log_info(\"Foo bar\", \"UnitName\")\n}}}\n\n'''Behaviour:'''\nThere should be no log entry with \"''INFO (UnitName): ...''\" since the level for UnitName was set to `WARN`. However, no filtering is applied to log statements in the log viewer, resulting in messages down to the lowest level.", "reporter": "sander.vanheule", "cc": "hdembinski", "resolution": "fixed", "_ts": "1471336712035911", "component": "cmake", "summary": "[steamshovel] ShovelLogger does not honor logging unit filter levels", "priority": "normal", "keywords": "steamshovel", "time": "2016-07-27T15:33:43", "milestone": "", "owner": "", "type": "defect" } ```
defect
shovellogger does not honor logging unit filter levels trac migrated from json status closed changetime description steamshovel s shovellogger does not check whether a log messages should be displayed or not resulting in a cluttered log if e g a lot of trace messages are logged n n example code n n python nfrom icecube icetray import logging n nlogging set level for unit unitname logging log warn nlogging log info foo bar unitname n n n behaviour nthere should be no log entry with info unitname since the level for unitname was set to warn however no filtering is applied to log statements in the log viewer resulting in messages down to the lowest level reporter sander vanheule cc hdembinski resolution fixed ts component cmake summary shovellogger does not honor logging unit filter levels priority normal keywords steamshovel time milestone owner type defect
1
170,105
6,424,123,577
IssuesEvent
2017-08-09 12:51:04
cms-gem-daq-project/vfatqc-python-scripts
https://api.github.com/repos/cms-gem-daq-project/vfatqc-python-scripts
closed
Extra parameters for latency scan
Priority: High Type: Feature Request
- need to propagate upstream to run_scans.py options to use the amc13 local trigger: - `--amc13local` - `--t3trig` - `--randoms` - need to add the option to set trigger throttle
1.0
Extra parameters for latency scan - - need to propagate upstream to run_scans.py options to use the amc13 local trigger: - `--amc13local` - `--t3trig` - `--randoms` - need to add the option to set trigger throttle
non_defect
extra parameters for latency scan need to propagate upstream to run scans py options to use the local trigger randoms need to add the option to set trigger throttle
0
347,306
10,428,294,098
IssuesEvent
2019-09-16 22:07:07
clearlinux/clr-installer
https://api.github.com/repos/clearlinux/clr-installer
closed
Advanced Configuration fails with existing paritions
TUI bug high priority
**Describe the bug** Using a disk with existing partitions can fail to install using Advanced Configuration **To Reproduce** Steps to reproduce the behavior: 1. Have a system with a disk with existing partitions 2. Launch TUI 3. Select Media Config 4. Select Advanced Config 5. Select the Auto Partition for the disk with existing partitions 6. Change one of the newly defined partitions 7. Complete the required install steps 8. Installation will fail **Expected behavior** Installation should create all of the newly selected partitions **Environment (please complete the following information):** - Clear Linux OS Version: Installed version: 29040 Version URL: https://cdn.download.clearlinux.org/update/ Content URL: https://cdn.download.clearlinux.org/update/ - Installer Version clr-installer: 1.2.1 **Additional context** The installer is not performing the mkpart mklabel gpt It knows that the mklabel is required after the Auto Partition, but loose that information/flag after doing any manual updates.
1.0
Advanced Configuration fails with existing paritions - **Describe the bug** Using a disk with existing partitions can fail to install using Advanced Configuration **To Reproduce** Steps to reproduce the behavior: 1. Have a system with a disk with existing partitions 2. Launch TUI 3. Select Media Config 4. Select Advanced Config 5. Select the Auto Partition for the disk with existing partitions 6. Change one of the newly defined partitions 7. Complete the required install steps 8. Installation will fail **Expected behavior** Installation should create all of the newly selected partitions **Environment (please complete the following information):** - Clear Linux OS Version: Installed version: 29040 Version URL: https://cdn.download.clearlinux.org/update/ Content URL: https://cdn.download.clearlinux.org/update/ - Installer Version clr-installer: 1.2.1 **Additional context** The installer is not performing the mkpart mklabel gpt It knows that the mklabel is required after the Auto Partition, but loose that information/flag after doing any manual updates.
non_defect
advanced configuration fails with existing paritions describe the bug using a disk with existing partitions can fail to install using advanced configuration to reproduce steps to reproduce the behavior have a system with a disk with existing partitions launch tui select media config select advanced config select the auto partition for the disk with existing partitions change one of the newly defined partitions complete the required install steps installation will fail expected behavior installation should create all of the newly selected partitions environment please complete the following information clear linux os version installed version version url content url installer version clr installer additional context the installer is not performing the mkpart mklabel gpt it knows that the mklabel is required after the auto partition but loose that information flag after doing any manual updates
0
35,598
17,145,184,445
IssuesEvent
2021-07-13 13:57:15
elementor/elementor
https://api.github.com/repos/elementor/elementor
closed
Enqueues unnecessary JS files on the frontend
status/merged type/performance
## Prerequisites <!-- MARK COMPLETED ITEMS WITH AN [x] --> - [x] I have searched for similar issues in both open and closed tickets and cannot find a duplicate. - [x] The issue still exists against the latest stable version of Elementor. ## Description The plugin adds extra JS files for functionalities which are not being used on the page. For example, on my [site's homepage](https://pixify.net), the plugin enqueues the following scripts which are not used: - lib/sticky/jquery.sticky.min.js (no sticky elements on the page; added by Elementor Pro) - jquery/ui/position.min.js (this one's probably a dependency for one of the below Elementor scripts) - lib/dialog/dialog.min.js (no dialog boxes either) - lib/swiper/swiper.min.js (no sliders used) This does not even include the `frontend-modules.min.js` by Elementor and `frontend.min.js` by Elementor Pro which I guess are files that define general functions. (...wondering what important task do they do which is supposed to be loaded on all the pages) So in short, Elementor (even just the free version) adds a lot of unnecessary JS files to the frontend whose functionalities are not being used. Affects site performance by quite a margin. I think you guys need to modularize asset enqueuing better so that you have a system in place to detect whether the particular features are being used or not and only then enqueue these. I appreciate the Element specific enqueuing model that Elementor has adapted which only includes those assets when those specific elements are used. So similarly, if that type of modular loading would be applied to these general elements as well, that would significantly cut down the number of assets enqueued and would only load the required files. Elementor has already become a very mature Page Builder and so, the modular asset enqueuing is something that is expected from it and should've been already implemented. I hope this would be prioritized and fixed soon. ## Isolating the problem <!-- MARK COMPLETED ITEMS WITH AN [x] --> - [x] This bug happens with only Elementor plugin active (and Elementor Pro). - [x] This bug happens with a default WordPress theme active. - [x] I can reproduce this bug consistently using the steps above.
True
Enqueues unnecessary JS files on the frontend - ## Prerequisites <!-- MARK COMPLETED ITEMS WITH AN [x] --> - [x] I have searched for similar issues in both open and closed tickets and cannot find a duplicate. - [x] The issue still exists against the latest stable version of Elementor. ## Description The plugin adds extra JS files for functionalities which are not being used on the page. For example, on my [site's homepage](https://pixify.net), the plugin enqueues the following scripts which are not used: - lib/sticky/jquery.sticky.min.js (no sticky elements on the page; added by Elementor Pro) - jquery/ui/position.min.js (this one's probably a dependency for one of the below Elementor scripts) - lib/dialog/dialog.min.js (no dialog boxes either) - lib/swiper/swiper.min.js (no sliders used) This does not even include the `frontend-modules.min.js` by Elementor and `frontend.min.js` by Elementor Pro which I guess are files that define general functions. (...wondering what important task do they do which is supposed to be loaded on all the pages) So in short, Elementor (even just the free version) adds a lot of unnecessary JS files to the frontend whose functionalities are not being used. Affects site performance by quite a margin. I think you guys need to modularize asset enqueuing better so that you have a system in place to detect whether the particular features are being used or not and only then enqueue these. I appreciate the Element specific enqueuing model that Elementor has adapted which only includes those assets when those specific elements are used. So similarly, if that type of modular loading would be applied to these general elements as well, that would significantly cut down the number of assets enqueued and would only load the required files. Elementor has already become a very mature Page Builder and so, the modular asset enqueuing is something that is expected from it and should've been already implemented. I hope this would be prioritized and fixed soon. ## Isolating the problem <!-- MARK COMPLETED ITEMS WITH AN [x] --> - [x] This bug happens with only Elementor plugin active (and Elementor Pro). - [x] This bug happens with a default WordPress theme active. - [x] I can reproduce this bug consistently using the steps above.
non_defect
enqueues unnecessary js files on the frontend prerequisites i have searched for similar issues in both open and closed tickets and cannot find a duplicate the issue still exists against the latest stable version of elementor description the plugin adds extra js files for functionalities which are not being used on the page for example on my the plugin enqueues the following scripts which are not used lib sticky jquery sticky min js no sticky elements on the page added by elementor pro jquery ui position min js this one s probably a dependency for one of the below elementor scripts lib dialog dialog min js no dialog boxes either lib swiper swiper min js no sliders used this does not even include the frontend modules min js by elementor and frontend min js by elementor pro which i guess are files that define general functions wondering what important task do they do which is supposed to be loaded on all the pages so in short elementor even just the free version adds a lot of unnecessary js files to the frontend whose functionalities are not being used affects site performance by quite a margin i think you guys need to modularize asset enqueuing better so that you have a system in place to detect whether the particular features are being used or not and only then enqueue these i appreciate the element specific enqueuing model that elementor has adapted which only includes those assets when those specific elements are used so similarly if that type of modular loading would be applied to these general elements as well that would significantly cut down the number of assets enqueued and would only load the required files elementor has already become a very mature page builder and so the modular asset enqueuing is something that is expected from it and should ve been already implemented i hope this would be prioritized and fixed soon isolating the problem this bug happens with only elementor plugin active and elementor pro this bug happens with a default wordpress theme active i can reproduce this bug consistently using the steps above
0
46,077
13,055,848,902
IssuesEvent
2020-07-30 02:55:05
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
dataio-shovel returns incomplete xml dump (Trac #549)
Incomplete Migration Migrated from Trac dataio defect
Migrated from https://code.icecube.wisc.edu/ticket/549 ```json { "status": "closed", "changetime": "2009-07-22T19:31:04", "description": "On systems with large memory, the \"x\" xml output is not complete (still using the pretty print function internally). On smaller memory systems, xml output fails and falls over to dump xml printing and you get the whole output.\n\nSome internal pretty print maximum buffersize?", "reporter": "anonymous", "cc": "", "resolution": "fixed", "_ts": "1248291064000000", "component": "dataio", "summary": "dataio-shovel returns incomplete xml dump", "priority": "normal", "keywords": "", "time": "2009-04-10T00:45:44", "milestone": "", "owner": "troy", "type": "defect" } ```
1.0
dataio-shovel returns incomplete xml dump (Trac #549) - Migrated from https://code.icecube.wisc.edu/ticket/549 ```json { "status": "closed", "changetime": "2009-07-22T19:31:04", "description": "On systems with large memory, the \"x\" xml output is not complete (still using the pretty print function internally). On smaller memory systems, xml output fails and falls over to dump xml printing and you get the whole output.\n\nSome internal pretty print maximum buffersize?", "reporter": "anonymous", "cc": "", "resolution": "fixed", "_ts": "1248291064000000", "component": "dataio", "summary": "dataio-shovel returns incomplete xml dump", "priority": "normal", "keywords": "", "time": "2009-04-10T00:45:44", "milestone": "", "owner": "troy", "type": "defect" } ```
defect
dataio shovel returns incomplete xml dump trac migrated from json status closed changetime description on systems with large memory the x xml output is not complete still using the pretty print function internally on smaller memory systems xml output fails and falls over to dump xml printing and you get the whole output n nsome internal pretty print maximum buffersize reporter anonymous cc resolution fixed ts component dataio summary dataio shovel returns incomplete xml dump priority normal keywords time milestone owner troy type defect
1
37,415
8,391,817,295
IssuesEvent
2018-10-09 15:53:46
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Generate dummy ORDER BY clause for ranking functions on databases that require them
C: Functionality E: All Editions P: Medium R: Fixed T: Defect
When writing ``` sql create.select(rowNumber().over()).fetch() ``` Oracle will complain that the ORDER BY clause is missing. We should generate a dummy order by clause as such: ``` sql [RANKING_FUNCTION()] OVER(ORDER BY (SELECT 1 /* from dual */)) ``` This will produce an arbitrary ordering, which is probably acceptable. ---- See also this discussion here: https://groups.google.com/d/topic/jooq-user/_FhoRCKJulE/discussion
1.0
Generate dummy ORDER BY clause for ranking functions on databases that require them - When writing ``` sql create.select(rowNumber().over()).fetch() ``` Oracle will complain that the ORDER BY clause is missing. We should generate a dummy order by clause as such: ``` sql [RANKING_FUNCTION()] OVER(ORDER BY (SELECT 1 /* from dual */)) ``` This will produce an arbitrary ordering, which is probably acceptable. ---- See also this discussion here: https://groups.google.com/d/topic/jooq-user/_FhoRCKJulE/discussion
defect
generate dummy order by clause for ranking functions on databases that require them when writing sql create select rownumber over fetch oracle will complain that the order by clause is missing we should generate a dummy order by clause as such sql over order by select from dual this will produce an arbitrary ordering which is probably acceptable see also this discussion here
1
64,130
18,238,723,433
IssuesEvent
2021-10-01 10:11:58
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Timeline wobbles vertically when viewing a new room for the first time as the history loads.
T-Defect X-Needs-Info X-Regression S-Minor A-Timeline O-Frequent
### Steps to reproduce View a room for the first time since launching the client (particularly an E2EE room, where the effect is either more noticeable because the history loads more slowly, or perhaps it only effects E2EE rooms). ### What happened? Watch the timeline carefully as the history is displayed on the room; notice that some messages sometimes 'dance' or 'wobble' up and down by 1px (or perhaps 0.5px?) as the history is added higher up the page, e.g. Anoa's one here. This is a regression in the last week or so, and makes things literally look wobbly and flakey. https://user-images.githubusercontent.com/1294269/129336575-33527256-e916-4f47-9aab-684d884ae9cc.mov ### What did you expect? The timeline should be rock solid as the history backfills. It's acceptable (but not desirable) for replies and URL previews and media thumbnails to push events *upwards* if they load below them in the timeline (but ideally we should be reserving the right vertical space in the timeline for these things). But it's not acceptable for things loading *above* the content to make the content wobble around. ### Operating system macOS 10.15.7 ### Application version nightly 2021-08-12 ### How did you install the app? nightly
1.0
Timeline wobbles vertically when viewing a new room for the first time as the history loads. - ### Steps to reproduce View a room for the first time since launching the client (particularly an E2EE room, where the effect is either more noticeable because the history loads more slowly, or perhaps it only effects E2EE rooms). ### What happened? Watch the timeline carefully as the history is displayed on the room; notice that some messages sometimes 'dance' or 'wobble' up and down by 1px (or perhaps 0.5px?) as the history is added higher up the page, e.g. Anoa's one here. This is a regression in the last week or so, and makes things literally look wobbly and flakey. https://user-images.githubusercontent.com/1294269/129336575-33527256-e916-4f47-9aab-684d884ae9cc.mov ### What did you expect? The timeline should be rock solid as the history backfills. It's acceptable (but not desirable) for replies and URL previews and media thumbnails to push events *upwards* if they load below them in the timeline (but ideally we should be reserving the right vertical space in the timeline for these things). But it's not acceptable for things loading *above* the content to make the content wobble around. ### Operating system macOS 10.15.7 ### Application version nightly 2021-08-12 ### How did you install the app? nightly
defect
timeline wobbles vertically when viewing a new room for the first time as the history loads steps to reproduce view a room for the first time since launching the client particularly an room where the effect is either more noticeable because the history loads more slowly or perhaps it only effects rooms what happened watch the timeline carefully as the history is displayed on the room notice that some messages sometimes dance or wobble up and down by or perhaps as the history is added higher up the page e g anoa s one here this is a regression in the last week or so and makes things literally look wobbly and flakey what did you expect the timeline should be rock solid as the history backfills it s acceptable but not desirable for replies and url previews and media thumbnails to push events upwards if they load below them in the timeline but ideally we should be reserving the right vertical space in the timeline for these things but it s not acceptable for things loading above the content to make the content wobble around operating system macos application version nightly how did you install the app nightly
1
31,550
6,549,216,879
IssuesEvent
2017-09-05 05:27:19
ms609/citation-bot
https://api.github.com/repos/ms609/citation-bot
closed
Wikilinks in author field
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? {{cite journal | author = [[Elias James Corey|Corey, E. J.]]; Chaykovsky, M. | journal = [[J. Am. Chem. Soc.]] | year = 1965 | volume = 87 | pages = 1345 | title = Methylsulfinyl Carbanion (CH3-SO-CH2<sup>&minus;</sup>). Formation and Applications to Organic Synthesis | doi = 10.1021/ja01084a033}}</ref>, a large number of publications have described additional uses for this reagent.<ref>{{cite journal | author = Durst, T. | journal = [[Adv. Org. Chem.]] | year = 1969 | volume = 6 | pages = 285 | format = review}} What is the expected output? What do you see instead? author1=[[Elias James Corey|Corey, E. J.]] author1=[[elias James Corey|corey, E. J.]] Please use labels and text to provide additional information. ``` Original issue reported on code.google.com by `MartinS` on 6 Oct 2009 at 10:52
1.0
Wikilinks in author field - ``` What steps will reproduce the problem? {{cite journal | author = [[Elias James Corey|Corey, E. J.]]; Chaykovsky, M. | journal = [[J. Am. Chem. Soc.]] | year = 1965 | volume = 87 | pages = 1345 | title = Methylsulfinyl Carbanion (CH3-SO-CH2<sup>&minus;</sup>). Formation and Applications to Organic Synthesis | doi = 10.1021/ja01084a033}}</ref>, a large number of publications have described additional uses for this reagent.<ref>{{cite journal | author = Durst, T. | journal = [[Adv. Org. Chem.]] | year = 1969 | volume = 6 | pages = 285 | format = review}} What is the expected output? What do you see instead? author1=[[Elias James Corey|Corey, E. J.]] author1=[[elias James Corey|corey, E. J.]] Please use labels and text to provide additional information. ``` Original issue reported on code.google.com by `MartinS` on 6 Oct 2009 at 10:52
defect
wikilinks in author field what steps will reproduce the problem cite journal author chaykovsky m journal year volume pages title methylsulfinyl carbanion so minus formation and applications to organic synthesis doi a large number of publications have described additional uses for this reagent cite journal author durst t journal year volume pages format review what is the expected output what do you see instead please use labels and text to provide additional information original issue reported on code google com by martins on oct at
1
241,345
26,256,750,916
IssuesEvent
2023-01-06 01:54:25
rgordon95/advanced-react-demo
https://api.github.com/repos/rgordon95/advanced-react-demo
opened
WS-2020-0342 (High) detected in is-my-json-valid-2.19.0.tgz
security vulnerability
## WS-2020-0342 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.19.0.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.19.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.19.0.tgz</a></p> <p>Path to dependency file: /advanced-react-demo/package.json</p> <p>Path to vulnerable library: /node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - eslint-4.1.0.tgz (Root Library) - :x: **is-my-json-valid-2.19.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Regular Expression Denial of Service (ReDoS) vulnerability was found in is-my-json-valid before 2.20.2 via the style format. <p>Publish Date: 2020-06-27 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb>WS-2020-0342</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-06-27</p> <p>Fix Resolution (is-my-json-valid): 2.20.2</p> <p>Direct dependency fix Resolution (eslint): 4.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2020-0342 (High) detected in is-my-json-valid-2.19.0.tgz - ## WS-2020-0342 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.19.0.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.19.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.19.0.tgz</a></p> <p>Path to dependency file: /advanced-react-demo/package.json</p> <p>Path to vulnerable library: /node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - eslint-4.1.0.tgz (Root Library) - :x: **is-my-json-valid-2.19.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Regular Expression Denial of Service (ReDoS) vulnerability was found in is-my-json-valid before 2.20.2 via the style format. <p>Publish Date: 2020-06-27 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb>WS-2020-0342</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-06-27</p> <p>Fix Resolution (is-my-json-valid): 2.20.2</p> <p>Direct dependency fix Resolution (eslint): 4.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
ws high detected in is my json valid tgz ws high severity vulnerability vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file advanced react demo package json path to vulnerable library node modules is my json valid package json dependency hierarchy eslint tgz root library x is my json valid tgz vulnerable library vulnerability details regular expression denial of service redos vulnerability was found in is my json valid before via the style format publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution is my json valid direct dependency fix resolution eslint step up your open source security game with mend
0
31,147
6,443,895,161
IssuesEvent
2017-08-12 02:13:40
opendatakit/opendatakit
https://api.github.com/repos/opendatakit/opendatakit
closed
Fusion Table publishing name, renaming breaks links, and multiple publishers
Aggregate Beta OpSys-All Priority-Critical Type-Defect
Originally reported on Google Code with ID 301 ``` Using Aggregate 1.0, comment from user ( danieljbridges@gmail.com ): 2) Fusion table names default to 'data'. Changing the name seems to break the link. Furthermore having multiple fusion tables each linked to a form breaks all but the first publishing link. Presumably there is a conflict due the other tables also being named 'data'? When publishing to Google Spreadsheet one can define the tablename, but not with Fusion. Is there a way of defining it in the form? --- We should verify that publishing to multiple fusion tables from the same form works and that different fusion table names can be specified (assuming Fusion Tables allows this). ``` Reported by `mitchellsundt` on 2011-08-12 15:42:22
1.0
Fusion Table publishing name, renaming breaks links, and multiple publishers - Originally reported on Google Code with ID 301 ``` Using Aggregate 1.0, comment from user ( danieljbridges@gmail.com ): 2) Fusion table names default to 'data'. Changing the name seems to break the link. Furthermore having multiple fusion tables each linked to a form breaks all but the first publishing link. Presumably there is a conflict due the other tables also being named 'data'? When publishing to Google Spreadsheet one can define the tablename, but not with Fusion. Is there a way of defining it in the form? --- We should verify that publishing to multiple fusion tables from the same form works and that different fusion table names can be specified (assuming Fusion Tables allows this). ``` Reported by `mitchellsundt` on 2011-08-12 15:42:22
defect
fusion table publishing name renaming breaks links and multiple publishers originally reported on google code with id using aggregate comment from user danieljbridges gmail com fusion table names default to data changing the name seems to break the link furthermore having multiple fusion tables each linked to a form breaks all but the first publishing link presumably there is a conflict due the other tables also being named data when publishing to google spreadsheet one can define the tablename but not with fusion is there a way of defining it in the form we should verify that publishing to multiple fusion tables from the same form works and that different fusion table names can be specified assuming fusion tables allows this reported by mitchellsundt on
1
653,433
21,582,120,404
IssuesEvent
2022-05-02 19:56:30
googleapis/storage-testbench
https://api.github.com/repos/googleapis/storage-testbench
closed
Review implementation of decompressive transcoding
type: feature request priority: p3
I am not sure the implementation of decompressive transcoding works as it should. It seems to decompress the data when uploaded: https://github.com/googleapis/storage-testbench/blob/465f85e604f2ee526e10a774b4f8ead6c667c386/testbench/handle_gzip.py#L29-L34 But that probably won't work with `uploadType=multipart`, where the `Content-Type` is `multipart/related;`. I think we need to decompress the results on the fly, and disable decompression if the `Accept-Encoding: gzip` flag is enabled. We also need to store the checksums of the compressed data, which suggests (again) that decompressing the request is not the right idea.
1.0
Review implementation of decompressive transcoding - I am not sure the implementation of decompressive transcoding works as it should. It seems to decompress the data when uploaded: https://github.com/googleapis/storage-testbench/blob/465f85e604f2ee526e10a774b4f8ead6c667c386/testbench/handle_gzip.py#L29-L34 But that probably won't work with `uploadType=multipart`, where the `Content-Type` is `multipart/related;`. I think we need to decompress the results on the fly, and disable decompression if the `Accept-Encoding: gzip` flag is enabled. We also need to store the checksums of the compressed data, which suggests (again) that decompressing the request is not the right idea.
non_defect
review implementation of decompressive transcoding i am not sure the implementation of decompressive transcoding works as it should it seems to decompress the data when uploaded but that probably won t work with uploadtype multipart where the content type is multipart related i think we need to decompress the results on the fly and disable decompression if the accept encoding gzip flag is enabled we also need to store the checksums of the compressed data which suggests again that decompressing the request is not the right idea
0
119,601
12,035,381,900
IssuesEvent
2020-04-13 17:47:07
roc-project/roc-java
https://api.github.com/repos/roc-project/roc-java
opened
Build and publish HTML docs
documentation
Automatically build and publish javadoc documentation. Are there any service allowing to do this on push? Or we can use travis + github pages, as we do in https://github.com/roc-project/roc.
1.0
Build and publish HTML docs - Automatically build and publish javadoc documentation. Are there any service allowing to do this on push? Or we can use travis + github pages, as we do in https://github.com/roc-project/roc.
non_defect
build and publish html docs automatically build and publish javadoc documentation are there any service allowing to do this on push or we can use travis github pages as we do in
0
234,375
17,952,810,722
IssuesEvent
2021-09-13 01:10:50
UnBArqDsw2021-1/2021.1_G6_Curumim
https://api.github.com/repos/UnBArqDsw2021-1/2021.1_G6_Curumim
closed
Guia de estilo
documentation
### Descrição: Issue direcionada para a criação do guia de estilo da aplicação. ### Tarefas: - [ ] Criar guia de estilo; - [ ] Todos os integrantes pontuarem a issue. ### Critérios de aceitação: - [ ] Guia de estilo criado; - [ ] Issue pontuada.
1.0
Guia de estilo - ### Descrição: Issue direcionada para a criação do guia de estilo da aplicação. ### Tarefas: - [ ] Criar guia de estilo; - [ ] Todos os integrantes pontuarem a issue. ### Critérios de aceitação: - [ ] Guia de estilo criado; - [ ] Issue pontuada.
non_defect
guia de estilo descrição issue direcionada para a criação do guia de estilo da aplicação tarefas criar guia de estilo todos os integrantes pontuarem a issue critérios de aceitação guia de estilo criado issue pontuada
0
10,962
2,622,855,874
IssuesEvent
2015-03-04 08:07:45
max99x/dict-lookup-chrome-ext
https://api.github.com/repos/max99x/dict-lookup-chrome-ext
closed
Can't get this extension to work at all.
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Better instruction? 2. Maybe a walk-thru of an example 3. I don't know. What is the expected output? What do you see instead? I expected the popup as described on the home page with a definition ect... I got absolutely nothing. What version of the product are you using? On what operating system? Version 4.0 Windows XP-SP3 Please provide any additional information below. Seems to install correctly but no functionality, Is there supposed to be a toolbar icon? Does it matter that my mouse is set up for single click select?,(instead of double clicking)? ``` Original issue reported on code.google.com by `scritt...@gmail.com` on 16 Mar 2010 at 11:21
1.0
Can't get this extension to work at all. - ``` What steps will reproduce the problem? 1. Better instruction? 2. Maybe a walk-thru of an example 3. I don't know. What is the expected output? What do you see instead? I expected the popup as described on the home page with a definition ect... I got absolutely nothing. What version of the product are you using? On what operating system? Version 4.0 Windows XP-SP3 Please provide any additional information below. Seems to install correctly but no functionality, Is there supposed to be a toolbar icon? Does it matter that my mouse is set up for single click select?,(instead of double clicking)? ``` Original issue reported on code.google.com by `scritt...@gmail.com` on 16 Mar 2010 at 11:21
defect
can t get this extension to work at all what steps will reproduce the problem better instruction maybe a walk thru of an example i don t know what is the expected output what do you see instead i expected the popup as described on the home page with a definition ect i got absolutely nothing what version of the product are you using on what operating system version windows xp please provide any additional information below seems to install correctly but no functionality is there supposed to be a toolbar icon does it matter that my mouse is set up for single click select instead of double clicking original issue reported on code google com by scritt gmail com on mar at
1
60,473
17,023,435,554
IssuesEvent
2021-07-03 02:01:15
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Rendering inconsistency with access tag on roads passing under bridges
Component: mapnik Priority: minor Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 2.59am, Saturday, 4th July 2009]** Example location which illustrates the problem: http://www.openstreetmap.org/?lat=34.282078&lon=-119.315806&zoom=18&layers=B000FTF A service road (Park Access Rd) with access=permissive crosses underneath a motorway bridge. The bridge is at layer=1, the service road at the default layer. The service road itself is properly rendered underneath the bridge, however the green 'dots' which denote the access type seem to float over the motorway. The same problem also occurs with access set to private or destination (with red and blue dots, respectively).
1.0
Rendering inconsistency with access tag on roads passing under bridges - **[Submitted to the original trac issue database at 2.59am, Saturday, 4th July 2009]** Example location which illustrates the problem: http://www.openstreetmap.org/?lat=34.282078&lon=-119.315806&zoom=18&layers=B000FTF A service road (Park Access Rd) with access=permissive crosses underneath a motorway bridge. The bridge is at layer=1, the service road at the default layer. The service road itself is properly rendered underneath the bridge, however the green 'dots' which denote the access type seem to float over the motorway. The same problem also occurs with access set to private or destination (with red and blue dots, respectively).
defect
rendering inconsistency with access tag on roads passing under bridges example location which illustrates the problem a service road park access rd with access permissive crosses underneath a motorway bridge the bridge is at layer the service road at the default layer the service road itself is properly rendered underneath the bridge however the green dots which denote the access type seem to float over the motorway the same problem also occurs with access set to private or destination with red and blue dots respectively
1
160,201
12,505,811,153
IssuesEvent
2020-06-02 11:26:04
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
opened
Use of possibly insecure function - consider using safer ast., /opt/ros_melodic_ws/src/qt_gui_core/qt_gui/src/qt_gui/perspective_manager.py:429
bandit bug static analysis testing triage
```yaml { "id": 1, "title": "Use of possibly insecure function - consider using safer ast., /opt/ros_melodic_ws/src/qt_gui_core/qt_gui/src/qt_gui/perspective_manager.py:429", "type": "bug", "description": "HIGH confidence of MEDIUM severity bug. Use of possibly insecure function - consider using safer ast.literal_eval. at /opt/ros_melodic_ws/src/qt_gui_core/qt_gui/src/qt_gui/perspective_manager.py:429 See links for more info on the bug.", "cwe": "None", "cve": "None", "keywords": [ "bandit", "bug", "static analysis", "testing", "triage", "bug" ], "system": "", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "subject-specific", "architectural-location": "application-specific", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-06-02 (11:25)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-06-02 (11:25)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "/opt/ros_melodic_ws/src/qt_gui_core/qt_gui/src/qt_gui/perspective_manager.py:429", "reproduction": "See artifacts below (if available)", "reproduction-image": "" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
1.0
Use of possibly insecure function - consider using safer ast., /opt/ros_melodic_ws/src/qt_gui_core/qt_gui/src/qt_gui/perspective_manager.py:429 - ```yaml { "id": 1, "title": "Use of possibly insecure function - consider using safer ast., /opt/ros_melodic_ws/src/qt_gui_core/qt_gui/src/qt_gui/perspective_manager.py:429", "type": "bug", "description": "HIGH confidence of MEDIUM severity bug. Use of possibly insecure function - consider using safer ast.literal_eval. at /opt/ros_melodic_ws/src/qt_gui_core/qt_gui/src/qt_gui/perspective_manager.py:429 See links for more info on the bug.", "cwe": "None", "cve": "None", "keywords": [ "bandit", "bug", "static analysis", "testing", "triage", "bug" ], "system": "", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "subject-specific", "architectural-location": "application-specific", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-06-02 (11:25)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-06-02 (11:25)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "/opt/ros_melodic_ws/src/qt_gui_core/qt_gui/src/qt_gui/perspective_manager.py:429", "reproduction": "See artifacts below (if available)", "reproduction-image": "" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
non_defect
use of possibly insecure function consider using safer ast opt ros melodic ws src qt gui core qt gui src qt gui perspective manager py yaml id title use of possibly insecure function consider using safer ast opt ros melodic ws src qt gui core qt gui src qt gui perspective manager py type bug description high confidence of medium severity bug use of possibly insecure function consider using safer ast literal eval at opt ros melodic ws src qt gui core qt gui src qt gui perspective manager py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace opt ros melodic ws src qt gui core qt gui src qt gui perspective manager py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
0
52,177
13,211,396,891
IssuesEvent
2020-08-15 22:50:46
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
[icetray] Assigned value is garbage or undefined (Trac #1798)
Incomplete Migration Migrated from Trac combo core defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1798">https://code.icecube.wisc.edu/projects/icecube/ticket/1798</a>, reported by kjmeagherand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2016-08-02T16:06:21", "_ts": "1470153981858316", "description": "found by static analysis: http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-f15af5.html#EndPath", "reporter": "kjmeagher", "cc": "", "resolution": "invalid", "time": "2016-07-27T08:00:22", "component": "combo core", "summary": "[icetray] Assigned value is garbage or undefined", "priority": "normal", "keywords": "", "milestone": "Long-Term Future", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
[icetray] Assigned value is garbage or undefined (Trac #1798) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1798">https://code.icecube.wisc.edu/projects/icecube/ticket/1798</a>, reported by kjmeagherand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2016-08-02T16:06:21", "_ts": "1470153981858316", "description": "found by static analysis: http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-f15af5.html#EndPath", "reporter": "kjmeagher", "cc": "", "resolution": "invalid", "time": "2016-07-27T08:00:22", "component": "combo core", "summary": "[icetray] Assigned value is garbage or undefined", "priority": "normal", "keywords": "", "milestone": "Long-Term Future", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
assigned value is garbage or undefined trac migrated from json status closed changetime ts description found by static analysis reporter kjmeagher cc resolution invalid time component combo core summary assigned value is garbage or undefined priority normal keywords milestone long term future owner olivas type defect
1
47,073
13,056,027,463
IssuesEvent
2020-07-30 03:26:01
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
[filterscripts] grecofilter crashes with Python3 (Trac #2391)
Incomplete Migration Migrated from Trac defect jeb + pnf
Migrated from https://code.icecube.wisc.edu/ticket/2391 ```json { "status": "closed", "changetime": "2020-06-24T12:31:42", "description": "I'm getting the following error and a throw:\n\n{{{\nTraceback (most recent call last):\n File \"/home/olivas/icecube/combo/trunk/build/lib/icecube/filterscripts/grecofilter.py\", line 174, in FirstHit\n x, y, z, t, q = grecovariables.GetHitInformation(frame['I3Geometry'], hits, 1)\n File \"/home/olivas/icecube/combo/trunk/build/lib/icecube/filterscripts/grecovariables.py\", line 69, in GetHitInformation\n times, charges = np.array(map(lambda pulse: [pulse.time, pulse.charge], pulses)).T\nTypeError: iteration over a 0-d array\n\nThe above exception was the direct cause of the following exception:\n\nSystemError: <built-in method get of dict object at 0x7fbeaa2c5dc8> returned a result with an error set\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/home/olivas/icecube/combo/trunk/build/filterscripts/resources/scripts/SimulationFiltering.py\", line 395, in <module>\n main(opts)\n File \"/home/olivas/icecube/combo/trunk/build/filterscripts/resources/scripts/SimulationFiltering.py\", line 352, in main\n tray.Execute()\n File \"/home/olivas/icecube/combo/trunk/build/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\nSystemError: <built-in method get of dict object at 0x7fbeaa2c5dc8> returned a result with an error set\n}}}\n\nSetting the milestone for the Spring release, since we'll likely not be able to support Python3 with this release.", "reporter": "olivas", "cc": "mjlarson@umd.edu", "resolution": "fixed", "_ts": "1593001902142004", "component": "jeb + pnf", "summary": "[filterscripts] grecofilter crashes with Python3", "priority": "blocker", "keywords": "", "time": "2019-12-19T04:55:52", "milestone": "Autumnal Equinox 2020", "owner": "blaufuss", "type": "defect" } ```
1.0
[filterscripts] grecofilter crashes with Python3 (Trac #2391) - Migrated from https://code.icecube.wisc.edu/ticket/2391 ```json { "status": "closed", "changetime": "2020-06-24T12:31:42", "description": "I'm getting the following error and a throw:\n\n{{{\nTraceback (most recent call last):\n File \"/home/olivas/icecube/combo/trunk/build/lib/icecube/filterscripts/grecofilter.py\", line 174, in FirstHit\n x, y, z, t, q = grecovariables.GetHitInformation(frame['I3Geometry'], hits, 1)\n File \"/home/olivas/icecube/combo/trunk/build/lib/icecube/filterscripts/grecovariables.py\", line 69, in GetHitInformation\n times, charges = np.array(map(lambda pulse: [pulse.time, pulse.charge], pulses)).T\nTypeError: iteration over a 0-d array\n\nThe above exception was the direct cause of the following exception:\n\nSystemError: <built-in method get of dict object at 0x7fbeaa2c5dc8> returned a result with an error set\n\nThe above exception was the direct cause of the following exception:\n\nTraceback (most recent call last):\n File \"/home/olivas/icecube/combo/trunk/build/filterscripts/resources/scripts/SimulationFiltering.py\", line 395, in <module>\n main(opts)\n File \"/home/olivas/icecube/combo/trunk/build/filterscripts/resources/scripts/SimulationFiltering.py\", line 352, in main\n tray.Execute()\n File \"/home/olivas/icecube/combo/trunk/build/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\nSystemError: <built-in method get of dict object at 0x7fbeaa2c5dc8> returned a result with an error set\n}}}\n\nSetting the milestone for the Spring release, since we'll likely not be able to support Python3 with this release.", "reporter": "olivas", "cc": "mjlarson@umd.edu", "resolution": "fixed", "_ts": "1593001902142004", "component": "jeb + pnf", "summary": "[filterscripts] grecofilter crashes with Python3", "priority": "blocker", "keywords": "", "time": "2019-12-19T04:55:52", "milestone": "Autumnal Equinox 2020", "owner": "blaufuss", "type": "defect" } ```
defect
grecofilter crashes with trac migrated from json status closed changetime description i m getting the following error and a throw n n ntraceback most recent call last n file home olivas icecube combo trunk build lib icecube filterscripts grecofilter py line in firsthit n x y z t q grecovariables gethitinformation frame hits n file home olivas icecube combo trunk build lib icecube filterscripts grecovariables py line in gethitinformation n times charges np array map lambda pulse pulses t ntypeerror iteration over a d array n nthe above exception was the direct cause of the following exception n nsystemerror returned a result with an error set n nthe above exception was the direct cause of the following exception n ntraceback most recent call last n file home olivas icecube combo trunk build filterscripts resources scripts simulationfiltering py line in n main opts n file home olivas icecube combo trunk build filterscripts resources scripts simulationfiltering py line in main n tray execute n file home olivas icecube combo trunk build lib py line in execute n super self execute nsystemerror returned a result with an error set n n nsetting the milestone for the spring release since we ll likely not be able to support with this release reporter olivas cc mjlarson umd edu resolution fixed ts component jeb pnf summary grecofilter crashes with priority blocker keywords time milestone autumnal equinox owner blaufuss type defect
1
664,446
22,270,301,362
IssuesEvent
2022-06-10 11:39:05
ignite-hq/cli
https://api.github.com/repos/ignite-hq/cli
opened
ci: investigate reliability of integration test workflows
bug priority/high CI&CD
In a previous commit tests do not pass but in the last commit they do, see: ![image](https://user-images.githubusercontent.com/6023117/173056714-38fa3f2f-c8aa-4e07-a658-d75215c4f847.png) Looks like in the last commit tests are skipped, maybe because cached, but marked as green, see the full logs: <img width="1153" alt="image" src="https://user-images.githubusercontent.com/6023117/173056810-1886319e-e651-4bae-aab5-1774dceaa3b9.png"> There is a possible error in the git diff step, it just skips if the files aren't modified but does not inherit the error from a previous run, just marks as green: <img width="1153" alt="image" src="https://user-images.githubusercontent.com/6023117/173056956-26fa4050-bffb-459f-bf67-b66e353b8e61.png"> https://github.com/ignite-hq/cli/runs/6821675345?check_suite_focus=true
1.0
ci: investigate reliability of integration test workflows - In a previous commit tests do not pass but in the last commit they do, see: ![image](https://user-images.githubusercontent.com/6023117/173056714-38fa3f2f-c8aa-4e07-a658-d75215c4f847.png) Looks like in the last commit tests are skipped, maybe because cached, but marked as green, see the full logs: <img width="1153" alt="image" src="https://user-images.githubusercontent.com/6023117/173056810-1886319e-e651-4bae-aab5-1774dceaa3b9.png"> There is a possible error in the git diff step, it just skips if the files aren't modified but does not inherit the error from a previous run, just marks as green: <img width="1153" alt="image" src="https://user-images.githubusercontent.com/6023117/173056956-26fa4050-bffb-459f-bf67-b66e353b8e61.png"> https://github.com/ignite-hq/cli/runs/6821675345?check_suite_focus=true
non_defect
ci investigate reliability of integration test workflows in a previous commit tests do not pass but in the last commit they do see looks like in the last commit tests are skipped maybe because cached but marked as green see the full logs img width alt image src there is a possible error in the git diff step it just skips if the files aren t modified but does not inherit the error from a previous run just marks as green img width alt image src
0
39,217
8,601,639,980
IssuesEvent
2018-11-16 11:30:46
mozilla/addons-frontend
https://api.github.com/repos/mozilla/addons-frontend
opened
Including .css files with @import is non-standard behaviour
component: code quality priority: p3 qa: not needed
Let's fix this warning: > Including .css files with @import is non-standard behaviour which will be removed in future versions of LibSass. > Use a custom importer to maintain this behaviour. Check your implementations documentation on how to create a custom importer. It happens when running: - `yarn storybook` - `yarn build`
1.0
Including .css files with @import is non-standard behaviour - Let's fix this warning: > Including .css files with @import is non-standard behaviour which will be removed in future versions of LibSass. > Use a custom importer to maintain this behaviour. Check your implementations documentation on how to create a custom importer. It happens when running: - `yarn storybook` - `yarn build`
non_defect
including css files with import is non standard behaviour let s fix this warning including css files with import is non standard behaviour which will be removed in future versions of libsass use a custom importer to maintain this behaviour check your implementations documentation on how to create a custom importer it happens when running yarn storybook yarn build
0
68,522
21,683,838,171
IssuesEvent
2022-05-09 09:19:47
primefaces/primereact
https://api.github.com/repos/primefaces/primereact
closed
InputNumber: On mobile, the focus on input makes using increment / decrement buttons difficult
defect mobile
### Describe the bug 1. Open InputNumber on mobile. https://www.primefaces.org/primereact/inputnumber/ 2. Click on the increment / decrement buttons. Button press causes the input box to focus causing the mobile keyboard to be shown and hidden repeatedly on multiple clicks. Also, mobile will zoom in and make using the button difficult. Also, buttons don’t use the spinner on mobile. ### Reproducer https://www.primefaces.org/primereact/inputnumber/ ### PrimeReact version 8.0.1 ### React version 18.x ### Language ES6 ### Build / Runtime Next.js ### Browser(s) iOS 15.4 ### Steps to reproduce the behavior 1. Open InputNumber on mobile. https://www.primefaces.org/primereact/inputnumber/ 2. Click on the increment / decrement buttons. ### Expected behavior Using buttons on mobile shouldn't make the keyboard appear but should change the value.
1.0
InputNumber: On mobile, the focus on input makes using increment / decrement buttons difficult - ### Describe the bug 1. Open InputNumber on mobile. https://www.primefaces.org/primereact/inputnumber/ 2. Click on the increment / decrement buttons. Button press causes the input box to focus causing the mobile keyboard to be shown and hidden repeatedly on multiple clicks. Also, mobile will zoom in and make using the button difficult. Also, buttons don’t use the spinner on mobile. ### Reproducer https://www.primefaces.org/primereact/inputnumber/ ### PrimeReact version 8.0.1 ### React version 18.x ### Language ES6 ### Build / Runtime Next.js ### Browser(s) iOS 15.4 ### Steps to reproduce the behavior 1. Open InputNumber on mobile. https://www.primefaces.org/primereact/inputnumber/ 2. Click on the increment / decrement buttons. ### Expected behavior Using buttons on mobile shouldn't make the keyboard appear but should change the value.
defect
inputnumber on mobile the focus on input makes using increment decrement buttons difficult describe the bug open inputnumber on mobile click on the increment decrement buttons button press causes the input box to focus causing the mobile keyboard to be shown and hidden repeatedly on multiple clicks also mobile will zoom in and make using the button difficult also buttons don’t use the spinner on mobile reproducer primereact version react version x language build runtime next js browser s ios steps to reproduce the behavior open inputnumber on mobile click on the increment decrement buttons expected behavior using buttons on mobile shouldn t make the keyboard appear but should change the value
1
3,987
2,610,085,330
IssuesEvent
2015-02-26 18:25:57
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
深圳长豆豆怎么办
auto-migrated Priority-Medium Type-Defect
``` 深圳长豆豆怎么办【深圳韩方科颜全国热线400-869-1818,24小时 QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘�� �——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方� ��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健 康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业�� �疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘� ��。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:04
1.0
深圳长豆豆怎么办 - ``` 深圳长豆豆怎么办【深圳韩方科颜全国热线400-869-1818,24小时 QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘�� �——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方� ��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健 康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业�� �疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘� ��。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:04
defect
深圳长豆豆怎么办 深圳长豆豆怎么办【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘�� �——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方� ��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健 康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业�� �疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘� ��。 original issue reported on code google com by szft com on may at
1
71,118
23,455,274,383
IssuesEvent
2022-08-16 08:19:33
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Some slash-commands like /myroomnick not working in „pre-DM“
T-Defect
### Steps to reproduce 1. Start a DM with someone, **do not** send the first message 2. Run a slash-command like `/myroomnick` Other commands are also affected. ### Outcome #### What did you expect? Some commands could be executed on room creation. For instance `/myroomnick` totally makes sense. Others should not be available. Example for that would be `/addwidget`. #### What happened instead? Server error - unknown room ### Operating system Ubuntu 22.04.1 LTS ### Browser information Firefox 103.0.2 ### URL for webapp https://develop.element.io/ ### Application version Element version: 39eee10c576f-react-0dffc582e7e5-js-3f6f5b69c7a1 Olm version: 3.2.12 ### Homeserver matrix.org ### Will you send logs? No
1.0
Some slash-commands like /myroomnick not working in „pre-DM“ - ### Steps to reproduce 1. Start a DM with someone, **do not** send the first message 2. Run a slash-command like `/myroomnick` Other commands are also affected. ### Outcome #### What did you expect? Some commands could be executed on room creation. For instance `/myroomnick` totally makes sense. Others should not be available. Example for that would be `/addwidget`. #### What happened instead? Server error - unknown room ### Operating system Ubuntu 22.04.1 LTS ### Browser information Firefox 103.0.2 ### URL for webapp https://develop.element.io/ ### Application version Element version: 39eee10c576f-react-0dffc582e7e5-js-3f6f5b69c7a1 Olm version: 3.2.12 ### Homeserver matrix.org ### Will you send logs? No
defect
some slash commands like myroomnick not working in „pre dm“ steps to reproduce start a dm with someone do not send the first message run a slash command like myroomnick other commands are also affected outcome what did you expect some commands could be executed on room creation for instance myroomnick totally makes sense others should not be available example for that would be addwidget what happened instead server error unknown room operating system ubuntu lts browser information firefox url for webapp application version element version react js olm version homeserver matrix org will you send logs no
1
25,915
4,506,671,593
IssuesEvent
2016-09-02 05:34:57
Cockatrice/Cockatrice
https://api.github.com/repos/Cockatrice/Cockatrice
closed
Add Deck Name and Comments to Save to/Load from clipboard functionality
App - Cockatrice Defect - Basic
Saving the deck to clipboard results in the main deck and the sideboard being copied, however we lose the deck name and comments in this process.
1.0
Add Deck Name and Comments to Save to/Load from clipboard functionality - Saving the deck to clipboard results in the main deck and the sideboard being copied, however we lose the deck name and comments in this process.
defect
add deck name and comments to save to load from clipboard functionality saving the deck to clipboard results in the main deck and the sideboard being copied however we lose the deck name and comments in this process
1
6,707
6,599,523,258
IssuesEvent
2017-09-16 20:53:01
AndreyLysenkov/dismoodicon
https://api.github.com/repos/AndreyLysenkov/dismoodicon
opened
add check update feature
#feature #security [info gathering]
- [ ] [alpha] every hour/day/week check git repository on commits; - [ ] [debug] every hour/day/week check git repository on any new release; - [ ] [stable] every hour/day/week check git repository on not-preview new release; - [ ] on default check stable releases; - [ ] store bot owner id in `config.json`; - [ ] on new release, write message to `bot owner`; - [ ] add update script; - [ ] add command to manage this; - [ ] show that command only to `bot owner`; - [ ] unsubscribe from notifications; - [ ] change update brunch; - [ ] run update on command;
True
add check update feature - - [ ] [alpha] every hour/day/week check git repository on commits; - [ ] [debug] every hour/day/week check git repository on any new release; - [ ] [stable] every hour/day/week check git repository on not-preview new release; - [ ] on default check stable releases; - [ ] store bot owner id in `config.json`; - [ ] on new release, write message to `bot owner`; - [ ] add update script; - [ ] add command to manage this; - [ ] show that command only to `bot owner`; - [ ] unsubscribe from notifications; - [ ] change update brunch; - [ ] run update on command;
non_defect
add check update feature every hour day week check git repository on commits every hour day week check git repository on any new release every hour day week check git repository on not preview new release on default check stable releases store bot owner id in config json on new release write message to bot owner add update script add command to manage this show that command only to bot owner unsubscribe from notifications change update brunch run update on command
0
1,736
2,603,970,414
IssuesEvent
2015-02-24 19:00:07
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳生殖疱疹能自愈吗
auto-migrated Priority-Medium Type-Defect
``` 沈阳生殖疱疹能自愈吗〓沈陽軍區政治部醫院性病〓TEL:024-3 1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。� ��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:28
1.0
沈阳生殖疱疹能自愈吗 - ``` 沈阳生殖疱疹能自愈吗〓沈陽軍區政治部醫院性病〓TEL:024-3 1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。� ��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:28
defect
沈阳生殖疱疹能自愈吗 沈阳生殖疱疹能自愈吗〓沈陽軍區政治部醫院性病〓tel: 〓 , 。� �� 。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 original issue reported on code google com by gmail com on jun at
1
62,810
26,161,940,276
IssuesEvent
2022-12-31 17:42:05
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
closed
Error: Provider produced inconsistent final plan for aws_ami_copy
bug service/ec2 stale
### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform CLI and Terraform AWS Provider Version ``` $ terraform -v Terraform v0.12.29 ``` but I'm pretty sure 0.13 / 0.14 wouldn't make a difference. <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> ### Affected Resource(s) * `aws_ami` ### Terraform Configuration Files ```hcl resource "aws_ami" "myami" { name = "myami2" virtualization_type = "hvm" root_device_name = "/dev/xvda" ebs_block_device { device_name = "/dev/xvda" snapshot_id = # pick a snapshot ID you have handy volume_size = 8 } } resource "aws_ami_copy" "myami-regional-eu-north-1" { name = aws_ami.myami.name description = aws_ami.myami.description source_ami_id = aws_ami.myami.id source_ami_region = var.region tags = aws_ami.myami.tags provider = aws.regional-eu-north-1 } ``` ### Debug Output https://gist.github.com/grahamc/f655eaed51d11e952833db39f33644f3 ### Panic Output <!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. ---> ### Expected Behavior Terraform should let me refer to the description and other related fields of aws_ami from aws_ami_copy. ### Actual Behavior ``` Error: Provider produced inconsistent final plan When expanding the plan for aws_ami_copy.myami-regional-eu-north-1 to include new values learned so far during apply, provider "registry.terraform.io/-/aws" produced an invalid new value for .description: was null, but now cty.StringVal(""). This is a bug in the provider, which should be reported in the provider's own issue tracker. ``` ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. `terraform apply` ### Important Factoids <!--- Are there anything atypical about your accounts that we should know? For example: Running in EC2 Classic? ---> ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example: --->
1.0
Error: Provider produced inconsistent final plan for aws_ami_copy - ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform CLI and Terraform AWS Provider Version ``` $ terraform -v Terraform v0.12.29 ``` but I'm pretty sure 0.13 / 0.14 wouldn't make a difference. <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> ### Affected Resource(s) * `aws_ami` ### Terraform Configuration Files ```hcl resource "aws_ami" "myami" { name = "myami2" virtualization_type = "hvm" root_device_name = "/dev/xvda" ebs_block_device { device_name = "/dev/xvda" snapshot_id = # pick a snapshot ID you have handy volume_size = 8 } } resource "aws_ami_copy" "myami-regional-eu-north-1" { name = aws_ami.myami.name description = aws_ami.myami.description source_ami_id = aws_ami.myami.id source_ami_region = var.region tags = aws_ami.myami.tags provider = aws.regional-eu-north-1 } ``` ### Debug Output https://gist.github.com/grahamc/f655eaed51d11e952833db39f33644f3 ### Panic Output <!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. ---> ### Expected Behavior Terraform should let me refer to the description and other related fields of aws_ami from aws_ami_copy. ### Actual Behavior ``` Error: Provider produced inconsistent final plan When expanding the plan for aws_ami_copy.myami-regional-eu-north-1 to include new values learned so far during apply, provider "registry.terraform.io/-/aws" produced an invalid new value for .description: was null, but now cty.StringVal(""). This is a bug in the provider, which should be reported in the provider's own issue tracker. ``` ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. `terraform apply` ### Important Factoids <!--- Are there anything atypical about your accounts that we should know? For example: Running in EC2 Classic? ---> ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example: --->
non_defect
error provider produced inconsistent final plan for aws ami copy community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform cli and terraform aws provider version terraform v terraform but i m pretty sure wouldn t make a difference affected resource s aws ami terraform configuration files hcl resource aws ami myami name virtualization type hvm root device name dev xvda ebs block device device name dev xvda snapshot id pick a snapshot id you have handy volume size resource aws ami copy myami regional eu north name aws ami myami name description aws ami myami description source ami id aws ami myami id source ami region var region tags aws ami myami tags provider aws regional eu north debug output panic output expected behavior terraform should let me refer to the description and other related fields of aws ami from aws ami copy actual behavior error provider produced inconsistent final plan when expanding the plan for aws ami copy myami regional eu north to include new values learned so far during apply provider registry terraform io aws produced an invalid new value for description was null but now cty stringval this is a bug in the provider which should be reported in the provider s own issue tracker steps to reproduce terraform apply important factoids references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor documentation for example
0
6,184
2,610,222,882
IssuesEvent
2015-02-26 19:10:42
chrsmith/somefinders
https://api.github.com/repos/chrsmith/somefinders
opened
ludovico einaudi ноты .txt
auto-migrated Priority-Medium Type-Defect
``` '''Геронтий Уваров''' День добрый никак не могу найти .ludovico einaudi ноты .txt. как то выкладывали уже '''Вольфрам Волков''' Вот держи линк http://bit.ly/1aXejI6 '''Бернард Рябов''' Просит ввести номер мобилы!Не опасно ли это? '''Велор Александров''' Не это не влияет на баланс '''Альбин Павлов''' Не это не влияет на баланс Информация о файле: ludovico einaudi ноты .txt Загружен: В этом месяце Скачан раз: 931 Рейтинг: 686 Средняя скорость скачивания: 1234 Похожих файлов: 31 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 5:15
1.0
ludovico einaudi ноты .txt - ``` '''Геронтий Уваров''' День добрый никак не могу найти .ludovico einaudi ноты .txt. как то выкладывали уже '''Вольфрам Волков''' Вот держи линк http://bit.ly/1aXejI6 '''Бернард Рябов''' Просит ввести номер мобилы!Не опасно ли это? '''Велор Александров''' Не это не влияет на баланс '''Альбин Павлов''' Не это не влияет на баланс Информация о файле: ludovico einaudi ноты .txt Загружен: В этом месяце Скачан раз: 931 Рейтинг: 686 Средняя скорость скачивания: 1234 Похожих файлов: 31 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 5:15
defect
ludovico einaudi ноты txt геронтий уваров день добрый никак не могу найти ludovico einaudi ноты txt как то выкладывали уже вольфрам волков вот держи линк бернард рябов просит ввести номер мобилы не опасно ли это велор александров не это не влияет на баланс альбин павлов не это не влияет на баланс информация о файле ludovico einaudi ноты txt загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
1
25,688
4,417,713,565
IssuesEvent
2016-08-15 07:25:03
snowie2000/mactype
https://api.github.com/repos/snowie2000/mactype
closed
MacType not rendering in Microsoft Office 2013 and in Windows 8 Metro
auto-migrated Priority-Medium Type-Defect
``` What is the expected output? What do you see instead? MacType does not render in Microsoft Office 2013 and on all Metro apps as well as the Start Screen What version of the product are you using? On what operating system? 1.2012.1022.0 on Windows 8 Pro (x64) Please provide any additional information below. Perhaps the code on which the Start screen and Metro apps and MS Office 2013 are not the same as the other Windows applications? Here are some of the screenshots showing failure to render in Word 2010 (side by side with Google Chrome) Thank you for looking into the issue and I hope this gets fixed in the next release. I really love MacType ``` Original issue reported on code.google.com by `neljason...@gmail.com` on 13 Dec 2012 at 8:57 Attachments: * [Untitled.png](https://storage.googleapis.com/google-code-attachments/mactype/issue-49/comment-0/Untitled.png)
1.0
MacType not rendering in Microsoft Office 2013 and in Windows 8 Metro - ``` What is the expected output? What do you see instead? MacType does not render in Microsoft Office 2013 and on all Metro apps as well as the Start Screen What version of the product are you using? On what operating system? 1.2012.1022.0 on Windows 8 Pro (x64) Please provide any additional information below. Perhaps the code on which the Start screen and Metro apps and MS Office 2013 are not the same as the other Windows applications? Here are some of the screenshots showing failure to render in Word 2010 (side by side with Google Chrome) Thank you for looking into the issue and I hope this gets fixed in the next release. I really love MacType ``` Original issue reported on code.google.com by `neljason...@gmail.com` on 13 Dec 2012 at 8:57 Attachments: * [Untitled.png](https://storage.googleapis.com/google-code-attachments/mactype/issue-49/comment-0/Untitled.png)
defect
mactype not rendering in microsoft office and in windows metro what is the expected output what do you see instead mactype does not render in microsoft office and on all metro apps as well as the start screen what version of the product are you using on what operating system on windows pro please provide any additional information below perhaps the code on which the start screen and metro apps and ms office are not the same as the other windows applications here are some of the screenshots showing failure to render in word side by side with google chrome thank you for looking into the issue and i hope this gets fixed in the next release i really love mactype original issue reported on code google com by neljason gmail com on dec at attachments
1
26,768
4,778,467,639
IssuesEvent
2016-10-27 19:23:25
wheeler-microfluidics/microdrop
https://api.github.com/repos/wheeler-microfluidics/microdrop
closed
Plugin crashes when normalizing by area and no electrodes on (Trac #173)
defect dmf_control_board_plugin Migrated from Trac
This error occurs when viewing frequency-sweep data and clicking on the "normalize by area" checkbox: [ERROR]: wheelerlab.dmf_control_board plugin crashed processing on_experiment_log_selection_changed signal. Reason: Data has no positive values, and therefore can not be log-scaled. Migrated from http://microfluidics.utoronto.ca/microdrop/ticket/173 ```json { "status": "closed", "changetime": "2015-01-13T08:25:37", "description": "This error occurs when viewing frequency-sweep data and clicking on the \"normalize by area\" checkbox:\n\n[ERROR]: wheelerlab.dmf_control_board plugin crashed processing on_experiment_log_selection_changed signal.\nReason: Data has no positive values, and therefore can not be log-scaled.", "reporter": "ryan", "cc": "", "resolution": "fixed", "_ts": "1421137537638116", "component": "dmf_control_board_plugin", "summary": "Plugin crashes when normalizing by area and no electrodes on", "priority": "minor", "keywords": "", "version": "0.1", "time": "2015-01-13T08:21:25", "milestone": "Control board plugin 1.0", "owner": "", "type": "defect" } ```
1.0
Plugin crashes when normalizing by area and no electrodes on (Trac #173) - This error occurs when viewing frequency-sweep data and clicking on the "normalize by area" checkbox: [ERROR]: wheelerlab.dmf_control_board plugin crashed processing on_experiment_log_selection_changed signal. Reason: Data has no positive values, and therefore can not be log-scaled. Migrated from http://microfluidics.utoronto.ca/microdrop/ticket/173 ```json { "status": "closed", "changetime": "2015-01-13T08:25:37", "description": "This error occurs when viewing frequency-sweep data and clicking on the \"normalize by area\" checkbox:\n\n[ERROR]: wheelerlab.dmf_control_board plugin crashed processing on_experiment_log_selection_changed signal.\nReason: Data has no positive values, and therefore can not be log-scaled.", "reporter": "ryan", "cc": "", "resolution": "fixed", "_ts": "1421137537638116", "component": "dmf_control_board_plugin", "summary": "Plugin crashes when normalizing by area and no electrodes on", "priority": "minor", "keywords": "", "version": "0.1", "time": "2015-01-13T08:21:25", "milestone": "Control board plugin 1.0", "owner": "", "type": "defect" } ```
defect
plugin crashes when normalizing by area and no electrodes on trac this error occurs when viewing frequency sweep data and clicking on the normalize by area checkbox wheelerlab dmf control board plugin crashed processing on experiment log selection changed signal reason data has no positive values and therefore can not be log scaled migrated from json status closed changetime description this error occurs when viewing frequency sweep data and clicking on the normalize by area checkbox n n wheelerlab dmf control board plugin crashed processing on experiment log selection changed signal nreason data has no positive values and therefore can not be log scaled reporter ryan cc resolution fixed ts component dmf control board plugin summary plugin crashes when normalizing by area and no electrodes on priority minor keywords version time milestone control board plugin owner type defect
1
40,848
10,188,432,107
IssuesEvent
2019-08-11 10:45:42
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
SuperLU is too slow
defect scipy.sparse.linalg
I have noticed that spsolve hangs during solve. Here are my system setting: OS: mac 10.9.4 Python version: 2.7.6 Scipy version: 0.14.0 I have created a simple test script to illustrate the bug: https://dl.dropboxusercontent.com/u/29899857/scipy_test.zip With scipy 0.13, the script finishes within seconds: ``` python > ./scipy_test.py Before solve After solve 545.636841049 ``` However, with scipy 0.14, it hangs after printing "Before solve". I have tried other matrices as well, besides the simple matrices (diagonal matrix), scipy hangs every time. I have also observed this problem on Linux systems.
1.0
SuperLU is too slow - I have noticed that spsolve hangs during solve. Here are my system setting: OS: mac 10.9.4 Python version: 2.7.6 Scipy version: 0.14.0 I have created a simple test script to illustrate the bug: https://dl.dropboxusercontent.com/u/29899857/scipy_test.zip With scipy 0.13, the script finishes within seconds: ``` python > ./scipy_test.py Before solve After solve 545.636841049 ``` However, with scipy 0.14, it hangs after printing "Before solve". I have tried other matrices as well, besides the simple matrices (diagonal matrix), scipy hangs every time. I have also observed this problem on Linux systems.
defect
superlu is too slow i have noticed that spsolve hangs during solve here are my system setting os mac python version scipy version i have created a simple test script to illustrate the bug with scipy the script finishes within seconds python scipy test py before solve after solve however with scipy it hangs after printing before solve i have tried other matrices as well besides the simple matrices diagonal matrix scipy hangs every time i have also observed this problem on linux systems
1
89,154
25,597,560,504
IssuesEvent
2022-12-01 17:18:47
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
Build of extra-utils (in stage-1) fails
0.kind: build failure
### Steps To Reproduce Steps to reproduce the behavior: 1. build *X* ### Build log Full log available at https://gist.github.com/mhwombat/d2ec165227b1d0a6adb20f29dbb3b6c3 ``` patching /nix/store/g13rzqqimgclvzb1414l8877k4qa7c9d-extra-utils/lib/libext2fs.so.2... patchelf: missing ELF header error: builder for '/nix/store/8ryi6wva8sycx228dzdg938ir50ivmp6-extra-utils.drv' failed with exit code 1 error: 1 dependencies of derivation '/nix/store/xyfzs6b6hv4gpgc1glpw69pdrch6pgiy-stage-1-init.sh.drv' failed to build error: 1 dependencies of derivation '/nix/store/s8pg4z5a0s4hr7aq83vj15dipi69n7pl-initrd-linux-5.15.79.drv' failed to build error: 1 dependencies of derivation '/nix/store/anv9387fbpy95pkpjkw7idsgrjvnd69h-nixos-system-wombat11k-23.05pre429868.a115bb9bd56.drv' failed to build ``` ### Additional context ### Notify maintainers @pennae @klemensn @trofi ### Metadata ```console ❯ nix-shell -p nix-info --run "nix-info -m" - system: `"x86_64-linux"` - host os: `Linux 5.15.76, NixOS, 22.11 (Raccoon), 22.11pre423448.a2a777538d9` - multi-user?: `yes` - sandbox: `yes` - version: `nix-env (Nix) 2.11.0` - channels(root): `"nixos"` - nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos` ```
1.0
Build of extra-utils (in stage-1) fails - ### Steps To Reproduce Steps to reproduce the behavior: 1. build *X* ### Build log Full log available at https://gist.github.com/mhwombat/d2ec165227b1d0a6adb20f29dbb3b6c3 ``` patching /nix/store/g13rzqqimgclvzb1414l8877k4qa7c9d-extra-utils/lib/libext2fs.so.2... patchelf: missing ELF header error: builder for '/nix/store/8ryi6wva8sycx228dzdg938ir50ivmp6-extra-utils.drv' failed with exit code 1 error: 1 dependencies of derivation '/nix/store/xyfzs6b6hv4gpgc1glpw69pdrch6pgiy-stage-1-init.sh.drv' failed to build error: 1 dependencies of derivation '/nix/store/s8pg4z5a0s4hr7aq83vj15dipi69n7pl-initrd-linux-5.15.79.drv' failed to build error: 1 dependencies of derivation '/nix/store/anv9387fbpy95pkpjkw7idsgrjvnd69h-nixos-system-wombat11k-23.05pre429868.a115bb9bd56.drv' failed to build ``` ### Additional context ### Notify maintainers @pennae @klemensn @trofi ### Metadata ```console ❯ nix-shell -p nix-info --run "nix-info -m" - system: `"x86_64-linux"` - host os: `Linux 5.15.76, NixOS, 22.11 (Raccoon), 22.11pre423448.a2a777538d9` - multi-user?: `yes` - sandbox: `yes` - version: `nix-env (Nix) 2.11.0` - channels(root): `"nixos"` - nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos` ```
non_defect
build of extra utils in stage fails steps to reproduce steps to reproduce the behavior build x build log full log available at patching nix store extra utils lib so patchelf missing elf header error builder for nix store extra utils drv failed with exit code error dependencies of derivation nix store stage init sh drv failed to build error dependencies of derivation nix store initrd linux drv failed to build error dependencies of derivation nix store nixos system drv failed to build additional context notify maintainers pennae klemensn trofi metadata console ❯ nix shell p nix info run nix info m system linux host os linux nixos raccoon multi user yes sandbox yes version nix env nix channels root nixos nixpkgs nix var nix profiles per user root channels nixos
0
4,513
2,610,111,523
IssuesEvent
2015-02-26 18:34:32
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
Cannot add a Posterous blog
auto-migrated Priority-Medium Type-Defect
``` What's the problem? I have a Posterous blog under a custom domain: http://stevecorneli.us. I tried to add my blog under this name and the default Posterous name: http://sscornelius.posterous.com. Neither name worked. I received this error message: Well, this is embarrassing... Sigh... ScribeFire couldn't get the information it needed about your blog. Helpfully, your blog returned this message: <html> <head><title>400 Bad Request</title></head> <body bgcolor="white"> <center><h1>400 Bad Request</h1></center> <hr><center>nginx/0.7.65</center> </body> </html> What version of ScribeFire for Chrome are you running? Chrome Alpha 2 ``` ----- Original issue reported on code.google.com by `sscornelius` on 17 Apr 2010 at 1:19
1.0
Cannot add a Posterous blog - ``` What's the problem? I have a Posterous blog under a custom domain: http://stevecorneli.us. I tried to add my blog under this name and the default Posterous name: http://sscornelius.posterous.com. Neither name worked. I received this error message: Well, this is embarrassing... Sigh... ScribeFire couldn't get the information it needed about your blog. Helpfully, your blog returned this message: <html> <head><title>400 Bad Request</title></head> <body bgcolor="white"> <center><h1>400 Bad Request</h1></center> <hr><center>nginx/0.7.65</center> </body> </html> What version of ScribeFire for Chrome are you running? Chrome Alpha 2 ``` ----- Original issue reported on code.google.com by `sscornelius` on 17 Apr 2010 at 1:19
defect
cannot add a posterous blog what s the problem i have a posterous blog under a custom domain i tried to add my blog under this name and the default posterous name neither name worked i received this error message well this is embarrassing sigh scribefire couldn t get the information it needed about your blog helpfully your blog returned this message bad request bad request nginx what version of scribefire for chrome are you running chrome alpha original issue reported on code google com by sscornelius on apr at
1
9,762
2,615,174,283
IssuesEvent
2015-03-01 06:57:13
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
WPS transaction failed (code: 0x02), Receive timeout occurred !
auto-migrated Priority-Triage Type-Defect
``` 1. What operating system are you using (Linux is the only supported OS)? # lsb_release -a Debian GNU/Linux Kali Linux 1.0 # uname -a Linux muriel 3.7-trunk-amd64 #1 SMP Debian 3.7.2-0+kali8 x86_64 GNU/Linux 2. Is your wireless card in monitor mode (yes/no)? # yes, Intel 3945ABG , and i've also tried switching to another instance in monitor mode interfaces (mon1,mon2). 3. What is the signal strength of the Access Point you are trying to crack? # 07:18:02 Ping (min/avg/max): 0.865ms/6.822ms/32.689ms Power: -80.45 07:18:02 29/30: 96% 4. What is the manufacturer and model # of the device you are trying to crack? # *Manufacturer: CIG *Model: CIG GPON ONT RG 5. What is the entire command line string you are supplying to reaver? Several, some examples ## $ reaver -i mon0 -b 74:31:70:EB:B4:F8 -vv --dh-small $ reaver -i mon0 -b 74:31:70:EB:B4:F8 -c 7 -a -S -N -vv --no-nacks --win7 --dh-small $ aireplay-ng -1 0 -a 74:31:70:EB:B4:F8 mon0 6. Please describe what you think the issue is. reaver prompts these messages repeatedly on every attempts , even when the traget was my own AP, which happens to be a few fets away .. [!] WPS transaction failed (code: 0x02), re-trying last pin [!] WARNING: Receive timeout occurred ## i just removed "<mac>.wpcs" in /etc/reaver/ .. i shall try again soon! 7. Paste the output from Reaver below. [+] Sending identity response [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M1 message [+] Sending WSC NACK [+] Sending WSC NACK [!] WPS transaction failed (code: 0x03), re-trying last pin [+] Nothing done, nothing to save. [+] 0.00% complete @ 2013-12-10 07:45:27 (0 seconds/pin) [+] Max time remaining at this rate: (undetermined) (11000 pins left to try) [+] Trying pin 12345670 [!] WARNING: Failed to associate with 74:31:70:EB:B4:F8 (ESSID: 7777777) [+] Sending EAPOL START request [!] WARNING: Receive timeout occurred [+] Sending identity response [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M1 message [+] Sending WSC NACK [+] Sending WSC NACK [!] WPS transaction failed (code: 0x03), re-trying last pin [+] Trying pin 12345670 [+] Sending EAPOL START request [!] WARNING: Receive timeout occurred [+] Sending EAPOL START request ^C [+] Nothing done, nothing to save. [+] Session saved. # and it kept repeating the same pin over & over again! ``` Original issue reported on code.google.com by `sary.sa@me.com` on 10 Dec 2013 at 2:17
1.0
WPS transaction failed (code: 0x02), Receive timeout occurred ! - ``` 1. What operating system are you using (Linux is the only supported OS)? # lsb_release -a Debian GNU/Linux Kali Linux 1.0 # uname -a Linux muriel 3.7-trunk-amd64 #1 SMP Debian 3.7.2-0+kali8 x86_64 GNU/Linux 2. Is your wireless card in monitor mode (yes/no)? # yes, Intel 3945ABG , and i've also tried switching to another instance in monitor mode interfaces (mon1,mon2). 3. What is the signal strength of the Access Point you are trying to crack? # 07:18:02 Ping (min/avg/max): 0.865ms/6.822ms/32.689ms Power: -80.45 07:18:02 29/30: 96% 4. What is the manufacturer and model # of the device you are trying to crack? # *Manufacturer: CIG *Model: CIG GPON ONT RG 5. What is the entire command line string you are supplying to reaver? Several, some examples ## $ reaver -i mon0 -b 74:31:70:EB:B4:F8 -vv --dh-small $ reaver -i mon0 -b 74:31:70:EB:B4:F8 -c 7 -a -S -N -vv --no-nacks --win7 --dh-small $ aireplay-ng -1 0 -a 74:31:70:EB:B4:F8 mon0 6. Please describe what you think the issue is. reaver prompts these messages repeatedly on every attempts , even when the traget was my own AP, which happens to be a few fets away .. [!] WPS transaction failed (code: 0x02), re-trying last pin [!] WARNING: Receive timeout occurred ## i just removed "<mac>.wpcs" in /etc/reaver/ .. i shall try again soon! 7. Paste the output from Reaver below. [+] Sending identity response [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M1 message [+] Sending WSC NACK [+] Sending WSC NACK [!] WPS transaction failed (code: 0x03), re-trying last pin [+] Nothing done, nothing to save. [+] 0.00% complete @ 2013-12-10 07:45:27 (0 seconds/pin) [+] Max time remaining at this rate: (undetermined) (11000 pins left to try) [+] Trying pin 12345670 [!] WARNING: Failed to associate with 74:31:70:EB:B4:F8 (ESSID: 7777777) [+] Sending EAPOL START request [!] WARNING: Receive timeout occurred [+] Sending identity response [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M1 message [+] Sending WSC NACK [+] Sending WSC NACK [!] WPS transaction failed (code: 0x03), re-trying last pin [+] Trying pin 12345670 [+] Sending EAPOL START request [!] WARNING: Receive timeout occurred [+] Sending EAPOL START request ^C [+] Nothing done, nothing to save. [+] Session saved. # and it kept repeating the same pin over & over again! ``` Original issue reported on code.google.com by `sary.sa@me.com` on 10 Dec 2013 at 2:17
defect
wps transaction failed code receive timeout occurred what operating system are you using linux is the only supported os lsb release a debian gnu linux kali linux uname a linux muriel trunk smp debian gnu linux is your wireless card in monitor mode yes no yes intel and i ve also tried switching to another instance in monitor mode interfaces what is the signal strength of the access point you are trying to crack ping min avg max power what is the manufacturer and model of the device you are trying to crack manufacturer cig model cig gpon ont rg what is the entire command line string you are supplying to reaver several some examples reaver i b eb vv dh small reaver i b eb c a s n vv no nacks dh small aireplay ng a eb please describe what you think the issue is reaver prompts these messages repeatedly on every attempts even when the traget was my own ap which happens to be a few fets away wps transaction failed code re trying last pin warning receive timeout occurred i just removed wpcs in etc reaver i shall try again soon paste the output from reaver below sending identity response received identity request sending identity response received message sending message received message sending wsc nack sending wsc nack wps transaction failed code re trying last pin nothing done nothing to save complete seconds pin max time remaining at this rate undetermined pins left to try trying pin warning failed to associate with eb essid sending eapol start request warning receive timeout occurred sending identity response received identity request sending identity response received message sending message received message sending wsc nack sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request warning receive timeout occurred sending eapol start request c nothing done nothing to save session saved and it kept repeating the same pin over over again original issue reported on code google com by sary sa me com on dec at
1
16,414
5,233,714,039
IssuesEvent
2017-01-30 13:46:57
SemsTestOrg/combinearchive-web
https://api.github.com/repos/SemsTestOrg/combinearchive-web
closed
"Edit" vs. "Rename" Button
code fixed migrated minor task
## Trac Ticket #125 **component:** code **owner:** somebody **reporter:** martinP **created:** 2015-09-22 12:37:52 **milestone:** **type:** task **version:** **keywords:** Think about changing the "Edit" Button into a "Rename" Button in some places. some user -> confused ## comment 1 **time:** 2015-10-06 10:34:47 **author:** martin copy ```edit` button to `rename` right behind the file name in case of `mouseover``` ## comment 2 **time:** 2015-10-06 10:35:41 **author:** martin Updated **_comment0** to **1444127769601742** ## comment 3 **time:** 2015-10-06 10:35:41 **author:** martin rename the ```edit` button to `rename``` in the archive editing section ## comment 4 **time:** 2015-10-06 14:33:38 **author:** mp487 <martin.peters3@uni-rostock.de> In [None](/8cacdf256a37619c4ad4b289d63428f743143b77): ```CommitTicketReference repository="" revision="8cacdf256a37619c4ad4b289d63428f743143b77" added hover 'rename' button after file name [see #125] ``` ## comment 5 **time:** 2015-10-06 14:34:30 **author:** martinP ## comment 6 **time:** 2015-10-06 14:34:30 **author:** martinP Updated **resolution** to **fixed** ## comment 7 **time:** 2015-10-06 14:34:30 **author:** martinP Updated **status** to **closed**
1.0
"Edit" vs. "Rename" Button - ## Trac Ticket #125 **component:** code **owner:** somebody **reporter:** martinP **created:** 2015-09-22 12:37:52 **milestone:** **type:** task **version:** **keywords:** Think about changing the "Edit" Button into a "Rename" Button in some places. some user -> confused ## comment 1 **time:** 2015-10-06 10:34:47 **author:** martin copy ```edit` button to `rename` right behind the file name in case of `mouseover``` ## comment 2 **time:** 2015-10-06 10:35:41 **author:** martin Updated **_comment0** to **1444127769601742** ## comment 3 **time:** 2015-10-06 10:35:41 **author:** martin rename the ```edit` button to `rename``` in the archive editing section ## comment 4 **time:** 2015-10-06 14:33:38 **author:** mp487 <martin.peters3@uni-rostock.de> In [None](/8cacdf256a37619c4ad4b289d63428f743143b77): ```CommitTicketReference repository="" revision="8cacdf256a37619c4ad4b289d63428f743143b77" added hover 'rename' button after file name [see #125] ``` ## comment 5 **time:** 2015-10-06 14:34:30 **author:** martinP ## comment 6 **time:** 2015-10-06 14:34:30 **author:** martinP Updated **resolution** to **fixed** ## comment 7 **time:** 2015-10-06 14:34:30 **author:** martinP Updated **status** to **closed**
non_defect
edit vs rename button trac ticket component code owner somebody reporter martinp created milestone type task version keywords think about changing the edit button into a rename button in some places some user confused comment time author martin copy edit button to rename right behind the file name in case of mouseover comment time author martin updated to comment time author martin rename the edit button to rename in the archive editing section comment time author in committicketreference repository revision added hover rename button after file name comment time author martinp comment time author martinp updated resolution to fixed comment time author martinp updated status to closed
0
47,657
6,062,213,708
IssuesEvent
2017-06-14 08:52:47
geetsisbac/WK2XQXCBCXIVMLBGXSPVU5EB
https://api.github.com/repos/geetsisbac/WK2XQXCBCXIVMLBGXSPVU5EB
reopened
LbJh6jFKkZ3p2/yrFYonZ8pTeE08GTHz+88CI0HhEizDuzxKEWTDN4Yvmh0ClvKqbrU9e14ChIMhgqjeYyXnwuKcYQEusq/vbHYgp7wGjAP6cIdLQhg+sNgy3/xSaMZ0Sq/qdRZl0+mYeBICkWVJtuAu41DuCBd7iBSULrNkLCI=
design
coLp1+lh2893JZacot947Qu6U8PQCG1VQcHmwYVKpxsz0sEs3G+M//xvUeapfrWfR0X/e/Od1MvZvKUhZKLbF7PNuT3clOXenp9dVK5+xx038SWMfY8aPHzIfGFUKmb8laKlDSKuo5F4BwytOsX8jlFAj90kVNbctLWWtjbOsbRRQI/dJFTW3LS1lrY2zrG0UUCP3SRU1ty0tZa2Ns6xtFFAj90kVNbctLWWtjbOsbTr5nBlibOUHzPj1kfirbZRpgJ7PgweO44pD831OGfMv1HNtR5gS7oZuq/Wq8nnJMjuITRYAe9FGkAtc5SbrBabUUCP3SRU1ty0tZa2Ns6xtFFAj90kVNbctLWWtjbOsbRRQI/dJFTW3LS1lrY2zrG0UUCP3SRU1ty0tZa2Ns6xtLga5SkzafnLJ2b0B3Z98qTsYYYDL3wRftU1k5HDKpQ7KaXii+znkPRI57ghXCoPAwpGP/32beWa5uhq9jgxzgRRQI/dJFTW3LS1lrY2zrG0UUCP3SRU1ty0tZa2Ns6xtFFAj90kVNbctLWWtjbOsbRRQI/dJFTW3LS1lrY2zrG081RwP+B3br3WLR98XLDbV4g9+PdqoVa/ntH1NaKIwHAX47UFFDNsOmd/0Mzv/P2FB24HgdzASM0AAOGEWMpVlsaM19IQly9oQxKjNXz7eCQFeb7ikxcJFhmC06W6L8D7UUCP3SRU1ty0tZa2Ns6xtFFAj90kVNbctLWWtjbOsbRRQI/dJFTW3LS1lrY2zrG0UUCP3SRU1ty0tZa2Ns6xtDMOcqzTKK/PsZHjsN8GIUlRQI/dJFTW3LS1lrY2zrG0UUCP3SRU1ty0tZa2Ns6xtFFAj90kVNbctLWWtjbOsbRRQI/dJFTW3LS1lrY2zrG0FtynX36ots2CIqZTCwH/0dmhAxxs/pGKSl5LT03Atk1lSOp8wAKmVNBNPAORuasc5DzWhYokXPc6eaDAfkAFOG36/dneV0l9UR+rAG34papctJmv+y7w//VJkna50sBrJfRiialS+5oRsuPyrSraP62eCkxAzo9Lz0ZjvKpbPZtAjxRhBnsMZ06espWo/WK2CzlgWqmVBwz49439TANmRTjQYx8KY1R9/8dWQMPndeqduBY55sHkh1BFl+6fQfzawKOIQ0UX+5v+tdVMNtifDiCeft7qrxPrnXREnD56udXq1s564d8aOcRyq6Gg1FMDrDlhaL/AdROMBQC+c5T5Y4HB4s+n+QdiZLgvcXr1d4KqEGRrSyQSu3QhPvQbgi3gYsIwucXcJg4/cmQZp+qTlk4pMUYbgmLyyQ6XS2RJk6AEx2DCfNrban9iHfrtGbX9lIJyq4O0uEThYDRCPtMm3BLRtWaHyADXF0T6v0sJHUeZymvL3pL0XENDZyI7GmJAWo9S7b6eNlX42rUgOJIlN7KrjoSokyk4kcHaeoxZ7LIfylUdfBbq69ZrAcvOjXsYBedthOBeIq/URZYVaHM5gGDvDnGU0UnyPoMwnb960WU5oe+8lN67BwpzsrLl39pG+vCPdNogR7WPeyBN/Apt4/z5UZspOnIn7LHaZVLZCLwIEeUEabzTUHSOperHO2GdpFELR7eqmhgjvn1hTkboXOjwM4JWZVbhmyzygHg5geKkfoRqa1QXLyWeeqCy+VUB14N6RPGCPqHRJZozRRPFE4yxapcW4Jf+j4m9JjKeufLMpbuhnXOXcJBOzACEmBnirWxU5bclGMWPVfG7VjAIOHScNHF2YZcxDt47GJptfIGL4BuhzksmxJx9koxdxwbo0atwq3k9S36gi7wlvbI7W1ugVlGRi16BTGQ/pV4lt6igs62FdFVK+M9P5ZvjLGwwQCSRlxw3GQUhmFSuYLfQbvFfDKYXUSC8HYZcJtG2eN3yk2Po6rAKY2ZVjacUDsZ9cjPMsxCJ9GLgSbmEaf8LdiW66NgmtRVWGyvUzVO5R1eCjqPpqwWouoZfdhNmN+38
1.0
LbJh6jFKkZ3p2/yrFYonZ8pTeE08GTHz+88CI0HhEizDuzxKEWTDN4Yvmh0ClvKqbrU9e14ChIMhgqjeYyXnwuKcYQEusq/vbHYgp7wGjAP6cIdLQhg+sNgy3/xSaMZ0Sq/qdRZl0+mYeBICkWVJtuAu41DuCBd7iBSULrNkLCI= - coLp1+lh2893JZacot947Qu6U8PQCG1VQcHmwYVKpxsz0sEs3G+M//xvUeapfrWfR0X/e/Od1MvZvKUhZKLbF7PNuT3clOXenp9dVK5+xx038SWMfY8aPHzIfGFUKmb8laKlDSKuo5F4BwytOsX8jlFAj90kVNbctLWWtjbOsbRRQI/dJFTW3LS1lrY2zrG0UUCP3SRU1ty0tZa2Ns6xtFFAj90kVNbctLWWtjbOsbTr5nBlibOUHzPj1kfirbZRpgJ7PgweO44pD831OGfMv1HNtR5gS7oZuq/Wq8nnJMjuITRYAe9FGkAtc5SbrBabUUCP3SRU1ty0tZa2Ns6xtFFAj90kVNbctLWWtjbOsbRRQI/dJFTW3LS1lrY2zrG0UUCP3SRU1ty0tZa2Ns6xtLga5SkzafnLJ2b0B3Z98qTsYYYDL3wRftU1k5HDKpQ7KaXii+znkPRI57ghXCoPAwpGP/32beWa5uhq9jgxzgRRQI/dJFTW3LS1lrY2zrG0UUCP3SRU1ty0tZa2Ns6xtFFAj90kVNbctLWWtjbOsbRRQI/dJFTW3LS1lrY2zrG081RwP+B3br3WLR98XLDbV4g9+PdqoVa/ntH1NaKIwHAX47UFFDNsOmd/0Mzv/P2FB24HgdzASM0AAOGEWMpVlsaM19IQly9oQxKjNXz7eCQFeb7ikxcJFhmC06W6L8D7UUCP3SRU1ty0tZa2Ns6xtFFAj90kVNbctLWWtjbOsbRRQI/dJFTW3LS1lrY2zrG0UUCP3SRU1ty0tZa2Ns6xtDMOcqzTKK/PsZHjsN8GIUlRQI/dJFTW3LS1lrY2zrG0UUCP3SRU1ty0tZa2Ns6xtFFAj90kVNbctLWWtjbOsbRRQI/dJFTW3LS1lrY2zrG0FtynX36ots2CIqZTCwH/0dmhAxxs/pGKSl5LT03Atk1lSOp8wAKmVNBNPAORuasc5DzWhYokXPc6eaDAfkAFOG36/dneV0l9UR+rAG34papctJmv+y7w//VJkna50sBrJfRiialS+5oRsuPyrSraP62eCkxAzo9Lz0ZjvKpbPZtAjxRhBnsMZ06espWo/WK2CzlgWqmVBwz49439TANmRTjQYx8KY1R9/8dWQMPndeqduBY55sHkh1BFl+6fQfzawKOIQ0UX+5v+tdVMNtifDiCeft7qrxPrnXREnD56udXq1s564d8aOcRyq6Gg1FMDrDlhaL/AdROMBQC+c5T5Y4HB4s+n+QdiZLgvcXr1d4KqEGRrSyQSu3QhPvQbgi3gYsIwucXcJg4/cmQZp+qTlk4pMUYbgmLyyQ6XS2RJk6AEx2DCfNrban9iHfrtGbX9lIJyq4O0uEThYDRCPtMm3BLRtWaHyADXF0T6v0sJHUeZymvL3pL0XENDZyI7GmJAWo9S7b6eNlX42rUgOJIlN7KrjoSokyk4kcHaeoxZ7LIfylUdfBbq69ZrAcvOjXsYBedthOBeIq/URZYVaHM5gGDvDnGU0UnyPoMwnb960WU5oe+8lN67BwpzsrLl39pG+vCPdNogR7WPeyBN/Apt4/z5UZspOnIn7LHaZVLZCLwIEeUEabzTUHSOperHO2GdpFELR7eqmhgjvn1hTkboXOjwM4JWZVbhmyzygHg5geKkfoRqa1QXLyWeeqCy+VUB14N6RPGCPqHRJZozRRPFE4yxapcW4Jf+j4m9JjKeufLMpbuhnXOXcJBOzACEmBnirWxU5bclGMWPVfG7VjAIOHScNHF2YZcxDt47GJptfIGL4BuhzksmxJx9koxdxwbo0atwq3k9S36gi7wlvbI7W1ugVlGRi16BTGQ/pV4lt6igs62FdFVK+M9P5ZvjLGwwQCSRlxw3GQUhmFSuYLfQbvFfDKYXUSC8HYZcJtG2eN3yk2Po6rAKY2ZVjacUDsZ9cjPMsxCJ9GLgSbmEaf8LdiW66NgmtRVWGyvUzVO5R1eCjqPpqwWouoZfdhNmN+38
non_defect
m e pdqova adrombqc n cmqzp
0
329,477
28,244,749,656
IssuesEvent
2023-04-06 09:51:39
emoncms/emoncms
https://api.github.com/repos/emoncms/emoncms
closed
Browser Support?
testing required
Do we care about IE < 9? As per https://github.com/flot/flot/blob/v0.8.3/README.md#installation we can remove this code ``` <!--[if lte IE 8]><script language="javascript" type="text/javascript" src="excanvas.min.js"></script><![endif]--> ```
1.0
Browser Support? - Do we care about IE < 9? As per https://github.com/flot/flot/blob/v0.8.3/README.md#installation we can remove this code ``` <!--[if lte IE 8]><script language="javascript" type="text/javascript" src="excanvas.min.js"></script><![endif]--> ```
non_defect
browser support do we care about ie as per we can remove this code
0
98,481
8,678,390,913
IssuesEvent
2018-11-30 19:46:01
WallarooLabs/wallaroo
https://api.github.com/repos/WallarooLabs/wallaroo
closed
Integration test harness: start workers too quickly
ci error test
#### Is this a bug, feature request, or feedback? Bug #### What is the current behavior? Intermittent bug in starting workers, e.g. https://circleci.com/gh/WallarooLabs/wallaroo/12730, where a worker is started AFIACT before the initializer worker is ready. The output from `worker1` is: ``` --- worker1 (pid: 7478, rc: None) -> ****CLUSTERING MODE is active**** ****AUTOSCALE MODE is active**** ||| Resilience directory: /tmp/res-data.o13dnsep||| Set up external channel listener on 127.0.0.1:20035 _create_control_connection: call from line 106 AUTOSCALE: Waiting for new autoscale event. worker1 external: listening on 127.0.0.1:20035 metrics outgoing connected worker1 control: listening on 127.0.0.1:20032 _create_control_connection: call from line 605 _create_control_connection: target_name initializer host 127.0.0.1 service 20010 Sent control message to initializer ControlSenderConnectNotifier (to initializer): connection failed! [... shutdown messages omitted ...] ``` ... which results in this failures by the test harness: ``` cd /root/project/examples/python/word_count_with_dynamic_keys/ && \ integration_test --newline-file-sender count_this.txt \ --validation-cmd 'python _test/validate.py --expected _expected.json --output' \ --log-level error \ --batch-size 1 \ --output 'received.txt' \ --command 'machida3 --application-module word_count_with_dynamic_keys' \ --sink-mode newlines \ --sink-expect 482 \ --workers 3 2018-11-01 22:09:37,724 integration ERROR [cluster.py:519] Encountered and error when starting up the cluster 2018-11-01 22:09:37,724 integration ERROR [cluster.py:520] Timed out after 30 seconds while waiting for cluster to resume processing. Traceback (most recent call last): File "/root/project/testing/tools/integration/cluster.py", line 510, in __init__ self.wait_to_resume_processing(self.is_ready_timeout) File "/root/project/testing/tools/integration/cluster.py", line 843, in wait_to_resume_processing raise w.error integration.errors.TimeoutError: Timed out after 30 seconds while waiting for cluster to resume processing. 2018-11-01 22:09:37,727 integration ERROR [integration.py:278] Integration pipeline_test encountered an error 2018-11-01 22:09:37,727 integration ERROR [integration.py:281] The last 10 lines of each worker were: --- initializer (pid: 7476, rc: None) -> ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed --- initializer (pid: 7476, rc: None) <- === --- worker1 (pid: 7478, rc: None) -> ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed --- worker1 (pid: 7478, rc: None) <- === --- worker2 (pid: 7484, rc: None) -> ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed --- worker2 (pid: 7484, rc: None) <- 2018-11-01 22:09:37,727 integration ERROR [integration_test:577] Encountered an error while running the test for 'machida3 --application-module word_count_with_dynamic_keys' === Traceback (most recent call last): File "/root/project/testing/tools/integration_test", line 559, in CLI persistent_data = persistent_data) File "/root/project/testing/tools/integration/integration.py", line 154, in pipeline_test persistent_data=persistent_data) as cluster: File "/root/project/testing/tools/integration/cluster.py", line 523, in __init__ raise err File "/root/project/testing/tools/integration/cluster.py", line 510, in __init__ self.wait_to_resume_processing(self.is_ready_timeout) File "/root/project/testing/tools/integration/cluster.py", line 843, in wait_to_resume_processing raise w.error integration.errors.TimeoutError: Timed out after 30 seconds while waiting for cluster to resume processing. examples/python/word_count_with_dynamic_keys/Makefile:81: recipe for target 'word_count_with_dynamic_keys_py3_test' failed make: *** [word_count_with_dynamic_keys_py3_test] Error 1 Exited with code 2 ``` #### What is the expected behavior? Perhaps we wait until all TCP ports listened to by initializer are actually open & listening? #### What OS and version of Wallaroo are you using? CircleCI environment, branch for PR #2621 #### Steps to reproduce? `make integration-tests debug=true`
1.0
Integration test harness: start workers too quickly - #### Is this a bug, feature request, or feedback? Bug #### What is the current behavior? Intermittent bug in starting workers, e.g. https://circleci.com/gh/WallarooLabs/wallaroo/12730, where a worker is started AFIACT before the initializer worker is ready. The output from `worker1` is: ``` --- worker1 (pid: 7478, rc: None) -> ****CLUSTERING MODE is active**** ****AUTOSCALE MODE is active**** ||| Resilience directory: /tmp/res-data.o13dnsep||| Set up external channel listener on 127.0.0.1:20035 _create_control_connection: call from line 106 AUTOSCALE: Waiting for new autoscale event. worker1 external: listening on 127.0.0.1:20035 metrics outgoing connected worker1 control: listening on 127.0.0.1:20032 _create_control_connection: call from line 605 _create_control_connection: target_name initializer host 127.0.0.1 service 20010 Sent control message to initializer ControlSenderConnectNotifier (to initializer): connection failed! [... shutdown messages omitted ...] ``` ... which results in this failures by the test harness: ``` cd /root/project/examples/python/word_count_with_dynamic_keys/ && \ integration_test --newline-file-sender count_this.txt \ --validation-cmd 'python _test/validate.py --expected _expected.json --output' \ --log-level error \ --batch-size 1 \ --output 'received.txt' \ --command 'machida3 --application-module word_count_with_dynamic_keys' \ --sink-mode newlines \ --sink-expect 482 \ --workers 3 2018-11-01 22:09:37,724 integration ERROR [cluster.py:519] Encountered and error when starting up the cluster 2018-11-01 22:09:37,724 integration ERROR [cluster.py:520] Timed out after 30 seconds while waiting for cluster to resume processing. Traceback (most recent call last): File "/root/project/testing/tools/integration/cluster.py", line 510, in __init__ self.wait_to_resume_processing(self.is_ready_timeout) File "/root/project/testing/tools/integration/cluster.py", line 843, in wait_to_resume_processing raise w.error integration.errors.TimeoutError: Timed out after 30 seconds while waiting for cluster to resume processing. 2018-11-01 22:09:37,727 integration ERROR [integration.py:278] Integration pipeline_test encountered an error 2018-11-01 22:09:37,727 integration ERROR [integration.py:281] The last 10 lines of each worker were: --- initializer (pid: 7476, rc: None) -> ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed ExternalChannelConnectNotifier (initializer): server closed --- initializer (pid: 7476, rc: None) <- === --- worker1 (pid: 7478, rc: None) -> ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed ExternalChannelConnectNotifier (worker1): server closed --- worker1 (pid: 7478, rc: None) <- === --- worker2 (pid: 7484, rc: None) -> ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed ExternalChannelConnectNotifier (worker2): server closed --- worker2 (pid: 7484, rc: None) <- 2018-11-01 22:09:37,727 integration ERROR [integration_test:577] Encountered an error while running the test for 'machida3 --application-module word_count_with_dynamic_keys' === Traceback (most recent call last): File "/root/project/testing/tools/integration_test", line 559, in CLI persistent_data = persistent_data) File "/root/project/testing/tools/integration/integration.py", line 154, in pipeline_test persistent_data=persistent_data) as cluster: File "/root/project/testing/tools/integration/cluster.py", line 523, in __init__ raise err File "/root/project/testing/tools/integration/cluster.py", line 510, in __init__ self.wait_to_resume_processing(self.is_ready_timeout) File "/root/project/testing/tools/integration/cluster.py", line 843, in wait_to_resume_processing raise w.error integration.errors.TimeoutError: Timed out after 30 seconds while waiting for cluster to resume processing. examples/python/word_count_with_dynamic_keys/Makefile:81: recipe for target 'word_count_with_dynamic_keys_py3_test' failed make: *** [word_count_with_dynamic_keys_py3_test] Error 1 Exited with code 2 ``` #### What is the expected behavior? Perhaps we wait until all TCP ports listened to by initializer are actually open & listening? #### What OS and version of Wallaroo are you using? CircleCI environment, branch for PR #2621 #### Steps to reproduce? `make integration-tests debug=true`
non_defect
integration test harness start workers too quickly is this a bug feature request or feedback bug what is the current behavior intermittent bug in starting workers e g where a worker is started afiact before the initializer worker is ready the output from is pid rc none clustering mode is active autoscale mode is active resilience directory tmp res data set up external channel listener on create control connection call from line autoscale waiting for new autoscale event external listening on metrics outgoing connected control listening on create control connection call from line create control connection target name initializer host service sent control message to initializer controlsenderconnectnotifier to initializer connection failed which results in this failures by the test harness cd root project examples python word count with dynamic keys integration test newline file sender count this txt validation cmd python test validate py expected expected json output log level error batch size output received txt command application module word count with dynamic keys sink mode newlines sink expect workers integration error encountered and error when starting up the cluster integration error timed out after seconds while waiting for cluster to resume processing traceback most recent call last file root project testing tools integration cluster py line in init self wait to resume processing self is ready timeout file root project testing tools integration cluster py line in wait to resume processing raise w error integration errors timeouterror timed out after seconds while waiting for cluster to resume processing integration error integration pipeline test encountered an error integration error the last lines of each worker were initializer pid rc none externalchannelconnectnotifier initializer server closed externalchannelconnectnotifier initializer server closed externalchannelconnectnotifier initializer server closed externalchannelconnectnotifier initializer server closed externalchannelconnectnotifier initializer server closed externalchannelconnectnotifier initializer server closed externalchannelconnectnotifier initializer server closed externalchannelconnectnotifier initializer server closed externalchannelconnectnotifier initializer server closed externalchannelconnectnotifier initializer server closed initializer pid rc none pid rc none externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed pid rc none pid rc none externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed externalchannelconnectnotifier server closed pid rc none integration error encountered an error while running the test for application module word count with dynamic keys traceback most recent call last file root project testing tools integration test line in cli persistent data persistent data file root project testing tools integration integration py line in pipeline test persistent data persistent data as cluster file root project testing tools integration cluster py line in init raise err file root project testing tools integration cluster py line in init self wait to resume processing self is ready timeout file root project testing tools integration cluster py line in wait to resume processing raise w error integration errors timeouterror timed out after seconds while waiting for cluster to resume processing examples python word count with dynamic keys makefile recipe for target word count with dynamic keys test failed make error exited with code what is the expected behavior perhaps we wait until all tcp ports listened to by initializer are actually open listening what os and version of wallaroo are you using circleci environment branch for pr steps to reproduce make integration tests debug true
0
58,586
16,613,237,259
IssuesEvent
2021-06-02 13:56:54
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Can't p2p 1:1 call if turnserver is disabled
T-Defect
Not sure if this is intended or not, but with the ``` Allow Peer-to-Peer for 1:1 calls (if you enable this, the other party might be able to see your IP address) ``` option enabled, it still requires the server's turnserver to be enabled. I thought that if both parties have this enabled, it would be a p2pe2ee call?
1.0
Can't p2p 1:1 call if turnserver is disabled - Not sure if this is intended or not, but with the ``` Allow Peer-to-Peer for 1:1 calls (if you enable this, the other party might be able to see your IP address) ``` option enabled, it still requires the server's turnserver to be enabled. I thought that if both parties have this enabled, it would be a p2pe2ee call?
defect
can t call if turnserver is disabled not sure if this is intended or not but with the allow peer to peer for calls if you enable this the other party might be able to see your ip address option enabled it still requires the server s turnserver to be enabled i thought that if both parties have this enabled it would be a call
1
359,108
25,220,073,067
IssuesEvent
2022-11-14 12:07:52
vercel/next.js
https://api.github.com/repos/vercel/next.js
closed
Docs: Update Next tutorial on the official website
template: documentation
### What is the improvement or update you wish to see? After the release of v13 of next, the tutorial present on the website has become outdated. ### Is there any context that might help us understand? [This page](https://nextjs.org/learn/basics) ### Does the docs page already exist? Please link to it. https://nextjs.org/learn/basics
1.0
Docs: Update Next tutorial on the official website - ### What is the improvement or update you wish to see? After the release of v13 of next, the tutorial present on the website has become outdated. ### Is there any context that might help us understand? [This page](https://nextjs.org/learn/basics) ### Does the docs page already exist? Please link to it. https://nextjs.org/learn/basics
non_defect
docs update next tutorial on the official website what is the improvement or update you wish to see after the release of of next the tutorial present on the website has become outdated is there any context that might help us understand does the docs page already exist please link to it
0
37,383
12,477,454,011
IssuesEvent
2020-05-29 14:59:12
LibrIT/passhport
https://api.github.com/repos/LibrIT/passhport
closed
CVE-2018-14042 (Medium) detected in bootstrap-3.3.2.min.js, bootstrap-3.3.7.min.js
New security vulnerability
## CVE-2018-14042 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bootstrap-3.3.2.min.js</b>, <b>bootstrap-3.3.7.min.js</b></p></summary> <p> <details><summary><b>bootstrap-3.3.2.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.2/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.2/js/bootstrap.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/passhport/passhweb/app/static/bower_components/bootstrap-daterangepicker/website/index.html</p> <p>Path to vulnerable library: /passhport/passhweb/app/static/bower_components/bootstrap-daterangepicker/website/index.html,/passhport/passhweb/app/static/bower_components/bootstrap-daterangepicker/demo.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.2.min.js** (Vulnerable Library) </details> <details><summary><b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/passhport/passhweb/app/static/bower_components/bootstrap-colorpicker/index.html</p> <p>Path to vulnerable library: /passhport/passhweb/app/static/bower_components/bootstrap-colorpicker/index.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/LibrIT/passhport/commit/280394daf60b8887c5eebccaca5e3c390a11b1f2">280394daf60b8887c5eebccaca5e3c390a11b1f2</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-14042 (Medium) detected in bootstrap-3.3.2.min.js, bootstrap-3.3.7.min.js - ## CVE-2018-14042 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bootstrap-3.3.2.min.js</b>, <b>bootstrap-3.3.7.min.js</b></p></summary> <p> <details><summary><b>bootstrap-3.3.2.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.2/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.2/js/bootstrap.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/passhport/passhweb/app/static/bower_components/bootstrap-daterangepicker/website/index.html</p> <p>Path to vulnerable library: /passhport/passhweb/app/static/bower_components/bootstrap-daterangepicker/website/index.html,/passhport/passhweb/app/static/bower_components/bootstrap-daterangepicker/demo.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.2.min.js** (Vulnerable Library) </details> <details><summary><b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/passhport/passhweb/app/static/bower_components/bootstrap-colorpicker/index.html</p> <p>Path to vulnerable library: /passhport/passhweb/app/static/bower_components/bootstrap-colorpicker/index.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/LibrIT/passhport/commit/280394daf60b8887c5eebccaca5e3c390a11b1f2">280394daf60b8887c5eebccaca5e3c390a11b1f2</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in bootstrap min js bootstrap min js cve medium severity vulnerability vulnerable libraries bootstrap min js bootstrap min js bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file tmp ws scm passhport passhweb app static bower components bootstrap daterangepicker website index html path to vulnerable library passhport passhweb app static bower components bootstrap daterangepicker website index html passhport passhweb app static bower components bootstrap daterangepicker demo html dependency hierarchy x bootstrap min js vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file tmp ws scm passhport passhweb app static bower components bootstrap colorpicker index html path to vulnerable library passhport passhweb app static bower components bootstrap colorpicker index html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href vulnerability details in bootstrap before xss is possible in the data container property of tooltip publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org webjars npm bootstrap org webjars bootstrap step up your open source security game with whitesource
0
51,837
12,814,612,434
IssuesEvent
2020-07-04 19:52:59
graknlabs/dependencies
https://api.github.com/repos/graknlabs/dependencies
opened
Transitive dependency diagram generator
type: build
Can we have a quick script to visualise all of our transitive dependencies in a diagram? We can take inspiration from this script in `@graknlabs_hypergraph`. https://github.com/graknlabs/hypergraph/blob/95ab5d7e6ae1d94fd7d46faf52234ae9af7efb65/diagrams/deps.sh If we can have a way to differentiate a direct dependency vs a transitive dependency in this diagram, that would be great. But if it's not doable with an immediately available API, don't worry about it. We can always tell the direct dependencies from the `artifacts.bzl` file.
1.0
Transitive dependency diagram generator - Can we have a quick script to visualise all of our transitive dependencies in a diagram? We can take inspiration from this script in `@graknlabs_hypergraph`. https://github.com/graknlabs/hypergraph/blob/95ab5d7e6ae1d94fd7d46faf52234ae9af7efb65/diagrams/deps.sh If we can have a way to differentiate a direct dependency vs a transitive dependency in this diagram, that would be great. But if it's not doable with an immediately available API, don't worry about it. We can always tell the direct dependencies from the `artifacts.bzl` file.
non_defect
transitive dependency diagram generator can we have a quick script to visualise all of our transitive dependencies in a diagram we can take inspiration from this script in graknlabs hypergraph if we can have a way to differentiate a direct dependency vs a transitive dependency in this diagram that would be great but if it s not doable with an immediately available api don t worry about it we can always tell the direct dependencies from the artifacts bzl file
0
77,826
27,183,072,382
IssuesEvent
2023-02-18 21:57:30
zed-industries/community
https://api.github.com/repos/zed-industries/community
closed
TypeScript rename action is broken when in Vim mode
defect vim quality week candidate
### Check for existing issues - [X] Completed ### Describe the bug / provide steps to reproduce it The user said: Renaming symbols in typescript works weird with Vim mode. If I press F2 while in normal mode, the rename input will appear. If I type 2 characters, the rename input will disappear and renaming is cancelled. I found out that if I switch back to insert mode by pressing "i", i can then edit the name, press enter, and the renaming will be done. ### Environment Zed: v0.73.3 (preview) OS: macOS 13.1.0 Memory: 64 GiB Architecture: aarch64 ### If applicable, add mockups / screenshots to help explain present your vision of the feature _No response_ ### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue. If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000. _No response_
1.0
TypeScript rename action is broken when in Vim mode - ### Check for existing issues - [X] Completed ### Describe the bug / provide steps to reproduce it The user said: Renaming symbols in typescript works weird with Vim mode. If I press F2 while in normal mode, the rename input will appear. If I type 2 characters, the rename input will disappear and renaming is cancelled. I found out that if I switch back to insert mode by pressing "i", i can then edit the name, press enter, and the renaming will be done. ### Environment Zed: v0.73.3 (preview) OS: macOS 13.1.0 Memory: 64 GiB Architecture: aarch64 ### If applicable, add mockups / screenshots to help explain present your vision of the feature _No response_ ### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue. If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000. _No response_
defect
typescript rename action is broken when in vim mode check for existing issues completed describe the bug provide steps to reproduce it the user said renaming symbols in typescript works weird with vim mode if i press while in normal mode the rename input will appear if i type characters the rename input will disappear and renaming is cancelled i found out that if i switch back to insert mode by pressing i i can then edit the name press enter and the renaming will be done environment zed preview os macos memory gib architecture if applicable add mockups screenshots to help explain present your vision of the feature no response if applicable attach your library logs zed zed log file to this issue if you only need the most recent lines you can run the zed open log command palette action to see the last no response
1
45,423
24,056,386,142
IssuesEvent
2022-09-16 17:20:28
python/cpython
https://api.github.com/repos/python/cpython
closed
pathlib.Path.glob's generator is not a real generator
performance expert-IO 3.12 expert-pathlib
BPO | [44069](https://bugs.python.org/issue44069) --- | :--- Nosy | @terryjreedy, @pitrou, @serhiy-storchaka, @erip <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2021-05-07.15:46:22.399> labels = ['3.11', 'expert-IO', 'performance'] title = "pathlib.Path.glob's generator is not a real generator" updated_at = <Date 2021-05-21.16:32:39.328> user = 'https://github.com/erip' ``` bugs.python.org fields: ```python activity = <Date 2021-05-21.16:32:39.328> actor = 'terry.reedy' assignee = 'none' closed = False closed_date = None closer = None components = ['IO'] creation = <Date 2021-05-07.15:46:22.399> creator = 'Elijah Rippeth' dependencies = [] files = [] hgrepos = [] issue_num = 44069 keywords = [] message_count = 3.0 messages = ['393190', '393562', '394125'] nosy_count = 4.0 nosy_names = ['terry.reedy', 'pitrou', 'serhiy.storchaka', 'Elijah Rippeth'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'performance' url = 'https://bugs.python.org/issue44069' versions = ['Python 3.11'] ``` </p></details>
True
pathlib.Path.glob's generator is not a real generator - BPO | [44069](https://bugs.python.org/issue44069) --- | :--- Nosy | @terryjreedy, @pitrou, @serhiy-storchaka, @erip <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2021-05-07.15:46:22.399> labels = ['3.11', 'expert-IO', 'performance'] title = "pathlib.Path.glob's generator is not a real generator" updated_at = <Date 2021-05-21.16:32:39.328> user = 'https://github.com/erip' ``` bugs.python.org fields: ```python activity = <Date 2021-05-21.16:32:39.328> actor = 'terry.reedy' assignee = 'none' closed = False closed_date = None closer = None components = ['IO'] creation = <Date 2021-05-07.15:46:22.399> creator = 'Elijah Rippeth' dependencies = [] files = [] hgrepos = [] issue_num = 44069 keywords = [] message_count = 3.0 messages = ['393190', '393562', '394125'] nosy_count = 4.0 nosy_names = ['terry.reedy', 'pitrou', 'serhiy.storchaka', 'Elijah Rippeth'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'performance' url = 'https://bugs.python.org/issue44069' versions = ['Python 3.11'] ``` </p></details>
non_defect
pathlib path glob s generator is not a real generator bpo nosy terryjreedy pitrou serhiy storchaka erip note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title pathlib path glob s generator is not a real generator updated at user bugs python org fields python activity actor terry reedy assignee none closed false closed date none closer none components creation creator elijah rippeth dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type performance url versions
0
47,646
19,686,379,116
IssuesEvent
2022-01-11 22:46:44
python-discord/infra
https://api.github.com/repos/python-discord/infra
closed
UFW Setup
type: service component: networking component: security
# Abstract We should setup UFW on our nodes to control inbound and outbound traffic. # Implementation There are several ways to implement this so we should talk about whether roles should control their own UFW rules or whether we should have a ufw role which manages all rules. Ansible has a [`ufw`](https://docs.ansible.com/ansible/latest/collections/community/general/ufw_module.html) module for rule management. We need to allow SSH, Wireguard and HTTP off the top of my head, there are probably other things here and there as well.
1.0
UFW Setup - # Abstract We should setup UFW on our nodes to control inbound and outbound traffic. # Implementation There are several ways to implement this so we should talk about whether roles should control their own UFW rules or whether we should have a ufw role which manages all rules. Ansible has a [`ufw`](https://docs.ansible.com/ansible/latest/collections/community/general/ufw_module.html) module for rule management. We need to allow SSH, Wireguard and HTTP off the top of my head, there are probably other things here and there as well.
non_defect
ufw setup abstract we should setup ufw on our nodes to control inbound and outbound traffic implementation there are several ways to implement this so we should talk about whether roles should control their own ufw rules or whether we should have a ufw role which manages all rules ansible has a module for rule management we need to allow ssh wireguard and http off the top of my head there are probably other things here and there as well
0
765,990
26,867,907,031
IssuesEvent
2023-02-04 04:31:08
Rugged-Thread/stitch
https://api.github.com/repos/Rugged-Thread/stitch
closed
Item: Listed Standard Repairs should be dependent on the item.
High Priority
Via 2022-10-28 Similar to how complications work in relation to standard repairs, the standard repairs available in the item creation should be limited based on the type of item.
1.0
Item: Listed Standard Repairs should be dependent on the item. - Via 2022-10-28 Similar to how complications work in relation to standard repairs, the standard repairs available in the item creation should be limited based on the type of item.
non_defect
item listed standard repairs should be dependent on the item via similar to how complications work in relation to standard repairs the standard repairs available in the item creation should be limited based on the type of item
0
74,813
9,807,825,534
IssuesEvent
2019-06-12 14:27:54
Accenture/AmpliGraph
https://api.github.com/repos/Accenture/AmpliGraph
closed
Number of epochs for benchmark dataset
quality & documentation
Hello, First, thanks a lot for such a nice library! I have a question on the performance reported [here](https://docs.ampligraph.org/en/latest/experiments.html#runtime-performance). I was wondering if 4000 epoch really necessary to achieve the reported result? My machine is not running as fast as yours, and It would be nice to know I can still replicate your result using fewer epochs. Also, are you able to replicate the results for most of the method you implemented? I just checked fb-237, it seems that for some old method, your result is better than that is reported in the paper.
1.0
Number of epochs for benchmark dataset - Hello, First, thanks a lot for such a nice library! I have a question on the performance reported [here](https://docs.ampligraph.org/en/latest/experiments.html#runtime-performance). I was wondering if 4000 epoch really necessary to achieve the reported result? My machine is not running as fast as yours, and It would be nice to know I can still replicate your result using fewer epochs. Also, are you able to replicate the results for most of the method you implemented? I just checked fb-237, it seems that for some old method, your result is better than that is reported in the paper.
non_defect
number of epochs for benchmark dataset hello first thanks a lot for such a nice library i have a question on the performance reported i was wondering if epoch really necessary to achieve the reported result my machine is not running as fast as yours and it would be nice to know i can still replicate your result using fewer epochs also are you able to replicate the results for most of the method you implemented i just checked fb it seems that for some old method your result is better than that is reported in the paper
0
22,277
3,619,793,015
IssuesEvent
2016-02-08 17:19:02
miracle091/transmission-remote-dotnet
https://api.github.com/repos/miracle091/transmission-remote-dotnet
closed
Crash
Priority-Medium Type-Defect
``` What steps will reproduce the problem? No idea - i just got crash report. What is the expected output? What do you see instead? NA What version of the products are you using? OS: Transmission: transmission - 2.42-1 Remote: 3.24(build 3) Please provide any additional information below. Feel free to attach screenshots or sample code which demonstrates the issue being described. ``` Original issue reported on code.google.com by `chetan...@gmail.com` on 12 Mar 2012 at 7:05 Attachments: * [trdcrash_20120312_062411.log](https://storage.googleapis.com/google-code-attachments/transmission-remote-dotnet/issue-438/comment-0/trdcrash_20120312_062411.log)
1.0
Crash - ``` What steps will reproduce the problem? No idea - i just got crash report. What is the expected output? What do you see instead? NA What version of the products are you using? OS: Transmission: transmission - 2.42-1 Remote: 3.24(build 3) Please provide any additional information below. Feel free to attach screenshots or sample code which demonstrates the issue being described. ``` Original issue reported on code.google.com by `chetan...@gmail.com` on 12 Mar 2012 at 7:05 Attachments: * [trdcrash_20120312_062411.log](https://storage.googleapis.com/google-code-attachments/transmission-remote-dotnet/issue-438/comment-0/trdcrash_20120312_062411.log)
defect
crash what steps will reproduce the problem no idea i just got crash report what is the expected output what do you see instead na what version of the products are you using os transmission transmission remote build please provide any additional information below feel free to attach screenshots or sample code which demonstrates the issue being described original issue reported on code google com by chetan gmail com on mar at attachments
1
252,288
21,569,429,908
IssuesEvent
2022-05-02 06:00:24
stores-cedcommerce/Internal-Christopher-Zorc-Store-Dev
https://api.github.com/repos/stores-cedcommerce/Internal-Christopher-Zorc-Store-Dev
closed
Page redirection issue on homepage.
Desktop Functional / bug Ready to test Home page content
Bug 1 - User click on 'Read more' button then get redirect on header. Exp - On about us section user click on read more button then it should get redirect to about us page. Ref Link - https://drive.google.com/file/d/1EupbSmYa9KZEulu0ypkwZmfDi49uiU6u/view Bug 2 - Meet our team button not working. Exp - Button should redirect user to its relevant respective page. Ref Link - https://drive.google.com/file/d/1BMjDLsqtUGlWUtpoM7s1hB4PViL61k0k/view
1.0
Page redirection issue on homepage. - Bug 1 - User click on 'Read more' button then get redirect on header. Exp - On about us section user click on read more button then it should get redirect to about us page. Ref Link - https://drive.google.com/file/d/1EupbSmYa9KZEulu0ypkwZmfDi49uiU6u/view Bug 2 - Meet our team button not working. Exp - Button should redirect user to its relevant respective page. Ref Link - https://drive.google.com/file/d/1BMjDLsqtUGlWUtpoM7s1hB4PViL61k0k/view
non_defect
page redirection issue on homepage bug user click on read more button then get redirect on header exp on about us section user click on read more button then it should get redirect to about us page ref link bug meet our team button not working exp button should redirect user to its relevant respective page ref link
0
69,394
22,337,285,426
IssuesEvent
2022-06-14 19:50:36
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
[Assistive tech and device support] Focus state isn't visible. (11.09.1)
508/Accessibility 508-defect-1 collab-cycle-feedback launch-blocking CCIssue11.09 eGain
### General Information #### VFS team name #### VFS product name eGain #### Point of Contact/Reviewers Sarah Koomson (@Saraharaba8) - Accessibility *For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website. --- ### Platform Issue Focus state isn't visible. ### Issue Details The first "X" mark icon has a very thin dotted line focus indicator shown on browsers such as Chrome, Safari, i.e. ### Link, screenshot or steps to recreate ### VA.gov Experience Standard [Category Number 11, Issue Number 09](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html) ### Other References WCAG SC 2.4.7_AA --- ### Platform Recommendation Dotted line focus indicators make it very difficult for users to see. In order to provide users with an easy-to-see focus indicator that is consistent across all browsers, use the :focus selector in CSS to define a style change that happens when an element has focus. ### VFS Team Tasks to Complete - [ ] Comment on the ticket if there are questions or concerns - [ ] Close the ticket when the issue has been resolved or validated by your Product Owner. If a team has additional questions or needs Platform help validating the issue, please comment in the ticket.
1.0
[Assistive tech and device support] Focus state isn't visible. (11.09.1) - ### General Information #### VFS team name #### VFS product name eGain #### Point of Contact/Reviewers Sarah Koomson (@Saraharaba8) - Accessibility *For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website. --- ### Platform Issue Focus state isn't visible. ### Issue Details The first "X" mark icon has a very thin dotted line focus indicator shown on browsers such as Chrome, Safari, i.e. ### Link, screenshot or steps to recreate ### VA.gov Experience Standard [Category Number 11, Issue Number 09](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html) ### Other References WCAG SC 2.4.7_AA --- ### Platform Recommendation Dotted line focus indicators make it very difficult for users to see. In order to provide users with an easy-to-see focus indicator that is consistent across all browsers, use the :focus selector in CSS to define a style change that happens when an element has focus. ### VFS Team Tasks to Complete - [ ] Comment on the ticket if there are questions or concerns - [ ] Close the ticket when the issue has been resolved or validated by your Product Owner. If a team has additional questions or needs Platform help validating the issue, please comment in the ticket.
defect
focus state isn t visible general information vfs team name vfs product name egain point of contact reviewers sarah koomson accessibility for more information on how to interpret this ticket please refer to the guidance on platform website platform issue focus state isn t visible issue details the first x mark icon has a very thin dotted line focus indicator shown on browsers such as chrome safari i e link screenshot or steps to recreate va gov experience standard other references wcag sc aa platform recommendation dotted line focus indicators make it very difficult for users to see in order to provide users with an easy to see focus indicator that is consistent across all browsers use the focus selector in css to define a style change that happens when an element has focus vfs team tasks to complete comment on the ticket if there are questions or concerns close the ticket when the issue has been resolved or validated by your product owner if a team has additional questions or needs platform help validating the issue please comment in the ticket
1
16,520
2,910,088,473
IssuesEvent
2015-06-21 11:58:29
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
Remove CSS scale from Advanced FileUpload
5.1.20 5.2.7 defect
The fileupload.css contains a very strange line that causes errors - for instance, see [this bug](https://github.com/TheCoder4eu/BootsFaces-OSP/issues/110) which has been reported on the BootsFaces page. As it turns out, it's a PrimeFaces bug, which in turn is a bug you've inherited from the FileUpload jQuery plugin. The offending line is: ``` CSS .ui-fileupload-buttonbar .ui-fileupload-choose input { ... transform: translate(-300px, 0) scale(4); ... } ``` As a quick fix, it suffices to remove the line. However, when I looked at the jQuery plugin page, I noted that the author removed this line a long time ago. He also published a lot of releases in the meantime. Would you mind to update to the current version?
1.0
Remove CSS scale from Advanced FileUpload - The fileupload.css contains a very strange line that causes errors - for instance, see [this bug](https://github.com/TheCoder4eu/BootsFaces-OSP/issues/110) which has been reported on the BootsFaces page. As it turns out, it's a PrimeFaces bug, which in turn is a bug you've inherited from the FileUpload jQuery plugin. The offending line is: ``` CSS .ui-fileupload-buttonbar .ui-fileupload-choose input { ... transform: translate(-300px, 0) scale(4); ... } ``` As a quick fix, it suffices to remove the line. However, when I looked at the jQuery plugin page, I noted that the author removed this line a long time ago. He also published a lot of releases in the meantime. Would you mind to update to the current version?
defect
remove css scale from advanced fileupload the fileupload css contains a very strange line that causes errors for instance see which has been reported on the bootsfaces page as it turns out it s a primefaces bug which in turn is a bug you ve inherited from the fileupload jquery plugin the offending line is css ui fileupload buttonbar ui fileupload choose input transform translate scale as a quick fix it suffices to remove the line however when i looked at the jquery plugin page i noted that the author removed this line a long time ago he also published a lot of releases in the meantime would you mind to update to the current version
1
22,403
3,643,500,936
IssuesEvent
2016-02-15 02:06:31
codingsnippets/prettytable
https://api.github.com/repos/codingsnippets/prettytable
closed
Upload new version to pypi
auto-migrated Priority-Medium Type-Defect
``` Currently, version 0.7.2 is the latest available on pypi, could you upload the latest version? Thanks. ``` Original issue reported on code.google.com by `rea...@asymmetricventures.com` on 4 Dec 2014 at 6:21
1.0
Upload new version to pypi - ``` Currently, version 0.7.2 is the latest available on pypi, could you upload the latest version? Thanks. ``` Original issue reported on code.google.com by `rea...@asymmetricventures.com` on 4 Dec 2014 at 6:21
defect
upload new version to pypi currently version is the latest available on pypi could you upload the latest version thanks original issue reported on code google com by rea asymmetricventures com on dec at
1
137,343
30,675,597,939
IssuesEvent
2023-07-26 04:49:10
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
tsblender 103.0.1 has 2 GuardDog issues
guarddog code-execution exec-base64
https://pypi.org/project/tsblender https://inspector.pypi.io/project/tsblender ```{ "dependency": "tsblender", "version": "103.0.1", "result": { "issues": 2, "errors": {}, "results": { "exec-base64": [ { "location": "tsblender-103.0.1/src/tsblender/tsblender.py:1963", "code": " series = eval(equation)", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" } ], "code-execution": [ { "location": "tsblender-103.0.1/setup.py:12", "code": " subprocess.run(shlex.split(\"cleanpy .\"), check=True)", "message": "This package is executing OS commands in the setup.py file" } ] }, "path": "/tmp/tmpot7u2_lo/tsblender" } }```
1.0
tsblender 103.0.1 has 2 GuardDog issues - https://pypi.org/project/tsblender https://inspector.pypi.io/project/tsblender ```{ "dependency": "tsblender", "version": "103.0.1", "result": { "issues": 2, "errors": {}, "results": { "exec-base64": [ { "location": "tsblender-103.0.1/src/tsblender/tsblender.py:1963", "code": " series = eval(equation)", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" } ], "code-execution": [ { "location": "tsblender-103.0.1/setup.py:12", "code": " subprocess.run(shlex.split(\"cleanpy .\"), check=True)", "message": "This package is executing OS commands in the setup.py file" } ] }, "path": "/tmp/tmpot7u2_lo/tsblender" } }```
non_defect
tsblender has guarddog issues dependency tsblender version result issues errors results exec location tsblender src tsblender tsblender py code series eval equation message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n code execution location tsblender setup py code subprocess run shlex split cleanpy check true message this package is executing os commands in the setup py file path tmp lo tsblender
0
131,220
18,234,865,738
IssuesEvent
2021-10-01 05:00:12
graywidjaya/snyk-scanning-testing
https://api.github.com/repos/graywidjaya/snyk-scanning-testing
opened
CVE-2021-36090 (High) detected in commons-compress-1.9.jar
security vulnerability
## CVE-2021-36090 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-compress-1.9.jar</b></p></summary> <p>Apache Commons Compress software defines an API for working with compression and archive formats. These include: bzip2, gzip, pack200, lzma, xz, Snappy, traditional Unix Compress, DEFLATE and ar, cpio, jar, tar, zip, dump, 7z, arj.</p> <p>Path to dependency file: snyk-scanning-testing/ProductManager/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/commons/commons-compress/1.9/commons-compress-1.9.jar</p> <p> Dependency Hierarchy: - webjars-locator-core-0.35.jar (Root Library) - :x: **commons-compress-1.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/graywidjaya/snyk-scanning-testing/commit/8e11d4935d4cae9cfc1d6d0b55433a3b1002a16e">8e11d4935d4cae9cfc1d6d0b55433a3b1002a16e</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When reading a specially crafted ZIP archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress' zip package. <p>Publish Date: 2021-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-36090>CVE-2021-36090</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://commons.apache.org/proper/commons-compress/security-reports.html">https://commons.apache.org/proper/commons-compress/security-reports.html</a></p> <p>Release Date: 2021-07-13</p> <p>Fix Resolution: org.apache.commons:commons-compress:1.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-36090 (High) detected in commons-compress-1.9.jar - ## CVE-2021-36090 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-compress-1.9.jar</b></p></summary> <p>Apache Commons Compress software defines an API for working with compression and archive formats. These include: bzip2, gzip, pack200, lzma, xz, Snappy, traditional Unix Compress, DEFLATE and ar, cpio, jar, tar, zip, dump, 7z, arj.</p> <p>Path to dependency file: snyk-scanning-testing/ProductManager/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/commons/commons-compress/1.9/commons-compress-1.9.jar</p> <p> Dependency Hierarchy: - webjars-locator-core-0.35.jar (Root Library) - :x: **commons-compress-1.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/graywidjaya/snyk-scanning-testing/commit/8e11d4935d4cae9cfc1d6d0b55433a3b1002a16e">8e11d4935d4cae9cfc1d6d0b55433a3b1002a16e</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When reading a specially crafted ZIP archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress' zip package. <p>Publish Date: 2021-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-36090>CVE-2021-36090</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://commons.apache.org/proper/commons-compress/security-reports.html">https://commons.apache.org/proper/commons-compress/security-reports.html</a></p> <p>Release Date: 2021-07-13</p> <p>Fix Resolution: org.apache.commons:commons-compress:1.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in commons compress jar cve high severity vulnerability vulnerable library commons compress jar apache commons compress software defines an api for working with compression and archive formats these include gzip lzma xz snappy traditional unix compress deflate and ar cpio jar tar zip dump arj path to dependency file snyk scanning testing productmanager pom xml path to vulnerable library home wss scanner repository org apache commons commons compress commons compress jar dependency hierarchy webjars locator core jar root library x commons compress jar vulnerable library found in head commit a href found in base branch main vulnerability details when reading a specially crafted zip archive compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs this could be used to mount a denial of service attack against services that use compress zip package publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache commons commons compress step up your open source security game with whitesource
0
125,569
26,687,279,271
IssuesEvent
2023-01-26 23:22:00
GPUOpen-Tools/compressonator
https://api.github.com/repos/GPUOpen-Tools/compressonator
closed
Is it possible to have windows CLI releases as zip files?
Code update
Currently, windows releases are executables that require administrator privileges to install the command line tool. The executable can't even be unpacked by 7zip. There's no reason to "install" a CLI tool, afaik. Edit: just found out the .exe can be called from the command line, but the administrator privileges requirement is still an issue. Edit: tried it. It extracted files somewhere, I don't know where. Not to the path I specified. Would it be possible to have a simple archive as it's provided for linux ? At least for the command line version. This would help a lot for automatic deployment of new versions, thank you
1.0
Is it possible to have windows CLI releases as zip files? - Currently, windows releases are executables that require administrator privileges to install the command line tool. The executable can't even be unpacked by 7zip. There's no reason to "install" a CLI tool, afaik. Edit: just found out the .exe can be called from the command line, but the administrator privileges requirement is still an issue. Edit: tried it. It extracted files somewhere, I don't know where. Not to the path I specified. Would it be possible to have a simple archive as it's provided for linux ? At least for the command line version. This would help a lot for automatic deployment of new versions, thank you
non_defect
is it possible to have windows cli releases as zip files currently windows releases are executables that require administrator privileges to install the command line tool the executable can t even be unpacked by there s no reason to install a cli tool afaik edit just found out the exe can be called from the command line but the administrator privileges requirement is still an issue edit tried it it extracted files somewhere i don t know where not to the path i specified would it be possible to have a simple archive as it s provided for linux at least for the command line version this would help a lot for automatic deployment of new versions thank you
0
90,559
3,823,250,432
IssuesEvent
2016-03-30 07:09:30
severnbronies/severnbronies-2016
https://api.github.com/repos/severnbronies/severnbronies-2016
opened
Redesign meet archive pagination
Priority: Low Status: Pending Type: Enhancement
Current version is fairly difficult to parse if searching for specific events or time periods as meet numbers have no mapping to date, nor necessarily chronology. The list also doesn't respond well, becoming a tall stack of links. Explore options for archives to work on a more date-centric system (by year? by quarter?) and ways of illustrating this that don't take up as much room (horizontally scrolling timeline?)
1.0
Redesign meet archive pagination - Current version is fairly difficult to parse if searching for specific events or time periods as meet numbers have no mapping to date, nor necessarily chronology. The list also doesn't respond well, becoming a tall stack of links. Explore options for archives to work on a more date-centric system (by year? by quarter?) and ways of illustrating this that don't take up as much room (horizontally scrolling timeline?)
non_defect
redesign meet archive pagination current version is fairly difficult to parse if searching for specific events or time periods as meet numbers have no mapping to date nor necessarily chronology the list also doesn t respond well becoming a tall stack of links explore options for archives to work on a more date centric system by year by quarter and ways of illustrating this that don t take up as much room horizontally scrolling timeline
0
239,247
26,220,032,073
IssuesEvent
2023-01-04 14:10:54
tabacws-sandbox/aspnet_accord
https://api.github.com/repos/tabacws-sandbox/aspnet_accord
closed
CVE-2017-0256 (Medium) detected in system.net.http.4.3.0.nupkg - autoclosed
security vulnerability
## CVE-2017-0256 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p></summary> <p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers. </p> <p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p> <p>Path to dependency file: /tests/Conduit.IntegrationTests/Conduit.IntegrationTests.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p> <p> Dependency Hierarchy: - Conduit-1.0.0 (Root Library) - htmlsanitizer.4.0.217.nupkg - anglesharp.0.9.11.nupkg - system.net.requests.4.0.11.nupkg - :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tabacws-sandbox/aspnet_accord/commit/8d79e3658239d0d1aff62d275310f772e0298fa0">8d79e3658239d0d1aff62d275310f772e0298fa0</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A spoofing vulnerability exists when the ASP.NET Core fails to properly sanitize web requests. <p>Publish Date: 2017-05-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-0256>CVE-2017-0256</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-0256">https://nvd.nist.gov/vuln/detail/CVE-2017-0256</a></p> <p>Release Date: 2017-05-12</p> <p>Fix Resolution: Microsoft.AspNetCore.Mvc.ApiExplorer - 1.1.3,1.0.4;Microsoft.AspNetCore.Mvc.Abstractions - 1.1.3,1.0.4;Microsoft.AspNetCore.Mvc.Core - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Cors - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Localization - 1.1.3,1.0.4;System.Net.Http - 4.1.2,4.3.2;Microsoft.AspNetCore.Mvc.Razor - 1.1.3,1.0.4;System.Net.Http.WinHttpHandler - 4.0.2,4.3.0-preview1-24530-04;System.Net.Security - 4.3.0-preview1-24530-04,4.0.1;Microsoft.AspNetCore.Mvc.ViewFeatures - 1.1.3,1.0.4;Microsoft.AspNetCore.Mvc.TagHelpers - 1.0.4,1.1.3;System.Text.Encodings.Web - 4.3.0-preview1-24530-04,4.0.1;Microsoft.AspNetCore.Mvc.Razor.Host - 1.1.3,1.0.4;Microsoft.AspNetCore.Mvc.Formatters.Json - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.WebApiCompatShim - 1.0.4,1.1.3;System.Net.WebSockets.Client - 4.3.0-preview1-24530-04,4.0.1;Microsoft.AspNetCore.Mvc.Formatters.Xml - 1.1.3,1.0.4;Microsoft.AspNetCore.Mvc.DataAnnotations - 1.0.4,1.1.3</p> </p> </details> <p></p>
True
CVE-2017-0256 (Medium) detected in system.net.http.4.3.0.nupkg - autoclosed - ## CVE-2017-0256 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p></summary> <p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers. </p> <p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p> <p>Path to dependency file: /tests/Conduit.IntegrationTests/Conduit.IntegrationTests.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p> <p> Dependency Hierarchy: - Conduit-1.0.0 (Root Library) - htmlsanitizer.4.0.217.nupkg - anglesharp.0.9.11.nupkg - system.net.requests.4.0.11.nupkg - :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tabacws-sandbox/aspnet_accord/commit/8d79e3658239d0d1aff62d275310f772e0298fa0">8d79e3658239d0d1aff62d275310f772e0298fa0</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A spoofing vulnerability exists when the ASP.NET Core fails to properly sanitize web requests. <p>Publish Date: 2017-05-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-0256>CVE-2017-0256</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-0256">https://nvd.nist.gov/vuln/detail/CVE-2017-0256</a></p> <p>Release Date: 2017-05-12</p> <p>Fix Resolution: Microsoft.AspNetCore.Mvc.ApiExplorer - 1.1.3,1.0.4;Microsoft.AspNetCore.Mvc.Abstractions - 1.1.3,1.0.4;Microsoft.AspNetCore.Mvc.Core - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Cors - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Localization - 1.1.3,1.0.4;System.Net.Http - 4.1.2,4.3.2;Microsoft.AspNetCore.Mvc.Razor - 1.1.3,1.0.4;System.Net.Http.WinHttpHandler - 4.0.2,4.3.0-preview1-24530-04;System.Net.Security - 4.3.0-preview1-24530-04,4.0.1;Microsoft.AspNetCore.Mvc.ViewFeatures - 1.1.3,1.0.4;Microsoft.AspNetCore.Mvc.TagHelpers - 1.0.4,1.1.3;System.Text.Encodings.Web - 4.3.0-preview1-24530-04,4.0.1;Microsoft.AspNetCore.Mvc.Razor.Host - 1.1.3,1.0.4;Microsoft.AspNetCore.Mvc.Formatters.Json - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.WebApiCompatShim - 1.0.4,1.1.3;System.Net.WebSockets.Client - 4.3.0-preview1-24530-04,4.0.1;Microsoft.AspNetCore.Mvc.Formatters.Xml - 1.1.3,1.0.4;Microsoft.AspNetCore.Mvc.DataAnnotations - 1.0.4,1.1.3</p> </p> </details> <p></p>
non_defect
cve medium detected in system net http nupkg autoclosed cve medium severity vulnerability vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that allow applications to consume web services over http and http components that can be used by both clients and servers for parsing http headers library home page a href path to dependency file tests conduit integrationtests conduit integrationtests csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg dependency hierarchy conduit root library htmlsanitizer nupkg anglesharp nupkg system net requests nupkg x system net http nupkg vulnerable library found in head commit a href found in base branch main vulnerability details a spoofing vulnerability exists when the asp net core fails to properly sanitize web requests publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution microsoft aspnetcore mvc apiexplorer microsoft aspnetcore mvc abstractions microsoft aspnetcore mvc core microsoft aspnetcore mvc cors microsoft aspnetcore mvc localization system net http microsoft aspnetcore mvc razor system net http winhttphandler system net security microsoft aspnetcore mvc viewfeatures microsoft aspnetcore mvc taghelpers system text encodings web microsoft aspnetcore mvc razor host microsoft aspnetcore mvc formatters json microsoft aspnetcore mvc webapicompatshim system net websockets client microsoft aspnetcore mvc formatters xml microsoft aspnetcore mvc dataannotations
0
20,586
27,245,929,762
IssuesEvent
2023-02-22 02:00:09
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Wed, 22 Feb 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### ViGU: Vision GNN U-Net for Fast MRI - **Authors:** Jiahao Huang, Angelica Aviles-Rivero, Carola-Bibiane Schonlieb, Guang Yang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.10273 - **Pdf link:** https://arxiv.org/pdf/2302.10273 - **Abstract** Deep learning models have been widely applied for fast MRI. The majority of existing deep learning models, e.g., convolutional neural networks, work on data with Euclidean or regular grids structures. However, high-dimensional features extracted from MR data could be encapsulated in non-Euclidean manifolds. This disparity between the go-to assumption of existing models and data requirements limits the flexibility to capture irregular anatomical features in MR data. In this work, we introduce a novel Vision GNN type network for fast MRI called Vision GNN U-Net (ViGU). More precisely, the pixel array is first embedded into patches and then converted into a graph. Secondly, a U-shape network is developed using several graph blocks in symmetrical encoder and decoder paths. Moreover, we show that the proposed ViGU can also benefit from Generative Adversarial Networks yielding to its variant ViGU-GAN. We demonstrate, through numerical and visual experiments, that the proposed ViGU and GAN variant outperform existing CNN and GAN-based methods. Moreover, we show that the proposed network readily competes with approaches based on Transformers while requiring a fraction of the computational cost. More importantly, the graph structure of the network reveals how the network extracts features from MR images, providing intuitive explainability. ### Semantic Feature Integration network for Fine-grained Visual Classification - **Authors:** Hui Wang, Yueyang li, Haichi Luo - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.10275 - **Pdf link:** https://arxiv.org/pdf/2302.10275 - **Abstract** Fine-Grained Visual Classification (FGVC) is known as a challenging task due to subtle differences among subordinate categories. Many current FGVC approaches focus on identifying and locating discriminative regions by using the attention mechanism, but neglect the presence of unnecessary features that hinder the understanding of object structure. These unnecessary features, including 1) ambiguous parts resulting from the visual similarity in object appearances and 2) noninformative parts (e.g., background noise), can have a significant adverse impact on classification results. In this paper, we propose the Semantic Feature Integration network (SFI-Net) to address the above difficulties. By eliminating unnecessary features and reconstructing the semantic relations among discriminative features, our SFI-Net has achieved satisfying performance. The network consists of two modules: 1) the multi-level feature filter (MFF) module is proposed to remove unnecessary features with different receptive field, and then concatenate the preserved features on pixel level for subsequent disposal; 2) the semantic information reconstitution (SIR) module is presented to further establish semantic relations among discriminative features obtained from the MFF module. These two modules are carefully designed to be light-weighted and can be trained end-to-end in a weakly-supervised way. Extensive experiments on four challenging fine-grained benchmarks demonstrate that our proposed SFI-Net achieves the state-of-the-arts performance. Especially, the classification accuracy of our model on CUB-200-2011 and Stanford Dogs reaches 92.64% and 93.03%, respectively. ### Combining Blockchain and Biometrics: A Survey on Technical Aspects and a First Legal Analysis - **Authors:** Mahdi Ghafourian, Bilgesu Sumer, Ruben Vera-Rodriguez, Julian Fierrez, Ruben Tolosana, Aythami Moralez, Els Kindt - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Cryptography and Security (cs.CR); Distributed, Parallel, and Cluster Computing (cs.DC); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2302.10883 - **Pdf link:** https://arxiv.org/pdf/2302.10883 - **Abstract** Biometric recognition as a unique, hard-to-forge, and efficient way of identification and verification has become an indispensable part of the current digital world. The fast evolution of this technology has been a strong incentive for integrating it into many applications. Meanwhile, blockchain, the very attractive decentralized ledger technology, has been widely received both by the research and industry in the past years and it is being increasingly deployed nowadays in many different applications, such as money transfer, IoT, healthcare, or logistics. Recently, researchers have started to speculate what would be the pros and cons and what would be the best applications when these two technologies cross paths. This paper provides a survey of technical literature research on the combination of blockchain and biometrics and includes a first legal analysis of this integration to shed light on challenges and potentials. While this combination is still in its infancy and a growing body of literature discusses specific blockchain applications and solutions in an advanced technological set-up, this paper presents a holistic understanding of blockchains applicability in the biometric sector. This study demonstrates that combining blockchain and biometrics would be beneficial for novel applications in biometrics such as the PKI mechanism, distributed trusted service, and identity management. However, blockchain networks at their current stage are not efficient and economical for real-time applications. From a legal point of view, the allocation of accountability remains a main issue, while other difficulties remain, such as conducting a proper Data Protection Impact Assessment. Finally, it supplies technical and legal recommendations to reap the benefits and mitigate the risks of the combination. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### On Interpretable Approaches to Cluster, Classify and Represent Multi-Subspace Data via Minimum Lossy Coding Length based on Rate-Distortion Theory - **Authors:** Kai-Liang Lu, Avraham Chapman - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.10383 - **Pdf link:** https://arxiv.org/pdf/2302.10383 - **Abstract** To cluster, classify and represent are three fundamental objectives of learning from high-dimensional data with intrinsic structure. To this end, this paper introduces three interpretable approaches, i.e., segmentation (clustering) via the Minimum Lossy Coding Length criterion, classification via the Minimum Incremental Coding Length criterion and representation via the Maximal Coding Rate Reduction criterion. These are derived based on the lossy data coding and compression framework from the principle of rate distortion in information theory. These algorithms are particularly suitable for dealing with finite-sample data (allowed to be sparse or almost degenerate) of mixed Gaussian distributions or subspaces. The theoretical value and attractive features of these methods are summarized by comparison with other learning methods or evaluation criteria. This summary note aims to provide a theoretical guide to researchers (also engineers) interested in understanding 'white-box' machine (deep) learning methods. ### Device Tuning for Multi-Task Large Model - **Authors:** Penghao Jiang, Xuanchen Hou, Yinsi Zhou - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.10820 - **Pdf link:** https://arxiv.org/pdf/2302.10820 - **Abstract** Unsupervised pre-training approaches have achieved great success in many fields such as Computer Vision (CV), Natural Language Processing (NLP) and so on. However, compared to typical deep learning models, pre-training or even fine-tuning the state-of-the-art self-attention models is extremely expensive, as they require much more computational and memory resources. It severely limits their applications and success in a variety of domains, especially for multi-task learning. To improve the efficiency, we propose Device Tuning for the efficient multi-task model, which is a massively multitask framework across the cloud and device and is designed to encourage learning of representations that generalize better to many different tasks. Specifically, we design Device Tuning architecture of a multi-task model that benefits both cloud modelling and device modelling, which reduces the communication between device and cloud by representation compression. Experimental results demonstrate the effectiveness of our proposed method. ## Keyword: RAW ### Automotive RADAR sub-sampling via object detection networks: Leveraging prior signal information - **Authors:** Madhumitha Sakthi, Ahmed Tewfik, Marius Arvinte, Haris Vikalo - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Signal Processing (eess.SP) - **Arxiv link:** https://arxiv.org/abs/2302.10450 - **Pdf link:** https://arxiv.org/pdf/2302.10450 - **Abstract** Automotive radar has increasingly attracted attention due to growing interest in autonomous driving technologies. Acquiring situational awareness using multimodal data collected at high sampling rates by various sensing devices including cameras, LiDAR, and radar requires considerable power, memory and compute resources which are often limited at an edge device. In this paper, we present a novel adaptive radar sub-sampling algorithm designed to identify regions that require more detailed/accurate reconstruction based on prior environmental conditions' knowledge, enabling near-optimal performance at considerably lower effective sampling rates. Designed to robustly perform under variable weather conditions, the algorithm was shown on the Oxford raw radar and RADIATE dataset to achieve accurate reconstruction utilizing only 10% of the original samples in good weather and 20% in extreme (snow, fog) weather conditions. A further modification of the algorithm incorporates object motion to enable reliable identification of important regions. This includes monitoring possible future occlusions caused by objects detected in the present frame. Finally, we train a YOLO network on the RADIATE dataset to perform object detection directly on RADAR data and obtain a 6.6% AP50 improvement over the baseline Faster R-CNN network. ### Bokeh Rendering Based on Adaptive Depth Calibration Network - **Authors:** Lu Liu, Lei Zhou, Yuhan Dong - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.10808 - **Pdf link:** https://arxiv.org/pdf/2302.10808 - **Abstract** Bokeh rendering is a popular and effective technique used in photography to create an aesthetically pleasing effect. It is widely used to blur the background and highlight the subject in the foreground, thereby drawing the viewer's attention to the main focus of the image. In traditional digital single-lens reflex cameras (DSLRs), this effect is achieved through the use of a large aperture lens. This allows the camera to capture images with shallow depth-of-field, in which only a small area of the image is in sharp focus, while the rest of the image is blurred. However, the hardware embedded in mobile phones is typically much smaller and more limited than that found in DSLRs. Consequently, mobile phones are not able to capture natural shallow depth-of-field photos, which can be a significant limitation for mobile photography. To address this challenge, in this paper, we propose a novel method for bokeh rendering using the Vision Transformer, a recent and powerful deep learning architecture. Our approach employs an adaptive depth calibration network that acts as a confidence level to compensate for errors in monocular depth estimation. This network is used to supervise the rendering process in conjunction with depth information, allowing for the generation of high-quality bokeh images at high resolutions. Our experiments demonstrate that our proposed method outperforms state-of-the-art methods, achieving about 24.7% improvements on LPIPS and obtaining higher PSNR scores. ## Keyword: raw image There is no result
2.0
New submissions for Wed, 22 Feb 23 - ## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### ViGU: Vision GNN U-Net for Fast MRI - **Authors:** Jiahao Huang, Angelica Aviles-Rivero, Carola-Bibiane Schonlieb, Guang Yang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.10273 - **Pdf link:** https://arxiv.org/pdf/2302.10273 - **Abstract** Deep learning models have been widely applied for fast MRI. The majority of existing deep learning models, e.g., convolutional neural networks, work on data with Euclidean or regular grids structures. However, high-dimensional features extracted from MR data could be encapsulated in non-Euclidean manifolds. This disparity between the go-to assumption of existing models and data requirements limits the flexibility to capture irregular anatomical features in MR data. In this work, we introduce a novel Vision GNN type network for fast MRI called Vision GNN U-Net (ViGU). More precisely, the pixel array is first embedded into patches and then converted into a graph. Secondly, a U-shape network is developed using several graph blocks in symmetrical encoder and decoder paths. Moreover, we show that the proposed ViGU can also benefit from Generative Adversarial Networks yielding to its variant ViGU-GAN. We demonstrate, through numerical and visual experiments, that the proposed ViGU and GAN variant outperform existing CNN and GAN-based methods. Moreover, we show that the proposed network readily competes with approaches based on Transformers while requiring a fraction of the computational cost. More importantly, the graph structure of the network reveals how the network extracts features from MR images, providing intuitive explainability. ### Semantic Feature Integration network for Fine-grained Visual Classification - **Authors:** Hui Wang, Yueyang li, Haichi Luo - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.10275 - **Pdf link:** https://arxiv.org/pdf/2302.10275 - **Abstract** Fine-Grained Visual Classification (FGVC) is known as a challenging task due to subtle differences among subordinate categories. Many current FGVC approaches focus on identifying and locating discriminative regions by using the attention mechanism, but neglect the presence of unnecessary features that hinder the understanding of object structure. These unnecessary features, including 1) ambiguous parts resulting from the visual similarity in object appearances and 2) noninformative parts (e.g., background noise), can have a significant adverse impact on classification results. In this paper, we propose the Semantic Feature Integration network (SFI-Net) to address the above difficulties. By eliminating unnecessary features and reconstructing the semantic relations among discriminative features, our SFI-Net has achieved satisfying performance. The network consists of two modules: 1) the multi-level feature filter (MFF) module is proposed to remove unnecessary features with different receptive field, and then concatenate the preserved features on pixel level for subsequent disposal; 2) the semantic information reconstitution (SIR) module is presented to further establish semantic relations among discriminative features obtained from the MFF module. These two modules are carefully designed to be light-weighted and can be trained end-to-end in a weakly-supervised way. Extensive experiments on four challenging fine-grained benchmarks demonstrate that our proposed SFI-Net achieves the state-of-the-arts performance. Especially, the classification accuracy of our model on CUB-200-2011 and Stanford Dogs reaches 92.64% and 93.03%, respectively. ### Combining Blockchain and Biometrics: A Survey on Technical Aspects and a First Legal Analysis - **Authors:** Mahdi Ghafourian, Bilgesu Sumer, Ruben Vera-Rodriguez, Julian Fierrez, Ruben Tolosana, Aythami Moralez, Els Kindt - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Cryptography and Security (cs.CR); Distributed, Parallel, and Cluster Computing (cs.DC); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2302.10883 - **Pdf link:** https://arxiv.org/pdf/2302.10883 - **Abstract** Biometric recognition as a unique, hard-to-forge, and efficient way of identification and verification has become an indispensable part of the current digital world. The fast evolution of this technology has been a strong incentive for integrating it into many applications. Meanwhile, blockchain, the very attractive decentralized ledger technology, has been widely received both by the research and industry in the past years and it is being increasingly deployed nowadays in many different applications, such as money transfer, IoT, healthcare, or logistics. Recently, researchers have started to speculate what would be the pros and cons and what would be the best applications when these two technologies cross paths. This paper provides a survey of technical literature research on the combination of blockchain and biometrics and includes a first legal analysis of this integration to shed light on challenges and potentials. While this combination is still in its infancy and a growing body of literature discusses specific blockchain applications and solutions in an advanced technological set-up, this paper presents a holistic understanding of blockchains applicability in the biometric sector. This study demonstrates that combining blockchain and biometrics would be beneficial for novel applications in biometrics such as the PKI mechanism, distributed trusted service, and identity management. However, blockchain networks at their current stage are not efficient and economical for real-time applications. From a legal point of view, the allocation of accountability remains a main issue, while other difficulties remain, such as conducting a proper Data Protection Impact Assessment. Finally, it supplies technical and legal recommendations to reap the benefits and mitigate the risks of the combination. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### On Interpretable Approaches to Cluster, Classify and Represent Multi-Subspace Data via Minimum Lossy Coding Length based on Rate-Distortion Theory - **Authors:** Kai-Liang Lu, Avraham Chapman - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.10383 - **Pdf link:** https://arxiv.org/pdf/2302.10383 - **Abstract** To cluster, classify and represent are three fundamental objectives of learning from high-dimensional data with intrinsic structure. To this end, this paper introduces three interpretable approaches, i.e., segmentation (clustering) via the Minimum Lossy Coding Length criterion, classification via the Minimum Incremental Coding Length criterion and representation via the Maximal Coding Rate Reduction criterion. These are derived based on the lossy data coding and compression framework from the principle of rate distortion in information theory. These algorithms are particularly suitable for dealing with finite-sample data (allowed to be sparse or almost degenerate) of mixed Gaussian distributions or subspaces. The theoretical value and attractive features of these methods are summarized by comparison with other learning methods or evaluation criteria. This summary note aims to provide a theoretical guide to researchers (also engineers) interested in understanding 'white-box' machine (deep) learning methods. ### Device Tuning for Multi-Task Large Model - **Authors:** Penghao Jiang, Xuanchen Hou, Yinsi Zhou - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.10820 - **Pdf link:** https://arxiv.org/pdf/2302.10820 - **Abstract** Unsupervised pre-training approaches have achieved great success in many fields such as Computer Vision (CV), Natural Language Processing (NLP) and so on. However, compared to typical deep learning models, pre-training or even fine-tuning the state-of-the-art self-attention models is extremely expensive, as they require much more computational and memory resources. It severely limits their applications and success in a variety of domains, especially for multi-task learning. To improve the efficiency, we propose Device Tuning for the efficient multi-task model, which is a massively multitask framework across the cloud and device and is designed to encourage learning of representations that generalize better to many different tasks. Specifically, we design Device Tuning architecture of a multi-task model that benefits both cloud modelling and device modelling, which reduces the communication between device and cloud by representation compression. Experimental results demonstrate the effectiveness of our proposed method. ## Keyword: RAW ### Automotive RADAR sub-sampling via object detection networks: Leveraging prior signal information - **Authors:** Madhumitha Sakthi, Ahmed Tewfik, Marius Arvinte, Haris Vikalo - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Signal Processing (eess.SP) - **Arxiv link:** https://arxiv.org/abs/2302.10450 - **Pdf link:** https://arxiv.org/pdf/2302.10450 - **Abstract** Automotive radar has increasingly attracted attention due to growing interest in autonomous driving technologies. Acquiring situational awareness using multimodal data collected at high sampling rates by various sensing devices including cameras, LiDAR, and radar requires considerable power, memory and compute resources which are often limited at an edge device. In this paper, we present a novel adaptive radar sub-sampling algorithm designed to identify regions that require more detailed/accurate reconstruction based on prior environmental conditions' knowledge, enabling near-optimal performance at considerably lower effective sampling rates. Designed to robustly perform under variable weather conditions, the algorithm was shown on the Oxford raw radar and RADIATE dataset to achieve accurate reconstruction utilizing only 10% of the original samples in good weather and 20% in extreme (snow, fog) weather conditions. A further modification of the algorithm incorporates object motion to enable reliable identification of important regions. This includes monitoring possible future occlusions caused by objects detected in the present frame. Finally, we train a YOLO network on the RADIATE dataset to perform object detection directly on RADAR data and obtain a 6.6% AP50 improvement over the baseline Faster R-CNN network. ### Bokeh Rendering Based on Adaptive Depth Calibration Network - **Authors:** Lu Liu, Lei Zhou, Yuhan Dong - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.10808 - **Pdf link:** https://arxiv.org/pdf/2302.10808 - **Abstract** Bokeh rendering is a popular and effective technique used in photography to create an aesthetically pleasing effect. It is widely used to blur the background and highlight the subject in the foreground, thereby drawing the viewer's attention to the main focus of the image. In traditional digital single-lens reflex cameras (DSLRs), this effect is achieved through the use of a large aperture lens. This allows the camera to capture images with shallow depth-of-field, in which only a small area of the image is in sharp focus, while the rest of the image is blurred. However, the hardware embedded in mobile phones is typically much smaller and more limited than that found in DSLRs. Consequently, mobile phones are not able to capture natural shallow depth-of-field photos, which can be a significant limitation for mobile photography. To address this challenge, in this paper, we propose a novel method for bokeh rendering using the Vision Transformer, a recent and powerful deep learning architecture. Our approach employs an adaptive depth calibration network that acts as a confidence level to compensate for errors in monocular depth estimation. This network is used to supervise the rendering process in conjunction with depth information, allowing for the generation of high-quality bokeh images at high resolutions. Our experiments demonstrate that our proposed method outperforms state-of-the-art methods, achieving about 24.7% improvements on LPIPS and obtaining higher PSNR scores. ## Keyword: raw image There is no result
non_defect
new submissions for wed feb keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp vigu vision gnn u net for fast mri authors jiahao huang angelica aviles rivero carola bibiane schonlieb guang yang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract deep learning models have been widely applied for fast mri the majority of existing deep learning models e g convolutional neural networks work on data with euclidean or regular grids structures however high dimensional features extracted from mr data could be encapsulated in non euclidean manifolds this disparity between the go to assumption of existing models and data requirements limits the flexibility to capture irregular anatomical features in mr data in this work we introduce a novel vision gnn type network for fast mri called vision gnn u net vigu more precisely the pixel array is first embedded into patches and then converted into a graph secondly a u shape network is developed using several graph blocks in symmetrical encoder and decoder paths moreover we show that the proposed vigu can also benefit from generative adversarial networks yielding to its variant vigu gan we demonstrate through numerical and visual experiments that the proposed vigu and gan variant outperform existing cnn and gan based methods moreover we show that the proposed network readily competes with approaches based on transformers while requiring a fraction of the computational cost more importantly the graph structure of the network reveals how the network extracts features from mr images providing intuitive explainability semantic feature integration network for fine grained visual classification authors hui wang yueyang li haichi luo subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract fine grained visual classification fgvc is known as a challenging task due to subtle differences among subordinate categories many current fgvc approaches focus on identifying and locating discriminative regions by using the attention mechanism but neglect the presence of unnecessary features that hinder the understanding of object structure these unnecessary features including ambiguous parts resulting from the visual similarity in object appearances and noninformative parts e g background noise can have a significant adverse impact on classification results in this paper we propose the semantic feature integration network sfi net to address the above difficulties by eliminating unnecessary features and reconstructing the semantic relations among discriminative features our sfi net has achieved satisfying performance the network consists of two modules the multi level feature filter mff module is proposed to remove unnecessary features with different receptive field and then concatenate the preserved features on pixel level for subsequent disposal the semantic information reconstitution sir module is presented to further establish semantic relations among discriminative features obtained from the mff module these two modules are carefully designed to be light weighted and can be trained end to end in a weakly supervised way extensive experiments on four challenging fine grained benchmarks demonstrate that our proposed sfi net achieves the state of the arts performance especially the classification accuracy of our model on cub and stanford dogs reaches and respectively combining blockchain and biometrics a survey on technical aspects and a first legal analysis authors mahdi ghafourian bilgesu sumer ruben vera rodriguez julian fierrez ruben tolosana aythami moralez els kindt subjects computer vision and pattern recognition cs cv cryptography and security cs cr distributed parallel and cluster computing cs dc machine learning cs lg arxiv link pdf link abstract biometric recognition as a unique hard to forge and efficient way of identification and verification has become an indispensable part of the current digital world the fast evolution of this technology has been a strong incentive for integrating it into many applications meanwhile blockchain the very attractive decentralized ledger technology has been widely received both by the research and industry in the past years and it is being increasingly deployed nowadays in many different applications such as money transfer iot healthcare or logistics recently researchers have started to speculate what would be the pros and cons and what would be the best applications when these two technologies cross paths this paper provides a survey of technical literature research on the combination of blockchain and biometrics and includes a first legal analysis of this integration to shed light on challenges and potentials while this combination is still in its infancy and a growing body of literature discusses specific blockchain applications and solutions in an advanced technological set up this paper presents a holistic understanding of blockchains applicability in the biometric sector this study demonstrates that combining blockchain and biometrics would be beneficial for novel applications in biometrics such as the pki mechanism distributed trusted service and identity management however blockchain networks at their current stage are not efficient and economical for real time applications from a legal point of view the allocation of accountability remains a main issue while other difficulties remain such as conducting a proper data protection impact assessment finally it supplies technical and legal recommendations to reap the benefits and mitigate the risks of the combination keyword image signal processing there is no result keyword image signal process there is no result keyword compression on interpretable approaches to cluster classify and represent multi subspace data via minimum lossy coding length based on rate distortion theory authors kai liang lu avraham chapman subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract to cluster classify and represent are three fundamental objectives of learning from high dimensional data with intrinsic structure to this end this paper introduces three interpretable approaches i e segmentation clustering via the minimum lossy coding length criterion classification via the minimum incremental coding length criterion and representation via the maximal coding rate reduction criterion these are derived based on the lossy data coding and compression framework from the principle of rate distortion in information theory these algorithms are particularly suitable for dealing with finite sample data allowed to be sparse or almost degenerate of mixed gaussian distributions or subspaces the theoretical value and attractive features of these methods are summarized by comparison with other learning methods or evaluation criteria this summary note aims to provide a theoretical guide to researchers also engineers interested in understanding white box machine deep learning methods device tuning for multi task large model authors penghao jiang xuanchen hou yinsi zhou subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract unsupervised pre training approaches have achieved great success in many fields such as computer vision cv natural language processing nlp and so on however compared to typical deep learning models pre training or even fine tuning the state of the art self attention models is extremely expensive as they require much more computational and memory resources it severely limits their applications and success in a variety of domains especially for multi task learning to improve the efficiency we propose device tuning for the efficient multi task model which is a massively multitask framework across the cloud and device and is designed to encourage learning of representations that generalize better to many different tasks specifically we design device tuning architecture of a multi task model that benefits both cloud modelling and device modelling which reduces the communication between device and cloud by representation compression experimental results demonstrate the effectiveness of our proposed method keyword raw automotive radar sub sampling via object detection networks leveraging prior signal information authors madhumitha sakthi ahmed tewfik marius arvinte haris vikalo subjects computer vision and pattern recognition cs cv signal processing eess sp arxiv link pdf link abstract automotive radar has increasingly attracted attention due to growing interest in autonomous driving technologies acquiring situational awareness using multimodal data collected at high sampling rates by various sensing devices including cameras lidar and radar requires considerable power memory and compute resources which are often limited at an edge device in this paper we present a novel adaptive radar sub sampling algorithm designed to identify regions that require more detailed accurate reconstruction based on prior environmental conditions knowledge enabling near optimal performance at considerably lower effective sampling rates designed to robustly perform under variable weather conditions the algorithm was shown on the oxford raw radar and radiate dataset to achieve accurate reconstruction utilizing only of the original samples in good weather and in extreme snow fog weather conditions a further modification of the algorithm incorporates object motion to enable reliable identification of important regions this includes monitoring possible future occlusions caused by objects detected in the present frame finally we train a yolo network on the radiate dataset to perform object detection directly on radar data and obtain a improvement over the baseline faster r cnn network bokeh rendering based on adaptive depth calibration network authors lu liu lei zhou yuhan dong subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract bokeh rendering is a popular and effective technique used in photography to create an aesthetically pleasing effect it is widely used to blur the background and highlight the subject in the foreground thereby drawing the viewer s attention to the main focus of the image in traditional digital single lens reflex cameras dslrs this effect is achieved through the use of a large aperture lens this allows the camera to capture images with shallow depth of field in which only a small area of the image is in sharp focus while the rest of the image is blurred however the hardware embedded in mobile phones is typically much smaller and more limited than that found in dslrs consequently mobile phones are not able to capture natural shallow depth of field photos which can be a significant limitation for mobile photography to address this challenge in this paper we propose a novel method for bokeh rendering using the vision transformer a recent and powerful deep learning architecture our approach employs an adaptive depth calibration network that acts as a confidence level to compensate for errors in monocular depth estimation this network is used to supervise the rendering process in conjunction with depth information allowing for the generation of high quality bokeh images at high resolutions our experiments demonstrate that our proposed method outperforms state of the art methods achieving about improvements on lpips and obtaining higher psnr scores keyword raw image there is no result
0
4,987
2,610,163,458
IssuesEvent
2015-02-26 18:51:50
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Text
auto-migrated Priority-Medium Type-Defect
``` GC salvage bonus displays [MISSING] for both title and description ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 3 May 2011 at 6:42
1.0
Text - ``` GC salvage bonus displays [MISSING] for both title and description ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 3 May 2011 at 6:42
defect
text gc salvage bonus displays for both title and description original issue reported on code google com by gmail com on may at
1
76,626
21,523,095,568
IssuesEvent
2022-04-28 15:46:43
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
opened
Move scikit-learn.org/dev into a seperate repo
Documentation Build / CI
Currently, https://github.com/scikit-learn/scikit-learn.github.io grows quickly because of the pushes from `main` and the binaries in the diff (images and the zipped HTML). When I clone the repo locally, the folder is 8.8 GB. I think it's better to follow what matplotlib does: https://github.com/matplotlib/devdocs, where we force push `main` onto a "devdocs" repo, and then have `scikit-learn.com/dev` point to that repo. In release branches, we can still push to https://github.com/scikit-learn/scikit-learn.github.io. Related to https://github.com/scikit-learn/scikit-learn/issues/21171
1.0
Move scikit-learn.org/dev into a seperate repo - Currently, https://github.com/scikit-learn/scikit-learn.github.io grows quickly because of the pushes from `main` and the binaries in the diff (images and the zipped HTML). When I clone the repo locally, the folder is 8.8 GB. I think it's better to follow what matplotlib does: https://github.com/matplotlib/devdocs, where we force push `main` onto a "devdocs" repo, and then have `scikit-learn.com/dev` point to that repo. In release branches, we can still push to https://github.com/scikit-learn/scikit-learn.github.io. Related to https://github.com/scikit-learn/scikit-learn/issues/21171
non_defect
move scikit learn org dev into a seperate repo currently grows quickly because of the pushes from main and the binaries in the diff images and the zipped html when i clone the repo locally the folder is gb i think it s better to follow what matplotlib does where we force push main onto a devdocs repo and then have scikit learn com dev point to that repo in release branches we can still push to related to
0
92,333
3,869,802,934
IssuesEvent
2016-04-10 20:24:09
nucular/kirei
https://api.github.com/repos/nucular/kirei
closed
Implement: Cursor
feature high priority in progress interface
[Wiki Page](https://osu.ppy.sh/wiki/Skinning_Interface#Cursor) - interface/cursor/ - [x] cursor.svg - [x] cursortrail.svg - [x] cursormiddle.svg - [ ] cursor-smoke.svg
1.0
Implement: Cursor - [Wiki Page](https://osu.ppy.sh/wiki/Skinning_Interface#Cursor) - interface/cursor/ - [x] cursor.svg - [x] cursortrail.svg - [x] cursormiddle.svg - [ ] cursor-smoke.svg
non_defect
implement cursor interface cursor cursor svg cursortrail svg cursormiddle svg cursor smoke svg
0
24,227
3,926,874,404
IssuesEvent
2016-04-23 06:04:59
netty/netty
https://api.github.com/repos/netty/netty
reopened
deregister and re-register, ClosedChannelException is thrown
defect
Hi, I'm doing a project on middle-ware for distributed database, and I want to use Netty as client connecting to MySQL reading data. But a problem gets me stuck for such a long time so I have to ask for help. The scenario is as below: 1. Netty dispatches SQL query to MySQL; 2. MySQL returns a bunch of resultset which might be very large so I have to prevent it from OOM; 3. When the data received exceed a user-defined high water mark threshold, I use channel.deregister() to stop the incoming data for a while; 4. The business logic consumes the data received, when it reaches a low water mark threshold, I use eventloop.register(channel) to continue to receive the remaining data; 5. Do the same thing until all data needed is received and processed. But here the strange thing is when I run a benchmark tool with N concurrent clients executing the same SQL constantly, and the NioEventLoopGroup is with M threads created(N>M), after a while the N-M connections are shown closed, leaving M connections active to the end. After debugging, I find during the deregister and re-register, the channel bounded to the later failed connection will be closed(seems like the SelectionKey is set invalid) , so when trying to re-register, ClosedChannelException will be thrown. And it should be mentioned that, if I remove the deregister and re-register logic it works fine, so obviously my usage of deregister and re-register could probably be wrong in some place, could you do me a favor instructing me to locate the reason? Thank you so much in advance. ```java Bootstrap bootstrap = new Bootstrap(); bootstrap .group(MySQLConnectorConfig.group) .channel(NioSocketChannel.class) .handler(new ChannelInitializer<SocketChannel>() { @Override protected void initChannel(SocketChannel ch) throws Exception { ch.pipeline().addLast(new MySQLClientDecoder(conn)); ch.pipeline().addLast(new MySQLClientHandler(conn)); ch.config().setAllocator(PooledByteBufAllocator.DEFAULT); } }) .option(ChannelOption.SO_KEEPALIVE, true) .option(ChannelOption.TCP_NODELAY, true) .option(ChannelOption.SO_REUSEADDR, true) .option(ChannelOption.SO_RCVBUF, MySQLConnectorConfig.DEFAULT_SO_RCVBUF); ``` I deregister the channel in MySQLClientDecoder, and re-register in my business logic when consuming the data received.
1.0
deregister and re-register, ClosedChannelException is thrown - Hi, I'm doing a project on middle-ware for distributed database, and I want to use Netty as client connecting to MySQL reading data. But a problem gets me stuck for such a long time so I have to ask for help. The scenario is as below: 1. Netty dispatches SQL query to MySQL; 2. MySQL returns a bunch of resultset which might be very large so I have to prevent it from OOM; 3. When the data received exceed a user-defined high water mark threshold, I use channel.deregister() to stop the incoming data for a while; 4. The business logic consumes the data received, when it reaches a low water mark threshold, I use eventloop.register(channel) to continue to receive the remaining data; 5. Do the same thing until all data needed is received and processed. But here the strange thing is when I run a benchmark tool with N concurrent clients executing the same SQL constantly, and the NioEventLoopGroup is with M threads created(N>M), after a while the N-M connections are shown closed, leaving M connections active to the end. After debugging, I find during the deregister and re-register, the channel bounded to the later failed connection will be closed(seems like the SelectionKey is set invalid) , so when trying to re-register, ClosedChannelException will be thrown. And it should be mentioned that, if I remove the deregister and re-register logic it works fine, so obviously my usage of deregister and re-register could probably be wrong in some place, could you do me a favor instructing me to locate the reason? Thank you so much in advance. ```java Bootstrap bootstrap = new Bootstrap(); bootstrap .group(MySQLConnectorConfig.group) .channel(NioSocketChannel.class) .handler(new ChannelInitializer<SocketChannel>() { @Override protected void initChannel(SocketChannel ch) throws Exception { ch.pipeline().addLast(new MySQLClientDecoder(conn)); ch.pipeline().addLast(new MySQLClientHandler(conn)); ch.config().setAllocator(PooledByteBufAllocator.DEFAULT); } }) .option(ChannelOption.SO_KEEPALIVE, true) .option(ChannelOption.TCP_NODELAY, true) .option(ChannelOption.SO_REUSEADDR, true) .option(ChannelOption.SO_RCVBUF, MySQLConnectorConfig.DEFAULT_SO_RCVBUF); ``` I deregister the channel in MySQLClientDecoder, and re-register in my business logic when consuming the data received.
defect
deregister and re register closedchannelexception is thrown hi i m doing a project on middle ware for distributed database and i want to use netty as client connecting to mysql reading data but a problem gets me stuck for such a long time so i have to ask for help the scenario is as below netty dispatches sql query to mysql mysql returns a bunch of resultset which might be very large so i have to prevent it from oom when the data received exceed a user defined high water mark threshold i use channel deregister to stop the incoming data for a while the business logic consumes the data received when it reaches a low water mark threshold i use eventloop register channel to continue to receive the remaining data do the same thing until all data needed is received and processed but here the strange thing is when i run a benchmark tool with n concurrent clients executing the same sql constantly and the nioeventloopgroup is with m threads created n m after a while the n m connections are shown closed leaving m connections active to the end after debugging i find during the deregister and re register the channel bounded to the later failed connection will be closed seems like the selectionkey is set invalid so when trying to re register closedchannelexception will be thrown and it should be mentioned that if i remove the deregister and re register logic it works fine so obviously my usage of deregister and re register could probably be wrong in some place could you do me a favor instructing me to locate the reason thank you so much in advance java bootstrap bootstrap new bootstrap bootstrap group mysqlconnectorconfig group channel niosocketchannel class handler new channelinitializer override protected void initchannel socketchannel ch throws exception ch pipeline addlast new mysqlclientdecoder conn ch pipeline addlast new mysqlclienthandler conn ch config setallocator pooledbytebufallocator default option channeloption so keepalive true option channeloption tcp nodelay true option channeloption so reuseaddr true option channeloption so rcvbuf mysqlconnectorconfig default so rcvbuf i deregister the channel in mysqlclientdecoder and re register in my business logic when consuming the data received
1
43,476
7,047,760,796
IssuesEvent
2018-01-02 14:58:51
camlistore/camlistore
https://api.github.com/repos/camlistore/camlistore
closed
remove /doc/search-ui.txt after v0.10.0 release
Documentation
up through v0.9.0, camlistore included a direct link to https://camlistore.googlesource.com/camlistore/+/master/doc/search-ui.txt in the help page. With the recent doc cleanup, this file got renamed to `search-ui.md` in 620d837a3d94cd939b16c56ece039639ef724df5, which broke that link. As a temporary fix, the `search-ui.txt` file was recreated in 29dcc70f3c6a9ad471eb30c3f08e70da3d570bc9 with a simple pointer to https://camlistore.org/doc/search-ui while also updating the help page inside camlistore. Once v0.10.0 is released (or soon after), we should be able to remove the temporary `search-ui.txt` file.
1.0
remove /doc/search-ui.txt after v0.10.0 release - up through v0.9.0, camlistore included a direct link to https://camlistore.googlesource.com/camlistore/+/master/doc/search-ui.txt in the help page. With the recent doc cleanup, this file got renamed to `search-ui.md` in 620d837a3d94cd939b16c56ece039639ef724df5, which broke that link. As a temporary fix, the `search-ui.txt` file was recreated in 29dcc70f3c6a9ad471eb30c3f08e70da3d570bc9 with a simple pointer to https://camlistore.org/doc/search-ui while also updating the help page inside camlistore. Once v0.10.0 is released (or soon after), we should be able to remove the temporary `search-ui.txt` file.
non_defect
remove doc search ui txt after release up through camlistore included a direct link to in the help page with the recent doc cleanup this file got renamed to search ui md in which broke that link as a temporary fix the search ui txt file was recreated in with a simple pointer to while also updating the help page inside camlistore once is released or soon after we should be able to remove the temporary search ui txt file
0
52,325
13,224,653,332
IssuesEvent
2020-08-17 19:34:14
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
[docs] switch to python-based documentation builder (Trac #2025)
Incomplete Migration Migrated from Trac analysis defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2025">https://code.icecube.wisc.edu/projects/icecube/ticket/2025</a>, reported by kjmeagherand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:13:57", "_ts": "1550067237750774", "description": "speed up documentation build by switching to python-based doc builder `docs-build`. Use the `-jN` option on multicore systems. this is in its own project which was added to combo \nhttp://code.icecube.wisc.edu/icetray/projects/docs/trunk", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2017-05-18T07:11:18", "component": "analysis", "summary": "[docs] switch to python-based documentation builder", "priority": "normal", "keywords": "", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
[docs] switch to python-based documentation builder (Trac #2025) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2025">https://code.icecube.wisc.edu/projects/icecube/ticket/2025</a>, reported by kjmeagherand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:13:57", "_ts": "1550067237750774", "description": "speed up documentation build by switching to python-based doc builder `docs-build`. Use the `-jN` option on multicore systems. this is in its own project which was added to combo \nhttp://code.icecube.wisc.edu/icetray/projects/docs/trunk", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2017-05-18T07:11:18", "component": "analysis", "summary": "[docs] switch to python-based documentation builder", "priority": "normal", "keywords": "", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
switch to python based documentation builder trac migrated from json status closed changetime ts description speed up documentation build by switching to python based doc builder docs build use the jn option on multicore systems this is in its own project which was added to combo n reporter kjmeagher cc resolution fixed time component analysis summary switch to python based documentation builder priority normal keywords milestone owner nega type defect
1
24,616
5,089,727,775
IssuesEvent
2017-01-01 20:44:48
pixijs/pixi.js
https://api.github.com/repos/pixijs/pixi.js
closed
Broken gallery picture link in README.md
Domain: Documentation Priority: Low Type: Bug Version: v4.x
There's a pointer to a non-existent jpg where I assume a picture of the gallery of demos should be: http://www.pixijs.com/wp-content/uploads/2013/05/headerPanel_projects-898x342.jpg
1.0
Broken gallery picture link in README.md - There's a pointer to a non-existent jpg where I assume a picture of the gallery of demos should be: http://www.pixijs.com/wp-content/uploads/2013/05/headerPanel_projects-898x342.jpg
non_defect
broken gallery picture link in readme md there s a pointer to a non existent jpg where i assume a picture of the gallery of demos should be
0
76,690
26,556,314,377
IssuesEvent
2023-01-20 12:20:38
DependencyTrack/dependency-track
https://api.github.com/repos/DependencyTrack/dependency-track
opened
Risk Outdated component shows a current version that seems misleading
defect in triage
### Current Behavior Cdxgen version 6.0.12 used to create a SBOM against that was ingested to Dependency Track v4.7.0 The component jQuery has version 3.2.1 but it is highlighted as risk: outdated component, latest version shows 1.7.4 Screenshot below <img width="859" alt="image" src="https://user-images.githubusercontent.com/73107830/213692712-03d4d674-052f-489a-8b92-16ec06e813f4.png"> Snippet of the sbom "components": [ { "publisher": "", "group": "", "name": "jQuery", "version": "3.2.1", "description": "jQuery: The Write Less, Do More, JavaScript Library (packaged for Node.JS)", "licenses": [], "purl": "pkg:npm/jQuery@3.2.1", "externalReferences": [ { "type": "website", "url": "https://github.com/coolaj86/node-jquery" }, { "type": "vcs", "url": "git://github.com/coolaj86/node-jquery.git" } ], "type": "library", "bom-ref": "pkg:npm/jQuery@3.2.1", "properties": [ { "name": "SrcFile", "value": "/var/jenkins_home/workspace/cdxgen-sbom-repo/shiftleft-java-demo/src/main/resources/static/jquery.min.js" } ] }, ### Steps to Reproduce 1. Clone repo https://github.com/ShiftLeftSecurity/shiftleft-java-demo 2. Generate a CycloneDX SBOM using cdxgen v6.0.12 3. Ingest the SBOM to Dependency Track v4.7.0 ### Expected Behavior The Risk Outdated component shown for the current version vis-a-vis latest version should not be misleading ### Dependency-Track Version 4.7.0 ### Dependency-Track Distribution Container Image ### Database Server PostgreSQL ### Database Server Version 13.2-alpine ### Browser Google Chrome ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
1.0
Risk Outdated component shows a current version that seems misleading - ### Current Behavior Cdxgen version 6.0.12 used to create a SBOM against that was ingested to Dependency Track v4.7.0 The component jQuery has version 3.2.1 but it is highlighted as risk: outdated component, latest version shows 1.7.4 Screenshot below <img width="859" alt="image" src="https://user-images.githubusercontent.com/73107830/213692712-03d4d674-052f-489a-8b92-16ec06e813f4.png"> Snippet of the sbom "components": [ { "publisher": "", "group": "", "name": "jQuery", "version": "3.2.1", "description": "jQuery: The Write Less, Do More, JavaScript Library (packaged for Node.JS)", "licenses": [], "purl": "pkg:npm/jQuery@3.2.1", "externalReferences": [ { "type": "website", "url": "https://github.com/coolaj86/node-jquery" }, { "type": "vcs", "url": "git://github.com/coolaj86/node-jquery.git" } ], "type": "library", "bom-ref": "pkg:npm/jQuery@3.2.1", "properties": [ { "name": "SrcFile", "value": "/var/jenkins_home/workspace/cdxgen-sbom-repo/shiftleft-java-demo/src/main/resources/static/jquery.min.js" } ] }, ### Steps to Reproduce 1. Clone repo https://github.com/ShiftLeftSecurity/shiftleft-java-demo 2. Generate a CycloneDX SBOM using cdxgen v6.0.12 3. Ingest the SBOM to Dependency Track v4.7.0 ### Expected Behavior The Risk Outdated component shown for the current version vis-a-vis latest version should not be misleading ### Dependency-Track Version 4.7.0 ### Dependency-Track Distribution Container Image ### Database Server PostgreSQL ### Database Server Version 13.2-alpine ### Browser Google Chrome ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
defect
risk outdated component shows a current version that seems misleading current behavior cdxgen version used to create a sbom against that was ingested to dependency track the component jquery has version but it is highlighted as risk outdated component latest version shows screenshot below img width alt image src snippet of the sbom components publisher group name jquery version description jquery the write less do more javascript library packaged for node js licenses purl pkg npm jquery externalreferences type website url type vcs url git github com node jquery git type library bom ref pkg npm jquery properties name srcfile value var jenkins home workspace cdxgen sbom repo shiftleft java demo src main resources static jquery min js steps to reproduce clone repo generate a cyclonedx sbom using cdxgen ingest the sbom to dependency track expected behavior the risk outdated component shown for the current version vis a vis latest version should not be misleading dependency track version dependency track distribution container image database server postgresql database server version alpine browser google chrome checklist i have read and understand the i have checked the for whether this defect was already reported
1
46,272
7,245,513,828
IssuesEvent
2018-02-14 18:21:32
google/blockly
https://api.github.com/repos/google/blockly
closed
Hard to implement a custom field
type: documentation type: feature request
Microsoft Make Code implements custom fields by injecting lots of code and logic. Is it possible to provide a super class for extending Fields?
1.0
Hard to implement a custom field - Microsoft Make Code implements custom fields by injecting lots of code and logic. Is it possible to provide a super class for extending Fields?
non_defect
hard to implement a custom field microsoft make code implements custom fields by injecting lots of code and logic is it possible to provide a super class for extending fields
0
67,093
20,870,495,557
IssuesEvent
2022-03-22 11:28:17
BOINC/boinc
https://api.github.com/repos/BOINC/boinc
closed
7.16.16 Size of the font for the tabs in Advanced view scales badly with the rest of BOINC Manager
C: Manager P: Minor T: Defect E: 1 day
7.16.16 All right, so the font size for the tabs has been adjusted and now the text fits. But the text size is the same whether I use a small window of BOINC Manager or a screen filling one (3840x2160). I mean... just look at it. Does this scale correctly? https://i.ibb.co/kJzY7J1/7-16-16-tabs-font-size-1.png https://i.ibb.co/PMkVdP8/7-16-16-tabs-font-size-2.png
1.0
7.16.16 Size of the font for the tabs in Advanced view scales badly with the rest of BOINC Manager - 7.16.16 All right, so the font size for the tabs has been adjusted and now the text fits. But the text size is the same whether I use a small window of BOINC Manager or a screen filling one (3840x2160). I mean... just look at it. Does this scale correctly? https://i.ibb.co/kJzY7J1/7-16-16-tabs-font-size-1.png https://i.ibb.co/PMkVdP8/7-16-16-tabs-font-size-2.png
defect
size of the font for the tabs in advanced view scales badly with the rest of boinc manager all right so the font size for the tabs has been adjusted and now the text fits but the text size is the same whether i use a small window of boinc manager or a screen filling one i mean just look at it does this scale correctly
1
429,681
30,085,235,840
IssuesEvent
2023-06-29 08:08:43
bokeh/dataviz-fundamentals
https://api.github.com/repos/bokeh/dataviz-fundamentals
closed
Introduction files
tag: documentation tag: notebook
The `Introduction.ipynb` and `Introduction.html` files till contain the old file paths. I have to modify them to refelct the new file paths and also reflect the style guide for subsequent posts.
1.0
Introduction files - The `Introduction.ipynb` and `Introduction.html` files till contain the old file paths. I have to modify them to refelct the new file paths and also reflect the style guide for subsequent posts.
non_defect
introduction files the introduction ipynb and introduction html files till contain the old file paths i have to modify them to refelct the new file paths and also reflect the style guide for subsequent posts
0
32,560
6,826,110,371
IssuesEvent
2017-11-08 13:03:48
ontop/ontop
https://api.github.com/repos/ontop/ontop
closed
Bug obtaining the datatypes of variables.
status: fixed topic: mapping processing type: defect
Hi, I found a bug a month ago in the file MappingDataTypeCompletion.java, concretely in the following function: ``` private static Map<String, List<IndexedPosition>> createIndex(List<Function> body) { Map<String, List<IndexedPosition>> termOccurenceIndex = new HashMap<>(); for (Function a : body) { List<Term> terms = a.getTerms(); int i = 1; // position index for (Term t : terms) { if (t instanceof Variable) { Variable var = (Variable) t; List<IndexedPosition> aux = termOccurenceIndex.get(var.getName()); if (aux == null) aux = new LinkedList<>(); aux.add(new IndexedPosition(a, i)); termOccurenceIndex.put(var.getName(), aux); i++; // increase the position index for the next variable } else if (t instanceof FunctionalTermImpl) { // NO-OP } else if (t instanceof ValueConstant) { // NO-OP } else if (t instanceof URIConstant) { // NO-OP } } } return termOccurenceIndex; } ``` The problem occurs when there are terms that are not variables (for example, a constant). In this case, the iterator "i" is not incremented, which causes a shift between the pointer and the position of the terms in the list of terms. I proposed the fix #213, which consist of the increment of i in all cases. I copy the text I wrote in the pull request for explanatory reasons: > There was a bug obtaining the data type of the variables. > > The application creates an index based on IndexedPosition object. This object have the property "pos", which indicates the index of a variable in a list of terms. The program iterates over all terms in the list and it checks if the term is a variable. If there exist constants in the list of terms, the pointer that iterates the list of terms was not updated. For instance, if we have the following terms: > var1 var2 const1 var3 > Then the created index was: > var pos > var1 0 > var2 1 > var3 2 > > This causes an error when var3 is accessed via terms.get(2) due to this position is occupied by "const1". What i've done to fix this is to increment the iterator for each term in the list and not only when the term is a variable. Cheers, Fran.
1.0
Bug obtaining the datatypes of variables. - Hi, I found a bug a month ago in the file MappingDataTypeCompletion.java, concretely in the following function: ``` private static Map<String, List<IndexedPosition>> createIndex(List<Function> body) { Map<String, List<IndexedPosition>> termOccurenceIndex = new HashMap<>(); for (Function a : body) { List<Term> terms = a.getTerms(); int i = 1; // position index for (Term t : terms) { if (t instanceof Variable) { Variable var = (Variable) t; List<IndexedPosition> aux = termOccurenceIndex.get(var.getName()); if (aux == null) aux = new LinkedList<>(); aux.add(new IndexedPosition(a, i)); termOccurenceIndex.put(var.getName(), aux); i++; // increase the position index for the next variable } else if (t instanceof FunctionalTermImpl) { // NO-OP } else if (t instanceof ValueConstant) { // NO-OP } else if (t instanceof URIConstant) { // NO-OP } } } return termOccurenceIndex; } ``` The problem occurs when there are terms that are not variables (for example, a constant). In this case, the iterator "i" is not incremented, which causes a shift between the pointer and the position of the terms in the list of terms. I proposed the fix #213, which consist of the increment of i in all cases. I copy the text I wrote in the pull request for explanatory reasons: > There was a bug obtaining the data type of the variables. > > The application creates an index based on IndexedPosition object. This object have the property "pos", which indicates the index of a variable in a list of terms. The program iterates over all terms in the list and it checks if the term is a variable. If there exist constants in the list of terms, the pointer that iterates the list of terms was not updated. For instance, if we have the following terms: > var1 var2 const1 var3 > Then the created index was: > var pos > var1 0 > var2 1 > var3 2 > > This causes an error when var3 is accessed via terms.get(2) due to this position is occupied by "const1". What i've done to fix this is to increment the iterator for each term in the list and not only when the term is a variable. Cheers, Fran.
defect
bug obtaining the datatypes of variables hi i found a bug a month ago in the file mappingdatatypecompletion java concretely in the following function private static map createindex list body map termoccurenceindex new hashmap for function a body list terms a getterms int i position index for term t terms if t instanceof variable variable var variable t list aux termoccurenceindex get var getname if aux null aux new linkedlist aux add new indexedposition a i termoccurenceindex put var getname aux i increase the position index for the next variable else if t instanceof functionaltermimpl no op else if t instanceof valueconstant no op else if t instanceof uriconstant no op return termoccurenceindex the problem occurs when there are terms that are not variables for example a constant in this case the iterator i is not incremented which causes a shift between the pointer and the position of the terms in the list of terms i proposed the fix which consist of the increment of i in all cases i copy the text i wrote in the pull request for explanatory reasons there was a bug obtaining the data type of the variables the application creates an index based on indexedposition object this object have the property pos which indicates the index of a variable in a list of terms the program iterates over all terms in the list and it checks if the term is a variable if there exist constants in the list of terms the pointer that iterates the list of terms was not updated for instance if we have the following terms then the created index was var pos this causes an error when is accessed via terms get due to this position is occupied by what i ve done to fix this is to increment the iterator for each term in the list and not only when the term is a variable cheers fran
1
75,916
26,149,434,434
IssuesEvent
2022-12-30 11:10:33
PyTables/PyTables
https://api.github.com/repos/PyTables/PyTables
closed
Please make it possible to use system c-blosc2 installation (without blosc2 Python package)
defect windows help wanted
Currently, `setup.py` forces `blosc2` Python package to be installed and forces blosc2 paths from it. Could you please make this optional and us use system-installed blosc2 like other compression libraries? To confirm that it works, I have used the following hack locally: ```diff diff --git a/setup.py b/setup.py index 966be028..2708deb9 100755 --- a/setup.py +++ b/setup.py @@ -262,11 +262,11 @@ if __name__ == "__main__": debug = "--debug" in sys.argv - blosc2_inc, blosc2_lib = get_blosc2_directories() + blosc2_inc, blosc2_lib = "", "" # Global variables lib_dirs = [blosc2_lib] - inc_dirs = [Path("hdf5-blosc/src"), Path("hdf5-blosc2/src"), blosc2_inc] + inc_dirs = [Path("hdf5-blosc/src"), Path("hdf5-blosc2/src")] optional_libs = [] copy_libs = [] ```
1.0
Please make it possible to use system c-blosc2 installation (without blosc2 Python package) - Currently, `setup.py` forces `blosc2` Python package to be installed and forces blosc2 paths from it. Could you please make this optional and us use system-installed blosc2 like other compression libraries? To confirm that it works, I have used the following hack locally: ```diff diff --git a/setup.py b/setup.py index 966be028..2708deb9 100755 --- a/setup.py +++ b/setup.py @@ -262,11 +262,11 @@ if __name__ == "__main__": debug = "--debug" in sys.argv - blosc2_inc, blosc2_lib = get_blosc2_directories() + blosc2_inc, blosc2_lib = "", "" # Global variables lib_dirs = [blosc2_lib] - inc_dirs = [Path("hdf5-blosc/src"), Path("hdf5-blosc2/src"), blosc2_inc] + inc_dirs = [Path("hdf5-blosc/src"), Path("hdf5-blosc2/src")] optional_libs = [] copy_libs = [] ```
defect
please make it possible to use system c installation without python package currently setup py forces python package to be installed and forces paths from it could you please make this optional and us use system installed like other compression libraries to confirm that it works i have used the following hack locally diff diff git a setup py b setup py index a setup py b setup py if name main debug debug in sys argv inc lib get directories inc lib global variables lib dirs inc dirs inc dirs optional libs copy libs
1
228,526
7,552,536,149
IssuesEvent
2018-04-19 00:53:13
kubernetes/kubeadm
https://api.github.com/repos/kubernetes/kubeadm
closed
Front proxy ca private key has to be shared accross all nodes
active area/security kind/bug priority/important-longterm sig/cluster-lifecycle
FEATURE REQUEST You can externalize the CA and delegate tls cert management to a third party But actually, you can't do it with the front proxy ca as kubeadm requires the private key to be installed on all master nodes. see: https://github.com/kubernetes/kubernetes/blob/master/cmd/kubeadm/app/phases/certs/certs.go#L639 ## Versions **kubeadm version** (use `kubeadm version`): 1.10.0 ## What you expected That the front ca key is not required as the front proxy client tls keypair can be managed by the third party.
1.0
Front proxy ca private key has to be shared accross all nodes - FEATURE REQUEST You can externalize the CA and delegate tls cert management to a third party But actually, you can't do it with the front proxy ca as kubeadm requires the private key to be installed on all master nodes. see: https://github.com/kubernetes/kubernetes/blob/master/cmd/kubeadm/app/phases/certs/certs.go#L639 ## Versions **kubeadm version** (use `kubeadm version`): 1.10.0 ## What you expected That the front ca key is not required as the front proxy client tls keypair can be managed by the third party.
non_defect
front proxy ca private key has to be shared accross all nodes feature request you can externalize the ca and delegate tls cert management to a third party but actually you can t do it with the front proxy ca as kubeadm requires the private key to be installed on all master nodes see versions kubeadm version use kubeadm version what you expected that the front ca key is not required as the front proxy client tls keypair can be managed by the third party
0
731,635
25,225,187,917
IssuesEvent
2022-11-14 15:29:52
DISSINET/InkVisitor
https://api.github.com/repos/DISSINET/InkVisitor
closed
Incorrect GUI handling of inverse relations in asymmetrical Relations
bug gui priority
@adammertel, @Ptrhnk: There is a misunderstanding re: reciprocation in _assymmetrical_ Relations in GUI detail (this is just GUI and not data model, at least I hope). Change is needed there. The problematic Relations are those which are marked as Assymmetrical = Y in [our table](https://docs.google.com/document/d/1cqr-kIF_W6voea7RHjf1vFv-BRKRdcFIiqDLLfjK7BE/edit#bookmark=id.lzu4f21zshfb). For those which are assymetrical = Y, the inverse relation should not be displayed in this section and under this Relation name at all. One is Holonymy (treated in #1293 - i.e. solving #1313 will solve also #1293), the other is Superclass: the subclasses _must not_ appear under the label Superclass (see image). ![obrazek](https://user-images.githubusercontent.com/55400794/197336624-7e664698-26b3-47d6-8dde-f748a06394f1.png) This, incl. the mixing of superclasses and subclasses under the same section, is incorrect. This concerns all those marked as Assymmetrical = Y, i.e.: - Superclass - SuperordinateLocation - SubjectSemantics - Actant1Semantics - Actant2Semantics - Classification - Holonym - Implication Instead of displaying them under the incorrect (inverse) name in the same section/table, let's create, after Relations section in Detail, a new section "Used in Relations", and in this section, list uneditably the inverse relation (similar logic to Used in Metaprops). You will find the heading of the inverse relation in the new col. "inverse relation name" in the [table in our gdoc](https://docs.google.com/document/d/1cqr-kIF_W6voea7RHjf1vFv-BRKRdcFIiqDLLfjK7BE/edit#bookmark=id.lzu4f21zshfb). As I say, this only concerns the assymmetrical relations. For symmetrical or cloud ones, no problem in the current display. That means: in this new section, only display the assymmetrical ones. Sorry about spotting this later than I should have (but I this "Used in" section was a part of my original idea, as implicitly noted in #1215). No problem, let's swiftly change this - it is a priority now since the current display is ontologically incorrect and editorially confusing.
1.0
Incorrect GUI handling of inverse relations in asymmetrical Relations - @adammertel, @Ptrhnk: There is a misunderstanding re: reciprocation in _assymmetrical_ Relations in GUI detail (this is just GUI and not data model, at least I hope). Change is needed there. The problematic Relations are those which are marked as Assymmetrical = Y in [our table](https://docs.google.com/document/d/1cqr-kIF_W6voea7RHjf1vFv-BRKRdcFIiqDLLfjK7BE/edit#bookmark=id.lzu4f21zshfb). For those which are assymetrical = Y, the inverse relation should not be displayed in this section and under this Relation name at all. One is Holonymy (treated in #1293 - i.e. solving #1313 will solve also #1293), the other is Superclass: the subclasses _must not_ appear under the label Superclass (see image). ![obrazek](https://user-images.githubusercontent.com/55400794/197336624-7e664698-26b3-47d6-8dde-f748a06394f1.png) This, incl. the mixing of superclasses and subclasses under the same section, is incorrect. This concerns all those marked as Assymmetrical = Y, i.e.: - Superclass - SuperordinateLocation - SubjectSemantics - Actant1Semantics - Actant2Semantics - Classification - Holonym - Implication Instead of displaying them under the incorrect (inverse) name in the same section/table, let's create, after Relations section in Detail, a new section "Used in Relations", and in this section, list uneditably the inverse relation (similar logic to Used in Metaprops). You will find the heading of the inverse relation in the new col. "inverse relation name" in the [table in our gdoc](https://docs.google.com/document/d/1cqr-kIF_W6voea7RHjf1vFv-BRKRdcFIiqDLLfjK7BE/edit#bookmark=id.lzu4f21zshfb). As I say, this only concerns the assymmetrical relations. For symmetrical or cloud ones, no problem in the current display. That means: in this new section, only display the assymmetrical ones. Sorry about spotting this later than I should have (but I this "Used in" section was a part of my original idea, as implicitly noted in #1215). No problem, let's swiftly change this - it is a priority now since the current display is ontologically incorrect and editorially confusing.
non_defect
incorrect gui handling of inverse relations in asymmetrical relations adammertel ptrhnk there is a misunderstanding re reciprocation in assymmetrical relations in gui detail this is just gui and not data model at least i hope change is needed there the problematic relations are those which are marked as assymmetrical y in for those which are assymetrical y the inverse relation should not be displayed in this section and under this relation name at all one is holonymy treated in i e solving will solve also the other is superclass the subclasses must not appear under the label superclass see image this incl the mixing of superclasses and subclasses under the same section is incorrect this concerns all those marked as assymmetrical y i e superclass superordinatelocation subjectsemantics classification holonym implication instead of displaying them under the incorrect inverse name in the same section table let s create after relations section in detail a new section used in relations and in this section list uneditably the inverse relation similar logic to used in metaprops you will find the heading of the inverse relation in the new col inverse relation name in the as i say this only concerns the assymmetrical relations for symmetrical or cloud ones no problem in the current display that means in this new section only display the assymmetrical ones sorry about spotting this later than i should have but i this used in section was a part of my original idea as implicitly noted in no problem let s swiftly change this it is a priority now since the current display is ontologically incorrect and editorially confusing
0