Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
241,897
| 20,170,893,265
|
IssuesEvent
|
2022-02-10 10:18:10
|
spacemeshos/go-spacemesh
|
https://api.github.com/repos/spacemeshos/go-spacemesh
|
closed
|
p2p node starts with new public key and chooses itself as a neighbor
|
bug P2P testnet
|
# Overview / Motivation
when a node restarts, if we didn't persist an identity we generate new p2p identity. this means we might get peers through discovery that are actually ourself. (same ip:port but different pubkey). we'll accept those peers and we might choose them as neighbors which will result in sending messages to ourselves.
# The Task
two levels of fixing that issue
- make sure nodes responding to discovery query won't return addresses with the requester's address
- think about a more defensive way of protecting against that from the restarted node's perspective.
# Implementation Notes
TODO: Add links to relevant resources, specs, related issues, etc...
# Contribution Guidelines
Important: Issue assignment to developers will be by the order of their application and proficiency level according to the tasks complexity. We will not assign tasks to developers who have'nt introduced themselves on our Gitter [dev channel](https://gitter.im/spacemesh-os/Lobby)
1. Introduce yourself on go-spacemesh [dev chat channel](https://gitter.im/spacemesh-os/Lobby) - ask our team any question you may have about this task
2. Fork branch `develop` to your own repo and work in your repo
3. You must document all methods, enums and types with [godoc comments](https://blog.golang.org/godoc-documenting-go-code)
4. You must write go unit tests for all types and methods when submitting a component, and integration tests if you submit a feature
5. When ready for code review, submit a PR from your repo back to branch `develop`
6. Attach relevant issue to PR
|
1.0
|
p2p node starts with new public key and chooses itself as a neighbor - # Overview / Motivation
when a node restarts, if we didn't persist an identity we generate new p2p identity. this means we might get peers through discovery that are actually ourself. (same ip:port but different pubkey). we'll accept those peers and we might choose them as neighbors which will result in sending messages to ourselves.
# The Task
two levels of fixing that issue
- make sure nodes responding to discovery query won't return addresses with the requester's address
- think about a more defensive way of protecting against that from the restarted node's perspective.
# Implementation Notes
TODO: Add links to relevant resources, specs, related issues, etc...
# Contribution Guidelines
Important: Issue assignment to developers will be by the order of their application and proficiency level according to the tasks complexity. We will not assign tasks to developers who have'nt introduced themselves on our Gitter [dev channel](https://gitter.im/spacemesh-os/Lobby)
1. Introduce yourself on go-spacemesh [dev chat channel](https://gitter.im/spacemesh-os/Lobby) - ask our team any question you may have about this task
2. Fork branch `develop` to your own repo and work in your repo
3. You must document all methods, enums and types with [godoc comments](https://blog.golang.org/godoc-documenting-go-code)
4. You must write go unit tests for all types and methods when submitting a component, and integration tests if you submit a feature
5. When ready for code review, submit a PR from your repo back to branch `develop`
6. Attach relevant issue to PR
|
non_process
|
node starts with new public key and chooses itself as a neighbor overview motivation when a node restarts if we didn t persist an identity we generate new identity this means we might get peers through discovery that are actually ourself same ip port but different pubkey we ll accept those peers and we might choose them as neighbors which will result in sending messages to ourselves the task two levels of fixing that issue make sure nodes responding to discovery query won t return addresses with the requester s address think about a more defensive way of protecting against that from the restarted node s perspective implementation notes todo add links to relevant resources specs related issues etc contribution guidelines important issue assignment to developers will be by the order of their application and proficiency level according to the tasks complexity we will not assign tasks to developers who have nt introduced themselves on our gitter introduce yourself on go spacemesh ask our team any question you may have about this task fork branch develop to your own repo and work in your repo you must document all methods enums and types with you must write go unit tests for all types and methods when submitting a component and integration tests if you submit a feature when ready for code review submit a pr from your repo back to branch develop attach relevant issue to pr
| 0
|
6,206
| 9,108,233,538
|
IssuesEvent
|
2019-02-21 07:53:44
|
FACK1/ReservationSystem
|
https://api.github.com/repos/FACK1/ReservationSystem
|
reopened
|
return service' events, and other details
|
inProcess technical
|
- [ ] update `/events/` endpoint to `/events/:serviceId` that to return all of the requested service' details.
- [ ] since from `homePage` which `services` page and by using `findAll` query that will return all details for all sercives.
- [ ] service details passing to the `bookEvent` page based on the new endpoint `/events/:serviceId.
|
1.0
|
return service' events, and other details - - [ ] update `/events/` endpoint to `/events/:serviceId` that to return all of the requested service' details.
- [ ] since from `homePage` which `services` page and by using `findAll` query that will return all details for all sercives.
- [ ] service details passing to the `bookEvent` page based on the new endpoint `/events/:serviceId.
|
process
|
return service events and other details update events endpoint to events serviceid that to return all of the requested service details since from homepage which services page and by using findall query that will return all details for all sercives service details passing to the bookevent page based on the new endpoint events serviceid
| 1
|
127,275
| 5,027,831,714
|
IssuesEvent
|
2016-12-15 16:31:29
|
graphcool/console
|
https://api.github.com/repos/graphcool/console
|
closed
|
No reactive behaviour for other side of new relation within the same model
|
area/models/databrowser bug onhold priority/P1
|
When creating a relation between one node and another node within the same model, the source field updates but the destination relation field does not reactively update with the relationship. Likely due to only the current row being re-rendered instead of the entire table?
|
1.0
|
No reactive behaviour for other side of new relation within the same model - When creating a relation between one node and another node within the same model, the source field updates but the destination relation field does not reactively update with the relationship. Likely due to only the current row being re-rendered instead of the entire table?
|
non_process
|
no reactive behaviour for other side of new relation within the same model when creating a relation between one node and another node within the same model the source field updates but the destination relation field does not reactively update with the relationship likely due to only the current row being re rendered instead of the entire table
| 0
|
444,800
| 31,148,463,159
|
IssuesEvent
|
2023-08-16 08:18:43
|
risingwavelabs/risingwave-docs
|
https://api.github.com/repos/risingwavelabs/risingwave-docs
|
closed
|
Document `GROUPING()`
|
documentation
|
### Related code PR
https://github.com/risingwavelabs/risingwave/pull/11006
### Which part(s) of the docs might be affected or should be updated? And how?
Document `GROUPING()` under Aggregrate functions. Also note how `GROUPING()` is used with the `GROUPING SETS` clause (documented with this [issue).](https://github.com/risingwavelabs/risingwave-docs/issues/1043)
### Reference
Table 9-53: Grouping Operations: https://www.postgresql.org/docs/9.5/functions-aggregate.html
|
1.0
|
Document `GROUPING()` - ### Related code PR
https://github.com/risingwavelabs/risingwave/pull/11006
### Which part(s) of the docs might be affected or should be updated? And how?
Document `GROUPING()` under Aggregrate functions. Also note how `GROUPING()` is used with the `GROUPING SETS` clause (documented with this [issue).](https://github.com/risingwavelabs/risingwave-docs/issues/1043)
### Reference
Table 9-53: Grouping Operations: https://www.postgresql.org/docs/9.5/functions-aggregate.html
|
non_process
|
document grouping related code pr which part s of the docs might be affected or should be updated and how document grouping under aggregrate functions also note how grouping is used with the grouping sets clause documented with this reference table grouping operations
| 0
|
49,850
| 7,543,650,697
|
IssuesEvent
|
2018-04-17 16:03:29
|
palantir/blueprint
|
https://api.github.com/repos/palantir/blueprint
|
closed
|
Documentation is a dead link
|
Browser: Firefox Domain: documentation Type: bug
|
#### Steps to reproduce
1. Go to [the documentation page](http://blueprintjs.com/docs/v2/)
2. Watch it be blank
3. Observe this message in the console; `TypeError: Not enough arguments to CSSStyleSheet.insertRule.`
#### Actual behavior
A working documentation page
#### Expected behavior
A blank page with a console error
|
1.0
|
Documentation is a dead link -
#### Steps to reproduce
1. Go to [the documentation page](http://blueprintjs.com/docs/v2/)
2. Watch it be blank
3. Observe this message in the console; `TypeError: Not enough arguments to CSSStyleSheet.insertRule.`
#### Actual behavior
A working documentation page
#### Expected behavior
A blank page with a console error
|
non_process
|
documentation is a dead link steps to reproduce go to watch it be blank observe this message in the console typeerror not enough arguments to cssstylesheet insertrule actual behavior a working documentation page expected behavior a blank page with a console error
| 0
|
6,904
| 10,056,362,785
|
IssuesEvent
|
2019-07-22 08:59:03
|
bisq-network/bisq
|
https://api.github.com/repos/bisq-network/bisq
|
closed
|
Show popup at shutdown if offer has been taken, but no action had been taken yet.
|
in:gui in:trade-process was:dropped
|
Could be a helpful reminder, to finish the trade before shutting down to avoid disputes. Also, could have the same popup when starting up again and a trade is still unfinished.
|
1.0
|
Show popup at shutdown if offer has been taken, but no action had been taken yet. - Could be a helpful reminder, to finish the trade before shutting down to avoid disputes. Also, could have the same popup when starting up again and a trade is still unfinished.
|
process
|
show popup at shutdown if offer has been taken but no action had been taken yet could be a helpful reminder to finish the trade before shutting down to avoid disputes also could have the same popup when starting up again and a trade is still unfinished
| 1
|
2,585
| 5,344,734,325
|
IssuesEvent
|
2017-02-17 15:16:58
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
closed
|
[Subtitles] [FR] MÉLENCHON - POUR UNE SÉCURITÉ SOCIALE DU LOGEMENT
|
Language: French Process: Someone is working on this issue Process: [4] Ready for review (2)
|
# Video title
MÉLENCHON - POUR UNE SÉCURITÉ SOCIALE DU LOGEMENT
# URL
https://www.youtube.com/watch?v=Bf0vb6mJTjg&t=5s
# Youtube subtitles language
Français
# Duration
30:42
# Subtitles URL
https://www.youtube.com/timedtext_editor?forceedit=captions&tab=captions&bl=vmp&ref=player&action_mde_edit_form=1&lang=fr&v=Bf0vb6mJTjg
|
2.0
|
[Subtitles] [FR] MÉLENCHON - POUR UNE SÉCURITÉ SOCIALE DU LOGEMENT - # Video title
MÉLENCHON - POUR UNE SÉCURITÉ SOCIALE DU LOGEMENT
# URL
https://www.youtube.com/watch?v=Bf0vb6mJTjg&t=5s
# Youtube subtitles language
Français
# Duration
30:42
# Subtitles URL
https://www.youtube.com/timedtext_editor?forceedit=captions&tab=captions&bl=vmp&ref=player&action_mde_edit_form=1&lang=fr&v=Bf0vb6mJTjg
|
process
|
mélenchon pour une sécurité sociale du logement video title mélenchon pour une sécurité sociale du logement url youtube subtitles language français duration subtitles url
| 1
|
16,651
| 21,715,566,601
|
IssuesEvent
|
2022-05-10 17:32:31
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
node child_process cann't normally catch error!
|
child_process
|
### Version
16.13.0
### Platform
Darwin tongdembp 17.7.0 Darwin Kernel Version 17.7.0: Fri Oct 30 13:34:27 PDT 2020; root:xnu-4570.71.82.8~1/RELEASE_X86_64 x86_64
### Subsystem
node:child_process
### What steps will reproduce the bug?
1. Use 'npm init' to create a project.
2. Deliberately correct package.json, make it not conform to the specification of JSON.(eg: Add an , at the end of the file)
```json
{
"name": "test",
"version": "1.0.0",
"description": "",
"main": "index.js",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"author": "",
"license": "ISC",
"dependencies": {
"ali-oss": "^6.17.1",
"axios": "^0.24.0",
"cos-nodejs-sdk-v5": "^2.11.8",
"path": "^0.12.7",
"qs": "^6.10.3"
},
"devDependencies": {
"prettier": "^2.6.0"
// see here, one more ","
},
}
```
4. Call b.js through a.js by require & child_process.Here is the code.
```js
// a.js
const { spawn } = require('node:child_process')
const test = () => {
return new Promise((resolve, reject) => {
const code = `require('./test')()`
const subprocess = spawn('node', ['-e', code])
subprocess.on('error', (err) => {
console.error('subprocess error')
reject('error')
})
subprocess.on('exit', (err) => {
console.error('subprocess error')
resolve('exit')
})
})
}
;(async () => {
try {
const res = await test()
console.log(`🚀 ~ res`, res)
} catch (e) {
console.log('e', e)
}
})()
```
```js
// b.js
;() => {
console.log('1111')
}
```
5. run, `node a.js` and the error thrown was not caught
```
node:internal/modules/cjs/loader:317
throw e;
^
SyntaxError: Error parsing /Users/tong/Desktop/待学习=v=资料/node-test/package.json: Unexpected token } in JSON at position 406
at parse (<anonymous>)
at readPackage (node:internal/modules/cjs/loader:304:20)
at readPackageScope (node:internal/modules/cjs/loader:329:19)
at shouldUseESMLoader (node:internal/modules/run_main:39:15)
at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:76:24)
at node:internal/main/run_main_module:17:47 {
path: '/Users/xxx/node-test/package.json'
}
```
### How often does it reproduce? Is there a required condition?
always
### What is the expected behavior?
Errors should be caught
### What do you see instead?
the error thrown was not caught
```
node:internal/modules/cjs/loader:317
throw e;
^
SyntaxError: Error parsing /Users/tong/Desktop/待学习=v=资料/node-test/package.json: Unexpected token } in JSON at position 406
at parse (<anonymous>)
at readPackage (node:internal/modules/cjs/loader:304:20)
at readPackageScope (node:internal/modules/cjs/loader:329:19)
at shouldUseESMLoader (node:internal/modules/run_main:39:15)
at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:76:24)
at node:internal/main/run_main_module:17:47 {
path: '/Users/xxx/node-test/package.json'
}
```
### Additional information
_No response_
|
1.0
|
node child_process cann't normally catch error! - ### Version
16.13.0
### Platform
Darwin tongdembp 17.7.0 Darwin Kernel Version 17.7.0: Fri Oct 30 13:34:27 PDT 2020; root:xnu-4570.71.82.8~1/RELEASE_X86_64 x86_64
### Subsystem
node:child_process
### What steps will reproduce the bug?
1. Use 'npm init' to create a project.
2. Deliberately correct package.json, make it not conform to the specification of JSON.(eg: Add an , at the end of the file)
```json
{
"name": "test",
"version": "1.0.0",
"description": "",
"main": "index.js",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"author": "",
"license": "ISC",
"dependencies": {
"ali-oss": "^6.17.1",
"axios": "^0.24.0",
"cos-nodejs-sdk-v5": "^2.11.8",
"path": "^0.12.7",
"qs": "^6.10.3"
},
"devDependencies": {
"prettier": "^2.6.0"
// see here, one more ","
},
}
```
4. Call b.js through a.js by require & child_process.Here is the code.
```js
// a.js
const { spawn } = require('node:child_process')
const test = () => {
return new Promise((resolve, reject) => {
const code = `require('./test')()`
const subprocess = spawn('node', ['-e', code])
subprocess.on('error', (err) => {
console.error('subprocess error')
reject('error')
})
subprocess.on('exit', (err) => {
console.error('subprocess error')
resolve('exit')
})
})
}
;(async () => {
try {
const res = await test()
console.log(`🚀 ~ res`, res)
} catch (e) {
console.log('e', e)
}
})()
```
```js
// b.js
;() => {
console.log('1111')
}
```
5. run, `node a.js` and the error thrown was not caught
```
node:internal/modules/cjs/loader:317
throw e;
^
SyntaxError: Error parsing /Users/tong/Desktop/待学习=v=资料/node-test/package.json: Unexpected token } in JSON at position 406
at parse (<anonymous>)
at readPackage (node:internal/modules/cjs/loader:304:20)
at readPackageScope (node:internal/modules/cjs/loader:329:19)
at shouldUseESMLoader (node:internal/modules/run_main:39:15)
at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:76:24)
at node:internal/main/run_main_module:17:47 {
path: '/Users/xxx/node-test/package.json'
}
```
### How often does it reproduce? Is there a required condition?
always
### What is the expected behavior?
Errors should be caught
### What do you see instead?
the error thrown was not caught
```
node:internal/modules/cjs/loader:317
throw e;
^
SyntaxError: Error parsing /Users/tong/Desktop/待学习=v=资料/node-test/package.json: Unexpected token } in JSON at position 406
at parse (<anonymous>)
at readPackage (node:internal/modules/cjs/loader:304:20)
at readPackageScope (node:internal/modules/cjs/loader:329:19)
at shouldUseESMLoader (node:internal/modules/run_main:39:15)
at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:76:24)
at node:internal/main/run_main_module:17:47 {
path: '/Users/xxx/node-test/package.json'
}
```
### Additional information
_No response_
|
process
|
node child process cann t normally catch error! version platform darwin tongdembp darwin kernel version fri oct pdt root xnu release subsystem node child process what steps will reproduce the bug use npm init to create a project deliberately correct package json make it not conform to the specification of json eg add an at the end of the file json name test version description main index js scripts test echo error no test specified exit author license isc dependencies ali oss axios cos nodejs sdk path qs devdependencies prettier see here one more call b js through a js by require child process here is the code js a js const spawn require node child process const test return new promise resolve reject const code require test const subprocess spawn node subprocess on error err console error subprocess error reject error subprocess on exit err console error subprocess error resolve exit async try const res await test console log 🚀 res res catch e console log e e js b js console log run node a js and the error thrown was not caught node internal modules cjs loader throw e syntaxerror error parsing users tong desktop 待学习 v 资料 node test package json unexpected token in json at position at parse at readpackage node internal modules cjs loader at readpackagescope node internal modules cjs loader at shoulduseesmloader node internal modules run main at function executeuserentrypoint node internal modules run main at node internal main run main module path users xxx node test package json how often does it reproduce is there a required condition always what is the expected behavior errors should be caught what do you see instead the error thrown was not caught node internal modules cjs loader throw e syntaxerror error parsing users tong desktop 待学习 v 资料 node test package json unexpected token in json at position at parse at readpackage node internal modules cjs loader at readpackagescope node internal modules cjs loader at shoulduseesmloader node internal modules run main at function executeuserentrypoint node internal modules run main at node internal main run main module path users xxx node test package json additional information no response
| 1
|
6,473
| 9,547,881,511
|
IssuesEvent
|
2019-05-02 01:49:38
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
HandleCountChanges test fails frequently with AssertException
|
area-System.Diagnostics.Process disabled-test test-run-core
|
Disabling the test. We should investigate if we make the test more robust without the need of a retry mechanism.
```
System.Diagnostics.Tests.ProcessTests.HandleCountChanges
Child exception:
Xunit.Sdk.TrueException: Assert.True() Failure
Expected: True
Actual: False
at System.Diagnostics.Tests.ProcessTests.<HandleCountChanges>b__103_0() in D:\a\1\s\src\System.Diagnostics.Process\tests\ProcessTests.cs:line 1495
Child process:
System.Diagnostics.Process.Tests, Version=4.2.1.0, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51 System.Diagnostics.Tests.ProcessTests Int32 <HandleCountChanges>b__103_0()
```
https://dev.azure.com/dnceng/public/_build/results?buildId=174161&view=ms.vss-test-web.build-test-results-tab
cc @stephentoub
|
1.0
|
HandleCountChanges test fails frequently with AssertException - Disabling the test. We should investigate if we make the test more robust without the need of a retry mechanism.
```
System.Diagnostics.Tests.ProcessTests.HandleCountChanges
Child exception:
Xunit.Sdk.TrueException: Assert.True() Failure
Expected: True
Actual: False
at System.Diagnostics.Tests.ProcessTests.<HandleCountChanges>b__103_0() in D:\a\1\s\src\System.Diagnostics.Process\tests\ProcessTests.cs:line 1495
Child process:
System.Diagnostics.Process.Tests, Version=4.2.1.0, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51 System.Diagnostics.Tests.ProcessTests Int32 <HandleCountChanges>b__103_0()
```
https://dev.azure.com/dnceng/public/_build/results?buildId=174161&view=ms.vss-test-web.build-test-results-tab
cc @stephentoub
|
process
|
handlecountchanges test fails frequently with assertexception disabling the test we should investigate if we make the test more robust without the need of a retry mechanism system diagnostics tests processtests handlecountchanges child exception xunit sdk trueexception assert true failure expected true actual false at system diagnostics tests processtests b in d a s src system diagnostics process tests processtests cs line child process system diagnostics process tests version culture neutral publickeytoken system diagnostics tests processtests b cc stephentoub
| 1
|
8,660
| 11,798,046,431
|
IssuesEvent
|
2020-03-18 13:47:19
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[feature][processing] Add dedicated parameter type for database schemas
|
3.12 Automatic new feature Processing User Manual
|
Original commit: https://github.com/qgis/QGIS/commit/93c917f7b6e93bae26a16078ace44879ff3930d8 by nyalldawson
Allows selection from the existing database schema for a specific
database provider connection (the provider must implement the connections API)
|
1.0
|
[feature][processing] Add dedicated parameter type for database schemas - Original commit: https://github.com/qgis/QGIS/commit/93c917f7b6e93bae26a16078ace44879ff3930d8 by nyalldawson
Allows selection from the existing database schema for a specific
database provider connection (the provider must implement the connections API)
|
process
|
add dedicated parameter type for database schemas original commit by nyalldawson allows selection from the existing database schema for a specific database provider connection the provider must implement the connections api
| 1
|
200,910
| 15,801,844,697
|
IssuesEvent
|
2021-04-03 06:49:27
|
tsh22/ped
|
https://api.github.com/repos/tsh22/ped
|
opened
|
Markdown not consistent or shown for certain sections of the UG
|
severity.VeryLow type.DocumentationBug
|
Just some minor markdown inconsistencies in the UG which might make it a little harder to read for users, otherwise instructions are ok!
<!--session: 1617429967547-821d5e25-def9-4f4f-b7c7-a341c7e0f8e7-->
|
1.0
|
Markdown not consistent or shown for certain sections of the UG - Just some minor markdown inconsistencies in the UG which might make it a little harder to read for users, otherwise instructions are ok!
<!--session: 1617429967547-821d5e25-def9-4f4f-b7c7-a341c7e0f8e7-->
|
non_process
|
markdown not consistent or shown for certain sections of the ug just some minor markdown inconsistencies in the ug which might make it a little harder to read for users otherwise instructions are ok
| 0
|
10,811
| 13,609,289,132
|
IssuesEvent
|
2020-09-23 04:50:21
|
googleapis/java-analytics-admin
|
https://api.github.com/repos/googleapis/java-analytics-admin
|
closed
|
Dependency Dashboard
|
api: analyticsadmin type: process
|
This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/org.codehaus.mojo-build-helper-maven-plugin-3.x -->build(deps): update dependency org.codehaus.mojo:build-helper-maven-plugin to v3.2.0
- [ ] <!-- rebase-branch=renovate/com.google.cloud.samples-shared-configuration-1.x -->chore(deps): update dependency com.google.cloud.samples:shared-configuration to v1.0.18
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/org.codehaus.mojo-build-helper-maven-plugin-3.x -->build(deps): update dependency org.codehaus.mojo:build-helper-maven-plugin to v3.2.0
- [ ] <!-- rebase-branch=renovate/com.google.cloud.samples-shared-configuration-1.x -->chore(deps): update dependency com.google.cloud.samples:shared-configuration to v1.0.18
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any build deps update dependency org codehaus mojo build helper maven plugin to chore deps update dependency com google cloud samples shared configuration to check this box to trigger a request for renovate to run again on this repository
| 1
|
10,128
| 13,044,162,336
|
IssuesEvent
|
2020-07-29 03:47:31
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `DateDiff` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `DateDiff` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @mapleFU
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `DateDiff` from TiDB -
## Description
Port the scalar function `DateDiff` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @mapleFU
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function datediff from tidb description port the scalar function datediff from tidb to coprocessor score mentor s maplefu recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
10,712
| 13,507,798,474
|
IssuesEvent
|
2020-09-14 06:40:07
|
pingcap/tidb
|
https://api.github.com/repos/pingcap/tidb
|
closed
|
reflect Copr Cache in explain analyze results
|
component/coprocessor epic/copr-cache type/enhancement
|
## Development Task
The execution info in explain analyze results is cached if Copr Cache is hit. We should find a way to show the cached execution info correctly.
Also it should be fixed in slow query log.
|
1.0
|
reflect Copr Cache in explain analyze results - ## Development Task
The execution info in explain analyze results is cached if Copr Cache is hit. We should find a way to show the cached execution info correctly.
Also it should be fixed in slow query log.
|
process
|
reflect copr cache in explain analyze results development task the execution info in explain analyze results is cached if copr cache is hit we should find a way to show the cached execution info correctly also it should be fixed in slow query log
| 1
|
10,031
| 13,044,161,488
|
IssuesEvent
|
2020-07-29 03:47:23
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `Char` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `Char` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @lonng
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `Char` from TiDB -
## Description
Port the scalar function `Char` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @lonng
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function char from tidb description port the scalar function char from tidb to coprocessor score mentor s lonng recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
465,503
| 13,387,065,730
|
IssuesEvent
|
2020-09-02 15:31:04
|
wso2/product-apim
|
https://api.github.com/repos/wso2/product-apim
|
opened
|
[3.0.0] UI error when a user with Internal/creator role tries to deploy the sample API
|
Priority/Normal Type/Bug
|
### Description:
When a user with Internal/creator tries to deploy the sample API, an error will be logged in the browser console.
**Tip:** This is happening because the sample API will try to get it published and it gets failed as the creator user does not have the permission to publish. This has been already fixed in APIM 3.2.0. Backporting those code segments would solve this issue.
### Steps to reproduce:
1. Create a user with the Internal/creator role.
2. Login to Publisher portal as that user.
3. Open up the browser console to check the console output (UI).
4. Click on "DEPLOY SAMPLE API" button.
5. The following UI error can be seen on the console as shown below.

6. The UI will not be updated, so the API will not be shown in the UI until you refresh the page.
### Affected Product Version:
APIM 3.0.0
### Environment details (with versions):
- OS: Ubuntu 20.04 LTS
|
1.0
|
[3.0.0] UI error when a user with Internal/creator role tries to deploy the sample API - ### Description:
When a user with Internal/creator tries to deploy the sample API, an error will be logged in the browser console.
**Tip:** This is happening because the sample API will try to get it published and it gets failed as the creator user does not have the permission to publish. This has been already fixed in APIM 3.2.0. Backporting those code segments would solve this issue.
### Steps to reproduce:
1. Create a user with the Internal/creator role.
2. Login to Publisher portal as that user.
3. Open up the browser console to check the console output (UI).
4. Click on "DEPLOY SAMPLE API" button.
5. The following UI error can be seen on the console as shown below.

6. The UI will not be updated, so the API will not be shown in the UI until you refresh the page.
### Affected Product Version:
APIM 3.0.0
### Environment details (with versions):
- OS: Ubuntu 20.04 LTS
|
non_process
|
ui error when a user with internal creator role tries to deploy the sample api description when a user with internal creator tries to deploy the sample api an error will be logged in the browser console tip this is happening because the sample api will try to get it published and it gets failed as the creator user does not have the permission to publish this has been already fixed in apim backporting those code segments would solve this issue steps to reproduce create a user with the internal creator role login to publisher portal as that user open up the browser console to check the console output ui click on deploy sample api button the following ui error can be seen on the console as shown below the ui will not be updated so the api will not be shown in the ui until you refresh the page affected product version apim environment details with versions os ubuntu lts
| 0
|
443,369
| 12,793,313,547
|
IssuesEvent
|
2020-07-02 03:38:28
|
Trustroots/trustroots
|
https://api.github.com/repos/Trustroots/trustroots
|
reopened
|
Update Agenda, background job worker from v1 to v2
|
Hacktoberfest [Priority] Normal framework server unloved
|
[Agenda](https://github.com/agenda/agenda) v2 was a complete rewrite to support promises instead of callbacks.
Agenda is run in the background worker to process jobs. Main app and the worker run separately in parallel with `npm start`.
v2 changelog: https://github.com/agenda/agenda/blob/master/History.md#200--2018-07-19
Rewrite needed mainly at:
- Worker entry script: https://github.com/Trustroots/trustroots/blob/15a8bcebee8286eca1591f950a120420b2e08dd3/config/lib/worker.js
- At bunch of `modules/*/server/jobs/*.server.job.js` -files
I'm not sure if we should wait for ES5 → ES6 migration before updating Agenda, because we could jump directly using `async/await` -format (that we don't support at the moment).
|
1.0
|
Update Agenda, background job worker from v1 to v2 - [Agenda](https://github.com/agenda/agenda) v2 was a complete rewrite to support promises instead of callbacks.
Agenda is run in the background worker to process jobs. Main app and the worker run separately in parallel with `npm start`.
v2 changelog: https://github.com/agenda/agenda/blob/master/History.md#200--2018-07-19
Rewrite needed mainly at:
- Worker entry script: https://github.com/Trustroots/trustroots/blob/15a8bcebee8286eca1591f950a120420b2e08dd3/config/lib/worker.js
- At bunch of `modules/*/server/jobs/*.server.job.js` -files
I'm not sure if we should wait for ES5 → ES6 migration before updating Agenda, because we could jump directly using `async/await` -format (that we don't support at the moment).
|
non_process
|
update agenda background job worker from to was a complete rewrite to support promises instead of callbacks agenda is run in the background worker to process jobs main app and the worker run separately in parallel with npm start changelog rewrite needed mainly at worker entry script at bunch of modules server jobs server job js files i m not sure if we should wait for → migration before updating agenda because we could jump directly using async await format that we don t support at the moment
| 0
|
16,332
| 20,988,702,912
|
IssuesEvent
|
2022-03-29 07:15:47
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Merge GO:0070359 actin polymerization-dependent cell motility involved in migration of symbiont in host and migration of symbiont within host by polymerization of host actin
|
multi-species process
|
- [x] GO:0070359 actin polymerization-dependent cell motility involved in migration of symbiont in host
Merge into GO:0070360 migration of symbiont within host by polymerization of host actin: 0 annotations
* GO:0070360 is a part_of child of GO:0070359; however this is inconsistent with the definitions, ie GO:0070360 mentions host actin, while GO:0070359.
* The data is papers cited and annotated do not provide in vivo evidence for the mechanisms of cell motiility
* The Multiorg group proposes to merge the two terms and change the label to 'actin **polymerization**-dependent migration in host'. We can create new terms later when more is known about this process.
* Def: update to remove "host" The directional movement of an organism, usually a bacterial cell, from one place to another within its host organism, by a process involving continuous polymerization of ~host~ actin at one pole of the symbiont cell. Some bacteria use host actin for migration. The host is defined as the larger of the organisms involved in a symbiotic interaction.
|
1.0
|
Merge GO:0070359 actin polymerization-dependent cell motility involved in migration of symbiont in host and migration of symbiont within host by polymerization of host actin - - [x] GO:0070359 actin polymerization-dependent cell motility involved in migration of symbiont in host
Merge into GO:0070360 migration of symbiont within host by polymerization of host actin: 0 annotations
* GO:0070360 is a part_of child of GO:0070359; however this is inconsistent with the definitions, ie GO:0070360 mentions host actin, while GO:0070359.
* The data is papers cited and annotated do not provide in vivo evidence for the mechanisms of cell motiility
* The Multiorg group proposes to merge the two terms and change the label to 'actin **polymerization**-dependent migration in host'. We can create new terms later when more is known about this process.
* Def: update to remove "host" The directional movement of an organism, usually a bacterial cell, from one place to another within its host organism, by a process involving continuous polymerization of ~host~ actin at one pole of the symbiont cell. Some bacteria use host actin for migration. The host is defined as the larger of the organisms involved in a symbiotic interaction.
|
process
|
merge go actin polymerization dependent cell motility involved in migration of symbiont in host and migration of symbiont within host by polymerization of host actin go actin polymerization dependent cell motility involved in migration of symbiont in host merge into go migration of symbiont within host by polymerization of host actin annotations go is a part of child of go however this is inconsistent with the definitions ie go mentions host actin while go the data is papers cited and annotated do not provide in vivo evidence for the mechanisms of cell motiility the multiorg group proposes to merge the two terms and change the label to actin polymerization dependent migration in host we can create new terms later when more is known about this process def update to remove host the directional movement of an organism usually a bacterial cell from one place to another within its host organism by a process involving continuous polymerization of host actin at one pole of the symbiont cell some bacteria use host actin for migration the host is defined as the larger of the organisms involved in a symbiotic interaction
| 1
|
13,776
| 3,356,863,517
|
IssuesEvent
|
2015-11-18 22:18:46
|
sellry/cardfool-issues
|
https://api.github.com/repos/sellry/cardfool-issues
|
closed
|
My Account Screen: Wording Change
|
enhancement Fixed and ready to test
|
Change the lower text to be:
"For more extensive account management (like removing credit cards), please visit www.cardfool.com"
|
1.0
|
My Account Screen: Wording Change - Change the lower text to be:
"For more extensive account management (like removing credit cards), please visit www.cardfool.com"
|
non_process
|
my account screen wording change change the lower text to be for more extensive account management like removing credit cards please visit
| 0
|
28,483
| 2,703,176,099
|
IssuesEvent
|
2015-04-06 15:46:28
|
NickolausDS/Unity-Free-Flight
|
https://api.github.com/repos/NickolausDS/Unity-Free-Flight
|
closed
|
Add flight stabilization mechanic for regular gliding
|
enhancement Low Priority Task
|
Current mechanics force the player to stabilize flight if they go into a dive, or slow their speed to stall.
It might be preferable to have a mechanism that will check angle of attack and speed, then apply the input for the player when they leave their hands off the controls.
Example: The player gains altitude, then flies level at near zero velocity. The mechanic would pitch them down before they reached stall speed.
Example2: The player dives off a cliff. The mechanic pitches them down while they gain velocity, then pitches up to level once the wings generate enough lift.
|
1.0
|
Add flight stabilization mechanic for regular gliding - Current mechanics force the player to stabilize flight if they go into a dive, or slow their speed to stall.
It might be preferable to have a mechanism that will check angle of attack and speed, then apply the input for the player when they leave their hands off the controls.
Example: The player gains altitude, then flies level at near zero velocity. The mechanic would pitch them down before they reached stall speed.
Example2: The player dives off a cliff. The mechanic pitches them down while they gain velocity, then pitches up to level once the wings generate enough lift.
|
non_process
|
add flight stabilization mechanic for regular gliding current mechanics force the player to stabilize flight if they go into a dive or slow their speed to stall it might be preferable to have a mechanism that will check angle of attack and speed then apply the input for the player when they leave their hands off the controls example the player gains altitude then flies level at near zero velocity the mechanic would pitch them down before they reached stall speed the player dives off a cliff the mechanic pitches them down while they gain velocity then pitches up to level once the wings generate enough lift
| 0
|
11,798
| 14,624,028,177
|
IssuesEvent
|
2020-12-23 05:13:27
|
rohanchandra30/Spectral-Trajectory-and-Behavior-Prediction
|
https://api.github.com/repos/rohanchandra30/Spectral-Trajectory-and-Behavior-Prediction
|
closed
|
Missing TrainSet file error during data preparation
|
Data Processing
|
Hello,
I am trying to run the models, but I am unable to prepare the formatted data. I have downloaded the resources during the installation, and have formatted the Apolloscape data as well. But while running the data_processing/data_stream.py, I am getting the following error:

I am unsure what is causing this, as it looks the program was expecting these training files to exist. Any help would be appreciated.
Thank you!
|
1.0
|
Missing TrainSet file error during data preparation - Hello,
I am trying to run the models, but I am unable to prepare the formatted data. I have downloaded the resources during the installation, and have formatted the Apolloscape data as well. But while running the data_processing/data_stream.py, I am getting the following error:

I am unsure what is causing this, as it looks the program was expecting these training files to exist. Any help would be appreciated.
Thank you!
|
process
|
missing trainset file error during data preparation hello i am trying to run the models but i am unable to prepare the formatted data i have downloaded the resources during the installation and have formatted the apolloscape data as well but while running the data processing data stream py i am getting the following error i am unsure what is causing this as it looks the program was expecting these training files to exist any help would be appreciated thank you
| 1
|
14,528
| 17,630,654,831
|
IssuesEvent
|
2021-08-19 07:31:13
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add "Fast Animals, Slow Children"
|
suggested title in process
|
Please add as much of the following info as you can:
Title: Fast Animals, Slow Children
Type (film/tv show): TV Show
Film or show in which it appears: Family Guy ( Season 1, Episode 2 - "I Never Met the Dead Man")
Is the parent film/show streaming anywhere? Yes. Netflix and Disney+
About when in the parent film/show does it appear? 04:31
Actual footage of the film/show can be seen (yes/no)? Yes. https://www.youtube.com/watch?v=abFllTlLqVI
|
1.0
|
Add "Fast Animals, Slow Children" - Please add as much of the following info as you can:
Title: Fast Animals, Slow Children
Type (film/tv show): TV Show
Film or show in which it appears: Family Guy ( Season 1, Episode 2 - "I Never Met the Dead Man")
Is the parent film/show streaming anywhere? Yes. Netflix and Disney+
About when in the parent film/show does it appear? 04:31
Actual footage of the film/show can be seen (yes/no)? Yes. https://www.youtube.com/watch?v=abFllTlLqVI
|
process
|
add fast animals slow children please add as much of the following info as you can title fast animals slow children type film tv show tv show film or show in which it appears family guy season episode i never met the dead man is the parent film show streaming anywhere yes netflix and disney about when in the parent film show does it appear actual footage of the film show can be seen yes no yes
| 1
|
3,550
| 6,587,491,168
|
IssuesEvent
|
2017-09-13 21:17:36
|
cliffparnitzky/ApparelManager
|
https://api.github.com/repos/cliffparnitzky/ApparelManager
|
closed
|
Add articlelink to order item
|
Comp - order item ⚙ - Processed
|
To get more info about an article, add an aarticlelink to oder items (with popup).
|
1.0
|
Add articlelink to order item - To get more info about an article, add an aarticlelink to oder items (with popup).
|
process
|
add articlelink to order item to get more info about an article add an aarticlelink to oder items with popup
| 1
|
379,042
| 11,212,418,089
|
IssuesEvent
|
2020-01-06 17:32:06
|
eclipse-researchlabs/scava
|
https://api.github.com/repos/eclipse-researchlabs/scava
|
opened
|
Use the same UI for both recommenders
|
User Interface enhancement low priority
|
It would be great if we could use the same View for presenting the recommendation results.
Actually, the Code Recommendation view is pretty cool because:
- you retrieve the code which was selected for asking a recommendation
- you keep the history of the recommendations

So we could list both kind of recommendations in the same view and,
- if we select a code recommendation, see on the right subview the snipped code,
- if we select a Q&A recommendation, see on the right subview the links to the StackOverFlow posts
My 2 cents
|
1.0
|
Use the same UI for both recommenders - It would be great if we could use the same View for presenting the recommendation results.
Actually, the Code Recommendation view is pretty cool because:
- you retrieve the code which was selected for asking a recommendation
- you keep the history of the recommendations

So we could list both kind of recommendations in the same view and,
- if we select a code recommendation, see on the right subview the snipped code,
- if we select a Q&A recommendation, see on the right subview the links to the StackOverFlow posts
My 2 cents
|
non_process
|
use the same ui for both recommenders it would be great if we could use the same view for presenting the recommendation results actually the code recommendation view is pretty cool because you retrieve the code which was selected for asking a recommendation you keep the history of the recommendations so we could list both kind of recommendations in the same view and if we select a code recommendation see on the right subview the snipped code if we select a q a recommendation see on the right subview the links to the stackoverflow posts my cents
| 0
|
817,417
| 30,640,280,926
|
IssuesEvent
|
2023-07-24 21:16:59
|
ethereum/ethereum-org-website
|
https://api.github.com/repos/ethereum/ethereum-org-website
|
closed
|
Automatically pull latest upgrades blog posts
|
feature :sparkles: medium priority Status: Up Next Status: Stale
|
**Is your feature request related to a problem? Please describe.**
Eth2 blog posts by Danny Ryan and Ben Edgington are currently listed at the bottom of https://ethereum.org/en/eth2. These are essentially hard-coded into our repo, and do not update when new blog posts by Danny or Ben are released.
**Describe the solution you'd like**
Research options for pulling latest Eth2 related blog posts in automatically, and reduce need for hard-coding this information.
**Describe alternatives you've considered**
- API requests (haven't looked into potential endpoints)
- Perhaps a hook/bot that puts up a PR to add any new articles to the list when they're posted
- Continue to manually keep an eye on these and update manually 👎🏻
**Additional context**
Issue #3627 brought this to my attention, as several months passed before this section way updated.
Perhaps solutions here could be generalized to alternative blog posts beyond Eth2 as well, for other parts of the site.
**Want to contribute?**
We love contributions from the Ethereum community! Please comment on an issue if you're interested in helping out with a PR.
|
1.0
|
Automatically pull latest upgrades blog posts - **Is your feature request related to a problem? Please describe.**
Eth2 blog posts by Danny Ryan and Ben Edgington are currently listed at the bottom of https://ethereum.org/en/eth2. These are essentially hard-coded into our repo, and do not update when new blog posts by Danny or Ben are released.
**Describe the solution you'd like**
Research options for pulling latest Eth2 related blog posts in automatically, and reduce need for hard-coding this information.
**Describe alternatives you've considered**
- API requests (haven't looked into potential endpoints)
- Perhaps a hook/bot that puts up a PR to add any new articles to the list when they're posted
- Continue to manually keep an eye on these and update manually 👎🏻
**Additional context**
Issue #3627 brought this to my attention, as several months passed before this section way updated.
Perhaps solutions here could be generalized to alternative blog posts beyond Eth2 as well, for other parts of the site.
**Want to contribute?**
We love contributions from the Ethereum community! Please comment on an issue if you're interested in helping out with a PR.
|
non_process
|
automatically pull latest upgrades blog posts is your feature request related to a problem please describe blog posts by danny ryan and ben edgington are currently listed at the bottom of these are essentially hard coded into our repo and do not update when new blog posts by danny or ben are released describe the solution you d like research options for pulling latest related blog posts in automatically and reduce need for hard coding this information describe alternatives you ve considered api requests haven t looked into potential endpoints perhaps a hook bot that puts up a pr to add any new articles to the list when they re posted continue to manually keep an eye on these and update manually 👎🏻 additional context issue brought this to my attention as several months passed before this section way updated perhaps solutions here could be generalized to alternative blog posts beyond as well for other parts of the site want to contribute we love contributions from the ethereum community please comment on an issue if you re interested in helping out with a pr
| 0
|
81,209
| 23,414,877,784
|
IssuesEvent
|
2022-08-12 22:34:28
|
spack/spack
|
https://api.github.com/repos/spack/spack
|
opened
|
Installation issue: llvm
|
build-error
|
### Steps to reproduce the issue
```console
$ spack spec -I llvm@14.0.6
Input spec
--------------------------------
llvm@14.0.6
Concretized
--------------------------------
cauws4d llvm@14.0.6%gcc@12.1.0+clang+compiler-rt~cuda~flang+gold+internal_unwind~ipo+libcxx~link_llvm_dylib+lld+lldb+llvm_dylib~mlir+omp_as_runtime~omp_debug~omp_tsan+polly~python~split_dwarf~z3 build_type=Release patches=6379168,d85ef51,f920173 shlib_symbol_version=none targets=none version_suffix=none arch=linux-rhel7-sandybridge
bz56uqf ^binutils@2.38%gcc@12.1.0~gas+gold~headers~interwork+ld~libiberty~lto+nls+plugins libs=shared,static arch=linux-rhel7-sandybridge
qanjaa2 ^diffutils@3.8%gcc@12.1.0 arch=linux-rhel7-sandybridge
aqkjcnx ^libiconv@1.16%gcc@12.1.0 libs=shared,static arch=linux-rhel7-sandybridge
xoc2mab ^gettext@0.21%gcc@12.1.0+bzip2+curses+git~libunistring+libxml2+tar+xz arch=linux-rhel7-sandybridge
v6vqvkk ^bzip2@1.0.8%gcc@12.1.0~debug~pic+shared arch=linux-rhel7-sandybridge
b324xkn ^libxml2@2.9.13%gcc@12.1.0~python arch=linux-rhel7-sandybridge
74jzup7 ^pkgconf@1.8.0%gcc@12.1.0 arch=linux-rhel7-sandybridge
mh24b7l ^xz@5.2.5%gcc@12.1.0~pic libs=shared,static arch=linux-rhel7-sandybridge
v7akdfg ^zlib@1.2.12%gcc@12.1.0+optimize+pic+shared patches=0d38234 arch=linux-rhel7-sandybridge
gnzo6m5 ^ncurses@6.2%gcc@12.1.0~symlinks+termlib abi=none arch=linux-rhel7-sandybridge
wvyqet2 ^tar@1.34%gcc@12.1.0 zip=pigz arch=linux-rhel7-sandybridge
hudp2d7 ^pigz@2.7%gcc@12.1.0 arch=linux-rhel7-sandybridge
wh37sg7 ^zstd@1.5.2%gcc@12.1.0+programs compression=none libs=shared,static arch=linux-rhel7-sandybridge
aewh5et ^cmake@3.23.3%gcc@12.1.0~doc+ncurses~ownlibs~qt build_type=Release arch=linux-rhel7-sandybridge
gizsjhb ^curl@7.84.0%gcc@12.1.0~gssapi~ldap~libidn2~librtmp~libssh~libssh2~nghttp2 libs=shared,static patches=9b011c9 tls=gnutls arch=linux-rhel7-sandybridge
s4lvb3x ^gnutls@3.6.15%gcc@12.1.0~guile+zlib arch=linux-rhel7-sandybridge
ztttlpd ^libidn2@2.3.0%gcc@12.1.0 arch=linux-rhel7-sandybridge
gxsvuva ^libunistring@0.9.10%gcc@12.1.0 arch=linux-rhel7-sandybridge
brncabs ^nettle@3.4.1%gcc@12.1.0 arch=linux-rhel7-sandybridge
b7e3ysy ^gmp@6.2.1%gcc@12.1.0 libs=shared,static arch=linux-rhel7-sandybridge
lak2y32 ^autoconf@2.69%gcc@12.1.0 patches=7793209 arch=linux-rhel7-sandybridge
yc6hkxy ^automake@1.16.5%gcc@12.1.0 arch=linux-rhel7-sandybridge
g2sk6lg ^perl@5.34.1%gcc@12.1.0+cpanm+shared+threads arch=linux-rhel7-sandybridge
oqbulwu ^berkeley-db@18.1.40%gcc@12.1.0+cxx~docs+stl patches=b231fcc arch=linux-rhel7-sandybridge
fze2c5x ^gdbm@1.19%gcc@12.1.0 arch=linux-rhel7-sandybridge
op7l5dw ^readline@8.1.2%gcc@12.1.0 arch=linux-rhel7-sandybridge
iei7lwt ^libtool@2.4.7%gcc@12.1.0 arch=linux-rhel7-sandybridge
g3ummln ^m4@1.4.19%gcc@12.1.0+sigsegv patches=9dc5fbd,bfdffa7 arch=linux-rhel7-sandybridge
hsgdcj2 ^libsigsegv@2.13%gcc@12.1.0 arch=linux-rhel7-sandybridge
c4t27bx ^expat@2.4.8%gcc@12.1.0+libbsd arch=linux-rhel7-sandybridge
sn5hu6p ^libbsd@0.11.5%gcc@12.1.0 arch=linux-rhel7-sandybridge
a45ocsx ^libmd@1.0.4%gcc@12.1.0 arch=linux-rhel7-sandybridge
htepgqg ^libarchive@3.5.2%gcc@12.1.0+iconv compression=bz2lib,lz4,lzma,lzo2,zlib,zstd crypto=mbedtls libs=shared,static programs=none xar=expat arch=linux-rhel7-sandybridge
b6lquau ^lz4@1.9.3%gcc@12.1.0 libs=shared,static arch=linux-rhel7-sandybridge
6yddcsh ^lzo@2.10%gcc@12.1.0 libs=shared,static arch=linux-rhel7-sandybridge
2q2mbkr ^mbedtls@2.28.0%gcc@12.1.0+pic build_type=Release libs=static arch=linux-rhel7-sandybridge
pa5eqnr ^libuv@1.44.1%gcc@12.1.0 arch=linux-rhel7-sandybridge
nz5gxtq ^rhash@1.4.2%gcc@12.1.0 patches=093518c,3fbfe46 arch=linux-rhel7-sandybridge
shugqek ^hwloc@2.8.0%gcc@12.1.0~cairo~cuda~gl~libudev+libxml2~netloc~nvml~oneapi-level-zero~opencl+pci~rocm+shared arch=linux-rhel7-sandybridge
pubuiwd ^libpciaccess@0.16%gcc@12.1.0 arch=linux-rhel7-sandybridge
iz3q3z6 ^util-macros@1.19.3%gcc@12.1.0 arch=linux-rhel7-sandybridge
cfy3kmq ^libedit@3.1-20210216%gcc@12.1.0 arch=linux-rhel7-sandybridge
rb2q6ky ^ninja@1.11.0%gcc@12.1.0 arch=linux-rhel7-sandybridge
lncwse2 ^python@3.9.13%gcc@12.1.0+bz2+ctypes+dbm~debug+ensurepip+libxml2+lzma~nis~optimizations+pic+pyexpat+pythoncmd+readline+shared+sqlite3~ssl~tix~tkinter~ucs4+uuid+zlib patches=0d98e93,4c24573,f2fd060 arch=linux-rhel7-sandybridge
cof4dip ^libffi@3.4.2%gcc@12.1.0 arch=linux-rhel7-sandybridge
2rhnl34 ^sqlite@3.38.5%gcc@12.1.0+column_metadata+dynamic_extensions+fts~functions+rtree arch=linux-rhel7-sandybridge
o74yelu ^util-linux-uuid@2.37.4%gcc@12.1.0 arch=linux-rhel7-sandybridge
vyov7jm ^perl-data-dumper@2.173%gcc@12.1.0 arch=linux-rhel7-sandybridge
```
### Error message
error message is too long, tail below - see attached spack-build-out.txt
<details><summary>Error message</summary><pre>
53569 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `curse
s::Window::~Window()':
>> 53570 IOHandlerCursesGUI.cpp:(.text._ZN6curses6WindowD2Ev[_ZN6curses6WindowD5Ev]+0x433): undefined ref
erence to `stdscr'
53571 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `curse
s::Window::~Window()':
>> 53572 IOHandlerCursesGUI.cpp:(.text._ZN6curses6WindowD0Ev[_ZN6curses6WindowD5Ev]+0x41b): undefined ref
erence to `stdscr'
53573 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `std::
_Sp_counted_ptr_inplace<curses::Window, std::allocator<void>, (__gnu_cxx::_Lock_policy)2>::_M_di
spose()':
>> 53574 IOHandlerCursesGUI.cpp:(.text._ZNSt23_Sp_counted_ptr_inplaceIN6curses6WindowESaIvELN9__gnu_cxx12
_Lock_policyE2EE10_M_disposeEv[_ZNSt23_Sp_counted_ptr_inplaceIN6curses6WindowESaIvELN9__gnu_cxx1
2_Lock_policyE2EE10_M_disposeEv]+0x423): undefined reference to `stdscr'
53575 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `curse
s::Application::GetMainWindow()':
>> 53576 IOHandlerCursesGUI.cpp:(.text._ZN6curses11Application13GetMainWindowEv[_ZN6curses11Application13
GetMainWindowEv]+0x6e): undefined reference to `stdscr'
53577 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `Row::
DrawTreeForChild(curses::Window&, Row*, unsigned int)':
>> 53578 IOHandlerCursesGUI.cpp:(.text._ZN3Row16DrawTreeForChildERN6curses6WindowEPS_j[_ZN3Row16DrawTreeF
orChildERN6curses6WindowEPS_j]+0x4b): undefined reference to `acs_map'
>> 53579 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN3Row16DrawTreeForChildERN6curs
es6WindowEPS_j[_ZN3Row16DrawTreeForChildERN6curses6WindowEPS_j]+0x7b): undefined reference to `a
cs_map'
>> 53580 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN3Row16DrawTreeForChildERN6curs
es6WindowEPS_j[_ZN3Row16DrawTreeForChildERN6curses6WindowEPS_j]+0xb0): undefined reference to `a
cs_map'
53581 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `Value
ObjectListDelegate::DisplayRows(curses::Window&, std::vector<Row, std::allocator<Row> >&, Displa
yOptions&)':
>> 53582 IOHandlerCursesGUI.cpp:(.text._ZN23ValueObjectListDelegate11DisplayRowsERN6curses6WindowERSt6vec
torI3RowSaIS4_EER14DisplayOptions[_ZN23ValueObjectListDelegate11DisplayRowsERN6curses6WindowERSt
6vectorI3RowSaIS4_EER14DisplayOptions]+0x261): undefined reference to `acs_map'
>> 53583 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN23ValueObjectListDelegate11Dis
playRowsERN6curses6WindowERSt6vectorI3RowSaIS4_EER14DisplayOptions[_ZN23ValueObjectListDelegate1
1DisplayRowsERN6curses6WindowERSt6vectorI3RowSaIS4_EER14DisplayOptions]+0x453): undefined refere
nce to `acs_map'
53584 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o):IOHandlerCursesGUI.
cpp:(.text._ZN23ValueObjectListDelegate11DisplayRowsERN6curses6WindowERSt6vectorI3RowSaIS4_EER14
DisplayOptions[_ZN23ValueObjectListDelegate11DisplayRowsERN6curses6WindowERSt6vectorI3RowSaIS4_E
ER14DisplayOptions]+0x493): more undefined references to `acs_map' follow
53585 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `lldb_
private::IOHandlerCursesGUI::Activate() [clone .localalias]':
>> 53586 IOHandlerCursesGUI.cpp:(.text._ZN12lldb_private18IOHandlerCursesGUI8ActivateEv+0xbc3): undefined
reference to `curs_set'
>> 53587 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN12lldb_private18IOHandlerCurse
sGUI8ActivateEv+0xbcf): undefined reference to `stdscr'
>> 53588 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN12lldb_private18IOHandlerCurse
sGUI8ActivateEv+0xbdc): undefined reference to `keypad'
>> 53589 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN12lldb_private18IOHandlerCurse
sGUI8ActivateEv+0x14f8): undefined reference to `define_key'
>> 53590 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN12lldb_private18IOHandlerCurse
sGUI8ActivateEv+0x1509): undefined reference to `define_key'
53591 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `Appli
cationDelegate::MenuDelegateAction(curses::Menu&)':
>> 53592 IOHandlerCursesGUI.cpp:(.text._ZN19ApplicationDelegate18MenuDelegateActionERN6curses4MenuE[_ZN19
ApplicationDelegate18MenuDelegateActionERN6curses4MenuE]+0xa55): undefined reference to `stdscr'
>> 53593 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN19ApplicationDelegate18MenuDel
egateActionERN6curses4MenuE[_ZN19ApplicationDelegate18MenuDelegateActionERN6curses4MenuE]+0xbd8)
: undefined reference to `stdscr'
>> 53594 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN19ApplicationDelegate18MenuDel
egateActionERN6curses4MenuE[_ZN19ApplicationDelegate18MenuDelegateActionERN6curses4MenuE]+0xd27)
: undefined reference to `stdscr'
>> 53595 collect2: error: ld returned 1 exit status
53596 ninja: build stopped: subcommand failed.
</pre></details>
### Information on your system
* **Spack:** 0.19.0.dev0
* **Python:** 3.6.8
* **Platform:** linux-rhel7-sandybridge
* **Concretizer:** clingo
### Additional information
[spack-build-out.txt](https://github.com/spack/spack/files/9330139/spack-build-out.txt)
[spack-build-env.txt](https://github.com/spack/spack/files/9330141/spack-build-env.txt)
@haampie
@trws
### General information
- [X] I have run `spack debug report` and reported the version of Spack/Python/Platform
- [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers
- [X] I have uploaded the build log and environment files
- [X] I have searched the issues of this repo and believe this is not a duplicate
|
1.0
|
Installation issue: llvm - ### Steps to reproduce the issue
```console
$ spack spec -I llvm@14.0.6
Input spec
--------------------------------
llvm@14.0.6
Concretized
--------------------------------
cauws4d llvm@14.0.6%gcc@12.1.0+clang+compiler-rt~cuda~flang+gold+internal_unwind~ipo+libcxx~link_llvm_dylib+lld+lldb+llvm_dylib~mlir+omp_as_runtime~omp_debug~omp_tsan+polly~python~split_dwarf~z3 build_type=Release patches=6379168,d85ef51,f920173 shlib_symbol_version=none targets=none version_suffix=none arch=linux-rhel7-sandybridge
bz56uqf ^binutils@2.38%gcc@12.1.0~gas+gold~headers~interwork+ld~libiberty~lto+nls+plugins libs=shared,static arch=linux-rhel7-sandybridge
qanjaa2 ^diffutils@3.8%gcc@12.1.0 arch=linux-rhel7-sandybridge
aqkjcnx ^libiconv@1.16%gcc@12.1.0 libs=shared,static arch=linux-rhel7-sandybridge
xoc2mab ^gettext@0.21%gcc@12.1.0+bzip2+curses+git~libunistring+libxml2+tar+xz arch=linux-rhel7-sandybridge
v6vqvkk ^bzip2@1.0.8%gcc@12.1.0~debug~pic+shared arch=linux-rhel7-sandybridge
b324xkn ^libxml2@2.9.13%gcc@12.1.0~python arch=linux-rhel7-sandybridge
74jzup7 ^pkgconf@1.8.0%gcc@12.1.0 arch=linux-rhel7-sandybridge
mh24b7l ^xz@5.2.5%gcc@12.1.0~pic libs=shared,static arch=linux-rhel7-sandybridge
v7akdfg ^zlib@1.2.12%gcc@12.1.0+optimize+pic+shared patches=0d38234 arch=linux-rhel7-sandybridge
gnzo6m5 ^ncurses@6.2%gcc@12.1.0~symlinks+termlib abi=none arch=linux-rhel7-sandybridge
wvyqet2 ^tar@1.34%gcc@12.1.0 zip=pigz arch=linux-rhel7-sandybridge
hudp2d7 ^pigz@2.7%gcc@12.1.0 arch=linux-rhel7-sandybridge
wh37sg7 ^zstd@1.5.2%gcc@12.1.0+programs compression=none libs=shared,static arch=linux-rhel7-sandybridge
aewh5et ^cmake@3.23.3%gcc@12.1.0~doc+ncurses~ownlibs~qt build_type=Release arch=linux-rhel7-sandybridge
gizsjhb ^curl@7.84.0%gcc@12.1.0~gssapi~ldap~libidn2~librtmp~libssh~libssh2~nghttp2 libs=shared,static patches=9b011c9 tls=gnutls arch=linux-rhel7-sandybridge
s4lvb3x ^gnutls@3.6.15%gcc@12.1.0~guile+zlib arch=linux-rhel7-sandybridge
ztttlpd ^libidn2@2.3.0%gcc@12.1.0 arch=linux-rhel7-sandybridge
gxsvuva ^libunistring@0.9.10%gcc@12.1.0 arch=linux-rhel7-sandybridge
brncabs ^nettle@3.4.1%gcc@12.1.0 arch=linux-rhel7-sandybridge
b7e3ysy ^gmp@6.2.1%gcc@12.1.0 libs=shared,static arch=linux-rhel7-sandybridge
lak2y32 ^autoconf@2.69%gcc@12.1.0 patches=7793209 arch=linux-rhel7-sandybridge
yc6hkxy ^automake@1.16.5%gcc@12.1.0 arch=linux-rhel7-sandybridge
g2sk6lg ^perl@5.34.1%gcc@12.1.0+cpanm+shared+threads arch=linux-rhel7-sandybridge
oqbulwu ^berkeley-db@18.1.40%gcc@12.1.0+cxx~docs+stl patches=b231fcc arch=linux-rhel7-sandybridge
fze2c5x ^gdbm@1.19%gcc@12.1.0 arch=linux-rhel7-sandybridge
op7l5dw ^readline@8.1.2%gcc@12.1.0 arch=linux-rhel7-sandybridge
iei7lwt ^libtool@2.4.7%gcc@12.1.0 arch=linux-rhel7-sandybridge
g3ummln ^m4@1.4.19%gcc@12.1.0+sigsegv patches=9dc5fbd,bfdffa7 arch=linux-rhel7-sandybridge
hsgdcj2 ^libsigsegv@2.13%gcc@12.1.0 arch=linux-rhel7-sandybridge
c4t27bx ^expat@2.4.8%gcc@12.1.0+libbsd arch=linux-rhel7-sandybridge
sn5hu6p ^libbsd@0.11.5%gcc@12.1.0 arch=linux-rhel7-sandybridge
a45ocsx ^libmd@1.0.4%gcc@12.1.0 arch=linux-rhel7-sandybridge
htepgqg ^libarchive@3.5.2%gcc@12.1.0+iconv compression=bz2lib,lz4,lzma,lzo2,zlib,zstd crypto=mbedtls libs=shared,static programs=none xar=expat arch=linux-rhel7-sandybridge
b6lquau ^lz4@1.9.3%gcc@12.1.0 libs=shared,static arch=linux-rhel7-sandybridge
6yddcsh ^lzo@2.10%gcc@12.1.0 libs=shared,static arch=linux-rhel7-sandybridge
2q2mbkr ^mbedtls@2.28.0%gcc@12.1.0+pic build_type=Release libs=static arch=linux-rhel7-sandybridge
pa5eqnr ^libuv@1.44.1%gcc@12.1.0 arch=linux-rhel7-sandybridge
nz5gxtq ^rhash@1.4.2%gcc@12.1.0 patches=093518c,3fbfe46 arch=linux-rhel7-sandybridge
shugqek ^hwloc@2.8.0%gcc@12.1.0~cairo~cuda~gl~libudev+libxml2~netloc~nvml~oneapi-level-zero~opencl+pci~rocm+shared arch=linux-rhel7-sandybridge
pubuiwd ^libpciaccess@0.16%gcc@12.1.0 arch=linux-rhel7-sandybridge
iz3q3z6 ^util-macros@1.19.3%gcc@12.1.0 arch=linux-rhel7-sandybridge
cfy3kmq ^libedit@3.1-20210216%gcc@12.1.0 arch=linux-rhel7-sandybridge
rb2q6ky ^ninja@1.11.0%gcc@12.1.0 arch=linux-rhel7-sandybridge
lncwse2 ^python@3.9.13%gcc@12.1.0+bz2+ctypes+dbm~debug+ensurepip+libxml2+lzma~nis~optimizations+pic+pyexpat+pythoncmd+readline+shared+sqlite3~ssl~tix~tkinter~ucs4+uuid+zlib patches=0d98e93,4c24573,f2fd060 arch=linux-rhel7-sandybridge
cof4dip ^libffi@3.4.2%gcc@12.1.0 arch=linux-rhel7-sandybridge
2rhnl34 ^sqlite@3.38.5%gcc@12.1.0+column_metadata+dynamic_extensions+fts~functions+rtree arch=linux-rhel7-sandybridge
o74yelu ^util-linux-uuid@2.37.4%gcc@12.1.0 arch=linux-rhel7-sandybridge
vyov7jm ^perl-data-dumper@2.173%gcc@12.1.0 arch=linux-rhel7-sandybridge
```
### Error message
error message is too long, tail below - see attached spack-build-out.txt
<details><summary>Error message</summary><pre>
53569 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `curse
s::Window::~Window()':
>> 53570 IOHandlerCursesGUI.cpp:(.text._ZN6curses6WindowD2Ev[_ZN6curses6WindowD5Ev]+0x433): undefined ref
erence to `stdscr'
53571 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `curse
s::Window::~Window()':
>> 53572 IOHandlerCursesGUI.cpp:(.text._ZN6curses6WindowD0Ev[_ZN6curses6WindowD5Ev]+0x41b): undefined ref
erence to `stdscr'
53573 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `std::
_Sp_counted_ptr_inplace<curses::Window, std::allocator<void>, (__gnu_cxx::_Lock_policy)2>::_M_di
spose()':
>> 53574 IOHandlerCursesGUI.cpp:(.text._ZNSt23_Sp_counted_ptr_inplaceIN6curses6WindowESaIvELN9__gnu_cxx12
_Lock_policyE2EE10_M_disposeEv[_ZNSt23_Sp_counted_ptr_inplaceIN6curses6WindowESaIvELN9__gnu_cxx1
2_Lock_policyE2EE10_M_disposeEv]+0x423): undefined reference to `stdscr'
53575 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `curse
s::Application::GetMainWindow()':
>> 53576 IOHandlerCursesGUI.cpp:(.text._ZN6curses11Application13GetMainWindowEv[_ZN6curses11Application13
GetMainWindowEv]+0x6e): undefined reference to `stdscr'
53577 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `Row::
DrawTreeForChild(curses::Window&, Row*, unsigned int)':
>> 53578 IOHandlerCursesGUI.cpp:(.text._ZN3Row16DrawTreeForChildERN6curses6WindowEPS_j[_ZN3Row16DrawTreeF
orChildERN6curses6WindowEPS_j]+0x4b): undefined reference to `acs_map'
>> 53579 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN3Row16DrawTreeForChildERN6curs
es6WindowEPS_j[_ZN3Row16DrawTreeForChildERN6curses6WindowEPS_j]+0x7b): undefined reference to `a
cs_map'
>> 53580 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN3Row16DrawTreeForChildERN6curs
es6WindowEPS_j[_ZN3Row16DrawTreeForChildERN6curses6WindowEPS_j]+0xb0): undefined reference to `a
cs_map'
53581 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `Value
ObjectListDelegate::DisplayRows(curses::Window&, std::vector<Row, std::allocator<Row> >&, Displa
yOptions&)':
>> 53582 IOHandlerCursesGUI.cpp:(.text._ZN23ValueObjectListDelegate11DisplayRowsERN6curses6WindowERSt6vec
torI3RowSaIS4_EER14DisplayOptions[_ZN23ValueObjectListDelegate11DisplayRowsERN6curses6WindowERSt
6vectorI3RowSaIS4_EER14DisplayOptions]+0x261): undefined reference to `acs_map'
>> 53583 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN23ValueObjectListDelegate11Dis
playRowsERN6curses6WindowERSt6vectorI3RowSaIS4_EER14DisplayOptions[_ZN23ValueObjectListDelegate1
1DisplayRowsERN6curses6WindowERSt6vectorI3RowSaIS4_EER14DisplayOptions]+0x453): undefined refere
nce to `acs_map'
53584 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o):IOHandlerCursesGUI.
cpp:(.text._ZN23ValueObjectListDelegate11DisplayRowsERN6curses6WindowERSt6vectorI3RowSaIS4_EER14
DisplayOptions[_ZN23ValueObjectListDelegate11DisplayRowsERN6curses6WindowERSt6vectorI3RowSaIS4_E
ER14DisplayOptions]+0x493): more undefined references to `acs_map' follow
53585 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `lldb_
private::IOHandlerCursesGUI::Activate() [clone .localalias]':
>> 53586 IOHandlerCursesGUI.cpp:(.text._ZN12lldb_private18IOHandlerCursesGUI8ActivateEv+0xbc3): undefined
reference to `curs_set'
>> 53587 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN12lldb_private18IOHandlerCurse
sGUI8ActivateEv+0xbcf): undefined reference to `stdscr'
>> 53588 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN12lldb_private18IOHandlerCurse
sGUI8ActivateEv+0xbdc): undefined reference to `keypad'
>> 53589 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN12lldb_private18IOHandlerCurse
sGUI8ActivateEv+0x14f8): undefined reference to `define_key'
>> 53590 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN12lldb_private18IOHandlerCurse
sGUI8ActivateEv+0x1509): undefined reference to `define_key'
53591 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: lib/liblldbCore.a(IOHandlerCursesGUI.cpp.o): in function `Appli
cationDelegate::MenuDelegateAction(curses::Menu&)':
>> 53592 IOHandlerCursesGUI.cpp:(.text._ZN19ApplicationDelegate18MenuDelegateActionERN6curses4MenuE[_ZN19
ApplicationDelegate18MenuDelegateActionERN6curses4MenuE]+0xa55): undefined reference to `stdscr'
>> 53593 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN19ApplicationDelegate18MenuDel
egateActionERN6curses4MenuE[_ZN19ApplicationDelegate18MenuDelegateActionERN6curses4MenuE]+0xbd8)
: undefined reference to `stdscr'
>> 53594 /nobackupnfs1/jcbecker/spack/opt/spack/linux-rhel7-sandybridge/gcc-12.1.0/binutils-2.38-bz56uqfv
g6rdr2dczyzo3jzm6jw2teu3/bin/ld: IOHandlerCursesGUI.cpp:(.text._ZN19ApplicationDelegate18MenuDel
egateActionERN6curses4MenuE[_ZN19ApplicationDelegate18MenuDelegateActionERN6curses4MenuE]+0xd27)
: undefined reference to `stdscr'
>> 53595 collect2: error: ld returned 1 exit status
53596 ninja: build stopped: subcommand failed.
</pre></details>
### Information on your system
* **Spack:** 0.19.0.dev0
* **Python:** 3.6.8
* **Platform:** linux-rhel7-sandybridge
* **Concretizer:** clingo
### Additional information
[spack-build-out.txt](https://github.com/spack/spack/files/9330139/spack-build-out.txt)
[spack-build-env.txt](https://github.com/spack/spack/files/9330141/spack-build-env.txt)
@haampie
@trws
### General information
- [X] I have run `spack debug report` and reported the version of Spack/Python/Platform
- [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers
- [X] I have uploaded the build log and environment files
- [X] I have searched the issues of this repo and believe this is not a duplicate
|
non_process
|
installation issue llvm steps to reproduce the issue console spack spec i llvm input spec llvm concretized llvm gcc clang compiler rt cuda flang gold internal unwind ipo libcxx link llvm dylib lld lldb llvm dylib mlir omp as runtime omp debug omp tsan polly python split dwarf build type release patches shlib symbol version none targets none version suffix none arch linux sandybridge binutils gcc gas gold headers interwork ld libiberty lto nls plugins libs shared static arch linux sandybridge diffutils gcc arch linux sandybridge aqkjcnx libiconv gcc libs shared static arch linux sandybridge gettext gcc curses git libunistring tar xz arch linux sandybridge gcc debug pic shared arch linux sandybridge gcc python arch linux sandybridge pkgconf gcc arch linux sandybridge xz gcc pic libs shared static arch linux sandybridge zlib gcc optimize pic shared patches arch linux sandybridge ncurses gcc symlinks termlib abi none arch linux sandybridge tar gcc zip pigz arch linux sandybridge pigz gcc arch linux sandybridge zstd gcc programs compression none libs shared static arch linux sandybridge cmake gcc doc ncurses ownlibs qt build type release arch linux sandybridge gizsjhb curl gcc gssapi ldap librtmp libssh libs shared static patches tls gnutls arch linux sandybridge gnutls gcc guile zlib arch linux sandybridge ztttlpd gcc arch linux sandybridge gxsvuva libunistring gcc arch linux sandybridge brncabs nettle gcc arch linux sandybridge gmp gcc libs shared static arch linux sandybridge autoconf gcc patches arch linux sandybridge automake gcc arch linux sandybridge perl gcc cpanm shared threads arch linux sandybridge oqbulwu berkeley db gcc cxx docs stl patches arch linux sandybridge gdbm gcc arch linux sandybridge readline gcc arch linux sandybridge libtool gcc arch linux sandybridge gcc sigsegv patches arch linux sandybridge libsigsegv gcc arch linux sandybridge expat gcc libbsd arch linux sandybridge libbsd gcc arch linux sandybridge libmd gcc arch linux sandybridge htepgqg libarchive gcc iconv compression lzma zlib zstd crypto mbedtls libs shared static programs none xar expat arch linux sandybridge gcc libs shared static arch linux sandybridge lzo gcc libs shared static arch linux sandybridge mbedtls gcc pic build type release libs static arch linux sandybridge libuv gcc arch linux sandybridge rhash gcc patches arch linux sandybridge shugqek hwloc gcc cairo cuda gl libudev netloc nvml oneapi level zero opencl pci rocm shared arch linux sandybridge pubuiwd libpciaccess gcc arch linux sandybridge util macros gcc arch linux sandybridge libedit gcc arch linux sandybridge ninja gcc arch linux sandybridge python gcc ctypes dbm debug ensurepip lzma nis optimizations pic pyexpat pythoncmd readline shared ssl tix tkinter uuid zlib patches arch linux sandybridge libffi gcc arch linux sandybridge sqlite gcc column metadata dynamic extensions fts functions rtree arch linux sandybridge util linux uuid gcc arch linux sandybridge perl data dumper gcc arch linux sandybridge error message error message is too long tail below see attached spack build out txt error message jcbecker spack opt spack linux sandybridge gcc binutils bin ld lib liblldbcore a iohandlercursesgui cpp o in function curse s window window iohandlercursesgui cpp text undefined ref erence to stdscr jcbecker spack opt spack linux sandybridge gcc binutils bin ld lib liblldbcore a iohandlercursesgui cpp o in function curse s window window iohandlercursesgui cpp text undefined ref erence to stdscr jcbecker spack opt spack linux sandybridge gcc binutils bin ld lib liblldbcore a iohandlercursesgui cpp o in function std sp counted ptr inplace gnu cxx lock policy m di spose iohandlercursesgui cpp text sp counted ptr gnu lock m disposeev sp counted ptr gnu lock m disposeev undefined reference to stdscr jcbecker spack opt spack linux sandybridge gcc binutils bin ld lib liblldbcore a iohandlercursesgui cpp o in function curse s application getmainwindow iohandlercursesgui cpp text getmainwindowev undefined reference to stdscr jcbecker spack opt spack linux sandybridge gcc binutils bin ld lib liblldbcore a iohandlercursesgui cpp o in function row drawtreeforchild curses window row unsigned int iohandlercursesgui cpp text j j undefined reference to acs map jcbecker spack opt spack linux sandybridge gcc binutils bin ld iohandlercursesgui cpp text j undefined reference to a cs map jcbecker spack opt spack linux sandybridge gcc binutils bin ld iohandlercursesgui cpp text j undefined reference to a cs map jcbecker spack opt spack linux sandybridge gcc binutils bin ld lib liblldbcore a iohandlercursesgui cpp o in function value objectlistdelegate displayrows curses window std vector displa yoptions iohandlercursesgui cpp text undefined reference to acs map jcbecker spack opt spack linux sandybridge gcc binutils bin ld iohandlercursesgui cpp text undefined refere nce to acs map jcbecker spack opt spack linux sandybridge gcc binutils bin ld lib liblldbcore a iohandlercursesgui cpp o iohandlercursesgui cpp text displayoptions e more undefined references to acs map follow jcbecker spack opt spack linux sandybridge gcc binutils bin ld lib liblldbcore a iohandlercursesgui cpp o in function lldb private iohandlercursesgui activate iohandlercursesgui cpp text undefined reference to curs set jcbecker spack opt spack linux sandybridge gcc binutils bin ld iohandlercursesgui cpp text undefined reference to stdscr jcbecker spack opt spack linux sandybridge gcc binutils bin ld iohandlercursesgui cpp text undefined reference to keypad jcbecker spack opt spack linux sandybridge gcc binutils bin ld iohandlercursesgui cpp text undefined reference to define key jcbecker spack opt spack linux sandybridge gcc binutils bin ld iohandlercursesgui cpp text undefined reference to define key jcbecker spack opt spack linux sandybridge gcc binutils bin ld lib liblldbcore a iohandlercursesgui cpp o in function appli cationdelegate menudelegateaction curses menu iohandlercursesgui cpp text undefined reference to stdscr jcbecker spack opt spack linux sandybridge gcc binutils bin ld iohandlercursesgui cpp text undefined reference to stdscr jcbecker spack opt spack linux sandybridge gcc binutils bin ld iohandlercursesgui cpp text undefined reference to stdscr error ld returned exit status ninja build stopped subcommand failed information on your system spack python platform linux sandybridge concretizer clingo additional information haampie trws general information i have run spack debug report and reported the version of spack python platform i have run spack maintainers and mentioned any maintainers i have uploaded the build log and environment files i have searched the issues of this repo and believe this is not a duplicate
| 0
|
2,171
| 5,019,878,526
|
IssuesEvent
|
2016-12-14 13:17:56
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
closed
|
[subtitles] [FR] AU FIL DE LA FRANCE - 1000 POINTS DE DIFFUSION DU PROGAMME : MERCI !
|
Language: French Process: [4] Ready for review (2)
|
# Video title
AU FIL DE LA FRANCE - 1000 POINTS DE DIFFUSION DU PROGAMME : MERCI !
# URL
https://www.youtube.com/watch?v=l6xhk9ltlCE
# Youtube subtitles language
Français
# Duration
2:36
# URL subtitles
https://www.youtube.com/timedtext_editor?action_mde_edit_form=1&ref=watch&lang=fr&v=l6xhk9ltlCE&tab=captions&bl=vmp&ui=hd
|
1.0
|
[subtitles] [FR] AU FIL DE LA FRANCE - 1000 POINTS DE DIFFUSION DU PROGAMME : MERCI ! - # Video title
AU FIL DE LA FRANCE - 1000 POINTS DE DIFFUSION DU PROGAMME : MERCI !
# URL
https://www.youtube.com/watch?v=l6xhk9ltlCE
# Youtube subtitles language
Français
# Duration
2:36
# URL subtitles
https://www.youtube.com/timedtext_editor?action_mde_edit_form=1&ref=watch&lang=fr&v=l6xhk9ltlCE&tab=captions&bl=vmp&ui=hd
|
process
|
au fil de la france points de diffusion du progamme merci video title au fil de la france points de diffusion du progamme merci url youtube subtitles language français duration url subtitles
| 1
|
68,982
| 3,294,688,412
|
IssuesEvent
|
2015-10-31 08:18:40
|
jgirald/ES2015C
|
https://api.github.com/repos/jgirald/ES2015C
|
closed
|
Implement the logic for a unit to attack another unit
|
Character Development High Priority Team C
|
**Product backlog item:** As a player, I want to create military units, so that I can attack/defend from other players. [High priority]
**Estimated effort:** 4 hours
**Ending condition:** Once the API function for attacking another unit is called (e.g. by right clicking , the unit must move to the desired enemy unit to attack. Once the unit reaches the enemy unit, it must change to an attacking state. In this state, the unit will inflict damage to the enemy unit while it is alive. If available, an attack animation must play. Care must be taken that range units can attack at a distance. Once the enemy unit is out of range or dead, the unit must return to the idle state.
|
1.0
|
Implement the logic for a unit to attack another unit - **Product backlog item:** As a player, I want to create military units, so that I can attack/defend from other players. [High priority]
**Estimated effort:** 4 hours
**Ending condition:** Once the API function for attacking another unit is called (e.g. by right clicking , the unit must move to the desired enemy unit to attack. Once the unit reaches the enemy unit, it must change to an attacking state. In this state, the unit will inflict damage to the enemy unit while it is alive. If available, an attack animation must play. Care must be taken that range units can attack at a distance. Once the enemy unit is out of range or dead, the unit must return to the idle state.
|
non_process
|
implement the logic for a unit to attack another unit product backlog item as a player i want to create military units so that i can attack defend from other players estimated effort hours ending condition once the api function for attacking another unit is called e g by right clicking the unit must move to the desired enemy unit to attack once the unit reaches the enemy unit it must change to an attacking state in this state the unit will inflict damage to the enemy unit while it is alive if available an attack animation must play care must be taken that range units can attack at a distance once the enemy unit is out of range or dead the unit must return to the idle state
| 0
|
471,699
| 13,608,444,451
|
IssuesEvent
|
2020-09-23 02:23:10
|
a2000-erp-team/WEBERP
|
https://api.github.com/repos/a2000-erp-team/WEBERP
|
opened
|
JAMES WILSON Sales Quotation is free text. Is there a setting for me to call out the inventory lists and all its attributes?
|
ABIGAIL High Priority
|
SALES-SALES-SALES QUOTATION

|
1.0
|
JAMES WILSON Sales Quotation is free text. Is there a setting for me to call out the inventory lists and all its attributes? - SALES-SALES-SALES QUOTATION

|
non_process
|
james wilson sales quotation is free text is there a setting for me to call out the inventory lists and all its attributes sales sales sales quotation
| 0
|
420,452
| 12,238,285,224
|
IssuesEvent
|
2020-05-04 19:32:08
|
ChainSafe/ChainBridge
|
https://api.github.com/repos/ChainSafe/ChainBridge
|
closed
|
Vote against invalid proposals
|
Priority: 4 - Low substrate
|
Initially we are just approving valid proposals, we need to additional vote against invalid proposals. We should reconstruct the call, retrieve a copy from the chain, and compare them.
|
1.0
|
Vote against invalid proposals - Initially we are just approving valid proposals, we need to additional vote against invalid proposals. We should reconstruct the call, retrieve a copy from the chain, and compare them.
|
non_process
|
vote against invalid proposals initially we are just approving valid proposals we need to additional vote against invalid proposals we should reconstruct the call retrieve a copy from the chain and compare them
| 0
|
13,708
| 16,469,094,129
|
IssuesEvent
|
2021-05-23 03:16:52
|
rdoddanavar/hpr-sim
|
https://api.github.com/repos/rdoddanavar/hpr-sim
|
closed
|
YAML processor: util_yaml.py
|
high-priority pre-processing
|
Extend YAML capabilities:
- Internal references: `ref( ... )`
- Arithmetic expressions: `eval( ... )` or `expr( ... )`
|
1.0
|
YAML processor: util_yaml.py - Extend YAML capabilities:
- Internal references: `ref( ... )`
- Arithmetic expressions: `eval( ... )` or `expr( ... )`
|
process
|
yaml processor util yaml py extend yaml capabilities internal references ref arithmetic expressions eval or expr
| 1
|
433,884
| 12,511,899,981
|
IssuesEvent
|
2020-06-02 21:29:34
|
single-spa/single-spa-angular
|
https://api.github.com/repos/single-spa/single-spa-angular
|
closed
|
URL Redirection Infinite Loop
|
priority: critical ready to release type: bug/fix
|
Issue seems to relate to this comment;
https://github.com/CanopyTax/single-spa-angular/issues/94#issuecomment-520550455
Creating a new ticket as I cannot reproduce using the browser back button.
This bug is reproducible in the vanilla example for coexisting angular micro-frontends found here:
https://github.com/joeldenning/coexisting-angular-microfrontends
Reproduction steps;
Using the navigation app, trigger a route change, and then quickly trigger another nav change.
Issue is reproducible on;
IE11 (Very common)
Firefox (Mediocre-Hard)
Chrome (Rare)
Notes;
In our application we have found that having more simultaneous angular apps running in parallel triggers the issue more commonly, however this bug is reproducible with only one app actively loaded by the single-spa (however with difficultly even in IE11). (Issue does not occur at all with the app running in angular standalone without the spa)
The issue appears to be also related to Router.forRoot in some way, upon removing this from an app, the issue becomes more difficult to reproduce even with IE11, a similar effect to whether the app wasn't running.
This seems to be related to the speed of the navigation. Navigating twice very quickly causes this effect.
If struggling to reproduce, we found that having a button which changes to swap between two pages and spam clicking it causes the bug to reproduce on IE11 every time.
The navigations do not need to be between separate apps, routing between two pages on the same app causes the issue.
|
1.0
|
URL Redirection Infinite Loop - Issue seems to relate to this comment;
https://github.com/CanopyTax/single-spa-angular/issues/94#issuecomment-520550455
Creating a new ticket as I cannot reproduce using the browser back button.
This bug is reproducible in the vanilla example for coexisting angular micro-frontends found here:
https://github.com/joeldenning/coexisting-angular-microfrontends
Reproduction steps;
Using the navigation app, trigger a route change, and then quickly trigger another nav change.
Issue is reproducible on;
IE11 (Very common)
Firefox (Mediocre-Hard)
Chrome (Rare)
Notes;
In our application we have found that having more simultaneous angular apps running in parallel triggers the issue more commonly, however this bug is reproducible with only one app actively loaded by the single-spa (however with difficultly even in IE11). (Issue does not occur at all with the app running in angular standalone without the spa)
The issue appears to be also related to Router.forRoot in some way, upon removing this from an app, the issue becomes more difficult to reproduce even with IE11, a similar effect to whether the app wasn't running.
This seems to be related to the speed of the navigation. Navigating twice very quickly causes this effect.
If struggling to reproduce, we found that having a button which changes to swap between two pages and spam clicking it causes the bug to reproduce on IE11 every time.
The navigations do not need to be between separate apps, routing between two pages on the same app causes the issue.
|
non_process
|
url redirection infinite loop issue seems to relate to this comment creating a new ticket as i cannot reproduce using the browser back button this bug is reproducible in the vanilla example for coexisting angular micro frontends found here reproduction steps using the navigation app trigger a route change and then quickly trigger another nav change issue is reproducible on very common firefox mediocre hard chrome rare notes in our application we have found that having more simultaneous angular apps running in parallel triggers the issue more commonly however this bug is reproducible with only one app actively loaded by the single spa however with difficultly even in issue does not occur at all with the app running in angular standalone without the spa the issue appears to be also related to router forroot in some way upon removing this from an app the issue becomes more difficult to reproduce even with a similar effect to whether the app wasn t running this seems to be related to the speed of the navigation navigating twice very quickly causes this effect if struggling to reproduce we found that having a button which changes to swap between two pages and spam clicking it causes the bug to reproduce on every time the navigations do not need to be between separate apps routing between two pages on the same app causes the issue
| 0
|
16,907
| 22,217,574,553
|
IssuesEvent
|
2022-06-08 04:25:26
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
install tulsi failed on Mac
|
P3 more data needed platform: apple type: support / not a bug (process) team-Rules-CPP
|
I build mediapipe examples using bazel on Mac. 'Hello world' is succeed. Then I need to build tulsi.
I think it is a question about bazel, not tulsi.
Mac os 10.13.6
bazel 4.1.0
liuchiyudeMacBook-Pro:tulsi lcy$ sh build_and_run.sh
Starting local Bazel server and connecting to it...
Analyzing: target //:tulsi (44 packages loaded, 217 targets configured)
Fetching @local_config_cc; fetching 99s
after 105s, I got this:
liuchiyudeMacBook-Pro:tulsi lcy$ sh build_and_run.sh
Starting local Bazel server and connecting to it...
ERROR: /private/var/tmp/_bazel_lcy/5c49963cb3cdabd3637667642b95bec0/external/local_config_cc/BUILD:48:19: in cc_toolchain_suite rule @local_config_cc//:toolchain: cc_toolchain_suite '@local_config_cc//:toolchain' does not contain a toolchain for cpu 'darwin_x86_64'
I could not solve the problem for a long time. I come from China, is this a network problem? How can I find this "@local_config_cc"
|
1.0
|
install tulsi failed on Mac - I build mediapipe examples using bazel on Mac. 'Hello world' is succeed. Then I need to build tulsi.
I think it is a question about bazel, not tulsi.
Mac os 10.13.6
bazel 4.1.0
liuchiyudeMacBook-Pro:tulsi lcy$ sh build_and_run.sh
Starting local Bazel server and connecting to it...
Analyzing: target //:tulsi (44 packages loaded, 217 targets configured)
Fetching @local_config_cc; fetching 99s
after 105s, I got this:
liuchiyudeMacBook-Pro:tulsi lcy$ sh build_and_run.sh
Starting local Bazel server and connecting to it...
ERROR: /private/var/tmp/_bazel_lcy/5c49963cb3cdabd3637667642b95bec0/external/local_config_cc/BUILD:48:19: in cc_toolchain_suite rule @local_config_cc//:toolchain: cc_toolchain_suite '@local_config_cc//:toolchain' does not contain a toolchain for cpu 'darwin_x86_64'
I could not solve the problem for a long time. I come from China, is this a network problem? How can I find this "@local_config_cc"
|
process
|
install tulsi failed on mac i build mediapipe examples using bazel on mac hello world is succeed then i need to build tulsi i think it is a question about bazel not tulsi mac os bazel liuchiyudemacbook pro tulsi lcy sh build and run sh starting local bazel server and connecting to it analyzing target tulsi packages loaded targets configured fetching local config cc fetching after i got this liuchiyudemacbook pro tulsi lcy sh build and run sh starting local bazel server and connecting to it error private var tmp bazel lcy external local config cc build in cc toolchain suite rule local config cc toolchain cc toolchain suite local config cc toolchain does not contain a toolchain for cpu darwin i could not solve the problem for a long time i come from china is this a network problem how can i find this local config cc
| 1
|
4,997
| 7,834,154,896
|
IssuesEvent
|
2018-06-16 09:36:18
|
hollowverse/hollowverse
|
https://api.github.com/repos/hollowverse/hollowverse
|
opened
|
Move management of process-image resources to the respective CloudFormation stack
|
repo: process-image
|
As part of our efforts to move all infrastructure management to code, we should move the management of the following resources used by `process-image` to the respective `serverless.yml` file:
* S3 bucket for unprocessed images
* S3 bucket for processed images
* CloudFront distribution for the processed images
* Route 53 records for `photos.hollowverse.com`
|
1.0
|
Move management of process-image resources to the respective CloudFormation stack - As part of our efforts to move all infrastructure management to code, we should move the management of the following resources used by `process-image` to the respective `serverless.yml` file:
* S3 bucket for unprocessed images
* S3 bucket for processed images
* CloudFront distribution for the processed images
* Route 53 records for `photos.hollowverse.com`
|
process
|
move management of process image resources to the respective cloudformation stack as part of our efforts to move all infrastructure management to code we should move the management of the following resources used by process image to the respective serverless yml file bucket for unprocessed images bucket for processed images cloudfront distribution for the processed images route records for photos hollowverse com
| 1
|
4,444
| 7,313,503,148
|
IssuesEvent
|
2018-03-01 01:29:03
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Instructions result in a project that starts with 15 errors ...
|
cxp functions in-process triaged
|
In the section entitled "Create an Azure Functions project" - I followed steps 1 through 4 to create the project. Unfortunately, the file "Function1.cs" reports the following errors:
Severity Code Description Project File Line Suppression State
Error CS0234 The type or namespace name 'Http' does not exist in the namespace 'System.Net' (are you missing an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 3 Active
Error CS0234 The type or namespace name 'Azure' does not exist in the namespace 'Microsoft' (are you missing an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 5 Active
Error CS0234 The type or namespace name 'Azure' does not exist in the namespace 'Microsoft' (are you missing an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 6 Active
Error CS0234 The type or namespace name 'Azure' does not exist in the namespace 'Microsoft' (are you missing an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 7 Active
Error CS0246 The type or namespace name 'FunctionNameAttribute' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 13 Active
Error CS0246 The type or namespace name 'FunctionName' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 13 Active
Error CS0246 The type or namespace name 'HttpResponseMessage' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error CS0246 The type or namespace name 'HttpTriggerAttribute' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error CS0246 The type or namespace name 'HttpTrigger' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error CS0103 The name 'AuthorizationLevel' does not exist in the current context test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error CS0246 The type or namespace name 'Route' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error CS0246 The type or namespace name 'HttpRequestMessage' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error CS0246 The type or namespace name 'TraceWriter' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error Assets file 'D:\vsprojects\serverless_functions\v1\test1\obj\project.assets.json' not found. Run a NuGet package restore to generate this file. test1 C:\Program Files\dotnet\sdk\2.1.4\Sdks\Microsoft.NET.Sdk\build\Microsoft.PackageDependencyResolution.targets 327
Error Assets file 'D:\vsprojects\serverless_functions\v1\test1\obj\project.assets.json' not found. Run a NuGet package restore to generate this file. test1 C:\Program Files\dotnet\sdk\2.1.4\Sdks\Microsoft.NET.Sdk\build\Microsoft.PackageDependencyResolution.targets 167
So, the template supplied by Visual Studio is either corrupt or mis-configured as it is obviously missing several dependencies!
Using Visual Studio 2017 15.5.7 Community.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 369aefa6-74da-3001-f835-15abdfffd5d6
* Version Independent ID: 2cc2a2db-d059-3252-7a1a-4261b7d3c37c
* [Content](https://docs.microsoft.com/en-us/azure/azure-functions/functions-develop-vs)
* [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/azure-functions/functions-develop-vs.md)
* Service: functions
|
1.0
|
Instructions result in a project that starts with 15 errors ... - In the section entitled "Create an Azure Functions project" - I followed steps 1 through 4 to create the project. Unfortunately, the file "Function1.cs" reports the following errors:
Severity Code Description Project File Line Suppression State
Error CS0234 The type or namespace name 'Http' does not exist in the namespace 'System.Net' (are you missing an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 3 Active
Error CS0234 The type or namespace name 'Azure' does not exist in the namespace 'Microsoft' (are you missing an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 5 Active
Error CS0234 The type or namespace name 'Azure' does not exist in the namespace 'Microsoft' (are you missing an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 6 Active
Error CS0234 The type or namespace name 'Azure' does not exist in the namespace 'Microsoft' (are you missing an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 7 Active
Error CS0246 The type or namespace name 'FunctionNameAttribute' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 13 Active
Error CS0246 The type or namespace name 'FunctionName' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 13 Active
Error CS0246 The type or namespace name 'HttpResponseMessage' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error CS0246 The type or namespace name 'HttpTriggerAttribute' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error CS0246 The type or namespace name 'HttpTrigger' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error CS0103 The name 'AuthorizationLevel' does not exist in the current context test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error CS0246 The type or namespace name 'Route' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error CS0246 The type or namespace name 'HttpRequestMessage' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error CS0246 The type or namespace name 'TraceWriter' could not be found (are you missing a using directive or an assembly reference?) test1 D:\vsprojects\serverless_functions\v1\test1\Function1.cs 14 Active
Error Assets file 'D:\vsprojects\serverless_functions\v1\test1\obj\project.assets.json' not found. Run a NuGet package restore to generate this file. test1 C:\Program Files\dotnet\sdk\2.1.4\Sdks\Microsoft.NET.Sdk\build\Microsoft.PackageDependencyResolution.targets 327
Error Assets file 'D:\vsprojects\serverless_functions\v1\test1\obj\project.assets.json' not found. Run a NuGet package restore to generate this file. test1 C:\Program Files\dotnet\sdk\2.1.4\Sdks\Microsoft.NET.Sdk\build\Microsoft.PackageDependencyResolution.targets 167
So, the template supplied by Visual Studio is either corrupt or mis-configured as it is obviously missing several dependencies!
Using Visual Studio 2017 15.5.7 Community.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 369aefa6-74da-3001-f835-15abdfffd5d6
* Version Independent ID: 2cc2a2db-d059-3252-7a1a-4261b7d3c37c
* [Content](https://docs.microsoft.com/en-us/azure/azure-functions/functions-develop-vs)
* [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/azure-functions/functions-develop-vs.md)
* Service: functions
|
process
|
instructions result in a project that starts with errors in the section entitled create an azure functions project i followed steps through to create the project unfortunately the file cs reports the following errors severity code description project file line suppression state error the type or namespace name http does not exist in the namespace system net are you missing an assembly reference d vsprojects serverless functions cs active error the type or namespace name azure does not exist in the namespace microsoft are you missing an assembly reference d vsprojects serverless functions cs active error the type or namespace name azure does not exist in the namespace microsoft are you missing an assembly reference d vsprojects serverless functions cs active error the type or namespace name azure does not exist in the namespace microsoft are you missing an assembly reference d vsprojects serverless functions cs active error the type or namespace name functionnameattribute could not be found are you missing a using directive or an assembly reference d vsprojects serverless functions cs active error the type or namespace name functionname could not be found are you missing a using directive or an assembly reference d vsprojects serverless functions cs active error the type or namespace name httpresponsemessage could not be found are you missing a using directive or an assembly reference d vsprojects serverless functions cs active error the type or namespace name httptriggerattribute could not be found are you missing a using directive or an assembly reference d vsprojects serverless functions cs active error the type or namespace name httptrigger could not be found are you missing a using directive or an assembly reference d vsprojects serverless functions cs active error the name authorizationlevel does not exist in the current context d vsprojects serverless functions cs active error the type or namespace name route could not be found are you missing a using directive or an assembly reference d vsprojects serverless functions cs active error the type or namespace name httprequestmessage could not be found are you missing a using directive or an assembly reference d vsprojects serverless functions cs active error the type or namespace name tracewriter could not be found are you missing a using directive or an assembly reference d vsprojects serverless functions cs active error assets file d vsprojects serverless functions obj project assets json not found run a nuget package restore to generate this file c program files dotnet sdk sdks microsoft net sdk build microsoft packagedependencyresolution targets error assets file d vsprojects serverless functions obj project assets json not found run a nuget package restore to generate this file c program files dotnet sdk sdks microsoft net sdk build microsoft packagedependencyresolution targets so the template supplied by visual studio is either corrupt or mis configured as it is obviously missing several dependencies using visual studio community document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id service functions
| 1
|
102,053
| 16,545,408,981
|
IssuesEvent
|
2021-05-27 23:01:28
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
Vulnerability roundup 102: openvpn-2.5.0: 2 advisories [7.5]
|
1.severity: security
|
[search](https://search.nix.gsc.io/?q=openvpn&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=openvpn+in%3Apath&type=Code)
* [ ] [CVE-2020-15078](https://nvd.nist.gov/vuln/detail/CVE-2020-15078) CVSSv3=7.5 (nixos-unstable)
* [x] [CVE-2020-27569](https://nvd.nist.gov/vuln/detail/CVE-2020-27569) CVSSv3=7.5 (nixos-unstable)
Scanned versions: nixos-unstable: ea7d4aa9b82.
Cc @peterhoeg
Cc @viric
|
True
|
Vulnerability roundup 102: openvpn-2.5.0: 2 advisories [7.5] - [search](https://search.nix.gsc.io/?q=openvpn&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=openvpn+in%3Apath&type=Code)
* [ ] [CVE-2020-15078](https://nvd.nist.gov/vuln/detail/CVE-2020-15078) CVSSv3=7.5 (nixos-unstable)
* [x] [CVE-2020-27569](https://nvd.nist.gov/vuln/detail/CVE-2020-27569) CVSSv3=7.5 (nixos-unstable)
Scanned versions: nixos-unstable: ea7d4aa9b82.
Cc @peterhoeg
Cc @viric
|
non_process
|
vulnerability roundup openvpn advisories nixos unstable nixos unstable scanned versions nixos unstable cc peterhoeg cc viric
| 0
|
104,496
| 16,616,843,958
|
IssuesEvent
|
2021-06-02 17:50:18
|
Dima2021/t-vault
|
https://api.github.com/repos/Dima2021/t-vault
|
opened
|
CVE-2014-7191 (Medium) detected in qs-0.4.2.tgz
|
security vulnerability
|
## CVE-2014-7191 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-0.4.2.tgz</b></p></summary>
<p>querystring parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-0.4.2.tgz">https://registry.npmjs.org/qs/-/qs-0.4.2.tgz</a></p>
<p>Path to dependency file: t-vault/tvaultui/package.json</p>
<p>Path to vulnerable library: t-vault/tvaultui/node_modules/express/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.9.12.tgz (Root Library)
- browser-sync-ui-0.5.19.tgz
- weinre-2.0.0-pre-I0Z7U9OV.tgz
- express-2.5.11.tgz
- :x: **qs-0.4.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Dima2021/t-vault/commit/259885b704776a5554c5d008b51b19c9b0ea9fd5">259885b704776a5554c5d008b51b19c9b0ea9fd5</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The qs module before 1.0.0 in Node.js does not call the compact function for array data, which allows remote attackers to cause a denial of service (memory consumption) by using a large index value to create a sparse array.
<p>Publish Date: 2014-10-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-7191>CVE-2014-7191</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-7191">https://nvd.nist.gov/vuln/detail/CVE-2014-7191</a></p>
<p>Release Date: 2014-10-19</p>
<p>Fix Resolution: 1.0.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"0.4.2","packageFilePaths":["/tvaultui/package.json"],"isTransitiveDependency":true,"dependencyTree":"browser-sync:2.9.12;browser-sync-ui:0.5.19;weinre:2.0.0-pre-I0Z7U9OV;express:2.5.11;qs:0.4.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.0"}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2014-7191","vulnerabilityDetails":"The qs module before 1.0.0 in Node.js does not call the compact function for array data, which allows remote attackers to cause a denial of service (memory consumption) by using a large index value to create a sparse array.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-7191","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2014-7191 (Medium) detected in qs-0.4.2.tgz - ## CVE-2014-7191 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-0.4.2.tgz</b></p></summary>
<p>querystring parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-0.4.2.tgz">https://registry.npmjs.org/qs/-/qs-0.4.2.tgz</a></p>
<p>Path to dependency file: t-vault/tvaultui/package.json</p>
<p>Path to vulnerable library: t-vault/tvaultui/node_modules/express/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.9.12.tgz (Root Library)
- browser-sync-ui-0.5.19.tgz
- weinre-2.0.0-pre-I0Z7U9OV.tgz
- express-2.5.11.tgz
- :x: **qs-0.4.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Dima2021/t-vault/commit/259885b704776a5554c5d008b51b19c9b0ea9fd5">259885b704776a5554c5d008b51b19c9b0ea9fd5</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The qs module before 1.0.0 in Node.js does not call the compact function for array data, which allows remote attackers to cause a denial of service (memory consumption) by using a large index value to create a sparse array.
<p>Publish Date: 2014-10-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-7191>CVE-2014-7191</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-7191">https://nvd.nist.gov/vuln/detail/CVE-2014-7191</a></p>
<p>Release Date: 2014-10-19</p>
<p>Fix Resolution: 1.0.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"0.4.2","packageFilePaths":["/tvaultui/package.json"],"isTransitiveDependency":true,"dependencyTree":"browser-sync:2.9.12;browser-sync-ui:0.5.19;weinre:2.0.0-pre-I0Z7U9OV;express:2.5.11;qs:0.4.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.0"}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2014-7191","vulnerabilityDetails":"The qs module before 1.0.0 in Node.js does not call the compact function for array data, which allows remote attackers to cause a denial of service (memory consumption) by using a large index value to create a sparse array.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-7191","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in qs tgz cve medium severity vulnerability vulnerable library qs tgz querystring parser library home page a href path to dependency file t vault tvaultui package json path to vulnerable library t vault tvaultui node modules express node modules qs package json dependency hierarchy browser sync tgz root library browser sync ui tgz weinre pre tgz express tgz x qs tgz vulnerable library found in head commit a href found in base branch dev vulnerability details the qs module before in node js does not call the compact function for array data which allows remote attackers to cause a denial of service memory consumption by using a large index value to create a sparse array publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree browser sync browser sync ui weinre pre express qs isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the qs module before in node js does not call the compact function for array data which allows remote attackers to cause a denial of service memory consumption by using a large index value to create a sparse array vulnerabilityurl
| 0
|
147
| 2,577,606,786
|
IssuesEvent
|
2015-02-12 18:01:16
|
cfpb/hmda-viz-prototype
|
https://api.github.com/repos/cfpb/hmda-viz-prototype
|
opened
|
Create a directory structure
|
Processing
|
Write the files to the correct URL path:
aggregate year/state name (full name)/msa name (friendly)/3-1.json
|
1.0
|
Create a directory structure - Write the files to the correct URL path:
aggregate year/state name (full name)/msa name (friendly)/3-1.json
|
process
|
create a directory structure write the files to the correct url path aggregate year state name full name msa name friendly json
| 1
|
18,819
| 24,718,972,401
|
IssuesEvent
|
2022-10-20 09:13:14
|
vectordotdev/vector
|
https://api.github.com/repos/vectordotdev/vector
|
opened
|
Allow rerouting discarded events
|
domain: processing Epic
|
**This is a tracking issue for RFC #12217 (PR #14708).**
We want to give Vector users the assurance that events are never lost unexpectedly. There are currently many circumstances in which events are simply discarded without recourse for users of Vector to recover those events. This implements a framework for handling those discarded events in a uniform fashion in all components.
closes #1772
closes #12217
closes #13549
|
1.0
|
Allow rerouting discarded events - **This is a tracking issue for RFC #12217 (PR #14708).**
We want to give Vector users the assurance that events are never lost unexpectedly. There are currently many circumstances in which events are simply discarded without recourse for users of Vector to recover those events. This implements a framework for handling those discarded events in a uniform fashion in all components.
closes #1772
closes #12217
closes #13549
|
process
|
allow rerouting discarded events this is a tracking issue for rfc pr we want to give vector users the assurance that events are never lost unexpectedly there are currently many circumstances in which events are simply discarded without recourse for users of vector to recover those events this implements a framework for handling those discarded events in a uniform fashion in all components closes closes closes
| 1
|
31,017
| 4,677,499,409
|
IssuesEvent
|
2016-10-07 15:16:04
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
opened
|
W should create a library for saturating the cluster that will be shared between density test and scheduler benchmark.
|
area/performance area/test
|
This way we'll be able to easily compare accuracy of the benchmark and make reduce the code duplication.
cc @wojtek-t @hongchaodeng
|
1.0
|
W should create a library for saturating the cluster that will be shared between density test and scheduler benchmark. - This way we'll be able to easily compare accuracy of the benchmark and make reduce the code duplication.
cc @wojtek-t @hongchaodeng
|
non_process
|
w should create a library for saturating the cluster that will be shared between density test and scheduler benchmark this way we ll be able to easily compare accuracy of the benchmark and make reduce the code duplication cc wojtek t hongchaodeng
| 0
|
448,903
| 31,817,627,948
|
IssuesEvent
|
2023-09-13 21:56:33
|
Xaptured/thejackfolio_dashboard
|
https://api.github.com/repos/Xaptured/thejackfolio_dashboard
|
opened
|
Create dummy object
|
documentation enhancement
|
Create dummy object which will render on page when getting error from backend with message
|
1.0
|
Create dummy object - Create dummy object which will render on page when getting error from backend with message
|
non_process
|
create dummy object create dummy object which will render on page when getting error from backend with message
| 0
|
605,971
| 18,752,753,997
|
IssuesEvent
|
2021-11-05 05:58:58
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
gadgets.ndtv.com - see bug description
|
priority-important browser-fenix engine-gecko
|
<!-- @browser: Firefox Mobile 95.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:95.0) Gecko/95.0 Firefox/95.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/89772 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://gadgets.ndtv.com/moto-g5s-4301
**Browser / Version**: Firefox Mobile 95.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: This site just crashes Firefox on Android
**Steps to Reproduce**:
The site does not load properly and Firefox just crashes
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/10/94e7a732-b1fa-42e6-8faa-ca244a52b4e9.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20211007091917</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/10/25588d1f-bc70-4ce7-abd9-362bc0115d31)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
gadgets.ndtv.com - see bug description - <!-- @browser: Firefox Mobile 95.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:95.0) Gecko/95.0 Firefox/95.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/89772 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://gadgets.ndtv.com/moto-g5s-4301
**Browser / Version**: Firefox Mobile 95.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: This site just crashes Firefox on Android
**Steps to Reproduce**:
The site does not load properly and Firefox just crashes
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/10/94e7a732-b1fa-42e6-8faa-ca244a52b4e9.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20211007091917</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/10/25588d1f-bc70-4ce7-abd9-362bc0115d31)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
gadgets ndtv com see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description this site just crashes firefox on android steps to reproduce the site does not load properly and firefox just crashes view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
44,760
| 11,499,631,095
|
IssuesEvent
|
2020-02-12 14:16:41
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Windows: Integration tests crash with native exception
|
broken-build electron electron-7-update important
|
Seeing this one on Windows just now: https://github.com/microsoft/vscode/runs/385552317 and https://github.com/microsoft/vscode/runs/388575039#step:13:306
```
v8_inspector::V8ContextInfo::executionContextId [0x00007FF64EF6310C+2715148]
v8::Unwinder::PCIsInV8 [0x00007FF64E7BB8A3+1379]
v8::Unwinder::PCIsInV8 [0x00007FF64E7BB466+294]
v8::Function::Call [0x00007FF64E5EBEF4+500]
node::InternalCallbackScope::Close [0x00007FF650B6ABF9+1001]
node::MakeCallback [0x00007FF650B6AFA5+213]
node::MakeCallback [0x00007FF650B6AE9E+158]
node::MakeCallback [0x00007FF650B6ADCA+138]
v8::ArrayBuffer::Contents::Deleter [0x00007FF64D6B778F+159535]
uv_backend_fd [0x00007FF64D6362BF+174959]
uv_backend_fd [0x00007FF64D636124+174548]
uv_backend_fd [0x00007FF64D626B22+111570]
uv_backend_fd [0x00007FF64D626A1C+111308]
uv_mutex_init_recursive [0x00007FF64F23E3A5+1075765]
GetHandleVerifier [0x00007FF64F4E91C3+631091]
GetHandleVerifier [0x00007FF64F4ED0EF+647263]
GetHandleVerifier [0x00007FF64F4ECD27+646295]
GetHandleVerifier [0x00007FF64F4E7872+624610]
GetHandleVerifier [0x00007FF64F4E804E+626622]
GetHandleVerifier [0x00007FF64F4F73A0+688912]
uv_cond_signal [0x00007FF64F3EC7C2+11506]
uv_cond_signal [0x00007FF64F3FCB85+78005]
uv_cond_signal [0x00007FF64F3FC80C+77116]
uv_loadavg [0x00007FF64F3ABBE4+761796]
uv_loadavg [0x00007FF64F3AB64E+760366]
uv_cond_signal [0x00007FF64F3FD4B6+80358]
uv_os_getpid [0x00007FF64F3D1B6E+68222]
v8::CpuProfile::GetTopDownRoot [0x00007FF64E00A260+1871536]
v8::CpuProfile::GetTopDownRoot [0x00007FF64E00A131+1871233]
v8::CpuProfile::GetTopDownRoot [0x00007FF64E00B851+1877153]
v8::CpuProfile::GetTopDownRoot [0x00007FF64E007702+1860434]
v8::CpuProfile::GetTopDownRoot [0x00007FF64DF10D76+850374]
v8::CpuProfile::GetTopDownRoot [0x00007FF64DF11777+852935]
v8::CpuProfile::GetTopDownRoot [0x00007FF64DF1148B+852187]
IsSandboxedProcess [0x00007FF64FC33A1A+234362]
v8::CpuProfile::GetTopDownRoot [0x00007FF64DF10CE6+850230]
Ordinal0 [0x00007FF64D5B13DC+5084]
node::AsyncResource::CallbackScope::CallbackScope [0x00007FF652390532+6778626]
BaseThreadInitThunk [0x00007FF9E17C84D4+20]
RtlUserThreadStart [0x00007FF9E193E8B1+33]
```
|
1.0
|
Windows: Integration tests crash with native exception - Seeing this one on Windows just now: https://github.com/microsoft/vscode/runs/385552317 and https://github.com/microsoft/vscode/runs/388575039#step:13:306
```
v8_inspector::V8ContextInfo::executionContextId [0x00007FF64EF6310C+2715148]
v8::Unwinder::PCIsInV8 [0x00007FF64E7BB8A3+1379]
v8::Unwinder::PCIsInV8 [0x00007FF64E7BB466+294]
v8::Function::Call [0x00007FF64E5EBEF4+500]
node::InternalCallbackScope::Close [0x00007FF650B6ABF9+1001]
node::MakeCallback [0x00007FF650B6AFA5+213]
node::MakeCallback [0x00007FF650B6AE9E+158]
node::MakeCallback [0x00007FF650B6ADCA+138]
v8::ArrayBuffer::Contents::Deleter [0x00007FF64D6B778F+159535]
uv_backend_fd [0x00007FF64D6362BF+174959]
uv_backend_fd [0x00007FF64D636124+174548]
uv_backend_fd [0x00007FF64D626B22+111570]
uv_backend_fd [0x00007FF64D626A1C+111308]
uv_mutex_init_recursive [0x00007FF64F23E3A5+1075765]
GetHandleVerifier [0x00007FF64F4E91C3+631091]
GetHandleVerifier [0x00007FF64F4ED0EF+647263]
GetHandleVerifier [0x00007FF64F4ECD27+646295]
GetHandleVerifier [0x00007FF64F4E7872+624610]
GetHandleVerifier [0x00007FF64F4E804E+626622]
GetHandleVerifier [0x00007FF64F4F73A0+688912]
uv_cond_signal [0x00007FF64F3EC7C2+11506]
uv_cond_signal [0x00007FF64F3FCB85+78005]
uv_cond_signal [0x00007FF64F3FC80C+77116]
uv_loadavg [0x00007FF64F3ABBE4+761796]
uv_loadavg [0x00007FF64F3AB64E+760366]
uv_cond_signal [0x00007FF64F3FD4B6+80358]
uv_os_getpid [0x00007FF64F3D1B6E+68222]
v8::CpuProfile::GetTopDownRoot [0x00007FF64E00A260+1871536]
v8::CpuProfile::GetTopDownRoot [0x00007FF64E00A131+1871233]
v8::CpuProfile::GetTopDownRoot [0x00007FF64E00B851+1877153]
v8::CpuProfile::GetTopDownRoot [0x00007FF64E007702+1860434]
v8::CpuProfile::GetTopDownRoot [0x00007FF64DF10D76+850374]
v8::CpuProfile::GetTopDownRoot [0x00007FF64DF11777+852935]
v8::CpuProfile::GetTopDownRoot [0x00007FF64DF1148B+852187]
IsSandboxedProcess [0x00007FF64FC33A1A+234362]
v8::CpuProfile::GetTopDownRoot [0x00007FF64DF10CE6+850230]
Ordinal0 [0x00007FF64D5B13DC+5084]
node::AsyncResource::CallbackScope::CallbackScope [0x00007FF652390532+6778626]
BaseThreadInitThunk [0x00007FF9E17C84D4+20]
RtlUserThreadStart [0x00007FF9E193E8B1+33]
```
|
non_process
|
windows integration tests crash with native exception seeing this one on windows just now and inspector executioncontextid unwinder unwinder function call node internalcallbackscope close node makecallback node makecallback node makecallback arraybuffer contents deleter uv backend fd uv backend fd uv backend fd uv backend fd uv mutex init recursive gethandleverifier gethandleverifier gethandleverifier gethandleverifier gethandleverifier gethandleverifier uv cond signal uv cond signal uv cond signal uv loadavg uv loadavg uv cond signal uv os getpid cpuprofile gettopdownroot cpuprofile gettopdownroot cpuprofile gettopdownroot cpuprofile gettopdownroot cpuprofile gettopdownroot cpuprofile gettopdownroot cpuprofile gettopdownroot issandboxedprocess cpuprofile gettopdownroot node asyncresource callbackscope callbackscope basethreadinitthunk rtluserthreadstart
| 0
|
18,802
| 24,702,828,409
|
IssuesEvent
|
2022-10-19 16:31:49
|
unicode-org/icu4x
|
https://api.github.com/repos/unicode-org/icu4x
|
opened
|
WASM demo takes 40 seconds to load and is not interactive
|
T-enhancement good first issue help wanted C-process S-medium
|
During the 40 seconds it takes to download the latest full-data WASM demo, the page is not interactive.
Short-term solutions:
1. Go back to using testdata
2. Add a loading spiral that goes away when WASM is finished loading (this would take about 10 lines of code)
3. Print an error upon input if the WASM is not loaded
Medium-term solution: download the locale data on demand. Steps to make that work:
1. The wasm-demo datagen should generate a directory of language-specific postcard files
2. A data provider should be hooked up that informs JS when root data is being used
3. JS code that `fetch`es the requested language pack when that callback occurs, and then re-tries the rendering
|
1.0
|
WASM demo takes 40 seconds to load and is not interactive - During the 40 seconds it takes to download the latest full-data WASM demo, the page is not interactive.
Short-term solutions:
1. Go back to using testdata
2. Add a loading spiral that goes away when WASM is finished loading (this would take about 10 lines of code)
3. Print an error upon input if the WASM is not loaded
Medium-term solution: download the locale data on demand. Steps to make that work:
1. The wasm-demo datagen should generate a directory of language-specific postcard files
2. A data provider should be hooked up that informs JS when root data is being used
3. JS code that `fetch`es the requested language pack when that callback occurs, and then re-tries the rendering
|
process
|
wasm demo takes seconds to load and is not interactive during the seconds it takes to download the latest full data wasm demo the page is not interactive short term solutions go back to using testdata add a loading spiral that goes away when wasm is finished loading this would take about lines of code print an error upon input if the wasm is not loaded medium term solution download the locale data on demand steps to make that work the wasm demo datagen should generate a directory of language specific postcard files a data provider should be hooked up that informs js when root data is being used js code that fetch es the requested language pack when that callback occurs and then re tries the rendering
| 1
|
21,134
| 28,106,561,771
|
IssuesEvent
|
2023-03-31 01:32:36
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Nested output symlinks behavior for RE
|
P3 type: process team-Remote-Exec stale
|
Let's discuss cornercase issues with symbolic links as outputs here, especially pertaining to remote execution. Examples:
1. Action declares outputs foo/bar and foo/. Expected: remote execution returns both the output directory digest and the output file digest. The file is not downloaded twice (as part of the output directory).
2. Action declares only bar output, but actually creates bar as a symlink to foo/bar. Expected behavior: remote execution returns a dangling symlink as output, because foo was not declared, therefore not uploaded to the CAS or returned.
3. Similarly, if action declared both foo/ and bar->foo/bar, then everything should work -- both the directory and the symlink pointing inside it should be returned.
4. Action declares outputs foo/bar and baz/. baz/ is a directory, and foo is a symlink pointing to baz. In the current API, there is no way for RE to tell Bazel that foo is actually a symlink. So the current behavior would be for the action to return foo/bar as a regular file, and the directory foo will then be created on the client as a regular directory, copying the bar file, which will be different than the local execution of the same action. If we think this edge-case is important enough to amend the API for, we can discuss options to fix it.
|
1.0
|
Nested output symlinks behavior for RE - Let's discuss cornercase issues with symbolic links as outputs here, especially pertaining to remote execution. Examples:
1. Action declares outputs foo/bar and foo/. Expected: remote execution returns both the output directory digest and the output file digest. The file is not downloaded twice (as part of the output directory).
2. Action declares only bar output, but actually creates bar as a symlink to foo/bar. Expected behavior: remote execution returns a dangling symlink as output, because foo was not declared, therefore not uploaded to the CAS or returned.
3. Similarly, if action declared both foo/ and bar->foo/bar, then everything should work -- both the directory and the symlink pointing inside it should be returned.
4. Action declares outputs foo/bar and baz/. baz/ is a directory, and foo is a symlink pointing to baz. In the current API, there is no way for RE to tell Bazel that foo is actually a symlink. So the current behavior would be for the action to return foo/bar as a regular file, and the directory foo will then be created on the client as a regular directory, copying the bar file, which will be different than the local execution of the same action. If we think this edge-case is important enough to amend the API for, we can discuss options to fix it.
|
process
|
nested output symlinks behavior for re let s discuss cornercase issues with symbolic links as outputs here especially pertaining to remote execution examples action declares outputs foo bar and foo expected remote execution returns both the output directory digest and the output file digest the file is not downloaded twice as part of the output directory action declares only bar output but actually creates bar as a symlink to foo bar expected behavior remote execution returns a dangling symlink as output because foo was not declared therefore not uploaded to the cas or returned similarly if action declared both foo and bar foo bar then everything should work both the directory and the symlink pointing inside it should be returned action declares outputs foo bar and baz baz is a directory and foo is a symlink pointing to baz in the current api there is no way for re to tell bazel that foo is actually a symlink so the current behavior would be for the action to return foo bar as a regular file and the directory foo will then be created on the client as a regular directory copying the bar file which will be different than the local execution of the same action if we think this edge case is important enough to amend the api for we can discuss options to fix it
| 1
|
9,802
| 12,814,801,401
|
IssuesEvent
|
2020-07-04 21:08:39
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Batch processing interface in 3.14 adds layer ids to the output file names
|
Bug Processing Regression
|
There is a regression in 3.14 where the batch processing interface adds layer ids to the output filenames. When autofilling the output with input layer parameter values, the file name contains layer ids. Which is unexpected and users should not see layer ids in their filenames.
<img width="993" alt="batch1" src="https://user-images.githubusercontent.com/5227506/86445462-08169900-bd30-11ea-95dd-151fb5a9eb36.png">
<img width="986" alt="batch2" src="https://user-images.githubusercontent.com/5227506/86445466-0baa2000-bd30-11ea-923b-47c55eef70d6.png">
Here's a geopackage with the source layers.
[basic_network_analysis.gpkg.zip](https://github.com/qgis/QGIS/files/4868459/basic_network_analysis.gpkg.zip)
**QGIS and OS versions**
<!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here -->
QGIS version | 3.14.0-Pi | QGIS code revision | 9f7028fd23
-- | -- | -- | --
Compiled against Qt | 5.12.3 | Running against Qt | 5.12.3
Compiled against GDAL/OGR | 2.4.1 | Running against GDAL/OGR | 2.4.1
Compiled against GEOS | 3.7.2-CAPI-1.11.2 | Running against GEOS | 3.7.2-CAPI-1.11.2 b55d2125
Compiled against SQLite | 3.28.0 | Running against SQLite | 3.28.0
PostgreSQL Client Version | 11.3 | SpatiaLite Version | 4.3.0a
QWT Version | 6.1.4 | QScintilla2 Version | 2.11.1
Compiled against PROJ | 5.2.0 | Running against PROJ | Rel. 5.2.0, September 15th, 2018
OS Version | macOS Mojave (10.14)
|
1.0
|
Batch processing interface in 3.14 adds layer ids to the output file names - There is a regression in 3.14 where the batch processing interface adds layer ids to the output filenames. When autofilling the output with input layer parameter values, the file name contains layer ids. Which is unexpected and users should not see layer ids in their filenames.
<img width="993" alt="batch1" src="https://user-images.githubusercontent.com/5227506/86445462-08169900-bd30-11ea-95dd-151fb5a9eb36.png">
<img width="986" alt="batch2" src="https://user-images.githubusercontent.com/5227506/86445466-0baa2000-bd30-11ea-923b-47c55eef70d6.png">
Here's a geopackage with the source layers.
[basic_network_analysis.gpkg.zip](https://github.com/qgis/QGIS/files/4868459/basic_network_analysis.gpkg.zip)
**QGIS and OS versions**
<!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here -->
QGIS version | 3.14.0-Pi | QGIS code revision | 9f7028fd23
-- | -- | -- | --
Compiled against Qt | 5.12.3 | Running against Qt | 5.12.3
Compiled against GDAL/OGR | 2.4.1 | Running against GDAL/OGR | 2.4.1
Compiled against GEOS | 3.7.2-CAPI-1.11.2 | Running against GEOS | 3.7.2-CAPI-1.11.2 b55d2125
Compiled against SQLite | 3.28.0 | Running against SQLite | 3.28.0
PostgreSQL Client Version | 11.3 | SpatiaLite Version | 4.3.0a
QWT Version | 6.1.4 | QScintilla2 Version | 2.11.1
Compiled against PROJ | 5.2.0 | Running against PROJ | Rel. 5.2.0, September 15th, 2018
OS Version | macOS Mojave (10.14)
|
process
|
batch processing interface in adds layer ids to the output file names there is a regression in where the batch processing interface adds layer ids to the output filenames when autofilling the output with input layer parameter values the file name contains layer ids which is unexpected and users should not see layer ids in their filenames img width alt src img width alt src here s a geopackage with the source layers qgis and os versions about click in the table ctrl a and then ctrl c finally paste here qgis version pi qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel september os version macos mojave
| 1
|
12,853
| 15,238,755,434
|
IssuesEvent
|
2021-02-19 02:43:05
|
GoogleCloudPlatform/cloud-sql-python-connector
|
https://api.github.com/repos/GoogleCloudPlatform/cloud-sql-python-connector
|
closed
|
Changing comment style from Google to Sphinx
|
priority: p1 type: process
|
The purpose of this issue is to track the change from Google-style comments to Sphinx-style comments.
|
1.0
|
Changing comment style from Google to Sphinx - The purpose of this issue is to track the change from Google-style comments to Sphinx-style comments.
|
process
|
changing comment style from google to sphinx the purpose of this issue is to track the change from google style comments to sphinx style comments
| 1
|
12,965
| 15,341,959,665
|
IssuesEvent
|
2021-02-27 14:18:21
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
I recently did the latest update now my automated test runs throw errors.
|
stage: needs information topic: preprocessors :wrench:
|
I recently did the latest update now my automated test runs throw errors.

Kalebmccaw@aviamedia.com
Thank You,
Kaleb Mccaw
|
1.0
|
I recently did the latest update now my automated test runs throw errors. - I recently did the latest update now my automated test runs throw errors.

Kalebmccaw@aviamedia.com
Thank You,
Kaleb Mccaw
|
process
|
i recently did the latest update now my automated test runs throw errors i recently did the latest update now my automated test runs throw errors kalebmccaw aviamedia com thank you kaleb mccaw
| 1
|
3,175
| 6,230,329,898
|
IssuesEvent
|
2017-07-11 07:42:21
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
opened
|
Improve definition of neurological system and nervous system
|
organism-level process
|
This was brought up in #13824
Uberon has 'neurological system' as an exact synonym for 'nervous system'.
http://www.ontobee.org/ontology/UBERON?iri=http://purl.obolibrary.org/obo/UBERON_0001016
@dosumis writes (by email)
> I just got the impression from the asserted child classes that something broader than what is usually understood by nervous system was intended. Might be worth investigating further - e.g. are the anatomical structures the child terms refer to part of (or overlapping) the nervous system according Uberon?
|
1.0
|
Improve definition of neurological system and nervous system - This was brought up in #13824
Uberon has 'neurological system' as an exact synonym for 'nervous system'.
http://www.ontobee.org/ontology/UBERON?iri=http://purl.obolibrary.org/obo/UBERON_0001016
@dosumis writes (by email)
> I just got the impression from the asserted child classes that something broader than what is usually understood by nervous system was intended. Might be worth investigating further - e.g. are the anatomical structures the child terms refer to part of (or overlapping) the nervous system according Uberon?
|
process
|
improve definition of neurological system and nervous system this was brought up in uberon has neurological system as an exact synonym for nervous system dosumis writes by email i just got the impression from the asserted child classes that something broader than what is usually understood by nervous system was intended might be worth investigating further e g are the anatomical structures the child terms refer to part of or overlapping the nervous system according uberon
| 1
|
22,403
| 6,245,970,991
|
IssuesEvent
|
2017-07-13 01:53:38
|
xceedsoftware/wpftoolkit
|
https://api.github.com/repos/xceedsoftware/wpftoolkit
|
closed
|
Message Box hangs second time.
|
CodePlex
|
<b>kamlendra[CodePlex]</b> <br />i am saving Employee data and updating it on click of save button again. First time when i click save it works Fine. Second time when i try to save it hangs on the MessageBox.
try
{
if(!string.IsNullOrEmpty(message))
Xceed.Wpf.Toolkit.MessageBox.Show(message);
}
catch(Exception ex )
{
nbsp
}
nbsp
it is not even going to the catch. here if i use Windows Message Box that works fine. earlier to this i was using same(WPF Extendedtoolkit) messageBox with different Name space i never observed this issue. may be with new namespace(dll) it is a regresion issue.
|
1.0
|
Message Box hangs second time. - <b>kamlendra[CodePlex]</b> <br />i am saving Employee data and updating it on click of save button again. First time when i click save it works Fine. Second time when i try to save it hangs on the MessageBox.
try
{
if(!string.IsNullOrEmpty(message))
Xceed.Wpf.Toolkit.MessageBox.Show(message);
}
catch(Exception ex )
{
nbsp
}
nbsp
it is not even going to the catch. here if i use Windows Message Box that works fine. earlier to this i was using same(WPF Extendedtoolkit) messageBox with different Name space i never observed this issue. may be with new namespace(dll) it is a regresion issue.
|
non_process
|
message box hangs second time kamlendra i am saving employee data and updating it on click of save button again first time when i click save it works fine second time when i try to save it hangs on the messagebox try if string isnullorempty message xceed wpf toolkit messagebox show message catch exception ex nbsp nbsp it is not even going to the catch here if i use windows message box that works fine earlier to this i was using same wpf extendedtoolkit messagebox with different name space i never observed this issue may be with new namespace dll it is a regresion issue
| 0
|
13,073
| 15,414,762,030
|
IssuesEvent
|
2021-03-05 01:01:14
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Process.Start() is stuck on Cent OS 7 and Ubuntu 18.04
|
area-System.Diagnostics.Process needs author feedback no recent activity
|
# Issue Title
Process.Start() is stuck sometimes on Cent OS 7 and Ubuntu 18.04
# General
Recently we updated Net core version used by [Azure pipelines agent](https://github.com/microsoft/azure-pipelines-agent) to 3.1.0, and it seems that for some customers issue appeared - Process.Start("chmod 600 <file in agent folder>") is stuck for some reason (on [line](https://github.com/microsoft/azure-pipelines-agent/blob/master/src/Agent.Sdk/ProcessInvoker.cs#L285))
At the same moment calling this command from terminal works fine for customer. Agent process is being run under the root, nothing else changed in code for this call.
Here are Process.StartInfo parameters for call:
File name: '/bin/chmod'
Arguments: '600 <file in agent folder>'
Working directory: '/root/azagent'
UseShellExecute: false
CreateNoWindow: true
RedirectStandardInput: true;
RedirectStandardError: true;
RedirectStandardOutput: true
OS version: Cent OS 7 (centos-release-7-6.1810.2.el7.centos.x86_64), Ubuntu 18.04
Net core version: 3.1.0
According to logs it seems that chmod process is in zombie state after the call, and there's no return value from Process.Start() method:
https://github.com/microsoft/azure-pipelines-agent/issues/3093#issuecomment-721630338
You can also find more info in thread for [issue](https://github.com/microsoft/azure-pipelines-agent/issues/3093).
Could you please take a look? Do you know if there were any similar issues - if so, is there any possible workaround/is it fixed?
Please let me know if you need some additional info.
|
1.0
|
Process.Start() is stuck on Cent OS 7 and Ubuntu 18.04 - # Issue Title
Process.Start() is stuck sometimes on Cent OS 7 and Ubuntu 18.04
# General
Recently we updated Net core version used by [Azure pipelines agent](https://github.com/microsoft/azure-pipelines-agent) to 3.1.0, and it seems that for some customers issue appeared - Process.Start("chmod 600 <file in agent folder>") is stuck for some reason (on [line](https://github.com/microsoft/azure-pipelines-agent/blob/master/src/Agent.Sdk/ProcessInvoker.cs#L285))
At the same moment calling this command from terminal works fine for customer. Agent process is being run under the root, nothing else changed in code for this call.
Here are Process.StartInfo parameters for call:
File name: '/bin/chmod'
Arguments: '600 <file in agent folder>'
Working directory: '/root/azagent'
UseShellExecute: false
CreateNoWindow: true
RedirectStandardInput: true;
RedirectStandardError: true;
RedirectStandardOutput: true
OS version: Cent OS 7 (centos-release-7-6.1810.2.el7.centos.x86_64), Ubuntu 18.04
Net core version: 3.1.0
According to logs it seems that chmod process is in zombie state after the call, and there's no return value from Process.Start() method:
https://github.com/microsoft/azure-pipelines-agent/issues/3093#issuecomment-721630338
You can also find more info in thread for [issue](https://github.com/microsoft/azure-pipelines-agent/issues/3093).
Could you please take a look? Do you know if there were any similar issues - if so, is there any possible workaround/is it fixed?
Please let me know if you need some additional info.
|
process
|
process start is stuck on cent os and ubuntu issue title process start is stuck sometimes on cent os and ubuntu general recently we updated net core version used by to and it seems that for some customers issue appeared process start chmod is stuck for some reason on at the same moment calling this command from terminal works fine for customer agent process is being run under the root nothing else changed in code for this call here are process startinfo parameters for call file name bin chmod arguments working directory root azagent useshellexecute false createnowindow true redirectstandardinput true redirectstandarderror true redirectstandardoutput true os version cent os centos release centos ubuntu net core version according to logs it seems that chmod process is in zombie state after the call and there s no return value from process start method you can also find more info in thread for could you please take a look do you know if there were any similar issues if so is there any possible workaround is it fixed please let me know if you need some additional info
| 1
|
1,042
| 3,511,059,513
|
IssuesEvent
|
2016-01-09 23:55:29
|
sysown/proxysql
|
https://api.github.com/repos/sysown/proxysql
|
opened
|
Use the Admin module as a backend
|
ADMIN AUTHENTICATION CLUSTER CONNECTION POOL GLOBAL MYSQL PROTOCOL QUERY PROCESSOR
|
Another requirement for ProxySQL Cluster.
MySQL_Thread should be able to communicate to Admin modules acting as backends.
MySQL_Monitor should be able to also monitor these backends.
Issue #463 is related to this one
|
1.0
|
Use the Admin module as a backend - Another requirement for ProxySQL Cluster.
MySQL_Thread should be able to communicate to Admin modules acting as backends.
MySQL_Monitor should be able to also monitor these backends.
Issue #463 is related to this one
|
process
|
use the admin module as a backend another requirement for proxysql cluster mysql thread should be able to communicate to admin modules acting as backends mysql monitor should be able to also monitor these backends issue is related to this one
| 1
|
16,051
| 20,194,341,644
|
IssuesEvent
|
2022-02-11 09:16:32
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
QgsProcessingParameterFile extension option
|
Processing Bug
|
### What is the bug or the crash?
When setting an input file with the QgsProcessingParameterFile in a plugin, for Linux if the selected file is a binary file it will not have an extension. When the plugin is executed in QGIS 3.20 on Debian 11.1, this will generate a "Wrong or missing parameter value: " error. If the executable file name altered to add an extension, such as .bin, it runs fine. This error does not occur on Debian 10.11 with QGIS 3.18 or in earlier versions of QGIS this problem does not occur.
I discovered this in using the QMarxan plugin I created, however I believe the error to be some undocumented change between 3.18 and 3.20 in the QGIS Python API.
### Steps to reproduce the issue
See above
### Versions
3.20
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
I see two possible options to resolve this problem:
1. If it is possible to set the file extension or filefilter parameter to accept files without extensions, then this needs to be documented in a clear way.
2. The default behaviour should be as was found in the 3.18 and earlier versions where files without extensions were accepted.
Note: I did not try a new QGIS profile as it is not relevant in this case.
|
1.0
|
QgsProcessingParameterFile extension option - ### What is the bug or the crash?
When setting an input file with the QgsProcessingParameterFile in a plugin, for Linux if the selected file is a binary file it will not have an extension. When the plugin is executed in QGIS 3.20 on Debian 11.1, this will generate a "Wrong or missing parameter value: " error. If the executable file name altered to add an extension, such as .bin, it runs fine. This error does not occur on Debian 10.11 with QGIS 3.18 or in earlier versions of QGIS this problem does not occur.
I discovered this in using the QMarxan plugin I created, however I believe the error to be some undocumented change between 3.18 and 3.20 in the QGIS Python API.
### Steps to reproduce the issue
See above
### Versions
3.20
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
I see two possible options to resolve this problem:
1. If it is possible to set the file extension or filefilter parameter to accept files without extensions, then this needs to be documented in a clear way.
2. The default behaviour should be as was found in the 3.18 and earlier versions where files without extensions were accepted.
Note: I did not try a new QGIS profile as it is not relevant in this case.
|
process
|
qgsprocessingparameterfile extension option what is the bug or the crash when setting an input file with the qgsprocessingparameterfile in a plugin for linux if the selected file is a binary file it will not have an extension when the plugin is executed in qgis on debian this will generate a wrong or missing parameter value error if the executable file name altered to add an extension such as bin it runs fine this error does not occur on debian with qgis or in earlier versions of qgis this problem does not occur i discovered this in using the qmarxan plugin i created however i believe the error to be some undocumented change between and in the qgis python api steps to reproduce the issue see above versions supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context i see two possible options to resolve this problem if it is possible to set the file extension or filefilter parameter to accept files without extensions then this needs to be documented in a clear way the default behaviour should be as was found in the and earlier versions where files without extensions were accepted note i did not try a new qgis profile as it is not relevant in this case
| 1
|
461,489
| 13,231,110,161
|
IssuesEvent
|
2020-08-18 11:02:44
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.ultreiadenver.com - The "ONLINE WAITLIST" button is not working
|
browser-fenix engine-gecko priority-normal severity-important
|
<!-- @browser: Firefox Mobile 70.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:70.0) Gecko/70.0 Firefox/70.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: http://www.ultreiadenver.com/
**Browser / Version**: Firefox Mobile 70.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: online wait-list button doesn't work
**Steps to Reproduce**:
Scroll down to see the button
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.ultreiadenver.com - The "ONLINE WAITLIST" button is not working - <!-- @browser: Firefox Mobile 70.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:70.0) Gecko/70.0 Firefox/70.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: http://www.ultreiadenver.com/
**Browser / Version**: Firefox Mobile 70.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: online wait-list button doesn't work
**Steps to Reproduce**:
Scroll down to see the button
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
the online waitlist button is not working url browser version firefox mobile operating system android tested another browser no problem type site is not usable description online wait list button doesn t work steps to reproduce scroll down to see the button browser configuration none from with ❤️
| 0
|
17,409
| 23,224,470,960
|
IssuesEvent
|
2022-08-02 21:51:13
|
vectordotdev/vector
|
https://api.github.com/repos/vectordotdev/vector
|
closed
|
Address processing order in docs
|
type: task have: nice domain: external docs domain: processing
|
We don't discuss processing order in the docs anywhere, which is necessary for some pipelines.
|
1.0
|
Address processing order in docs - We don't discuss processing order in the docs anywhere, which is necessary for some pipelines.
|
process
|
address processing order in docs we don t discuss processing order in the docs anywhere which is necessary for some pipelines
| 1
|
15,247
| 19,183,459,312
|
IssuesEvent
|
2021-12-04 20:12:54
|
ethereum/EIPs
|
https://api.github.com/repos/ethereum/EIPs
|
closed
|
Patent covenant for EIP submissions
|
type: Meta type: EIP1 (Process) stale
|
In several discussions about the EIP process, a concern has been raised about patent coverage.
Specifically, while the text of the EIP is contributed by the authors under CC0, this does not speak to whether or not they hold patents against it.
Do we need to add patent covenant language to the EIP process? Can we pay for a review and assistance from lawyers to help with this?
Please use the [EthMagicians forum](https://ethereum-magicians.org/t/patent-covenant-for-eip-submissions/2901) for long form discussion of this item.
|
1.0
|
Patent covenant for EIP submissions - In several discussions about the EIP process, a concern has been raised about patent coverage.
Specifically, while the text of the EIP is contributed by the authors under CC0, this does not speak to whether or not they hold patents against it.
Do we need to add patent covenant language to the EIP process? Can we pay for a review and assistance from lawyers to help with this?
Please use the [EthMagicians forum](https://ethereum-magicians.org/t/patent-covenant-for-eip-submissions/2901) for long form discussion of this item.
|
process
|
patent covenant for eip submissions in several discussions about the eip process a concern has been raised about patent coverage specifically while the text of the eip is contributed by the authors under this does not speak to whether or not they hold patents against it do we need to add patent covenant language to the eip process can we pay for a review and assistance from lawyers to help with this please use the for long form discussion of this item
| 1
|
7,462
| 10,562,889,974
|
IssuesEvent
|
2019-10-04 19:29:29
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
closed
|
BigQuery: bucket creation fails with 429 in systest 'test_extract_table'.
|
api: bigquery flaky testing type: process
|
Similar to #5746, #6096.
See [yesterday's failure](https://source.cloud.google.com/results/invocations/530881fa-f34a-45dc-a062-615dbabe7f67/targets/cloud-devrel%2Fclient-libraries%2Fgoogle-cloud-python%2Fpresubmit%2Fbigquery/log).
```python
_____________ ERROR at teardown of TestBigQuery.test_extract_table _____________
self = <tests.system.TestBigQuery testMethod=test_extract_table>
def tearDown(self):
def _still_in_use(bad_request):
return any(
error["reason"] == "resourceInUse" for error in bad_request._errors
)
retry_in_use = RetryErrors(BadRequest, error_predicate=_still_in_use)
retry_storage_errors_conflict = RetryErrors(
(Conflict, TooManyRequests, InternalServerError, ServiceUnavailable)
)
for doomed in self.to_delete:
if isinstance(doomed, storage.Bucket):
retry_storage_errors_conflict(doomed.delete)(force=True)
elif isinstance(doomed, (Dataset, bigquery.DatasetReference)):
retry_in_use(Config.CLIENT.delete_dataset)(doomed, delete_contents=True)
elif isinstance(doomed, (Table, bigquery.TableReference)):
> retry_in_use(Config.CLIENT.delete_table)(doomed)
tests/system.py:173:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
../test_utils/test_utils/retry.py:95: in wrapped_function
return to_wrap(*args, **kwargs)
google/cloud/bigquery/client.py:627: in delete_table
self._call_api(retry, method="DELETE", path=table.path)
google/cloud/bigquery/client.py:381: in _call_api
return call()
../api_core/google/api_core/retry.py:270: in retry_wrapped_func
on_error=on_error,
../api_core/google/api_core/retry.py:179: in retry_target
return target()
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.cloud.bigquery._http.Connection object at 0x7f7bd6389050>
method = 'DELETE'
path = '/projects/precise-truck-742/datasets/load_gcs_then_extract_1549494433207/tables/test_table'
query_params = None, data = None, content_type = None, headers = None
api_base_url = None, api_version = None, expect_json = True
_target_object = None
def api_request(
self,
method,
path,
query_params=None,
data=None,
content_type=None,
headers=None,
api_base_url=None,
api_version=None,
expect_json=True,
_target_object=None,
):
"""Make a request over the HTTP transport to the API.
You shouldn't need to use this method, but if you plan to
interact with the API using these primitives, this is the
correct one to use.
:type method: str
:param method: The HTTP method name (ie, ``GET``, ``POST``, etc).
Required.
:type path: str
:param path: The path to the resource (ie, ``'/b/bucket-name'``).
Required.
:type query_params: dict or list
:param query_params: A dictionary of keys and values (or list of
key-value pairs) to insert into the query
string of the URL.
:type data: str
:param data: The data to send as the body of the request. Default is
the empty string.
:type content_type: str
:param content_type: The proper MIME type of the data provided. Default
is None.
:type headers: dict
:param headers: extra HTTP headers to be sent with the request.
:type api_base_url: str
:param api_base_url: The base URL for the API endpoint.
Typically you won't have to provide this.
Default is the standard API base URL.
:type api_version: str
:param api_version: The version of the API to call. Typically
you shouldn't provide this and instead use
the default for the library. Default is the
latest API version supported by
google-cloud-python.
:type expect_json: bool
:param expect_json: If True, this method will try to parse the
response as JSON and raise an exception if
that cannot be done. Default is True.
:type _target_object: :class:`object`
:param _target_object:
(Optional) Protected argument to be used by library callers. This
can allow custom behavior, for example, to defer an HTTP request
and complete initialization of the object at a later time.
:raises ~google.cloud.exceptions.GoogleCloudError: if the response code
is not 200 OK.
:raises ValueError: if the response content type is not JSON.
:rtype: dict or str
:returns: The API response payload, either as a raw string or
a dictionary if the response is valid JSON.
"""
url = self.build_api_url(
path=path,
query_params=query_params,
api_base_url=api_base_url,
api_version=api_version,
)
# Making the executive decision that any dictionary
# data will be sent properly as JSON.
if data and isinstance(data, dict):
data = json.dumps(data)
content_type = "application/json"
response = self._make_request(
method=method,
url=url,
data=data,
content_type=content_type,
headers=headers,
target_object=_target_object,
)
if not 200 <= response.status_code < 300:
> raise exceptions.from_http_response(response)
E NotFound: 404 DELETE https://www.googleapis.com/bigquery/v2/projects/precise-truck-742/datasets/load_gcs_then_extract_1549494433207/tables/test_table: Not found: Dataset precise-truck-742:load_gcs_then_extract_1549494433207
../core/google/cloud/_http.py:319: NotFound
----------------------------- Captured stdout call -----------------------------
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 1 seconds...
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 2 seconds...
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 4 seconds...
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 8 seconds...
=================================== FAILURES ===================================
_______________________ TestBigQuery.test_extract_table ________________________
self = <tests.system.TestBigQuery testMethod=test_extract_table>
def test_extract_table(self):
local_id = unique_resource_id()
bucket_name = "bq_extract_test" + local_id
source_blob_name = "person_ages.csv"
dataset_id = _make_dataset_id("load_gcs_then_extract")
table_id = "test_table"
table_ref = Config.CLIENT.dataset(dataset_id).table(table_id)
table = Table(table_ref)
self.to_delete.insert(0, table)
> bucket = self._create_bucket(bucket_name)
tests/system.py:852:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
tests/system.py:188: in _create_bucket
retry_storage_errors(bucket.create)(location=location)
../test_utils/test_utils/retry.py:108: in wrapped_function
return to_wrap(*args, **kwargs)
../storage/google/cloud/storage/bucket.py:585: in create
_target_object=self,
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.cloud.storage._http.Connection object at 0x7f7bd5800bd0>
method = 'POST', path = '/b', query_params = {'project': 'precise-truck-742'}
data = '{"name": "bq_extract_test_1549494433207"}'
content_type = 'application/json', headers = None, api_base_url = None
api_version = None, expect_json = True
_target_object = <Bucket: bq_extract_test_1549494433207>
def api_request(
self,
method,
path,
query_params=None,
data=None,
content_type=None,
headers=None,
api_base_url=None,
api_version=None,
expect_json=True,
_target_object=None,
):
"""Make a request over the HTTP transport to the API.
You shouldn't need to use this method, but if you plan to
interact with the API using these primitives, this is the
correct one to use.
:type method: str
:param method: The HTTP method name (ie, ``GET``, ``POST``, etc).
Required.
:type path: str
:param path: The path to the resource (ie, ``'/b/bucket-name'``).
Required.
:type query_params: dict or list
:param query_params: A dictionary of keys and values (or list of
key-value pairs) to insert into the query
string of the URL.
:type data: str
:param data: The data to send as the body of the request. Default is
the empty string.
:type content_type: str
:param content_type: The proper MIME type of the data provided. Default
is None.
:type headers: dict
:param headers: extra HTTP headers to be sent with the request.
:type api_base_url: str
:param api_base_url: The base URL for the API endpoint.
Typically you won't have to provide this.
Default is the standard API base URL.
:type api_version: str
:param api_version: The version of the API to call. Typically
you shouldn't provide this and instead use
the default for the library. Default is the
latest API version supported by
google-cloud-python.
:type expect_json: bool
:param expect_json: If True, this method will try to parse the
response as JSON and raise an exception if
that cannot be done. Default is True.
:type _target_object: :class:`object`
:param _target_object:
(Optional) Protected argument to be used by library callers. This
can allow custom behavior, for example, to defer an HTTP request
and complete initialization of the object at a later time.
:raises ~google.cloud.exceptions.GoogleCloudError: if the response code
is not 200 OK.
:raises ValueError: if the response content type is not JSON.
:rtype: dict or str
:returns: The API response payload, either as a raw string or
a dictionary if the response is valid JSON.
"""
url = self.build_api_url(
path=path,
query_params=query_params,
api_base_url=api_base_url,
api_version=api_version,
)
# Making the executive decision that any dictionary
# data will be sent properly as JSON.
if data and isinstance(data, dict):
data = json.dumps(data)
content_type = "application/json"
response = self._make_request(
method=method,
url=url,
data=data,
content_type=content_type,
headers=headers,
target_object=_target_object,
)
if not 200 <= response.status_code < 300:
> raise exceptions.from_http_response(response)
E TooManyRequests: 429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets.
../core/google/cloud/_http.py:319: TooManyRequests
----------------------------- Captured stdout call -----------------------------
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 1 seconds...
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 2 seconds...
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 4 seconds...
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 8 seconds...
_____________ TestBigQuery.test_load_avro_from_uri_then_dump_table _____________
self = <tests.system.TestBigQuery testMethod=test_load_avro_from_uri_then_dump_table>
def tearDown(self):
def _still_in_use(bad_request):
return any(
error["reason"] == "resourceInUse" for error in bad_request._errors
)
retry_in_use = RetryErrors(BadRequest, error_predicate=_still_in_use)
retry_storage_errors_conflict = RetryErrors(
(Conflict, TooManyRequests, InternalServerError, ServiceUnavailable)
)
for doomed in self.to_delete:
if isinstance(doomed, storage.Bucket):
> retry_storage_errors_conflict(doomed.delete)(force=True)
tests/system.py:169:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
../test_utils/test_utils/retry.py:108: in wrapped_function
return to_wrap(*args, **kwargs)
../storage/google/cloud/storage/bucket.py:881: in delete
_target_object=None,
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.cloud.storage._http.Connection object at 0x7f7bd5768e10>
method = 'DELETE', path = '/b/bq_load_test_1549494478676', query_params = {}
data = None, content_type = None, headers = None, api_base_url = None
api_version = None, expect_json = True, _target_object = None
def api_request(
self,
method,
path,
query_params=None,
data=None,
content_type=None,
headers=None,
api_base_url=None,
api_version=None,
expect_json=True,
_target_object=None,
):
"""Make a request over the HTTP transport to the API.
You shouldn't need to use this method, but if you plan to
interact with the API using these primitives, this is the
correct one to use.
:type method: str
:param method: The HTTP method name (ie, ``GET``, ``POST``, etc).
Required.
:type path: str
:param path: The path to the resource (ie, ``'/b/bucket-name'``).
Required.
:type query_params: dict or list
:param query_params: A dictionary of keys and values (or list of
key-value pairs) to insert into the query
string of the URL.
:type data: str
:param data: The data to send as the body of the request. Default is
the empty string.
:type content_type: str
:param content_type: The proper MIME type of the data provided. Default
is None.
:type headers: dict
:param headers: extra HTTP headers to be sent with the request.
:type api_base_url: str
:param api_base_url: The base URL for the API endpoint.
Typically you won't have to provide this.
Default is the standard API base URL.
:type api_version: str
:param api_version: The version of the API to call. Typically
you shouldn't provide this and instead use
the default for the library. Default is the
latest API version supported by
google-cloud-python.
:type expect_json: bool
:param expect_json: If True, this method will try to parse the
response as JSON and raise an exception if
that cannot be done. Default is True.
:type _target_object: :class:`object`
:param _target_object:
(Optional) Protected argument to be used by library callers. This
can allow custom behavior, for example, to defer an HTTP request
and complete initialization of the object at a later time.
:raises ~google.cloud.exceptions.GoogleCloudError: if the response code
is not 200 OK.
:raises ValueError: if the response content type is not JSON.
:rtype: dict or str
:returns: The API response payload, either as a raw string or
a dictionary if the response is valid JSON.
"""
url = self.build_api_url(
path=path,
query_params=query_params,
api_base_url=api_base_url,
api_version=api_version,
)
# Making the executive decision that any dictionary
# data will be sent properly as JSON.
if data and isinstance(data, dict):
data = json.dumps(data)
content_type = "application/json"
response = self._make_request(
method=method,
url=url,
data=data,
content_type=content_type,
headers=headers,
target_object=_target_object,
)
if not 200 <= response.status_code < 300:
> raise exceptions.from_http_response(response)
E TooManyRequests: 429 DELETE https://www.googleapis.com/storage/v1/b/bq_load_test_1549494478676: The project exceeded the rate limit for creating and deleting buckets.
../core/google/cloud/_http.py:319: TooManyRequests
```
|
1.0
|
BigQuery: bucket creation fails with 429 in systest 'test_extract_table'. - Similar to #5746, #6096.
See [yesterday's failure](https://source.cloud.google.com/results/invocations/530881fa-f34a-45dc-a062-615dbabe7f67/targets/cloud-devrel%2Fclient-libraries%2Fgoogle-cloud-python%2Fpresubmit%2Fbigquery/log).
```python
_____________ ERROR at teardown of TestBigQuery.test_extract_table _____________
self = <tests.system.TestBigQuery testMethod=test_extract_table>
def tearDown(self):
def _still_in_use(bad_request):
return any(
error["reason"] == "resourceInUse" for error in bad_request._errors
)
retry_in_use = RetryErrors(BadRequest, error_predicate=_still_in_use)
retry_storage_errors_conflict = RetryErrors(
(Conflict, TooManyRequests, InternalServerError, ServiceUnavailable)
)
for doomed in self.to_delete:
if isinstance(doomed, storage.Bucket):
retry_storage_errors_conflict(doomed.delete)(force=True)
elif isinstance(doomed, (Dataset, bigquery.DatasetReference)):
retry_in_use(Config.CLIENT.delete_dataset)(doomed, delete_contents=True)
elif isinstance(doomed, (Table, bigquery.TableReference)):
> retry_in_use(Config.CLIENT.delete_table)(doomed)
tests/system.py:173:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
../test_utils/test_utils/retry.py:95: in wrapped_function
return to_wrap(*args, **kwargs)
google/cloud/bigquery/client.py:627: in delete_table
self._call_api(retry, method="DELETE", path=table.path)
google/cloud/bigquery/client.py:381: in _call_api
return call()
../api_core/google/api_core/retry.py:270: in retry_wrapped_func
on_error=on_error,
../api_core/google/api_core/retry.py:179: in retry_target
return target()
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.cloud.bigquery._http.Connection object at 0x7f7bd6389050>
method = 'DELETE'
path = '/projects/precise-truck-742/datasets/load_gcs_then_extract_1549494433207/tables/test_table'
query_params = None, data = None, content_type = None, headers = None
api_base_url = None, api_version = None, expect_json = True
_target_object = None
def api_request(
self,
method,
path,
query_params=None,
data=None,
content_type=None,
headers=None,
api_base_url=None,
api_version=None,
expect_json=True,
_target_object=None,
):
"""Make a request over the HTTP transport to the API.
You shouldn't need to use this method, but if you plan to
interact with the API using these primitives, this is the
correct one to use.
:type method: str
:param method: The HTTP method name (ie, ``GET``, ``POST``, etc).
Required.
:type path: str
:param path: The path to the resource (ie, ``'/b/bucket-name'``).
Required.
:type query_params: dict or list
:param query_params: A dictionary of keys and values (or list of
key-value pairs) to insert into the query
string of the URL.
:type data: str
:param data: The data to send as the body of the request. Default is
the empty string.
:type content_type: str
:param content_type: The proper MIME type of the data provided. Default
is None.
:type headers: dict
:param headers: extra HTTP headers to be sent with the request.
:type api_base_url: str
:param api_base_url: The base URL for the API endpoint.
Typically you won't have to provide this.
Default is the standard API base URL.
:type api_version: str
:param api_version: The version of the API to call. Typically
you shouldn't provide this and instead use
the default for the library. Default is the
latest API version supported by
google-cloud-python.
:type expect_json: bool
:param expect_json: If True, this method will try to parse the
response as JSON and raise an exception if
that cannot be done. Default is True.
:type _target_object: :class:`object`
:param _target_object:
(Optional) Protected argument to be used by library callers. This
can allow custom behavior, for example, to defer an HTTP request
and complete initialization of the object at a later time.
:raises ~google.cloud.exceptions.GoogleCloudError: if the response code
is not 200 OK.
:raises ValueError: if the response content type is not JSON.
:rtype: dict or str
:returns: The API response payload, either as a raw string or
a dictionary if the response is valid JSON.
"""
url = self.build_api_url(
path=path,
query_params=query_params,
api_base_url=api_base_url,
api_version=api_version,
)
# Making the executive decision that any dictionary
# data will be sent properly as JSON.
if data and isinstance(data, dict):
data = json.dumps(data)
content_type = "application/json"
response = self._make_request(
method=method,
url=url,
data=data,
content_type=content_type,
headers=headers,
target_object=_target_object,
)
if not 200 <= response.status_code < 300:
> raise exceptions.from_http_response(response)
E NotFound: 404 DELETE https://www.googleapis.com/bigquery/v2/projects/precise-truck-742/datasets/load_gcs_then_extract_1549494433207/tables/test_table: Not found: Dataset precise-truck-742:load_gcs_then_extract_1549494433207
../core/google/cloud/_http.py:319: NotFound
----------------------------- Captured stdout call -----------------------------
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 1 seconds...
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 2 seconds...
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 4 seconds...
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 8 seconds...
=================================== FAILURES ===================================
_______________________ TestBigQuery.test_extract_table ________________________
self = <tests.system.TestBigQuery testMethod=test_extract_table>
def test_extract_table(self):
local_id = unique_resource_id()
bucket_name = "bq_extract_test" + local_id
source_blob_name = "person_ages.csv"
dataset_id = _make_dataset_id("load_gcs_then_extract")
table_id = "test_table"
table_ref = Config.CLIENT.dataset(dataset_id).table(table_id)
table = Table(table_ref)
self.to_delete.insert(0, table)
> bucket = self._create_bucket(bucket_name)
tests/system.py:852:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
tests/system.py:188: in _create_bucket
retry_storage_errors(bucket.create)(location=location)
../test_utils/test_utils/retry.py:108: in wrapped_function
return to_wrap(*args, **kwargs)
../storage/google/cloud/storage/bucket.py:585: in create
_target_object=self,
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.cloud.storage._http.Connection object at 0x7f7bd5800bd0>
method = 'POST', path = '/b', query_params = {'project': 'precise-truck-742'}
data = '{"name": "bq_extract_test_1549494433207"}'
content_type = 'application/json', headers = None, api_base_url = None
api_version = None, expect_json = True
_target_object = <Bucket: bq_extract_test_1549494433207>
def api_request(
self,
method,
path,
query_params=None,
data=None,
content_type=None,
headers=None,
api_base_url=None,
api_version=None,
expect_json=True,
_target_object=None,
):
"""Make a request over the HTTP transport to the API.
You shouldn't need to use this method, but if you plan to
interact with the API using these primitives, this is the
correct one to use.
:type method: str
:param method: The HTTP method name (ie, ``GET``, ``POST``, etc).
Required.
:type path: str
:param path: The path to the resource (ie, ``'/b/bucket-name'``).
Required.
:type query_params: dict or list
:param query_params: A dictionary of keys and values (or list of
key-value pairs) to insert into the query
string of the URL.
:type data: str
:param data: The data to send as the body of the request. Default is
the empty string.
:type content_type: str
:param content_type: The proper MIME type of the data provided. Default
is None.
:type headers: dict
:param headers: extra HTTP headers to be sent with the request.
:type api_base_url: str
:param api_base_url: The base URL for the API endpoint.
Typically you won't have to provide this.
Default is the standard API base URL.
:type api_version: str
:param api_version: The version of the API to call. Typically
you shouldn't provide this and instead use
the default for the library. Default is the
latest API version supported by
google-cloud-python.
:type expect_json: bool
:param expect_json: If True, this method will try to parse the
response as JSON and raise an exception if
that cannot be done. Default is True.
:type _target_object: :class:`object`
:param _target_object:
(Optional) Protected argument to be used by library callers. This
can allow custom behavior, for example, to defer an HTTP request
and complete initialization of the object at a later time.
:raises ~google.cloud.exceptions.GoogleCloudError: if the response code
is not 200 OK.
:raises ValueError: if the response content type is not JSON.
:rtype: dict or str
:returns: The API response payload, either as a raw string or
a dictionary if the response is valid JSON.
"""
url = self.build_api_url(
path=path,
query_params=query_params,
api_base_url=api_base_url,
api_version=api_version,
)
# Making the executive decision that any dictionary
# data will be sent properly as JSON.
if data and isinstance(data, dict):
data = json.dumps(data)
content_type = "application/json"
response = self._make_request(
method=method,
url=url,
data=data,
content_type=content_type,
headers=headers,
target_object=_target_object,
)
if not 200 <= response.status_code < 300:
> raise exceptions.from_http_response(response)
E TooManyRequests: 429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets.
../core/google/cloud/_http.py:319: TooManyRequests
----------------------------- Captured stdout call -----------------------------
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 1 seconds...
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 2 seconds...
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 4 seconds...
429 POST https://www.googleapis.com/storage/v1/b?project=precise-truck-742: The project exceeded the rate limit for creating and deleting buckets., Trying again in 8 seconds...
_____________ TestBigQuery.test_load_avro_from_uri_then_dump_table _____________
self = <tests.system.TestBigQuery testMethod=test_load_avro_from_uri_then_dump_table>
def tearDown(self):
def _still_in_use(bad_request):
return any(
error["reason"] == "resourceInUse" for error in bad_request._errors
)
retry_in_use = RetryErrors(BadRequest, error_predicate=_still_in_use)
retry_storage_errors_conflict = RetryErrors(
(Conflict, TooManyRequests, InternalServerError, ServiceUnavailable)
)
for doomed in self.to_delete:
if isinstance(doomed, storage.Bucket):
> retry_storage_errors_conflict(doomed.delete)(force=True)
tests/system.py:169:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
../test_utils/test_utils/retry.py:108: in wrapped_function
return to_wrap(*args, **kwargs)
../storage/google/cloud/storage/bucket.py:881: in delete
_target_object=None,
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.cloud.storage._http.Connection object at 0x7f7bd5768e10>
method = 'DELETE', path = '/b/bq_load_test_1549494478676', query_params = {}
data = None, content_type = None, headers = None, api_base_url = None
api_version = None, expect_json = True, _target_object = None
def api_request(
self,
method,
path,
query_params=None,
data=None,
content_type=None,
headers=None,
api_base_url=None,
api_version=None,
expect_json=True,
_target_object=None,
):
"""Make a request over the HTTP transport to the API.
You shouldn't need to use this method, but if you plan to
interact with the API using these primitives, this is the
correct one to use.
:type method: str
:param method: The HTTP method name (ie, ``GET``, ``POST``, etc).
Required.
:type path: str
:param path: The path to the resource (ie, ``'/b/bucket-name'``).
Required.
:type query_params: dict or list
:param query_params: A dictionary of keys and values (or list of
key-value pairs) to insert into the query
string of the URL.
:type data: str
:param data: The data to send as the body of the request. Default is
the empty string.
:type content_type: str
:param content_type: The proper MIME type of the data provided. Default
is None.
:type headers: dict
:param headers: extra HTTP headers to be sent with the request.
:type api_base_url: str
:param api_base_url: The base URL for the API endpoint.
Typically you won't have to provide this.
Default is the standard API base URL.
:type api_version: str
:param api_version: The version of the API to call. Typically
you shouldn't provide this and instead use
the default for the library. Default is the
latest API version supported by
google-cloud-python.
:type expect_json: bool
:param expect_json: If True, this method will try to parse the
response as JSON and raise an exception if
that cannot be done. Default is True.
:type _target_object: :class:`object`
:param _target_object:
(Optional) Protected argument to be used by library callers. This
can allow custom behavior, for example, to defer an HTTP request
and complete initialization of the object at a later time.
:raises ~google.cloud.exceptions.GoogleCloudError: if the response code
is not 200 OK.
:raises ValueError: if the response content type is not JSON.
:rtype: dict or str
:returns: The API response payload, either as a raw string or
a dictionary if the response is valid JSON.
"""
url = self.build_api_url(
path=path,
query_params=query_params,
api_base_url=api_base_url,
api_version=api_version,
)
# Making the executive decision that any dictionary
# data will be sent properly as JSON.
if data and isinstance(data, dict):
data = json.dumps(data)
content_type = "application/json"
response = self._make_request(
method=method,
url=url,
data=data,
content_type=content_type,
headers=headers,
target_object=_target_object,
)
if not 200 <= response.status_code < 300:
> raise exceptions.from_http_response(response)
E TooManyRequests: 429 DELETE https://www.googleapis.com/storage/v1/b/bq_load_test_1549494478676: The project exceeded the rate limit for creating and deleting buckets.
../core/google/cloud/_http.py:319: TooManyRequests
```
|
process
|
bigquery bucket creation fails with in systest test extract table similar to see python error at teardown of testbigquery test extract table self def teardown self def still in use bad request return any error resourceinuse for error in bad request errors retry in use retryerrors badrequest error predicate still in use retry storage errors conflict retryerrors conflict toomanyrequests internalservererror serviceunavailable for doomed in self to delete if isinstance doomed storage bucket retry storage errors conflict doomed delete force true elif isinstance doomed dataset bigquery datasetreference retry in use config client delete dataset doomed delete contents true elif isinstance doomed table bigquery tablereference retry in use config client delete table doomed tests system py test utils test utils retry py in wrapped function return to wrap args kwargs google cloud bigquery client py in delete table self call api retry method delete path table path google cloud bigquery client py in call api return call api core google api core retry py in retry wrapped func on error on error api core google api core retry py in retry target return target self method delete path projects precise truck datasets load gcs then extract tables test table query params none data none content type none headers none api base url none api version none expect json true target object none def api request self method path query params none data none content type none headers none api base url none api version none expect json true target object none make a request over the http transport to the api you shouldn t need to use this method but if you plan to interact with the api using these primitives this is the correct one to use type method str param method the http method name ie get post etc required type path str param path the path to the resource ie b bucket name required type query params dict or list param query params a dictionary of keys and values or list of key value pairs to insert into the query string of the url type data str param data the data to send as the body of the request default is the empty string type content type str param content type the proper mime type of the data provided default is none type headers dict param headers extra http headers to be sent with the request type api base url str param api base url the base url for the api endpoint typically you won t have to provide this default is the standard api base url type api version str param api version the version of the api to call typically you shouldn t provide this and instead use the default for the library default is the latest api version supported by google cloud python type expect json bool param expect json if true this method will try to parse the response as json and raise an exception if that cannot be done default is true type target object class object param target object optional protected argument to be used by library callers this can allow custom behavior for example to defer an http request and complete initialization of the object at a later time raises google cloud exceptions googleclouderror if the response code is not ok raises valueerror if the response content type is not json rtype dict or str returns the api response payload either as a raw string or a dictionary if the response is valid json url self build api url path path query params query params api base url api base url api version api version making the executive decision that any dictionary data will be sent properly as json if data and isinstance data dict data json dumps data content type application json response self make request method method url url data data content type content type headers headers target object target object if not response status code raise exceptions from http response response e notfound delete not found dataset precise truck load gcs then extract core google cloud http py notfound captured stdout call post the project exceeded the rate limit for creating and deleting buckets trying again in seconds post the project exceeded the rate limit for creating and deleting buckets trying again in seconds post the project exceeded the rate limit for creating and deleting buckets trying again in seconds post the project exceeded the rate limit for creating and deleting buckets trying again in seconds failures testbigquery test extract table self def test extract table self local id unique resource id bucket name bq extract test local id source blob name person ages csv dataset id make dataset id load gcs then extract table id test table table ref config client dataset dataset id table table id table table table ref self to delete insert table bucket self create bucket bucket name tests system py tests system py in create bucket retry storage errors bucket create location location test utils test utils retry py in wrapped function return to wrap args kwargs storage google cloud storage bucket py in create target object self self method post path b query params project precise truck data name bq extract test content type application json headers none api base url none api version none expect json true target object def api request self method path query params none data none content type none headers none api base url none api version none expect json true target object none make a request over the http transport to the api you shouldn t need to use this method but if you plan to interact with the api using these primitives this is the correct one to use type method str param method the http method name ie get post etc required type path str param path the path to the resource ie b bucket name required type query params dict or list param query params a dictionary of keys and values or list of key value pairs to insert into the query string of the url type data str param data the data to send as the body of the request default is the empty string type content type str param content type the proper mime type of the data provided default is none type headers dict param headers extra http headers to be sent with the request type api base url str param api base url the base url for the api endpoint typically you won t have to provide this default is the standard api base url type api version str param api version the version of the api to call typically you shouldn t provide this and instead use the default for the library default is the latest api version supported by google cloud python type expect json bool param expect json if true this method will try to parse the response as json and raise an exception if that cannot be done default is true type target object class object param target object optional protected argument to be used by library callers this can allow custom behavior for example to defer an http request and complete initialization of the object at a later time raises google cloud exceptions googleclouderror if the response code is not ok raises valueerror if the response content type is not json rtype dict or str returns the api response payload either as a raw string or a dictionary if the response is valid json url self build api url path path query params query params api base url api base url api version api version making the executive decision that any dictionary data will be sent properly as json if data and isinstance data dict data json dumps data content type application json response self make request method method url url data data content type content type headers headers target object target object if not response status code raise exceptions from http response response e toomanyrequests post the project exceeded the rate limit for creating and deleting buckets core google cloud http py toomanyrequests captured stdout call post the project exceeded the rate limit for creating and deleting buckets trying again in seconds post the project exceeded the rate limit for creating and deleting buckets trying again in seconds post the project exceeded the rate limit for creating and deleting buckets trying again in seconds post the project exceeded the rate limit for creating and deleting buckets trying again in seconds testbigquery test load avro from uri then dump table self def teardown self def still in use bad request return any error resourceinuse for error in bad request errors retry in use retryerrors badrequest error predicate still in use retry storage errors conflict retryerrors conflict toomanyrequests internalservererror serviceunavailable for doomed in self to delete if isinstance doomed storage bucket retry storage errors conflict doomed delete force true tests system py test utils test utils retry py in wrapped function return to wrap args kwargs storage google cloud storage bucket py in delete target object none self method delete path b bq load test query params data none content type none headers none api base url none api version none expect json true target object none def api request self method path query params none data none content type none headers none api base url none api version none expect json true target object none make a request over the http transport to the api you shouldn t need to use this method but if you plan to interact with the api using these primitives this is the correct one to use type method str param method the http method name ie get post etc required type path str param path the path to the resource ie b bucket name required type query params dict or list param query params a dictionary of keys and values or list of key value pairs to insert into the query string of the url type data str param data the data to send as the body of the request default is the empty string type content type str param content type the proper mime type of the data provided default is none type headers dict param headers extra http headers to be sent with the request type api base url str param api base url the base url for the api endpoint typically you won t have to provide this default is the standard api base url type api version str param api version the version of the api to call typically you shouldn t provide this and instead use the default for the library default is the latest api version supported by google cloud python type expect json bool param expect json if true this method will try to parse the response as json and raise an exception if that cannot be done default is true type target object class object param target object optional protected argument to be used by library callers this can allow custom behavior for example to defer an http request and complete initialization of the object at a later time raises google cloud exceptions googleclouderror if the response code is not ok raises valueerror if the response content type is not json rtype dict or str returns the api response payload either as a raw string or a dictionary if the response is valid json url self build api url path path query params query params api base url api base url api version api version making the executive decision that any dictionary data will be sent properly as json if data and isinstance data dict data json dumps data content type application json response self make request method method url url data data content type content type headers headers target object target object if not response status code raise exceptions from http response response e toomanyrequests delete the project exceeded the rate limit for creating and deleting buckets core google cloud http py toomanyrequests
| 1
|
185,262
| 15,016,734,080
|
IssuesEvent
|
2021-02-01 09:57:48
|
ArchiveBox/ArchiveBox
|
https://api.github.com/repos/ArchiveBox/ArchiveBox
|
closed
|
Bugfix: docker-compose instructions create a sonic container that fails to start
|
bug changes: documentation complexity: low
|
#### Describe the bug
I followed the docker-compose instructions from the README. This is the result:
```
[root@Acheron archivebox]# docker-compose ps
Name Command State Ports
--------------------------------------------------------------------------------------------
archivebox_archivebox_1 dumb-init -- /app/bin/dock ... Up 0.0.0.0:8000->8000/tcp
archivebox_sonic_1 sonic -c /etc/sonic.cfg Exit 101
```
```
[root@Acheron archivebox]# docker-compose logs sonic
Attaching to archivebox_sonic_1
sonic_1 | thread 'main' panicked at 'cannot read config file: Os { code: 21, kind: Other, message: "Is a directory" }', src/config/reader.rs:24:14
sonic_1 | note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
sonic_1 | thread 'main' panicked at 'cannot read config file: Os { code: 21, kind: Other, message: "Is a directory" }', src/config/reader.rs:24:14
sonic_1 | note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
```
Search seems to work anyway.
I would expect one of:
a. `sonic` container is not created by default if it requires the user to manually create a config and is not necessary to run ArchiveBox
b. `config.cfg` is created for me by the init script, using the environment variable I set in the docker-compose file
c. `config.cfg` is not required by sonic (however, this is not the case: https://github.com/valeriansaliou/sonic/issues/197)
#### Steps to reproduce
From the README:
```
# create a new empty directory and initalize your collection (can be anywhere)
mkdir ~/archivebox && cd ~/archivebox
curl -O https://raw.githubusercontent.com/ArchiveBox/ArchiveBox/master/docker-compose.yml
docker-compose run archivebox init
docker-compose run archivebox --version
# start the webserver and open the UI (optional)
docker-compose run archivebox manage createsuperuser
docker-compose up -d
open http://127.0.0.1:8000
# you can also add links and manage your archive via the CLI:
docker-compose run archivebox add 'https://example.com'
docker-compose run archivebox status
docker-compose run archivebox help # to see more options
```
#### ArchiveBox version
```
[root@Acheron archivebox]# docker-compose run archivebox --version
Starting archivebox_sonic_1 ... done
Creating archivebox_archivebox_run ... done
ArchiveBox v0.5.3
Cpython Linux Linux-5.9.1-arch1-1-x86_64-with-glibc2.28 x86_64 (in Docker)
[i] Dependency versions:
√ ARCHIVEBOX_BINARY v0.5.3 valid /usr/local/bin/archivebox
√ PYTHON_BINARY v3.9.1 valid /usr/local/bin/python3.9
√ DJANGO_BINARY v3.1.3 valid /usr/local/lib/python3.9/site-packages/django/bin/django-admin.py
√ CURL_BINARY v7.64.0 valid /usr/bin/curl
√ WGET_BINARY v1.20.1 valid /usr/bin/wget
√ NODE_BINARY v15.5.1 valid /usr/bin/node
√ SINGLEFILE_BINARY v0.1.14 valid /node/node_modules/single-file/cli/single-file
√ READABILITY_BINARY v0.1.0 valid /node/node_modules/readability-extractor/readability-extractor
√ MERCURY_BINARY v1.0.0 valid /node/node_modules/@postlight/mercury-parser/cli.js
√ GIT_BINARY v2.20.1 valid /usr/bin/git
√ YOUTUBEDL_BINARY v2021.01.03 valid /usr/local/bin/youtube-dl
√ CHROME_BINARY v87.0.4280.88 valid /usr/bin/chromium
√ RIPGREP_BINARY v0.10.0 valid /usr/bin/rg
[i] Source-code locations:
√ PACKAGE_DIR 22 files valid /app/archivebox
√ TEMPLATES_DIR 3 files valid /app/archivebox/themes
[i] Secrets locations:
- CHROME_USER_DATA_DIR - disabled
- COOKIES_FILE - disabled
[i] Data locations:
√ OUTPUT_DIR 6 files valid /data
√ SOURCES_DIR 1 files valid ./sources
√ LOGS_DIR 0 files valid ./logs
√ ARCHIVE_DIR 1 files valid ./archive
√ CONFIG_FILE 81.0 Bytes valid ./ArchiveBox.conf
√ SQL_INDEX 204.0 KB valid ./index.sqlite3
```
```
[root@Acheron archivebox]# docker version
Client:
Version: 20.10.2
API version: 1.40
Go version: go1.15.6
Git commit: 2291f610ae
Built: Tue Jan 5 19:56:21 2021
OS/Arch: linux/amd64
Context: default
Experimental: true
Server:
Engine:
Version: 19.03.13-ce
API version: 1.40 (minimum version 1.12)
Go version: go1.15.2
Git commit: 4484c46d9d
Built: Sat Sep 26 12:03:35 2020
OS/Arch: linux/amd64
Experimental: false
containerd:
Version: v1.4.1.m
GitCommit: c623d1b36f09f8ef6536a057bd658b3aa8632828.m
runc:
Version: 1.0.0-rc92
GitCommit: ff819c7e9184c13b7c2607fe6c30ae19403a7aff
docker-init:
Version: 0.19.0
GitCommit: de40ad0
```
```
[root@Acheron archivebox]# docker-compose version
docker-compose version 1.27.4, build 40524192
docker-py version: 4.3.1
CPython version: 3.7.7
OpenSSL version: OpenSSL 1.1.0l 10 Sep 2019
```
|
1.0
|
Bugfix: docker-compose instructions create a sonic container that fails to start - #### Describe the bug
I followed the docker-compose instructions from the README. This is the result:
```
[root@Acheron archivebox]# docker-compose ps
Name Command State Ports
--------------------------------------------------------------------------------------------
archivebox_archivebox_1 dumb-init -- /app/bin/dock ... Up 0.0.0.0:8000->8000/tcp
archivebox_sonic_1 sonic -c /etc/sonic.cfg Exit 101
```
```
[root@Acheron archivebox]# docker-compose logs sonic
Attaching to archivebox_sonic_1
sonic_1 | thread 'main' panicked at 'cannot read config file: Os { code: 21, kind: Other, message: "Is a directory" }', src/config/reader.rs:24:14
sonic_1 | note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
sonic_1 | thread 'main' panicked at 'cannot read config file: Os { code: 21, kind: Other, message: "Is a directory" }', src/config/reader.rs:24:14
sonic_1 | note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
```
Search seems to work anyway.
I would expect one of:
a. `sonic` container is not created by default if it requires the user to manually create a config and is not necessary to run ArchiveBox
b. `config.cfg` is created for me by the init script, using the environment variable I set in the docker-compose file
c. `config.cfg` is not required by sonic (however, this is not the case: https://github.com/valeriansaliou/sonic/issues/197)
#### Steps to reproduce
From the README:
```
# create a new empty directory and initalize your collection (can be anywhere)
mkdir ~/archivebox && cd ~/archivebox
curl -O https://raw.githubusercontent.com/ArchiveBox/ArchiveBox/master/docker-compose.yml
docker-compose run archivebox init
docker-compose run archivebox --version
# start the webserver and open the UI (optional)
docker-compose run archivebox manage createsuperuser
docker-compose up -d
open http://127.0.0.1:8000
# you can also add links and manage your archive via the CLI:
docker-compose run archivebox add 'https://example.com'
docker-compose run archivebox status
docker-compose run archivebox help # to see more options
```
#### ArchiveBox version
```
[root@Acheron archivebox]# docker-compose run archivebox --version
Starting archivebox_sonic_1 ... done
Creating archivebox_archivebox_run ... done
ArchiveBox v0.5.3
Cpython Linux Linux-5.9.1-arch1-1-x86_64-with-glibc2.28 x86_64 (in Docker)
[i] Dependency versions:
√ ARCHIVEBOX_BINARY v0.5.3 valid /usr/local/bin/archivebox
√ PYTHON_BINARY v3.9.1 valid /usr/local/bin/python3.9
√ DJANGO_BINARY v3.1.3 valid /usr/local/lib/python3.9/site-packages/django/bin/django-admin.py
√ CURL_BINARY v7.64.0 valid /usr/bin/curl
√ WGET_BINARY v1.20.1 valid /usr/bin/wget
√ NODE_BINARY v15.5.1 valid /usr/bin/node
√ SINGLEFILE_BINARY v0.1.14 valid /node/node_modules/single-file/cli/single-file
√ READABILITY_BINARY v0.1.0 valid /node/node_modules/readability-extractor/readability-extractor
√ MERCURY_BINARY v1.0.0 valid /node/node_modules/@postlight/mercury-parser/cli.js
√ GIT_BINARY v2.20.1 valid /usr/bin/git
√ YOUTUBEDL_BINARY v2021.01.03 valid /usr/local/bin/youtube-dl
√ CHROME_BINARY v87.0.4280.88 valid /usr/bin/chromium
√ RIPGREP_BINARY v0.10.0 valid /usr/bin/rg
[i] Source-code locations:
√ PACKAGE_DIR 22 files valid /app/archivebox
√ TEMPLATES_DIR 3 files valid /app/archivebox/themes
[i] Secrets locations:
- CHROME_USER_DATA_DIR - disabled
- COOKIES_FILE - disabled
[i] Data locations:
√ OUTPUT_DIR 6 files valid /data
√ SOURCES_DIR 1 files valid ./sources
√ LOGS_DIR 0 files valid ./logs
√ ARCHIVE_DIR 1 files valid ./archive
√ CONFIG_FILE 81.0 Bytes valid ./ArchiveBox.conf
√ SQL_INDEX 204.0 KB valid ./index.sqlite3
```
```
[root@Acheron archivebox]# docker version
Client:
Version: 20.10.2
API version: 1.40
Go version: go1.15.6
Git commit: 2291f610ae
Built: Tue Jan 5 19:56:21 2021
OS/Arch: linux/amd64
Context: default
Experimental: true
Server:
Engine:
Version: 19.03.13-ce
API version: 1.40 (minimum version 1.12)
Go version: go1.15.2
Git commit: 4484c46d9d
Built: Sat Sep 26 12:03:35 2020
OS/Arch: linux/amd64
Experimental: false
containerd:
Version: v1.4.1.m
GitCommit: c623d1b36f09f8ef6536a057bd658b3aa8632828.m
runc:
Version: 1.0.0-rc92
GitCommit: ff819c7e9184c13b7c2607fe6c30ae19403a7aff
docker-init:
Version: 0.19.0
GitCommit: de40ad0
```
```
[root@Acheron archivebox]# docker-compose version
docker-compose version 1.27.4, build 40524192
docker-py version: 4.3.1
CPython version: 3.7.7
OpenSSL version: OpenSSL 1.1.0l 10 Sep 2019
```
|
non_process
|
bugfix docker compose instructions create a sonic container that fails to start describe the bug i followed the docker compose instructions from the readme this is the result docker compose ps name command state ports archivebox archivebox dumb init app bin dock up tcp archivebox sonic sonic c etc sonic cfg exit docker compose logs sonic attaching to archivebox sonic sonic thread main panicked at cannot read config file os code kind other message is a directory src config reader rs sonic note run with rust backtrace environment variable to display a backtrace sonic thread main panicked at cannot read config file os code kind other message is a directory src config reader rs sonic note run with rust backtrace environment variable to display a backtrace search seems to work anyway i would expect one of a sonic container is not created by default if it requires the user to manually create a config and is not necessary to run archivebox b config cfg is created for me by the init script using the environment variable i set in the docker compose file c config cfg is not required by sonic however this is not the case steps to reproduce from the readme create a new empty directory and initalize your collection can be anywhere mkdir archivebox cd archivebox curl o docker compose run archivebox init docker compose run archivebox version start the webserver and open the ui optional docker compose run archivebox manage createsuperuser docker compose up d open you can also add links and manage your archive via the cli docker compose run archivebox add docker compose run archivebox status docker compose run archivebox help to see more options archivebox version docker compose run archivebox version starting archivebox sonic done creating archivebox archivebox run done archivebox cpython linux linux with in docker dependency versions √ archivebox binary valid usr local bin archivebox √ python binary valid usr local bin √ django binary valid usr local lib site packages django bin django admin py √ curl binary valid usr bin curl √ wget binary valid usr bin wget √ node binary valid usr bin node √ singlefile binary valid node node modules single file cli single file √ readability binary valid node node modules readability extractor readability extractor √ mercury binary valid node node modules postlight mercury parser cli js √ git binary valid usr bin git √ youtubedl binary valid usr local bin youtube dl √ chrome binary valid usr bin chromium √ ripgrep binary valid usr bin rg source code locations √ package dir files valid app archivebox √ templates dir files valid app archivebox themes secrets locations chrome user data dir disabled cookies file disabled data locations √ output dir files valid data √ sources dir files valid sources √ logs dir files valid logs √ archive dir files valid archive √ config file bytes valid archivebox conf √ sql index kb valid index docker version client version api version go version git commit built tue jan os arch linux context default experimental true server engine version ce api version minimum version go version git commit built sat sep os arch linux experimental false containerd version m gitcommit m runc version gitcommit docker init version gitcommit docker compose version docker compose version build docker py version cpython version openssl version openssl sep
| 0
|
92,777
| 3,873,793,830
|
IssuesEvent
|
2016-04-11 18:13:45
|
CPLamb/RouteTracker
|
https://api.github.com/repos/CPLamb/RouteTracker
|
closed
|
Discuss architecture for writing of sheets back up to Google Drive
|
Priority - High
|
Decide on a simple architecture we can use to provide the **write** capability of read/write of the sheets.
API reference - https://developers.google.com/google-apps/spreadsheets/
https://developers.google.com/drive/ios/devguide/files
|
1.0
|
Discuss architecture for writing of sheets back up to Google Drive - Decide on a simple architecture we can use to provide the **write** capability of read/write of the sheets.
API reference - https://developers.google.com/google-apps/spreadsheets/
https://developers.google.com/drive/ios/devguide/files
|
non_process
|
discuss architecture for writing of sheets back up to google drive decide on a simple architecture we can use to provide the write capability of read write of the sheets api reference
| 0
|
12,582
| 14,991,210,586
|
IssuesEvent
|
2021-01-29 07:57:23
|
panther-labs/panther
|
https://api.github.com/repos/panther-labs/panther
|
opened
|
Fix rule error: AttributeError("'ImmutableList' object has no attribute 'get'")
|
bug p1 team:data processing
|
### Describe the bug
Rule `AWS.CloudTrail.ResourceMadePublic` is reporting the following error:
AttributeError("'ImmutableList' object has no attribute 'get'")
### Steps to reproduce
Steps to reproduce the behavior:
1. Go to a Panther deployment
2. Go to Rules page, find rule `AWS.CloudTrail.ResourceMadePublic`
3. Click edit on the rule
4. Click "Run all tests" and see error
### Expected behavior
The rule should't report an error
### Environment
How are you deploying or using Panther?
- Panther version or commit: 1.15.2
|
1.0
|
Fix rule error: AttributeError("'ImmutableList' object has no attribute 'get'") - ### Describe the bug
Rule `AWS.CloudTrail.ResourceMadePublic` is reporting the following error:
AttributeError("'ImmutableList' object has no attribute 'get'")
### Steps to reproduce
Steps to reproduce the behavior:
1. Go to a Panther deployment
2. Go to Rules page, find rule `AWS.CloudTrail.ResourceMadePublic`
3. Click edit on the rule
4. Click "Run all tests" and see error
### Expected behavior
The rule should't report an error
### Environment
How are you deploying or using Panther?
- Panther version or commit: 1.15.2
|
process
|
fix rule error attributeerror immutablelist object has no attribute get describe the bug rule aws cloudtrail resourcemadepublic is reporting the following error attributeerror immutablelist object has no attribute get steps to reproduce steps to reproduce the behavior go to a panther deployment go to rules page find rule aws cloudtrail resourcemadepublic click edit on the rule click run all tests and see error expected behavior the rule should t report an error environment how are you deploying or using panther panther version or commit
| 1
|
106,951
| 16,736,719,804
|
IssuesEvent
|
2021-06-11 03:14:06
|
atlslscsrv-app/upgraded-fiesta
|
https://api.github.com/repos/atlslscsrv-app/upgraded-fiesta
|
closed
|
CVE-2020-11023 (Medium) detected in jquery-1.11.1.js
|
no-issue-activity security vulnerability
|
## CVE-2020-11023 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.11.1.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.js</a></p>
<p>Path to dependency file: upgraded-fiesta/node_modules/unix-crypt-td-js/test/test.html</p>
<p>Path to vulnerable library: upgraded-fiesta/node_modules/unix-crypt-td-js/test/test.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.11.1.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/atlslscsrv-app/upgraded-fiesta/commit/ec68291ce25200691b7a48431969b7c8021b35e2">ec68291ce25200691b7a48431969b7c8021b35e2</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-11023 (Medium) detected in jquery-1.11.1.js - ## CVE-2020-11023 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.11.1.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.js</a></p>
<p>Path to dependency file: upgraded-fiesta/node_modules/unix-crypt-td-js/test/test.html</p>
<p>Path to vulnerable library: upgraded-fiesta/node_modules/unix-crypt-td-js/test/test.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.11.1.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/atlslscsrv-app/upgraded-fiesta/commit/ec68291ce25200691b7a48431969b7c8021b35e2">ec68291ce25200691b7a48431969b7c8021b35e2</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in jquery js cve medium severity vulnerability vulnerable library jquery js javascript library for dom operations library home page a href path to dependency file upgraded fiesta node modules unix crypt td js test test html path to vulnerable library upgraded fiesta node modules unix crypt td js test test html dependency hierarchy x jquery js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
| 0
|
20,152
| 26,702,766,397
|
IssuesEvent
|
2023-01-27 15:38:06
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
siRNA processing defintion
|
RNA processes textual definition
|
It seems that this term has been made narrower in scope.
The term processing refers to the production of a mature transcript. but the definition includes
"MAy also include amplification step of the siRNA by RNA-directed RNA polymerase."
This part of the process isn't processing.
|
1.0
|
siRNA processing defintion -
It seems that this term has been made narrower in scope.
The term processing refers to the production of a mature transcript. but the definition includes
"MAy also include amplification step of the siRNA by RNA-directed RNA polymerase."
This part of the process isn't processing.
|
process
|
sirna processing defintion it seems that this term has been made narrower in scope the term processing refers to the production of a mature transcript but the definition includes may also include amplification step of the sirna by rna directed rna polymerase this part of the process isn t processing
| 1
|
91,456
| 10,720,940,117
|
IssuesEvent
|
2019-10-26 21:29:20
|
dukeify/fake-jni
|
https://api.github.com/repos/dukeify/fake-jni
|
opened
|
Add an explicit example for registering function overloads via fake-jni
|
documentation high priority
|
```cpp
using namespace FakeJni;
struct NativeClass {
DEFINE_CLASS_NAME("com/example/NativeClass")
void func() {}
void func(JInt) {}
};
BEGIN_NATIVE_DESCRIPTOR(NativeClass)
{Constructor<NativeClass> {}},
{(void (*)())&NativeClass::func, "func"},
{(void (*)(JInt))&NativeClass::func, "func"}
END_NATIVE_DESCRIPTOR
```
|
1.0
|
Add an explicit example for registering function overloads via fake-jni - ```cpp
using namespace FakeJni;
struct NativeClass {
DEFINE_CLASS_NAME("com/example/NativeClass")
void func() {}
void func(JInt) {}
};
BEGIN_NATIVE_DESCRIPTOR(NativeClass)
{Constructor<NativeClass> {}},
{(void (*)())&NativeClass::func, "func"},
{(void (*)(JInt))&NativeClass::func, "func"}
END_NATIVE_DESCRIPTOR
```
|
non_process
|
add an explicit example for registering function overloads via fake jni cpp using namespace fakejni struct nativeclass define class name com example nativeclass void func void func jint begin native descriptor nativeclass constructor void nativeclass func func void jint nativeclass func func end native descriptor
| 0
|
4,924
| 7,795,442,815
|
IssuesEvent
|
2018-06-08 08:08:09
|
StrikeNP/trac_test
|
https://api.github.com/repos/StrikeNP/trac_test
|
closed
|
Implement .case files for CLUBB and SAM_CLUBB cases (Trac #138)
|
Migrated from Trac post_processing senkbeil@uwm.edu task
|
All the cases for CLUBB and SAM_CLUBB need to be added to Plotgen 3. To do this, we need to create .case files. Most .case files should be almost identical (case names will be different), so this should be as simple as copying the files.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/138
```json
{
"status": "closed",
"changetime": "2009-09-02T20:40:36",
"description": "All the cases for CLUBB and SAM_CLUBB need to be added to Plotgen 3. To do this, we need to create .case files. Most .case files should be almost identical (case names will be different), so this should be as simple as copying the files.",
"reporter": "senkbeil@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1251924036000000",
"component": "post_processing",
"summary": "Implement .case files for CLUBB and SAM_CLUBB cases",
"priority": "minor",
"keywords": "plotgen",
"time": "2009-07-21T23:16:58",
"milestone": "Plotgen 3.0",
"owner": "senkbeil@uwm.edu",
"type": "task"
}
```
|
1.0
|
Implement .case files for CLUBB and SAM_CLUBB cases (Trac #138) - All the cases for CLUBB and SAM_CLUBB need to be added to Plotgen 3. To do this, we need to create .case files. Most .case files should be almost identical (case names will be different), so this should be as simple as copying the files.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/138
```json
{
"status": "closed",
"changetime": "2009-09-02T20:40:36",
"description": "All the cases for CLUBB and SAM_CLUBB need to be added to Plotgen 3. To do this, we need to create .case files. Most .case files should be almost identical (case names will be different), so this should be as simple as copying the files.",
"reporter": "senkbeil@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1251924036000000",
"component": "post_processing",
"summary": "Implement .case files for CLUBB and SAM_CLUBB cases",
"priority": "minor",
"keywords": "plotgen",
"time": "2009-07-21T23:16:58",
"milestone": "Plotgen 3.0",
"owner": "senkbeil@uwm.edu",
"type": "task"
}
```
|
process
|
implement case files for clubb and sam clubb cases trac all the cases for clubb and sam clubb need to be added to plotgen to do this we need to create case files most case files should be almost identical case names will be different so this should be as simple as copying the files attachments migrated from json status closed changetime description all the cases for clubb and sam clubb need to be added to plotgen to do this we need to create case files most case files should be almost identical case names will be different so this should be as simple as copying the files reporter senkbeil uwm edu cc resolution verified by v larson ts component post processing summary implement case files for clubb and sam clubb cases priority minor keywords plotgen time milestone plotgen owner senkbeil uwm edu type task
| 1
|
138,813
| 12,830,087,291
|
IssuesEvent
|
2020-07-07 01:02:45
|
drashland/rhum
|
https://api.github.com/repos/drashland/rhum
|
opened
|
Video Tutorial: General
|
documentation
|
## Summary
What: Video tutorial (acts of visual documentation) for how to use Rhum, and how it works on certain level. It's open to whoever takes on this issue how they'd like to do the video and what content it will have. But there will be other issues to address hooks, and stubs/mocks
Could be useful to include in the description of the video, times of topics, eg:
0:00 - introduction
0:23 - what is sockets
5:10 - creating rooms
Why: Supports visual learners, and thus creates another platform for people to understand Rhum and how to use it - further enforcing "documentation driven"
## Information
* OBS can be used as the software to screenshare and video recording
* Includes the drashland introduction? (speak to @crookse )
* Includes the sockets logo with the greenish background as the thumbnail
* Video can be added to a "Rhum" playlist (can also be created if it doesn't exist)
* Make it chill :)
|
1.0
|
Video Tutorial: General - ## Summary
What: Video tutorial (acts of visual documentation) for how to use Rhum, and how it works on certain level. It's open to whoever takes on this issue how they'd like to do the video and what content it will have. But there will be other issues to address hooks, and stubs/mocks
Could be useful to include in the description of the video, times of topics, eg:
0:00 - introduction
0:23 - what is sockets
5:10 - creating rooms
Why: Supports visual learners, and thus creates another platform for people to understand Rhum and how to use it - further enforcing "documentation driven"
## Information
* OBS can be used as the software to screenshare and video recording
* Includes the drashland introduction? (speak to @crookse )
* Includes the sockets logo with the greenish background as the thumbnail
* Video can be added to a "Rhum" playlist (can also be created if it doesn't exist)
* Make it chill :)
|
non_process
|
video tutorial general summary what video tutorial acts of visual documentation for how to use rhum and how it works on certain level it s open to whoever takes on this issue how they d like to do the video and what content it will have but there will be other issues to address hooks and stubs mocks could be useful to include in the description of the video times of topics eg introduction what is sockets creating rooms why supports visual learners and thus creates another platform for people to understand rhum and how to use it further enforcing documentation driven information obs can be used as the software to screenshare and video recording includes the drashland introduction speak to crookse includes the sockets logo with the greenish background as the thumbnail video can be added to a rhum playlist can also be created if it doesn t exist make it chill
| 0
|
22,748
| 32,064,060,993
|
IssuesEvent
|
2023-09-25 00:17:12
|
solop-develop/frontend-core
|
https://api.github.com/repos/solop-develop/frontend-core
|
closed
|
[Feature Request] Mostrar Cantidad del Histórico Procesos
|
bug enhancement (PRC) Processes (RPT) Reports (UX) User Experience
|
## Feature request
Actualmente siempre muestra el numero encima de la Etiqueta de la Pestañas, dando la impresión al usuario que existen procesos con errores o en ejecución.
Debería solo mostrar cuando exista mas de un registro en el **Histórico Procesos**
### GIF

|
1.0
|
[Feature Request] Mostrar Cantidad del Histórico Procesos - ## Feature request
Actualmente siempre muestra el numero encima de la Etiqueta de la Pestañas, dando la impresión al usuario que existen procesos con errores o en ejecución.
Debería solo mostrar cuando exista mas de un registro en el **Histórico Procesos**
### GIF

|
process
|
mostrar cantidad del histórico procesos feature request actualmente siempre muestra el numero encima de la etiqueta de la pestañas dando la impresión al usuario que existen procesos con errores o en ejecución debería solo mostrar cuando exista mas de un registro en el histórico procesos gif
| 1
|
17,669
| 23,493,781,983
|
IssuesEvent
|
2022-08-17 21:42:22
|
alchemistry/alchemlyb
|
https://api.github.com/repos/alchemistry/alchemlyb
|
closed
|
Remove the alchemlyb.concat when the dependency for pandas raise to v1.4.0
|
preprocessors
|
WIth https://github.com/pandas-dev/pandas/pull/42252,
The pandas.concat is fixed from v1.4.0 (at the develop branch), so I placed the issue here to remind me to remove alchemlyb.concat when we raise the dependency of pandas to v1.4.0.
|
1.0
|
Remove the alchemlyb.concat when the dependency for pandas raise to v1.4.0 - WIth https://github.com/pandas-dev/pandas/pull/42252,
The pandas.concat is fixed from v1.4.0 (at the develop branch), so I placed the issue here to remind me to remove alchemlyb.concat when we raise the dependency of pandas to v1.4.0.
|
process
|
remove the alchemlyb concat when the dependency for pandas raise to with the pandas concat is fixed from at the develop branch so i placed the issue here to remind me to remove alchemlyb concat when we raise the dependency of pandas to
| 1
|
19,120
| 3,420,945,609
|
IssuesEvent
|
2015-12-08 16:45:03
|
lbroudoux/eip-designer
|
https://api.github.com/repos/lbroudoux/eip-designer
|
closed
|
Parse resource and persist to existing EIP model
|
designer feature
|
As we use parsers to build an in-memory model for comparing with original design, we'd like to allow this in-memory model to be merged and persist into an existing one. This will open the way to migration scenarios where you are using EIP tooling to migrate from one technology to another while retro-documenting existing resources.
|
1.0
|
Parse resource and persist to existing EIP model - As we use parsers to build an in-memory model for comparing with original design, we'd like to allow this in-memory model to be merged and persist into an existing one. This will open the way to migration scenarios where you are using EIP tooling to migrate from one technology to another while retro-documenting existing resources.
|
non_process
|
parse resource and persist to existing eip model as we use parsers to build an in memory model for comparing with original design we d like to allow this in memory model to be merged and persist into an existing one this will open the way to migration scenarios where you are using eip tooling to migrate from one technology to another while retro documenting existing resources
| 0
|
78,071
| 10,038,235,213
|
IssuesEvent
|
2019-07-18 14:46:15
|
edgexfoundry/edgex-go
|
https://api.github.com/repos/edgexfoundry/edgex-go
|
closed
|
core-metadata.raml causing edgexfoundry/edgex-docs build failure
|
bug documentation
|
# 📚 Docs or Wiki Bug Report
### Description
core-metadata.raml is consumed by the edgexfoundry/edgex-docs build pipeline. #1523 introduced defects into the raml file that cause the edgex-docs build to fail.
|
1.0
|
core-metadata.raml causing edgexfoundry/edgex-docs build failure - # 📚 Docs or Wiki Bug Report
### Description
core-metadata.raml is consumed by the edgexfoundry/edgex-docs build pipeline. #1523 introduced defects into the raml file that cause the edgex-docs build to fail.
|
non_process
|
core metadata raml causing edgexfoundry edgex docs build failure 📚 docs or wiki bug report description core metadata raml is consumed by the edgexfoundry edgex docs build pipeline introduced defects into the raml file that cause the edgex docs build to fail
| 0
|
8,137
| 11,339,771,328
|
IssuesEvent
|
2020-01-23 03:27:49
|
ryankeefe92/Episodes
|
https://api.github.com/repos/ryankeefe92/Episodes
|
closed
|
Add incoming 5.1 audio to already-added higher-video-resolution stereo file
|
feature process: question
|
if an incoming episode has 5.1 audio but a lower video resolution, and that same episode that has already been added to iTunes has just stereo audio but a higher video resolution than the incoming file, and most importantly, the episodes are the same length (within perhaps a 1 second difference at most…do some testing to see if this works in most situations, and the most time it can be different by without it being noticeably out of sync), have it take the 5.1 audio from the incoming file and add it to the file already in iTunes (to the file’s finder location [tell iTunes set variable to LOCATION of every track of playlist TV shows whose name contains myname])
|
1.0
|
Add incoming 5.1 audio to already-added higher-video-resolution stereo file - if an incoming episode has 5.1 audio but a lower video resolution, and that same episode that has already been added to iTunes has just stereo audio but a higher video resolution than the incoming file, and most importantly, the episodes are the same length (within perhaps a 1 second difference at most…do some testing to see if this works in most situations, and the most time it can be different by without it being noticeably out of sync), have it take the 5.1 audio from the incoming file and add it to the file already in iTunes (to the file’s finder location [tell iTunes set variable to LOCATION of every track of playlist TV shows whose name contains myname])
|
process
|
add incoming audio to already added higher video resolution stereo file if an incoming episode has audio but a lower video resolution and that same episode that has already been added to itunes has just stereo audio but a higher video resolution than the incoming file and most importantly the episodes are the same length within perhaps a second difference at most…do some testing to see if this works in most situations and the most time it can be different by without it being noticeably out of sync have it take the audio from the incoming file and add it to the file already in itunes to the file’s finder location
| 1
|
568,706
| 16,986,742,458
|
IssuesEvent
|
2021-06-30 15:09:55
|
WowRarity/Rarity
|
https://api.github.com/repos/WowRarity/Rarity
|
opened
|
Add collectibles that were introduced in patch 9.1
|
Complexity: TBD Module: Database Priority: Urgent Status: Accepted Type: Task
|
Source: [Discord](https://discord.com/channels/788119147740790854/788145559654760458/859809353233858575)
> Just wondering when support for the 9.1 new shinies will be available in the addon?
I don't have a lot of time right now, but maybe they didn't add a metric ton this time? Either way it would be a little easier adding them now that we have a more structured item database!
|
1.0
|
Add collectibles that were introduced in patch 9.1 - Source: [Discord](https://discord.com/channels/788119147740790854/788145559654760458/859809353233858575)
> Just wondering when support for the 9.1 new shinies will be available in the addon?
I don't have a lot of time right now, but maybe they didn't add a metric ton this time? Either way it would be a little easier adding them now that we have a more structured item database!
|
non_process
|
add collectibles that were introduced in patch source just wondering when support for the new shinies will be available in the addon i don t have a lot of time right now but maybe they didn t add a metric ton this time either way it would be a little easier adding them now that we have a more structured item database
| 0
|
297,290
| 22,348,126,060
|
IssuesEvent
|
2022-06-15 09:35:17
|
vercel/next.js
|
https://api.github.com/repos/vercel/next.js
|
closed
|
Docs: The example code is wrong.
|
template: documentation
|
### What is the improvement or update you wish to see?
(In the middle of this document)
This feature is called [Fast Refresh](https://nextjs.org/docs/basic-features/fast-refresh). It gives you instantaneous feedback on any edits you make and comes pre-configured with Next.js.
To recap, your code went from this..
```jsx
~~~ skip ~~~
<ul>
{name.map((name) => (
<li key={name}>{name}</li>
))}
</ul>
```
### Is there any context that might help us understand?
```jsx
~~~ skip ~~~
<ul>
{name.map((name) => (
<li key={name}>{name}</li>
))}
</ul>
```
should be **_names_**
```jsx
~~~ skip ~~~
<ul>
{names.map((name) => (
<li key={name}>{name}</li>
))}
</ul>
```
### Does the docs page already exist? Please link to it.
https://nextjs.org/learn/foundations/from-react-to-nextjs/getting-started-with-nextjs
|
1.0
|
Docs: The example code is wrong. - ### What is the improvement or update you wish to see?
(In the middle of this document)
This feature is called [Fast Refresh](https://nextjs.org/docs/basic-features/fast-refresh). It gives you instantaneous feedback on any edits you make and comes pre-configured with Next.js.
To recap, your code went from this..
```jsx
~~~ skip ~~~
<ul>
{name.map((name) => (
<li key={name}>{name}</li>
))}
</ul>
```
### Is there any context that might help us understand?
```jsx
~~~ skip ~~~
<ul>
{name.map((name) => (
<li key={name}>{name}</li>
))}
</ul>
```
should be **_names_**
```jsx
~~~ skip ~~~
<ul>
{names.map((name) => (
<li key={name}>{name}</li>
))}
</ul>
```
### Does the docs page already exist? Please link to it.
https://nextjs.org/learn/foundations/from-react-to-nextjs/getting-started-with-nextjs
|
non_process
|
docs the example code is wrong what is the improvement or update you wish to see in the middle of this document this feature is called it gives you instantaneous feedback on any edits you make and comes pre configured with next js to recap your code went from this jsx skip name map name name is there any context that might help us understand jsx skip name map name name should be names jsx skip names map name name does the docs page already exist please link to it
| 0
|
12,911
| 2,728,781,025
|
IssuesEvent
|
2015-04-16 01:13:39
|
codenameone/CodenameOne
|
https://api.github.com/repos/codenameone/CodenameOne
|
closed
|
Graphics performance on NewVM laggy.
|
Priority-Low Type-Defect
|
Original [issue 1327](https://code.google.com/p/codenameone/issues/detail?id=1327) created by codenameone on 2015-01-31T01:57:36.000Z:
Responsiveness on new vm is noticeably slower than on the old vm. I will post a simple sample app to demonstrate - right now I'm using some real-world apps and noticing this.
Steps:
1. A list with a custom renderer (just an image and text) with enough items to scroll ( at least 20 or so).
2. Scroll list up and down with index finger at varying speeds.
In old VM the list scroll is quite responsive, staying with the finger.
On new VM the list scroll is laggy, often times a full up-down movement behind.
This issue should lead to two activities:
1. Work on improving graphics pipeline efficiency.
2. Work on improving performance of VM in general.
|
1.0
|
Graphics performance on NewVM laggy. - Original [issue 1327](https://code.google.com/p/codenameone/issues/detail?id=1327) created by codenameone on 2015-01-31T01:57:36.000Z:
Responsiveness on new vm is noticeably slower than on the old vm. I will post a simple sample app to demonstrate - right now I'm using some real-world apps and noticing this.
Steps:
1. A list with a custom renderer (just an image and text) with enough items to scroll ( at least 20 or so).
2. Scroll list up and down with index finger at varying speeds.
In old VM the list scroll is quite responsive, staying with the finger.
On new VM the list scroll is laggy, often times a full up-down movement behind.
This issue should lead to two activities:
1. Work on improving graphics pipeline efficiency.
2. Work on improving performance of VM in general.
|
non_process
|
graphics performance on newvm laggy original created by codenameone on responsiveness on new vm is noticeably slower than on the old vm i will post a simple sample app to demonstrate right now i m using some real world apps and noticing this steps a list with a custom renderer just an image and text with enough items to scroll at least or so scroll list up and down with index finger at varying speeds in old vm the list scroll is quite responsive staying with the finger on new vm the list scroll is laggy often times a full up down movement behind this issue should lead to two activities work on improving graphics pipeline efficiency work on improving performance of vm in general
| 0
|
80,679
| 15,586,314,469
|
IssuesEvent
|
2021-03-18 01:39:51
|
soumya132/java-code
|
https://api.github.com/repos/soumya132/java-code
|
closed
|
CVE-2018-14720 (High) detected in jackson-databind-2.8.1.jar - autoclosed
|
security vulnerability
|
## CVE-2018-14720 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/java-code/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-jersey-1.4.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/soumya132/java-code/commit/bd6deafa717543d14242a80f30b2189c4dfe4f6c">bd6deafa717543d14242a80f30b2189c4dfe4f6c</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.7 might allow attackers to conduct external XML entity (XXE) attacks by leveraging failure to block unspecified JDK classes from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14720>CVE-2018-14720</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14720">https://nvd.nist.gov/vuln/detail/CVE-2018-14720</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-14720 (High) detected in jackson-databind-2.8.1.jar - autoclosed - ## CVE-2018-14720 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/java-code/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-jersey-1.4.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/soumya132/java-code/commit/bd6deafa717543d14242a80f30b2189c4dfe4f6c">bd6deafa717543d14242a80f30b2189c4dfe4f6c</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.7 might allow attackers to conduct external XML entity (XXE) attacks by leveraging failure to block unspecified JDK classes from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14720>CVE-2018-14720</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14720">https://nvd.nist.gov/vuln/detail/CVE-2018-14720</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm java code pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter jersey release jar root library x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before might allow attackers to conduct external xml entity xxe attacks by leveraging failure to block unspecified jdk classes from polymorphic deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
233,062
| 17,841,504,714
|
IssuesEvent
|
2021-09-03 10:43:05
|
scikit-learn-contrib/MAPIE
|
https://api.github.com/repos/scikit-learn-contrib/MAPIE
|
opened
|
[DOCUMENTATION] Fix errors in code examples in tutorial section of documentation
|
documentation
|
**Is your documentation request related to a problem? Please describe.**
I found this library interesting and was going through the library code examples where I found a code error in the plotting of interval widths for the section "Estimating the uncertainty with different sklearn-compatible regressors". It uses a variable y_preds which is not defined earlier anywhere in the code. Besides, a few import statements are missing from sklearn methods, I have made the changes. Should I send a PR for the same?
**Describe the solution you'd like**
The incorrect line of code can be fixed by using y_pis instead of undefined y_preds and using the correct dimensions for the data. The import statements for the missing sklearn methods can be added as well.
|
1.0
|
[DOCUMENTATION] Fix errors in code examples in tutorial section of documentation - **Is your documentation request related to a problem? Please describe.**
I found this library interesting and was going through the library code examples where I found a code error in the plotting of interval widths for the section "Estimating the uncertainty with different sklearn-compatible regressors". It uses a variable y_preds which is not defined earlier anywhere in the code. Besides, a few import statements are missing from sklearn methods, I have made the changes. Should I send a PR for the same?
**Describe the solution you'd like**
The incorrect line of code can be fixed by using y_pis instead of undefined y_preds and using the correct dimensions for the data. The import statements for the missing sklearn methods can be added as well.
|
non_process
|
fix errors in code examples in tutorial section of documentation is your documentation request related to a problem please describe i found this library interesting and was going through the library code examples where i found a code error in the plotting of interval widths for the section estimating the uncertainty with different sklearn compatible regressors it uses a variable y preds which is not defined earlier anywhere in the code besides a few import statements are missing from sklearn methods i have made the changes should i send a pr for the same describe the solution you d like the incorrect line of code can be fixed by using y pis instead of undefined y preds and using the correct dimensions for the data the import statements for the missing sklearn methods can be added as well
| 0
|
2,900
| 5,887,866,190
|
IssuesEvent
|
2017-05-17 08:41:37
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Has a problem with child_process.exec.
|
child_process errors question
|
<!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **6.10**:
* **OSX**:
* **10.11.3**:
<!-- Enter your issue details below this comment. -->
Run child_process.exec to call webpack2 building, and webpack write log to stderr, then child_process.exec get a callback error, but the error object has no code field.
It is a problem of nodejs?
|
1.0
|
Has a problem with child_process.exec. - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **6.10**:
* **OSX**:
* **10.11.3**:
<!-- Enter your issue details below this comment. -->
Run child_process.exec to call webpack2 building, and webpack write log to stderr, then child_process.exec get a callback error, but the error object has no code field.
It is a problem of nodejs?
|
process
|
has a problem with child process exec thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able osx run child process exec to call building and webpack write log to stderr then child process exec get a callback error but the error object has no code field it is a problem of nodejs
| 1
|
3,024
| 6,028,382,650
|
IssuesEvent
|
2017-06-08 15:37:46
|
wpninjas/ninja-forms
|
https://api.github.com/repos/wpninjas/ninja-forms
|
closed
|
PHP Warning on Submission
|
BUG: Minor FRONT: Processing
|
PHP Warning: preg_match_all() expects parameter 2 to be string, array given in /nas/content/live/fr2000/wp-content/plugins/ninja-forms/includes/MergeTags/WP.php on line 48
|
1.0
|
PHP Warning on Submission - PHP Warning: preg_match_all() expects parameter 2 to be string, array given in /nas/content/live/fr2000/wp-content/plugins/ninja-forms/includes/MergeTags/WP.php on line 48
|
process
|
php warning on submission php warning preg match all expects parameter to be string array given in nas content live wp content plugins ninja forms includes mergetags wp php on line
| 1
|
21,099
| 28,053,956,675
|
IssuesEvent
|
2023-03-29 08:08:29
|
scikit-learn/scikit-learn
|
https://api.github.com/repos/scikit-learn/scikit-learn
|
closed
|
Add the max_categories functionality from OneHotEncoder in OrdinalEncoder
|
New Feature module:preprocessing
|
### Describe the workflow you want to enable
I want to be able to limit the number of output levels from the OrdinalEncoder and also have a new "unknown_value" mapped to this same encoded level. This is motivated in the following bug report
https://github.com/scikit-learn/scikit-learn/issues/25627
### Describe your proposed solution
The functionality should work the same as in OneHotEncoder for consistency
https://scikit-learn.org/stable/modules/generated/sklearn.preprocessing.OneHotEncoder.html
There are three components:
1) max_categories parameter
2) min_frequency parameter
3) Option for ‘infrequent_if_exist’ in handle_unknown parameter
### Describe alternatives you've considered, if relevant
I currently do this before the encoding but it is limited by lack of integration with handle_unknown
### Additional context
_No response_
|
1.0
|
Add the max_categories functionality from OneHotEncoder in OrdinalEncoder - ### Describe the workflow you want to enable
I want to be able to limit the number of output levels from the OrdinalEncoder and also have a new "unknown_value" mapped to this same encoded level. This is motivated in the following bug report
https://github.com/scikit-learn/scikit-learn/issues/25627
### Describe your proposed solution
The functionality should work the same as in OneHotEncoder for consistency
https://scikit-learn.org/stable/modules/generated/sklearn.preprocessing.OneHotEncoder.html
There are three components:
1) max_categories parameter
2) min_frequency parameter
3) Option for ‘infrequent_if_exist’ in handle_unknown parameter
### Describe alternatives you've considered, if relevant
I currently do this before the encoding but it is limited by lack of integration with handle_unknown
### Additional context
_No response_
|
process
|
add the max categories functionality from onehotencoder in ordinalencoder describe the workflow you want to enable i want to be able to limit the number of output levels from the ordinalencoder and also have a new unknown value mapped to this same encoded level this is motivated in the following bug report describe your proposed solution the functionality should work the same as in onehotencoder for consistency there are three components max categories parameter min frequency parameter option for ‘infrequent if exist’ in handle unknown parameter describe alternatives you ve considered if relevant i currently do this before the encoding but it is limited by lack of integration with handle unknown additional context no response
| 1
|
11,441
| 7,238,405,901
|
IssuesEvent
|
2018-02-13 14:31:55
|
nciinc/fs-permit-platform
|
https://api.github.com/repos/nciinc/fs-permit-platform
|
opened
|
As a user, I would like to see all the Christmas tree cutting rules of a National Forest on one page so I can easily view, acknowledge and print the rules before buying my permit.
|
usability - high
|
This comes from SME/User usability testing from sprints 06/08.
|
True
|
As a user, I would like to see all the Christmas tree cutting rules of a National Forest on one page so I can easily view, acknowledge and print the rules before buying my permit. - This comes from SME/User usability testing from sprints 06/08.
|
non_process
|
as a user i would like to see all the christmas tree cutting rules of a national forest on one page so i can easily view acknowledge and print the rules before buying my permit this comes from sme user usability testing from sprints
| 0
|
1,475
| 4,053,878,408
|
IssuesEvent
|
2016-05-24 10:10:38
|
opentrials/opentrials
|
https://api.github.com/repos/opentrials/opentrials
|
opened
|
Extract age information from trials
|
Processors
|
For the registries and trials that have it, we should extract the age range of the participants.
|
1.0
|
Extract age information from trials - For the registries and trials that have it, we should extract the age range of the participants.
|
process
|
extract age information from trials for the registries and trials that have it we should extract the age range of the participants
| 1
|
774,129
| 27,184,326,205
|
IssuesEvent
|
2023-02-19 02:06:25
|
Together-Java/TJ-Bot
|
https://api.github.com/repos/Together-Java/TJ-Bot
|
closed
|
ScamBlocker DM failure shouldnt result in an error
|
bug good first issue priority: normal help wanted
|
The `ScamBlocker` sends DMs to the authors of a message to inform them about the action. Thats totally fine and expected, but failure to DM a user (for example if they blocked DMs), it should not yield to an exception being thrown in the logs:
> ```java
> net.dv8tion.jda.api.requests.RestAction
> ERROR
> RestAction queue returned failure: [ErrorResponseException] 50007: Cannot send messages to this user
> net.dv8tion.jda.api.exceptions.ContextException
> at net.dv8tion.jda.api.exceptions.ContextException.here(ContextException.java:54)
> at net.dv8tion.jda.internal.requests.restaction.operator.RestActionOperator.contextWrap(RestActionOperator.java:123)
> at net.dv8tion.jda.internal.requests.restaction.operator.FlatMapRestAction.queue(FlatMapRestAction.java:50)
> at net.dv8tion.jda.api.requests.RestAction.queue(RestAction.java:572)
> at net.dv8tion.jda.api.requests.RestAction.queue(RestAction.java:538)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.dmUser(ScamBlocker.java:256)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.dmUser(ScamBlocker.java:252)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.takeActionAutoDeleteAndQuarantine(ScamBlocker.java:189)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.takeAction(ScamBl...
> 01/03/2023 5:52 PM
> ```
The actual DM sending should probably use `mapToResults()` and then maybe send a DEBUG level log message:

|
1.0
|
ScamBlocker DM failure shouldnt result in an error - The `ScamBlocker` sends DMs to the authors of a message to inform them about the action. Thats totally fine and expected, but failure to DM a user (for example if they blocked DMs), it should not yield to an exception being thrown in the logs:
> ```java
> net.dv8tion.jda.api.requests.RestAction
> ERROR
> RestAction queue returned failure: [ErrorResponseException] 50007: Cannot send messages to this user
> net.dv8tion.jda.api.exceptions.ContextException
> at net.dv8tion.jda.api.exceptions.ContextException.here(ContextException.java:54)
> at net.dv8tion.jda.internal.requests.restaction.operator.RestActionOperator.contextWrap(RestActionOperator.java:123)
> at net.dv8tion.jda.internal.requests.restaction.operator.FlatMapRestAction.queue(FlatMapRestAction.java:50)
> at net.dv8tion.jda.api.requests.RestAction.queue(RestAction.java:572)
> at net.dv8tion.jda.api.requests.RestAction.queue(RestAction.java:538)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.dmUser(ScamBlocker.java:256)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.dmUser(ScamBlocker.java:252)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.takeActionAutoDeleteAndQuarantine(ScamBlocker.java:189)
> at org.togetherjava.tjbot.commands.moderation.scam.ScamBlocker.takeAction(ScamBl...
> 01/03/2023 5:52 PM
> ```
The actual DM sending should probably use `mapToResults()` and then maybe send a DEBUG level log message:

|
non_process
|
scamblocker dm failure shouldnt result in an error the scamblocker sends dms to the authors of a message to inform them about the action thats totally fine and expected but failure to dm a user for example if they blocked dms it should not yield to an exception being thrown in the logs java net jda api requests restaction error restaction queue returned failure cannot send messages to this user net jda api exceptions contextexception at net jda api exceptions contextexception here contextexception java at net jda internal requests restaction operator restactionoperator contextwrap restactionoperator java at net jda internal requests restaction operator flatmaprestaction queue flatmaprestaction java at net jda api requests restaction queue restaction java at net jda api requests restaction queue restaction java at org togetherjava tjbot commands moderation scam scamblocker dmuser scamblocker java at org togetherjava tjbot commands moderation scam scamblocker dmuser scamblocker java at org togetherjava tjbot commands moderation scam scamblocker takeactionautodeleteandquarantine scamblocker java at org togetherjava tjbot commands moderation scam scamblocker takeaction scambl pm the actual dm sending should probably use maptoresults and then maybe send a debug level log message
| 0
|
13,357
| 15,819,635,351
|
IssuesEvent
|
2021-04-05 17:47:18
|
Figma-Linux/figma-linux
|
https://api.github.com/repos/Figma-Linux/figma-linux
|
closed
|
Log out button does nothing
|
Kind: Main Process Priority: High Status: To Do Type: Bug
|
I have tried to log out of my personal account to enter to my job account, the button for log out does nothing.
I also tried with the option inside the arrow menu and also does nothing.
|
1.0
|
Log out button does nothing - I have tried to log out of my personal account to enter to my job account, the button for log out does nothing.
I also tried with the option inside the arrow menu and also does nothing.
|
process
|
log out button does nothing i have tried to log out of my personal account to enter to my job account the button for log out does nothing i also tried with the option inside the arrow menu and also does nothing
| 1
|
323,112
| 9,842,999,303
|
IssuesEvent
|
2019-06-18 10:33:11
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
syndication.exdynsrv.com - see bug description
|
browser-firefox engine-gecko priority-important
|
<!-- @browser: Firefox 64.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:64.0) Gecko/20100101 Firefox/64.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://syndication.exdynsrv.com/splash.php?idzone=2836516&sub=9&type=8&p=https%3A%2F%2Fthegoodcaster.com%2Fredirect%2F57a764d042bf8&tested=1&check=75ee04bc3e21f7178f48b1268dff6794&screen_resolution=1366x768&container_resolution=1366x654&iframe=0
**Browser / Version**: Firefox 64.0
**Operating System**: Windows 7
**Tested Another Browser**: Unknown
**Problem type**: Something else
**Description**: very bad linked have issue open suddenly
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/6/9e57665c-2555-49aa-ad93-caec96d2dfb1.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190108160530</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: release</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "unreachable code after return statement" {file: "https://decentsite.xyz/inject.js?tag=ww_60_2&did=1&v_tag=false&k_tag=false" line: 594 column: 8 source: " scriptRoot.setAttribute(\'id\', \'M392137ScriptRootC\' + blockId"}]', u'[JavaScript Warning: "unreachable code after return statement" {file: "https://decentsite.xyz/inject.js?tag=ww_60_2&did=1&v_tag=false&k_tag=false" line: 618}]', u'[JavaScript Warning: "unreachable code after return statement" {file: "https://decentsite.xyz/inject.js?tag=ww_60_2&did=1&v_tag=false&k_tag=false" line: 1163 column: 8 source: " url += "bs=" + encodeURIComponent(blockSize);"}]', u'[JavaScript Warning: "unreachable code after return statement" {file: "https://decentsite.xyz/inject.js?tag=ww_60_2&did=1&v_tag=false&k_tag=false" line: 618}]', u'[console.log(onMessage extension) moz-extension://17087568-508b-4a45-9c9d-8f4b8abf0e4d/scripts/extension.js:24:17]', u'[console.log(Clean the cache of the scraper (new onComplete event)) moz-extension://17087568-508b-4a45-9c9d-8f4b8abf0e4d/common/scripts/ial.js:425:13]', u'[console.log(SecurityError: Permission denied to access property "addEventListener" on cross-origin object) https://decentsite.xyz/inject.js?tag=ww_60_2&did=1&v_tag=false&k_tag=false:1574:9]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
syndication.exdynsrv.com - see bug description - <!-- @browser: Firefox 64.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:64.0) Gecko/20100101 Firefox/64.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://syndication.exdynsrv.com/splash.php?idzone=2836516&sub=9&type=8&p=https%3A%2F%2Fthegoodcaster.com%2Fredirect%2F57a764d042bf8&tested=1&check=75ee04bc3e21f7178f48b1268dff6794&screen_resolution=1366x768&container_resolution=1366x654&iframe=0
**Browser / Version**: Firefox 64.0
**Operating System**: Windows 7
**Tested Another Browser**: Unknown
**Problem type**: Something else
**Description**: very bad linked have issue open suddenly
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/6/9e57665c-2555-49aa-ad93-caec96d2dfb1.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190108160530</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: release</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "unreachable code after return statement" {file: "https://decentsite.xyz/inject.js?tag=ww_60_2&did=1&v_tag=false&k_tag=false" line: 594 column: 8 source: " scriptRoot.setAttribute(\'id\', \'M392137ScriptRootC\' + blockId"}]', u'[JavaScript Warning: "unreachable code after return statement" {file: "https://decentsite.xyz/inject.js?tag=ww_60_2&did=1&v_tag=false&k_tag=false" line: 618}]', u'[JavaScript Warning: "unreachable code after return statement" {file: "https://decentsite.xyz/inject.js?tag=ww_60_2&did=1&v_tag=false&k_tag=false" line: 1163 column: 8 source: " url += "bs=" + encodeURIComponent(blockSize);"}]', u'[JavaScript Warning: "unreachable code after return statement" {file: "https://decentsite.xyz/inject.js?tag=ww_60_2&did=1&v_tag=false&k_tag=false" line: 618}]', u'[console.log(onMessage extension) moz-extension://17087568-508b-4a45-9c9d-8f4b8abf0e4d/scripts/extension.js:24:17]', u'[console.log(Clean the cache of the scraper (new onComplete event)) moz-extension://17087568-508b-4a45-9c9d-8f4b8abf0e4d/common/scripts/ial.js:425:13]', u'[console.log(SecurityError: Permission denied to access property "addEventListener" on cross-origin object) https://decentsite.xyz/inject.js?tag=ww_60_2&did=1&v_tag=false&k_tag=false:1574:9]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
syndication exdynsrv com see bug description url browser version firefox operating system windows tested another browser unknown problem type something else description very bad linked have issue open suddenly steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel release console messages u u u u u u from with ❤️
| 0
|
1,203
| 3,702,737,047
|
IssuesEvent
|
2016-02-29 17:51:52
|
pelias/fuzzy-tester
|
https://api.github.com/repos/pelias/fuzzy-tester
|
closed
|
allow expected properties to be object or array
|
processed
|
Currently, if it's an array of size 1 (or more) everything is good. but simply an object fails.
|
1.0
|
allow expected properties to be object or array - Currently, if it's an array of size 1 (or more) everything is good. but simply an object fails.
|
process
|
allow expected properties to be object or array currently if it s an array of size or more everything is good but simply an object fails
| 1
|
11,709
| 14,545,660,809
|
IssuesEvent
|
2020-12-15 19:59:58
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Does Demand specification allow for "-contains" string?
|
Pri1 devops-cicd-process/tech devops/prod doc-enhancement product-question
|
I was wondering if it is possible to substring-match a capability within a demand.
The current version of the docs suggests that this could potentially be possible, it's still not pretty clear though:
* the comment says "# check for specific string in capability" (sounds like substring match)
* the code says "agent.os -equals Darwin" (sounds like full string equality match)

Are there any other supported operators in addition to `-equals` when specifying a demand, e.g. such as `-contains`, `-gt`, etc?
I was hoping for a subset of [PowerShell Operators](https://docs.microsoft.com/en-us/powershell/module/microsoft.powershell.core/about/about_operators) or similar (mostly logical and comparison operators), but found nothing indicative of that in the docs...
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: e7541ee6-d2bb-84c0-fead-1aa8ee7d2372
* Version Independent ID: 5cf7c51e-37e1-6c67-e6c6-80262c4eb662
* Content: [Demands - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/demands?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/demands.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/demands.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @steved0x
* Microsoft Alias: **sdanie**
|
1.0
|
Does Demand specification allow for "-contains" string? - I was wondering if it is possible to substring-match a capability within a demand.
The current version of the docs suggests that this could potentially be possible, it's still not pretty clear though:
* the comment says "# check for specific string in capability" (sounds like substring match)
* the code says "agent.os -equals Darwin" (sounds like full string equality match)

Are there any other supported operators in addition to `-equals` when specifying a demand, e.g. such as `-contains`, `-gt`, etc?
I was hoping for a subset of [PowerShell Operators](https://docs.microsoft.com/en-us/powershell/module/microsoft.powershell.core/about/about_operators) or similar (mostly logical and comparison operators), but found nothing indicative of that in the docs...
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: e7541ee6-d2bb-84c0-fead-1aa8ee7d2372
* Version Independent ID: 5cf7c51e-37e1-6c67-e6c6-80262c4eb662
* Content: [Demands - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/demands?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/demands.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/demands.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @steved0x
* Microsoft Alias: **sdanie**
|
process
|
does demand specification allow for contains string i was wondering if it is possible to substring match a capability within a demand the current version of the docs suggests that this could potentially be possible it s still not pretty clear though the comment says check for specific string in capability sounds like substring match the code says agent os equals darwin sounds like full string equality match are there any other supported operators in addition to equals when specifying a demand e g such as contains gt etc i was hoping for a subset of or similar mostly logical and comparison operators but found nothing indicative of that in the docs document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id fead version independent id content content source product devops technology devops cicd process github login microsoft alias sdanie
| 1
|
727,528
| 25,038,344,113
|
IssuesEvent
|
2022-11-04 18:04:25
|
umgc/fall2022
|
https://api.github.com/repos/umgc/fall2022
|
closed
|
Investigate technical approach for Google Assistant integration
|
*** HIGH PRIORITY *** support
|
Consider integration with Flutter/Android studio
Consider the functionality and constraints placed on us by Google Assistant.
Determine set of realistic functions that would be useful for our purposes
Determine if we could potentially fulfill our "Chatbot" requirements with the voice assistant instead
Document findings of approaches considered, and why they should be selected or excluded as we need this information in the TDD. Place findings in the documents as a draft, no need for extensive polishing within this task. Some content may be moved out later to the Programmers Guide.
Additionally, if implementation appears to be simple, developing a proof of concept (or locating an existing one built by someone else) would also be great for our next presentation.
|
1.0
|
Investigate technical approach for Google Assistant integration - Consider integration with Flutter/Android studio
Consider the functionality and constraints placed on us by Google Assistant.
Determine set of realistic functions that would be useful for our purposes
Determine if we could potentially fulfill our "Chatbot" requirements with the voice assistant instead
Document findings of approaches considered, and why they should be selected or excluded as we need this information in the TDD. Place findings in the documents as a draft, no need for extensive polishing within this task. Some content may be moved out later to the Programmers Guide.
Additionally, if implementation appears to be simple, developing a proof of concept (or locating an existing one built by someone else) would also be great for our next presentation.
|
non_process
|
investigate technical approach for google assistant integration consider integration with flutter android studio consider the functionality and constraints placed on us by google assistant determine set of realistic functions that would be useful for our purposes determine if we could potentially fulfill our chatbot requirements with the voice assistant instead document findings of approaches considered and why they should be selected or excluded as we need this information in the tdd place findings in the documents as a draft no need for extensive polishing within this task some content may be moved out later to the programmers guide additionally if implementation appears to be simple developing a proof of concept or locating an existing one built by someone else would also be great for our next presentation
| 0
|
7,027
| 10,188,843,794
|
IssuesEvent
|
2019-08-11 14:35:05
|
atlasmap/atlasmap
|
https://api.github.com/repos/atlasmap/atlasmap
|
closed
|
Create community release job for jenkins
|
cat/process status/stale
|
Here is the build script for monorepo, it builds everything including atlasmap/ui
https://github.com/atlasmap/atlasmap/blob/master/build.sh
|
1.0
|
Create community release job for jenkins - Here is the build script for monorepo, it builds everything including atlasmap/ui
https://github.com/atlasmap/atlasmap/blob/master/build.sh
|
process
|
create community release job for jenkins here is the build script for monorepo it builds everything including atlasmap ui
| 1
|
7,471
| 10,567,544,956
|
IssuesEvent
|
2019-10-06 05:24:44
|
KerasKorea/KerasObjectDetector
|
https://api.github.com/repos/KerasKorea/KerasObjectDetector
|
opened
|
[Preprocessing] Dataset - COCO
|
dataset preprocessing
|
- [ ] Explanation of the dataset
- [ ] Download dataset
- [ ] Generate generator
- [ ] Process label map
|
1.0
|
[Preprocessing] Dataset - COCO - - [ ] Explanation of the dataset
- [ ] Download dataset
- [ ] Generate generator
- [ ] Process label map
|
process
|
dataset coco explanation of the dataset download dataset generate generator process label map
| 1
|
18,857
| 24,775,554,647
|
IssuesEvent
|
2022-10-23 17:35:43
|
Blazebit/blaze-persistence
|
https://api.github.com/repos/Blazebit/blaze-persistence
|
closed
|
Using MappingParameter for collection type doesn't work
|
kind: bug workaround available component: entity-view worth: low component: entity-view-annotation-processor
|
From the question https://github.com/Blazebit/blaze-persistence/discussions/1496
We can see an error:
```
The initialization of the static metamodel class 'nz.co.submitkit.blazeviews.AppointmentView_' failed: java.lang.IllegalArgumentException: Can not set static com.blazebit.persistence.view.metamodel.MethodListAttribute field nz.co.submitkit.blazeviews.AppointmentView_.serviceCodes to com.blazebit.persistence.view.impl.metamodel.attribute.MappingMethodSingularAttribute
```
when using
```java
@MappingParameter("serviceCodes")
List<ServiceCodeDto> getServiceCodes();
void setServiceCodes(List<ServiceCodeDto> set);
```
|
1.0
|
Using MappingParameter for collection type doesn't work - From the question https://github.com/Blazebit/blaze-persistence/discussions/1496
We can see an error:
```
The initialization of the static metamodel class 'nz.co.submitkit.blazeviews.AppointmentView_' failed: java.lang.IllegalArgumentException: Can not set static com.blazebit.persistence.view.metamodel.MethodListAttribute field nz.co.submitkit.blazeviews.AppointmentView_.serviceCodes to com.blazebit.persistence.view.impl.metamodel.attribute.MappingMethodSingularAttribute
```
when using
```java
@MappingParameter("serviceCodes")
List<ServiceCodeDto> getServiceCodes();
void setServiceCodes(List<ServiceCodeDto> set);
```
|
process
|
using mappingparameter for collection type doesn t work from the question we can see an error the initialization of the static metamodel class nz co submitkit blazeviews appointmentview failed java lang illegalargumentexception can not set static com blazebit persistence view metamodel methodlistattribute field nz co submitkit blazeviews appointmentview servicecodes to com blazebit persistence view impl metamodel attribute mappingmethodsingularattribute when using java mappingparameter servicecodes list getservicecodes void setservicecodes list set
| 1
|
253,092
| 19,091,148,036
|
IssuesEvent
|
2021-11-29 12:15:03
|
hashicorp/terraform-provider-azurerm
|
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
|
closed
|
Twitter block is undocumented for azurerm_function_app
|
documentation good first issue service/functions
|
<!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform (and AzureRM Provider) Version
TF v0.13.6
AzureRM v2.51.0
(but this is irrelevant as this is a documentation issue)
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* `azurerm_function_app`
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
N/A
```
### Debug Output
<!---
Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist.
To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html).
--->
### Panic Output
<!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. --->
### Expected Behaviour
<!--- What should have happened? --->
twitter sub-block for auth_settings block should exist in documentation.
### Actual Behaviour
<!--- What actually happened? --->
No twitter sub-block for auth_settings block exist in documentation.
### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. Visit `https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/function_app`
2. Read text: `twitter - (Optional) A twitter block as defined below.`
3. try to find the twitter block documentation
### Important Factoids
<!--- Are there anything atypical about your accounts that we should know? For example: Running in a Azure China/Germany/Government? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Such as vendor documentation?
--->
|
1.0
|
Twitter block is undocumented for azurerm_function_app - <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform (and AzureRM Provider) Version
TF v0.13.6
AzureRM v2.51.0
(but this is irrelevant as this is a documentation issue)
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* `azurerm_function_app`
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
N/A
```
### Debug Output
<!---
Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist.
To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html).
--->
### Panic Output
<!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. --->
### Expected Behaviour
<!--- What should have happened? --->
twitter sub-block for auth_settings block should exist in documentation.
### Actual Behaviour
<!--- What actually happened? --->
No twitter sub-block for auth_settings block exist in documentation.
### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. Visit `https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/function_app`
2. Read text: `twitter - (Optional) A twitter block as defined below.`
3. try to find the twitter block documentation
### Important Factoids
<!--- Are there anything atypical about your accounts that we should know? For example: Running in a Azure China/Germany/Government? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Such as vendor documentation?
--->
|
non_process
|
twitter block is undocumented for azurerm function app please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform and azurerm provider version tf azurerm but this is irrelevant as this is a documentation issue affected resource s azurerm function app terraform configuration files hcl n a debug output please provide a link to a github gist containing the complete debug output please do not paste the debug output in the issue just paste a link to the gist to obtain the debug output see the panic output expected behaviour twitter sub block for auth settings block should exist in documentation actual behaviour no twitter sub block for auth settings block exist in documentation steps to reproduce visit read text twitter optional a twitter block as defined below try to find the twitter block documentation important factoids references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here such as vendor documentation
| 0
|
161,888
| 25,414,522,870
|
IssuesEvent
|
2022-11-22 22:18:06
|
WordPress/pattern-directory
|
https://api.github.com/repos/WordPress/pattern-directory
|
closed
|
Duplicate "My Favorite" items in the top bar
|
[Component] Theme [Status] Needs Design Feedback
|
When viewing the "My Favorites" screen the top, blue bar has some duplicated text:
<img width="1023" alt="image" src="https://user-images.githubusercontent.com/191598/126696389-e55572e6-5161-4048-8ac8-d935f32cc59c.png">
I don't think we should include the bolded breadcrumb indicator. Instead we should bold the button near the search input:
<img width="1002" alt="image" src="https://user-images.githubusercontent.com/191598/126696521-99cfcd61-c38f-488f-84ed-b9bc57ec3788.png">
|
1.0
|
Duplicate "My Favorite" items in the top bar - When viewing the "My Favorites" screen the top, blue bar has some duplicated text:
<img width="1023" alt="image" src="https://user-images.githubusercontent.com/191598/126696389-e55572e6-5161-4048-8ac8-d935f32cc59c.png">
I don't think we should include the bolded breadcrumb indicator. Instead we should bold the button near the search input:
<img width="1002" alt="image" src="https://user-images.githubusercontent.com/191598/126696521-99cfcd61-c38f-488f-84ed-b9bc57ec3788.png">
|
non_process
|
duplicate my favorite items in the top bar when viewing the my favorites screen the top blue bar has some duplicated text img width alt image src i don t think we should include the bolded breadcrumb indicator instead we should bold the button near the search input img width alt image src
| 0
|
62,177
| 17,023,866,969
|
IssuesEvent
|
2021-07-03 04:15:55
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Registration on osm.org fails with HTTP 500 if cookies are not accepted in the user's browser
|
Component: website Priority: minor Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 1.17pm, Thursday, 4th July 2013]**
in the standard user registration process (with email, username and password) https://www.openstreetmap.org/user/new after accepting the contributor's terms https://www.openstreetmap.org/user/terms the next page is https://www.openstreetmap.org/user/save displays which displays a HTTP 500 error without further details. Tested with Opera 12, cookies not accepted, no old cookies were existent, "rest of world" was preselected, "public domain" checked. Confirmed by another user who stumbled over this problem first.
Expected: if cookies are really required then there should be some user friendly message telling about this.
|
1.0
|
Registration on osm.org fails with HTTP 500 if cookies are not accepted in the user's browser - **[Submitted to the original trac issue database at 1.17pm, Thursday, 4th July 2013]**
in the standard user registration process (with email, username and password) https://www.openstreetmap.org/user/new after accepting the contributor's terms https://www.openstreetmap.org/user/terms the next page is https://www.openstreetmap.org/user/save displays which displays a HTTP 500 error without further details. Tested with Opera 12, cookies not accepted, no old cookies were existent, "rest of world" was preselected, "public domain" checked. Confirmed by another user who stumbled over this problem first.
Expected: if cookies are really required then there should be some user friendly message telling about this.
|
non_process
|
registration on osm org fails with http if cookies are not accepted in the user s browser in the standard user registration process with email username and password after accepting the contributor s terms the next page is displays which displays a http error without further details tested with opera cookies not accepted no old cookies were existent rest of world was preselected public domain checked confirmed by another user who stumbled over this problem first expected if cookies are really required then there should be some user friendly message telling about this
| 0
|
21,379
| 29,202,228,677
|
IssuesEvent
|
2023-05-21 00:36:57
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[Remoto] Fullstack PHP Developer na Coodesh
|
SALVADOR PHP MYSQL FULL-STACK LARAVEL SQL ANGULAR VUE REQUISITOS REMOTO PROCESSOS BACKEND GITHUB UMA APIs MANUTENÇÃO Stale
|
## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/fullstack-developer-155154905?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>Techsocial</strong> está em busca de <strong><ins>Fullstack Developer</ins></strong> para compor seu time!</p>
<p></p>
<p>Somos uma empresa de Soluções Tecnológicas, que busca transformar os dados e informações de nossos clientes em resultados. Evoluímos a partir de consultoria em Gestão Empresarial, somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas. A Tech é uma empresa inovadora! Desenvolvemos e aportamos inteligência em softwares, aplicativos, RPAs, APIs entre outras soluções digitais. </p>
## Techsocial:
<p>Somos uma empresa de Soluções Tecnológicas, que busca transformar os dados e informações de nossos clientes em resultados. Evoluímos a partir de consultoria em Gestão Empresarial, somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas.</p>
<p>A Tech é uma empresa inovadora! Desenvolvemos e aportamos inteligência em softwares, aplicativos, RPAs, APIs entre outras soluções digitais. Nossa missão é simplificar os processos de nossos clientes por meio da tecnologia e estruturar grandes bancos de dados para garimparmos e lapidarmos as melhores informações para as empresas.</p><a href='https://coodesh.com/empresas/techsocial-tecnologia-e-consultoria-ltda'>Veja mais no site</a>
## Habilidades:
- PHP
- Laravel
- MySQL
- Web Services
## Local:
100% Remoto
## Requisitos:
- Conhecimento em PHP;
- Experiência em desenvolvimento de aplicações backend e frontend;
- Conhecimento em SQL;
- Experiência em desenvolvimento, operação e manutenção de sistemas web;
- Raciocínio lógico, capacidade analítica e boas práticas para sistemas escaláveis e de processamento paralelo.
## Diferenciais:
- Angular;
- Vue.js.
## Benefícios:
- Convênio Médico;
- Vale refeição;
- Vale alimentação.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack PHP Developer na Techsocial](https://coodesh.com/vagas/fullstack-developer-155154905?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Categoria
Full-Stack
|
1.0
|
[Remoto] Fullstack PHP Developer na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/fullstack-developer-155154905?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>Techsocial</strong> está em busca de <strong><ins>Fullstack Developer</ins></strong> para compor seu time!</p>
<p></p>
<p>Somos uma empresa de Soluções Tecnológicas, que busca transformar os dados e informações de nossos clientes em resultados. Evoluímos a partir de consultoria em Gestão Empresarial, somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas. A Tech é uma empresa inovadora! Desenvolvemos e aportamos inteligência em softwares, aplicativos, RPAs, APIs entre outras soluções digitais. </p>
## Techsocial:
<p>Somos uma empresa de Soluções Tecnológicas, que busca transformar os dados e informações de nossos clientes em resultados. Evoluímos a partir de consultoria em Gestão Empresarial, somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas.</p>
<p>A Tech é uma empresa inovadora! Desenvolvemos e aportamos inteligência em softwares, aplicativos, RPAs, APIs entre outras soluções digitais. Nossa missão é simplificar os processos de nossos clientes por meio da tecnologia e estruturar grandes bancos de dados para garimparmos e lapidarmos as melhores informações para as empresas.</p><a href='https://coodesh.com/empresas/techsocial-tecnologia-e-consultoria-ltda'>Veja mais no site</a>
## Habilidades:
- PHP
- Laravel
- MySQL
- Web Services
## Local:
100% Remoto
## Requisitos:
- Conhecimento em PHP;
- Experiência em desenvolvimento de aplicações backend e frontend;
- Conhecimento em SQL;
- Experiência em desenvolvimento, operação e manutenção de sistemas web;
- Raciocínio lógico, capacidade analítica e boas práticas para sistemas escaláveis e de processamento paralelo.
## Diferenciais:
- Angular;
- Vue.js.
## Benefícios:
- Convênio Médico;
- Vale refeição;
- Vale alimentação.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack PHP Developer na Techsocial](https://coodesh.com/vagas/fullstack-developer-155154905?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Categoria
Full-Stack
|
process
|
fullstack php developer na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a techsocial está em busca de fullstack developer para compor seu time somos uma empresa de soluções tecnológicas que busca transformar os dados e informações de nossos clientes em resultados evoluímos a partir de consultoria em gestão empresarial somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas a tech é uma empresa inovadora desenvolvemos e aportamos inteligência em softwares aplicativos rpas apis entre outras soluções digitais nbsp techsocial somos uma empresa de soluções tecnológicas que busca transformar os dados e informações de nossos clientes em resultados evoluímos a partir de consultoria em gestão empresarial somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas a tech é uma empresa inovadora desenvolvemos e aportamos inteligência em softwares aplicativos rpas apis entre outras soluções digitais nossa missão é simplificar os processos de nossos clientes por meio da tecnologia e estruturar grandes bancos de dados para garimparmos e lapidarmos as melhores informações para as empresas habilidades php laravel mysql web services local remoto requisitos conhecimento em php experiência em desenvolvimento de aplicações backend e frontend conhecimento em sql experiência em desenvolvimento operação e manutenção de sistemas web raciocínio lógico capacidade analítica e boas práticas para sistemas escaláveis e de processamento paralelo diferenciais angular vue js benefícios convênio médico vale refeição vale alimentação como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto categoria full stack
| 1
|
63,352
| 14,656,703,905
|
IssuesEvent
|
2020-12-28 14:00:52
|
fu1771695yongxie/wepy
|
https://api.github.com/repos/fu1771695yongxie/wepy
|
opened
|
CVE-2019-11358 (Medium) detected in jquery-2.1.4.min.js
|
security vulnerability
|
## CVE-2019-11358 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: wepy/packages/cli/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>Path to vulnerable library: wepy/packages/cli/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/wepy/commit/01d3dc60ddf82d562895250655097a09e9db1196">01d3dc60ddf82d562895250655097a09e9db1196</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
<p>Publish Date: 2019-04-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p>
<p>Release Date: 2019-04-20</p>
<p>Fix Resolution: 3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-11358 (Medium) detected in jquery-2.1.4.min.js - ## CVE-2019-11358 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: wepy/packages/cli/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>Path to vulnerable library: wepy/packages/cli/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/wepy/commit/01d3dc60ddf82d562895250655097a09e9db1196">01d3dc60ddf82d562895250655097a09e9db1196</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
<p>Publish Date: 2019-04-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p>
<p>Release Date: 2019-04-20</p>
<p>Fix Resolution: 3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file wepy packages cli node modules js attic test moment index html path to vulnerable library wepy packages cli node modules js attic test moment index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
222,307
| 17,063,628,601
|
IssuesEvent
|
2021-07-07 02:50:48
|
sqlcollaborative/dbatools
|
https://api.github.com/repos/sqlcollaborative/dbatools
|
closed
|
[Blog] Copy-DbaLogin write-up
|
Documentation good first issue
|
I didn't even remember LoginRenameHashtable existed! Can someone please write an in-depth blog post about Copy-DbaLogin that we can put on dbatools blog?
thanks !
|
1.0
|
[Blog] Copy-DbaLogin write-up - I didn't even remember LoginRenameHashtable existed! Can someone please write an in-depth blog post about Copy-DbaLogin that we can put on dbatools blog?
thanks !
|
non_process
|
copy dbalogin write up i didn t even remember loginrenamehashtable existed can someone please write an in depth blog post about copy dbalogin that we can put on dbatools blog thanks
| 0
|
2,561
| 5,315,734,729
|
IssuesEvent
|
2017-02-13 18:03:59
|
scieloorg/search-journals
|
https://api.github.com/repos/scieloorg/search-journals
|
closed
|
Avaliar os campos utilizado na indexação
|
Processamento
|
O arquivo de esquema do Solr está com vários campos que não estão sendo utilizados tão pouco é indexado, devemos avaliar esses campos documentar e remover do arquivo de esquema.
|
1.0
|
Avaliar os campos utilizado na indexação - O arquivo de esquema do Solr está com vários campos que não estão sendo utilizados tão pouco é indexado, devemos avaliar esses campos documentar e remover do arquivo de esquema.
|
process
|
avaliar os campos utilizado na indexação o arquivo de esquema do solr está com vários campos que não estão sendo utilizados tão pouco é indexado devemos avaliar esses campos documentar e remover do arquivo de esquema
| 1
|
14,030
| 16,827,387,828
|
IssuesEvent
|
2021-06-17 20:36:08
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
closed
|
PH calcs in TH
|
Process Heating Treasure Hunt
|
For any of the PH calc's in TH that have the flue gas calculation (either the flue gas calc or the modal for available heat)
* Default for "Type of fuel" == "Gas
* Default for "Fuel" == "Typical Natual Gas - US"
* stretch goal: if "Energy source type" == "Natural gas" then "Typical Natural Gas - US", else, "Coke Oven Gas"
|
1.0
|
PH calcs in TH - For any of the PH calc's in TH that have the flue gas calculation (either the flue gas calc or the modal for available heat)
* Default for "Type of fuel" == "Gas
* Default for "Fuel" == "Typical Natual Gas - US"
* stretch goal: if "Energy source type" == "Natural gas" then "Typical Natural Gas - US", else, "Coke Oven Gas"
|
process
|
ph calcs in th for any of the ph calc s in th that have the flue gas calculation either the flue gas calc or the modal for available heat default for type of fuel gas default for fuel typical natual gas us stretch goal if energy source type natural gas then typical natural gas us else coke oven gas
| 1
|
5,271
| 8,059,623,808
|
IssuesEvent
|
2018-08-02 22:51:05
|
GoogleCloudPlatform/google-cloud-python
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
|
closed
|
Release BigQuery 1.5.0
|
api: bigquery packaging type: process
|
For clustering feature.
@alixhami Do you want to try your hand at releasing this one?
|
1.0
|
Release BigQuery 1.5.0 - For clustering feature.
@alixhami Do you want to try your hand at releasing this one?
|
process
|
release bigquery for clustering feature alixhami do you want to try your hand at releasing this one
| 1
|
159,350
| 20,048,363,697
|
IssuesEvent
|
2022-02-03 01:09:22
|
kapseliboi/token-wizard
|
https://api.github.com/repos/kapseliboi/token-wizard
|
opened
|
CVE-2021-25949 (High) detected in set-getter-0.1.0.tgz
|
security vulnerability
|
## CVE-2021-25949 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>set-getter-0.1.0.tgz</b></p></summary>
<p>Create nested getter properties and any intermediary dot notation (`'a.b.c'`) paths</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-getter/-/set-getter-0.1.0.tgz">https://registry.npmjs.org/set-getter/-/set-getter-0.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/set-getter/package.json</p>
<p>
Dependency Hierarchy:
- markdown-toc-1.2.0.tgz (Root Library)
- lazy-cache-2.0.2.tgz
- :x: **set-getter-0.1.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in 'set-getter' version 0.1.0 allows an attacker to cause a denial of service and may lead to remote code execution.
<p>Publish Date: 2021-06-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25949>CVE-2021-25949</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/doowb/set-getter/commit/66eb3f0d4686a4a8c7c3d6f7ecd8e570b580edc4">https://github.com/doowb/set-getter/commit/66eb3f0d4686a4a8c7c3d6f7ecd8e570b580edc4</a></p>
<p>Release Date: 2021-06-10</p>
<p>Fix Resolution: set-getter - 0.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-25949 (High) detected in set-getter-0.1.0.tgz - ## CVE-2021-25949 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>set-getter-0.1.0.tgz</b></p></summary>
<p>Create nested getter properties and any intermediary dot notation (`'a.b.c'`) paths</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-getter/-/set-getter-0.1.0.tgz">https://registry.npmjs.org/set-getter/-/set-getter-0.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/set-getter/package.json</p>
<p>
Dependency Hierarchy:
- markdown-toc-1.2.0.tgz (Root Library)
- lazy-cache-2.0.2.tgz
- :x: **set-getter-0.1.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in 'set-getter' version 0.1.0 allows an attacker to cause a denial of service and may lead to remote code execution.
<p>Publish Date: 2021-06-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25949>CVE-2021-25949</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/doowb/set-getter/commit/66eb3f0d4686a4a8c7c3d6f7ecd8e570b580edc4">https://github.com/doowb/set-getter/commit/66eb3f0d4686a4a8c7c3d6f7ecd8e570b580edc4</a></p>
<p>Release Date: 2021-06-10</p>
<p>Fix Resolution: set-getter - 0.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in set getter tgz cve high severity vulnerability vulnerable library set getter tgz create nested getter properties and any intermediary dot notation a b c paths library home page a href path to dependency file package json path to vulnerable library node modules set getter package json dependency hierarchy markdown toc tgz root library lazy cache tgz x set getter tgz vulnerable library found in base branch master vulnerability details prototype pollution vulnerability in set getter version allows an attacker to cause a denial of service and may lead to remote code execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution set getter step up your open source security game with whitesource
| 0
|
275,813
| 30,308,201,141
|
IssuesEvent
|
2023-07-10 10:58:00
|
berviantoleo/telegram-bot-ai
|
https://api.github.com/repos/berviantoleo/telegram-bot-ai
|
opened
|
xunit.2.5.0.nupkg: 1 vulnerabilities (highest severity is: 7.5)
|
Mend: dependency security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xunit.2.5.0.nupkg</b></p></summary>
<p></p>
<p>Path to dependency file: /TelegramBot.Test/TelegramBot.Test.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/berviantoleo/telegram-bot-ai/commit/5dca2efc5192faf8a853e40b84550d4180aa86bc">5dca2efc5192faf8a853e40b84550d4180aa86bc</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (xunit.2.5.0.nupkg version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2018-8292](https://www.mend.io/vulnerability-database/CVE-2018-8292) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | system.net.http.4.3.0.nupkg | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2018-8292</summary>
### Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers.
</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /TelegramBot.Test/TelegramBot.Test.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- xunit.2.5.0.nupkg (Root Library)
- xunit.assert.2.5.0.nupkg
- netstandard.library.1.6.1.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/berviantoleo/telegram-bot-ai/commit/5dca2efc5192faf8a853e40b84550d4180aa86bc">5dca2efc5192faf8a853e40b84550d4180aa86bc</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0.
<p>Publish Date: 2018-10-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8292>CVE-2018-8292</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-10-10</p>
<p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
xunit.2.5.0.nupkg: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xunit.2.5.0.nupkg</b></p></summary>
<p></p>
<p>Path to dependency file: /TelegramBot.Test/TelegramBot.Test.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/berviantoleo/telegram-bot-ai/commit/5dca2efc5192faf8a853e40b84550d4180aa86bc">5dca2efc5192faf8a853e40b84550d4180aa86bc</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (xunit.2.5.0.nupkg version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2018-8292](https://www.mend.io/vulnerability-database/CVE-2018-8292) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | system.net.http.4.3.0.nupkg | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2018-8292</summary>
### Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers.
</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /TelegramBot.Test/TelegramBot.Test.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- xunit.2.5.0.nupkg (Root Library)
- xunit.assert.2.5.0.nupkg
- netstandard.library.1.6.1.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/berviantoleo/telegram-bot-ai/commit/5dca2efc5192faf8a853e40b84550d4180aa86bc">5dca2efc5192faf8a853e40b84550d4180aa86bc</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0.
<p>Publish Date: 2018-10-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8292>CVE-2018-8292</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-10-10</p>
<p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_process
|
xunit nupkg vulnerabilities highest severity is vulnerable library xunit nupkg path to dependency file telegrambot test telegrambot test csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg found in head commit a href vulnerabilities cve severity cvss dependency type fixed in xunit nupkg version remediation available high system net http nupkg transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the details section below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that allow applications to consume web services over http and http components that can be used by both clients and servers for parsing http headers library home page a href path to dependency file telegrambot test telegrambot test csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg dependency hierarchy xunit nupkg root library xunit assert nupkg netstandard library nupkg x system net http nupkg vulnerable library found in head commit a href found in base branch main vulnerability details an information disclosure vulnerability exists in net core when authentication information is inadvertently exposed in a redirect aka net core information disclosure vulnerability this affects net core net core net core powershell core publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution system net http microsoft powershell commands utility rc step up your open source security game with mend
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.