Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
9,536
| 12,504,875,738
|
IssuesEvent
|
2020-06-02 09:44:20
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
Add a test case for project paths with spaces
|
process/candidate
|
Some user's project paths contain spaces. This broke client generation various times already, e.g. https://github.com/prisma/prisma/issues/1973 and https://github.com/prisma/prisma/issues/2612.
We should add at least one dedicated test for this so this won't break again for our users.
|
1.0
|
Add a test case for project paths with spaces - Some user's project paths contain spaces. This broke client generation various times already, e.g. https://github.com/prisma/prisma/issues/1973 and https://github.com/prisma/prisma/issues/2612.
We should add at least one dedicated test for this so this won't break again for our users.
|
process
|
add a test case for project paths with spaces some user s project paths contain spaces this broke client generation various times already e g and we should add at least one dedicated test for this so this won t break again for our users
| 1
|
21,982
| 30,474,010,994
|
IssuesEvent
|
2023-07-17 15:17:10
|
The-Data-Alchemists-Manipal/MindWave
|
https://api.github.com/repos/The-Data-Alchemists-Manipal/MindWave
|
closed
|
Thumbnailator
|
gssoc23 level2 image-processing
|
### Is your feature request related to a problem? Please describe.
Thumbnailator's fluent interface can be used to perform fairly complicated thumbnail processing task in one simple step.
### Describe the solution you'd like
creating JPEG thumbnails of image files in a directory, all resized to a maximum dimension of 640 pixels by 480 pixels while preserving the aspect ratio of the original image can be performed
### Describe alternatives you've considered
_No response_
### Additional context
Assign this issue to me under GSSOC'23
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct
|
1.0
|
Thumbnailator - ### Is your feature request related to a problem? Please describe.
Thumbnailator's fluent interface can be used to perform fairly complicated thumbnail processing task in one simple step.
### Describe the solution you'd like
creating JPEG thumbnails of image files in a directory, all resized to a maximum dimension of 640 pixels by 480 pixels while preserving the aspect ratio of the original image can be performed
### Describe alternatives you've considered
_No response_
### Additional context
Assign this issue to me under GSSOC'23
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct
|
process
|
thumbnailator is your feature request related to a problem please describe thumbnailator s fluent interface can be used to perform fairly complicated thumbnail processing task in one simple step describe the solution you d like creating jpeg thumbnails of image files in a directory all resized to a maximum dimension of pixels by pixels while preserving the aspect ratio of the original image can be performed describe alternatives you ve considered no response additional context assign this issue to me under gssoc code of conduct i agree to follow this project s code of conduct
| 1
|
124,378
| 17,772,541,543
|
IssuesEvent
|
2021-08-30 15:10:40
|
kapseliboi/evergreen
|
https://api.github.com/repos/kapseliboi/evergreen
|
opened
|
CVE-2019-8331 (Medium) detected in bootstrap-3.2.0.min.js
|
security vulnerability
|
## CVE-2019-8331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.2.0.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.2.0/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.2.0/js/bootstrap.min.js</a></p>
<p>Path to dependency file: evergreen/services/node_modules/remarkable/demo/index.html</p>
<p>Path to vulnerable library: /services/node_modules/remarkable/demo/index.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.2.0.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/evergreen/commit/13675096220f0e986aa94cafc5f57de6b38e38cd">13675096220f0e986aa94cafc5f57de6b38e38cd</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/28236">https://github.com/twbs/bootstrap/pull/28236</a></p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-8331 (Medium) detected in bootstrap-3.2.0.min.js - ## CVE-2019-8331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.2.0.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.2.0/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.2.0/js/bootstrap.min.js</a></p>
<p>Path to dependency file: evergreen/services/node_modules/remarkable/demo/index.html</p>
<p>Path to vulnerable library: /services/node_modules/remarkable/demo/index.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.2.0.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/evergreen/commit/13675096220f0e986aa94cafc5f57de6b38e38cd">13675096220f0e986aa94cafc5f57de6b38e38cd</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/28236">https://github.com/twbs/bootstrap/pull/28236</a></p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file evergreen services node modules remarkable demo index html path to vulnerable library services node modules remarkable demo index html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap bootstrap sass step up your open source security game with whitesource
| 0
|
281,037
| 8,689,924,593
|
IssuesEvent
|
2018-12-03 20:03:37
|
SpaceNetChallenge/utilities
|
https://api.github.com/repos/SpaceNetChallenge/utilities
|
closed
|
You might want a .gitignore file in this repo
|
Difficulty: Easy Priority: High Status: Unassigned Type: Maintenance
|
`.pyc` files don't need to be committed; they can be filtered out with .gitignore.
|
1.0
|
You might want a .gitignore file in this repo - `.pyc` files don't need to be committed; they can be filtered out with .gitignore.
|
non_process
|
you might want a gitignore file in this repo pyc files don t need to be committed they can be filtered out with gitignore
| 0
|
417,964
| 28,112,517,879
|
IssuesEvent
|
2023-03-31 08:19:31
|
venuslimm/ped
|
https://api.github.com/repos/venuslimm/ped
|
opened
|
Screenshot of find command on user guide does not match with the current UI
|
severity.VeryLow type.DocumentationBug
|
Screenshot on the user guide:

Actual UI:

<!--session: 1680242405607-d67b960d-7651-4494-be68-de1656255e7d-->
<!--Version: Web v3.4.7-->
|
1.0
|
Screenshot of find command on user guide does not match with the current UI - Screenshot on the user guide:

Actual UI:

<!--session: 1680242405607-d67b960d-7651-4494-be68-de1656255e7d-->
<!--Version: Web v3.4.7-->
|
non_process
|
screenshot of find command on user guide does not match with the current ui screenshot on the user guide actual ui
| 0
|
301,569
| 9,221,755,877
|
IssuesEvent
|
2019-03-11 20:48:46
|
RobotLocomotion/drake
|
https://api.github.com/repos/RobotLocomotion/drake
|
opened
|
ci: Turn off `xenial-valgrind-memcheck-weekly`
|
priority: medium team: kitware
|
See Slack convo:
https://drakedevelopers.slack.com/archives/C270MN28G/p1552336898014300?thread_ts=1552312717.012700&cid=C270MN28G
Summary: It's difficult to reproduce in Bionic, and not providing too much value as the Enabling team has moved primary development / usage to Bionic.
\cc @jwnimmer-tri
|
1.0
|
ci: Turn off `xenial-valgrind-memcheck-weekly` - See Slack convo:
https://drakedevelopers.slack.com/archives/C270MN28G/p1552336898014300?thread_ts=1552312717.012700&cid=C270MN28G
Summary: It's difficult to reproduce in Bionic, and not providing too much value as the Enabling team has moved primary development / usage to Bionic.
\cc @jwnimmer-tri
|
non_process
|
ci turn off xenial valgrind memcheck weekly see slack convo summary it s difficult to reproduce in bionic and not providing too much value as the enabling team has moved primary development usage to bionic cc jwnimmer tri
| 0
|
562
| 3,023,861,228
|
IssuesEvent
|
2015-08-01 23:56:31
|
HazyResearch/dd-genomics
|
https://api.github.com/repos/HazyResearch/dd-genomics
|
opened
|
Set up and document AWS pipeline for pre-processing
|
Preprocessing PRIORITY
|
This is / should be already documented in `HazyResearch/bazaar` but make sure we have some notes here / run this for everything today...
|
1.0
|
Set up and document AWS pipeline for pre-processing - This is / should be already documented in `HazyResearch/bazaar` but make sure we have some notes here / run this for everything today...
|
process
|
set up and document aws pipeline for pre processing this is should be already documented in hazyresearch bazaar but make sure we have some notes here run this for everything today
| 1
|
29,811
| 13,173,137,014
|
IssuesEvent
|
2020-08-11 19:44:13
|
thkl/hap-homematic
|
https://api.github.com/repos/thkl/hap-homematic
|
closed
|
Variablen mit mehr Services
|
DeviceService enhancement
|
Ist es Möglich im Bereich Variablen mehr Services unter zubringen. z.B. Luftfeuchtigkeit, Bewegungsmelder, Belegtmelder, Feuchtigkeitssensor etc.
|
1.0
|
Variablen mit mehr Services - Ist es Möglich im Bereich Variablen mehr Services unter zubringen. z.B. Luftfeuchtigkeit, Bewegungsmelder, Belegtmelder, Feuchtigkeitssensor etc.
|
non_process
|
variablen mit mehr services ist es möglich im bereich variablen mehr services unter zubringen z b luftfeuchtigkeit bewegungsmelder belegtmelder feuchtigkeitssensor etc
| 0
|
11,488
| 5,011,878,407
|
IssuesEvent
|
2016-12-13 09:32:43
|
LLNL/spack
|
https://api.github.com/repos/LLNL/spack
|
closed
|
SLEPc fails to configure with Spack's python
|
bug build-error package python
|
I just wiped my installation of Spack to re-install and check things and got the error `Symbol not found: __PyCodecInfo_GetIncrementalDecoder`:
```
==> './configure' '--prefix=/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/slepc-3.7.3-rhcxmg2ntqe3v6epgljeseffnpa4gla2' '--with-arpack-dir=/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/arpack-ng-3.4.0-g76ncwdpqcyx5lm5e65ydwaetbx5sulo/lib' '--with-arpack-flags=-lparpack,-larpack'
Traceback (most recent call last):
File "./configure", line 10, in <module>
execfile(os.path.join(os.path.dirname(__file__), 'config', 'configure.py'))
File "./config/configure.py", line 140, in <module>
import slepc, petsc, arpack, blzpack, trlan, feast, primme, blopex, sowing, lapack
File "/private/var/folders/5k/sqpp24tx3ylds4fgm13pfht00000gn/T/davydden/spack-stage/spack-stage-ZOF1pH/slepc-3.7.3/config/packages/petsc.py", line 22, in <module>
import package, os, sys, commands
File "/private/var/folders/5k/sqpp24tx3ylds4fgm13pfht00000gn/T/davydden/spack-stage/spack-stage-ZOF1pH/slepc-3.7.3/config/package.py", line 22, in <module>
import os, sys, commands, tempfile, shutil, urllib, urlparse, tarfile
File "/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/tempfile.py", line 32, in <module>
import io as _io
File "/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/io.py", line 51, in <module>
import _io
ImportError: dlopen(/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/lib-dynload/_io.so, 2): Symbol not found: __PyCodecInfo_GetIncrementalDecoder
Referenced from: /Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/lib-dynload/_io.so
Expected in: flat namespace
in /Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/lib-dynload/_io.so
```
Looking at the history of `python` package, i don't see what could have led to this.
For now will be using
```
python:
version: [2.7.10]
paths:
python@2.7.10: /usr
buildable: False
```
|
1.0
|
SLEPc fails to configure with Spack's python - I just wiped my installation of Spack to re-install and check things and got the error `Symbol not found: __PyCodecInfo_GetIncrementalDecoder`:
```
==> './configure' '--prefix=/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/slepc-3.7.3-rhcxmg2ntqe3v6epgljeseffnpa4gla2' '--with-arpack-dir=/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/arpack-ng-3.4.0-g76ncwdpqcyx5lm5e65ydwaetbx5sulo/lib' '--with-arpack-flags=-lparpack,-larpack'
Traceback (most recent call last):
File "./configure", line 10, in <module>
execfile(os.path.join(os.path.dirname(__file__), 'config', 'configure.py'))
File "./config/configure.py", line 140, in <module>
import slepc, petsc, arpack, blzpack, trlan, feast, primme, blopex, sowing, lapack
File "/private/var/folders/5k/sqpp24tx3ylds4fgm13pfht00000gn/T/davydden/spack-stage/spack-stage-ZOF1pH/slepc-3.7.3/config/packages/petsc.py", line 22, in <module>
import package, os, sys, commands
File "/private/var/folders/5k/sqpp24tx3ylds4fgm13pfht00000gn/T/davydden/spack-stage/spack-stage-ZOF1pH/slepc-3.7.3/config/package.py", line 22, in <module>
import os, sys, commands, tempfile, shutil, urllib, urlparse, tarfile
File "/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/tempfile.py", line 32, in <module>
import io as _io
File "/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/io.py", line 51, in <module>
import _io
ImportError: dlopen(/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/lib-dynload/_io.so, 2): Symbol not found: __PyCodecInfo_GetIncrementalDecoder
Referenced from: /Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/lib-dynload/_io.so
Expected in: flat namespace
in /Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/lib-dynload/_io.so
```
Looking at the history of `python` package, i don't see what could have led to this.
For now will be using
```
python:
version: [2.7.10]
paths:
python@2.7.10: /usr
buildable: False
```
|
non_process
|
slepc fails to configure with spack s python i just wiped my installation of spack to re install and check things and got the error symbol not found pycodecinfo getincrementaldecoder configure prefix users davydden spack opt spack darwin sierra clang apple slepc with arpack dir users davydden spack opt spack darwin sierra clang apple arpack ng lib with arpack flags lparpack larpack traceback most recent call last file configure line in execfile os path join os path dirname file config configure py file config configure py line in import slepc petsc arpack blzpack trlan feast primme blopex sowing lapack file private var folders t davydden spack stage spack stage slepc config packages petsc py line in import package os sys commands file private var folders t davydden spack stage spack stage slepc config package py line in import os sys commands tempfile shutil urllib urlparse tarfile file users davydden spack opt spack darwin sierra clang apple python lib tempfile py line in import io as io file users davydden spack opt spack darwin sierra clang apple python lib io py line in import io importerror dlopen users davydden spack opt spack darwin sierra clang apple python lib lib dynload io so symbol not found pycodecinfo getincrementaldecoder referenced from users davydden spack opt spack darwin sierra clang apple python lib lib dynload io so expected in flat namespace in users davydden spack opt spack darwin sierra clang apple python lib lib dynload io so looking at the history of python package i don t see what could have led to this for now will be using python version paths python usr buildable false
| 0
|
328,938
| 10,007,238,770
|
IssuesEvent
|
2019-07-14 08:56:47
|
answeropedia/answeropedia.org
|
https://api.github.com/repos/answeropedia/answeropedia.org
|
closed
|
Abolish short answer (extraction from common answer)
|
priority-critical
|
@gomzyakov
>Короткий ответ будем хранить в отдельном поле сущности, не пытаясь выцарапать из общего ответа (т.к. это иногд просто бессмыслено)
|
1.0
|
Abolish short answer (extraction from common answer) - @gomzyakov
>Короткий ответ будем хранить в отдельном поле сущности, не пытаясь выцарапать из общего ответа (т.к. это иногд просто бессмыслено)
|
non_process
|
abolish short answer extraction from common answer gomzyakov короткий ответ будем хранить в отдельном поле сущности не пытаясь выцарапать из общего ответа т к это иногд просто бессмыслено
| 0
|
13,297
| 22,574,830,062
|
IssuesEvent
|
2022-06-28 06:11:47
|
FederatedAI/KubeFATE
|
https://api.github.com/repos/FederatedAI/KubeFATE
|
closed
|
希望将fate-Serving的Ingress的api由networking.k8s.io/v1beta1升级到networking.k8s.io/v1,以适配kubernates-1.22以上版本
|
kind/requirement
|
**Is your feature request related to a problem? Please describe.**
最新版本的k8s安装cluster-serving时报错:https://github.com/FederatedAI/KubeFATE/issues/618
**Describe the solution you'd like**
支持最新的k8s,由networking.k8s.io/v1beta1升级到networking.k8s.io/v1
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
|
1.0
|
希望将fate-Serving的Ingress的api由networking.k8s.io/v1beta1升级到networking.k8s.io/v1,以适配kubernates-1.22以上版本 - **Is your feature request related to a problem? Please describe.**
最新版本的k8s安装cluster-serving时报错:https://github.com/FederatedAI/KubeFATE/issues/618
**Describe the solution you'd like**
支持最新的k8s,由networking.k8s.io/v1beta1升级到networking.k8s.io/v1
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
|
non_process
|
希望将fate serving的ingress的api由networking io io ,以适配kubernates is your feature request related to a problem please describe serving时报错: describe the solution you d like ,由networking io io describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
| 0
|
2,358
| 5,165,992,743
|
IssuesEvent
|
2017-01-17 15:11:17
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
Clarity of `filterFormatted` in Documentation
|
inprocess
|
## Issue
The `filterFormatted` prop has an unclear description in the documentation. I spent hours trying to figure out why my select filter wouldn't work. It turns out I had included `filterFormatted` in the column (because of the [examples for the select filter](http://allenfang.github.io/react-bootstrap-table/example.html#column-filter)) without fully understanding what it did. Looking at the documentation it seemed like I just needed to include it in the column to enable filtering.
## Proposed fix
Make the documentation for `filterFormatted` clearer. To get started in the right direction, I'd suggest something as follows.
> When true, the column will filter using the value returned by the column's formatter. When false (default), the column will filter using the pre-formatted value.
Also consider changing the [filter examples](http://allenfang.github.io/react-bootstrap-table/example.html#column-filter) to not include `filterFormatted` and possibly include one extra example to showcase the use of `filterFormatted` explicitly.
Anyways, thanks to everyone who's worked on this component. It's extremely powerful and versatile.
|
1.0
|
Clarity of `filterFormatted` in Documentation - ## Issue
The `filterFormatted` prop has an unclear description in the documentation. I spent hours trying to figure out why my select filter wouldn't work. It turns out I had included `filterFormatted` in the column (because of the [examples for the select filter](http://allenfang.github.io/react-bootstrap-table/example.html#column-filter)) without fully understanding what it did. Looking at the documentation it seemed like I just needed to include it in the column to enable filtering.
## Proposed fix
Make the documentation for `filterFormatted` clearer. To get started in the right direction, I'd suggest something as follows.
> When true, the column will filter using the value returned by the column's formatter. When false (default), the column will filter using the pre-formatted value.
Also consider changing the [filter examples](http://allenfang.github.io/react-bootstrap-table/example.html#column-filter) to not include `filterFormatted` and possibly include one extra example to showcase the use of `filterFormatted` explicitly.
Anyways, thanks to everyone who's worked on this component. It's extremely powerful and versatile.
|
process
|
clarity of filterformatted in documentation issue the filterformatted prop has an unclear description in the documentation i spent hours trying to figure out why my select filter wouldn t work it turns out i had included filterformatted in the column because of the without fully understanding what it did looking at the documentation it seemed like i just needed to include it in the column to enable filtering proposed fix make the documentation for filterformatted clearer to get started in the right direction i d suggest something as follows when true the column will filter using the value returned by the column s formatter when false default the column will filter using the pre formatted value also consider changing the to not include filterformatted and possibly include one extra example to showcase the use of filterformatted explicitly anyways thanks to everyone who s worked on this component it s extremely powerful and versatile
| 1
|
334,497
| 10,142,069,030
|
IssuesEvent
|
2019-08-03 20:07:38
|
tensorwerk/hangar-py
|
https://api.github.com/repos/tensorwerk/hangar-py
|
closed
|
[BUG REPORT] Commit inside context manager throws RuntimeError
|
Bug: Priority 2 PR In Progress
|
**Describe the bug**
If we try to commit inside the context manager (before `__exit__()`), hangar throws RuntimeError saying `No changes made in the staging area. Cannot commit.`. We should allow the user to do commits inside the context manager IMO but probably with a warning about the performance hit
**Severity**
<!--- fill in the space between `[ ]` with and `x` (ie. `[x]`) --->
Select an option:
- [ ] Data Corruption / Loss of Any Kind
- [x] Unexpected Behavior, Exceptions or Error Thrown
- [ ] Performance Bottleneck
**To Reproduce**
```python
import numpy as np
from hangar import Repository
repo = Repository(path='myhangarrepo')
repo.init(user_name='Sherin Thomas', user_email='sherin@gmail.com', remove_old=True)
# generate data
data = []
for i in range(1000):
data.append(np.random.rand(28, 28))
data = np.array(data)
co = repo.checkout(write=True)
data_dset = co.datasets.init_dataset('mnist_data', prototype=data[0])
co.commit('datasets init')
co.close()
co = repo.checkout(write=True)
data_dset = co.datasets['mnist_data']
with data_dset:
for i in range(len(data)):
sample_name = str(i)
data_dset[sample_name] = data[i]
co.commit('dataset curation: stage 1') # this throws error
co.close()
```
**Expected behavior**
It should not break the program instead raise a warning about the performance hit
|
1.0
|
[BUG REPORT] Commit inside context manager throws RuntimeError - **Describe the bug**
If we try to commit inside the context manager (before `__exit__()`), hangar throws RuntimeError saying `No changes made in the staging area. Cannot commit.`. We should allow the user to do commits inside the context manager IMO but probably with a warning about the performance hit
**Severity**
<!--- fill in the space between `[ ]` with and `x` (ie. `[x]`) --->
Select an option:
- [ ] Data Corruption / Loss of Any Kind
- [x] Unexpected Behavior, Exceptions or Error Thrown
- [ ] Performance Bottleneck
**To Reproduce**
```python
import numpy as np
from hangar import Repository
repo = Repository(path='myhangarrepo')
repo.init(user_name='Sherin Thomas', user_email='sherin@gmail.com', remove_old=True)
# generate data
data = []
for i in range(1000):
data.append(np.random.rand(28, 28))
data = np.array(data)
co = repo.checkout(write=True)
data_dset = co.datasets.init_dataset('mnist_data', prototype=data[0])
co.commit('datasets init')
co.close()
co = repo.checkout(write=True)
data_dset = co.datasets['mnist_data']
with data_dset:
for i in range(len(data)):
sample_name = str(i)
data_dset[sample_name] = data[i]
co.commit('dataset curation: stage 1') # this throws error
co.close()
```
**Expected behavior**
It should not break the program instead raise a warning about the performance hit
|
non_process
|
commit inside context manager throws runtimeerror describe the bug if we try to commit inside the context manager before exit hangar throws runtimeerror saying no changes made in the staging area cannot commit we should allow the user to do commits inside the context manager imo but probably with a warning about the performance hit severity select an option data corruption loss of any kind unexpected behavior exceptions or error thrown performance bottleneck to reproduce python import numpy as np from hangar import repository repo repository path myhangarrepo repo init user name sherin thomas user email sherin gmail com remove old true generate data data for i in range data append np random rand data np array data co repo checkout write true data dset co datasets init dataset mnist data prototype data co commit datasets init co close co repo checkout write true data dset co datasets with data dset for i in range len data sample name str i data dset data co commit dataset curation stage this throws error co close expected behavior it should not break the program instead raise a warning about the performance hit
| 0
|
35
| 2,505,372,297
|
IssuesEvent
|
2015-01-11 12:42:06
|
Graylog2/graylog2-server
|
https://api.github.com/repos/Graylog2/graylog2-server
|
closed
|
Ship a better/higher default configuration for output_batch_size
|
processing
|
We see significantly improved/less CPU usage by Elasticsearch when setting output_batch_size = 1000 or higher even when dealing with lower loads (around 250 msg/s). The default configuration value of 25 seems to be quite low if not too low, it might be good to raise it a bit. It probably would be also useful to have the batch size adjust dynamically based on relevant metric values.
|
1.0
|
Ship a better/higher default configuration for output_batch_size - We see significantly improved/less CPU usage by Elasticsearch when setting output_batch_size = 1000 or higher even when dealing with lower loads (around 250 msg/s). The default configuration value of 25 seems to be quite low if not too low, it might be good to raise it a bit. It probably would be also useful to have the batch size adjust dynamically based on relevant metric values.
|
process
|
ship a better higher default configuration for output batch size we see significantly improved less cpu usage by elasticsearch when setting output batch size or higher even when dealing with lower loads around msg s the default configuration value of seems to be quite low if not too low it might be good to raise it a bit it probably would be also useful to have the batch size adjust dynamically based on relevant metric values
| 1
|
19,059
| 25,078,181,879
|
IssuesEvent
|
2022-11-07 17:01:06
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
templateContext not well explained / exampled
|
doc-enhancement devops/prod Pri1 devops-cicd-process/tech
|
https://learn.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#use-templatecontext-to-pass-properties-to-templates
I’d never noticed this functionality before but I’m struggling to understand the use case. I don’t feel it’s hugely well explained as it really just shows an if switch.
I’ve done much the same type of thing as this example before with the if conditionals but it referenced a variable or parameters.
Why do I want to use templateContext over those methods of if-ing off a variable or parameter.
What advantage is it bringing to the table ?
I seen the original blog post and again don’t feel it gave a good example. Can anyone give me more examples ?
I posted this on azure devops on Reddit and the excellent Ming pointed out you can use this on top of a variable template but that still left me wondering why ?
there’s clearly something I’m missing as you’d not create this functionality without a need to fill.
thank you !
Much appreciated
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
templateContext not well explained / exampled - https://learn.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#use-templatecontext-to-pass-properties-to-templates
I’d never noticed this functionality before but I’m struggling to understand the use case. I don’t feel it’s hugely well explained as it really just shows an if switch.
I’ve done much the same type of thing as this example before with the if conditionals but it referenced a variable or parameters.
Why do I want to use templateContext over those methods of if-ing off a variable or parameter.
What advantage is it bringing to the table ?
I seen the original blog post and again don’t feel it gave a good example. Can anyone give me more examples ?
I posted this on azure devops on Reddit and the excellent Ming pointed out you can use this on top of a variable template but that still left me wondering why ?
there’s clearly something I’m missing as you’d not create this functionality without a need to fill.
thank you !
Much appreciated
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
templatecontext not well explained exampled i’d never noticed this functionality before but i’m struggling to understand the use case i don’t feel it’s hugely well explained as it really just shows an if switch i’ve done much the same type of thing as this example before with the if conditionals but it referenced a variable or parameters why do i want to use templatecontext over those methods of if ing off a variable or parameter what advantage is it bringing to the table i seen the original blog post and again don’t feel it gave a good example can anyone give me more examples i posted this on azure devops on reddit and the excellent ming pointed out you can use this on top of a variable template but that still left me wondering why there’s clearly something i’m missing as you’d not create this functionality without a need to fill thank you much appreciated document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id bbdc version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
214,324
| 7,268,889,980
|
IssuesEvent
|
2018-02-20 11:43:49
|
STEP-tw/battleship-phoenix
|
https://api.github.com/repos/STEP-tw/battleship-phoenix
|
closed
|
Game starts when both players are ready.
|
High Priority small
|
As a _player_
I want the _game to start_
So that I _can play_
**Additional Details**
1.Both players are ready.(assumption)
**Acceptance Criteria**
- [x] Criteria 1
- Given _opponent is ready and_
- When _I'm ready_
- Then _the game should start with a game started message_
|
1.0
|
Game starts when both players are ready. - As a _player_
I want the _game to start_
So that I _can play_
**Additional Details**
1.Both players are ready.(assumption)
**Acceptance Criteria**
- [x] Criteria 1
- Given _opponent is ready and_
- When _I'm ready_
- Then _the game should start with a game started message_
|
non_process
|
game starts when both players are ready as a player i want the game to start so that i can play additional details both players are ready assumption acceptance criteria criteria given opponent is ready and when i m ready then the game should start with a game started message
| 0
|
10,387
| 13,196,387,330
|
IssuesEvent
|
2020-08-13 20:32:59
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Triggers are not scheduled in UTC
|
Pri2 devops-cicd-process/tech devops/prod doc-bug investigating
|
Triggers are scheduled in the org's timezone, not UTC.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 2ea2c851-bd1e-cddc-b4d0-e9f4112b8565
* Version Independent ID: 07c23fdd-14b5-985b-1c63-3f26f3a216ad
* Content: [Configure schedules to run pipelines - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/scheduled-triggers?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/scheduled-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/scheduled-triggers.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @steved0x
* Microsoft Alias: **sdanie**
|
1.0
|
Triggers are not scheduled in UTC -
Triggers are scheduled in the org's timezone, not UTC.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 2ea2c851-bd1e-cddc-b4d0-e9f4112b8565
* Version Independent ID: 07c23fdd-14b5-985b-1c63-3f26f3a216ad
* Content: [Configure schedules to run pipelines - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/scheduled-triggers?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/scheduled-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/scheduled-triggers.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @steved0x
* Microsoft Alias: **sdanie**
|
process
|
triggers are not scheduled in utc triggers are scheduled in the org s timezone not utc document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id cddc version independent id content content source product devops technology devops cicd process github login microsoft alias sdanie
| 1
|
4,182
| 7,114,540,309
|
IssuesEvent
|
2018-01-18 01:18:11
|
sysown/proxysql
|
https://api.github.com/repos/sysown/proxysql
|
closed
|
Extract modifiers from comment
|
GLOBAL MYSQL PROTOCOL QUERY PROCESSOR cxx_pa development enhancement
|
Application should be able to send instructions and modify the behavior of the proxy using key/value pairs inside a comment.
We need to define the list of variables.
|
1.0
|
Extract modifiers from comment - Application should be able to send instructions and modify the behavior of the proxy using key/value pairs inside a comment.
We need to define the list of variables.
|
process
|
extract modifiers from comment application should be able to send instructions and modify the behavior of the proxy using key value pairs inside a comment we need to define the list of variables
| 1
|
15,976
| 20,188,183,795
|
IssuesEvent
|
2022-02-11 01:16:05
|
savitamittalmsft/WAS-SEC-TEST
|
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
|
opened
|
Scan container workloads for vulnerabilities
|
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Deployment & Testing Testing & Validation
|
<a href="https://docs.microsoft.com/azure/security-center/container-security">Scan container workloads for vulnerabilities</a>
<p><b>Why Consider This?</b></p>
To build secure containerized workloads, ensure the images that they're based on are free of known vulnerabilities. Shipping a known vulnerability in a container isn't significantly safer than on a VM.
<p><b>Context</b></p>
<p><span>Azure Security Center / Azure Defender is the Azure-native solution for securing containers. Azure Defender can protect virtual machines that are running Docker, Azure Kubernetes Service clusters, Azure Container Registry registries. Azure Defender is able to scan container images and identify security issues, or provide real-time threat detection for containerized environments.</span></p>
<p><b>Suggested Actions</b></p>
<p><span>Consider using Azure Defender for securing containerized workloads.</span></p>
<p><b>Learn More</b></p>
<p><a href="https://docs.microsoft.com/en-us/azure/security-center/container-security" target="_blank"><span>https://docs.microsoft.com/en-us/azure/security-center/container-security</span></a><span /></p>
|
1.0
|
Scan container workloads for vulnerabilities - <a href="https://docs.microsoft.com/azure/security-center/container-security">Scan container workloads for vulnerabilities</a>
<p><b>Why Consider This?</b></p>
To build secure containerized workloads, ensure the images that they're based on are free of known vulnerabilities. Shipping a known vulnerability in a container isn't significantly safer than on a VM.
<p><b>Context</b></p>
<p><span>Azure Security Center / Azure Defender is the Azure-native solution for securing containers. Azure Defender can protect virtual machines that are running Docker, Azure Kubernetes Service clusters, Azure Container Registry registries. Azure Defender is able to scan container images and identify security issues, or provide real-time threat detection for containerized environments.</span></p>
<p><b>Suggested Actions</b></p>
<p><span>Consider using Azure Defender for securing containerized workloads.</span></p>
<p><b>Learn More</b></p>
<p><a href="https://docs.microsoft.com/en-us/azure/security-center/container-security" target="_blank"><span>https://docs.microsoft.com/en-us/azure/security-center/container-security</span></a><span /></p>
|
process
|
scan container workloads for vulnerabilities why consider this to build secure containerized workloads ensure the images that they re based on are free of known vulnerabilities shipping a known vulnerability in a container isn t significantly safer than on a vm context azure security center azure defender is the azure native solution for securing containers azure defender can protect virtual machines that are running docker azure kubernetes service clusters azure container registry registries azure defender is able to scan container images and identify security issues or provide real time threat detection for containerized environments suggested actions consider using azure defender for securing containerized workloads learn more
| 1
|
82,383
| 10,279,473,812
|
IssuesEvent
|
2019-08-25 23:33:54
|
fga-desenho-2019-2/Wiki
|
https://api.github.com/repos/fga-desenho-2019-2/Wiki
|
opened
|
API de pagamento
|
Iniciativa extra Pesquisas documentation
|
## Descrição da Issue
Procurar API's de pagamento e verificar como elas funcionam e precificação
### Tasks:
- [ ] Listar API's de pagamento
- [ ] Descrever o funcionamento delas e precificação
|
1.0
|
API de pagamento - ## Descrição da Issue
Procurar API's de pagamento e verificar como elas funcionam e precificação
### Tasks:
- [ ] Listar API's de pagamento
- [ ] Descrever o funcionamento delas e precificação
|
non_process
|
api de pagamento descrição da issue procurar api s de pagamento e verificar como elas funcionam e precificação tasks listar api s de pagamento descrever o funcionamento delas e precificação
| 0
|
15,874
| 20,049,635,950
|
IssuesEvent
|
2022-02-03 03:43:33
|
q191201771/lal
|
https://api.github.com/repos/q191201771/lal
|
closed
|
开启 hls 并启用 use_memory_as_disk_flag 后,cleanup_mode 为0或者1时,会出现内存持续增长
|
#Question #Need doc *In process *Waiting reply
|
开启 hls 并启用 use_memory_as_disk_flag 后,m3u8 和 ts 会存储在内存中
当 cleanup_mode 为1或者2时,不会持续删除过期的 ts 文件,导致内存持续增长
建议在手册的 lalserver 配置文件说明 中注明
|
1.0
|
开启 hls 并启用 use_memory_as_disk_flag 后,cleanup_mode 为0或者1时,会出现内存持续增长 - 开启 hls 并启用 use_memory_as_disk_flag 后,m3u8 和 ts 会存储在内存中
当 cleanup_mode 为1或者2时,不会持续删除过期的 ts 文件,导致内存持续增长
建议在手册的 lalserver 配置文件说明 中注明
|
process
|
开启 hls 并启用 use memory as disk flag 后,cleanup mode ,会出现内存持续增长 开启 hls 并启用 use memory as disk flag 后, 和 ts 会存储在内存中 当 cleanup mode ,不会持续删除过期的 ts 文件,导致内存持续增长 建议在手册的 lalserver 配置文件说明 中注明
| 1
|
60,357
| 7,333,143,774
|
IssuesEvent
|
2018-03-05 18:27:45
|
juliett-golf-hotel/web-app
|
https://api.github.com/repos/juliett-golf-hotel/web-app
|
opened
|
Add hero banner to home page
|
content design dev
|
- [ ] Add hero banner images for different screen sizes
- [ ] Add current temperature
- [ ] Add feels like temperature
- [ ] Add what the weather is (partly cloudy, hail, sun showers, etc)
- [ ] Add a greeting
|
1.0
|
Add hero banner to home page - - [ ] Add hero banner images for different screen sizes
- [ ] Add current temperature
- [ ] Add feels like temperature
- [ ] Add what the weather is (partly cloudy, hail, sun showers, etc)
- [ ] Add a greeting
|
non_process
|
add hero banner to home page add hero banner images for different screen sizes add current temperature add feels like temperature add what the weather is partly cloudy hail sun showers etc add a greeting
| 0
|
1,324
| 3,874,111,167
|
IssuesEvent
|
2016-04-11 19:20:28
|
neuropoly/spinalcordtoolbox
|
https://api.github.com/repos/neuropoly/spinalcordtoolbox
|
closed
|
IndexError for computing CSA
|
bug sct_process_segmentation
|
data:
dropbox simon/results/2015-12-14
~~~
sct_process_segmentation -i T2_seg.nii -p csa -t label/template/MNI-Poly-AMU_level.nii.gz -vert 1:1
Check parameters:
.. segmentation file: T2_seg.nii
Create temporary folder...
mkdir tmp.151214225906_621711/
Copying input data to tmp folder and convert to nii...
sct_convert -i /Users/julien/data/biospective/2015-12-14_sct2.1-2015-12-14/100-011_s2_T2/T2_seg.nii -o tmp.151214225906_621711/segmentation.nii.gz
Change orientation to RPI...
sct_image -i segmentation.nii.gz -setorient RPI -o segmentation_RPI.nii.gz
Open segmentation volume...
Get data dimensions...
56 x 288 x 288
Smooth centerline/segmentation...
.. Get center of mass of the centerline/segmentation...
.. Smoothing algo = hanning
.. Windows length = 50
Compute CSA...
Traceback (most recent call last):
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 815, in <module>
main(sys.argv[1:])
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 235, in main
compute_csa(fname_segmentation, verbose, remove_temp_files, step, smoothing_param, figure_fit, param.file_csa_volume, slices, vert_lev, fname_vertebral_labeling, algo_fitting = param.algo_fitting, type_window= param.type_window, window_length=param.window_length)
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 487, in compute_csa
normal = normalize(np.array([x_centerline_deriv[iz-min_z_index], y_centerline_deriv[iz-min_z_index], z_centerline_deriv[iz-min_z_index]]))
IndexError: index 278 is out of bounds for axis 0 with size 278
~~~
|
1.0
|
IndexError for computing CSA - data:
dropbox simon/results/2015-12-14
~~~
sct_process_segmentation -i T2_seg.nii -p csa -t label/template/MNI-Poly-AMU_level.nii.gz -vert 1:1
Check parameters:
.. segmentation file: T2_seg.nii
Create temporary folder...
mkdir tmp.151214225906_621711/
Copying input data to tmp folder and convert to nii...
sct_convert -i /Users/julien/data/biospective/2015-12-14_sct2.1-2015-12-14/100-011_s2_T2/T2_seg.nii -o tmp.151214225906_621711/segmentation.nii.gz
Change orientation to RPI...
sct_image -i segmentation.nii.gz -setorient RPI -o segmentation_RPI.nii.gz
Open segmentation volume...
Get data dimensions...
56 x 288 x 288
Smooth centerline/segmentation...
.. Get center of mass of the centerline/segmentation...
.. Smoothing algo = hanning
.. Windows length = 50
Compute CSA...
Traceback (most recent call last):
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 815, in <module>
main(sys.argv[1:])
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 235, in main
compute_csa(fname_segmentation, verbose, remove_temp_files, step, smoothing_param, figure_fit, param.file_csa_volume, slices, vert_lev, fname_vertebral_labeling, algo_fitting = param.algo_fitting, type_window= param.type_window, window_length=param.window_length)
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 487, in compute_csa
normal = normalize(np.array([x_centerline_deriv[iz-min_z_index], y_centerline_deriv[iz-min_z_index], z_centerline_deriv[iz-min_z_index]]))
IndexError: index 278 is out of bounds for axis 0 with size 278
~~~
|
process
|
indexerror for computing csa data dropbox simon results sct process segmentation i seg nii p csa t label template mni poly amu level nii gz vert check parameters segmentation file seg nii create temporary folder mkdir tmp copying input data to tmp folder and convert to nii sct convert i users julien data biospective seg nii o tmp segmentation nii gz change orientation to rpi sct image i segmentation nii gz setorient rpi o segmentation rpi nii gz open segmentation volume get data dimensions x x smooth centerline segmentation get center of mass of the centerline segmentation smoothing algo hanning windows length compute csa traceback most recent call last file users julien code spinalcordtoolbox bin sct process segmentation line in main sys argv file users julien code spinalcordtoolbox bin sct process segmentation line in main compute csa fname segmentation verbose remove temp files step smoothing param figure fit param file csa volume slices vert lev fname vertebral labeling algo fitting param algo fitting type window param type window window length param window length file users julien code spinalcordtoolbox bin sct process segmentation line in compute csa normal normalize np array y centerline deriv z centerline deriv indexerror index is out of bounds for axis with size
| 1
|
3,535
| 6,572,687,732
|
IssuesEvent
|
2017-09-11 04:25:42
|
zero-os/0-Disk
|
https://api.github.com/repos/zero-os/0-Disk
|
closed
|
add clusterID to ardb (failure) 0-log messages
|
process_wontfix type_feature
|
Currently only the address and db of the ardb storage server is given, but it would also be useful to give its clusterID alongside [ardb-storage-server-issues](https://github.com/zero-os/0-Disk/blob/master/docs/log.md#ardb-storage-server-issues) messages.
|
1.0
|
add clusterID to ardb (failure) 0-log messages - Currently only the address and db of the ardb storage server is given, but it would also be useful to give its clusterID alongside [ardb-storage-server-issues](https://github.com/zero-os/0-Disk/blob/master/docs/log.md#ardb-storage-server-issues) messages.
|
process
|
add clusterid to ardb failure log messages currently only the address and db of the ardb storage server is given but it would also be useful to give its clusterid alongside messages
| 1
|
17,763
| 23,691,657,953
|
IssuesEvent
|
2022-08-29 11:22:56
|
ankidroid/Anki-Android
|
https://api.github.com/repos/ankidroid/Anki-Android
|
opened
|
Make code coverage reporting stable
|
Dev Test process
|
`codecov` often reports random fluctuations in the coverage. This is likely because tests are nondeterministic, but may be a bug in codecov.
* Determine why this occurs (typically by finding the classes which fluctuate)
* Fix this, so we no longer get `codecov`-based CI failures on changes which should be no-ops
|
1.0
|
Make code coverage reporting stable - `codecov` often reports random fluctuations in the coverage. This is likely because tests are nondeterministic, but may be a bug in codecov.
* Determine why this occurs (typically by finding the classes which fluctuate)
* Fix this, so we no longer get `codecov`-based CI failures on changes which should be no-ops
|
process
|
make code coverage reporting stable codecov often reports random fluctuations in the coverage this is likely because tests are nondeterministic but may be a bug in codecov determine why this occurs typically by finding the classes which fluctuate fix this so we no longer get codecov based ci failures on changes which should be no ops
| 1
|
75,081
| 3,455,052,267
|
IssuesEvent
|
2015-12-17 18:26:21
|
ThoughtWorksInc/registrolivre
|
https://api.github.com/repos/ThoughtWorksInc/registrolivre
|
closed
|
Bug no highlight do formulário de cadastro de empresa
|
Bug Priority
|
Bug no highlight do formulário de cadastro de empresas.
Ao tentar adicionar sócios na empresa, caso um CPF esteja errado o highlight aparece nos dois formulários mesmo que o segundo CPF esteja correto.

|
1.0
|
Bug no highlight do formulário de cadastro de empresa - Bug no highlight do formulário de cadastro de empresas.
Ao tentar adicionar sócios na empresa, caso um CPF esteja errado o highlight aparece nos dois formulários mesmo que o segundo CPF esteja correto.

|
non_process
|
bug no highlight do formulário de cadastro de empresa bug no highlight do formulário de cadastro de empresas ao tentar adicionar sócios na empresa caso um cpf esteja errado o highlight aparece nos dois formulários mesmo que o segundo cpf esteja correto
| 0
|
18,775
| 24,678,010,403
|
IssuesEvent
|
2022-10-18 18:41:02
|
hashgraph/hedera-mirror-node
|
https://api.github.com/repos/hashgraph/hedera-mirror-node
|
closed
|
Monitor slow to startup in Kubernetes
|
bug process monitor
|
### Description
The monitor is now taking too long to startup that its liveness probe is causing it to restart.
### Steps to reproduce
Run monitor on Kubernetes
### Additional context
_No response_
### Hedera network
other
### Version
main
### Operating system
_No response_
|
1.0
|
Monitor slow to startup in Kubernetes - ### Description
The monitor is now taking too long to startup that its liveness probe is causing it to restart.
### Steps to reproduce
Run monitor on Kubernetes
### Additional context
_No response_
### Hedera network
other
### Version
main
### Operating system
_No response_
|
process
|
monitor slow to startup in kubernetes description the monitor is now taking too long to startup that its liveness probe is causing it to restart steps to reproduce run monitor on kubernetes additional context no response hedera network other version main operating system no response
| 1
|
311,218
| 23,376,086,577
|
IssuesEvent
|
2022-08-11 03:19:10
|
singularity-data/risingwave-docs
|
https://api.github.com/repos/singularity-data/risingwave-docs
|
closed
|
Update the lookup join behavior
|
documentation
|
### Related code PR
PR: https://github.com/singularity-data/risingwave/pull/4207
Code issue: https://github.com/singularity-data/risingwave/issues/4044
### Which part(s) of the docs might be affected or should be updated? And how?
Add this code change (issue fixes) to Release Notes.
When using lookup joins (JOIN ... ON ...), RW now will cast the data types if the data type of the two key columns are different, instead of throwing an error.
### Reference
_No response_
|
1.0
|
Update the lookup join behavior - ### Related code PR
PR: https://github.com/singularity-data/risingwave/pull/4207
Code issue: https://github.com/singularity-data/risingwave/issues/4044
### Which part(s) of the docs might be affected or should be updated? And how?
Add this code change (issue fixes) to Release Notes.
When using lookup joins (JOIN ... ON ...), RW now will cast the data types if the data type of the two key columns are different, instead of throwing an error.
### Reference
_No response_
|
non_process
|
update the lookup join behavior related code pr pr code issue which part s of the docs might be affected or should be updated and how add this code change issue fixes to release notes when using lookup joins join on rw now will cast the data types if the data type of the two key columns are different instead of throwing an error reference no response
| 0
|
82,645
| 10,300,605,138
|
IssuesEvent
|
2019-08-28 16:14:13
|
ISibboI/graphrepresentations
|
https://api.github.com/repos/ISibboI/graphrepresentations
|
opened
|
Fix documentation asserts
|
documentation
|
Asserting that the node ids are output in order does not prove anything or help the user understand. It should be asserted that the data is at the correct place.
|
1.0
|
Fix documentation asserts - Asserting that the node ids are output in order does not prove anything or help the user understand. It should be asserted that the data is at the correct place.
|
non_process
|
fix documentation asserts asserting that the node ids are output in order does not prove anything or help the user understand it should be asserted that the data is at the correct place
| 0
|
11,315
| 14,134,856,491
|
IssuesEvent
|
2020-11-10 00:15:56
|
googleapis/python-speech
|
https://api.github.com/repos/googleapis/python-speech
|
closed
|
speec2srt.py is not working because of removal of submodules enums and types
|
api: speech type: process
|
Following [this](https://www.youtube.com/watch?v=uBzp5xGSZ6o&ab_channel=GoogleCloudAPAC) for creating speech to srt using google apis. When running the command `python3 speech2srt.py --storage_uri gs://subtitlingsc/en.wav` is showing that
```
File "speech2srt.py", line 19, in <module>
from google.cloud.speech_v1 import enums
ImportError: cannot import name 'enums' from 'google.cloud.speech_v1'
```
I think this is due to removal of enums and types. Please do the needful and update those videos and [GitHub file](https://github.com/GoogleCloudPlatform/community/blob/master/tutorials/speech2srt/speech2srt.py)
|
1.0
|
speec2srt.py is not working because of removal of submodules enums and types - Following [this](https://www.youtube.com/watch?v=uBzp5xGSZ6o&ab_channel=GoogleCloudAPAC) for creating speech to srt using google apis. When running the command `python3 speech2srt.py --storage_uri gs://subtitlingsc/en.wav` is showing that
```
File "speech2srt.py", line 19, in <module>
from google.cloud.speech_v1 import enums
ImportError: cannot import name 'enums' from 'google.cloud.speech_v1'
```
I think this is due to removal of enums and types. Please do the needful and update those videos and [GitHub file](https://github.com/GoogleCloudPlatform/community/blob/master/tutorials/speech2srt/speech2srt.py)
|
process
|
py is not working because of removal of submodules enums and types following for creating speech to srt using google apis when running the command py storage uri gs subtitlingsc en wav is showing that file py line in from google cloud speech import enums importerror cannot import name enums from google cloud speech i think this is due to removal of enums and types please do the needful and update those videos and
| 1
|
136,030
| 30,462,136,059
|
IssuesEvent
|
2023-07-17 07:47:25
|
FerretDB/FerretDB
|
https://api.github.com/repos/FerretDB/FerretDB
|
opened
|
Integration tests should report their progress
|
code/chore not ready
|
### What should be done?
The integration tests should print something meaningful to the console once in a while, like "running test 24/1032".
Right now they can take 10-30 minutes on a dev machine without printing a thing. The terminal looks like it hangs, there is no indication of progress. A natural reaction to that is `ctrl+C` (which doesn't allow the tests to complete).
It could be a good idea to disable this progress printing when running the tests from CI (to avoid some redundant logs in a non-interactive environment).
### Where?
The integration tests runner maybe?
I haven't looked at the relevant test runner yet. :swe
### Definition of Done
- test runner is updated
|
1.0
|
Integration tests should report their progress - ### What should be done?
The integration tests should print something meaningful to the console once in a while, like "running test 24/1032".
Right now they can take 10-30 minutes on a dev machine without printing a thing. The terminal looks like it hangs, there is no indication of progress. A natural reaction to that is `ctrl+C` (which doesn't allow the tests to complete).
It could be a good idea to disable this progress printing when running the tests from CI (to avoid some redundant logs in a non-interactive environment).
### Where?
The integration tests runner maybe?
I haven't looked at the relevant test runner yet. :swe
### Definition of Done
- test runner is updated
|
non_process
|
integration tests should report their progress what should be done the integration tests should print something meaningful to the console once in a while like running test right now they can take minutes on a dev machine without printing a thing the terminal looks like it hangs there is no indication of progress a natural reaction to that is ctrl c which doesn t allow the tests to complete it could be a good idea to disable this progress printing when running the tests from ci to avoid some redundant logs in a non interactive environment where the integration tests runner maybe i haven t looked at the relevant test runner yet swe definition of done test runner is updated
| 0
|
11,304
| 14,107,274,324
|
IssuesEvent
|
2020-11-06 16:03:57
|
panther-labs/panther
|
https://api.github.com/repos/panther-labs/panther
|
opened
|
Invalid JSON in test crashes entire test suite
|
bug p1 team:data processing
|
### Describe the bug
Currently, when adding an invalid input as a test body, I expect to get back a structured response including a particular set of keys (like `ruleError` and `genericError`), stating that I have added invalid content.
Instead, the entire lambda fails, causes a runtime issue and AppSync doesn't return any structured data.
### Steps to reproduce
Steps to reproduce the behavior:
1. Create a New Rule
2. Add a test
3. Fill the body of the test with the content
```
{ "uuid": "123" }
{ "uuid": "123" }
```
4. Click "Run all"
### Expected behavior
There should be a structured API response with a specific key containing the error
### Environment
How are you deploying or using Panther?
- Panther version or commit: v1.12
- OS: Mac
- Browser: Chrome
### Additional context
If I have 20 tests and 1 of them has invalid body, I will never get feedback for **any of the other 19 tests**, since the entire API response gets replaced by the runtime error
### Screenshots
<img width="1287" alt="Screen Shot 2020-11-06 at 5 55 55 PM" src="https://user-images.githubusercontent.com/10436045/98387474-19630f80-205a-11eb-88ed-5c358783b789.png">
|
1.0
|
Invalid JSON in test crashes entire test suite -
### Describe the bug
Currently, when adding an invalid input as a test body, I expect to get back a structured response including a particular set of keys (like `ruleError` and `genericError`), stating that I have added invalid content.
Instead, the entire lambda fails, causes a runtime issue and AppSync doesn't return any structured data.
### Steps to reproduce
Steps to reproduce the behavior:
1. Create a New Rule
2. Add a test
3. Fill the body of the test with the content
```
{ "uuid": "123" }
{ "uuid": "123" }
```
4. Click "Run all"
### Expected behavior
There should be a structured API response with a specific key containing the error
### Environment
How are you deploying or using Panther?
- Panther version or commit: v1.12
- OS: Mac
- Browser: Chrome
### Additional context
If I have 20 tests and 1 of them has invalid body, I will never get feedback for **any of the other 19 tests**, since the entire API response gets replaced by the runtime error
### Screenshots
<img width="1287" alt="Screen Shot 2020-11-06 at 5 55 55 PM" src="https://user-images.githubusercontent.com/10436045/98387474-19630f80-205a-11eb-88ed-5c358783b789.png">
|
process
|
invalid json in test crashes entire test suite describe the bug currently when adding an invalid input as a test body i expect to get back a structured response including a particular set of keys like ruleerror and genericerror stating that i have added invalid content instead the entire lambda fails causes a runtime issue and appsync doesn t return any structured data steps to reproduce steps to reproduce the behavior create a new rule add a test fill the body of the test with the content uuid uuid click run all expected behavior there should be a structured api response with a specific key containing the error environment how are you deploying or using panther panther version or commit os mac browser chrome additional context if i have tests and of them has invalid body i will never get feedback for any of the other tests since the entire api response gets replaced by the runtime error screenshots img width alt screen shot at pm src
| 1
|
20,590
| 27,254,450,947
|
IssuesEvent
|
2023-02-22 10:27:51
|
python/cpython
|
https://api.github.com/repos/python/cpython
|
reopened
|
Deadlock when using fork whilst multiprocessing.resource_tracker._resource_tracker._lock is held
|
type-bug stdlib expert-multiprocessing
|
# Bug report
Given the following situation: You have a Process "a" with two threads ("aa", "ab"). aa is currently creating a shared memory segment and is holding `multiprocessing.resource_tracker._resource_tracker._lock` for that reason. Note that that lock is not acquired by user code, but deep inside the SharedMemory constructor.
ab now wants to create a second process using the `fork` method. Process b is created with one thread: bb. As per `man fork` the thread aa is not duplicated in process b.
If aa now finishes creating its SharedMemory it releases `multiprocessing.resource_tracker._resource_tracker._lock`. But that lock is only a `threading.Lock`, which means it is released only in process a. In process b it will not be released. If b now wants to create a SharedMemory object and therefore tries to acquire `multiprocessing.resource_tracker._resource_tracker._lock` this will take forever, as it will never be free in process b. `man fork` explicitly mentions that such situations are a potential source of issues.
As this is a race condition there is no minimal, reproducible example I can give.
# Possible solutions
One could replace this `threading.Lock` with a `multiprocessing.Lock`. Alternatively one could replace that lock with a new one in the child process after a fork. Im not sure about the intended behavior of the resource tracker in such situations.
# Your environment
<!-- Include as many relevant details as possible about the environment you experienced the bug in -->
- CPython versions tested on: 3.9, 3.10
- Operating system and architecture: Unix (Arch, Kernel 5.19.7-1)
<!--
You can freely edit this text. Remove any lines you believe are unnecessary.
-->
|
1.0
|
Deadlock when using fork whilst multiprocessing.resource_tracker._resource_tracker._lock is held - # Bug report
Given the following situation: You have a Process "a" with two threads ("aa", "ab"). aa is currently creating a shared memory segment and is holding `multiprocessing.resource_tracker._resource_tracker._lock` for that reason. Note that that lock is not acquired by user code, but deep inside the SharedMemory constructor.
ab now wants to create a second process using the `fork` method. Process b is created with one thread: bb. As per `man fork` the thread aa is not duplicated in process b.
If aa now finishes creating its SharedMemory it releases `multiprocessing.resource_tracker._resource_tracker._lock`. But that lock is only a `threading.Lock`, which means it is released only in process a. In process b it will not be released. If b now wants to create a SharedMemory object and therefore tries to acquire `multiprocessing.resource_tracker._resource_tracker._lock` this will take forever, as it will never be free in process b. `man fork` explicitly mentions that such situations are a potential source of issues.
As this is a race condition there is no minimal, reproducible example I can give.
# Possible solutions
One could replace this `threading.Lock` with a `multiprocessing.Lock`. Alternatively one could replace that lock with a new one in the child process after a fork. Im not sure about the intended behavior of the resource tracker in such situations.
# Your environment
<!-- Include as many relevant details as possible about the environment you experienced the bug in -->
- CPython versions tested on: 3.9, 3.10
- Operating system and architecture: Unix (Arch, Kernel 5.19.7-1)
<!--
You can freely edit this text. Remove any lines you believe are unnecessary.
-->
|
process
|
deadlock when using fork whilst multiprocessing resource tracker resource tracker lock is held bug report given the following situation you have a process a with two threads aa ab aa is currently creating a shared memory segment and is holding multiprocessing resource tracker resource tracker lock for that reason note that that lock is not acquired by user code but deep inside the sharedmemory constructor ab now wants to create a second process using the fork method process b is created with one thread bb as per man fork the thread aa is not duplicated in process b if aa now finishes creating its sharedmemory it releases multiprocessing resource tracker resource tracker lock but that lock is only a threading lock which means it is released only in process a in process b it will not be released if b now wants to create a sharedmemory object and therefore tries to acquire multiprocessing resource tracker resource tracker lock this will take forever as it will never be free in process b man fork explicitly mentions that such situations are a potential source of issues as this is a race condition there is no minimal reproducible example i can give possible solutions one could replace this threading lock with a multiprocessing lock alternatively one could replace that lock with a new one in the child process after a fork im not sure about the intended behavior of the resource tracker in such situations your environment cpython versions tested on operating system and architecture unix arch kernel you can freely edit this text remove any lines you believe are unnecessary
| 1
|
5,475
| 8,354,087,548
|
IssuesEvent
|
2018-10-02 12:18:23
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
copy-to ignored in preprocess2
|
P1 bug preprocess2
|
## Expected Behavior
Using `@copy-to` should result in an extra copy of the referenced topic. If file names are relevant for output, the specified names will be used.
## Actual Behavior
The `@copy-to` attribute is not processed, so a copy is not created.
In PDF output, this results in a broken link and missing file. There are several errors when different steps try to read the copy that does not exist. In 3.1.2 this ends in a build failure caused by the missing file; #3082 removes the build failure and replaces it with an error like the other steps.
## Possible Solution
The error comes from this line: https://github.com/dita-ot/dita-ot/blob/develop/src/main/java/org/dita/dost/module/CopyToModule.java#L148
In an HTML build with `preprocess`, the file listed in `.job.xml` does not have `@src` until after `copy-to` is evaluated. That listing immediately before copy-to:
`<file uri="copyto.dita" path="copyto.dita" result="file:/C:/DITA-OT/TESTGITHUB/copyto/copyto.dita">`
and after:
`<file src="file:/C:/DITA-OT/TESTGITHUB/copyto/original.dita" uri="copyto.dita" path="copyto.dita" result="file:/C:/DITA-OT/TESTGITHUB/copyto/copyto.dita" format="dita" target="true">`
In a PDF build with `preprocess2`, that same file has `@src` after the first step (`map-reader`). That line in `preprocess2` assumes that if `@src` is listed, the file was referenced as a source file, so we shouldn't copy anything to it. That listing before copy-to:
`<file src="file:/C:/DITA-OT/TESTGITHUB/copyto/original.dita" uri="f7ac72d7ff9763eb29f90236b7c20b67fce3a84a.dita" path="f7ac72d7ff9763eb29f90236b7c20b67fce3a84a.dita" result="file:/C:/DITA-OT/TESTGITHUB/copyto/original.dita" format="dita" target="true">
</file>`
I'm not yet sure where in `map-reader` the `@src` attribute is set up, otherwise this would be coming in as a pull request. I _suspect_ that if we just leave off `@src` for the copies at this point, it will give the correct results, but I haven't been able to test that so I'm not sure if it would affect other map steps.
## Steps to Reproduce
Build this to PDF - it will result in the build failures with 3.1.2, but you can see in temp that the copies are not created, and other steps will warn when trying to access them. The map has commented out copies of the `<files>` section from HTML (before/after copy-to runs), and PDF (before `copy-to` runs). With the fix for #3082 in place, it will produce a PDF that has 1 topic instead of 3.
[copyto.zip](https://github.com/dita-ot/dita-ot/files/2429029/copyto.zip)
## Environment
<!-- Include relevant details about the environment you experienced this in. -->
* DITA-OT version: 3.1.2 and `hotfix/3.1.3`
* Operating system and version: Windows
* How did you run DITA-OT? `dita`
* Transformation type: PDF
|
1.0
|
copy-to ignored in preprocess2 - ## Expected Behavior
Using `@copy-to` should result in an extra copy of the referenced topic. If file names are relevant for output, the specified names will be used.
## Actual Behavior
The `@copy-to` attribute is not processed, so a copy is not created.
In PDF output, this results in a broken link and missing file. There are several errors when different steps try to read the copy that does not exist. In 3.1.2 this ends in a build failure caused by the missing file; #3082 removes the build failure and replaces it with an error like the other steps.
## Possible Solution
The error comes from this line: https://github.com/dita-ot/dita-ot/blob/develop/src/main/java/org/dita/dost/module/CopyToModule.java#L148
In an HTML build with `preprocess`, the file listed in `.job.xml` does not have `@src` until after `copy-to` is evaluated. That listing immediately before copy-to:
`<file uri="copyto.dita" path="copyto.dita" result="file:/C:/DITA-OT/TESTGITHUB/copyto/copyto.dita">`
and after:
`<file src="file:/C:/DITA-OT/TESTGITHUB/copyto/original.dita" uri="copyto.dita" path="copyto.dita" result="file:/C:/DITA-OT/TESTGITHUB/copyto/copyto.dita" format="dita" target="true">`
In a PDF build with `preprocess2`, that same file has `@src` after the first step (`map-reader`). That line in `preprocess2` assumes that if `@src` is listed, the file was referenced as a source file, so we shouldn't copy anything to it. That listing before copy-to:
`<file src="file:/C:/DITA-OT/TESTGITHUB/copyto/original.dita" uri="f7ac72d7ff9763eb29f90236b7c20b67fce3a84a.dita" path="f7ac72d7ff9763eb29f90236b7c20b67fce3a84a.dita" result="file:/C:/DITA-OT/TESTGITHUB/copyto/original.dita" format="dita" target="true">
</file>`
I'm not yet sure where in `map-reader` the `@src` attribute is set up, otherwise this would be coming in as a pull request. I _suspect_ that if we just leave off `@src` for the copies at this point, it will give the correct results, but I haven't been able to test that so I'm not sure if it would affect other map steps.
## Steps to Reproduce
Build this to PDF - it will result in the build failures with 3.1.2, but you can see in temp that the copies are not created, and other steps will warn when trying to access them. The map has commented out copies of the `<files>` section from HTML (before/after copy-to runs), and PDF (before `copy-to` runs). With the fix for #3082 in place, it will produce a PDF that has 1 topic instead of 3.
[copyto.zip](https://github.com/dita-ot/dita-ot/files/2429029/copyto.zip)
## Environment
<!-- Include relevant details about the environment you experienced this in. -->
* DITA-OT version: 3.1.2 and `hotfix/3.1.3`
* Operating system and version: Windows
* How did you run DITA-OT? `dita`
* Transformation type: PDF
|
process
|
copy to ignored in expected behavior using copy to should result in an extra copy of the referenced topic if file names are relevant for output the specified names will be used actual behavior the copy to attribute is not processed so a copy is not created in pdf output this results in a broken link and missing file there are several errors when different steps try to read the copy that does not exist in this ends in a build failure caused by the missing file removes the build failure and replaces it with an error like the other steps possible solution the error comes from this line in an html build with preprocess the file listed in job xml does not have src until after copy to is evaluated that listing immediately before copy to and after in a pdf build with that same file has src after the first step map reader that line in assumes that if src is listed the file was referenced as a source file so we shouldn t copy anything to it that listing before copy to i m not yet sure where in map reader the src attribute is set up otherwise this would be coming in as a pull request i suspect that if we just leave off src for the copies at this point it will give the correct results but i haven t been able to test that so i m not sure if it would affect other map steps steps to reproduce build this to pdf it will result in the build failures with but you can see in temp that the copies are not created and other steps will warn when trying to access them the map has commented out copies of the section from html before after copy to runs and pdf before copy to runs with the fix for in place it will produce a pdf that has topic instead of environment dita ot version and hotfix operating system and version windows how did you run dita ot dita transformation type pdf
| 1
|
62,481
| 6,797,828,979
|
IssuesEvent
|
2017-11-02 01:18:31
|
brave/browser-laptop
|
https://api.github.com/repos/brave/browser-laptop
|
reopened
|
Fingerprinting protections bypassable
|
feature/shields privacy QA/checked-Linux QA/checked-macOS QA/checked-Win64 QA/test-plan-specified release-notes/include release/blocking security
|
## Test plan
https://github.com/brave/browser-laptop/issues/11683#issuecomment-340532797
----
### Description
Many of the current fingerprinting protections rely on removing references to methods from the global object. However, you can get alternate references to the same methods by inserting a configured iframe element and grabbing the references off `iframe.contentWindow` or `iframe.contentDocument`.
### Steps to Reproduce
```
//blocked
console.log(HTMLCanvasElement.prototype.toDataURL);
let iframe = document.createElement("iframe");
iframe.src = "https://www.brave.com/";
document.body.appendChild(iframe);
// Not blocked
console.log(iframe.contentWindow.HTMLCanvasElement.prototype.toDataURL);
// Do the bad stuff
let canvasToDataUrl = iframe.contentWindow.HTMLCanvasElement.prototype.toDataURL;
canvasToDataUrl.apply(someParentFrameCanvasElement);
```
**Actual result:**
You can get references to the fingerprinting related methods.
**Expected result:**
You shouldn't be able to get to these methods.
**Reproduces how often:** 100%
### Brave Version
All versions
**Reproducible on current live release:**
Yes
|
1.0
|
Fingerprinting protections bypassable - ## Test plan
https://github.com/brave/browser-laptop/issues/11683#issuecomment-340532797
----
### Description
Many of the current fingerprinting protections rely on removing references to methods from the global object. However, you can get alternate references to the same methods by inserting a configured iframe element and grabbing the references off `iframe.contentWindow` or `iframe.contentDocument`.
### Steps to Reproduce
```
//blocked
console.log(HTMLCanvasElement.prototype.toDataURL);
let iframe = document.createElement("iframe");
iframe.src = "https://www.brave.com/";
document.body.appendChild(iframe);
// Not blocked
console.log(iframe.contentWindow.HTMLCanvasElement.prototype.toDataURL);
// Do the bad stuff
let canvasToDataUrl = iframe.contentWindow.HTMLCanvasElement.prototype.toDataURL;
canvasToDataUrl.apply(someParentFrameCanvasElement);
```
**Actual result:**
You can get references to the fingerprinting related methods.
**Expected result:**
You shouldn't be able to get to these methods.
**Reproduces how often:** 100%
### Brave Version
All versions
**Reproducible on current live release:**
Yes
|
non_process
|
fingerprinting protections bypassable test plan description many of the current fingerprinting protections rely on removing references to methods from the global object however you can get alternate references to the same methods by inserting a configured iframe element and grabbing the references off iframe contentwindow or iframe contentdocument steps to reproduce blocked console log htmlcanvaselement prototype todataurl let iframe document createelement iframe iframe src document body appendchild iframe not blocked console log iframe contentwindow htmlcanvaselement prototype todataurl do the bad stuff let canvastodataurl iframe contentwindow htmlcanvaselement prototype todataurl canvastodataurl apply someparentframecanvaselement actual result you can get references to the fingerprinting related methods expected result you shouldn t be able to get to these methods reproduces how often brave version all versions reproducible on current live release yes
| 0
|
16,594
| 21,645,947,690
|
IssuesEvent
|
2022-05-06 01:54:55
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
k8sattributes processor: support flexible labels/annotation regex mapping
|
proc: k8sprocessor
|
[This PR](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/5780) adds ability to fetch pod/namespace annotations and labels based on regex. Attribute values created as the result have only one predefined form `k8s.pod(namespace).label.$0`.
This issue is to track a follow up improvement to support more flexible construction of the attribute values as suggested in [this comment](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/5780#discussion_r732390230)
|
1.0
|
k8sattributes processor: support flexible labels/annotation regex mapping - [This PR](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/5780) adds ability to fetch pod/namespace annotations and labels based on regex. Attribute values created as the result have only one predefined form `k8s.pod(namespace).label.$0`.
This issue is to track a follow up improvement to support more flexible construction of the attribute values as suggested in [this comment](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/5780#discussion_r732390230)
|
process
|
processor support flexible labels annotation regex mapping adds ability to fetch pod namespace annotations and labels based on regex attribute values created as the result have only one predefined form pod namespace label this issue is to track a follow up improvement to support more flexible construction of the attribute values as suggested in
| 1
|
3,091
| 6,107,772,018
|
IssuesEvent
|
2017-06-21 08:55:59
|
rec-framework/rec-core
|
https://api.github.com/repos/rec-framework/rec-core
|
closed
|
Add encoding convertor?
|
feature/data-processing priority/Medium request/suggestion
|
Issue found in #35 , when specified CSV file in utf-8 mode, it will cause potential encoding problem.
Suggest add a conversion method to convert file from one encoding to another.
|
1.0
|
Add encoding convertor? - Issue found in #35 , when specified CSV file in utf-8 mode, it will cause potential encoding problem.
Suggest add a conversion method to convert file from one encoding to another.
|
process
|
add encoding convertor issue found in when specified csv file in utf mode it will cause potential encoding problem suggest add a conversion method to convert file from one encoding to another
| 1
|
17,430
| 23,249,029,817
|
IssuesEvent
|
2022-08-04 00:20:21
|
beyondhb1079/s4us
|
https://api.github.com/repos/beyondhb1079/s4us
|
opened
|
Email draft for reaching out
|
process
|
Come up with a compelling email draft for reaching out to new contacts
|
1.0
|
Email draft for reaching out - Come up with a compelling email draft for reaching out to new contacts
|
process
|
email draft for reaching out come up with a compelling email draft for reaching out to new contacts
| 1
|
165,307
| 26,142,449,088
|
IssuesEvent
|
2022-12-29 20:46:42
|
audacity/audacity
|
https://api.github.com/repos/audacity/audacity
|
closed
|
Track resampling affects to track speed
|
P4 Design / UX
|
**Describe the bug**
Track resampling works incorrectly
**To Reproduce**
1. Launch Audacity
2. Change rate in audio-interface settings
3. Click Play in Audacity
**Expected result:** Track rate has changed correctly and does not affect track speed.
Actual Result: Track speed had been changed
Win 10
Audacity 3.0.3.
Focusrite Scarlett 2i2
https://user-images.githubusercontent.com/25851359/127655827-c7798f6e-cbdf-409d-806c-79c42271391b.mp4
|
1.0
|
Track resampling affects to track speed - **Describe the bug**
Track resampling works incorrectly
**To Reproduce**
1. Launch Audacity
2. Change rate in audio-interface settings
3. Click Play in Audacity
**Expected result:** Track rate has changed correctly and does not affect track speed.
Actual Result: Track speed had been changed
Win 10
Audacity 3.0.3.
Focusrite Scarlett 2i2
https://user-images.githubusercontent.com/25851359/127655827-c7798f6e-cbdf-409d-806c-79c42271391b.mp4
|
non_process
|
track resampling affects to track speed describe the bug track resampling works incorrectly to reproduce launch audacity change rate in audio interface settings click play in audacity expected result track rate has changed correctly and does not affect track speed actual result track speed had been changed win audacity focusrite scarlett
| 0
|
121,452
| 10,168,817,266
|
IssuesEvent
|
2019-08-07 21:55:20
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
pull-kubernetes-e2e-aks-engine-azure consistently failing
|
area/provider/azure kind/failing-test priority/important-soon sig/cloud-provider
|
<!-- Please only use this template for submitting reports about failing tests in Kubernetes CI jobs -->
**Which jobs are failing**:
pull-kubernetes-e2e-aks-engine-azure
**Which test(s) are failing**:
`Up`
**Since when has it been failing**:
[2019-05-10 05:20:51 +0000 UTC](https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/77719/pull-kubernetes-e2e-aks-engine-azure/1126718651890667520)
**Testgrid link**:
No Testgrid, but job history: https://prow.k8s.io/job-history/kubernetes-jenkins/pr-logs/directory/pull-kubernetes-e2e-aks-engine-azure
**Reason for failure**:
```
error creating cluster: cannot deploy: cannot get the create deployment future response: Code="DeploymentFailed" Message="At least one resource deployment operation failed. Please list deployment operations for details. Please see https://aka.ms/arm-debug for usage details." Details=[{"code":"Conflict","message":"{\r\n \"status\": \"Failed\",\r\n \"error\": {\r\n \"code\": \"ResourceDeploymentFailure\",\r\n \"message\": \"The resource operation completed with terminal provisioning state 'Failed'.\",\r\n \"details\": [\r\n {\r\n \"code\": \"VMExtensionProvisioningError\",\r\n \"message\": \"VM has reported a failure when processing extension 'cse-master-0'. Error message: \\\"Enable failed: failed to execute command: command terminated with exit status=30\\n[stdout]\\n\\n[stderr]\\nConnection to k8s.gcr.io 443 port [tcp/https] succeeded!\\nConnection to gcr.io 443 port [tcp/https] succeeded!\\nConnection to docker.io 443 port [tcp/https] succeeded!\\n\\\".\"\r\n }\r\n ]\r\n }\r\n}"}]
```
**Anything else we need to know**:
ref: https://github.com/kubernetes/kubernetes/pull/79583#issuecomment-507224973
|
1.0
|
pull-kubernetes-e2e-aks-engine-azure consistently failing - <!-- Please only use this template for submitting reports about failing tests in Kubernetes CI jobs -->
**Which jobs are failing**:
pull-kubernetes-e2e-aks-engine-azure
**Which test(s) are failing**:
`Up`
**Since when has it been failing**:
[2019-05-10 05:20:51 +0000 UTC](https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/77719/pull-kubernetes-e2e-aks-engine-azure/1126718651890667520)
**Testgrid link**:
No Testgrid, but job history: https://prow.k8s.io/job-history/kubernetes-jenkins/pr-logs/directory/pull-kubernetes-e2e-aks-engine-azure
**Reason for failure**:
```
error creating cluster: cannot deploy: cannot get the create deployment future response: Code="DeploymentFailed" Message="At least one resource deployment operation failed. Please list deployment operations for details. Please see https://aka.ms/arm-debug for usage details." Details=[{"code":"Conflict","message":"{\r\n \"status\": \"Failed\",\r\n \"error\": {\r\n \"code\": \"ResourceDeploymentFailure\",\r\n \"message\": \"The resource operation completed with terminal provisioning state 'Failed'.\",\r\n \"details\": [\r\n {\r\n \"code\": \"VMExtensionProvisioningError\",\r\n \"message\": \"VM has reported a failure when processing extension 'cse-master-0'. Error message: \\\"Enable failed: failed to execute command: command terminated with exit status=30\\n[stdout]\\n\\n[stderr]\\nConnection to k8s.gcr.io 443 port [tcp/https] succeeded!\\nConnection to gcr.io 443 port [tcp/https] succeeded!\\nConnection to docker.io 443 port [tcp/https] succeeded!\\n\\\".\"\r\n }\r\n ]\r\n }\r\n}"}]
```
**Anything else we need to know**:
ref: https://github.com/kubernetes/kubernetes/pull/79583#issuecomment-507224973
|
non_process
|
pull kubernetes aks engine azure consistently failing which jobs are failing pull kubernetes aks engine azure which test s are failing up since when has it been failing testgrid link no testgrid but job history reason for failure error creating cluster cannot deploy cannot get the create deployment future response code deploymentfailed message at least one resource deployment operation failed please list deployment operations for details please see for usage details details n n nconnection to gcr io port succeeded nconnection to gcr io port succeeded nconnection to docker io port succeeded n r n r n r n r n anything else we need to know ref
| 0
|
99,715
| 21,032,253,669
|
IssuesEvent
|
2022-03-31 02:33:34
|
rss2email/rss2email
|
https://api.github.com/repos/rss2email/rss2email
|
closed
|
Feature Request: Modify the subject line
|
is:enhancement good first issue needs:code help wanted
|
I checked the docs and tried a lot of things but I couldn't seem to figure out a way to do this.
Some email providers, such as Microsoft Office 365 will strip / replace the From name & address with that of the account that's being used to send.
In these cases it would be better to use the subject line to display the feed name, author etc as opposed to trying to modify the From address. In addition, it would be good to be able to add your own subject items so R2E messages can be tagged for setting up mail sorting filters.
If there is a way to do this then I apologize in advance for opening an issue - if that is the case, please inform / update docs on how to do this.
Thank you,
|
1.0
|
Feature Request: Modify the subject line - I checked the docs and tried a lot of things but I couldn't seem to figure out a way to do this.
Some email providers, such as Microsoft Office 365 will strip / replace the From name & address with that of the account that's being used to send.
In these cases it would be better to use the subject line to display the feed name, author etc as opposed to trying to modify the From address. In addition, it would be good to be able to add your own subject items so R2E messages can be tagged for setting up mail sorting filters.
If there is a way to do this then I apologize in advance for opening an issue - if that is the case, please inform / update docs on how to do this.
Thank you,
|
non_process
|
feature request modify the subject line i checked the docs and tried a lot of things but i couldn t seem to figure out a way to do this some email providers such as microsoft office will strip replace the from name address with that of the account that s being used to send in these cases it would be better to use the subject line to display the feed name author etc as opposed to trying to modify the from address in addition it would be good to be able to add your own subject items so messages can be tagged for setting up mail sorting filters if there is a way to do this then i apologize in advance for opening an issue if that is the case please inform update docs on how to do this thank you
| 0
|
179,092
| 21,514,008,813
|
IssuesEvent
|
2022-04-28 08:13:26
|
elastic/security-docs
|
https://api.github.com/repos/elastic/security-docs
|
closed
|
[BUG] Authentications tab should not present on Hosts Page.
|
bug Team:Docs v8.2.0 Team:AWP Team: SecuritySolution
|
**Description**
Authentications tab should not present on Hosts Page.
**Actual Result and Guide links**
- Hosts Page : https://security-docs_1824.docs-preview.app.elstc.co/guide/en/security/master/session-view.html

**Expected Result**

|
True
|
[BUG] Authentications tab should not present on Hosts Page. - **Description**
Authentications tab should not present on Hosts Page.
**Actual Result and Guide links**
- Hosts Page : https://security-docs_1824.docs-preview.app.elstc.co/guide/en/security/master/session-view.html

**Expected Result**

|
non_process
|
authentications tab should not present on hosts page description authentications tab should not present on hosts page actual result and guide links hosts page expected result
| 0
|
9,057
| 12,131,948,116
|
IssuesEvent
|
2020-04-23 06:11:59
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
QgsProcessingParameterFolderDestination is not visible as input for other algs in modeler
|
Bug Processing PyQGIS
|
**Describe the bug**
processing parameter using QgsProcessingParameterFolderDestination is nott available as input in following algorithms used in a modeler.
Despite others derived from QgsProcessingDestinationParameter as QgsProcessingParameterFileDestination, this parameter is not available as input.
**How to Reproduce**
1. Create a pyqgis alg using:
`param = QgsProcessingParameterFolderDestination('OUTPUT', 'Output')
self.addParameter(param)`
2. Create a model il graphical modeler that use the als as in 1)
3. Create a new alg having a parameter that have to be initialize as "Algorithm output"
4. Scroll th elist of available Algorithm output => the output of algorithm added in 2 is not available
**QGIS and OS versions**
!!!The same happen in master https://github.com/qgis/QGIS/commit/8addbde0a5 !!!
QGIS version | 3.10.4-A Coruña | QGIS code revision | 961b3eee10
-- | -- | -- | --
Compiled against Qt | 5.9.5 | Running against Qt | 5.9.5
Compiled against GDAL/OGR | 2.2.3 | Running against GDAL/OGR | 2.2.3
Compiled against GEOS | 3.7.1-CAPI-1.11.1 | Running against GEOS | 3.7.1-CAPI-1.11.1 27a5e771
Compiled against SQLite | 3.22.0 | Running against SQLite | 3.22.0
PostgreSQL Client Version | 12.2 (Ubuntu 12.2-2.pgdg18.04+1) | SpatiaLite Version | 4.3.0a
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.2
PROJ.4 Version | 493
OS Version | Ubuntu 18.04.4 LTS | This copy of QGIS writes debugging output.
Active python plugins | plugin_reloader; IPyConsole; pluginbuilder3; copernicushub; firstaid; epyris; db_manager; processing; MetaSearch
|
1.0
|
QgsProcessingParameterFolderDestination is not visible as input for other algs in modeler - **Describe the bug**
processing parameter using QgsProcessingParameterFolderDestination is nott available as input in following algorithms used in a modeler.
Despite others derived from QgsProcessingDestinationParameter as QgsProcessingParameterFileDestination, this parameter is not available as input.
**How to Reproduce**
1. Create a pyqgis alg using:
`param = QgsProcessingParameterFolderDestination('OUTPUT', 'Output')
self.addParameter(param)`
2. Create a model il graphical modeler that use the als as in 1)
3. Create a new alg having a parameter that have to be initialize as "Algorithm output"
4. Scroll th elist of available Algorithm output => the output of algorithm added in 2 is not available
**QGIS and OS versions**
!!!The same happen in master https://github.com/qgis/QGIS/commit/8addbde0a5 !!!
QGIS version | 3.10.4-A Coruña | QGIS code revision | 961b3eee10
-- | -- | -- | --
Compiled against Qt | 5.9.5 | Running against Qt | 5.9.5
Compiled against GDAL/OGR | 2.2.3 | Running against GDAL/OGR | 2.2.3
Compiled against GEOS | 3.7.1-CAPI-1.11.1 | Running against GEOS | 3.7.1-CAPI-1.11.1 27a5e771
Compiled against SQLite | 3.22.0 | Running against SQLite | 3.22.0
PostgreSQL Client Version | 12.2 (Ubuntu 12.2-2.pgdg18.04+1) | SpatiaLite Version | 4.3.0a
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.2
PROJ.4 Version | 493
OS Version | Ubuntu 18.04.4 LTS | This copy of QGIS writes debugging output.
Active python plugins | plugin_reloader; IPyConsole; pluginbuilder3; copernicushub; firstaid; epyris; db_manager; processing; MetaSearch
|
process
|
qgsprocessingparameterfolderdestination is not visible as input for other algs in modeler describe the bug processing parameter using qgsprocessingparameterfolderdestination is nott available as input in following algorithms used in a modeler despite others derived from qgsprocessingdestinationparameter as qgsprocessingparameterfiledestination this parameter is not available as input how to reproduce create a pyqgis alg using param qgsprocessingparameterfolderdestination output output self addparameter param create a model il graphical modeler that use the als as in create a new alg having a parameter that have to be initialize as algorithm output scroll th elist of available algorithm output the output of algorithm added in is not available qgis and os versions the same happen in master qgis version a coruña qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version ubuntu spatialite version qwt version version proj version os version ubuntu lts this copy of qgis writes debugging output active python plugins plugin reloader ipyconsole copernicushub firstaid epyris db manager processing metasearch
| 1
|
40,789
| 12,799,736,489
|
IssuesEvent
|
2020-07-02 15:51:45
|
mwilliams7197/hollow
|
https://api.github.com/repos/mwilliams7197/hollow
|
opened
|
CVE-2019-12384 (Medium) detected in jackson-databind-2.4.3.jar, jackson-databind-2.2.2.jar
|
security vulnerability
|
## CVE-2019-12384 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.4.3.jar</b>, <b>jackson-databind-2.2.2.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.4.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to vulnerable library: canner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.4.3/feff63199be7b8f495c2f3e2096dcb6bd5e5b0b3/jackson-databind-2.4.3.jar,canner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.4.3/feff63199be7b8f495c2f3e2096dcb6bd5e5b0b3/jackson-databind-2.4.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.4.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.2.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to vulnerable library: /tmp/ws-scm/hollow/hollow-zenoadapter/build.gradle</p>
<p>
Dependency Hierarchy:
- netflix-zeno-2.22.3.jar (Root Library)
- :x: **jackson-databind-2.2.2.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/mwilliams7197/hollow/commits/ae866fd0a69bac4d9f321f0b05dafcde9a3d6e69">ae866fd0a69bac4d9f321f0b05dafcde9a3d6e69</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.9.1 might allow attackers to have a variety of impacts by leveraging failure to block the logback-core class from polymorphic deserialization. Depending on the classpath content, remote code execution may be possible.
<p>Publish Date: 2019-06-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12384>CVE-2019-12384</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384</a></p>
<p>Release Date: 2019-06-24</p>
<p>Fix Resolution: 2.9.9.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.4.3","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.4.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.9.1"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.2.2","isTransitiveDependency":true,"dependencyTree":"com.netflix.zeno:netflix-zeno:2.22.3;com.fasterxml.jackson.core:jackson-databind:2.2.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.9.1"}],"vulnerabilityIdentifier":"CVE-2019-12384","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.9.1 might allow attackers to have a variety of impacts by leveraging failure to block the logback-core class from polymorphic deserialization. Depending on the classpath content, remote code execution may be possible.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12384","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-12384 (Medium) detected in jackson-databind-2.4.3.jar, jackson-databind-2.2.2.jar - ## CVE-2019-12384 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.4.3.jar</b>, <b>jackson-databind-2.2.2.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.4.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to vulnerable library: canner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.4.3/feff63199be7b8f495c2f3e2096dcb6bd5e5b0b3/jackson-databind-2.4.3.jar,canner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.4.3/feff63199be7b8f495c2f3e2096dcb6bd5e5b0b3/jackson-databind-2.4.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.4.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.2.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to vulnerable library: /tmp/ws-scm/hollow/hollow-zenoadapter/build.gradle</p>
<p>
Dependency Hierarchy:
- netflix-zeno-2.22.3.jar (Root Library)
- :x: **jackson-databind-2.2.2.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/mwilliams7197/hollow/commits/ae866fd0a69bac4d9f321f0b05dafcde9a3d6e69">ae866fd0a69bac4d9f321f0b05dafcde9a3d6e69</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.9.1 might allow attackers to have a variety of impacts by leveraging failure to block the logback-core class from polymorphic deserialization. Depending on the classpath content, remote code execution may be possible.
<p>Publish Date: 2019-06-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12384>CVE-2019-12384</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384</a></p>
<p>Release Date: 2019-06-24</p>
<p>Fix Resolution: 2.9.9.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.4.3","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.4.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.9.1"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.2.2","isTransitiveDependency":true,"dependencyTree":"com.netflix.zeno:netflix-zeno:2.22.3;com.fasterxml.jackson.core:jackson-databind:2.2.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.9.1"}],"vulnerabilityIdentifier":"CVE-2019-12384","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.9.1 might allow attackers to have a variety of impacts by leveraging failure to block the logback-core class from polymorphic deserialization. Depending on the classpath content, remote code execution may be possible.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12384","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in jackson databind jar jackson databind jar cve medium severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api path to vulnerable library canner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar canner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to vulnerable library tmp ws scm hollow hollow zenoadapter build gradle dependency hierarchy netflix zeno jar root library x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before might allow attackers to have a variety of impacts by leveraging failure to block the logback core class from polymorphic deserialization depending on the classpath content remote code execution may be possible publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before might allow attackers to have a variety of impacts by leveraging failure to block the logback core class from polymorphic deserialization depending on the classpath content remote code execution may be possible vulnerabilityurl
| 0
|
660
| 3,129,825,780
|
IssuesEvent
|
2015-09-09 05:01:43
|
e-government-ua/i
|
https://api.github.com/repos/e-government-ua/i
|
closed
|
На главном портале для Киева убрать пункты меню
|
active hi priority In process of testing test
|
8. В версию возможно В Работе… Убрать разделы "Документи", "Мій журнал", "Про портал"
https://docs.google.com/document/d/1fUJlMptp0npeXNShMwZedfcqnhfz6mhyDcw7qoxtZqU/edit#
|
1.0
|
На главном портале для Киева убрать пункты меню - 8. В версию возможно В Работе… Убрать разделы "Документи", "Мій журнал", "Про портал"
https://docs.google.com/document/d/1fUJlMptp0npeXNShMwZedfcqnhfz6mhyDcw7qoxtZqU/edit#
|
process
|
на главном портале для киева убрать пункты меню в версию возможно в работе… убрать разделы документи мій журнал про портал
| 1
|
284,510
| 30,913,639,654
|
IssuesEvent
|
2023-08-05 02:28:22
|
Nivaskumark/kernel_v4.19.72_old
|
https://api.github.com/repos/Nivaskumark/kernel_v4.19.72_old
|
reopened
|
CVE-2022-41222 (High) detected in linux-yoctov5.4.51
|
Mend: dependency security vulnerability
|
## CVE-2022-41222 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/kernel_v4.19.72/commit/ce49083a1c14be2d13cb5e878257d293e6c748bc">ce49083a1c14be2d13cb5e878257d293e6c748bc</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/mremap.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/mremap.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
mm/mremap.c in the Linux kernel before 5.13.3 has a use-after-free via a stale TLB because an rmap lock is not held during a PUD move.
<p>Publish Date: 2022-09-21
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41222>CVE-2022-41222</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-41222">https://www.linuxkernelcves.com/cves/CVE-2022-41222</a></p>
<p>Release Date: 2022-09-21</p>
<p>Fix Resolution: v5.4.211,v5.10.137,v5.12.18,v5.13.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-41222 (High) detected in linux-yoctov5.4.51 - ## CVE-2022-41222 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/kernel_v4.19.72/commit/ce49083a1c14be2d13cb5e878257d293e6c748bc">ce49083a1c14be2d13cb5e878257d293e6c748bc</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/mremap.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/mremap.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
mm/mremap.c in the Linux kernel before 5.13.3 has a use-after-free via a stale TLB because an rmap lock is not held during a PUD move.
<p>Publish Date: 2022-09-21
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41222>CVE-2022-41222</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-41222">https://www.linuxkernelcves.com/cves/CVE-2022-41222</a></p>
<p>Release Date: 2022-09-21</p>
<p>Fix Resolution: v5.4.211,v5.10.137,v5.12.18,v5.13.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in linux cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files mm mremap c mm mremap c vulnerability details mm mremap c in the linux kernel before has a use after free via a stale tlb because an rmap lock is not held during a pud move publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
2,402
| 5,193,014,871
|
IssuesEvent
|
2017-01-22 15:12:02
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
when field validator message is displayed and clicking New button will cause infinite loop
|
bug inprocess
|
Hi, this is an edge case which will cause the browser to hang.
# Repro steps:
1) Run the react-bootstrap-table example project locally
2) Navigate to the "Advanced data edit & Insert" option under Examples
3) Scroll to the last example
4) click the job name cell. Remove some characters so validator will display validation msg
5) before the msg disappear, click the "+new" button.
# Result
browser will hang
|
1.0
|
when field validator message is displayed and clicking New button will cause infinite loop - Hi, this is an edge case which will cause the browser to hang.
# Repro steps:
1) Run the react-bootstrap-table example project locally
2) Navigate to the "Advanced data edit & Insert" option under Examples
3) Scroll to the last example
4) click the job name cell. Remove some characters so validator will display validation msg
5) before the msg disappear, click the "+new" button.
# Result
browser will hang
|
process
|
when field validator message is displayed and clicking new button will cause infinite loop hi this is an edge case which will cause the browser to hang repro steps run the react bootstrap table example project locally navigate to the advanced data edit insert option under examples scroll to the last example click the job name cell remove some characters so validator will display validation msg before the msg disappear click the new button result browser will hang
| 1
|
4,969
| 7,806,702,570
|
IssuesEvent
|
2018-06-11 14:46:15
|
StackSavingsTeam/stacksavings.com_templates
|
https://api.github.com/repos/StackSavingsTeam/stacksavings.com_templates
|
opened
|
Crear el layout en el repo
|
On Process
|
Hay que crea el siguiente layout en este repositorio:
http://stacksavings.com/detail-post/the-best-new-products-every-day-gDuXo/en?skipcache
Hay que chequear si le falta algo
|
1.0
|
Crear el layout en el repo - Hay que crea el siguiente layout en este repositorio:
http://stacksavings.com/detail-post/the-best-new-products-every-day-gDuXo/en?skipcache
Hay que chequear si le falta algo
|
process
|
crear el layout en el repo hay que crea el siguiente layout en este repositorio hay que chequear si le falta algo
| 1
|
13,143
| 15,560,290,183
|
IssuesEvent
|
2021-03-16 12:33:21
|
prisma/language-tools
|
https://api.github.com/repos/prisma/language-tools
|
closed
|
Autocomplete of @@map is not perfect
|
bug/1-repro-available bug/2-confirmed kind/bug process/candidate topic: autocompletion
|
## Bug description
When I have an autocomplete for @@map, an array is automatically added
## How to reproduce
Create a model like this:
```prisma
model Persona {
uid String @id @default(uuid())
nome String
cognome String
sesso Sesso
data_nascita DateTime
}
```
Then start typing `@@ma...`
A suggestion of autocompletion for the **@@map** should appear. But if I press **enter**, what I obtain is:
```prisma
model Persona {
uid String @id @default(uuid())
nome String
cognome String
sesso Sesso
data_nascita DateTime
@@map([""]) // array should not be here
}
```
It creates an array, but it should not be an array. I always have to delete it manually.
## Expected behavior
Instead of ` @@map([""]) ` I should have ` @@map("")`
## Environment & setup
- OS: Ubuntu
- Editor: VSCode
```
```
|
1.0
|
Autocomplete of @@map is not perfect - ## Bug description
When I have an autocomplete for @@map, an array is automatically added
## How to reproduce
Create a model like this:
```prisma
model Persona {
uid String @id @default(uuid())
nome String
cognome String
sesso Sesso
data_nascita DateTime
}
```
Then start typing `@@ma...`
A suggestion of autocompletion for the **@@map** should appear. But if I press **enter**, what I obtain is:
```prisma
model Persona {
uid String @id @default(uuid())
nome String
cognome String
sesso Sesso
data_nascita DateTime
@@map([""]) // array should not be here
}
```
It creates an array, but it should not be an array. I always have to delete it manually.
## Expected behavior
Instead of ` @@map([""]) ` I should have ` @@map("")`
## Environment & setup
- OS: Ubuntu
- Editor: VSCode
```
```
|
process
|
autocomplete of map is not perfect bug description when i have an autocomplete for map an array is automatically added how to reproduce create a model like this prisma model persona uid string id default uuid nome string cognome string sesso sesso data nascita datetime then start typing ma a suggestion of autocompletion for the map should appear but if i press enter what i obtain is prisma model persona uid string id default uuid nome string cognome string sesso sesso data nascita datetime map array should not be here it creates an array but it should not be an array i always have to delete it manually expected behavior instead of map i should have map environment setup os ubuntu editor vscode
| 1
|
99,261
| 16,439,882,365
|
IssuesEvent
|
2021-05-20 13:19:31
|
nextcloud/server
|
https://api.github.com/repos/nextcloud/server
|
closed
|
Lack of SFTP (SSH) Host Key Verification
|
0. Needs triage bug feature: external storage security
|
Connections to SFTP external storage were made without verifying the remote host's SSH host key.
This behaviour did not appear to be documented anywhere so that users may factor it into their decision to use the External Storage application for remote SFTP resources. This is important since a change of the remote host's SSH host key may indicate that a man-in-the-middle attack is in progress.
From the application [logic](https://github.com/nextcloud/server/blob/master/apps/files_external/lib/Lib/Storage/SFTP.php), host key checking would only be performed if a "_ssh_hostKeys_" file exists for the given user. However, this file was not created by default. Furthermore, the method "_writeHostKeys()_" would only update the list of SSH host keys if the user's "_ssh_hostKeys_" file already existed.
To conclude, the verification logic does already exist within the current Nextcloud code base but is not used by default. I do recognise that the reason for this is because of the associated functionality that is presently missing such as the user's ability to manage their own "_ssh_hostKeys_" file.
### Steps to reproduce
1. On Nextcloud, create a SFTP external storage as normal
2. Through Nextcloud, access the files stored on the remote SFTP storage
3. On the remote SSH host, change the SSH host key and restart the SSH service
4. Through Nextcloud, once again access the files stored on the remote SFTP storage
### Expected behaviour
The Nextcloud instance should alert the user with an error message stating "_Host public key does not match known key_".
It should be noted that this message does actually appear if the user's "_ssh_HostKeys_" file exists and contains a host key that differs the remote host's current one (i.e. verification has failed).
### Actual behaviour
The Nextcloud instance permits the user to access the remote SFTP storage without any indication that the remote host's SSH host key has been changed.
### Server configuration
**Operating system:** Ubuntu GNU/Linux 18.04.1 LTS
**Web server:** Apache 2.4.29-1ubuntu4.5
**Database:** MariaDB 10.1.34-0ubuntu0.18.04.1
**PHP version:** 7.2+60ubuntu1
**Nextcloud version:** 15.0.2
**Updated from an older Nextcloud/ownCloud or fresh install:** Fresh
**Where did you install Nextcloud from:** Official website archive
**Signing status:**
<details>
<summary>Signing status</summary>
No errors have been found.
</details>
**Are you using external storage, if yes which one:** SFTP
**Are you using encryption:** No
**Are you using an external user-backend, if yes which one:** No
|
True
|
Lack of SFTP (SSH) Host Key Verification - Connections to SFTP external storage were made without verifying the remote host's SSH host key.
This behaviour did not appear to be documented anywhere so that users may factor it into their decision to use the External Storage application for remote SFTP resources. This is important since a change of the remote host's SSH host key may indicate that a man-in-the-middle attack is in progress.
From the application [logic](https://github.com/nextcloud/server/blob/master/apps/files_external/lib/Lib/Storage/SFTP.php), host key checking would only be performed if a "_ssh_hostKeys_" file exists for the given user. However, this file was not created by default. Furthermore, the method "_writeHostKeys()_" would only update the list of SSH host keys if the user's "_ssh_hostKeys_" file already existed.
To conclude, the verification logic does already exist within the current Nextcloud code base but is not used by default. I do recognise that the reason for this is because of the associated functionality that is presently missing such as the user's ability to manage their own "_ssh_hostKeys_" file.
### Steps to reproduce
1. On Nextcloud, create a SFTP external storage as normal
2. Through Nextcloud, access the files stored on the remote SFTP storage
3. On the remote SSH host, change the SSH host key and restart the SSH service
4. Through Nextcloud, once again access the files stored on the remote SFTP storage
### Expected behaviour
The Nextcloud instance should alert the user with an error message stating "_Host public key does not match known key_".
It should be noted that this message does actually appear if the user's "_ssh_HostKeys_" file exists and contains a host key that differs the remote host's current one (i.e. verification has failed).
### Actual behaviour
The Nextcloud instance permits the user to access the remote SFTP storage without any indication that the remote host's SSH host key has been changed.
### Server configuration
**Operating system:** Ubuntu GNU/Linux 18.04.1 LTS
**Web server:** Apache 2.4.29-1ubuntu4.5
**Database:** MariaDB 10.1.34-0ubuntu0.18.04.1
**PHP version:** 7.2+60ubuntu1
**Nextcloud version:** 15.0.2
**Updated from an older Nextcloud/ownCloud or fresh install:** Fresh
**Where did you install Nextcloud from:** Official website archive
**Signing status:**
<details>
<summary>Signing status</summary>
No errors have been found.
</details>
**Are you using external storage, if yes which one:** SFTP
**Are you using encryption:** No
**Are you using an external user-backend, if yes which one:** No
|
non_process
|
lack of sftp ssh host key verification connections to sftp external storage were made without verifying the remote host s ssh host key this behaviour did not appear to be documented anywhere so that users may factor it into their decision to use the external storage application for remote sftp resources this is important since a change of the remote host s ssh host key may indicate that a man in the middle attack is in progress from the application host key checking would only be performed if a ssh hostkeys file exists for the given user however this file was not created by default furthermore the method writehostkeys would only update the list of ssh host keys if the user s ssh hostkeys file already existed to conclude the verification logic does already exist within the current nextcloud code base but is not used by default i do recognise that the reason for this is because of the associated functionality that is presently missing such as the user s ability to manage their own ssh hostkeys file steps to reproduce on nextcloud create a sftp external storage as normal through nextcloud access the files stored on the remote sftp storage on the remote ssh host change the ssh host key and restart the ssh service through nextcloud once again access the files stored on the remote sftp storage expected behaviour the nextcloud instance should alert the user with an error message stating host public key does not match known key it should be noted that this message does actually appear if the user s ssh hostkeys file exists and contains a host key that differs the remote host s current one i e verification has failed actual behaviour the nextcloud instance permits the user to access the remote sftp storage without any indication that the remote host s ssh host key has been changed server configuration operating system ubuntu gnu linux lts web server apache database mariadb php version nextcloud version updated from an older nextcloud owncloud or fresh install fresh where did you install nextcloud from official website archive signing status signing status no errors have been found are you using external storage if yes which one sftp are you using encryption no are you using an external user backend if yes which one no
| 0
|
23,513
| 16,361,431,193
|
IssuesEvent
|
2021-05-14 10:04:30
|
tarantool/tarantool
|
https://api.github.com/repos/tarantool/tarantool
|
opened
|
infra: add Fedora 34 builds
|
2sp infrastructure qa teamQ
|
> Fedora Linux 34 was released on 2021-04-27.
https://fedoraproject.org/wiki/Releases/34/ChangeSet
A Fedora release [usually](https://fedoraproject.org/wiki/End_of_life) maintained ~400 days.
|
1.0
|
infra: add Fedora 34 builds - > Fedora Linux 34 was released on 2021-04-27.
https://fedoraproject.org/wiki/Releases/34/ChangeSet
A Fedora release [usually](https://fedoraproject.org/wiki/End_of_life) maintained ~400 days.
|
non_process
|
infra add fedora builds fedora linux was released on a fedora release maintained days
| 0
|
2,255
| 5,088,657,062
|
IssuesEvent
|
2017-01-01 00:09:42
|
sw4j-org/tool-jpa-processor
|
https://api.github.com/repos/sw4j-org/tool-jpa-processor
|
opened
|
Handle @UniqueConstraint Annotation
|
annotation processor task
|
Handle the `@UniqueConstraint` annotation for a property or field.
See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf)
- 11.1.53 UniqueConstraint Annotation
|
1.0
|
Handle @UniqueConstraint Annotation - Handle the `@UniqueConstraint` annotation for a property or field.
See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf)
- 11.1.53 UniqueConstraint Annotation
|
process
|
handle uniqueconstraint annotation handle the uniqueconstraint annotation for a property or field see uniqueconstraint annotation
| 1
|
25,426
| 11,171,419,071
|
IssuesEvent
|
2019-12-28 19:35:20
|
eldorplus/conception-website
|
https://api.github.com/repos/eldorplus/conception-website
|
opened
|
CVE-2019-18797 (Medium) detected in node-sass-v4.11.0
|
security vulnerability
|
## CVE-2019-18797 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.11.0</b></p></summary>
<p>
<p>:rainbow: Node.js bindings to libsass</p>
<p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/eldorplus/conception-website/commit/d01989566d2a70e542b08bc943c5de9d223ce39d">d01989566d2a70e542b08bc943c5de9d223ce39d</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (66)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /conception-website/node_modules/node-sass/src/libsass/src/expand.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/expand.cpp
- /conception-website/node_modules/node-sass/src/sass_types/factory.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/operators.cpp
- /conception-website/node_modules/node-sass/src/sass_types/boolean.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/util.hpp
- /conception-website/node_modules/node-sass/src/sass_types/value.h
- /conception-website/node_modules/node-sass/src/libsass/src/emitter.hpp
- /conception-website/node_modules/node-sass/src/callback_bridge.h
- /conception-website/node_modules/node-sass/src/libsass/src/file.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/sass.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/operation.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/operators.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/constants.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/error_handling.hpp
- /conception-website/node_modules/node-sass/src/custom_importer_bridge.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/parser.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/constants.cpp
- /conception-website/node_modules/node-sass/src/sass_types/list.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/cssize.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/functions.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/util.cpp
- /conception-website/node_modules/node-sass/src/custom_function_bridge.cpp
- /conception-website/node_modules/node-sass/src/custom_importer_bridge.h
- /conception-website/node_modules/node-sass/src/libsass/src/bind.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/eval.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/inspect.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/backtrace.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/extend.cpp
- /conception-website/node_modules/node-sass/src/sass_types/sass_value_wrapper.h
- /conception-website/node_modules/node-sass/src/libsass/src/error_handling.cpp
- /conception-website/node_modules/node-sass/src/sass_context_wrapper.h
- /conception-website/node_modules/node-sass/src/libsass/src/parser.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/debugger.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/emitter.cpp
- /conception-website/node_modules/node-sass/src/sass_types/number.cpp
- /conception-website/node_modules/node-sass/src/sass_types/color.h
- /conception-website/node_modules/node-sass/src/libsass/src/ast.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/sass_values.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/output.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/check_nesting.cpp
- /conception-website/node_modules/node-sass/src/sass_types/null.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/cssize.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/functions.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/prelexer.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/ast.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/to_c.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/to_value.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp
- /conception-website/node_modules/node-sass/src/sass_types/color.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/inspect.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/values.cpp
- /conception-website/node_modules/node-sass/src/sass_context_wrapper.cpp
- /conception-website/node_modules/node-sass/src/sass_types/list.h
- /conception-website/node_modules/node-sass/src/libsass/src/check_nesting.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/to_value.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/context.cpp
- /conception-website/node_modules/node-sass/src/sass_types/map.cpp
- /conception-website/node_modules/node-sass/src/binding.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/sass_context.cpp
- /conception-website/node_modules/node-sass/src/sass_types/string.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/context.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/prelexer.hpp
- /conception-website/node_modules/node-sass/src/sass_types/boolean.h
- /conception-website/node_modules/node-sass/src/libsass/src/eval.cpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp.
<p>Publish Date: 2019-11-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797>CVE-2019-18797</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797</a></p>
<p>Release Date: 2019-11-06</p>
<p>Fix Resolution: 3.6.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-18797 (Medium) detected in node-sass-v4.11.0 - ## CVE-2019-18797 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.11.0</b></p></summary>
<p>
<p>:rainbow: Node.js bindings to libsass</p>
<p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/eldorplus/conception-website/commit/d01989566d2a70e542b08bc943c5de9d223ce39d">d01989566d2a70e542b08bc943c5de9d223ce39d</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (66)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /conception-website/node_modules/node-sass/src/libsass/src/expand.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/expand.cpp
- /conception-website/node_modules/node-sass/src/sass_types/factory.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/operators.cpp
- /conception-website/node_modules/node-sass/src/sass_types/boolean.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/util.hpp
- /conception-website/node_modules/node-sass/src/sass_types/value.h
- /conception-website/node_modules/node-sass/src/libsass/src/emitter.hpp
- /conception-website/node_modules/node-sass/src/callback_bridge.h
- /conception-website/node_modules/node-sass/src/libsass/src/file.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/sass.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/operation.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/operators.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/constants.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/error_handling.hpp
- /conception-website/node_modules/node-sass/src/custom_importer_bridge.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/parser.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/constants.cpp
- /conception-website/node_modules/node-sass/src/sass_types/list.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/cssize.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/functions.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/util.cpp
- /conception-website/node_modules/node-sass/src/custom_function_bridge.cpp
- /conception-website/node_modules/node-sass/src/custom_importer_bridge.h
- /conception-website/node_modules/node-sass/src/libsass/src/bind.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/eval.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/inspect.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/backtrace.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/extend.cpp
- /conception-website/node_modules/node-sass/src/sass_types/sass_value_wrapper.h
- /conception-website/node_modules/node-sass/src/libsass/src/error_handling.cpp
- /conception-website/node_modules/node-sass/src/sass_context_wrapper.h
- /conception-website/node_modules/node-sass/src/libsass/src/parser.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/debugger.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/emitter.cpp
- /conception-website/node_modules/node-sass/src/sass_types/number.cpp
- /conception-website/node_modules/node-sass/src/sass_types/color.h
- /conception-website/node_modules/node-sass/src/libsass/src/ast.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/sass_values.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/output.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/check_nesting.cpp
- /conception-website/node_modules/node-sass/src/sass_types/null.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/cssize.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/functions.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/prelexer.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/ast.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/to_c.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/to_value.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp
- /conception-website/node_modules/node-sass/src/sass_types/color.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/inspect.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/values.cpp
- /conception-website/node_modules/node-sass/src/sass_context_wrapper.cpp
- /conception-website/node_modules/node-sass/src/sass_types/list.h
- /conception-website/node_modules/node-sass/src/libsass/src/check_nesting.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/to_value.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/context.cpp
- /conception-website/node_modules/node-sass/src/sass_types/map.cpp
- /conception-website/node_modules/node-sass/src/binding.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/sass_context.cpp
- /conception-website/node_modules/node-sass/src/sass_types/string.cpp
- /conception-website/node_modules/node-sass/src/libsass/src/context.hpp
- /conception-website/node_modules/node-sass/src/libsass/src/prelexer.hpp
- /conception-website/node_modules/node-sass/src/sass_types/boolean.h
- /conception-website/node_modules/node-sass/src/libsass/src/eval.cpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp.
<p>Publish Date: 2019-11-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797>CVE-2019-18797</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797</a></p>
<p>Release Date: 2019-11-06</p>
<p>Fix Resolution: 3.6.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in node sass cve medium severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries conception website node modules node sass src libsass src expand hpp conception website node modules node sass src libsass src expand cpp conception website node modules node sass src sass types factory cpp conception website node modules node sass src libsass src operators cpp conception website node modules node sass src sass types boolean cpp conception website node modules node sass src libsass src util hpp conception website node modules node sass src sass types value h conception website node modules node sass src libsass src emitter hpp conception website node modules node sass src callback bridge h conception website node modules node sass src libsass src file cpp conception website node modules node sass src libsass src sass cpp conception website node modules node sass src libsass src operation hpp conception website node modules node sass src libsass src operators hpp conception website node modules node sass src libsass src constants hpp conception website node modules node sass src libsass src error handling hpp conception website node modules node sass src custom importer bridge cpp conception website node modules node sass src libsass src parser hpp conception website node modules node sass src libsass src constants cpp conception website node modules node sass src sass types list cpp conception website node modules node sass src libsass src cssize cpp conception website node modules node sass src libsass src functions hpp conception website node modules node sass src libsass src util cpp conception website node modules node sass src custom function bridge cpp conception website node modules node sass src custom importer bridge h conception website node modules node sass src libsass src bind cpp conception website node modules node sass src libsass src eval hpp conception website node modules node sass src libsass src inspect cpp conception website node modules node sass src libsass src backtrace cpp conception website node modules node sass src libsass src extend cpp conception website node modules node sass src sass types sass value wrapper h conception website node modules node sass src libsass src error handling cpp conception website node modules node sass src sass context wrapper h conception website node modules node sass src libsass src parser cpp conception website node modules node sass src libsass src debugger hpp conception website node modules node sass src libsass src emitter cpp conception website node modules node sass src sass types number cpp conception website node modules node sass src sass types color h conception website node modules node sass src libsass src ast hpp conception website node modules node sass src libsass src sass values cpp conception website node modules node sass src libsass src output cpp conception website node modules node sass src libsass src check nesting cpp conception website node modules node sass src sass types null cpp conception website node modules node sass src libsass src ast def macros hpp conception website node modules node sass src libsass src cssize hpp conception website node modules node sass src libsass src functions cpp conception website node modules node sass src libsass src prelexer cpp conception website node modules node sass src libsass src ast cpp conception website node modules node sass src libsass src to c cpp conception website node modules node sass src libsass src to value hpp conception website node modules node sass src libsass src ast fwd decl hpp conception website node modules node sass src sass types color cpp conception website node modules node sass src libsass src inspect hpp conception website node modules node sass src libsass src values cpp conception website node modules node sass src sass context wrapper cpp conception website node modules node sass src sass types list h conception website node modules node sass src libsass src check nesting hpp conception website node modules node sass src libsass src to value cpp conception website node modules node sass src libsass src context cpp conception website node modules node sass src sass types map cpp conception website node modules node sass src binding cpp conception website node modules node sass src libsass src sass context cpp conception website node modules node sass src sass types string cpp conception website node modules node sass src libsass src context hpp conception website node modules node sass src libsass src prelexer hpp conception website node modules node sass src sass types boolean h conception website node modules node sass src libsass src eval cpp vulnerability details libsass has uncontrolled recursion in sass eval operator sass binary expression in eval cpp publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
17,831
| 23,769,964,743
|
IssuesEvent
|
2022-09-01 15:30:41
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
[processor/cumulativetodelta] Convert cumulative Histograms to delta temporality
|
priority:p2 processor/cumulativetodelta
|
**Is your feature request related to a problem? Please describe.**
It seems that cumulative-to-delta at the moment only processes Sums, and just skips over other data types. This is a problem for us, as we have Histograms in our pipeline with cumulative temporality that we want to convert.
**Describe the solution you'd like**
Rather than bypass Histogram metrics, have the cumulative-to-delta processor convert their data points in a similar way to sums.
**Describe alternatives you've considered**
None, other than continuing to ignore non-Sum metrics.
|
1.0
|
[processor/cumulativetodelta] Convert cumulative Histograms to delta temporality - **Is your feature request related to a problem? Please describe.**
It seems that cumulative-to-delta at the moment only processes Sums, and just skips over other data types. This is a problem for us, as we have Histograms in our pipeline with cumulative temporality that we want to convert.
**Describe the solution you'd like**
Rather than bypass Histogram metrics, have the cumulative-to-delta processor convert their data points in a similar way to sums.
**Describe alternatives you've considered**
None, other than continuing to ignore non-Sum metrics.
|
process
|
convert cumulative histograms to delta temporality is your feature request related to a problem please describe it seems that cumulative to delta at the moment only processes sums and just skips over other data types this is a problem for us as we have histograms in our pipeline with cumulative temporality that we want to convert describe the solution you d like rather than bypass histogram metrics have the cumulative to delta processor convert their data points in a similar way to sums describe alternatives you ve considered none other than continuing to ignore non sum metrics
| 1
|
14,513
| 5,683,318,116
|
IssuesEvent
|
2017-04-13 12:22:46
|
alicevision/openMVG
|
https://api.github.com/repos/alicevision/openMVG
|
closed
|
[cmake] fix ceres dependency in openMVGConfig
|
review scope:build type:bug
|
Currently, when we use openMVG from another repo we need to add `find_package(Ceres)`.
|
1.0
|
[cmake] fix ceres dependency in openMVGConfig - Currently, when we use openMVG from another repo we need to add `find_package(Ceres)`.
|
non_process
|
fix ceres dependency in openmvgconfig currently when we use openmvg from another repo we need to add find package ceres
| 0
|
11,209
| 13,957,706,427
|
IssuesEvent
|
2020-10-24 08:14:50
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
IT - RNDT: IT - National discovery service (RNDT) under technical maintenance in order to switch to the new release
|
Geoportal Harvesting process IT - Italy
|
Dear Angelo,
I inform you that the RNDT services are under technical maintenance for the time being in order to allow to perform all the activities required for the release of the new version of the discovery service.
Consequently, please keep the records harvested in the last session.
In the meantime, if you agree, a harvesting test session on the new endpoint of the national discovery service could be performed. If so, I'll report to you the URL of the new endpoint.
In any case, that endpoint will be formally registered by the competent Authority according to the known procedure.
Best regards,
Antonio
|
1.0
|
IT - RNDT: IT - National discovery service (RNDT) under technical maintenance in order to switch to the new release - Dear Angelo,
I inform you that the RNDT services are under technical maintenance for the time being in order to allow to perform all the activities required for the release of the new version of the discovery service.
Consequently, please keep the records harvested in the last session.
In the meantime, if you agree, a harvesting test session on the new endpoint of the national discovery service could be performed. If so, I'll report to you the URL of the new endpoint.
In any case, that endpoint will be formally registered by the competent Authority according to the known procedure.
Best regards,
Antonio
|
process
|
it rndt it national discovery service rndt under technical maintenance in order to switch to the new release dear angelo i inform you that the rndt services are under technical maintenance for the time being in order to allow to perform all the activities required for the release of the new version of the discovery service consequently please keep the records harvested in the last session in the meantime if you agree a harvesting test session on the new endpoint of the national discovery service could be performed if so i ll report to you the url of the new endpoint in any case that endpoint will be formally registered by the competent authority according to the known procedure best regards antonio
| 1
|
13,297
| 15,770,095,062
|
IssuesEvent
|
2021-03-31 19:03:41
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Add project editorconfig
|
process: contributing stage: proposal 💡
|
Add a project [`.editorconfig`](http://editorconfig.org/) file, to make it easier for contributors.
Suggested (`/.editorconfig`):
```
root = true
[*]
end_of_line = lf
insert_final_newline = true
indent_style = space
indent_size = 2
```
Not sure if those apply project-wide.
|
1.0
|
Add project editorconfig - Add a project [`.editorconfig`](http://editorconfig.org/) file, to make it easier for contributors.
Suggested (`/.editorconfig`):
```
root = true
[*]
end_of_line = lf
insert_final_newline = true
indent_style = space
indent_size = 2
```
Not sure if those apply project-wide.
|
process
|
add project editorconfig add a project file to make it easier for contributors suggested editorconfig root true end of line lf insert final newline true indent style space indent size not sure if those apply project wide
| 1
|
1,260
| 3,792,106,229
|
IssuesEvent
|
2016-03-22 08:00:50
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
About process.nextTick
|
c++ process question
|
It's not a problem or bug. I just view code in node.cc, and could not find where to invoke the `node::MakeCallback` function which actually execute callbacks bound with `process.nextTick`.
It looks like those code make sense:
```
if (!env->KickNextTick(&callback_scope)) {
return Undefined(env->isolate());
}
```
* **Version**: v4.2.6
* **Platform**: Darwin Kernel Version 14.1.1
|
1.0
|
About process.nextTick - It's not a problem or bug. I just view code in node.cc, and could not find where to invoke the `node::MakeCallback` function which actually execute callbacks bound with `process.nextTick`.
It looks like those code make sense:
```
if (!env->KickNextTick(&callback_scope)) {
return Undefined(env->isolate());
}
```
* **Version**: v4.2.6
* **Platform**: Darwin Kernel Version 14.1.1
|
process
|
about process nexttick it s not a problem or bug i just view code in node cc and could not find where to invoke the node makecallback function which actually execute callbacks bound with process nexttick it looks like those code make sense if env kicknexttick callback scope return undefined env isolate version platform darwin kernel version
| 1
|
11,642
| 2,660,012,351
|
IssuesEvent
|
2015-03-19 01:34:13
|
perfsonar/project
|
https://api.github.com/repos/perfsonar/project
|
closed
|
Display input-source name in psGraph.cgi
|
Milestone-Release3.5 Priority-Medium Type-Defect
|
Original [issue 1027](https://code.google.com/p/perfsonar-ps/issues/detail?id=1027) created by arlake228 on 2014-12-22T16:54:41.000Z:
<b>What steps will reproduce the problem?</b>
Create test for a target host (foo.example) using its hostname, but which does not have reverse DNS
<b>What is the expected output? What do you see instead?</b>
What I has hoping to see in the psGraph.cgi listing is:
foo.example (192.0.2.252)
What I actually see is:
192.0.2.252 (192.0.2.252)
<b>What version of the product are you using? On what operating system?</b>
perfsonar PS 3.4.1
<b>Please provide any additional information below.</b>
If I look at the top-level JSON info in esmond:
curl http://x.x.x.x/esmond/perfsonar/archive/?format=json | python -m json.tool
I see:
"input-source": "foo.example",
so it looks like the information is available. ("input-source" is the name, and "input" is the resolved name).
If I add an entry in /etc/hosts for the remote host then the name is displayed in the psGraph.cgi output, which suggests that it is indeed doing a reverse lookup to get the name.
Suggested approach 1:
* if input-source is an IP literal, do a reverse lookup to find the name
* Display "input-source (input)"
* Ditto for "output-source (output)"
This displays the name configured in the test, in preference to what's in reverse DNS.
Suggested approach 2:
* Do a reverse lookup to find the name
* If this fails, use input-source or output-source instead
This preserves current behavior as closely as possible, and only differs if reverse lookup is unavailable
|
1.0
|
Display input-source name in psGraph.cgi - Original [issue 1027](https://code.google.com/p/perfsonar-ps/issues/detail?id=1027) created by arlake228 on 2014-12-22T16:54:41.000Z:
<b>What steps will reproduce the problem?</b>
Create test for a target host (foo.example) using its hostname, but which does not have reverse DNS
<b>What is the expected output? What do you see instead?</b>
What I has hoping to see in the psGraph.cgi listing is:
foo.example (192.0.2.252)
What I actually see is:
192.0.2.252 (192.0.2.252)
<b>What version of the product are you using? On what operating system?</b>
perfsonar PS 3.4.1
<b>Please provide any additional information below.</b>
If I look at the top-level JSON info in esmond:
curl http://x.x.x.x/esmond/perfsonar/archive/?format=json | python -m json.tool
I see:
"input-source": "foo.example",
so it looks like the information is available. ("input-source" is the name, and "input" is the resolved name).
If I add an entry in /etc/hosts for the remote host then the name is displayed in the psGraph.cgi output, which suggests that it is indeed doing a reverse lookup to get the name.
Suggested approach 1:
* if input-source is an IP literal, do a reverse lookup to find the name
* Display "input-source (input)"
* Ditto for "output-source (output)"
This displays the name configured in the test, in preference to what's in reverse DNS.
Suggested approach 2:
* Do a reverse lookup to find the name
* If this fails, use input-source or output-source instead
This preserves current behavior as closely as possible, and only differs if reverse lookup is unavailable
|
non_process
|
display input source name in psgraph cgi original created by on what steps will reproduce the problem create test for a target host foo example using its hostname but which does not have reverse dns what is the expected output what do you see instead what i has hoping to see in the psgraph cgi listing is foo example what i actually see is what version of the product are you using on what operating system perfsonar ps please provide any additional information below if i look at the top level json info in esmond curl python m json tool i see quot input source quot quot foo example quot so it looks like the information is available quot input source quot is the name and quot input quot is the resolved name if i add an entry in etc hosts for the remote host then the name is displayed in the psgraph cgi output which suggests that it is indeed doing a reverse lookup to get the name suggested approach if input source is an ip literal do a reverse lookup to find the name display quot input source input quot ditto for quot output source output quot this displays the name configured in the test in preference to what s in reverse dns suggested approach do a reverse lookup to find the name if this fails use input source or output source instead this preserves current behavior as closely as possible and only differs if reverse lookup is unavailable
| 0
|
77,658
| 7,594,609,378
|
IssuesEvent
|
2018-04-27 00:04:38
|
nuxsmin/sysPass
|
https://api.github.com/repos/nuxsmin/sysPass
|
closed
|
Can't edit or create new accounts on v3.0-beta (300.18042501)
|
NeedTests v3
|
When I try to update an account or create new accounts after I updated to the latest release I'm having this error:
Notice: Undefined offset: 0 in /var/www/html/syspass/lib/SP/Storage/DbWrapper.php on line 103
Argument 1 passed to SP\DataModel\Dto\AccountDetailsResponse::setUsers() must be of the type array, null given, called in /var/www/html/syspass/lib/SP/Services/Account/AccountService.php on line 105.
Here is what I had in syspass.log:
2018-04-26 12:04:22 - SQLSTATE[42S22]: Column not found: 1054 Unknown column 'AU.isEdit' in 'field list'
2018-04-26 12:04:22 - #0 /var/www/html/syspass/lib/SP/Storage/Database.php(186): PDOStatement->execute()
#1 /var/www/html/syspass/lib/SP/Storage/Database.php(117): SP\Storage\Database->prepareQueryData(Object(SP\Storage\QueryData))
#2 /var/www/html/syspass/lib/SP/Storage/DbWrapper.php(89): SP\Storage\Database->doQuery(Object(SP\Storage\QueryData))
#3 /var/www/html/syspass/lib/SP/Storage/DbWrapper.php(65): SP\Storage\DbWrapper::getResults(Object(SP\Storage\QueryData), Object(SP\Storage\Database))
#4 /var/www/html/syspass/lib/SP/Repositories/Account/AccountToUserRepository.php(189): SP\Storage\DbWrapper::getResultsArray(Object(SP\Storage\QueryData), Object(SP\Storage\Database))
#5 /var/www/html/syspass/lib/SP/Services/Account/AccountService.php(105): SP\Repositories\Account\AccountToUserRepository->getUsersByAccountId(5)
#6 /var/www/html/syspass/app/modules/web/Controllers/AccountController.php(329): SP\Services\Account\AccountService->withUsersById(Object(SP\DataModel\Dto\AccountDetailsResponse))
#7 [internal function]: SP\Modules\Web\Controllers\AccountController->editAction(5)
#8 /var/www/html/syspass/lib/SP/Bootstrap.php(239): call_user_func_array(Array, Array)
#9 [internal function]: SP\Bootstrap->SP\{closure}(Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array)
#10 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(886): call_user_func(Object(Closure), Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array)
#11 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(588): Klein\Klein->handleRouteCallback(Object(Klein\Route), Object(Klein\DataCollection\RouteCollection), Array)
#12 /var/www/html/syspass/lib/SP/Bootstrap.php(446): Klein\Klein->dispatch()
#13 /var/www/html/syspass/lib/Base.php(74): SP\Bootstrap::run(Object(DI\Container))
#14 /var/www/html/syspass/index.php(28): require('/var/www/html/s...')
#15 {main}
2018-04-26 12:04:22 - Routing error: Argument 1 passed to SP\DataModel\Dto\AccountDetailsResponse::setUsers() must be of the type array, null given, called in /var/www/html/syspass/lib/SP/Services/Account/AccountService.php on line 105
2018-04-26 12:04:22 - Routing error: #0 /var/www/html/syspass/lib/SP/Services/Account/AccountService.php(105): SP\DataModel\Dto\AccountDetailsResponse->setUsers(NULL)
#1 /var/www/html/syspass/app/modules/web/Controllers/AccountController.php(329): SP\Services\Account\AccountService->withUsersById(Object(SP\DataModel\Dto\AccountDetailsResponse))
#2 [internal function]: SP\Modules\Web\Controllers\AccountController->editAction(5)
#3 /var/www/html/syspass/lib/SP/Bootstrap.php(239): call_user_func_array(Array, Array)
#4 [internal function]: SP\Bootstrap->SP\{closure}(Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array)
#5 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(886): call_user_func(Object(Closure), Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array)
#6 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(588): Klein\Klein->handleRouteCallback(Object(Klein\Route), Object(Klein\DataCollection\RouteCollection), Array)
#7 /var/www/html/syspass/lib/SP/Bootstrap.php(446): Klein\Klein->dispatch()
#8 /var/www/html/syspass/lib/Base.php(74): SP\Bootstrap::run(Object(DI\Container))
#9 /var/www/html/syspass/index.php(28): require('/var/www/html/s...')
|
1.0
|
Can't edit or create new accounts on v3.0-beta (300.18042501) - When I try to update an account or create new accounts after I updated to the latest release I'm having this error:
Notice: Undefined offset: 0 in /var/www/html/syspass/lib/SP/Storage/DbWrapper.php on line 103
Argument 1 passed to SP\DataModel\Dto\AccountDetailsResponse::setUsers() must be of the type array, null given, called in /var/www/html/syspass/lib/SP/Services/Account/AccountService.php on line 105.
Here is what I had in syspass.log:
2018-04-26 12:04:22 - SQLSTATE[42S22]: Column not found: 1054 Unknown column 'AU.isEdit' in 'field list'
2018-04-26 12:04:22 - #0 /var/www/html/syspass/lib/SP/Storage/Database.php(186): PDOStatement->execute()
#1 /var/www/html/syspass/lib/SP/Storage/Database.php(117): SP\Storage\Database->prepareQueryData(Object(SP\Storage\QueryData))
#2 /var/www/html/syspass/lib/SP/Storage/DbWrapper.php(89): SP\Storage\Database->doQuery(Object(SP\Storage\QueryData))
#3 /var/www/html/syspass/lib/SP/Storage/DbWrapper.php(65): SP\Storage\DbWrapper::getResults(Object(SP\Storage\QueryData), Object(SP\Storage\Database))
#4 /var/www/html/syspass/lib/SP/Repositories/Account/AccountToUserRepository.php(189): SP\Storage\DbWrapper::getResultsArray(Object(SP\Storage\QueryData), Object(SP\Storage\Database))
#5 /var/www/html/syspass/lib/SP/Services/Account/AccountService.php(105): SP\Repositories\Account\AccountToUserRepository->getUsersByAccountId(5)
#6 /var/www/html/syspass/app/modules/web/Controllers/AccountController.php(329): SP\Services\Account\AccountService->withUsersById(Object(SP\DataModel\Dto\AccountDetailsResponse))
#7 [internal function]: SP\Modules\Web\Controllers\AccountController->editAction(5)
#8 /var/www/html/syspass/lib/SP/Bootstrap.php(239): call_user_func_array(Array, Array)
#9 [internal function]: SP\Bootstrap->SP\{closure}(Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array)
#10 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(886): call_user_func(Object(Closure), Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array)
#11 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(588): Klein\Klein->handleRouteCallback(Object(Klein\Route), Object(Klein\DataCollection\RouteCollection), Array)
#12 /var/www/html/syspass/lib/SP/Bootstrap.php(446): Klein\Klein->dispatch()
#13 /var/www/html/syspass/lib/Base.php(74): SP\Bootstrap::run(Object(DI\Container))
#14 /var/www/html/syspass/index.php(28): require('/var/www/html/s...')
#15 {main}
2018-04-26 12:04:22 - Routing error: Argument 1 passed to SP\DataModel\Dto\AccountDetailsResponse::setUsers() must be of the type array, null given, called in /var/www/html/syspass/lib/SP/Services/Account/AccountService.php on line 105
2018-04-26 12:04:22 - Routing error: #0 /var/www/html/syspass/lib/SP/Services/Account/AccountService.php(105): SP\DataModel\Dto\AccountDetailsResponse->setUsers(NULL)
#1 /var/www/html/syspass/app/modules/web/Controllers/AccountController.php(329): SP\Services\Account\AccountService->withUsersById(Object(SP\DataModel\Dto\AccountDetailsResponse))
#2 [internal function]: SP\Modules\Web\Controllers\AccountController->editAction(5)
#3 /var/www/html/syspass/lib/SP/Bootstrap.php(239): call_user_func_array(Array, Array)
#4 [internal function]: SP\Bootstrap->SP\{closure}(Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array)
#5 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(886): call_user_func(Object(Closure), Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array)
#6 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(588): Klein\Klein->handleRouteCallback(Object(Klein\Route), Object(Klein\DataCollection\RouteCollection), Array)
#7 /var/www/html/syspass/lib/SP/Bootstrap.php(446): Klein\Klein->dispatch()
#8 /var/www/html/syspass/lib/Base.php(74): SP\Bootstrap::run(Object(DI\Container))
#9 /var/www/html/syspass/index.php(28): require('/var/www/html/s...')
|
non_process
|
can t edit or create new accounts on beta when i try to update an account or create new accounts after i updated to the latest release i m having this error notice undefined offset in var www html syspass lib sp storage dbwrapper php on line argument passed to sp datamodel dto accountdetailsresponse setusers must be of the type array null given called in var www html syspass lib sp services account accountservice php on line here is what i had in syspass log sqlstate column not found unknown column au isedit in field list var www html syspass lib sp storage database php pdostatement execute var www html syspass lib sp storage database php sp storage database preparequerydata object sp storage querydata var www html syspass lib sp storage dbwrapper php sp storage database doquery object sp storage querydata var www html syspass lib sp storage dbwrapper php sp storage dbwrapper getresults object sp storage querydata object sp storage database var www html syspass lib sp repositories account accounttouserrepository php sp storage dbwrapper getresultsarray object sp storage querydata object sp storage database var www html syspass lib sp services account accountservice php sp repositories account accounttouserrepository getusersbyaccountid var www html syspass app modules web controllers accountcontroller php sp services account accountservice withusersbyid object sp datamodel dto accountdetailsresponse sp modules web controllers accountcontroller editaction var www html syspass lib sp bootstrap php call user func array array array sp bootstrap sp closure object klein request object klein response object klein serviceprovider object klein app object klein klein object klein datacollection routecollection array var www html syspass vendor klein klein src klein klein php call user func object closure object klein request object klein response object klein serviceprovider object klein app object klein klein object klein datacollection routecollection array var www html syspass vendor klein klein src klein klein php klein klein handleroutecallback object klein route object klein datacollection routecollection array var www html syspass lib sp bootstrap php klein klein dispatch var www html syspass lib base php sp bootstrap run object di container var www html syspass index php require var www html s main routing error argument passed to sp datamodel dto accountdetailsresponse setusers must be of the type array null given called in var www html syspass lib sp services account accountservice php on line routing error var www html syspass lib sp services account accountservice php sp datamodel dto accountdetailsresponse setusers null var www html syspass app modules web controllers accountcontroller php sp services account accountservice withusersbyid object sp datamodel dto accountdetailsresponse sp modules web controllers accountcontroller editaction var www html syspass lib sp bootstrap php call user func array array array sp bootstrap sp closure object klein request object klein response object klein serviceprovider object klein app object klein klein object klein datacollection routecollection array var www html syspass vendor klein klein src klein klein php call user func object closure object klein request object klein response object klein serviceprovider object klein app object klein klein object klein datacollection routecollection array var www html syspass vendor klein klein src klein klein php klein klein handleroutecallback object klein route object klein datacollection routecollection array var www html syspass lib sp bootstrap php klein klein dispatch var www html syspass lib base php sp bootstrap run object di container var www html syspass index php require var www html s
| 0
|
174,907
| 27,750,569,828
|
IssuesEvent
|
2023-03-15 20:22:31
|
spring-projects/spring-boot
|
https://api.github.com/repos/spring-projects/spring-boot
|
opened
|
Allow fine-grain control of the getters/setters considered in @ConfigurationProperties
|
type: enhancement status: pending-design-work
|
See #34309 for background.
Currently our `@ConfigurationProperties` only considers getter/setter combinations when generating meta-data. The binder is a little more lenient and can bind to beans that only have setters.
It would be helpful if users could indicate that certain setters without getters should be included. It would also be helpful to be able to exclude properties that have getters/setters that can't actually be set in `application.properties`. For example `SimpleMessageListenerContainer` generates meta-data that includes `connection-factory` of type `org.springframework.amqp.rabbit.connection.ConnectionFactory`.
|
1.0
|
Allow fine-grain control of the getters/setters considered in @ConfigurationProperties - See #34309 for background.
Currently our `@ConfigurationProperties` only considers getter/setter combinations when generating meta-data. The binder is a little more lenient and can bind to beans that only have setters.
It would be helpful if users could indicate that certain setters without getters should be included. It would also be helpful to be able to exclude properties that have getters/setters that can't actually be set in `application.properties`. For example `SimpleMessageListenerContainer` generates meta-data that includes `connection-factory` of type `org.springframework.amqp.rabbit.connection.ConnectionFactory`.
|
non_process
|
allow fine grain control of the getters setters considered in configurationproperties see for background currently our configurationproperties only considers getter setter combinations when generating meta data the binder is a little more lenient and can bind to beans that only have setters it would be helpful if users could indicate that certain setters without getters should be included it would also be helpful to be able to exclude properties that have getters setters that can t actually be set in application properties for example simplemessagelistenercontainer generates meta data that includes connection factory of type org springframework amqp rabbit connection connectionfactory
| 0
|
97,457
| 28,302,007,218
|
IssuesEvent
|
2023-04-10 07:10:18
|
denoland/deno
|
https://api.github.com/repos/denoland/deno
|
closed
|
Homebrew not released Deno 1.32.3 due a compile failure
|
build
|
Just curious why homebrew still does not have the latest 1.32.3, then I found https://github.com/Homebrew/homebrew-core/pull/127307
```
[v8 0.68.0] common/icudtl.dat
[v8 0.68.0] thread 'main' panicked at 'assertion failed: Command::new(gn()).arg(format!(\"--script-executable={}\",\n python())).arg(\"args\").arg(gn_out_dir).arg(\"--list\").status().unwrap().success()', v8/build.rs:218:3
```
Could anyone help to take a look?
|
1.0
|
Homebrew not released Deno 1.32.3 due a compile failure - Just curious why homebrew still does not have the latest 1.32.3, then I found https://github.com/Homebrew/homebrew-core/pull/127307
```
[v8 0.68.0] common/icudtl.dat
[v8 0.68.0] thread 'main' panicked at 'assertion failed: Command::new(gn()).arg(format!(\"--script-executable={}\",\n python())).arg(\"args\").arg(gn_out_dir).arg(\"--list\").status().unwrap().success()', v8/build.rs:218:3
```
Could anyone help to take a look?
|
non_process
|
homebrew not released deno due a compile failure just curious why homebrew still does not have the latest then i found common icudtl dat thread main panicked at assertion failed command new gn arg format script executable n python arg args arg gn out dir arg list status unwrap success build rs could anyone help to take a look
| 0
|
203,757
| 15,385,856,243
|
IssuesEvent
|
2021-03-03 07:18:47
|
pingcap/br
|
https://api.github.com/repos/pingcap/br
|
opened
|
Test testChecksumSuite.TestChecksum fails
|
component/test type/bug
|
1. What did you do?
If possible, provide a recipe for reproducing the error.
```
[2021-03-03T07:06:36.488Z] FAIL: executor_test.go:52: testChecksumSuite.TestChecksum
[2021-03-03T07:06:36.488Z]
[2021-03-03T07:06:36.488Z] executor_test.go:53:
[2021-03-03T07:06:36.488Z] c.Assert(s.mock.Start(), IsNil)
[2021-03-03T07:06:36.488Z] ... value *errors.withStack = listen tcp 0.0.0.0:41602: bind: address already in use ("listen tcp 0.0.0.0:41602: bind: address already in use")
[2021-03-03T07:06:36.488Z]
[2021-03-03T07:06:36.488Z] OOPS: 0 passed, 1 FAILED
```
https://internal.pingcap.net/idc-jenkins/blue/organizations/jenkins/br_ghpr_unit_and_integration_test/detail/br_ghpr_unit_and_integration_test/5150/pipeline#step-876-log-270
|
1.0
|
Test testChecksumSuite.TestChecksum fails - 1. What did you do?
If possible, provide a recipe for reproducing the error.
```
[2021-03-03T07:06:36.488Z] FAIL: executor_test.go:52: testChecksumSuite.TestChecksum
[2021-03-03T07:06:36.488Z]
[2021-03-03T07:06:36.488Z] executor_test.go:53:
[2021-03-03T07:06:36.488Z] c.Assert(s.mock.Start(), IsNil)
[2021-03-03T07:06:36.488Z] ... value *errors.withStack = listen tcp 0.0.0.0:41602: bind: address already in use ("listen tcp 0.0.0.0:41602: bind: address already in use")
[2021-03-03T07:06:36.488Z]
[2021-03-03T07:06:36.488Z] OOPS: 0 passed, 1 FAILED
```
https://internal.pingcap.net/idc-jenkins/blue/organizations/jenkins/br_ghpr_unit_and_integration_test/detail/br_ghpr_unit_and_integration_test/5150/pipeline#step-876-log-270
|
non_process
|
test testchecksumsuite testchecksum fails what did you do if possible provide a recipe for reproducing the error fail executor test go testchecksumsuite testchecksum executor test go c assert s mock start isnil value errors withstack listen tcp bind address already in use listen tcp bind address already in use oops passed failed
| 0
|
9,875
| 12,886,325,043
|
IssuesEvent
|
2020-07-13 09:18:36
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Missing source and CI name information from Zeit Now GitHub integration app
|
CI: Zeit cli process: dependencies stage: awaiting external fix
|
When recording tests from Zeit Now GitHub application on a pull request, the dashboard is missing all Git commit information and CI name
See https://dashboard.cypress.io/#/projects/6v6qq1/runs/4/specs and the blog post https://www.cypress.io/blog/2018/08/28/Record-Test-Artifacts-From-Any-CI/
<img width="1024" alt="screen shot 2018-09-05 at 9 19 00 am" src="https://user-images.githubusercontent.com/2212006/45095815-ff7db900-b0ec-11e8-96d4-f9db27abb78b.png">
|
1.0
|
Missing source and CI name information from Zeit Now GitHub integration app - When recording tests from Zeit Now GitHub application on a pull request, the dashboard is missing all Git commit information and CI name
See https://dashboard.cypress.io/#/projects/6v6qq1/runs/4/specs and the blog post https://www.cypress.io/blog/2018/08/28/Record-Test-Artifacts-From-Any-CI/
<img width="1024" alt="screen shot 2018-09-05 at 9 19 00 am" src="https://user-images.githubusercontent.com/2212006/45095815-ff7db900-b0ec-11e8-96d4-f9db27abb78b.png">
|
process
|
missing source and ci name information from zeit now github integration app when recording tests from zeit now github application on a pull request the dashboard is missing all git commit information and ci name see and the blog post img width alt screen shot at am src
| 1
|
182,070
| 14,101,302,798
|
IssuesEvent
|
2020-11-06 06:32:24
|
JuliaLang/julia
|
https://api.github.com/repos/JuliaLang/julia
|
closed
|
Provide a way to print the reports of the nested tests
|
testsystem
|
Hi guys!
I have a package with lots of tests. I am using nested test sets to improve organization, something like:
```julia
@testset "Test Set" begin
@testset "Test Sub Set"
...
end
```
However, if nothing fails, then nothing about the subtests is printed. This is not good, because I cannot be sure if every test was executed. Is it possible to provide an option to `@testset` print the reports of the subsets all the time?
|
1.0
|
Provide a way to print the reports of the nested tests - Hi guys!
I have a package with lots of tests. I am using nested test sets to improve organization, something like:
```julia
@testset "Test Set" begin
@testset "Test Sub Set"
...
end
```
However, if nothing fails, then nothing about the subtests is printed. This is not good, because I cannot be sure if every test was executed. Is it possible to provide an option to `@testset` print the reports of the subsets all the time?
|
non_process
|
provide a way to print the reports of the nested tests hi guys i have a package with lots of tests i am using nested test sets to improve organization something like julia testset test set begin testset test sub set end however if nothing fails then nothing about the subtests is printed this is not good because i cannot be sure if every test was executed is it possible to provide an option to testset print the reports of the subsets all the time
| 0
|
2,098
| 4,932,119,019
|
IssuesEvent
|
2016-11-28 12:36:04
|
Alfresco/alfresco-ng2-components
|
https://api.github.com/repos/Alfresco/alfresco-ng2-components
|
opened
|
Typeahead suggestions options are cut short on form with attached to start event, if last widget on form
|
browser: all bug comp: activiti-processList
|
1. Import process
2. Start a process
N.b. does not occur with the dropdown widget
**App to import**
[typeaheads.zip](https://github.com/Alfresco/alfresco-ng2-components/files/616240/typeaheads.zip)

|
1.0
|
Typeahead suggestions options are cut short on form with attached to start event, if last widget on form - 1. Import process
2. Start a process
N.b. does not occur with the dropdown widget
**App to import**
[typeaheads.zip](https://github.com/Alfresco/alfresco-ng2-components/files/616240/typeaheads.zip)

|
process
|
typeahead suggestions options are cut short on form with attached to start event if last widget on form import process start a process n b does not occur with the dropdown widget app to import
| 1
|
197,852
| 22,607,331,083
|
IssuesEvent
|
2022-06-29 14:15:27
|
elikkatzgit/TestingPOM
|
https://api.github.com/repos/elikkatzgit/TestingPOM
|
opened
|
WS-2021-0153 (High) detected in ejs-2.7.4.tgz
|
security vulnerability
|
## WS-2021-0153 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ejs-2.7.4.tgz</b></p></summary>
<p>Embedded JavaScript templates</p>
<p>Library home page: <a href="https://registry.npmjs.org/ejs/-/ejs-2.7.4.tgz">https://registry.npmjs.org/ejs/-/ejs-2.7.4.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/ejs/package.json</p>
<p>
Dependency Hierarchy:
- webpack-bundle-analyzer-3.3.2.tgz (Root Library)
- :x: **ejs-2.7.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/TestingPOM/commit/349fffeed7cf25f2cf5b8b6a05b5e4367130406e">349fffeed7cf25f2cf5b8b6a05b5e4367130406e</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Arbitrary Code Injection vulnerability was found in ejs before 3.1.6. Caused by filename which isn't sanitized for display.
<p>Publish Date: 2021-01-22
<p>URL: <a href=https://github.com/mde/ejs/commit/abaee2be937236b1b8da9a1f55096c17dda905fd>WS-2021-0153</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-01-22</p>
<p>Fix Resolution (ejs): 3.1.6</p>
<p>Direct dependency fix Resolution (webpack-bundle-analyzer): 4.0.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
|
True
|
WS-2021-0153 (High) detected in ejs-2.7.4.tgz - ## WS-2021-0153 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ejs-2.7.4.tgz</b></p></summary>
<p>Embedded JavaScript templates</p>
<p>Library home page: <a href="https://registry.npmjs.org/ejs/-/ejs-2.7.4.tgz">https://registry.npmjs.org/ejs/-/ejs-2.7.4.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/ejs/package.json</p>
<p>
Dependency Hierarchy:
- webpack-bundle-analyzer-3.3.2.tgz (Root Library)
- :x: **ejs-2.7.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/TestingPOM/commit/349fffeed7cf25f2cf5b8b6a05b5e4367130406e">349fffeed7cf25f2cf5b8b6a05b5e4367130406e</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Arbitrary Code Injection vulnerability was found in ejs before 3.1.6. Caused by filename which isn't sanitized for display.
<p>Publish Date: 2021-01-22
<p>URL: <a href=https://github.com/mde/ejs/commit/abaee2be937236b1b8da9a1f55096c17dda905fd>WS-2021-0153</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-01-22</p>
<p>Fix Resolution (ejs): 3.1.6</p>
<p>Direct dependency fix Resolution (webpack-bundle-analyzer): 4.0.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
|
non_process
|
ws high detected in ejs tgz ws high severity vulnerability vulnerable library ejs tgz embedded javascript templates library home page a href path to dependency file package json path to vulnerable library node modules ejs package json dependency hierarchy webpack bundle analyzer tgz root library x ejs tgz vulnerable library found in head commit a href found in base branch dev vulnerability details arbitrary code injection vulnerability was found in ejs before caused by filename which isn t sanitized for display publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution ejs direct dependency fix resolution webpack bundle analyzer rescue worker helmet automatic remediation is available for this issue
| 0
|
15,313
| 19,406,013,845
|
IssuesEvent
|
2021-12-20 00:48:45
|
emily-writes-poems/emily-writes-poems-processing
|
https://api.github.com/repos/emily-writes-poems/emily-writes-poems-processing
|
opened
|
migrate: create new feature
|
script migration processing
|
build Javascript/Electron functionality from existing Python script mongo_feature.py
insert new feature given poem id, poem title, featured text and whether the feature should be the current feature (optional)
|
1.0
|
migrate: create new feature - build Javascript/Electron functionality from existing Python script mongo_feature.py
insert new feature given poem id, poem title, featured text and whether the feature should be the current feature (optional)
|
process
|
migrate create new feature build javascript electron functionality from existing python script mongo feature py insert new feature given poem id poem title featured text and whether the feature should be the current feature optional
| 1
|
18,656
| 24,581,325,103
|
IssuesEvent
|
2022-10-13 15:49:38
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[FHIR] Questionnaire resources >JSON > 'Initial' key not getting displayed when admin adds placeholder text in the study builder
|
Bug P1 Response datastore Process: Fixed Process: Tested dev
|
Questionnaire resources > 'Initial' key not getting displayed when admin adds placeholder text in the study builder
|
2.0
|
[FHIR] Questionnaire resources >JSON > 'Initial' key not getting displayed when admin adds placeholder text in the study builder - Questionnaire resources > 'Initial' key not getting displayed when admin adds placeholder text in the study builder
|
process
|
questionnaire resources json initial key not getting displayed when admin adds placeholder text in the study builder questionnaire resources initial key not getting displayed when admin adds placeholder text in the study builder
| 1
|
3,340
| 6,474,890,092
|
IssuesEvent
|
2017-08-17 19:06:52
|
Azure/azure-event-hubs-java
|
https://api.github.com/repos/Azure/azure-event-hubs-java
|
closed
|
log exceptions in one line
|
EventProcessorHost
|
This is an issue filed by customer in an internal thread:
Logging stack traces
– when the EPH logs error messages, it logs every line of the stack trace separately, Instead of logging everything in one log message. Why?
This makes is harder to understand what happened, especially when there are multiple partitions and each partition may log errors in parallel.

|
1.0
|
log exceptions in one line - This is an issue filed by customer in an internal thread:
Logging stack traces
– when the EPH logs error messages, it logs every line of the stack trace separately, Instead of logging everything in one log message. Why?
This makes is harder to understand what happened, especially when there are multiple partitions and each partition may log errors in parallel.

|
process
|
log exceptions in one line this is an issue filed by customer in an internal thread logging stack traces – when the eph logs error messages it logs every line of the stack trace separately instead of logging everything in one log message why this makes is harder to understand what happened especially when there are multiple partitions and each partition may log errors in parallel
| 1
|
515,307
| 14,959,702,120
|
IssuesEvent
|
2021-01-27 03:53:22
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
[Coverity CID :216793] Division or modulo by zero in tests/ztest/error_hook/src/main.c
|
Coverity bug priority: low
|
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/f91e9fba51e5da46ee5c6822f8656713d74a6ecf/tests/ztest/error_hook/src/main.c#L68
Category: Integer handling issues
Function: `trigger_fault_divide_zero`
Component: Tests
CID: [216793](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=216793)
Details:
```
62 static void trigger_fault_divide_zero(void)
63 {
64 int a = 1;
65 int b = 0;
66
67 /* divde zero */
>>> CID 216793: Integer handling issues (DIVIDE_BY_ZERO)
>>> In expression "a / b", division by expression "b" which may be zero has undefined behavior.
68 a = a / b;
69 printk("a is %d\n", a);
70 }
71
72 static void trigger_fault_oops(void)
73 {
```
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
|
1.0
|
[Coverity CID :216793] Division or modulo by zero in tests/ztest/error_hook/src/main.c -
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/f91e9fba51e5da46ee5c6822f8656713d74a6ecf/tests/ztest/error_hook/src/main.c#L68
Category: Integer handling issues
Function: `trigger_fault_divide_zero`
Component: Tests
CID: [216793](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=216793)
Details:
```
62 static void trigger_fault_divide_zero(void)
63 {
64 int a = 1;
65 int b = 0;
66
67 /* divde zero */
>>> CID 216793: Integer handling issues (DIVIDE_BY_ZERO)
>>> In expression "a / b", division by expression "b" which may be zero has undefined behavior.
68 a = a / b;
69 printk("a is %d\n", a);
70 }
71
72 static void trigger_fault_oops(void)
73 {
```
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
|
non_process
|
division or modulo by zero in tests ztest error hook src main c static code scan issues found in file category integer handling issues function trigger fault divide zero component tests cid details static void trigger fault divide zero void int a int b divde zero cid integer handling issues divide by zero in expression a b division by expression b which may be zero has undefined behavior a a b printk a is d n a static void trigger fault oops void please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file
| 0
|
7,952
| 11,137,562,422
|
IssuesEvent
|
2019-12-20 19:42:17
|
openopps/openopps-platform
|
https://api.github.com/repos/openopps/openopps-platform
|
closed
|
Allow for sorting of education data in Open Opps application
|
Apply Process Approved Requirements Ready State Dept.
|
Who: Applicants
What: ability to sort education data
Why: in order to highlight specific education
Acceptance Criteria:
- Education data will be pulled from USAJOBS and displayed in the sort order that comes over. Allow for applicants to sort that data in Open Opps.
- Add arrow sorting per design system.
Sample of USAJOBS education sort arrows:

|
1.0
|
Allow for sorting of education data in Open Opps application - Who: Applicants
What: ability to sort education data
Why: in order to highlight specific education
Acceptance Criteria:
- Education data will be pulled from USAJOBS and displayed in the sort order that comes over. Allow for applicants to sort that data in Open Opps.
- Add arrow sorting per design system.
Sample of USAJOBS education sort arrows:

|
process
|
allow for sorting of education data in open opps application who applicants what ability to sort education data why in order to highlight specific education acceptance criteria education data will be pulled from usajobs and displayed in the sort order that comes over allow for applicants to sort that data in open opps add arrow sorting per design system sample of usajobs education sort arrows
| 1
|
5,159
| 7,933,331,646
|
IssuesEvent
|
2018-07-08 04:14:13
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
closed
|
Speedup for displaying non-permanent data
|
libs-etherlib status-inprocess type-enhancement
|
getTrans -v -t 2283440.0 is very slow. Mostly because it creates the trace array (which is huge) in memory by parsing the result of the curl call. If, instead, the curl call was parsed and displayed directly for traces (which we never store on disc anyway), we could (a) start the display immediately, (b) not have to allocate all that memory just to throw it away, (c) not have to build an array
|
1.0
|
Speedup for displaying non-permanent data - getTrans -v -t 2283440.0 is very slow. Mostly because it creates the trace array (which is huge) in memory by parsing the result of the curl call. If, instead, the curl call was parsed and displayed directly for traces (which we never store on disc anyway), we could (a) start the display immediately, (b) not have to allocate all that memory just to throw it away, (c) not have to build an array
|
process
|
speedup for displaying non permanent data gettrans v t is very slow mostly because it creates the trace array which is huge in memory by parsing the result of the curl call if instead the curl call was parsed and displayed directly for traces which we never store on disc anyway we could a start the display immediately b not have to allocate all that memory just to throw it away c not have to build an array
| 1
|
17,954
| 23,959,917,599
|
IssuesEvent
|
2022-09-12 18:07:08
|
2i2c-org/infrastructure
|
https://api.github.com/repos/2i2c-org/infrastructure
|
closed
|
Design a secure method to receive secret keys
|
type: enhancement :label: team-process
|
### Description
Sometimes, we need to receive secret info (like OAuth keys!) from community champions / university IT to finish setting up our infrastructure. This needs to be encrypted both at rest and in transit.
### Value / benefit
- We can receive secrets in a secret fashion
- We should receive it in a way that allows anyone from the 2i2c team to decrypt it, and isn't tied to anyone in particular
### Implementation details
_No response_
### Tasks to complete
- [ ] Choose what kinds of transfer options we want to officially recommend
- [ ] Work out a process for each and document it
### Updates
_No response_
|
1.0
|
Design a secure method to receive secret keys - ### Description
Sometimes, we need to receive secret info (like OAuth keys!) from community champions / university IT to finish setting up our infrastructure. This needs to be encrypted both at rest and in transit.
### Value / benefit
- We can receive secrets in a secret fashion
- We should receive it in a way that allows anyone from the 2i2c team to decrypt it, and isn't tied to anyone in particular
### Implementation details
_No response_
### Tasks to complete
- [ ] Choose what kinds of transfer options we want to officially recommend
- [ ] Work out a process for each and document it
### Updates
_No response_
|
process
|
design a secure method to receive secret keys description sometimes we need to receive secret info like oauth keys from community champions university it to finish setting up our infrastructure this needs to be encrypted both at rest and in transit value benefit we can receive secrets in a secret fashion we should receive it in a way that allows anyone from the team to decrypt it and isn t tied to anyone in particular implementation details no response tasks to complete choose what kinds of transfer options we want to officially recommend work out a process for each and document it updates no response
| 1
|
17,620
| 23,436,972,351
|
IssuesEvent
|
2022-08-15 11:00:17
|
Tencent/tdesign-miniprogram
|
https://api.github.com/repos/Tencent/tdesign-miniprogram
|
closed
|
[upload] 上传失败无状态
|
bug in process
|
### tdesign-miniprogram 版本
latest
### 重现链接
_No response_
### 重现步骤

上传失败,无法触发重试
### 期望结果
_No response_
### 实际结果
上传失败,无法触发重试
### 框架版本
_No response_
### 浏览器版本
_No response_
### 系统版本
_No response_
### Node版本
_No response_
### 补充说明
_No response_
|
1.0
|
[upload] 上传失败无状态 - ### tdesign-miniprogram 版本
latest
### 重现链接
_No response_
### 重现步骤

上传失败,无法触发重试
### 期望结果
_No response_
### 实际结果
上传失败,无法触发重试
### 框架版本
_No response_
### 浏览器版本
_No response_
### 系统版本
_No response_
### Node版本
_No response_
### 补充说明
_No response_
|
process
|
上传失败无状态 tdesign miniprogram 版本 latest 重现链接 no response 重现步骤 上传失败,无法触发重试 期望结果 no response 实际结果 上传失败,无法触发重试 框架版本 no response 浏览器版本 no response 系统版本 no response node版本 no response 补充说明 no response
| 1
|
283,106
| 30,889,582,573
|
IssuesEvent
|
2023-08-04 02:56:39
|
maddyCode23/linux-4.1.15
|
https://api.github.com/repos/maddyCode23/linux-4.1.15
|
reopened
|
CVE-2021-0448 (Medium) detected in linux-stable-rtv4.1.33
|
Mend: dependency security vulnerability
|
## CVE-2021-0448 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_conntrack_netlink.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A security vulnerability was found in Linux Kernel before 4.4.239, 4.9.239, 4.14.201, 4.19.150, 5.4.70, and 5.8.13. Missing range check for l3/l4 protonum in netfilter.c
<p>Publish Date: 2020-11-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-0448>CVE-2021-0448</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-0448">https://www.linuxkernelcves.com/cves/CVE-2021-0448</a></p>
<p>Release Date: 2020-11-07</p>
<p>Fix Resolution: v4.4.239,v4.9.239,v4.14.201,v4.19.150,v5.4.70,v5.8.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-0448 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2021-0448 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_conntrack_netlink.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A security vulnerability was found in Linux Kernel before 4.4.239, 4.9.239, 4.14.201, 4.19.150, 5.4.70, and 5.8.13. Missing range check for l3/l4 protonum in netfilter.c
<p>Publish Date: 2020-11-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-0448>CVE-2021-0448</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-0448">https://www.linuxkernelcves.com/cves/CVE-2021-0448</a></p>
<p>Release Date: 2020-11-07</p>
<p>Fix Resolution: v4.4.239,v4.9.239,v4.14.201,v4.19.150,v5.4.70,v5.8.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net netfilter nf conntrack netlink c vulnerability details a security vulnerability was found in linux kernel before and missing range check for protonum in netfilter c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
168,285
| 14,144,998,162
|
IssuesEvent
|
2020-11-10 17:08:43
|
andreascalzo99/ProgettoBlackFriendly
|
https://api.github.com/repos/andreascalzo99/ProgettoBlackFriendly
|
closed
|
Aggiunta Diagrammi UML
|
documentation
|
Sviluppare **diagramma delle classi con prospettiva software** e **diagramma dei package**.
|
1.0
|
Aggiunta Diagrammi UML - Sviluppare **diagramma delle classi con prospettiva software** e **diagramma dei package**.
|
non_process
|
aggiunta diagrammi uml sviluppare diagramma delle classi con prospettiva software e diagramma dei package
| 0
|
255,728
| 19,323,729,591
|
IssuesEvent
|
2021-12-14 09:10:05
|
kubernetes-sigs/cluster-api-provider-kubevirt
|
https://api.github.com/repos/kubernetes-sigs/cluster-api-provider-kubevirt
|
closed
|
Typo in README.md
|
good first issue help wanted kind/documentation
|
The `README.md` file has a simple typo at `line 48`. A go-through of the whole .md file is advised in order to correct any other typos.
|
1.0
|
Typo in README.md - The `README.md` file has a simple typo at `line 48`. A go-through of the whole .md file is advised in order to correct any other typos.
|
non_process
|
typo in readme md the readme md file has a simple typo at line a go through of the whole md file is advised in order to correct any other typos
| 0
|
14,380
| 17,401,179,228
|
IssuesEvent
|
2021-08-02 19:56:23
|
googleapis/python-storage
|
https://api.github.com/repos/googleapis/python-storage
|
opened
|
New 'page_size' argument breaks unit tests under Python 3.6
|
type: process
|
PR #520 added passing `page_size` to the `google.api_core.page_iterator.HTTPIterator` constructor, which is a feature added only in `google-api-core 1.29.0`. `testing/constraints-3.6.txt` pins `google-cloud-core==1.6.0`, the minimum version specified in `setup.py`, but does not pin `google-api-core`, which results in:
```bash
$ git remote -v
origin git@github.com:googleapis/python-storage (fetch)
origin git@github.com:googleapis/python-storage (push)
$ git status
On branch master
Your branch is up to date with 'origin/master'.
nothing to commit, working tree clean
$ git log -1
commit 4abb40310eca7ec45afc4bc5e4dfafbe083e74d2 (HEAD -> master, origin/master, origin/HEAD)
Author: Tres Seaver <tseaver@palladion.com>
Date: Tue Jul 27 13:43:42 2021 -0400
fix: make 'requests.exceptions.ChunkedEncodingError retryable by default (#526)
Closes #525.
$ .nox/unit-3-6/bin/pip list | grep google
google-api-core 1.27.0
google-auth 1.24.0
google-cloud-core 1.6.0
google-cloud-storage 1.41.1 /path/to/python-storage
google-crc32c 1.1.2
google-resumable-media 1.3.0
googleapis-common-protos 1.53.0
$ nox -re unit-3.6
nox > Running session unit-3.6
nox > Creating virtual environment (virtualenv) using python3.6 in .nox/unit-3-6
nox > python -m pip install mock pytest pytest-cov -c /home/tseaver/projects/agendaless/Google/src/python-storage/testing/constraints-3.6.txt
nox > python -m pip install -e . -c /home/tseaver/projects/agendaless/Google/src/python-storage/testing/constraints-3.6.txt
nox > py.test --quiet --cov=google.cloud.storage --cov=google.cloud --cov=tests.unit --cov-append --cov-config=.coveragerc --cov-report= --cov-fail-under=0 tests/unit
..............................................s......................... [ 6%]
........................................................................ [ 13%]
........................................................................ [ 20%]
........................................................................ [ 27%]
........................................................................ [ 34%]
........................................................................ [ 41%]
........................................................................ [ 48%]
........................................................................ [ 55%]
........................................................................ [ 62%]
........................................................................ [ 69%]
........................................................................ [ 76%]
.........................FF............................................. [ 83%]
........................................................................ [ 90%]
........................................................................ [ 97%]
..................... [100%]Coverage.py warning: Module google.cloud was previously imported, but not measured (module-not-measured)
=================================== FAILURES ===================================
__________________ TestClient.test__list_resource_w_defaults ___________________
self = <tests.unit.test_client.TestClient testMethod=test__list_resource_w_defaults>
def test__list_resource_w_defaults(self):
import functools
from google.api_core.page_iterator import HTTPIterator
from google.api_core.page_iterator import _do_nothing_page_start
project = "PROJECT"
path = "/path/to/list/resource"
item_to_value = mock.Mock(spec=[])
credentials = _make_credentials()
client = self._make_one(project=project, credentials=credentials)
connection = client._base_connection = _make_connection()
> iterator = client._list_resource(path=path, item_to_value=item_to_value,)
tests/unit/test_client.py:485:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.cloud.storage.client.Client object at 0x7efcfb9f3b38>
path = '/path/to/list/resource', item_to_value = <Mock id='139625018221344'>
page_token = None, max_results = None, extra_params = None
page_start = <function _do_nothing_page_start at 0x7efcfc69fae8>
page_size = None, timeout = 60
retry = <google.api_core.retry.Retry object at 0x7efcfc756c50>
def _list_resource(
self,
path,
item_to_value,
page_token=None,
max_results=None,
extra_params=None,
page_start=page_iterator._do_nothing_page_start,
page_size=None,
timeout=_DEFAULT_TIMEOUT,
retry=DEFAULT_RETRY,
):
api_request = functools.partial(
self._connection.api_request, timeout=timeout, retry=retry
)
return page_iterator.HTTPIterator(
client=self,
api_request=api_request,
path=path,
item_to_value=item_to_value,
page_token=page_token,
max_results=max_results,
extra_params=extra_params,
page_start=page_start,
> page_size=page_size,
)
E TypeError: __init__() got an unexpected keyword argument 'page_size'
google/cloud/storage/client.py:403: TypeError
__________________ TestClient.test__list_resource_w_explicit ___________________
self = <tests.unit.test_client.TestClient testMethod=test__list_resource_w_explicit>
def test__list_resource_w_explicit(self):
import functools
from google.api_core.page_iterator import HTTPIterator
project = "PROJECT"
path = "/path/to/list/resource"
item_to_value = mock.Mock(spec=[])
page_token = "PAGE-TOKEN"
max_results = 47
extra_params = {"foo": "Foo"}
page_start = mock.Mock(spec=[])
credentials = _make_credentials()
client = self._make_one(project=project, credentials=credentials)
connection = client._base_connection = _make_connection()
iterator = client._list_resource(
path=path,
item_to_value=item_to_value,
page_token=page_token,
max_results=max_results,
extra_params=extra_params,
> page_start=page_start,
)
tests/unit/test_client.py:523:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.cloud.storage.client.Client object at 0x7efcfbdfe470>
path = '/path/to/list/resource', item_to_value = <Mock id='139625022614160'>
page_token = 'PAGE-TOKEN', max_results = 47, extra_params = {'foo': 'Foo'}
page_start = <Mock id='139625022612368'>, page_size = None, timeout = 60
retry = <google.api_core.retry.Retry object at 0x7efcfc756c50>
def _list_resource(
self,
path,
item_to_value,
page_token=None,
max_results=None,
extra_params=None,
page_start=page_iterator._do_nothing_page_start,
page_size=None,
timeout=_DEFAULT_TIMEOUT,
retry=DEFAULT_RETRY,
):
api_request = functools.partial(
self._connection.api_request, timeout=timeout, retry=retry
)
return page_iterator.HTTPIterator(
client=self,
api_request=api_request,
path=path,
item_to_value=item_to_value,
page_token=page_token,
max_results=max_results,
extra_params=extra_params,
page_start=page_start,
> page_size=page_size,
)
E TypeError: __init__() got an unexpected keyword argument 'page_size'
google/cloud/storage/client.py:403: TypeError
=========================== short test summary info ============================
FAILED tests/unit/test_client.py::TestClient::test__list_resource_w_defaults
FAILED tests/unit/test_client.py::TestClient::test__list_resource_w_explicit
2 failed, 1026 passed, 1 skipped in 19.16s
nox > Command py.test --quiet --cov=google.cloud.storage --cov=google.cloud --cov=tests.unit --cov-append --cov-config=.coveragerc --cov-report= --cov-fail-under=0 tests/unit failed with exit code 1
nox > Session unit-3.6 failed.
```
The fix would be to add ranges for `google-api-core` and `google-auth` (another transitive dependency) to `setup.py`, and then pin the minimum versions for those ranges in `testing/constraints-3.6.txt`.
|
1.0
|
New 'page_size' argument breaks unit tests under Python 3.6 - PR #520 added passing `page_size` to the `google.api_core.page_iterator.HTTPIterator` constructor, which is a feature added only in `google-api-core 1.29.0`. `testing/constraints-3.6.txt` pins `google-cloud-core==1.6.0`, the minimum version specified in `setup.py`, but does not pin `google-api-core`, which results in:
```bash
$ git remote -v
origin git@github.com:googleapis/python-storage (fetch)
origin git@github.com:googleapis/python-storage (push)
$ git status
On branch master
Your branch is up to date with 'origin/master'.
nothing to commit, working tree clean
$ git log -1
commit 4abb40310eca7ec45afc4bc5e4dfafbe083e74d2 (HEAD -> master, origin/master, origin/HEAD)
Author: Tres Seaver <tseaver@palladion.com>
Date: Tue Jul 27 13:43:42 2021 -0400
fix: make 'requests.exceptions.ChunkedEncodingError retryable by default (#526)
Closes #525.
$ .nox/unit-3-6/bin/pip list | grep google
google-api-core 1.27.0
google-auth 1.24.0
google-cloud-core 1.6.0
google-cloud-storage 1.41.1 /path/to/python-storage
google-crc32c 1.1.2
google-resumable-media 1.3.0
googleapis-common-protos 1.53.0
$ nox -re unit-3.6
nox > Running session unit-3.6
nox > Creating virtual environment (virtualenv) using python3.6 in .nox/unit-3-6
nox > python -m pip install mock pytest pytest-cov -c /home/tseaver/projects/agendaless/Google/src/python-storage/testing/constraints-3.6.txt
nox > python -m pip install -e . -c /home/tseaver/projects/agendaless/Google/src/python-storage/testing/constraints-3.6.txt
nox > py.test --quiet --cov=google.cloud.storage --cov=google.cloud --cov=tests.unit --cov-append --cov-config=.coveragerc --cov-report= --cov-fail-under=0 tests/unit
..............................................s......................... [ 6%]
........................................................................ [ 13%]
........................................................................ [ 20%]
........................................................................ [ 27%]
........................................................................ [ 34%]
........................................................................ [ 41%]
........................................................................ [ 48%]
........................................................................ [ 55%]
........................................................................ [ 62%]
........................................................................ [ 69%]
........................................................................ [ 76%]
.........................FF............................................. [ 83%]
........................................................................ [ 90%]
........................................................................ [ 97%]
..................... [100%]Coverage.py warning: Module google.cloud was previously imported, but not measured (module-not-measured)
=================================== FAILURES ===================================
__________________ TestClient.test__list_resource_w_defaults ___________________
self = <tests.unit.test_client.TestClient testMethod=test__list_resource_w_defaults>
def test__list_resource_w_defaults(self):
import functools
from google.api_core.page_iterator import HTTPIterator
from google.api_core.page_iterator import _do_nothing_page_start
project = "PROJECT"
path = "/path/to/list/resource"
item_to_value = mock.Mock(spec=[])
credentials = _make_credentials()
client = self._make_one(project=project, credentials=credentials)
connection = client._base_connection = _make_connection()
> iterator = client._list_resource(path=path, item_to_value=item_to_value,)
tests/unit/test_client.py:485:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.cloud.storage.client.Client object at 0x7efcfb9f3b38>
path = '/path/to/list/resource', item_to_value = <Mock id='139625018221344'>
page_token = None, max_results = None, extra_params = None
page_start = <function _do_nothing_page_start at 0x7efcfc69fae8>
page_size = None, timeout = 60
retry = <google.api_core.retry.Retry object at 0x7efcfc756c50>
def _list_resource(
self,
path,
item_to_value,
page_token=None,
max_results=None,
extra_params=None,
page_start=page_iterator._do_nothing_page_start,
page_size=None,
timeout=_DEFAULT_TIMEOUT,
retry=DEFAULT_RETRY,
):
api_request = functools.partial(
self._connection.api_request, timeout=timeout, retry=retry
)
return page_iterator.HTTPIterator(
client=self,
api_request=api_request,
path=path,
item_to_value=item_to_value,
page_token=page_token,
max_results=max_results,
extra_params=extra_params,
page_start=page_start,
> page_size=page_size,
)
E TypeError: __init__() got an unexpected keyword argument 'page_size'
google/cloud/storage/client.py:403: TypeError
__________________ TestClient.test__list_resource_w_explicit ___________________
self = <tests.unit.test_client.TestClient testMethod=test__list_resource_w_explicit>
def test__list_resource_w_explicit(self):
import functools
from google.api_core.page_iterator import HTTPIterator
project = "PROJECT"
path = "/path/to/list/resource"
item_to_value = mock.Mock(spec=[])
page_token = "PAGE-TOKEN"
max_results = 47
extra_params = {"foo": "Foo"}
page_start = mock.Mock(spec=[])
credentials = _make_credentials()
client = self._make_one(project=project, credentials=credentials)
connection = client._base_connection = _make_connection()
iterator = client._list_resource(
path=path,
item_to_value=item_to_value,
page_token=page_token,
max_results=max_results,
extra_params=extra_params,
> page_start=page_start,
)
tests/unit/test_client.py:523:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.cloud.storage.client.Client object at 0x7efcfbdfe470>
path = '/path/to/list/resource', item_to_value = <Mock id='139625022614160'>
page_token = 'PAGE-TOKEN', max_results = 47, extra_params = {'foo': 'Foo'}
page_start = <Mock id='139625022612368'>, page_size = None, timeout = 60
retry = <google.api_core.retry.Retry object at 0x7efcfc756c50>
def _list_resource(
self,
path,
item_to_value,
page_token=None,
max_results=None,
extra_params=None,
page_start=page_iterator._do_nothing_page_start,
page_size=None,
timeout=_DEFAULT_TIMEOUT,
retry=DEFAULT_RETRY,
):
api_request = functools.partial(
self._connection.api_request, timeout=timeout, retry=retry
)
return page_iterator.HTTPIterator(
client=self,
api_request=api_request,
path=path,
item_to_value=item_to_value,
page_token=page_token,
max_results=max_results,
extra_params=extra_params,
page_start=page_start,
> page_size=page_size,
)
E TypeError: __init__() got an unexpected keyword argument 'page_size'
google/cloud/storage/client.py:403: TypeError
=========================== short test summary info ============================
FAILED tests/unit/test_client.py::TestClient::test__list_resource_w_defaults
FAILED tests/unit/test_client.py::TestClient::test__list_resource_w_explicit
2 failed, 1026 passed, 1 skipped in 19.16s
nox > Command py.test --quiet --cov=google.cloud.storage --cov=google.cloud --cov=tests.unit --cov-append --cov-config=.coveragerc --cov-report= --cov-fail-under=0 tests/unit failed with exit code 1
nox > Session unit-3.6 failed.
```
The fix would be to add ranges for `google-api-core` and `google-auth` (another transitive dependency) to `setup.py`, and then pin the minimum versions for those ranges in `testing/constraints-3.6.txt`.
|
process
|
new page size argument breaks unit tests under python pr added passing page size to the google api core page iterator httpiterator constructor which is a feature added only in google api core testing constraints txt pins google cloud core the minimum version specified in setup py but does not pin google api core which results in bash git remote v origin git github com googleapis python storage fetch origin git github com googleapis python storage push git status on branch master your branch is up to date with origin master nothing to commit working tree clean git log commit head master origin master origin head author tres seaver date tue jul fix make requests exceptions chunkedencodingerror retryable by default closes nox unit bin pip list grep google google api core google auth google cloud core google cloud storage path to python storage google google resumable media googleapis common protos nox re unit nox running session unit nox creating virtual environment virtualenv using in nox unit nox python m pip install mock pytest pytest cov c home tseaver projects agendaless google src python storage testing constraints txt nox python m pip install e c home tseaver projects agendaless google src python storage testing constraints txt nox py test quiet cov google cloud storage cov google cloud cov tests unit cov append cov config coveragerc cov report cov fail under tests unit s ff coverage py warning module google cloud was previously imported but not measured module not measured failures testclient test list resource w defaults self def test list resource w defaults self import functools from google api core page iterator import httpiterator from google api core page iterator import do nothing page start project project path path to list resource item to value mock mock spec credentials make credentials client self make one project project credentials credentials connection client base connection make connection iterator client list resource path path item to value item to value tests unit test client py self path path to list resource item to value page token none max results none extra params none page start page size none timeout retry def list resource self path item to value page token none max results none extra params none page start page iterator do nothing page start page size none timeout default timeout retry default retry api request functools partial self connection api request timeout timeout retry retry return page iterator httpiterator client self api request api request path path item to value item to value page token page token max results max results extra params extra params page start page start page size page size e typeerror init got an unexpected keyword argument page size google cloud storage client py typeerror testclient test list resource w explicit self def test list resource w explicit self import functools from google api core page iterator import httpiterator project project path path to list resource item to value mock mock spec page token page token max results extra params foo foo page start mock mock spec credentials make credentials client self make one project project credentials credentials connection client base connection make connection iterator client list resource path path item to value item to value page token page token max results max results extra params extra params page start page start tests unit test client py self path path to list resource item to value page token page token max results extra params foo foo page start page size none timeout retry def list resource self path item to value page token none max results none extra params none page start page iterator do nothing page start page size none timeout default timeout retry default retry api request functools partial self connection api request timeout timeout retry retry return page iterator httpiterator client self api request api request path path item to value item to value page token page token max results max results extra params extra params page start page start page size page size e typeerror init got an unexpected keyword argument page size google cloud storage client py typeerror short test summary info failed tests unit test client py testclient test list resource w defaults failed tests unit test client py testclient test list resource w explicit failed passed skipped in nox command py test quiet cov google cloud storage cov google cloud cov tests unit cov append cov config coveragerc cov report cov fail under tests unit failed with exit code nox session unit failed the fix would be to add ranges for google api core and google auth another transitive dependency to setup py and then pin the minimum versions for those ranges in testing constraints txt
| 1
|
11,862
| 14,665,645,043
|
IssuesEvent
|
2020-12-29 14:40:53
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
Mobile app (gateway) > Study List > Modifications to filters
|
Android P1 Process: Enhancement Process: Track 3 iOS
|
Modify filtering options to not include 'upcoming studies' and 'favorites'. Also, remove the bookmarking action on the study list and overview screens.
|
2.0
|
Mobile app (gateway) > Study List > Modifications to filters - Modify filtering options to not include 'upcoming studies' and 'favorites'. Also, remove the bookmarking action on the study list and overview screens.
|
process
|
mobile app gateway study list modifications to filters modify filtering options to not include upcoming studies and favorites also remove the bookmarking action on the study list and overview screens
| 1
|
469
| 2,905,797,977
|
IssuesEvent
|
2015-06-19 04:05:54
|
mitchellh/packer
|
https://api.github.com/repos/mitchellh/packer
|
closed
|
Vagrant-Cloud: Retry box upload on failure
|
bug post-processor/atlas
|
I currently maintain a Vagrant box for 3 providers. I have set up a workflow where a single command builds the machines and pushes them to Vagrant Cloud (Atlas).
Occasionally, one of the boxes fails to upload (for unknown reason) and *it seems* that the whole version is deleted upon this single failure, regardless of whether other boxes are still in transit or not. I then have to create the new version and upload the three boxes manually.
**I propose that instead of failing on "first sight of problem", the upload is retried (perhaps a configurable amount of times, maybe 3 times by default).**
I tracked down the upload step to [this line](https://github.com/mitchellh/packer/blob/master/post-processor/vagrant-cloud/step_upload.go#L23), but my Go skills are basically non-existent to submit a PR. Also, it seems like it would be better to implement the retrial at the `client.Upload` function level as I suspect that function to be shared on more places, thus having direct benefit to other parts of the program.
Thanks for feedback! Let me know if I can provide more details or help anyhow.
|
1.0
|
Vagrant-Cloud: Retry box upload on failure - I currently maintain a Vagrant box for 3 providers. I have set up a workflow where a single command builds the machines and pushes them to Vagrant Cloud (Atlas).
Occasionally, one of the boxes fails to upload (for unknown reason) and *it seems* that the whole version is deleted upon this single failure, regardless of whether other boxes are still in transit or not. I then have to create the new version and upload the three boxes manually.
**I propose that instead of failing on "first sight of problem", the upload is retried (perhaps a configurable amount of times, maybe 3 times by default).**
I tracked down the upload step to [this line](https://github.com/mitchellh/packer/blob/master/post-processor/vagrant-cloud/step_upload.go#L23), but my Go skills are basically non-existent to submit a PR. Also, it seems like it would be better to implement the retrial at the `client.Upload` function level as I suspect that function to be shared on more places, thus having direct benefit to other parts of the program.
Thanks for feedback! Let me know if I can provide more details or help anyhow.
|
process
|
vagrant cloud retry box upload on failure i currently maintain a vagrant box for providers i have set up a workflow where a single command builds the machines and pushes them to vagrant cloud atlas occasionally one of the boxes fails to upload for unknown reason and it seems that the whole version is deleted upon this single failure regardless of whether other boxes are still in transit or not i then have to create the new version and upload the three boxes manually i propose that instead of failing on first sight of problem the upload is retried perhaps a configurable amount of times maybe times by default i tracked down the upload step to but my go skills are basically non existent to submit a pr also it seems like it would be better to implement the retrial at the client upload function level as i suspect that function to be shared on more places thus having direct benefit to other parts of the program thanks for feedback let me know if i can provide more details or help anyhow
| 1
|
363
| 2,797,621,697
|
IssuesEvent
|
2015-05-12 14:40:58
|
DynareTeam/dynare
|
https://api.github.com/repos/DynareTeam/dynare
|
closed
|
Investigate preprocessor crash with external function in model-local variable
|
bug preprocessor
|
See ```use_local_variable_for_beta.mod``` at http://www.dynare.org/phpBB3/viewtopic.php?f=1&t=6772
|
1.0
|
Investigate preprocessor crash with external function in model-local variable - See ```use_local_variable_for_beta.mod``` at http://www.dynare.org/phpBB3/viewtopic.php?f=1&t=6772
|
process
|
investigate preprocessor crash with external function in model local variable see use local variable for beta mod at
| 1
|
563,935
| 16,706,571,443
|
IssuesEvent
|
2021-06-09 10:41:17
|
googleapis/google-api-ruby-client
|
https://api.github.com/repos/googleapis/google-api-ruby-client
|
closed
|
Synthesis failed for gkehub-v1alpha
|
autosynth failure priority: p1 type: bug
|
Hello! Autosynth couldn't regenerate gkehub-v1alpha. :broken_heart:
Please investigate and fix this issue within 5 business days. While it remains broken,
this library cannot be updated with changes to the gkehub-v1alpha API, and the library grows
stale.
See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md
for trouble shooting tips.
Here's the output from running `synth.py`:
```
2021-06-08 03:13:21,720 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/google-api-ruby-client
2021-06-08 03:13:22,610 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore
2021-06-08 03:13:22,612 autosynth [DEBUG] > Running: git config user.name yoshi-automation
2021-06-08 03:13:22,615 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com
2021-06-08 03:13:22,617 autosynth [DEBUG] > Running: git config push.default simple
2021-06-08 03:13:22,620 autosynth [DEBUG] > Running: git branch -f autosynth-gkehub-v1alpha
2021-06-08 03:13:22,623 autosynth [DEBUG] > Running: git checkout autosynth-gkehub-v1alpha
Switched to branch 'autosynth-gkehub-v1alpha'
2021-06-08 03:13:22,837 autosynth [INFO] > Running synthtool
2021-06-08 03:13:22,837 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-gkehub_v1alpha/synth.metadata', 'synth.py', '--']
2021-06-08 03:13:22,837 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/google-api-ruby-client/gkehub/v1alpha/sponge_log.log
2021-06-08 03:13:22,839 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata generated/google-apis-gkehub_v1alpha/synth.metadata synth.py -- gkehub v1alpha
2021-06-08 03:13:23,034 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py.
On branch autosynth-gkehub-v1alpha
nothing to commit, working tree clean
2021-06-08 03:13:23,092 synthtool [DEBUG] > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha
DEBUG:synthtool:Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha
git clean -df
bundle install
Don't run Bundler as root. Bundler can ask for sudo if it is needed, and
installing your bundle as root will break this application for all non-root
users on this machine.
Fetching source index from https://rubygems.org/
Net::HTTPServiceUnavailable:
<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN"
"http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd">
<html>
<head>
<title>503 Service Unavailable</title>
</head>
<body>
<h1>Error 503 Service Unavailable</h1>
<p>Service Unavailable</p>
<h3>Guru Mediation:</h3>
<p>Details: cache-sea4457-SEA 1623147211 1892265244</p>
<hr>
<p>Varnish cache server</p>
</body>
</html>
chown -R 1000:1000 /workspace/generated
2021-06-08 03:13:31,327 synthtool [ERROR] > Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha:
None
ERROR:synthtool:Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha:
None
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1137, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1062, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1404, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 763, in invoke
return __callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py", line 41, in <module>
shell.run(command, hide_output=False)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-w', '/workspace', '-e', 'USER_GROUP=1000:1000', '--entrypoint', 'script/synth.rb', 'gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth', 'gkehub', 'v1alpha']' returned non-zero exit status 1.
2021-06-08 03:13:31,356 autosynth [ERROR] > Synthesis failed
2021-06-08 03:13:31,357 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 293, in _inner_main
).synthesize(synth_log_path / "sponge_log.log")
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-gkehub_v1alpha/synth.metadata', 'synth.py', '--', 'gkehub', 'v1alpha']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/bdd2ce63-997e-4095-81b3-cd54e5fccbe4/targets/github%2Fsynthtool;config=default/tests;query=google-api-ruby-client;failed=false).
|
1.0
|
Synthesis failed for gkehub-v1alpha - Hello! Autosynth couldn't regenerate gkehub-v1alpha. :broken_heart:
Please investigate and fix this issue within 5 business days. While it remains broken,
this library cannot be updated with changes to the gkehub-v1alpha API, and the library grows
stale.
See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md
for trouble shooting tips.
Here's the output from running `synth.py`:
```
2021-06-08 03:13:21,720 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/google-api-ruby-client
2021-06-08 03:13:22,610 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore
2021-06-08 03:13:22,612 autosynth [DEBUG] > Running: git config user.name yoshi-automation
2021-06-08 03:13:22,615 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com
2021-06-08 03:13:22,617 autosynth [DEBUG] > Running: git config push.default simple
2021-06-08 03:13:22,620 autosynth [DEBUG] > Running: git branch -f autosynth-gkehub-v1alpha
2021-06-08 03:13:22,623 autosynth [DEBUG] > Running: git checkout autosynth-gkehub-v1alpha
Switched to branch 'autosynth-gkehub-v1alpha'
2021-06-08 03:13:22,837 autosynth [INFO] > Running synthtool
2021-06-08 03:13:22,837 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-gkehub_v1alpha/synth.metadata', 'synth.py', '--']
2021-06-08 03:13:22,837 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/google-api-ruby-client/gkehub/v1alpha/sponge_log.log
2021-06-08 03:13:22,839 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata generated/google-apis-gkehub_v1alpha/synth.metadata synth.py -- gkehub v1alpha
2021-06-08 03:13:23,034 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py.
On branch autosynth-gkehub-v1alpha
nothing to commit, working tree clean
2021-06-08 03:13:23,092 synthtool [DEBUG] > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha
DEBUG:synthtool:Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha
git clean -df
bundle install
Don't run Bundler as root. Bundler can ask for sudo if it is needed, and
installing your bundle as root will break this application for all non-root
users on this machine.
Fetching source index from https://rubygems.org/
Net::HTTPServiceUnavailable:
<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN"
"http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd">
<html>
<head>
<title>503 Service Unavailable</title>
</head>
<body>
<h1>Error 503 Service Unavailable</h1>
<p>Service Unavailable</p>
<h3>Guru Mediation:</h3>
<p>Details: cache-sea4457-SEA 1623147211 1892265244</p>
<hr>
<p>Varnish cache server</p>
</body>
</html>
chown -R 1000:1000 /workspace/generated
2021-06-08 03:13:31,327 synthtool [ERROR] > Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha:
None
ERROR:synthtool:Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha:
None
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1137, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1062, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1404, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 763, in invoke
return __callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py", line 41, in <module>
shell.run(command, hide_output=False)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-w', '/workspace', '-e', 'USER_GROUP=1000:1000', '--entrypoint', 'script/synth.rb', 'gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth', 'gkehub', 'v1alpha']' returned non-zero exit status 1.
2021-06-08 03:13:31,356 autosynth [ERROR] > Synthesis failed
2021-06-08 03:13:31,357 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 293, in _inner_main
).synthesize(synth_log_path / "sponge_log.log")
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-gkehub_v1alpha/synth.metadata', 'synth.py', '--', 'gkehub', 'v1alpha']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/bdd2ce63-997e-4095-81b3-cd54e5fccbe4/targets/github%2Fsynthtool;config=default/tests;query=google-api-ruby-client;failed=false).
|
non_process
|
synthesis failed for gkehub hello autosynth couldn t regenerate gkehub broken heart please investigate and fix this issue within business days while it remains broken this library cannot be updated with changes to the gkehub api and the library grows stale see for trouble shooting tips here s the output from running synth py autosynth logs will be written to tmpfs src logs google api ruby client autosynth running git config global core excludesfile home kbuilder autosynth gitignore autosynth running git config user name yoshi automation autosynth running git config user email yoshi automation google com autosynth running git config push default simple autosynth running git branch f autosynth gkehub autosynth running git checkout autosynth gkehub switched to branch autosynth gkehub autosynth running synthtool autosynth autosynth log file path tmpfs src logs google api ruby client gkehub sponge log log autosynth running tmpfs src github synthtool env bin m synthtool metadata generated google apis gkehub synth metadata synth py gkehub synthtool executing home kbuilder cache synthtool google api ruby client synth py on branch autosynth gkehub nothing to commit working tree clean synthtool running docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth gkehub debug synthtool running docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth gkehub git clean df bundle install don t run bundler as root bundler can ask for sudo if it is needed and installing your bundle as root will break this application for all non root users on this machine fetching source index from net httpserviceunavailable doctype html public dtd xhtml strict en service unavailable error service unavailable service unavailable guru mediation details cache sea varnish cache server chown r workspace generated synthtool failed executing docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth gkehub none error synthtool failed executing docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth gkehub none traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool google api ruby client synth py line in shell run command hide output false file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize synth log path sponge log log file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
| 0
|
17,145
| 22,692,397,782
|
IssuesEvent
|
2022-07-04 23:03:54
|
sparc4-dev/astropop
|
https://api.github.com/repos/sparc4-dev/astropop
|
closed
|
Imcombine: memory overflow
|
bug image-processing critical
|
When combining a large number of images, `imcombine` is using more memory then expected. This is critical and causing errors.
|
1.0
|
Imcombine: memory overflow - When combining a large number of images, `imcombine` is using more memory then expected. This is critical and causing errors.
|
process
|
imcombine memory overflow when combining a large number of images imcombine is using more memory then expected this is critical and causing errors
| 1
|
5,558
| 2,948,397,394
|
IssuesEvent
|
2015-07-06 02:41:21
|
progrium/dokku
|
https://api.github.com/repos/progrium/dokku
|
opened
|
Document Release Instructions
|
documentation
|
Maintainers other than myself should have the ability to make dokku releases.
- [ ] Find a way to circulate packagecloud auth tokens.
- [ ] Setup a build farm somewhere. DO has been willing to provide an account with credits, maybe something I can look into?
- [ ] Add documentation to docs folder.
|
1.0
|
Document Release Instructions - Maintainers other than myself should have the ability to make dokku releases.
- [ ] Find a way to circulate packagecloud auth tokens.
- [ ] Setup a build farm somewhere. DO has been willing to provide an account with credits, maybe something I can look into?
- [ ] Add documentation to docs folder.
|
non_process
|
document release instructions maintainers other than myself should have the ability to make dokku releases find a way to circulate packagecloud auth tokens setup a build farm somewhere do has been willing to provide an account with credits maybe something i can look into add documentation to docs folder
| 0
|
275
| 2,704,399,535
|
IssuesEvent
|
2015-04-07 00:06:36
|
MultiMC/MultiMC5
|
https://api.github.com/repos/MultiMC/MultiMC5
|
closed
|
Silent Failure with malformed qmod
|
bug confirmed needs processing quickmods
|
A malformed qmod causes an error in the log, but no message is shown, so there is no indication of failure
|
1.0
|
Silent Failure with malformed qmod - A malformed qmod causes an error in the log, but no message is shown, so there is no indication of failure
|
process
|
silent failure with malformed qmod a malformed qmod causes an error in the log but no message is shown so there is no indication of failure
| 1
|
5,471
| 8,337,912,571
|
IssuesEvent
|
2018-09-28 12:48:47
|
sysown/proxysql
|
https://api.github.com/repos/sysown/proxysql
|
closed
|
Do not cache empty resultset - per rule setting
|
CACHE QUERY PROCESSOR
|
This is a follow up from #1693.
Instead of just having a global variable, we can also have a per-rule setting.
|
1.0
|
Do not cache empty resultset - per rule setting - This is a follow up from #1693.
Instead of just having a global variable, we can also have a per-rule setting.
|
process
|
do not cache empty resultset per rule setting this is a follow up from instead of just having a global variable we can also have a per rule setting
| 1
|
20,160
| 26,713,775,073
|
IssuesEvent
|
2023-01-28 08:00:05
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
DISABLED test_fs_sharing (__main__.TestMultiprocessing)
|
module: windows module: multiprocessing triaged skipped
|
Flaky failures in the last week: https://fburl.com/scuba/opensource_ci_jobs/inmj698k. They only appear to be on windows
Platforms: windows, mac, linux, dynamo, rocm
cc @peterjc123 @mszhanyi @skyline75489 @nbcsm @VitalyFedyunin
|
1.0
|
DISABLED test_fs_sharing (__main__.TestMultiprocessing) - Flaky failures in the last week: https://fburl.com/scuba/opensource_ci_jobs/inmj698k. They only appear to be on windows
Platforms: windows, mac, linux, dynamo, rocm
cc @peterjc123 @mszhanyi @skyline75489 @nbcsm @VitalyFedyunin
|
process
|
disabled test fs sharing main testmultiprocessing flaky failures in the last week they only appear to be on windows platforms windows mac linux dynamo rocm cc mszhanyi nbcsm vitalyfedyunin
| 1
|
26,129
| 6,755,375,993
|
IssuesEvent
|
2017-10-24 00:06:40
|
jascam/CodePlexFoo
|
https://api.github.com/repos/jascam/CodePlexFoo
|
closed
|
Create Example: CSWPFAnimation
|
bug CodePlexMigrationInitiated impact: Low
|
Demonstrate animation in WPF
#### Migrated CodePlex Work Item Details
CodePlex Work Item ID: '2251'
Vote count: '2'
|
1.0
|
Create Example: CSWPFAnimation - Demonstrate animation in WPF
#### Migrated CodePlex Work Item Details
CodePlex Work Item ID: '2251'
Vote count: '2'
|
non_process
|
create example cswpfanimation demonstrate animation in wpf migrated codeplex work item details codeplex work item id vote count
| 0
|
10,815
| 13,609,290,695
|
IssuesEvent
|
2020-09-23 04:50:37
|
googleapis/java-billing
|
https://api.github.com/repos/googleapis/java-billing
|
closed
|
Dependency Dashboard
|
api: cloudbilling type: process
|
This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any chore deps update dependency com google cloud libraries bom to check this box to trigger a request for renovate to run again on this repository
| 1
|
15,162
| 18,912,589,445
|
IssuesEvent
|
2021-11-16 15:27:37
|
prisma/language-tools
|
https://api.github.com/repos/prisma/language-tools
|
opened
|
Add configuration option to turn of file watcher
|
kind/feature process/candidate
|
File watcher is causing problems we can not figure out, so let's build in a switch to turn it off and see if this fixes problems for users.
|
1.0
|
Add configuration option to turn of file watcher - File watcher is causing problems we can not figure out, so let's build in a switch to turn it off and see if this fixes problems for users.
|
process
|
add configuration option to turn of file watcher file watcher is causing problems we can not figure out so let s build in a switch to turn it off and see if this fixes problems for users
| 1
|
8,479
| 11,643,460,862
|
IssuesEvent
|
2020-02-29 13:41:53
|
SE-Garden/tms-webserver
|
https://api.github.com/repos/SE-Garden/tms-webserver
|
closed
|
DockerイメージをDockerHubへPushできるようにする
|
kind:アーキ process:検討
|
## 概要
DockerイメージをDockerHubへPushできるようにする 。
対象のリポジトリは、`SE-Gardenn` 専用のOrganizationとすること。
## ゴール
- DockerイメージをDockerHubにPushして利用できるようにする。
- ローカル環境で動かす際も自分がビルドしたDockerイメージでDockerComposeを利用して立ち上げられるようにする
## 成果物
- docker-compose.yaml
- gradle.yaml
## 関連Issue
- none
|
1.0
|
DockerイメージをDockerHubへPushできるようにする - ## 概要
DockerイメージをDockerHubへPushできるようにする 。
対象のリポジトリは、`SE-Gardenn` 専用のOrganizationとすること。
## ゴール
- DockerイメージをDockerHubにPushして利用できるようにする。
- ローカル環境で動かす際も自分がビルドしたDockerイメージでDockerComposeを利用して立ち上げられるようにする
## 成果物
- docker-compose.yaml
- gradle.yaml
## 関連Issue
- none
|
process
|
dockerイメージをdockerhubへpushできるようにする 概要 dockerイメージをdockerhubへpushできるようにする 。 対象のリポジトリは、 se gardenn 専用のorganizationとすること。 ゴール dockerイメージをdockerhubにpushして利用できるようにする。 ローカル環境で動かす際も自分がビルドしたdockerイメージでdockercomposeを利用して立ち上げられるようにする 成果物 docker compose yaml gradle yaml 関連issue none
| 1
|
4,575
| 7,397,974,802
|
IssuesEvent
|
2018-03-19 03:01:11
|
nokia-wroclaw/innovativeprojects-healthiness-of-data
|
https://api.github.com/repos/nokia-wroclaw/innovativeprojects-healthiness-of-data
|
closed
|
Create a function for unknown kpi names.
|
backend data processing enhancement help wanted
|
There are cases when the kpi name extracted from database doesn't match any key in our csv file. We need to create a function that handles these cases:
1. The kpi name is nowhere to be found in the file. No similiar name. These entries will be inserted into a separate database.
2. The kpi name has a similiar name in csv file. Find the newest and best matching one. Example:
Extracted from database: sgsn_2012. In our file we can find sgsn_2012a, sgsn_2012b etc.
|
1.0
|
Create a function for unknown kpi names. - There are cases when the kpi name extracted from database doesn't match any key in our csv file. We need to create a function that handles these cases:
1. The kpi name is nowhere to be found in the file. No similiar name. These entries will be inserted into a separate database.
2. The kpi name has a similiar name in csv file. Find the newest and best matching one. Example:
Extracted from database: sgsn_2012. In our file we can find sgsn_2012a, sgsn_2012b etc.
|
process
|
create a function for unknown kpi names there are cases when the kpi name extracted from database doesn t match any key in our csv file we need to create a function that handles these cases the kpi name is nowhere to be found in the file no similiar name these entries will be inserted into a separate database the kpi name has a similiar name in csv file find the newest and best matching one example extracted from database sgsn in our file we can find sgsn sgsn etc
| 1
|
559,334
| 16,556,161,498
|
IssuesEvent
|
2021-05-28 14:10:20
|
sopra-fs21-group-15/server
|
https://api.github.com/repos/sopra-fs21-group-15/server
|
closed
|
OPTIONAL: Make friend-relations symmetric
|
low priority task
|
Part of #16
Includes friends-request system
---
Time estimate: 6h
|
1.0
|
OPTIONAL: Make friend-relations symmetric - Part of #16
Includes friends-request system
---
Time estimate: 6h
|
non_process
|
optional make friend relations symmetric part of includes friends request system time estimate
| 0
|
310,347
| 9,488,923,171
|
IssuesEvent
|
2019-04-22 20:54:50
|
pulumi/pulumi
|
https://api.github.com/repos/pulumi/pulumi
|
closed
|
Allow changes to property values to be ignored
|
area/core customer/feedback priority/P1
|
Currently there is no way to mark that we don't care about changed values of a particular property - similar to Terraform's `ignore_changes` block.
Although not required often, there are some use cases which would benefit. For example, enabling DynamoDB AutoScaling alters the read and write capacities at the provider end. Since these are required, and are set to their initial values in the Pulumi program, a refresh will give a diff to reset the values to their initial state.
|
1.0
|
Allow changes to property values to be ignored - Currently there is no way to mark that we don't care about changed values of a particular property - similar to Terraform's `ignore_changes` block.
Although not required often, there are some use cases which would benefit. For example, enabling DynamoDB AutoScaling alters the read and write capacities at the provider end. Since these are required, and are set to their initial values in the Pulumi program, a refresh will give a diff to reset the values to their initial state.
|
non_process
|
allow changes to property values to be ignored currently there is no way to mark that we don t care about changed values of a particular property similar to terraform s ignore changes block although not required often there are some use cases which would benefit for example enabling dynamodb autoscaling alters the read and write capacities at the provider end since these are required and are set to their initial values in the pulumi program a refresh will give a diff to reset the values to their initial state
| 0
|
89,386
| 3,793,372,258
|
IssuesEvent
|
2016-03-22 13:44:54
|
DynamoRIO/dynamorio
|
https://api.github.com/repos/DynamoRIO/dynamorio
|
closed
|
CRASH caused by generated LDR PC from unaligned address
|
Bug-DRCrash Hotlist-Release OpSys-ARM Priority-High
|
The ARM architectural manual specifies that the address used by an LDR instruction into PC has to be word aligned:
```
[...]
if t == 15 then
if address<1:0> == '00' then LoadWritePC(data); else UNPREDICTABLE;
[...]
```
However, DynamoRIO seems to sometimes generate loads from unaligned addresses into PC. For example:
```
0x47e6c318: ldr.w r0, [r10]
0x47e6c31c: ldr r3, [r7, #48] ; 0x30
0x47e6c31e: ldr r3, [r3, #12]
0x47e6c320: cmp r3, #0
0x47e6c322: beq.w 0x47e6c32a
0x47e6c326: b.w 0x47e6c35c
=> 0x47e6c32a: ldr.w pc, [pc, #14] ; 0x47e6c33a
0x47e6c32e: movw r0, #51396 ; 0xc8c4
0x47e6c332: movt r0, #18411 ; 0x47eb
0x47e6c336: ldr.w pc, [r10, #24]
0x47e6c33a: pop {r0, r3, r5, r7}
```
This is causing a bus error on certain cores, for example APM X-Gene.
How to reproduce the crash: run a few Thumb applications via DR on a system using X-Gene cores. Many will crash immediately.
|
1.0
|
CRASH caused by generated LDR PC from unaligned address - The ARM architectural manual specifies that the address used by an LDR instruction into PC has to be word aligned:
```
[...]
if t == 15 then
if address<1:0> == '00' then LoadWritePC(data); else UNPREDICTABLE;
[...]
```
However, DynamoRIO seems to sometimes generate loads from unaligned addresses into PC. For example:
```
0x47e6c318: ldr.w r0, [r10]
0x47e6c31c: ldr r3, [r7, #48] ; 0x30
0x47e6c31e: ldr r3, [r3, #12]
0x47e6c320: cmp r3, #0
0x47e6c322: beq.w 0x47e6c32a
0x47e6c326: b.w 0x47e6c35c
=> 0x47e6c32a: ldr.w pc, [pc, #14] ; 0x47e6c33a
0x47e6c32e: movw r0, #51396 ; 0xc8c4
0x47e6c332: movt r0, #18411 ; 0x47eb
0x47e6c336: ldr.w pc, [r10, #24]
0x47e6c33a: pop {r0, r3, r5, r7}
```
This is causing a bus error on certain cores, for example APM X-Gene.
How to reproduce the crash: run a few Thumb applications via DR on a system using X-Gene cores. Many will crash immediately.
|
non_process
|
crash caused by generated ldr pc from unaligned address the arm architectural manual specifies that the address used by an ldr instruction into pc has to be word aligned if t then if address then loadwritepc data else unpredictable however dynamorio seems to sometimes generate loads from unaligned addresses into pc for example ldr w ldr ldr cmp beq w b w ldr w pc movw movt ldr w pc pop this is causing a bus error on certain cores for example apm x gene how to reproduce the crash run a few thumb applications via dr on a system using x gene cores many will crash immediately
| 0
|
219,837
| 16,851,876,054
|
IssuesEvent
|
2021-06-20 17:23:42
|
CFEL-CMI/richmol
|
https://api.github.com/repos/CFEL-CMI/richmol
|
closed
|
Provide user and developer documentation on readthedocs
|
documentation
|
Create automatically updated RTD docs for both the develop/ branch as well as the latest release.
|
1.0
|
Provide user and developer documentation on readthedocs - Create automatically updated RTD docs for both the develop/ branch as well as the latest release.
|
non_process
|
provide user and developer documentation on readthedocs create automatically updated rtd docs for both the develop branch as well as the latest release
| 0
|
14,004
| 16,812,492,838
|
IssuesEvent
|
2021-06-17 00:52:42
|
w3c/webauthn
|
https://api.github.com/repos/w3c/webauthn
|
closed
|
Update README with new Bikeshed build instructions
|
type:process
|
Follow-up to #1625. Build instructions in README need to be updated to match.
|
1.0
|
Update README with new Bikeshed build instructions - Follow-up to #1625. Build instructions in README need to be updated to match.
|
process
|
update readme with new bikeshed build instructions follow up to build instructions in readme need to be updated to match
| 1
|
737,734
| 25,529,307,893
|
IssuesEvent
|
2022-11-29 06:50:06
|
rstudio/gt
|
https://api.github.com/repos/rstudio/gt
|
closed
|
Audit all internal and unsafe uses of `glue::glue()`
|
Difficulty: [2] Intermediate Effort: [2] Medium Priority: [3] High Type: ٩(͡๏̯͡๏)۶ Refactor
|
There are several instances where `glue::glue()` is used to generate strings with local variables. The problem is that, by default, the search path for `glue()` includes all environments. There are some steps to make the (convenient) use of `glue()` safer:
(1) Use the `base::get()` function as the transformer function (supplying it to the `.transformer` arg).
(2) Specify an environment for the `.envir` arg; `new.env()` is a pretty good choice, `emptyenv()` is even better if that's possible.
(3) Use named arguments in the `glue()` call for replacements, if possible; these will take precedence over vars in the search path.
The dev version of **glue** has the `glue_safe()` function. Once that's available in a CRAN release of **glue**, it should be used.
|
1.0
|
Audit all internal and unsafe uses of `glue::glue()` - There are several instances where `glue::glue()` is used to generate strings with local variables. The problem is that, by default, the search path for `glue()` includes all environments. There are some steps to make the (convenient) use of `glue()` safer:
(1) Use the `base::get()` function as the transformer function (supplying it to the `.transformer` arg).
(2) Specify an environment for the `.envir` arg; `new.env()` is a pretty good choice, `emptyenv()` is even better if that's possible.
(3) Use named arguments in the `glue()` call for replacements, if possible; these will take precedence over vars in the search path.
The dev version of **glue** has the `glue_safe()` function. Once that's available in a CRAN release of **glue**, it should be used.
|
non_process
|
audit all internal and unsafe uses of glue glue there are several instances where glue glue is used to generate strings with local variables the problem is that by default the search path for glue includes all environments there are some steps to make the convenient use of glue safer use the base get function as the transformer function supplying it to the transformer arg specify an environment for the envir arg new env is a pretty good choice emptyenv is even better if that s possible use named arguments in the glue call for replacements if possible these will take precedence over vars in the search path the dev version of glue has the glue safe function once that s available in a cran release of glue it should be used
| 0
|
9,465
| 12,450,510,476
|
IssuesEvent
|
2020-05-27 08:55:37
|
ESMValGroup/ESMValCore
|
https://api.github.com/repos/ESMValGroup/ESMValCore
|
closed
|
Empty cubes after volume_statistics for certain datasets
|
bug preprocessor
|
**Describe the bug**
CMIP6 EC-Earth ocean 3D data has the whole bottom depth layer set to masked values. Both the variables and the volcello. The volume_statistics preproc computes the weighted mean per layer just fine, and for the bottom layer the results are masked as well. But when it comes to computing the weighted mean over the water volume, the presence of masked values sets the results to NaNs and the data comes out of the preprocessing empty:
`result.append(np.average(column, weights=depth_volume))`
It looks like the function `np.average` is not handling well the presence of layers with only masked values. Would it be correct check and remove this sort of layers before calling `np.average` ?
**Please attach**
I'm working with EC-Earth local data and can't seem to find Ecearth data with volcello on /badc, sorry.
|
1.0
|
Empty cubes after volume_statistics for certain datasets - **Describe the bug**
CMIP6 EC-Earth ocean 3D data has the whole bottom depth layer set to masked values. Both the variables and the volcello. The volume_statistics preproc computes the weighted mean per layer just fine, and for the bottom layer the results are masked as well. But when it comes to computing the weighted mean over the water volume, the presence of masked values sets the results to NaNs and the data comes out of the preprocessing empty:
`result.append(np.average(column, weights=depth_volume))`
It looks like the function `np.average` is not handling well the presence of layers with only masked values. Would it be correct check and remove this sort of layers before calling `np.average` ?
**Please attach**
I'm working with EC-Earth local data and can't seem to find Ecearth data with volcello on /badc, sorry.
|
process
|
empty cubes after volume statistics for certain datasets describe the bug ec earth ocean data has the whole bottom depth layer set to masked values both the variables and the volcello the volume statistics preproc computes the weighted mean per layer just fine and for the bottom layer the results are masked as well but when it comes to computing the weighted mean over the water volume the presence of masked values sets the results to nans and the data comes out of the preprocessing empty result append np average column weights depth volume it looks like the function np average is not handling well the presence of layers with only masked values would it be correct check and remove this sort of layers before calling np average please attach i m working with ec earth local data and can t seem to find ecearth data with volcello on badc sorry
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.