Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
9,536
12,504,875,738
IssuesEvent
2020-06-02 09:44:20
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Add a test case for project paths with spaces
process/candidate
Some user's project paths contain spaces. This broke client generation various times already, e.g. https://github.com/prisma/prisma/issues/1973 and https://github.com/prisma/prisma/issues/2612. We should add at least one dedicated test for this so this won't break again for our users.
1.0
Add a test case for project paths with spaces - Some user's project paths contain spaces. This broke client generation various times already, e.g. https://github.com/prisma/prisma/issues/1973 and https://github.com/prisma/prisma/issues/2612. We should add at least one dedicated test for this so this won't break again for our users.
process
add a test case for project paths with spaces some user s project paths contain spaces this broke client generation various times already e g and we should add at least one dedicated test for this so this won t break again for our users
1
21,982
30,474,010,994
IssuesEvent
2023-07-17 15:17:10
The-Data-Alchemists-Manipal/MindWave
https://api.github.com/repos/The-Data-Alchemists-Manipal/MindWave
closed
Thumbnailator
gssoc23 level2 image-processing
### Is your feature request related to a problem? Please describe. Thumbnailator's fluent interface can be used to perform fairly complicated thumbnail processing task in one simple step. ### Describe the solution you'd like creating JPEG thumbnails of image files in a directory, all resized to a maximum dimension of 640 pixels by 480 pixels while preserving the aspect ratio of the original image can be performed ### Describe alternatives you've considered _No response_ ### Additional context Assign this issue to me under GSSOC'23 ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
Thumbnailator - ### Is your feature request related to a problem? Please describe. Thumbnailator's fluent interface can be used to perform fairly complicated thumbnail processing task in one simple step. ### Describe the solution you'd like creating JPEG thumbnails of image files in a directory, all resized to a maximum dimension of 640 pixels by 480 pixels while preserving the aspect ratio of the original image can be performed ### Describe alternatives you've considered _No response_ ### Additional context Assign this issue to me under GSSOC'23 ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
process
thumbnailator is your feature request related to a problem please describe thumbnailator s fluent interface can be used to perform fairly complicated thumbnail processing task in one simple step describe the solution you d like creating jpeg thumbnails of image files in a directory all resized to a maximum dimension of pixels by pixels while preserving the aspect ratio of the original image can be performed describe alternatives you ve considered no response additional context assign this issue to me under gssoc code of conduct i agree to follow this project s code of conduct
1
124,378
17,772,541,543
IssuesEvent
2021-08-30 15:10:40
kapseliboi/evergreen
https://api.github.com/repos/kapseliboi/evergreen
opened
CVE-2019-8331 (Medium) detected in bootstrap-3.2.0.min.js
security vulnerability
## CVE-2019-8331 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.2.0.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.2.0/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.2.0/js/bootstrap.min.js</a></p> <p>Path to dependency file: evergreen/services/node_modules/remarkable/demo/index.html</p> <p>Path to vulnerable library: /services/node_modules/remarkable/demo/index.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.2.0.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/evergreen/commit/13675096220f0e986aa94cafc5f57de6b38e38cd">13675096220f0e986aa94cafc5f57de6b38e38cd</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/pull/28236">https://github.com/twbs/bootstrap/pull/28236</a></p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-8331 (Medium) detected in bootstrap-3.2.0.min.js - ## CVE-2019-8331 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.2.0.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.2.0/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.2.0/js/bootstrap.min.js</a></p> <p>Path to dependency file: evergreen/services/node_modules/remarkable/demo/index.html</p> <p>Path to vulnerable library: /services/node_modules/remarkable/demo/index.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.2.0.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/evergreen/commit/13675096220f0e986aa94cafc5f57de6b38e38cd">13675096220f0e986aa94cafc5f57de6b38e38cd</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/pull/28236">https://github.com/twbs/bootstrap/pull/28236</a></p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file evergreen services node modules remarkable demo index html path to vulnerable library services node modules remarkable demo index html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap bootstrap sass step up your open source security game with whitesource
0
281,037
8,689,924,593
IssuesEvent
2018-12-03 20:03:37
SpaceNetChallenge/utilities
https://api.github.com/repos/SpaceNetChallenge/utilities
closed
You might want a .gitignore file in this repo
Difficulty: Easy Priority: High Status: Unassigned Type: Maintenance
`.pyc` files don't need to be committed; they can be filtered out with .gitignore.
1.0
You might want a .gitignore file in this repo - `.pyc` files don't need to be committed; they can be filtered out with .gitignore.
non_process
you might want a gitignore file in this repo pyc files don t need to be committed they can be filtered out with gitignore
0
417,964
28,112,517,879
IssuesEvent
2023-03-31 08:19:31
venuslimm/ped
https://api.github.com/repos/venuslimm/ped
opened
Screenshot of find command on user guide does not match with the current UI
severity.VeryLow type.DocumentationBug
Screenshot on the user guide: ![image.png](https://raw.githubusercontent.com/venuslimm/ped/main/files/07254cf5-81a8-4eb0-a9b1-ad11a789cf6c.png) Actual UI: ![image.png](https://raw.githubusercontent.com/venuslimm/ped/main/files/8970156e-6b3a-4a79-a931-2e2859a195c6.png) <!--session: 1680242405607-d67b960d-7651-4494-be68-de1656255e7d--> <!--Version: Web v3.4.7-->
1.0
Screenshot of find command on user guide does not match with the current UI - Screenshot on the user guide: ![image.png](https://raw.githubusercontent.com/venuslimm/ped/main/files/07254cf5-81a8-4eb0-a9b1-ad11a789cf6c.png) Actual UI: ![image.png](https://raw.githubusercontent.com/venuslimm/ped/main/files/8970156e-6b3a-4a79-a931-2e2859a195c6.png) <!--session: 1680242405607-d67b960d-7651-4494-be68-de1656255e7d--> <!--Version: Web v3.4.7-->
non_process
screenshot of find command on user guide does not match with the current ui screenshot on the user guide actual ui
0
301,569
9,221,755,877
IssuesEvent
2019-03-11 20:48:46
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
opened
ci: Turn off `xenial-valgrind-memcheck-weekly`
priority: medium team: kitware
See Slack convo: https://drakedevelopers.slack.com/archives/C270MN28G/p1552336898014300?thread_ts=1552312717.012700&cid=C270MN28G Summary: It's difficult to reproduce in Bionic, and not providing too much value as the Enabling team has moved primary development / usage to Bionic. \cc @jwnimmer-tri
1.0
ci: Turn off `xenial-valgrind-memcheck-weekly` - See Slack convo: https://drakedevelopers.slack.com/archives/C270MN28G/p1552336898014300?thread_ts=1552312717.012700&cid=C270MN28G Summary: It's difficult to reproduce in Bionic, and not providing too much value as the Enabling team has moved primary development / usage to Bionic. \cc @jwnimmer-tri
non_process
ci turn off xenial valgrind memcheck weekly see slack convo summary it s difficult to reproduce in bionic and not providing too much value as the enabling team has moved primary development usage to bionic cc jwnimmer tri
0
562
3,023,861,228
IssuesEvent
2015-08-01 23:56:31
HazyResearch/dd-genomics
https://api.github.com/repos/HazyResearch/dd-genomics
opened
Set up and document AWS pipeline for pre-processing
Preprocessing PRIORITY
This is / should be already documented in `HazyResearch/bazaar` but make sure we have some notes here / run this for everything today...
1.0
Set up and document AWS pipeline for pre-processing - This is / should be already documented in `HazyResearch/bazaar` but make sure we have some notes here / run this for everything today...
process
set up and document aws pipeline for pre processing this is should be already documented in hazyresearch bazaar but make sure we have some notes here run this for everything today
1
29,811
13,173,137,014
IssuesEvent
2020-08-11 19:44:13
thkl/hap-homematic
https://api.github.com/repos/thkl/hap-homematic
closed
Variablen mit mehr Services
DeviceService enhancement
Ist es Möglich im Bereich Variablen mehr Services unter zubringen. z.B. Luftfeuchtigkeit, Bewegungsmelder, Belegtmelder, Feuchtigkeitssensor etc.
1.0
Variablen mit mehr Services - Ist es Möglich im Bereich Variablen mehr Services unter zubringen. z.B. Luftfeuchtigkeit, Bewegungsmelder, Belegtmelder, Feuchtigkeitssensor etc.
non_process
variablen mit mehr services ist es möglich im bereich variablen mehr services unter zubringen z b luftfeuchtigkeit bewegungsmelder belegtmelder feuchtigkeitssensor etc
0
11,488
5,011,878,407
IssuesEvent
2016-12-13 09:32:43
LLNL/spack
https://api.github.com/repos/LLNL/spack
closed
SLEPc fails to configure with Spack's python
bug build-error package python
I just wiped my installation of Spack to re-install and check things and got the error `Symbol not found: __PyCodecInfo_GetIncrementalDecoder`: ``` ==> './configure' '--prefix=/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/slepc-3.7.3-rhcxmg2ntqe3v6epgljeseffnpa4gla2' '--with-arpack-dir=/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/arpack-ng-3.4.0-g76ncwdpqcyx5lm5e65ydwaetbx5sulo/lib' '--with-arpack-flags=-lparpack,-larpack' Traceback (most recent call last): File "./configure", line 10, in <module> execfile(os.path.join(os.path.dirname(__file__), 'config', 'configure.py')) File "./config/configure.py", line 140, in <module> import slepc, petsc, arpack, blzpack, trlan, feast, primme, blopex, sowing, lapack File "/private/var/folders/5k/sqpp24tx3ylds4fgm13pfht00000gn/T/davydden/spack-stage/spack-stage-ZOF1pH/slepc-3.7.3/config/packages/petsc.py", line 22, in <module> import package, os, sys, commands File "/private/var/folders/5k/sqpp24tx3ylds4fgm13pfht00000gn/T/davydden/spack-stage/spack-stage-ZOF1pH/slepc-3.7.3/config/package.py", line 22, in <module> import os, sys, commands, tempfile, shutil, urllib, urlparse, tarfile File "/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/tempfile.py", line 32, in <module> import io as _io File "/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/io.py", line 51, in <module> import _io ImportError: dlopen(/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/lib-dynload/_io.so, 2): Symbol not found: __PyCodecInfo_GetIncrementalDecoder Referenced from: /Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/lib-dynload/_io.so Expected in: flat namespace in /Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/lib-dynload/_io.so ``` Looking at the history of `python` package, i don't see what could have led to this. For now will be using ``` python: version: [2.7.10] paths: python@2.7.10: /usr buildable: False ```
1.0
SLEPc fails to configure with Spack's python - I just wiped my installation of Spack to re-install and check things and got the error `Symbol not found: __PyCodecInfo_GetIncrementalDecoder`: ``` ==> './configure' '--prefix=/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/slepc-3.7.3-rhcxmg2ntqe3v6epgljeseffnpa4gla2' '--with-arpack-dir=/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/arpack-ng-3.4.0-g76ncwdpqcyx5lm5e65ydwaetbx5sulo/lib' '--with-arpack-flags=-lparpack,-larpack' Traceback (most recent call last): File "./configure", line 10, in <module> execfile(os.path.join(os.path.dirname(__file__), 'config', 'configure.py')) File "./config/configure.py", line 140, in <module> import slepc, petsc, arpack, blzpack, trlan, feast, primme, blopex, sowing, lapack File "/private/var/folders/5k/sqpp24tx3ylds4fgm13pfht00000gn/T/davydden/spack-stage/spack-stage-ZOF1pH/slepc-3.7.3/config/packages/petsc.py", line 22, in <module> import package, os, sys, commands File "/private/var/folders/5k/sqpp24tx3ylds4fgm13pfht00000gn/T/davydden/spack-stage/spack-stage-ZOF1pH/slepc-3.7.3/config/package.py", line 22, in <module> import os, sys, commands, tempfile, shutil, urllib, urlparse, tarfile File "/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/tempfile.py", line 32, in <module> import io as _io File "/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/io.py", line 51, in <module> import _io ImportError: dlopen(/Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/lib-dynload/_io.so, 2): Symbol not found: __PyCodecInfo_GetIncrementalDecoder Referenced from: /Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/lib-dynload/_io.so Expected in: flat namespace in /Users/davydden/spack/opt/spack/darwin-sierra-x86_64/clang-8.0.0-apple/python-2.7.12-6dtr7kw2sj5zu7z7v3ox3agrmpw5cndt/lib/python2.7/lib-dynload/_io.so ``` Looking at the history of `python` package, i don't see what could have led to this. For now will be using ``` python: version: [2.7.10] paths: python@2.7.10: /usr buildable: False ```
non_process
slepc fails to configure with spack s python i just wiped my installation of spack to re install and check things and got the error symbol not found pycodecinfo getincrementaldecoder configure prefix users davydden spack opt spack darwin sierra clang apple slepc with arpack dir users davydden spack opt spack darwin sierra clang apple arpack ng lib with arpack flags lparpack larpack traceback most recent call last file configure line in execfile os path join os path dirname file config configure py file config configure py line in import slepc petsc arpack blzpack trlan feast primme blopex sowing lapack file private var folders t davydden spack stage spack stage slepc config packages petsc py line in import package os sys commands file private var folders t davydden spack stage spack stage slepc config package py line in import os sys commands tempfile shutil urllib urlparse tarfile file users davydden spack opt spack darwin sierra clang apple python lib tempfile py line in import io as io file users davydden spack opt spack darwin sierra clang apple python lib io py line in import io importerror dlopen users davydden spack opt spack darwin sierra clang apple python lib lib dynload io so symbol not found pycodecinfo getincrementaldecoder referenced from users davydden spack opt spack darwin sierra clang apple python lib lib dynload io so expected in flat namespace in users davydden spack opt spack darwin sierra clang apple python lib lib dynload io so looking at the history of python package i don t see what could have led to this for now will be using python version paths python usr buildable false
0
328,938
10,007,238,770
IssuesEvent
2019-07-14 08:56:47
answeropedia/answeropedia.org
https://api.github.com/repos/answeropedia/answeropedia.org
closed
Abolish short answer (extraction from common answer)
priority-critical
@gomzyakov >Короткий ответ будем хранить в отдельном поле сущности, не пытаясь выцарапать из общего ответа (т.к. это иногд просто бессмыслено)
1.0
Abolish short answer (extraction from common answer) - @gomzyakov >Короткий ответ будем хранить в отдельном поле сущности, не пытаясь выцарапать из общего ответа (т.к. это иногд просто бессмыслено)
non_process
abolish short answer extraction from common answer gomzyakov короткий ответ будем хранить в отдельном поле сущности не пытаясь выцарапать из общего ответа т к это иногд просто бессмыслено
0
13,297
22,574,830,062
IssuesEvent
2022-06-28 06:11:47
FederatedAI/KubeFATE
https://api.github.com/repos/FederatedAI/KubeFATE
closed
希望将fate-Serving的Ingress的api由networking.k8s.io/v1beta1升级到networking.k8s.io/v1,以适配kubernates-1.22以上版本
kind/requirement
**Is your feature request related to a problem? Please describe.** 最新版本的k8s安装cluster-serving时报错:https://github.com/FederatedAI/KubeFATE/issues/618 **Describe the solution you'd like** 支持最新的k8s,由networking.k8s.io/v1beta1升级到networking.k8s.io/v1 **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
1.0
希望将fate-Serving的Ingress的api由networking.k8s.io/v1beta1升级到networking.k8s.io/v1,以适配kubernates-1.22以上版本 - **Is your feature request related to a problem? Please describe.** 最新版本的k8s安装cluster-serving时报错:https://github.com/FederatedAI/KubeFATE/issues/618 **Describe the solution you'd like** 支持最新的k8s,由networking.k8s.io/v1beta1升级到networking.k8s.io/v1 **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
non_process
希望将fate serving的ingress的api由networking io io ,以适配kubernates is your feature request related to a problem please describe serving时报错: describe the solution you d like ,由networking io io describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
0
2,358
5,165,992,743
IssuesEvent
2017-01-17 15:11:17
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
Clarity of `filterFormatted` in Documentation
inprocess
## Issue The `filterFormatted` prop has an unclear description in the documentation. I spent hours trying to figure out why my select filter wouldn't work. It turns out I had included `filterFormatted` in the column (because of the [examples for the select filter](http://allenfang.github.io/react-bootstrap-table/example.html#column-filter)) without fully understanding what it did. Looking at the documentation it seemed like I just needed to include it in the column to enable filtering. ## Proposed fix Make the documentation for `filterFormatted` clearer. To get started in the right direction, I'd suggest something as follows. > When true, the column will filter using the value returned by the column's formatter. When false (default), the column will filter using the pre-formatted value. Also consider changing the [filter examples](http://allenfang.github.io/react-bootstrap-table/example.html#column-filter) to not include `filterFormatted` and possibly include one extra example to showcase the use of `filterFormatted` explicitly. Anyways, thanks to everyone who's worked on this component. It's extremely powerful and versatile.
1.0
Clarity of `filterFormatted` in Documentation - ## Issue The `filterFormatted` prop has an unclear description in the documentation. I spent hours trying to figure out why my select filter wouldn't work. It turns out I had included `filterFormatted` in the column (because of the [examples for the select filter](http://allenfang.github.io/react-bootstrap-table/example.html#column-filter)) without fully understanding what it did. Looking at the documentation it seemed like I just needed to include it in the column to enable filtering. ## Proposed fix Make the documentation for `filterFormatted` clearer. To get started in the right direction, I'd suggest something as follows. > When true, the column will filter using the value returned by the column's formatter. When false (default), the column will filter using the pre-formatted value. Also consider changing the [filter examples](http://allenfang.github.io/react-bootstrap-table/example.html#column-filter) to not include `filterFormatted` and possibly include one extra example to showcase the use of `filterFormatted` explicitly. Anyways, thanks to everyone who's worked on this component. It's extremely powerful and versatile.
process
clarity of filterformatted in documentation issue the filterformatted prop has an unclear description in the documentation i spent hours trying to figure out why my select filter wouldn t work it turns out i had included filterformatted in the column because of the without fully understanding what it did looking at the documentation it seemed like i just needed to include it in the column to enable filtering proposed fix make the documentation for filterformatted clearer to get started in the right direction i d suggest something as follows when true the column will filter using the value returned by the column s formatter when false default the column will filter using the pre formatted value also consider changing the to not include filterformatted and possibly include one extra example to showcase the use of filterformatted explicitly anyways thanks to everyone who s worked on this component it s extremely powerful and versatile
1
334,497
10,142,069,030
IssuesEvent
2019-08-03 20:07:38
tensorwerk/hangar-py
https://api.github.com/repos/tensorwerk/hangar-py
closed
[BUG REPORT] Commit inside context manager throws RuntimeError
Bug: Priority 2 PR In Progress
**Describe the bug** If we try to commit inside the context manager (before `__exit__()`), hangar throws RuntimeError saying `No changes made in the staging area. Cannot commit.`. We should allow the user to do commits inside the context manager IMO but probably with a warning about the performance hit **Severity** <!--- fill in the space between `[ ]` with and `x` (ie. `[x]`) ---> Select an option: - [ ] Data Corruption / Loss of Any Kind - [x] Unexpected Behavior, Exceptions or Error Thrown - [ ] Performance Bottleneck **To Reproduce** ```python import numpy as np from hangar import Repository repo = Repository(path='myhangarrepo') repo.init(user_name='Sherin Thomas', user_email='sherin@gmail.com', remove_old=True) # generate data data = [] for i in range(1000): data.append(np.random.rand(28, 28)) data = np.array(data) co = repo.checkout(write=True) data_dset = co.datasets.init_dataset('mnist_data', prototype=data[0]) co.commit('datasets init') co.close() co = repo.checkout(write=True) data_dset = co.datasets['mnist_data'] with data_dset: for i in range(len(data)): sample_name = str(i) data_dset[sample_name] = data[i] co.commit('dataset curation: stage 1') # this throws error co.close() ``` **Expected behavior** It should not break the program instead raise a warning about the performance hit
1.0
[BUG REPORT] Commit inside context manager throws RuntimeError - **Describe the bug** If we try to commit inside the context manager (before `__exit__()`), hangar throws RuntimeError saying `No changes made in the staging area. Cannot commit.`. We should allow the user to do commits inside the context manager IMO but probably with a warning about the performance hit **Severity** <!--- fill in the space between `[ ]` with and `x` (ie. `[x]`) ---> Select an option: - [ ] Data Corruption / Loss of Any Kind - [x] Unexpected Behavior, Exceptions or Error Thrown - [ ] Performance Bottleneck **To Reproduce** ```python import numpy as np from hangar import Repository repo = Repository(path='myhangarrepo') repo.init(user_name='Sherin Thomas', user_email='sherin@gmail.com', remove_old=True) # generate data data = [] for i in range(1000): data.append(np.random.rand(28, 28)) data = np.array(data) co = repo.checkout(write=True) data_dset = co.datasets.init_dataset('mnist_data', prototype=data[0]) co.commit('datasets init') co.close() co = repo.checkout(write=True) data_dset = co.datasets['mnist_data'] with data_dset: for i in range(len(data)): sample_name = str(i) data_dset[sample_name] = data[i] co.commit('dataset curation: stage 1') # this throws error co.close() ``` **Expected behavior** It should not break the program instead raise a warning about the performance hit
non_process
commit inside context manager throws runtimeerror describe the bug if we try to commit inside the context manager before exit hangar throws runtimeerror saying no changes made in the staging area cannot commit we should allow the user to do commits inside the context manager imo but probably with a warning about the performance hit severity select an option data corruption loss of any kind unexpected behavior exceptions or error thrown performance bottleneck to reproduce python import numpy as np from hangar import repository repo repository path myhangarrepo repo init user name sherin thomas user email sherin gmail com remove old true generate data data for i in range data append np random rand data np array data co repo checkout write true data dset co datasets init dataset mnist data prototype data co commit datasets init co close co repo checkout write true data dset co datasets with data dset for i in range len data sample name str i data dset data co commit dataset curation stage this throws error co close expected behavior it should not break the program instead raise a warning about the performance hit
0
35
2,505,372,297
IssuesEvent
2015-01-11 12:42:06
Graylog2/graylog2-server
https://api.github.com/repos/Graylog2/graylog2-server
closed
Ship a better/higher default configuration for output_batch_size
processing
We see significantly improved/less CPU usage by Elasticsearch when setting output_batch_size = 1000 or higher even when dealing with lower loads (around 250 msg/s). The default configuration value of 25 seems to be quite low if not too low, it might be good to raise it a bit. It probably would be also useful to have the batch size adjust dynamically based on relevant metric values.
1.0
Ship a better/higher default configuration for output_batch_size - We see significantly improved/less CPU usage by Elasticsearch when setting output_batch_size = 1000 or higher even when dealing with lower loads (around 250 msg/s). The default configuration value of 25 seems to be quite low if not too low, it might be good to raise it a bit. It probably would be also useful to have the batch size adjust dynamically based on relevant metric values.
process
ship a better higher default configuration for output batch size we see significantly improved less cpu usage by elasticsearch when setting output batch size or higher even when dealing with lower loads around msg s the default configuration value of seems to be quite low if not too low it might be good to raise it a bit it probably would be also useful to have the batch size adjust dynamically based on relevant metric values
1
19,059
25,078,181,879
IssuesEvent
2022-11-07 17:01:06
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
templateContext not well explained / exampled
doc-enhancement devops/prod Pri1 devops-cicd-process/tech
https://learn.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#use-templatecontext-to-pass-properties-to-templates I’d never noticed this functionality before but I’m struggling to understand the use case. I don’t feel it’s hugely well explained as it really just shows an if switch. I’ve done much the same type of thing as this example before with the if conditionals but it referenced a variable or parameters. Why do I want to use templateContext over those methods of if-ing off a variable or parameter. What advantage is it bringing to the table ? I seen the original blog post and again don’t feel it gave a good example. Can anyone give me more examples ? I posted this on azure devops on Reddit and the excellent Ming pointed out you can use this on top of a variable template but that still left me wondering why ? there’s clearly something I’m missing as you’d not create this functionality without a need to fill. thank you ! Much appreciated --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66 * Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065 * Content: [Templates - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops) * Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/templates.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
templateContext not well explained / exampled - https://learn.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#use-templatecontext-to-pass-properties-to-templates I’d never noticed this functionality before but I’m struggling to understand the use case. I don’t feel it’s hugely well explained as it really just shows an if switch. I’ve done much the same type of thing as this example before with the if conditionals but it referenced a variable or parameters. Why do I want to use templateContext over those methods of if-ing off a variable or parameter. What advantage is it bringing to the table ? I seen the original blog post and again don’t feel it gave a good example. Can anyone give me more examples ? I posted this on azure devops on Reddit and the excellent Ming pointed out you can use this on top of a variable template but that still left me wondering why ? there’s clearly something I’m missing as you’d not create this functionality without a need to fill. thank you ! Much appreciated --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66 * Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065 * Content: [Templates - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops) * Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/templates.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
templatecontext not well explained exampled i’d never noticed this functionality before but i’m struggling to understand the use case i don’t feel it’s hugely well explained as it really just shows an if switch i’ve done much the same type of thing as this example before with the if conditionals but it referenced a variable or parameters why do i want to use templatecontext over those methods of if ing off a variable or parameter what advantage is it bringing to the table i seen the original blog post and again don’t feel it gave a good example can anyone give me more examples i posted this on azure devops on reddit and the excellent ming pointed out you can use this on top of a variable template but that still left me wondering why there’s clearly something i’m missing as you’d not create this functionality without a need to fill thank you much appreciated document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id bbdc version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
214,324
7,268,889,980
IssuesEvent
2018-02-20 11:43:49
STEP-tw/battleship-phoenix
https://api.github.com/repos/STEP-tw/battleship-phoenix
closed
Game starts when both players are ready.
High Priority small
As a _player_ I want the _game to start_ So that I _can play_ **Additional Details** 1.Both players are ready.(assumption) **Acceptance Criteria** - [x] Criteria 1 - Given _opponent is ready and_ - When _I'm ready_ - Then _the game should start with a game started message_
1.0
Game starts when both players are ready. - As a _player_ I want the _game to start_ So that I _can play_ **Additional Details** 1.Both players are ready.(assumption) **Acceptance Criteria** - [x] Criteria 1 - Given _opponent is ready and_ - When _I'm ready_ - Then _the game should start with a game started message_
non_process
game starts when both players are ready as a player i want the game to start so that i can play additional details both players are ready assumption acceptance criteria criteria given opponent is ready and when i m ready then the game should start with a game started message
0
10,387
13,196,387,330
IssuesEvent
2020-08-13 20:32:59
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Triggers are not scheduled in UTC
Pri2 devops-cicd-process/tech devops/prod doc-bug investigating
Triggers are scheduled in the org's timezone, not UTC. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2ea2c851-bd1e-cddc-b4d0-e9f4112b8565 * Version Independent ID: 07c23fdd-14b5-985b-1c63-3f26f3a216ad * Content: [Configure schedules to run pipelines - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/scheduled-triggers?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/scheduled-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/scheduled-triggers.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @steved0x * Microsoft Alias: **sdanie**
1.0
Triggers are not scheduled in UTC - Triggers are scheduled in the org's timezone, not UTC. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2ea2c851-bd1e-cddc-b4d0-e9f4112b8565 * Version Independent ID: 07c23fdd-14b5-985b-1c63-3f26f3a216ad * Content: [Configure schedules to run pipelines - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/scheduled-triggers?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/scheduled-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/scheduled-triggers.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @steved0x * Microsoft Alias: **sdanie**
process
triggers are not scheduled in utc triggers are scheduled in the org s timezone not utc document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id cddc version independent id content content source product devops technology devops cicd process github login microsoft alias sdanie
1
4,182
7,114,540,309
IssuesEvent
2018-01-18 01:18:11
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
closed
Extract modifiers from comment
GLOBAL MYSQL PROTOCOL QUERY PROCESSOR cxx_pa development enhancement
Application should be able to send instructions and modify the behavior of the proxy using key/value pairs inside a comment. We need to define the list of variables.
1.0
Extract modifiers from comment - Application should be able to send instructions and modify the behavior of the proxy using key/value pairs inside a comment. We need to define the list of variables.
process
extract modifiers from comment application should be able to send instructions and modify the behavior of the proxy using key value pairs inside a comment we need to define the list of variables
1
15,976
20,188,183,795
IssuesEvent
2022-02-11 01:16:05
savitamittalmsft/WAS-SEC-TEST
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
opened
Scan container workloads for vulnerabilities
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Deployment & Testing Testing & Validation
<a href="https://docs.microsoft.com/azure/security-center/container-security">Scan container workloads for vulnerabilities</a> <p><b>Why Consider This?</b></p> To build secure containerized workloads, ensure the images that they're based on are free of known vulnerabilities. Shipping a known vulnerability in a container isn't significantly safer than on a VM. <p><b>Context</b></p> <p><span>Azure Security Center / Azure Defender is the Azure-native solution for securing containers. Azure Defender can protect virtual machines that are running Docker, Azure Kubernetes Service clusters, Azure Container Registry registries. Azure Defender is able to scan container images and identify security issues, or provide real-time threat detection for containerized environments.</span></p> <p><b>Suggested Actions</b></p> <p><span>Consider using Azure Defender for securing containerized workloads.</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/azure/security-center/container-security" target="_blank"><span>https://docs.microsoft.com/en-us/azure/security-center/container-security</span></a><span /></p>
1.0
Scan container workloads for vulnerabilities - <a href="https://docs.microsoft.com/azure/security-center/container-security">Scan container workloads for vulnerabilities</a> <p><b>Why Consider This?</b></p> To build secure containerized workloads, ensure the images that they're based on are free of known vulnerabilities. Shipping a known vulnerability in a container isn't significantly safer than on a VM. <p><b>Context</b></p> <p><span>Azure Security Center / Azure Defender is the Azure-native solution for securing containers. Azure Defender can protect virtual machines that are running Docker, Azure Kubernetes Service clusters, Azure Container Registry registries. Azure Defender is able to scan container images and identify security issues, or provide real-time threat detection for containerized environments.</span></p> <p><b>Suggested Actions</b></p> <p><span>Consider using Azure Defender for securing containerized workloads.</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/azure/security-center/container-security" target="_blank"><span>https://docs.microsoft.com/en-us/azure/security-center/container-security</span></a><span /></p>
process
scan container workloads for vulnerabilities why consider this to build secure containerized workloads ensure the images that they re based on are free of known vulnerabilities shipping a known vulnerability in a container isn t significantly safer than on a vm context azure security center azure defender is the azure native solution for securing containers azure defender can protect virtual machines that are running docker azure kubernetes service clusters azure container registry registries azure defender is able to scan container images and identify security issues or provide real time threat detection for containerized environments suggested actions consider using azure defender for securing containerized workloads learn more
1
82,383
10,279,473,812
IssuesEvent
2019-08-25 23:33:54
fga-desenho-2019-2/Wiki
https://api.github.com/repos/fga-desenho-2019-2/Wiki
opened
API de pagamento
Iniciativa extra Pesquisas documentation
## Descrição da Issue Procurar API's de pagamento e verificar como elas funcionam e precificação ### Tasks: - [ ] Listar API's de pagamento - [ ] Descrever o funcionamento delas e precificação
1.0
API de pagamento - ## Descrição da Issue Procurar API's de pagamento e verificar como elas funcionam e precificação ### Tasks: - [ ] Listar API's de pagamento - [ ] Descrever o funcionamento delas e precificação
non_process
api de pagamento descrição da issue procurar api s de pagamento e verificar como elas funcionam e precificação tasks listar api s de pagamento descrever o funcionamento delas e precificação
0
15,874
20,049,635,950
IssuesEvent
2022-02-03 03:43:33
q191201771/lal
https://api.github.com/repos/q191201771/lal
closed
开启 hls 并启用 use_memory_as_disk_flag 后,cleanup_mode 为0或者1时,会出现内存持续增长
#Question #Need doc *In process *Waiting reply
开启 hls 并启用 use_memory_as_disk_flag 后,m3u8 和 ts 会存储在内存中 当 cleanup_mode 为1或者2时,不会持续删除过期的 ts 文件,导致内存持续增长 建议在手册的 lalserver 配置文件说明 中注明
1.0
开启 hls 并启用 use_memory_as_disk_flag 后,cleanup_mode 为0或者1时,会出现内存持续增长 - 开启 hls 并启用 use_memory_as_disk_flag 后,m3u8 和 ts 会存储在内存中 当 cleanup_mode 为1或者2时,不会持续删除过期的 ts 文件,导致内存持续增长 建议在手册的 lalserver 配置文件说明 中注明
process
开启 hls 并启用 use memory as disk flag 后,cleanup mode ,会出现内存持续增长 开启 hls 并启用 use memory as disk flag 后, 和 ts 会存储在内存中 当 cleanup mode ,不会持续删除过期的 ts 文件,导致内存持续增长 建议在手册的 lalserver 配置文件说明 中注明
1
60,357
7,333,143,774
IssuesEvent
2018-03-05 18:27:45
juliett-golf-hotel/web-app
https://api.github.com/repos/juliett-golf-hotel/web-app
opened
Add hero banner to home page
content design dev
- [ ] Add hero banner images for different screen sizes - [ ] Add current temperature - [ ] Add feels like temperature - [ ] Add what the weather is (partly cloudy, hail, sun showers, etc) - [ ] Add a greeting
1.0
Add hero banner to home page - - [ ] Add hero banner images for different screen sizes - [ ] Add current temperature - [ ] Add feels like temperature - [ ] Add what the weather is (partly cloudy, hail, sun showers, etc) - [ ] Add a greeting
non_process
add hero banner to home page add hero banner images for different screen sizes add current temperature add feels like temperature add what the weather is partly cloudy hail sun showers etc add a greeting
0
1,324
3,874,111,167
IssuesEvent
2016-04-11 19:20:28
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
closed
IndexError for computing CSA
bug sct_process_segmentation
data: dropbox simon/results/2015-12-14 ~~~ sct_process_segmentation -i T2_seg.nii -p csa -t label/template/MNI-Poly-AMU_level.nii.gz -vert 1:1 Check parameters: .. segmentation file: T2_seg.nii Create temporary folder... mkdir tmp.151214225906_621711/ Copying input data to tmp folder and convert to nii... sct_convert -i /Users/julien/data/biospective/2015-12-14_sct2.1-2015-12-14/100-011_s2_T2/T2_seg.nii -o tmp.151214225906_621711/segmentation.nii.gz Change orientation to RPI... sct_image -i segmentation.nii.gz -setorient RPI -o segmentation_RPI.nii.gz Open segmentation volume... Get data dimensions... 56 x 288 x 288 Smooth centerline/segmentation... .. Get center of mass of the centerline/segmentation... .. Smoothing algo = hanning .. Windows length = 50 Compute CSA... Traceback (most recent call last): File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 815, in <module> main(sys.argv[1:]) File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 235, in main compute_csa(fname_segmentation, verbose, remove_temp_files, step, smoothing_param, figure_fit, param.file_csa_volume, slices, vert_lev, fname_vertebral_labeling, algo_fitting = param.algo_fitting, type_window= param.type_window, window_length=param.window_length) File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 487, in compute_csa normal = normalize(np.array([x_centerline_deriv[iz-min_z_index], y_centerline_deriv[iz-min_z_index], z_centerline_deriv[iz-min_z_index]])) IndexError: index 278 is out of bounds for axis 0 with size 278 ~~~
1.0
IndexError for computing CSA - data: dropbox simon/results/2015-12-14 ~~~ sct_process_segmentation -i T2_seg.nii -p csa -t label/template/MNI-Poly-AMU_level.nii.gz -vert 1:1 Check parameters: .. segmentation file: T2_seg.nii Create temporary folder... mkdir tmp.151214225906_621711/ Copying input data to tmp folder and convert to nii... sct_convert -i /Users/julien/data/biospective/2015-12-14_sct2.1-2015-12-14/100-011_s2_T2/T2_seg.nii -o tmp.151214225906_621711/segmentation.nii.gz Change orientation to RPI... sct_image -i segmentation.nii.gz -setorient RPI -o segmentation_RPI.nii.gz Open segmentation volume... Get data dimensions... 56 x 288 x 288 Smooth centerline/segmentation... .. Get center of mass of the centerline/segmentation... .. Smoothing algo = hanning .. Windows length = 50 Compute CSA... Traceback (most recent call last): File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 815, in <module> main(sys.argv[1:]) File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 235, in main compute_csa(fname_segmentation, verbose, remove_temp_files, step, smoothing_param, figure_fit, param.file_csa_volume, slices, vert_lev, fname_vertebral_labeling, algo_fitting = param.algo_fitting, type_window= param.type_window, window_length=param.window_length) File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 487, in compute_csa normal = normalize(np.array([x_centerline_deriv[iz-min_z_index], y_centerline_deriv[iz-min_z_index], z_centerline_deriv[iz-min_z_index]])) IndexError: index 278 is out of bounds for axis 0 with size 278 ~~~
process
indexerror for computing csa data dropbox simon results sct process segmentation i seg nii p csa t label template mni poly amu level nii gz vert check parameters segmentation file seg nii create temporary folder mkdir tmp copying input data to tmp folder and convert to nii sct convert i users julien data biospective seg nii o tmp segmentation nii gz change orientation to rpi sct image i segmentation nii gz setorient rpi o segmentation rpi nii gz open segmentation volume get data dimensions x x smooth centerline segmentation get center of mass of the centerline segmentation smoothing algo hanning windows length compute csa traceback most recent call last file users julien code spinalcordtoolbox bin sct process segmentation line in main sys argv file users julien code spinalcordtoolbox bin sct process segmentation line in main compute csa fname segmentation verbose remove temp files step smoothing param figure fit param file csa volume slices vert lev fname vertebral labeling algo fitting param algo fitting type window param type window window length param window length file users julien code spinalcordtoolbox bin sct process segmentation line in compute csa normal normalize np array y centerline deriv z centerline deriv indexerror index is out of bounds for axis with size
1
3,535
6,572,687,732
IssuesEvent
2017-09-11 04:25:42
zero-os/0-Disk
https://api.github.com/repos/zero-os/0-Disk
closed
add clusterID to ardb (failure) 0-log messages
process_wontfix type_feature
Currently only the address and db of the ardb storage server is given, but it would also be useful to give its clusterID alongside [ardb-storage-server-issues](https://github.com/zero-os/0-Disk/blob/master/docs/log.md#ardb-storage-server-issues) messages.
1.0
add clusterID to ardb (failure) 0-log messages - Currently only the address and db of the ardb storage server is given, but it would also be useful to give its clusterID alongside [ardb-storage-server-issues](https://github.com/zero-os/0-Disk/blob/master/docs/log.md#ardb-storage-server-issues) messages.
process
add clusterid to ardb failure log messages currently only the address and db of the ardb storage server is given but it would also be useful to give its clusterid alongside messages
1
17,763
23,691,657,953
IssuesEvent
2022-08-29 11:22:56
ankidroid/Anki-Android
https://api.github.com/repos/ankidroid/Anki-Android
opened
Make code coverage reporting stable
Dev Test process
`codecov` often reports random fluctuations in the coverage. This is likely because tests are nondeterministic, but may be a bug in codecov. * Determine why this occurs (typically by finding the classes which fluctuate) * Fix this, so we no longer get `codecov`-based CI failures on changes which should be no-ops
1.0
Make code coverage reporting stable - `codecov` often reports random fluctuations in the coverage. This is likely because tests are nondeterministic, but may be a bug in codecov. * Determine why this occurs (typically by finding the classes which fluctuate) * Fix this, so we no longer get `codecov`-based CI failures on changes which should be no-ops
process
make code coverage reporting stable codecov often reports random fluctuations in the coverage this is likely because tests are nondeterministic but may be a bug in codecov determine why this occurs typically by finding the classes which fluctuate fix this so we no longer get codecov based ci failures on changes which should be no ops
1
75,081
3,455,052,267
IssuesEvent
2015-12-17 18:26:21
ThoughtWorksInc/registrolivre
https://api.github.com/repos/ThoughtWorksInc/registrolivre
closed
Bug no highlight do formulário de cadastro de empresa
Bug Priority
Bug no highlight do formulário de cadastro de empresas. Ao tentar adicionar sócios na empresa, caso um CPF esteja errado o highlight aparece nos dois formulários mesmo que o segundo CPF esteja correto. ![screen shot 2015-12-07 at 6 22 51 pm](https://cloud.githubusercontent.com/assets/5021565/11638721/92ce8a08-9d0f-11e5-80e5-282eed445e70.png)
1.0
Bug no highlight do formulário de cadastro de empresa - Bug no highlight do formulário de cadastro de empresas. Ao tentar adicionar sócios na empresa, caso um CPF esteja errado o highlight aparece nos dois formulários mesmo que o segundo CPF esteja correto. ![screen shot 2015-12-07 at 6 22 51 pm](https://cloud.githubusercontent.com/assets/5021565/11638721/92ce8a08-9d0f-11e5-80e5-282eed445e70.png)
non_process
bug no highlight do formulário de cadastro de empresa bug no highlight do formulário de cadastro de empresas ao tentar adicionar sócios na empresa caso um cpf esteja errado o highlight aparece nos dois formulários mesmo que o segundo cpf esteja correto
0
18,775
24,678,010,403
IssuesEvent
2022-10-18 18:41:02
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Monitor slow to startup in Kubernetes
bug process monitor
### Description The monitor is now taking too long to startup that its liveness probe is causing it to restart. ### Steps to reproduce Run monitor on Kubernetes ### Additional context _No response_ ### Hedera network other ### Version main ### Operating system _No response_
1.0
Monitor slow to startup in Kubernetes - ### Description The monitor is now taking too long to startup that its liveness probe is causing it to restart. ### Steps to reproduce Run monitor on Kubernetes ### Additional context _No response_ ### Hedera network other ### Version main ### Operating system _No response_
process
monitor slow to startup in kubernetes description the monitor is now taking too long to startup that its liveness probe is causing it to restart steps to reproduce run monitor on kubernetes additional context no response hedera network other version main operating system no response
1
311,218
23,376,086,577
IssuesEvent
2022-08-11 03:19:10
singularity-data/risingwave-docs
https://api.github.com/repos/singularity-data/risingwave-docs
closed
Update the lookup join behavior
documentation
### Related code PR PR: https://github.com/singularity-data/risingwave/pull/4207 Code issue: https://github.com/singularity-data/risingwave/issues/4044 ### Which part(s) of the docs might be affected or should be updated? And how? Add this code change (issue fixes) to Release Notes. When using lookup joins (JOIN ... ON ...), RW now will cast the data types if the data type of the two key columns are different, instead of throwing an error. ### Reference _No response_
1.0
Update the lookup join behavior - ### Related code PR PR: https://github.com/singularity-data/risingwave/pull/4207 Code issue: https://github.com/singularity-data/risingwave/issues/4044 ### Which part(s) of the docs might be affected or should be updated? And how? Add this code change (issue fixes) to Release Notes. When using lookup joins (JOIN ... ON ...), RW now will cast the data types if the data type of the two key columns are different, instead of throwing an error. ### Reference _No response_
non_process
update the lookup join behavior related code pr pr code issue which part s of the docs might be affected or should be updated and how add this code change issue fixes to release notes when using lookup joins join on rw now will cast the data types if the data type of the two key columns are different instead of throwing an error reference no response
0
82,645
10,300,605,138
IssuesEvent
2019-08-28 16:14:13
ISibboI/graphrepresentations
https://api.github.com/repos/ISibboI/graphrepresentations
opened
Fix documentation asserts
documentation
Asserting that the node ids are output in order does not prove anything or help the user understand. It should be asserted that the data is at the correct place.
1.0
Fix documentation asserts - Asserting that the node ids are output in order does not prove anything or help the user understand. It should be asserted that the data is at the correct place.
non_process
fix documentation asserts asserting that the node ids are output in order does not prove anything or help the user understand it should be asserted that the data is at the correct place
0
11,315
14,134,856,491
IssuesEvent
2020-11-10 00:15:56
googleapis/python-speech
https://api.github.com/repos/googleapis/python-speech
closed
speec2srt.py is not working because of removal of submodules enums and types
api: speech type: process
Following [this](https://www.youtube.com/watch?v=uBzp5xGSZ6o&ab_channel=GoogleCloudAPAC) for creating speech to srt using google apis. When running the command `python3 speech2srt.py --storage_uri gs://subtitlingsc/en.wav` is showing that ``` File "speech2srt.py", line 19, in <module> from google.cloud.speech_v1 import enums ImportError: cannot import name 'enums' from 'google.cloud.speech_v1' ``` I think this is due to removal of enums and types. Please do the needful and update those videos and [GitHub file](https://github.com/GoogleCloudPlatform/community/blob/master/tutorials/speech2srt/speech2srt.py)
1.0
speec2srt.py is not working because of removal of submodules enums and types - Following [this](https://www.youtube.com/watch?v=uBzp5xGSZ6o&ab_channel=GoogleCloudAPAC) for creating speech to srt using google apis. When running the command `python3 speech2srt.py --storage_uri gs://subtitlingsc/en.wav` is showing that ``` File "speech2srt.py", line 19, in <module> from google.cloud.speech_v1 import enums ImportError: cannot import name 'enums' from 'google.cloud.speech_v1' ``` I think this is due to removal of enums and types. Please do the needful and update those videos and [GitHub file](https://github.com/GoogleCloudPlatform/community/blob/master/tutorials/speech2srt/speech2srt.py)
process
py is not working because of removal of submodules enums and types following for creating speech to srt using google apis when running the command py storage uri gs subtitlingsc en wav is showing that file py line in from google cloud speech import enums importerror cannot import name enums from google cloud speech i think this is due to removal of enums and types please do the needful and update those videos and
1
136,030
30,462,136,059
IssuesEvent
2023-07-17 07:47:25
FerretDB/FerretDB
https://api.github.com/repos/FerretDB/FerretDB
opened
Integration tests should report their progress
code/chore not ready
### What should be done? The integration tests should print something meaningful to the console once in a while, like "running test 24/1032". Right now they can take 10-30 minutes on a dev machine without printing a thing. The terminal looks like it hangs, there is no indication of progress. A natural reaction to that is `ctrl+C` (which doesn't allow the tests to complete). It could be a good idea to disable this progress printing when running the tests from CI (to avoid some redundant logs in a non-interactive environment). ### Where? The integration tests runner maybe? I haven't looked at the relevant test runner yet. :swe ### Definition of Done - test runner is updated
1.0
Integration tests should report their progress - ### What should be done? The integration tests should print something meaningful to the console once in a while, like "running test 24/1032". Right now they can take 10-30 minutes on a dev machine without printing a thing. The terminal looks like it hangs, there is no indication of progress. A natural reaction to that is `ctrl+C` (which doesn't allow the tests to complete). It could be a good idea to disable this progress printing when running the tests from CI (to avoid some redundant logs in a non-interactive environment). ### Where? The integration tests runner maybe? I haven't looked at the relevant test runner yet. :swe ### Definition of Done - test runner is updated
non_process
integration tests should report their progress what should be done the integration tests should print something meaningful to the console once in a while like running test right now they can take minutes on a dev machine without printing a thing the terminal looks like it hangs there is no indication of progress a natural reaction to that is ctrl c which doesn t allow the tests to complete it could be a good idea to disable this progress printing when running the tests from ci to avoid some redundant logs in a non interactive environment where the integration tests runner maybe i haven t looked at the relevant test runner yet swe definition of done test runner is updated
0
11,304
14,107,274,324
IssuesEvent
2020-11-06 16:03:57
panther-labs/panther
https://api.github.com/repos/panther-labs/panther
opened
Invalid JSON in test crashes entire test suite
bug p1 team:data processing
### Describe the bug Currently, when adding an invalid input as a test body, I expect to get back a structured response including a particular set of keys (like `ruleError` and `genericError`), stating that I have added invalid content. Instead, the entire lambda fails, causes a runtime issue and AppSync doesn't return any structured data. ### Steps to reproduce Steps to reproduce the behavior: 1. Create a New Rule 2. Add a test 3. Fill the body of the test with the content ``` { "uuid": "123" } { "uuid": "123" } ``` 4. Click "Run all" ### Expected behavior There should be a structured API response with a specific key containing the error ### Environment How are you deploying or using Panther? - Panther version or commit: v1.12 - OS: Mac - Browser: Chrome ### Additional context If I have 20 tests and 1 of them has invalid body, I will never get feedback for **any of the other 19 tests**, since the entire API response gets replaced by the runtime error ### Screenshots <img width="1287" alt="Screen Shot 2020-11-06 at 5 55 55 PM" src="https://user-images.githubusercontent.com/10436045/98387474-19630f80-205a-11eb-88ed-5c358783b789.png">
1.0
Invalid JSON in test crashes entire test suite - ### Describe the bug Currently, when adding an invalid input as a test body, I expect to get back a structured response including a particular set of keys (like `ruleError` and `genericError`), stating that I have added invalid content. Instead, the entire lambda fails, causes a runtime issue and AppSync doesn't return any structured data. ### Steps to reproduce Steps to reproduce the behavior: 1. Create a New Rule 2. Add a test 3. Fill the body of the test with the content ``` { "uuid": "123" } { "uuid": "123" } ``` 4. Click "Run all" ### Expected behavior There should be a structured API response with a specific key containing the error ### Environment How are you deploying or using Panther? - Panther version or commit: v1.12 - OS: Mac - Browser: Chrome ### Additional context If I have 20 tests and 1 of them has invalid body, I will never get feedback for **any of the other 19 tests**, since the entire API response gets replaced by the runtime error ### Screenshots <img width="1287" alt="Screen Shot 2020-11-06 at 5 55 55 PM" src="https://user-images.githubusercontent.com/10436045/98387474-19630f80-205a-11eb-88ed-5c358783b789.png">
process
invalid json in test crashes entire test suite describe the bug currently when adding an invalid input as a test body i expect to get back a structured response including a particular set of keys like ruleerror and genericerror stating that i have added invalid content instead the entire lambda fails causes a runtime issue and appsync doesn t return any structured data steps to reproduce steps to reproduce the behavior create a new rule add a test fill the body of the test with the content uuid uuid click run all expected behavior there should be a structured api response with a specific key containing the error environment how are you deploying or using panther panther version or commit os mac browser chrome additional context if i have tests and of them has invalid body i will never get feedback for any of the other tests since the entire api response gets replaced by the runtime error screenshots img width alt screen shot at pm src
1
20,590
27,254,450,947
IssuesEvent
2023-02-22 10:27:51
python/cpython
https://api.github.com/repos/python/cpython
reopened
Deadlock when using fork whilst multiprocessing.resource_tracker._resource_tracker._lock is held
type-bug stdlib expert-multiprocessing
# Bug report Given the following situation: You have a Process "a" with two threads ("aa", "ab"). aa is currently creating a shared memory segment and is holding `multiprocessing.resource_tracker._resource_tracker._lock` for that reason. Note that that lock is not acquired by user code, but deep inside the SharedMemory constructor. ab now wants to create a second process using the `fork` method. Process b is created with one thread: bb. As per `man fork` the thread aa is not duplicated in process b. If aa now finishes creating its SharedMemory it releases `multiprocessing.resource_tracker._resource_tracker._lock`. But that lock is only a `threading.Lock`, which means it is released only in process a. In process b it will not be released. If b now wants to create a SharedMemory object and therefore tries to acquire `multiprocessing.resource_tracker._resource_tracker._lock` this will take forever, as it will never be free in process b. `man fork` explicitly mentions that such situations are a potential source of issues. As this is a race condition there is no minimal, reproducible example I can give. # Possible solutions One could replace this `threading.Lock` with a `multiprocessing.Lock`. Alternatively one could replace that lock with a new one in the child process after a fork. Im not sure about the intended behavior of the resource tracker in such situations. # Your environment <!-- Include as many relevant details as possible about the environment you experienced the bug in --> - CPython versions tested on: 3.9, 3.10 - Operating system and architecture: Unix (Arch, Kernel 5.19.7-1) <!-- You can freely edit this text. Remove any lines you believe are unnecessary. -->
1.0
Deadlock when using fork whilst multiprocessing.resource_tracker._resource_tracker._lock is held - # Bug report Given the following situation: You have a Process "a" with two threads ("aa", "ab"). aa is currently creating a shared memory segment and is holding `multiprocessing.resource_tracker._resource_tracker._lock` for that reason. Note that that lock is not acquired by user code, but deep inside the SharedMemory constructor. ab now wants to create a second process using the `fork` method. Process b is created with one thread: bb. As per `man fork` the thread aa is not duplicated in process b. If aa now finishes creating its SharedMemory it releases `multiprocessing.resource_tracker._resource_tracker._lock`. But that lock is only a `threading.Lock`, which means it is released only in process a. In process b it will not be released. If b now wants to create a SharedMemory object and therefore tries to acquire `multiprocessing.resource_tracker._resource_tracker._lock` this will take forever, as it will never be free in process b. `man fork` explicitly mentions that such situations are a potential source of issues. As this is a race condition there is no minimal, reproducible example I can give. # Possible solutions One could replace this `threading.Lock` with a `multiprocessing.Lock`. Alternatively one could replace that lock with a new one in the child process after a fork. Im not sure about the intended behavior of the resource tracker in such situations. # Your environment <!-- Include as many relevant details as possible about the environment you experienced the bug in --> - CPython versions tested on: 3.9, 3.10 - Operating system and architecture: Unix (Arch, Kernel 5.19.7-1) <!-- You can freely edit this text. Remove any lines you believe are unnecessary. -->
process
deadlock when using fork whilst multiprocessing resource tracker resource tracker lock is held bug report given the following situation you have a process a with two threads aa ab aa is currently creating a shared memory segment and is holding multiprocessing resource tracker resource tracker lock for that reason note that that lock is not acquired by user code but deep inside the sharedmemory constructor ab now wants to create a second process using the fork method process b is created with one thread bb as per man fork the thread aa is not duplicated in process b if aa now finishes creating its sharedmemory it releases multiprocessing resource tracker resource tracker lock but that lock is only a threading lock which means it is released only in process a in process b it will not be released if b now wants to create a sharedmemory object and therefore tries to acquire multiprocessing resource tracker resource tracker lock this will take forever as it will never be free in process b man fork explicitly mentions that such situations are a potential source of issues as this is a race condition there is no minimal reproducible example i can give possible solutions one could replace this threading lock with a multiprocessing lock alternatively one could replace that lock with a new one in the child process after a fork im not sure about the intended behavior of the resource tracker in such situations your environment cpython versions tested on operating system and architecture unix arch kernel you can freely edit this text remove any lines you believe are unnecessary
1
5,475
8,354,087,548
IssuesEvent
2018-10-02 12:18:23
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
copy-to ignored in preprocess2
P1 bug preprocess2
## Expected Behavior Using `@copy-to` should result in an extra copy of the referenced topic. If file names are relevant for output, the specified names will be used. ## Actual Behavior The `@copy-to` attribute is not processed, so a copy is not created. In PDF output, this results in a broken link and missing file. There are several errors when different steps try to read the copy that does not exist. In 3.1.2 this ends in a build failure caused by the missing file; #3082 removes the build failure and replaces it with an error like the other steps. ## Possible Solution The error comes from this line: https://github.com/dita-ot/dita-ot/blob/develop/src/main/java/org/dita/dost/module/CopyToModule.java#L148 In an HTML build with `preprocess`, the file listed in `.job.xml` does not have `@src` until after `copy-to` is evaluated. That listing immediately before copy-to: `<file uri="copyto.dita" path="copyto.dita" result="file:/C:/DITA-OT/TESTGITHUB/copyto/copyto.dita">` and after: `<file src="file:/C:/DITA-OT/TESTGITHUB/copyto/original.dita" uri="copyto.dita" path="copyto.dita" result="file:/C:/DITA-OT/TESTGITHUB/copyto/copyto.dita" format="dita" target="true">` In a PDF build with `preprocess2`, that same file has `@src` after the first step (`map-reader`). That line in `preprocess2` assumes that if `@src` is listed, the file was referenced as a source file, so we shouldn't copy anything to it. That listing before copy-to: `<file src="file:/C:/DITA-OT/TESTGITHUB/copyto/original.dita" uri="f7ac72d7ff9763eb29f90236b7c20b67fce3a84a.dita" path="f7ac72d7ff9763eb29f90236b7c20b67fce3a84a.dita" result="file:/C:/DITA-OT/TESTGITHUB/copyto/original.dita" format="dita" target="true"> </file>` I'm not yet sure where in `map-reader` the `@src` attribute is set up, otherwise this would be coming in as a pull request. I _suspect_ that if we just leave off `@src` for the copies at this point, it will give the correct results, but I haven't been able to test that so I'm not sure if it would affect other map steps. ## Steps to Reproduce Build this to PDF - it will result in the build failures with 3.1.2, but you can see in temp that the copies are not created, and other steps will warn when trying to access them. The map has commented out copies of the `<files>` section from HTML (before/after copy-to runs), and PDF (before `copy-to` runs). With the fix for #3082 in place, it will produce a PDF that has 1 topic instead of 3. [copyto.zip](https://github.com/dita-ot/dita-ot/files/2429029/copyto.zip) ## Environment <!-- Include relevant details about the environment you experienced this in. --> * DITA-OT version: 3.1.2 and `hotfix/3.1.3` * Operating system and version: Windows * How did you run DITA-OT? `dita` * Transformation type: PDF
1.0
copy-to ignored in preprocess2 - ## Expected Behavior Using `@copy-to` should result in an extra copy of the referenced topic. If file names are relevant for output, the specified names will be used. ## Actual Behavior The `@copy-to` attribute is not processed, so a copy is not created. In PDF output, this results in a broken link and missing file. There are several errors when different steps try to read the copy that does not exist. In 3.1.2 this ends in a build failure caused by the missing file; #3082 removes the build failure and replaces it with an error like the other steps. ## Possible Solution The error comes from this line: https://github.com/dita-ot/dita-ot/blob/develop/src/main/java/org/dita/dost/module/CopyToModule.java#L148 In an HTML build with `preprocess`, the file listed in `.job.xml` does not have `@src` until after `copy-to` is evaluated. That listing immediately before copy-to: `<file uri="copyto.dita" path="copyto.dita" result="file:/C:/DITA-OT/TESTGITHUB/copyto/copyto.dita">` and after: `<file src="file:/C:/DITA-OT/TESTGITHUB/copyto/original.dita" uri="copyto.dita" path="copyto.dita" result="file:/C:/DITA-OT/TESTGITHUB/copyto/copyto.dita" format="dita" target="true">` In a PDF build with `preprocess2`, that same file has `@src` after the first step (`map-reader`). That line in `preprocess2` assumes that if `@src` is listed, the file was referenced as a source file, so we shouldn't copy anything to it. That listing before copy-to: `<file src="file:/C:/DITA-OT/TESTGITHUB/copyto/original.dita" uri="f7ac72d7ff9763eb29f90236b7c20b67fce3a84a.dita" path="f7ac72d7ff9763eb29f90236b7c20b67fce3a84a.dita" result="file:/C:/DITA-OT/TESTGITHUB/copyto/original.dita" format="dita" target="true"> </file>` I'm not yet sure where in `map-reader` the `@src` attribute is set up, otherwise this would be coming in as a pull request. I _suspect_ that if we just leave off `@src` for the copies at this point, it will give the correct results, but I haven't been able to test that so I'm not sure if it would affect other map steps. ## Steps to Reproduce Build this to PDF - it will result in the build failures with 3.1.2, but you can see in temp that the copies are not created, and other steps will warn when trying to access them. The map has commented out copies of the `<files>` section from HTML (before/after copy-to runs), and PDF (before `copy-to` runs). With the fix for #3082 in place, it will produce a PDF that has 1 topic instead of 3. [copyto.zip](https://github.com/dita-ot/dita-ot/files/2429029/copyto.zip) ## Environment <!-- Include relevant details about the environment you experienced this in. --> * DITA-OT version: 3.1.2 and `hotfix/3.1.3` * Operating system and version: Windows * How did you run DITA-OT? `dita` * Transformation type: PDF
process
copy to ignored in expected behavior using copy to should result in an extra copy of the referenced topic if file names are relevant for output the specified names will be used actual behavior the copy to attribute is not processed so a copy is not created in pdf output this results in a broken link and missing file there are several errors when different steps try to read the copy that does not exist in this ends in a build failure caused by the missing file removes the build failure and replaces it with an error like the other steps possible solution the error comes from this line in an html build with preprocess the file listed in job xml does not have src until after copy to is evaluated that listing immediately before copy to and after in a pdf build with that same file has src after the first step map reader that line in assumes that if src is listed the file was referenced as a source file so we shouldn t copy anything to it that listing before copy to i m not yet sure where in map reader the src attribute is set up otherwise this would be coming in as a pull request i suspect that if we just leave off src for the copies at this point it will give the correct results but i haven t been able to test that so i m not sure if it would affect other map steps steps to reproduce build this to pdf it will result in the build failures with but you can see in temp that the copies are not created and other steps will warn when trying to access them the map has commented out copies of the section from html before after copy to runs and pdf before copy to runs with the fix for in place it will produce a pdf that has topic instead of environment dita ot version and hotfix operating system and version windows how did you run dita ot dita transformation type pdf
1
62,481
6,797,828,979
IssuesEvent
2017-11-02 01:18:31
brave/browser-laptop
https://api.github.com/repos/brave/browser-laptop
reopened
Fingerprinting protections bypassable
feature/shields privacy QA/checked-Linux QA/checked-macOS QA/checked-Win64 QA/test-plan-specified release-notes/include release/blocking security
## Test plan https://github.com/brave/browser-laptop/issues/11683#issuecomment-340532797 ---- ### Description Many of the current fingerprinting protections rely on removing references to methods from the global object. However, you can get alternate references to the same methods by inserting a configured iframe element and grabbing the references off `iframe.contentWindow` or `iframe.contentDocument`. ### Steps to Reproduce ``` //blocked console.log(HTMLCanvasElement.prototype.toDataURL); let iframe = document.createElement("iframe"); iframe.src = "https://www.brave.com/"; document.body.appendChild(iframe); // Not blocked console.log(iframe.contentWindow.HTMLCanvasElement.prototype.toDataURL); // Do the bad stuff let canvasToDataUrl = iframe.contentWindow.HTMLCanvasElement.prototype.toDataURL; canvasToDataUrl.apply(someParentFrameCanvasElement); ``` **Actual result:** You can get references to the fingerprinting related methods. **Expected result:** You shouldn't be able to get to these methods. **Reproduces how often:** 100% ### Brave Version All versions **Reproducible on current live release:** Yes
1.0
Fingerprinting protections bypassable - ## Test plan https://github.com/brave/browser-laptop/issues/11683#issuecomment-340532797 ---- ### Description Many of the current fingerprinting protections rely on removing references to methods from the global object. However, you can get alternate references to the same methods by inserting a configured iframe element and grabbing the references off `iframe.contentWindow` or `iframe.contentDocument`. ### Steps to Reproduce ``` //blocked console.log(HTMLCanvasElement.prototype.toDataURL); let iframe = document.createElement("iframe"); iframe.src = "https://www.brave.com/"; document.body.appendChild(iframe); // Not blocked console.log(iframe.contentWindow.HTMLCanvasElement.prototype.toDataURL); // Do the bad stuff let canvasToDataUrl = iframe.contentWindow.HTMLCanvasElement.prototype.toDataURL; canvasToDataUrl.apply(someParentFrameCanvasElement); ``` **Actual result:** You can get references to the fingerprinting related methods. **Expected result:** You shouldn't be able to get to these methods. **Reproduces how often:** 100% ### Brave Version All versions **Reproducible on current live release:** Yes
non_process
fingerprinting protections bypassable test plan description many of the current fingerprinting protections rely on removing references to methods from the global object however you can get alternate references to the same methods by inserting a configured iframe element and grabbing the references off iframe contentwindow or iframe contentdocument steps to reproduce blocked console log htmlcanvaselement prototype todataurl let iframe document createelement iframe iframe src document body appendchild iframe not blocked console log iframe contentwindow htmlcanvaselement prototype todataurl do the bad stuff let canvastodataurl iframe contentwindow htmlcanvaselement prototype todataurl canvastodataurl apply someparentframecanvaselement actual result you can get references to the fingerprinting related methods expected result you shouldn t be able to get to these methods reproduces how often brave version all versions reproducible on current live release yes
0
16,594
21,645,947,690
IssuesEvent
2022-05-06 01:54:55
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
k8sattributes processor: support flexible labels/annotation regex mapping
proc: k8sprocessor
[This PR](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/5780) adds ability to fetch pod/namespace annotations and labels based on regex. Attribute values created as the result have only one predefined form `k8s.pod(namespace).label.$0`. This issue is to track a follow up improvement to support more flexible construction of the attribute values as suggested in [this comment](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/5780#discussion_r732390230)
1.0
k8sattributes processor: support flexible labels/annotation regex mapping - [This PR](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/5780) adds ability to fetch pod/namespace annotations and labels based on regex. Attribute values created as the result have only one predefined form `k8s.pod(namespace).label.$0`. This issue is to track a follow up improvement to support more flexible construction of the attribute values as suggested in [this comment](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/5780#discussion_r732390230)
process
processor support flexible labels annotation regex mapping adds ability to fetch pod namespace annotations and labels based on regex attribute values created as the result have only one predefined form pod namespace label this issue is to track a follow up improvement to support more flexible construction of the attribute values as suggested in
1
3,091
6,107,772,018
IssuesEvent
2017-06-21 08:55:59
rec-framework/rec-core
https://api.github.com/repos/rec-framework/rec-core
closed
Add encoding convertor?
feature/data-processing priority/Medium request/suggestion
Issue found in #35 , when specified CSV file in utf-8 mode, it will cause potential encoding problem. Suggest add a conversion method to convert file from one encoding to another.
1.0
Add encoding convertor? - Issue found in #35 , when specified CSV file in utf-8 mode, it will cause potential encoding problem. Suggest add a conversion method to convert file from one encoding to another.
process
add encoding convertor issue found in when specified csv file in utf mode it will cause potential encoding problem suggest add a conversion method to convert file from one encoding to another
1
17,430
23,249,029,817
IssuesEvent
2022-08-04 00:20:21
beyondhb1079/s4us
https://api.github.com/repos/beyondhb1079/s4us
opened
Email draft for reaching out
process
Come up with a compelling email draft for reaching out to new contacts
1.0
Email draft for reaching out - Come up with a compelling email draft for reaching out to new contacts
process
email draft for reaching out come up with a compelling email draft for reaching out to new contacts
1
165,307
26,142,449,088
IssuesEvent
2022-12-29 20:46:42
audacity/audacity
https://api.github.com/repos/audacity/audacity
closed
Track resampling affects to track speed
P4 Design / UX
**Describe the bug** Track resampling works incorrectly **To Reproduce** 1. Launch Audacity 2. Change rate in audio-interface settings 3. Click Play in Audacity **Expected result:** Track rate has changed correctly and does not affect track speed. Actual Result: Track speed had been changed Win 10 Audacity 3.0.3. Focusrite Scarlett 2i2 https://user-images.githubusercontent.com/25851359/127655827-c7798f6e-cbdf-409d-806c-79c42271391b.mp4
1.0
Track resampling affects to track speed - **Describe the bug** Track resampling works incorrectly **To Reproduce** 1. Launch Audacity 2. Change rate in audio-interface settings 3. Click Play in Audacity **Expected result:** Track rate has changed correctly and does not affect track speed. Actual Result: Track speed had been changed Win 10 Audacity 3.0.3. Focusrite Scarlett 2i2 https://user-images.githubusercontent.com/25851359/127655827-c7798f6e-cbdf-409d-806c-79c42271391b.mp4
non_process
track resampling affects to track speed describe the bug track resampling works incorrectly to reproduce launch audacity change rate in audio interface settings click play in audacity expected result track rate has changed correctly and does not affect track speed actual result track speed had been changed win audacity focusrite scarlett
0
121,452
10,168,817,266
IssuesEvent
2019-08-07 21:55:20
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
pull-kubernetes-e2e-aks-engine-azure consistently failing
area/provider/azure kind/failing-test priority/important-soon sig/cloud-provider
<!-- Please only use this template for submitting reports about failing tests in Kubernetes CI jobs --> **Which jobs are failing**: pull-kubernetes-e2e-aks-engine-azure **Which test(s) are failing**: `Up` **Since when has it been failing**: [2019-05-10 05:20:51 +0000 UTC](https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/77719/pull-kubernetes-e2e-aks-engine-azure/1126718651890667520) **Testgrid link**: No Testgrid, but job history: https://prow.k8s.io/job-history/kubernetes-jenkins/pr-logs/directory/pull-kubernetes-e2e-aks-engine-azure **Reason for failure**: ``` error creating cluster: cannot deploy: cannot get the create deployment future response: Code="DeploymentFailed" Message="At least one resource deployment operation failed. Please list deployment operations for details. Please see https://aka.ms/arm-debug for usage details." Details=[{"code":"Conflict","message":"{\r\n \"status\": \"Failed\",\r\n \"error\": {\r\n \"code\": \"ResourceDeploymentFailure\",\r\n \"message\": \"The resource operation completed with terminal provisioning state 'Failed'.\",\r\n \"details\": [\r\n {\r\n \"code\": \"VMExtensionProvisioningError\",\r\n \"message\": \"VM has reported a failure when processing extension 'cse-master-0'. Error message: \\\"Enable failed: failed to execute command: command terminated with exit status=30\\n[stdout]\\n\\n[stderr]\\nConnection to k8s.gcr.io 443 port [tcp/https] succeeded!\\nConnection to gcr.io 443 port [tcp/https] succeeded!\\nConnection to docker.io 443 port [tcp/https] succeeded!\\n\\\".\"\r\n }\r\n ]\r\n }\r\n}"}] ``` **Anything else we need to know**: ref: https://github.com/kubernetes/kubernetes/pull/79583#issuecomment-507224973
1.0
pull-kubernetes-e2e-aks-engine-azure consistently failing - <!-- Please only use this template for submitting reports about failing tests in Kubernetes CI jobs --> **Which jobs are failing**: pull-kubernetes-e2e-aks-engine-azure **Which test(s) are failing**: `Up` **Since when has it been failing**: [2019-05-10 05:20:51 +0000 UTC](https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/77719/pull-kubernetes-e2e-aks-engine-azure/1126718651890667520) **Testgrid link**: No Testgrid, but job history: https://prow.k8s.io/job-history/kubernetes-jenkins/pr-logs/directory/pull-kubernetes-e2e-aks-engine-azure **Reason for failure**: ``` error creating cluster: cannot deploy: cannot get the create deployment future response: Code="DeploymentFailed" Message="At least one resource deployment operation failed. Please list deployment operations for details. Please see https://aka.ms/arm-debug for usage details." Details=[{"code":"Conflict","message":"{\r\n \"status\": \"Failed\",\r\n \"error\": {\r\n \"code\": \"ResourceDeploymentFailure\",\r\n \"message\": \"The resource operation completed with terminal provisioning state 'Failed'.\",\r\n \"details\": [\r\n {\r\n \"code\": \"VMExtensionProvisioningError\",\r\n \"message\": \"VM has reported a failure when processing extension 'cse-master-0'. Error message: \\\"Enable failed: failed to execute command: command terminated with exit status=30\\n[stdout]\\n\\n[stderr]\\nConnection to k8s.gcr.io 443 port [tcp/https] succeeded!\\nConnection to gcr.io 443 port [tcp/https] succeeded!\\nConnection to docker.io 443 port [tcp/https] succeeded!\\n\\\".\"\r\n }\r\n ]\r\n }\r\n}"}] ``` **Anything else we need to know**: ref: https://github.com/kubernetes/kubernetes/pull/79583#issuecomment-507224973
non_process
pull kubernetes aks engine azure consistently failing which jobs are failing pull kubernetes aks engine azure which test s are failing up since when has it been failing testgrid link no testgrid but job history reason for failure error creating cluster cannot deploy cannot get the create deployment future response code deploymentfailed message at least one resource deployment operation failed please list deployment operations for details please see for usage details details n n nconnection to gcr io port succeeded nconnection to gcr io port succeeded nconnection to docker io port succeeded n r n r n r n r n anything else we need to know ref
0
99,715
21,032,253,669
IssuesEvent
2022-03-31 02:33:34
rss2email/rss2email
https://api.github.com/repos/rss2email/rss2email
closed
Feature Request: Modify the subject line
is:enhancement good first issue needs:code help wanted
I checked the docs and tried a lot of things but I couldn't seem to figure out a way to do this. Some email providers, such as Microsoft Office 365 will strip / replace the From name & address with that of the account that's being used to send. In these cases it would be better to use the subject line to display the feed name, author etc as opposed to trying to modify the From address. In addition, it would be good to be able to add your own subject items so R2E messages can be tagged for setting up mail sorting filters. If there is a way to do this then I apologize in advance for opening an issue - if that is the case, please inform / update docs on how to do this. Thank you,
1.0
Feature Request: Modify the subject line - I checked the docs and tried a lot of things but I couldn't seem to figure out a way to do this. Some email providers, such as Microsoft Office 365 will strip / replace the From name & address with that of the account that's being used to send. In these cases it would be better to use the subject line to display the feed name, author etc as opposed to trying to modify the From address. In addition, it would be good to be able to add your own subject items so R2E messages can be tagged for setting up mail sorting filters. If there is a way to do this then I apologize in advance for opening an issue - if that is the case, please inform / update docs on how to do this. Thank you,
non_process
feature request modify the subject line i checked the docs and tried a lot of things but i couldn t seem to figure out a way to do this some email providers such as microsoft office will strip replace the from name address with that of the account that s being used to send in these cases it would be better to use the subject line to display the feed name author etc as opposed to trying to modify the from address in addition it would be good to be able to add your own subject items so messages can be tagged for setting up mail sorting filters if there is a way to do this then i apologize in advance for opening an issue if that is the case please inform update docs on how to do this thank you
0
179,092
21,514,008,813
IssuesEvent
2022-04-28 08:13:26
elastic/security-docs
https://api.github.com/repos/elastic/security-docs
closed
[BUG] Authentications tab should not present on Hosts Page.
bug Team:Docs v8.2.0 Team:AWP Team: SecuritySolution
**Description** Authentications tab should not present on Hosts Page. **Actual Result and Guide links** - Hosts Page : https://security-docs_1824.docs-preview.app.elstc.co/guide/en/security/master/session-view.html ![image](https://user-images.githubusercontent.com/91867110/165446604-1a53a020-2554-47a1-af32-4378ca0fdc04.png) **Expected Result** ![image](https://user-images.githubusercontent.com/91867110/165446711-7a2ea90d-ce8f-4468-91f1-ff258d131513.png)
True
[BUG] Authentications tab should not present on Hosts Page. - **Description** Authentications tab should not present on Hosts Page. **Actual Result and Guide links** - Hosts Page : https://security-docs_1824.docs-preview.app.elstc.co/guide/en/security/master/session-view.html ![image](https://user-images.githubusercontent.com/91867110/165446604-1a53a020-2554-47a1-af32-4378ca0fdc04.png) **Expected Result** ![image](https://user-images.githubusercontent.com/91867110/165446711-7a2ea90d-ce8f-4468-91f1-ff258d131513.png)
non_process
authentications tab should not present on hosts page description authentications tab should not present on hosts page actual result and guide links hosts page expected result
0
9,057
12,131,948,116
IssuesEvent
2020-04-23 06:11:59
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
QgsProcessingParameterFolderDestination is not visible as input for other algs in modeler
Bug Processing PyQGIS
**Describe the bug** processing parameter using QgsProcessingParameterFolderDestination is nott available as input in following algorithms used in a modeler. Despite others derived from QgsProcessingDestinationParameter as QgsProcessingParameterFileDestination, this parameter is not available as input. **How to Reproduce** 1. Create a pyqgis alg using: `param = QgsProcessingParameterFolderDestination('OUTPUT', 'Output') self.addParameter(param)` 2. Create a model il graphical modeler that use the als as in 1) 3. Create a new alg having a parameter that have to be initialize as "Algorithm output" 4. Scroll th elist of available Algorithm output => the output of algorithm added in 2 is not available **QGIS and OS versions** !!!The same happen in master https://github.com/qgis/QGIS/commit/8addbde0a5 !!! QGIS version | 3.10.4-A Coruña | QGIS code revision | 961b3eee10 -- | -- | -- | -- Compiled against Qt | 5.9.5 | Running against Qt | 5.9.5 Compiled against GDAL/OGR | 2.2.3 | Running against GDAL/OGR | 2.2.3 Compiled against GEOS | 3.7.1-CAPI-1.11.1 | Running against GEOS | 3.7.1-CAPI-1.11.1 27a5e771 Compiled against SQLite | 3.22.0 | Running against SQLite | 3.22.0 PostgreSQL Client Version | 12.2 (Ubuntu 12.2-2.pgdg18.04+1) | SpatiaLite Version | 4.3.0a QWT Version | 6.1.3 | QScintilla2 Version | 2.10.2 PROJ.4 Version | 493 OS Version | Ubuntu 18.04.4 LTS | This copy of QGIS writes debugging output. Active python plugins | plugin_reloader; IPyConsole; pluginbuilder3; copernicushub; firstaid; epyris; db_manager; processing; MetaSearch
1.0
QgsProcessingParameterFolderDestination is not visible as input for other algs in modeler - **Describe the bug** processing parameter using QgsProcessingParameterFolderDestination is nott available as input in following algorithms used in a modeler. Despite others derived from QgsProcessingDestinationParameter as QgsProcessingParameterFileDestination, this parameter is not available as input. **How to Reproduce** 1. Create a pyqgis alg using: `param = QgsProcessingParameterFolderDestination('OUTPUT', 'Output') self.addParameter(param)` 2. Create a model il graphical modeler that use the als as in 1) 3. Create a new alg having a parameter that have to be initialize as "Algorithm output" 4. Scroll th elist of available Algorithm output => the output of algorithm added in 2 is not available **QGIS and OS versions** !!!The same happen in master https://github.com/qgis/QGIS/commit/8addbde0a5 !!! QGIS version | 3.10.4-A Coruña | QGIS code revision | 961b3eee10 -- | -- | -- | -- Compiled against Qt | 5.9.5 | Running against Qt | 5.9.5 Compiled against GDAL/OGR | 2.2.3 | Running against GDAL/OGR | 2.2.3 Compiled against GEOS | 3.7.1-CAPI-1.11.1 | Running against GEOS | 3.7.1-CAPI-1.11.1 27a5e771 Compiled against SQLite | 3.22.0 | Running against SQLite | 3.22.0 PostgreSQL Client Version | 12.2 (Ubuntu 12.2-2.pgdg18.04+1) | SpatiaLite Version | 4.3.0a QWT Version | 6.1.3 | QScintilla2 Version | 2.10.2 PROJ.4 Version | 493 OS Version | Ubuntu 18.04.4 LTS | This copy of QGIS writes debugging output. Active python plugins | plugin_reloader; IPyConsole; pluginbuilder3; copernicushub; firstaid; epyris; db_manager; processing; MetaSearch
process
qgsprocessingparameterfolderdestination is not visible as input for other algs in modeler describe the bug processing parameter using qgsprocessingparameterfolderdestination is nott available as input in following algorithms used in a modeler despite others derived from qgsprocessingdestinationparameter as qgsprocessingparameterfiledestination this parameter is not available as input how to reproduce create a pyqgis alg using param qgsprocessingparameterfolderdestination output output self addparameter param create a model il graphical modeler that use the als as in create a new alg having a parameter that have to be initialize as algorithm output scroll th elist of available algorithm output the output of algorithm added in is not available qgis and os versions the same happen in master qgis version a coruña qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version ubuntu spatialite version qwt version version proj version os version ubuntu lts this copy of qgis writes debugging output active python plugins plugin reloader ipyconsole copernicushub firstaid epyris db manager processing metasearch
1
40,789
12,799,736,489
IssuesEvent
2020-07-02 15:51:45
mwilliams7197/hollow
https://api.github.com/repos/mwilliams7197/hollow
opened
CVE-2019-12384 (Medium) detected in jackson-databind-2.4.3.jar, jackson-databind-2.2.2.jar
security vulnerability
## CVE-2019-12384 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.4.3.jar</b>, <b>jackson-databind-2.2.2.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.4.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Path to vulnerable library: canner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.4.3/feff63199be7b8f495c2f3e2096dcb6bd5e5b0b3/jackson-databind-2.4.3.jar,canner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.4.3/feff63199be7b8f495c2f3e2096dcb6bd5e5b0b3/jackson-databind-2.4.3.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.4.3.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.2.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Path to vulnerable library: /tmp/ws-scm/hollow/hollow-zenoadapter/build.gradle</p> <p> Dependency Hierarchy: - netflix-zeno-2.22.3.jar (Root Library) - :x: **jackson-databind-2.2.2.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://api.github.com/repos/mwilliams7197/hollow/commits/ae866fd0a69bac4d9f321f0b05dafcde9a3d6e69">ae866fd0a69bac4d9f321f0b05dafcde9a3d6e69</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.9.1 might allow attackers to have a variety of impacts by leveraging failure to block the logback-core class from polymorphic deserialization. Depending on the classpath content, remote code execution may be possible. <p>Publish Date: 2019-06-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12384>CVE-2019-12384</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384</a></p> <p>Release Date: 2019-06-24</p> <p>Fix Resolution: 2.9.9.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.4.3","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.4.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.9.1"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.2.2","isTransitiveDependency":true,"dependencyTree":"com.netflix.zeno:netflix-zeno:2.22.3;com.fasterxml.jackson.core:jackson-databind:2.2.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.9.1"}],"vulnerabilityIdentifier":"CVE-2019-12384","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.9.1 might allow attackers to have a variety of impacts by leveraging failure to block the logback-core class from polymorphic deserialization. Depending on the classpath content, remote code execution may be possible.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12384","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-12384 (Medium) detected in jackson-databind-2.4.3.jar, jackson-databind-2.2.2.jar - ## CVE-2019-12384 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.4.3.jar</b>, <b>jackson-databind-2.2.2.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.4.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Path to vulnerable library: canner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.4.3/feff63199be7b8f495c2f3e2096dcb6bd5e5b0b3/jackson-databind-2.4.3.jar,canner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.4.3/feff63199be7b8f495c2f3e2096dcb6bd5e5b0b3/jackson-databind-2.4.3.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.4.3.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.2.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Path to vulnerable library: /tmp/ws-scm/hollow/hollow-zenoadapter/build.gradle</p> <p> Dependency Hierarchy: - netflix-zeno-2.22.3.jar (Root Library) - :x: **jackson-databind-2.2.2.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://api.github.com/repos/mwilliams7197/hollow/commits/ae866fd0a69bac4d9f321f0b05dafcde9a3d6e69">ae866fd0a69bac4d9f321f0b05dafcde9a3d6e69</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.9.1 might allow attackers to have a variety of impacts by leveraging failure to block the logback-core class from polymorphic deserialization. Depending on the classpath content, remote code execution may be possible. <p>Publish Date: 2019-06-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12384>CVE-2019-12384</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384</a></p> <p>Release Date: 2019-06-24</p> <p>Fix Resolution: 2.9.9.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.4.3","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.4.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.9.1"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.2.2","isTransitiveDependency":true,"dependencyTree":"com.netflix.zeno:netflix-zeno:2.22.3;com.fasterxml.jackson.core:jackson-databind:2.2.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.9.1"}],"vulnerabilityIdentifier":"CVE-2019-12384","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.9.1 might allow attackers to have a variety of impacts by leveraging failure to block the logback-core class from polymorphic deserialization. Depending on the classpath content, remote code execution may be possible.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12384","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in jackson databind jar jackson databind jar cve medium severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api path to vulnerable library canner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar canner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to vulnerable library tmp ws scm hollow hollow zenoadapter build gradle dependency hierarchy netflix zeno jar root library x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before might allow attackers to have a variety of impacts by leveraging failure to block the logback core class from polymorphic deserialization depending on the classpath content remote code execution may be possible publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before might allow attackers to have a variety of impacts by leveraging failure to block the logback core class from polymorphic deserialization depending on the classpath content remote code execution may be possible vulnerabilityurl
0
660
3,129,825,780
IssuesEvent
2015-09-09 05:01:43
e-government-ua/i
https://api.github.com/repos/e-government-ua/i
closed
На главном портале для Киева убрать пункты меню
active hi priority In process of testing test
8. В версию возможно В Работе… Убрать разделы "Документи", "Мій журнал", "Про портал" https://docs.google.com/document/d/1fUJlMptp0npeXNShMwZedfcqnhfz6mhyDcw7qoxtZqU/edit#
1.0
На главном портале для Киева убрать пункты меню - 8. В версию возможно В Работе… Убрать разделы "Документи", "Мій журнал", "Про портал" https://docs.google.com/document/d/1fUJlMptp0npeXNShMwZedfcqnhfz6mhyDcw7qoxtZqU/edit#
process
на главном портале для киева убрать пункты меню в версию возможно в работе… убрать разделы документи мій журнал про портал
1
284,510
30,913,639,654
IssuesEvent
2023-08-05 02:28:22
Nivaskumark/kernel_v4.19.72_old
https://api.github.com/repos/Nivaskumark/kernel_v4.19.72_old
reopened
CVE-2022-41222 (High) detected in linux-yoctov5.4.51
Mend: dependency security vulnerability
## CVE-2022-41222 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/kernel_v4.19.72/commit/ce49083a1c14be2d13cb5e878257d293e6c748bc">ce49083a1c14be2d13cb5e878257d293e6c748bc</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/mremap.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/mremap.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> mm/mremap.c in the Linux kernel before 5.13.3 has a use-after-free via a stale TLB because an rmap lock is not held during a PUD move. <p>Publish Date: 2022-09-21 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41222>CVE-2022-41222</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-41222">https://www.linuxkernelcves.com/cves/CVE-2022-41222</a></p> <p>Release Date: 2022-09-21</p> <p>Fix Resolution: v5.4.211,v5.10.137,v5.12.18,v5.13.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-41222 (High) detected in linux-yoctov5.4.51 - ## CVE-2022-41222 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/kernel_v4.19.72/commit/ce49083a1c14be2d13cb5e878257d293e6c748bc">ce49083a1c14be2d13cb5e878257d293e6c748bc</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/mremap.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/mremap.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> mm/mremap.c in the Linux kernel before 5.13.3 has a use-after-free via a stale TLB because an rmap lock is not held during a PUD move. <p>Publish Date: 2022-09-21 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41222>CVE-2022-41222</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-41222">https://www.linuxkernelcves.com/cves/CVE-2022-41222</a></p> <p>Release Date: 2022-09-21</p> <p>Fix Resolution: v5.4.211,v5.10.137,v5.12.18,v5.13.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linux cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files mm mremap c mm mremap c vulnerability details mm mremap c in the linux kernel before has a use after free via a stale tlb because an rmap lock is not held during a pud move publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
2,402
5,193,014,871
IssuesEvent
2017-01-22 15:12:02
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
when field validator message is displayed and clicking New button will cause infinite loop
bug inprocess
Hi, this is an edge case which will cause the browser to hang. # Repro steps: 1) Run the react-bootstrap-table example project locally 2) Navigate to the "Advanced data edit & Insert" option under Examples 3) Scroll to the last example 4) click the job name cell. Remove some characters so validator will display validation msg 5) before the msg disappear, click the "+new" button. # Result browser will hang
1.0
when field validator message is displayed and clicking New button will cause infinite loop - Hi, this is an edge case which will cause the browser to hang. # Repro steps: 1) Run the react-bootstrap-table example project locally 2) Navigate to the "Advanced data edit & Insert" option under Examples 3) Scroll to the last example 4) click the job name cell. Remove some characters so validator will display validation msg 5) before the msg disappear, click the "+new" button. # Result browser will hang
process
when field validator message is displayed and clicking new button will cause infinite loop hi this is an edge case which will cause the browser to hang repro steps run the react bootstrap table example project locally navigate to the advanced data edit insert option under examples scroll to the last example click the job name cell remove some characters so validator will display validation msg before the msg disappear click the new button result browser will hang
1
4,969
7,806,702,570
IssuesEvent
2018-06-11 14:46:15
StackSavingsTeam/stacksavings.com_templates
https://api.github.com/repos/StackSavingsTeam/stacksavings.com_templates
opened
Crear el layout en el repo
On Process
Hay que crea el siguiente layout en este repositorio: http://stacksavings.com/detail-post/the-best-new-products-every-day-gDuXo/en?skipcache Hay que chequear si le falta algo
1.0
Crear el layout en el repo - Hay que crea el siguiente layout en este repositorio: http://stacksavings.com/detail-post/the-best-new-products-every-day-gDuXo/en?skipcache Hay que chequear si le falta algo
process
crear el layout en el repo hay que crea el siguiente layout en este repositorio hay que chequear si le falta algo
1
13,143
15,560,290,183
IssuesEvent
2021-03-16 12:33:21
prisma/language-tools
https://api.github.com/repos/prisma/language-tools
closed
Autocomplete of @@map is not perfect
bug/1-repro-available bug/2-confirmed kind/bug process/candidate topic: autocompletion
## Bug description When I have an autocomplete for @@map, an array is automatically added ## How to reproduce Create a model like this: ```prisma model Persona { uid String @id @default(uuid()) nome String cognome String sesso Sesso data_nascita DateTime } ``` Then start typing `@@ma...` A suggestion of autocompletion for the **@@map** should appear. But if I press **enter**, what I obtain is: ```prisma model Persona { uid String @id @default(uuid()) nome String cognome String sesso Sesso data_nascita DateTime @@map([""]) // array should not be here } ``` It creates an array, but it should not be an array. I always have to delete it manually. ## Expected behavior Instead of ` @@map([""]) ` I should have ` @@map("")` ## Environment & setup - OS: Ubuntu - Editor: VSCode ``` ```
1.0
Autocomplete of @@map is not perfect - ## Bug description When I have an autocomplete for @@map, an array is automatically added ## How to reproduce Create a model like this: ```prisma model Persona { uid String @id @default(uuid()) nome String cognome String sesso Sesso data_nascita DateTime } ``` Then start typing `@@ma...` A suggestion of autocompletion for the **@@map** should appear. But if I press **enter**, what I obtain is: ```prisma model Persona { uid String @id @default(uuid()) nome String cognome String sesso Sesso data_nascita DateTime @@map([""]) // array should not be here } ``` It creates an array, but it should not be an array. I always have to delete it manually. ## Expected behavior Instead of ` @@map([""]) ` I should have ` @@map("")` ## Environment & setup - OS: Ubuntu - Editor: VSCode ``` ```
process
autocomplete of map is not perfect bug description when i have an autocomplete for map an array is automatically added how to reproduce create a model like this prisma model persona uid string id default uuid nome string cognome string sesso sesso data nascita datetime then start typing ma a suggestion of autocompletion for the map should appear but if i press enter what i obtain is prisma model persona uid string id default uuid nome string cognome string sesso sesso data nascita datetime map array should not be here it creates an array but it should not be an array i always have to delete it manually expected behavior instead of map i should have map environment setup os ubuntu editor vscode
1
99,261
16,439,882,365
IssuesEvent
2021-05-20 13:19:31
nextcloud/server
https://api.github.com/repos/nextcloud/server
closed
Lack of SFTP (SSH) Host Key Verification
0. Needs triage bug feature: external storage security
Connections to SFTP external storage were made without verifying the remote host's SSH host key. This behaviour did not appear to be documented anywhere so that users may factor it into their decision to use the External Storage application for remote SFTP resources. This is important since a change of the remote host's SSH host key may indicate that a man-in-the-middle attack is in progress. From the application [logic](https://github.com/nextcloud/server/blob/master/apps/files_external/lib/Lib/Storage/SFTP.php), host key checking would only be performed if a "_ssh_hostKeys_" file exists for the given user. However, this file was not created by default. Furthermore, the method "_writeHostKeys()_" would only update the list of SSH host keys if the user's "_ssh_hostKeys_" file already existed. To conclude, the verification logic does already exist within the current Nextcloud code base but is not used by default. I do recognise that the reason for this is because of the associated functionality that is presently missing such as the user's ability to manage their own "_ssh_hostKeys_" file. ### Steps to reproduce 1. On Nextcloud, create a SFTP external storage as normal 2. Through Nextcloud, access the files stored on the remote SFTP storage 3. On the remote SSH host, change the SSH host key and restart the SSH service 4. Through Nextcloud, once again access the files stored on the remote SFTP storage ### Expected behaviour The Nextcloud instance should alert the user with an error message stating "_Host public key does not match known key_". It should be noted that this message does actually appear if the user's "_ssh_HostKeys_" file exists and contains a host key that differs the remote host's current one (i.e. verification has failed). ### Actual behaviour The Nextcloud instance permits the user to access the remote SFTP storage without any indication that the remote host's SSH host key has been changed. ### Server configuration **Operating system:** Ubuntu GNU/Linux 18.04.1 LTS **Web server:** Apache 2.4.29-1ubuntu4.5 **Database:** MariaDB 10.1.34-0ubuntu0.18.04.1 **PHP version:** 7.2+60ubuntu1 **Nextcloud version:** 15.0.2 **Updated from an older Nextcloud/ownCloud or fresh install:** Fresh **Where did you install Nextcloud from:** Official website archive **Signing status:** <details> <summary>Signing status</summary> No errors have been found. </details> **Are you using external storage, if yes which one:** SFTP **Are you using encryption:** No **Are you using an external user-backend, if yes which one:** No
True
Lack of SFTP (SSH) Host Key Verification - Connections to SFTP external storage were made without verifying the remote host's SSH host key. This behaviour did not appear to be documented anywhere so that users may factor it into their decision to use the External Storage application for remote SFTP resources. This is important since a change of the remote host's SSH host key may indicate that a man-in-the-middle attack is in progress. From the application [logic](https://github.com/nextcloud/server/blob/master/apps/files_external/lib/Lib/Storage/SFTP.php), host key checking would only be performed if a "_ssh_hostKeys_" file exists for the given user. However, this file was not created by default. Furthermore, the method "_writeHostKeys()_" would only update the list of SSH host keys if the user's "_ssh_hostKeys_" file already existed. To conclude, the verification logic does already exist within the current Nextcloud code base but is not used by default. I do recognise that the reason for this is because of the associated functionality that is presently missing such as the user's ability to manage their own "_ssh_hostKeys_" file. ### Steps to reproduce 1. On Nextcloud, create a SFTP external storage as normal 2. Through Nextcloud, access the files stored on the remote SFTP storage 3. On the remote SSH host, change the SSH host key and restart the SSH service 4. Through Nextcloud, once again access the files stored on the remote SFTP storage ### Expected behaviour The Nextcloud instance should alert the user with an error message stating "_Host public key does not match known key_". It should be noted that this message does actually appear if the user's "_ssh_HostKeys_" file exists and contains a host key that differs the remote host's current one (i.e. verification has failed). ### Actual behaviour The Nextcloud instance permits the user to access the remote SFTP storage without any indication that the remote host's SSH host key has been changed. ### Server configuration **Operating system:** Ubuntu GNU/Linux 18.04.1 LTS **Web server:** Apache 2.4.29-1ubuntu4.5 **Database:** MariaDB 10.1.34-0ubuntu0.18.04.1 **PHP version:** 7.2+60ubuntu1 **Nextcloud version:** 15.0.2 **Updated from an older Nextcloud/ownCloud or fresh install:** Fresh **Where did you install Nextcloud from:** Official website archive **Signing status:** <details> <summary>Signing status</summary> No errors have been found. </details> **Are you using external storage, if yes which one:** SFTP **Are you using encryption:** No **Are you using an external user-backend, if yes which one:** No
non_process
lack of sftp ssh host key verification connections to sftp external storage were made without verifying the remote host s ssh host key this behaviour did not appear to be documented anywhere so that users may factor it into their decision to use the external storage application for remote sftp resources this is important since a change of the remote host s ssh host key may indicate that a man in the middle attack is in progress from the application host key checking would only be performed if a ssh hostkeys file exists for the given user however this file was not created by default furthermore the method writehostkeys would only update the list of ssh host keys if the user s ssh hostkeys file already existed to conclude the verification logic does already exist within the current nextcloud code base but is not used by default i do recognise that the reason for this is because of the associated functionality that is presently missing such as the user s ability to manage their own ssh hostkeys file steps to reproduce on nextcloud create a sftp external storage as normal through nextcloud access the files stored on the remote sftp storage on the remote ssh host change the ssh host key and restart the ssh service through nextcloud once again access the files stored on the remote sftp storage expected behaviour the nextcloud instance should alert the user with an error message stating host public key does not match known key it should be noted that this message does actually appear if the user s ssh hostkeys file exists and contains a host key that differs the remote host s current one i e verification has failed actual behaviour the nextcloud instance permits the user to access the remote sftp storage without any indication that the remote host s ssh host key has been changed server configuration operating system ubuntu gnu linux lts web server apache database mariadb php version nextcloud version updated from an older nextcloud owncloud or fresh install fresh where did you install nextcloud from official website archive signing status signing status no errors have been found are you using external storage if yes which one sftp are you using encryption no are you using an external user backend if yes which one no
0
23,513
16,361,431,193
IssuesEvent
2021-05-14 10:04:30
tarantool/tarantool
https://api.github.com/repos/tarantool/tarantool
opened
infra: add Fedora 34 builds
2sp infrastructure qa teamQ
> Fedora Linux 34 was released on 2021-04-27. https://fedoraproject.org/wiki/Releases/34/ChangeSet A Fedora release [usually](https://fedoraproject.org/wiki/End_of_life) maintained ~400 days.
1.0
infra: add Fedora 34 builds - > Fedora Linux 34 was released on 2021-04-27. https://fedoraproject.org/wiki/Releases/34/ChangeSet A Fedora release [usually](https://fedoraproject.org/wiki/End_of_life) maintained ~400 days.
non_process
infra add fedora builds fedora linux was released on a fedora release maintained days
0
2,255
5,088,657,062
IssuesEvent
2017-01-01 00:09:42
sw4j-org/tool-jpa-processor
https://api.github.com/repos/sw4j-org/tool-jpa-processor
opened
Handle @UniqueConstraint Annotation
annotation processor task
Handle the `@UniqueConstraint` annotation for a property or field. See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf) - 11.1.53 UniqueConstraint Annotation
1.0
Handle @UniqueConstraint Annotation - Handle the `@UniqueConstraint` annotation for a property or field. See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf) - 11.1.53 UniqueConstraint Annotation
process
handle uniqueconstraint annotation handle the uniqueconstraint annotation for a property or field see uniqueconstraint annotation
1
25,426
11,171,419,071
IssuesEvent
2019-12-28 19:35:20
eldorplus/conception-website
https://api.github.com/repos/eldorplus/conception-website
opened
CVE-2019-18797 (Medium) detected in node-sass-v4.11.0
security vulnerability
## CVE-2019-18797 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.11.0</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/eldorplus/conception-website/commit/d01989566d2a70e542b08bc943c5de9d223ce39d">d01989566d2a70e542b08bc943c5de9d223ce39d</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (66)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /conception-website/node_modules/node-sass/src/libsass/src/expand.hpp - /conception-website/node_modules/node-sass/src/libsass/src/expand.cpp - /conception-website/node_modules/node-sass/src/sass_types/factory.cpp - /conception-website/node_modules/node-sass/src/libsass/src/operators.cpp - /conception-website/node_modules/node-sass/src/sass_types/boolean.cpp - /conception-website/node_modules/node-sass/src/libsass/src/util.hpp - /conception-website/node_modules/node-sass/src/sass_types/value.h - /conception-website/node_modules/node-sass/src/libsass/src/emitter.hpp - /conception-website/node_modules/node-sass/src/callback_bridge.h - /conception-website/node_modules/node-sass/src/libsass/src/file.cpp - /conception-website/node_modules/node-sass/src/libsass/src/sass.cpp - /conception-website/node_modules/node-sass/src/libsass/src/operation.hpp - /conception-website/node_modules/node-sass/src/libsass/src/operators.hpp - /conception-website/node_modules/node-sass/src/libsass/src/constants.hpp - /conception-website/node_modules/node-sass/src/libsass/src/error_handling.hpp - /conception-website/node_modules/node-sass/src/custom_importer_bridge.cpp - /conception-website/node_modules/node-sass/src/libsass/src/parser.hpp - /conception-website/node_modules/node-sass/src/libsass/src/constants.cpp - /conception-website/node_modules/node-sass/src/sass_types/list.cpp - /conception-website/node_modules/node-sass/src/libsass/src/cssize.cpp - /conception-website/node_modules/node-sass/src/libsass/src/functions.hpp - /conception-website/node_modules/node-sass/src/libsass/src/util.cpp - /conception-website/node_modules/node-sass/src/custom_function_bridge.cpp - /conception-website/node_modules/node-sass/src/custom_importer_bridge.h - /conception-website/node_modules/node-sass/src/libsass/src/bind.cpp - /conception-website/node_modules/node-sass/src/libsass/src/eval.hpp - /conception-website/node_modules/node-sass/src/libsass/src/inspect.cpp - /conception-website/node_modules/node-sass/src/libsass/src/backtrace.cpp - /conception-website/node_modules/node-sass/src/libsass/src/extend.cpp - /conception-website/node_modules/node-sass/src/sass_types/sass_value_wrapper.h - /conception-website/node_modules/node-sass/src/libsass/src/error_handling.cpp - /conception-website/node_modules/node-sass/src/sass_context_wrapper.h - /conception-website/node_modules/node-sass/src/libsass/src/parser.cpp - /conception-website/node_modules/node-sass/src/libsass/src/debugger.hpp - /conception-website/node_modules/node-sass/src/libsass/src/emitter.cpp - /conception-website/node_modules/node-sass/src/sass_types/number.cpp - /conception-website/node_modules/node-sass/src/sass_types/color.h - /conception-website/node_modules/node-sass/src/libsass/src/ast.hpp - /conception-website/node_modules/node-sass/src/libsass/src/sass_values.cpp - /conception-website/node_modules/node-sass/src/libsass/src/output.cpp - /conception-website/node_modules/node-sass/src/libsass/src/check_nesting.cpp - /conception-website/node_modules/node-sass/src/sass_types/null.cpp - /conception-website/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp - /conception-website/node_modules/node-sass/src/libsass/src/cssize.hpp - /conception-website/node_modules/node-sass/src/libsass/src/functions.cpp - /conception-website/node_modules/node-sass/src/libsass/src/prelexer.cpp - /conception-website/node_modules/node-sass/src/libsass/src/ast.cpp - /conception-website/node_modules/node-sass/src/libsass/src/to_c.cpp - /conception-website/node_modules/node-sass/src/libsass/src/to_value.hpp - /conception-website/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp - /conception-website/node_modules/node-sass/src/sass_types/color.cpp - /conception-website/node_modules/node-sass/src/libsass/src/inspect.hpp - /conception-website/node_modules/node-sass/src/libsass/src/values.cpp - /conception-website/node_modules/node-sass/src/sass_context_wrapper.cpp - /conception-website/node_modules/node-sass/src/sass_types/list.h - /conception-website/node_modules/node-sass/src/libsass/src/check_nesting.hpp - /conception-website/node_modules/node-sass/src/libsass/src/to_value.cpp - /conception-website/node_modules/node-sass/src/libsass/src/context.cpp - /conception-website/node_modules/node-sass/src/sass_types/map.cpp - /conception-website/node_modules/node-sass/src/binding.cpp - /conception-website/node_modules/node-sass/src/libsass/src/sass_context.cpp - /conception-website/node_modules/node-sass/src/sass_types/string.cpp - /conception-website/node_modules/node-sass/src/libsass/src/context.hpp - /conception-website/node_modules/node-sass/src/libsass/src/prelexer.hpp - /conception-website/node_modules/node-sass/src/sass_types/boolean.h - /conception-website/node_modules/node-sass/src/libsass/src/eval.cpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp. <p>Publish Date: 2019-11-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797>CVE-2019-18797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797</a></p> <p>Release Date: 2019-11-06</p> <p>Fix Resolution: 3.6.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-18797 (Medium) detected in node-sass-v4.11.0 - ## CVE-2019-18797 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.11.0</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/eldorplus/conception-website/commit/d01989566d2a70e542b08bc943c5de9d223ce39d">d01989566d2a70e542b08bc943c5de9d223ce39d</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (66)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /conception-website/node_modules/node-sass/src/libsass/src/expand.hpp - /conception-website/node_modules/node-sass/src/libsass/src/expand.cpp - /conception-website/node_modules/node-sass/src/sass_types/factory.cpp - /conception-website/node_modules/node-sass/src/libsass/src/operators.cpp - /conception-website/node_modules/node-sass/src/sass_types/boolean.cpp - /conception-website/node_modules/node-sass/src/libsass/src/util.hpp - /conception-website/node_modules/node-sass/src/sass_types/value.h - /conception-website/node_modules/node-sass/src/libsass/src/emitter.hpp - /conception-website/node_modules/node-sass/src/callback_bridge.h - /conception-website/node_modules/node-sass/src/libsass/src/file.cpp - /conception-website/node_modules/node-sass/src/libsass/src/sass.cpp - /conception-website/node_modules/node-sass/src/libsass/src/operation.hpp - /conception-website/node_modules/node-sass/src/libsass/src/operators.hpp - /conception-website/node_modules/node-sass/src/libsass/src/constants.hpp - /conception-website/node_modules/node-sass/src/libsass/src/error_handling.hpp - /conception-website/node_modules/node-sass/src/custom_importer_bridge.cpp - /conception-website/node_modules/node-sass/src/libsass/src/parser.hpp - /conception-website/node_modules/node-sass/src/libsass/src/constants.cpp - /conception-website/node_modules/node-sass/src/sass_types/list.cpp - /conception-website/node_modules/node-sass/src/libsass/src/cssize.cpp - /conception-website/node_modules/node-sass/src/libsass/src/functions.hpp - /conception-website/node_modules/node-sass/src/libsass/src/util.cpp - /conception-website/node_modules/node-sass/src/custom_function_bridge.cpp - /conception-website/node_modules/node-sass/src/custom_importer_bridge.h - /conception-website/node_modules/node-sass/src/libsass/src/bind.cpp - /conception-website/node_modules/node-sass/src/libsass/src/eval.hpp - /conception-website/node_modules/node-sass/src/libsass/src/inspect.cpp - /conception-website/node_modules/node-sass/src/libsass/src/backtrace.cpp - /conception-website/node_modules/node-sass/src/libsass/src/extend.cpp - /conception-website/node_modules/node-sass/src/sass_types/sass_value_wrapper.h - /conception-website/node_modules/node-sass/src/libsass/src/error_handling.cpp - /conception-website/node_modules/node-sass/src/sass_context_wrapper.h - /conception-website/node_modules/node-sass/src/libsass/src/parser.cpp - /conception-website/node_modules/node-sass/src/libsass/src/debugger.hpp - /conception-website/node_modules/node-sass/src/libsass/src/emitter.cpp - /conception-website/node_modules/node-sass/src/sass_types/number.cpp - /conception-website/node_modules/node-sass/src/sass_types/color.h - /conception-website/node_modules/node-sass/src/libsass/src/ast.hpp - /conception-website/node_modules/node-sass/src/libsass/src/sass_values.cpp - /conception-website/node_modules/node-sass/src/libsass/src/output.cpp - /conception-website/node_modules/node-sass/src/libsass/src/check_nesting.cpp - /conception-website/node_modules/node-sass/src/sass_types/null.cpp - /conception-website/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp - /conception-website/node_modules/node-sass/src/libsass/src/cssize.hpp - /conception-website/node_modules/node-sass/src/libsass/src/functions.cpp - /conception-website/node_modules/node-sass/src/libsass/src/prelexer.cpp - /conception-website/node_modules/node-sass/src/libsass/src/ast.cpp - /conception-website/node_modules/node-sass/src/libsass/src/to_c.cpp - /conception-website/node_modules/node-sass/src/libsass/src/to_value.hpp - /conception-website/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp - /conception-website/node_modules/node-sass/src/sass_types/color.cpp - /conception-website/node_modules/node-sass/src/libsass/src/inspect.hpp - /conception-website/node_modules/node-sass/src/libsass/src/values.cpp - /conception-website/node_modules/node-sass/src/sass_context_wrapper.cpp - /conception-website/node_modules/node-sass/src/sass_types/list.h - /conception-website/node_modules/node-sass/src/libsass/src/check_nesting.hpp - /conception-website/node_modules/node-sass/src/libsass/src/to_value.cpp - /conception-website/node_modules/node-sass/src/libsass/src/context.cpp - /conception-website/node_modules/node-sass/src/sass_types/map.cpp - /conception-website/node_modules/node-sass/src/binding.cpp - /conception-website/node_modules/node-sass/src/libsass/src/sass_context.cpp - /conception-website/node_modules/node-sass/src/sass_types/string.cpp - /conception-website/node_modules/node-sass/src/libsass/src/context.hpp - /conception-website/node_modules/node-sass/src/libsass/src/prelexer.hpp - /conception-website/node_modules/node-sass/src/sass_types/boolean.h - /conception-website/node_modules/node-sass/src/libsass/src/eval.cpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> LibSass 3.6.1 has uncontrolled recursion in Sass::Eval::operator()(Sass::Binary_Expression*) in eval.cpp. <p>Publish Date: 2019-11-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-18797>CVE-2019-18797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-18797</a></p> <p>Release Date: 2019-11-06</p> <p>Fix Resolution: 3.6.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in node sass cve medium severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries conception website node modules node sass src libsass src expand hpp conception website node modules node sass src libsass src expand cpp conception website node modules node sass src sass types factory cpp conception website node modules node sass src libsass src operators cpp conception website node modules node sass src sass types boolean cpp conception website node modules node sass src libsass src util hpp conception website node modules node sass src sass types value h conception website node modules node sass src libsass src emitter hpp conception website node modules node sass src callback bridge h conception website node modules node sass src libsass src file cpp conception website node modules node sass src libsass src sass cpp conception website node modules node sass src libsass src operation hpp conception website node modules node sass src libsass src operators hpp conception website node modules node sass src libsass src constants hpp conception website node modules node sass src libsass src error handling hpp conception website node modules node sass src custom importer bridge cpp conception website node modules node sass src libsass src parser hpp conception website node modules node sass src libsass src constants cpp conception website node modules node sass src sass types list cpp conception website node modules node sass src libsass src cssize cpp conception website node modules node sass src libsass src functions hpp conception website node modules node sass src libsass src util cpp conception website node modules node sass src custom function bridge cpp conception website node modules node sass src custom importer bridge h conception website node modules node sass src libsass src bind cpp conception website node modules node sass src libsass src eval hpp conception website node modules node sass src libsass src inspect cpp conception website node modules node sass src libsass src backtrace cpp conception website node modules node sass src libsass src extend cpp conception website node modules node sass src sass types sass value wrapper h conception website node modules node sass src libsass src error handling cpp conception website node modules node sass src sass context wrapper h conception website node modules node sass src libsass src parser cpp conception website node modules node sass src libsass src debugger hpp conception website node modules node sass src libsass src emitter cpp conception website node modules node sass src sass types number cpp conception website node modules node sass src sass types color h conception website node modules node sass src libsass src ast hpp conception website node modules node sass src libsass src sass values cpp conception website node modules node sass src libsass src output cpp conception website node modules node sass src libsass src check nesting cpp conception website node modules node sass src sass types null cpp conception website node modules node sass src libsass src ast def macros hpp conception website node modules node sass src libsass src cssize hpp conception website node modules node sass src libsass src functions cpp conception website node modules node sass src libsass src prelexer cpp conception website node modules node sass src libsass src ast cpp conception website node modules node sass src libsass src to c cpp conception website node modules node sass src libsass src to value hpp conception website node modules node sass src libsass src ast fwd decl hpp conception website node modules node sass src sass types color cpp conception website node modules node sass src libsass src inspect hpp conception website node modules node sass src libsass src values cpp conception website node modules node sass src sass context wrapper cpp conception website node modules node sass src sass types list h conception website node modules node sass src libsass src check nesting hpp conception website node modules node sass src libsass src to value cpp conception website node modules node sass src libsass src context cpp conception website node modules node sass src sass types map cpp conception website node modules node sass src binding cpp conception website node modules node sass src libsass src sass context cpp conception website node modules node sass src sass types string cpp conception website node modules node sass src libsass src context hpp conception website node modules node sass src libsass src prelexer hpp conception website node modules node sass src sass types boolean h conception website node modules node sass src libsass src eval cpp vulnerability details libsass has uncontrolled recursion in sass eval operator sass binary expression in eval cpp publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
17,831
23,769,964,743
IssuesEvent
2022-09-01 15:30:41
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/cumulativetodelta] Convert cumulative Histograms to delta temporality
priority:p2 processor/cumulativetodelta
**Is your feature request related to a problem? Please describe.** It seems that cumulative-to-delta at the moment only processes Sums, and just skips over other data types. This is a problem for us, as we have Histograms in our pipeline with cumulative temporality that we want to convert. **Describe the solution you'd like** Rather than bypass Histogram metrics, have the cumulative-to-delta processor convert their data points in a similar way to sums. **Describe alternatives you've considered** None, other than continuing to ignore non-Sum metrics.
1.0
[processor/cumulativetodelta] Convert cumulative Histograms to delta temporality - **Is your feature request related to a problem? Please describe.** It seems that cumulative-to-delta at the moment only processes Sums, and just skips over other data types. This is a problem for us, as we have Histograms in our pipeline with cumulative temporality that we want to convert. **Describe the solution you'd like** Rather than bypass Histogram metrics, have the cumulative-to-delta processor convert their data points in a similar way to sums. **Describe alternatives you've considered** None, other than continuing to ignore non-Sum metrics.
process
convert cumulative histograms to delta temporality is your feature request related to a problem please describe it seems that cumulative to delta at the moment only processes sums and just skips over other data types this is a problem for us as we have histograms in our pipeline with cumulative temporality that we want to convert describe the solution you d like rather than bypass histogram metrics have the cumulative to delta processor convert their data points in a similar way to sums describe alternatives you ve considered none other than continuing to ignore non sum metrics
1
14,513
5,683,318,116
IssuesEvent
2017-04-13 12:22:46
alicevision/openMVG
https://api.github.com/repos/alicevision/openMVG
closed
[cmake] fix ceres dependency in openMVGConfig
review scope:build type:bug
Currently, when we use openMVG from another repo we need to add `find_package(Ceres)`.
1.0
[cmake] fix ceres dependency in openMVGConfig - Currently, when we use openMVG from another repo we need to add `find_package(Ceres)`.
non_process
fix ceres dependency in openmvgconfig currently when we use openmvg from another repo we need to add find package ceres
0
11,209
13,957,706,427
IssuesEvent
2020-10-24 08:14:50
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
IT - RNDT: IT - National discovery service (RNDT) under technical maintenance in order to switch to the new release
Geoportal Harvesting process IT - Italy
Dear Angelo, I inform you that the RNDT services are under technical maintenance for the time being in order to allow to perform all the activities required for the release of the new version of the discovery service. Consequently, please keep the records harvested in the last session. In the meantime, if you agree, a harvesting test session on the new endpoint of the national discovery service could be performed. If so, I&#39;ll report to you the URL of the new endpoint. In any case, that endpoint will be formally registered by the competent Authority according to the known procedure. Best regards, Antonio
1.0
IT - RNDT: IT - National discovery service (RNDT) under technical maintenance in order to switch to the new release - Dear Angelo, I inform you that the RNDT services are under technical maintenance for the time being in order to allow to perform all the activities required for the release of the new version of the discovery service. Consequently, please keep the records harvested in the last session. In the meantime, if you agree, a harvesting test session on the new endpoint of the national discovery service could be performed. If so, I&#39;ll report to you the URL of the new endpoint. In any case, that endpoint will be formally registered by the competent Authority according to the known procedure. Best regards, Antonio
process
it rndt it national discovery service rndt under technical maintenance in order to switch to the new release dear angelo i inform you that the rndt services are under technical maintenance for the time being in order to allow to perform all the activities required for the release of the new version of the discovery service consequently please keep the records harvested in the last session in the meantime if you agree a harvesting test session on the new endpoint of the national discovery service could be performed if so i ll report to you the url of the new endpoint in any case that endpoint will be formally registered by the competent authority according to the known procedure best regards antonio
1
13,297
15,770,095,062
IssuesEvent
2021-03-31 19:03:41
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Add project editorconfig
process: contributing stage: proposal 💡
Add a project [`.editorconfig`](http://editorconfig.org/) file, to make it easier for contributors. Suggested (`/.editorconfig`): ``` root = true [*] end_of_line = lf insert_final_newline = true indent_style = space indent_size = 2 ``` Not sure if those apply project-wide.
1.0
Add project editorconfig - Add a project [`.editorconfig`](http://editorconfig.org/) file, to make it easier for contributors. Suggested (`/.editorconfig`): ``` root = true [*] end_of_line = lf insert_final_newline = true indent_style = space indent_size = 2 ``` Not sure if those apply project-wide.
process
add project editorconfig add a project file to make it easier for contributors suggested editorconfig root true end of line lf insert final newline true indent style space indent size not sure if those apply project wide
1
1,260
3,792,106,229
IssuesEvent
2016-03-22 08:00:50
nodejs/node
https://api.github.com/repos/nodejs/node
closed
About process.nextTick
c++ process question
It's not a problem or bug. I just view code in node.cc, and could not find where to invoke the `node::MakeCallback` function which actually execute callbacks bound with `process.nextTick`. It looks like those code make sense: ``` if (!env->KickNextTick(&callback_scope)) { return Undefined(env->isolate()); } ``` * **Version**: v4.2.6 * **Platform**: Darwin Kernel Version 14.1.1
1.0
About process.nextTick - It's not a problem or bug. I just view code in node.cc, and could not find where to invoke the `node::MakeCallback` function which actually execute callbacks bound with `process.nextTick`. It looks like those code make sense: ``` if (!env->KickNextTick(&callback_scope)) { return Undefined(env->isolate()); } ``` * **Version**: v4.2.6 * **Platform**: Darwin Kernel Version 14.1.1
process
about process nexttick it s not a problem or bug i just view code in node cc and could not find where to invoke the node makecallback function which actually execute callbacks bound with process nexttick it looks like those code make sense if env kicknexttick callback scope return undefined env isolate version platform darwin kernel version
1
11,642
2,660,012,351
IssuesEvent
2015-03-19 01:34:13
perfsonar/project
https://api.github.com/repos/perfsonar/project
closed
Display input-source name in psGraph.cgi
Milestone-Release3.5 Priority-Medium Type-Defect
Original [issue 1027](https://code.google.com/p/perfsonar-ps/issues/detail?id=1027) created by arlake228 on 2014-12-22T16:54:41.000Z: <b>What steps will reproduce the problem?</b> Create test for a target host (foo.example) using its hostname, but which does not have reverse DNS <b>What is the expected output? What do you see instead?</b> What I has hoping to see in the psGraph.cgi listing is: foo.example (192.0.2.252) What I actually see is: 192.0.2.252 (192.0.2.252) <b>What version of the product are you using? On what operating system?</b> perfsonar PS 3.4.1 <b>Please provide any additional information below.</b> If I look at the top-level JSON info in esmond: curl http://x.x.x.x/esmond/perfsonar/archive/?format=json | python -m json.tool I see: &quot;input-source&quot;: &quot;foo.example&quot;, so it looks like the information is available. (&quot;input-source&quot; is the name, and &quot;input&quot; is the resolved name). If I add an entry in /etc/hosts for the remote host then the name is displayed in the psGraph.cgi output, which suggests that it is indeed doing a reverse lookup to get the name. Suggested approach 1: * if input-source is an IP literal, do a reverse lookup to find the name * Display &quot;input-source (input)&quot; * Ditto for &quot;output-source (output)&quot; This displays the name configured in the test, in preference to what's in reverse DNS. Suggested approach 2: * Do a reverse lookup to find the name * If this fails, use input-source or output-source instead This preserves current behavior as closely as possible, and only differs if reverse lookup is unavailable
1.0
Display input-source name in psGraph.cgi - Original [issue 1027](https://code.google.com/p/perfsonar-ps/issues/detail?id=1027) created by arlake228 on 2014-12-22T16:54:41.000Z: <b>What steps will reproduce the problem?</b> Create test for a target host (foo.example) using its hostname, but which does not have reverse DNS <b>What is the expected output? What do you see instead?</b> What I has hoping to see in the psGraph.cgi listing is: foo.example (192.0.2.252) What I actually see is: 192.0.2.252 (192.0.2.252) <b>What version of the product are you using? On what operating system?</b> perfsonar PS 3.4.1 <b>Please provide any additional information below.</b> If I look at the top-level JSON info in esmond: curl http://x.x.x.x/esmond/perfsonar/archive/?format=json | python -m json.tool I see: &quot;input-source&quot;: &quot;foo.example&quot;, so it looks like the information is available. (&quot;input-source&quot; is the name, and &quot;input&quot; is the resolved name). If I add an entry in /etc/hosts for the remote host then the name is displayed in the psGraph.cgi output, which suggests that it is indeed doing a reverse lookup to get the name. Suggested approach 1: * if input-source is an IP literal, do a reverse lookup to find the name * Display &quot;input-source (input)&quot; * Ditto for &quot;output-source (output)&quot; This displays the name configured in the test, in preference to what's in reverse DNS. Suggested approach 2: * Do a reverse lookup to find the name * If this fails, use input-source or output-source instead This preserves current behavior as closely as possible, and only differs if reverse lookup is unavailable
non_process
display input source name in psgraph cgi original created by on what steps will reproduce the problem create test for a target host foo example using its hostname but which does not have reverse dns what is the expected output what do you see instead what i has hoping to see in the psgraph cgi listing is foo example what i actually see is what version of the product are you using on what operating system perfsonar ps please provide any additional information below if i look at the top level json info in esmond curl python m json tool i see quot input source quot quot foo example quot so it looks like the information is available quot input source quot is the name and quot input quot is the resolved name if i add an entry in etc hosts for the remote host then the name is displayed in the psgraph cgi output which suggests that it is indeed doing a reverse lookup to get the name suggested approach if input source is an ip literal do a reverse lookup to find the name display quot input source input quot ditto for quot output source output quot this displays the name configured in the test in preference to what s in reverse dns suggested approach do a reverse lookup to find the name if this fails use input source or output source instead this preserves current behavior as closely as possible and only differs if reverse lookup is unavailable
0
77,658
7,594,609,378
IssuesEvent
2018-04-27 00:04:38
nuxsmin/sysPass
https://api.github.com/repos/nuxsmin/sysPass
closed
Can't edit or create new accounts on v3.0-beta (300.18042501)
NeedTests v3
When I try to update an account or create new accounts after I updated to the latest release I'm having this error: Notice: Undefined offset: 0 in /var/www/html/syspass/lib/SP/Storage/DbWrapper.php on line 103 Argument 1 passed to SP\DataModel\Dto\AccountDetailsResponse::setUsers() must be of the type array, null given, called in /var/www/html/syspass/lib/SP/Services/Account/AccountService.php on line 105. Here is what I had in syspass.log: 2018-04-26 12:04:22 - SQLSTATE[42S22]: Column not found: 1054 Unknown column 'AU.isEdit' in 'field list' 2018-04-26 12:04:22 - #0 /var/www/html/syspass/lib/SP/Storage/Database.php(186): PDOStatement->execute() #1 /var/www/html/syspass/lib/SP/Storage/Database.php(117): SP\Storage\Database->prepareQueryData(Object(SP\Storage\QueryData)) #2 /var/www/html/syspass/lib/SP/Storage/DbWrapper.php(89): SP\Storage\Database->doQuery(Object(SP\Storage\QueryData)) #3 /var/www/html/syspass/lib/SP/Storage/DbWrapper.php(65): SP\Storage\DbWrapper::getResults(Object(SP\Storage\QueryData), Object(SP\Storage\Database)) #4 /var/www/html/syspass/lib/SP/Repositories/Account/AccountToUserRepository.php(189): SP\Storage\DbWrapper::getResultsArray(Object(SP\Storage\QueryData), Object(SP\Storage\Database)) #5 /var/www/html/syspass/lib/SP/Services/Account/AccountService.php(105): SP\Repositories\Account\AccountToUserRepository->getUsersByAccountId(5) #6 /var/www/html/syspass/app/modules/web/Controllers/AccountController.php(329): SP\Services\Account\AccountService->withUsersById(Object(SP\DataModel\Dto\AccountDetailsResponse)) #7 [internal function]: SP\Modules\Web\Controllers\AccountController->editAction(5) #8 /var/www/html/syspass/lib/SP/Bootstrap.php(239): call_user_func_array(Array, Array) #9 [internal function]: SP\Bootstrap->SP\{closure}(Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array) #10 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(886): call_user_func(Object(Closure), Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array) #11 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(588): Klein\Klein->handleRouteCallback(Object(Klein\Route), Object(Klein\DataCollection\RouteCollection), Array) #12 /var/www/html/syspass/lib/SP/Bootstrap.php(446): Klein\Klein->dispatch() #13 /var/www/html/syspass/lib/Base.php(74): SP\Bootstrap::run(Object(DI\Container)) #14 /var/www/html/syspass/index.php(28): require('/var/www/html/s...') #15 {main} 2018-04-26 12:04:22 - Routing error: Argument 1 passed to SP\DataModel\Dto\AccountDetailsResponse::setUsers() must be of the type array, null given, called in /var/www/html/syspass/lib/SP/Services/Account/AccountService.php on line 105 2018-04-26 12:04:22 - Routing error: #0 /var/www/html/syspass/lib/SP/Services/Account/AccountService.php(105): SP\DataModel\Dto\AccountDetailsResponse->setUsers(NULL) #1 /var/www/html/syspass/app/modules/web/Controllers/AccountController.php(329): SP\Services\Account\AccountService->withUsersById(Object(SP\DataModel\Dto\AccountDetailsResponse)) #2 [internal function]: SP\Modules\Web\Controllers\AccountController->editAction(5) #3 /var/www/html/syspass/lib/SP/Bootstrap.php(239): call_user_func_array(Array, Array) #4 [internal function]: SP\Bootstrap->SP\{closure}(Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array) #5 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(886): call_user_func(Object(Closure), Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array) #6 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(588): Klein\Klein->handleRouteCallback(Object(Klein\Route), Object(Klein\DataCollection\RouteCollection), Array) #7 /var/www/html/syspass/lib/SP/Bootstrap.php(446): Klein\Klein->dispatch() #8 /var/www/html/syspass/lib/Base.php(74): SP\Bootstrap::run(Object(DI\Container)) #9 /var/www/html/syspass/index.php(28): require('/var/www/html/s...')
1.0
Can't edit or create new accounts on v3.0-beta (300.18042501) - When I try to update an account or create new accounts after I updated to the latest release I'm having this error: Notice: Undefined offset: 0 in /var/www/html/syspass/lib/SP/Storage/DbWrapper.php on line 103 Argument 1 passed to SP\DataModel\Dto\AccountDetailsResponse::setUsers() must be of the type array, null given, called in /var/www/html/syspass/lib/SP/Services/Account/AccountService.php on line 105. Here is what I had in syspass.log: 2018-04-26 12:04:22 - SQLSTATE[42S22]: Column not found: 1054 Unknown column 'AU.isEdit' in 'field list' 2018-04-26 12:04:22 - #0 /var/www/html/syspass/lib/SP/Storage/Database.php(186): PDOStatement->execute() #1 /var/www/html/syspass/lib/SP/Storage/Database.php(117): SP\Storage\Database->prepareQueryData(Object(SP\Storage\QueryData)) #2 /var/www/html/syspass/lib/SP/Storage/DbWrapper.php(89): SP\Storage\Database->doQuery(Object(SP\Storage\QueryData)) #3 /var/www/html/syspass/lib/SP/Storage/DbWrapper.php(65): SP\Storage\DbWrapper::getResults(Object(SP\Storage\QueryData), Object(SP\Storage\Database)) #4 /var/www/html/syspass/lib/SP/Repositories/Account/AccountToUserRepository.php(189): SP\Storage\DbWrapper::getResultsArray(Object(SP\Storage\QueryData), Object(SP\Storage\Database)) #5 /var/www/html/syspass/lib/SP/Services/Account/AccountService.php(105): SP\Repositories\Account\AccountToUserRepository->getUsersByAccountId(5) #6 /var/www/html/syspass/app/modules/web/Controllers/AccountController.php(329): SP\Services\Account\AccountService->withUsersById(Object(SP\DataModel\Dto\AccountDetailsResponse)) #7 [internal function]: SP\Modules\Web\Controllers\AccountController->editAction(5) #8 /var/www/html/syspass/lib/SP/Bootstrap.php(239): call_user_func_array(Array, Array) #9 [internal function]: SP\Bootstrap->SP\{closure}(Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array) #10 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(886): call_user_func(Object(Closure), Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array) #11 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(588): Klein\Klein->handleRouteCallback(Object(Klein\Route), Object(Klein\DataCollection\RouteCollection), Array) #12 /var/www/html/syspass/lib/SP/Bootstrap.php(446): Klein\Klein->dispatch() #13 /var/www/html/syspass/lib/Base.php(74): SP\Bootstrap::run(Object(DI\Container)) #14 /var/www/html/syspass/index.php(28): require('/var/www/html/s...') #15 {main} 2018-04-26 12:04:22 - Routing error: Argument 1 passed to SP\DataModel\Dto\AccountDetailsResponse::setUsers() must be of the type array, null given, called in /var/www/html/syspass/lib/SP/Services/Account/AccountService.php on line 105 2018-04-26 12:04:22 - Routing error: #0 /var/www/html/syspass/lib/SP/Services/Account/AccountService.php(105): SP\DataModel\Dto\AccountDetailsResponse->setUsers(NULL) #1 /var/www/html/syspass/app/modules/web/Controllers/AccountController.php(329): SP\Services\Account\AccountService->withUsersById(Object(SP\DataModel\Dto\AccountDetailsResponse)) #2 [internal function]: SP\Modules\Web\Controllers\AccountController->editAction(5) #3 /var/www/html/syspass/lib/SP/Bootstrap.php(239): call_user_func_array(Array, Array) #4 [internal function]: SP\Bootstrap->SP\{closure}(Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array) #5 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(886): call_user_func(Object(Closure), Object(Klein\Request), Object(Klein\Response), Object(Klein\ServiceProvider), Object(Klein\App), Object(Klein\Klein), Object(Klein\DataCollection\RouteCollection), Array) #6 /var/www/html/syspass/vendor/klein/klein/src/Klein/Klein.php(588): Klein\Klein->handleRouteCallback(Object(Klein\Route), Object(Klein\DataCollection\RouteCollection), Array) #7 /var/www/html/syspass/lib/SP/Bootstrap.php(446): Klein\Klein->dispatch() #8 /var/www/html/syspass/lib/Base.php(74): SP\Bootstrap::run(Object(DI\Container)) #9 /var/www/html/syspass/index.php(28): require('/var/www/html/s...')
non_process
can t edit or create new accounts on beta when i try to update an account or create new accounts after i updated to the latest release i m having this error notice undefined offset in var www html syspass lib sp storage dbwrapper php on line argument passed to sp datamodel dto accountdetailsresponse setusers must be of the type array null given called in var www html syspass lib sp services account accountservice php on line here is what i had in syspass log sqlstate column not found unknown column au isedit in field list var www html syspass lib sp storage database php pdostatement execute var www html syspass lib sp storage database php sp storage database preparequerydata object sp storage querydata var www html syspass lib sp storage dbwrapper php sp storage database doquery object sp storage querydata var www html syspass lib sp storage dbwrapper php sp storage dbwrapper getresults object sp storage querydata object sp storage database var www html syspass lib sp repositories account accounttouserrepository php sp storage dbwrapper getresultsarray object sp storage querydata object sp storage database var www html syspass lib sp services account accountservice php sp repositories account accounttouserrepository getusersbyaccountid var www html syspass app modules web controllers accountcontroller php sp services account accountservice withusersbyid object sp datamodel dto accountdetailsresponse sp modules web controllers accountcontroller editaction var www html syspass lib sp bootstrap php call user func array array array sp bootstrap sp closure object klein request object klein response object klein serviceprovider object klein app object klein klein object klein datacollection routecollection array var www html syspass vendor klein klein src klein klein php call user func object closure object klein request object klein response object klein serviceprovider object klein app object klein klein object klein datacollection routecollection array var www html syspass vendor klein klein src klein klein php klein klein handleroutecallback object klein route object klein datacollection routecollection array var www html syspass lib sp bootstrap php klein klein dispatch var www html syspass lib base php sp bootstrap run object di container var www html syspass index php require var www html s main routing error argument passed to sp datamodel dto accountdetailsresponse setusers must be of the type array null given called in var www html syspass lib sp services account accountservice php on line routing error var www html syspass lib sp services account accountservice php sp datamodel dto accountdetailsresponse setusers null var www html syspass app modules web controllers accountcontroller php sp services account accountservice withusersbyid object sp datamodel dto accountdetailsresponse sp modules web controllers accountcontroller editaction var www html syspass lib sp bootstrap php call user func array array array sp bootstrap sp closure object klein request object klein response object klein serviceprovider object klein app object klein klein object klein datacollection routecollection array var www html syspass vendor klein klein src klein klein php call user func object closure object klein request object klein response object klein serviceprovider object klein app object klein klein object klein datacollection routecollection array var www html syspass vendor klein klein src klein klein php klein klein handleroutecallback object klein route object klein datacollection routecollection array var www html syspass lib sp bootstrap php klein klein dispatch var www html syspass lib base php sp bootstrap run object di container var www html syspass index php require var www html s
0
174,907
27,750,569,828
IssuesEvent
2023-03-15 20:22:31
spring-projects/spring-boot
https://api.github.com/repos/spring-projects/spring-boot
opened
Allow fine-grain control of the getters/setters considered in @ConfigurationProperties
type: enhancement status: pending-design-work
See #34309 for background. Currently our `@ConfigurationProperties` only considers getter/setter combinations when generating meta-data. The binder is a little more lenient and can bind to beans that only have setters. It would be helpful if users could indicate that certain setters without getters should be included. It would also be helpful to be able to exclude properties that have getters/setters that can't actually be set in `application.properties`. For example `SimpleMessageListenerContainer` generates meta-data that includes `connection-factory` of type `org.springframework.amqp.rabbit.connection.ConnectionFactory`.
1.0
Allow fine-grain control of the getters/setters considered in @ConfigurationProperties - See #34309 for background. Currently our `@ConfigurationProperties` only considers getter/setter combinations when generating meta-data. The binder is a little more lenient and can bind to beans that only have setters. It would be helpful if users could indicate that certain setters without getters should be included. It would also be helpful to be able to exclude properties that have getters/setters that can't actually be set in `application.properties`. For example `SimpleMessageListenerContainer` generates meta-data that includes `connection-factory` of type `org.springframework.amqp.rabbit.connection.ConnectionFactory`.
non_process
allow fine grain control of the getters setters considered in configurationproperties see for background currently our configurationproperties only considers getter setter combinations when generating meta data the binder is a little more lenient and can bind to beans that only have setters it would be helpful if users could indicate that certain setters without getters should be included it would also be helpful to be able to exclude properties that have getters setters that can t actually be set in application properties for example simplemessagelistenercontainer generates meta data that includes connection factory of type org springframework amqp rabbit connection connectionfactory
0
97,457
28,302,007,218
IssuesEvent
2023-04-10 07:10:18
denoland/deno
https://api.github.com/repos/denoland/deno
closed
Homebrew not released Deno 1.32.3 due a compile failure
build
Just curious why homebrew still does not have the latest 1.32.3, then I found https://github.com/Homebrew/homebrew-core/pull/127307 ``` [v8 0.68.0] common/icudtl.dat [v8 0.68.0] thread 'main' panicked at 'assertion failed: Command::new(gn()).arg(format!(\"--script-executable={}\",\n python())).arg(\"args\").arg(gn_out_dir).arg(\"--list\").status().unwrap().success()', v8/build.rs:218:3 ``` Could anyone help to take a look?
1.0
Homebrew not released Deno 1.32.3 due a compile failure - Just curious why homebrew still does not have the latest 1.32.3, then I found https://github.com/Homebrew/homebrew-core/pull/127307 ``` [v8 0.68.0] common/icudtl.dat [v8 0.68.0] thread 'main' panicked at 'assertion failed: Command::new(gn()).arg(format!(\"--script-executable={}\",\n python())).arg(\"args\").arg(gn_out_dir).arg(\"--list\").status().unwrap().success()', v8/build.rs:218:3 ``` Could anyone help to take a look?
non_process
homebrew not released deno due a compile failure just curious why homebrew still does not have the latest then i found common icudtl dat thread main panicked at assertion failed command new gn arg format script executable n python arg args arg gn out dir arg list status unwrap success build rs could anyone help to take a look
0
203,757
15,385,856,243
IssuesEvent
2021-03-03 07:18:47
pingcap/br
https://api.github.com/repos/pingcap/br
opened
Test testChecksumSuite.TestChecksum fails
component/test type/bug
1. What did you do? If possible, provide a recipe for reproducing the error. ``` [2021-03-03T07:06:36.488Z] FAIL: executor_test.go:52: testChecksumSuite.TestChecksum [2021-03-03T07:06:36.488Z] [2021-03-03T07:06:36.488Z] executor_test.go:53: [2021-03-03T07:06:36.488Z] c.Assert(s.mock.Start(), IsNil) [2021-03-03T07:06:36.488Z] ... value *errors.withStack = listen tcp 0.0.0.0:41602: bind: address already in use ("listen tcp 0.0.0.0:41602: bind: address already in use") [2021-03-03T07:06:36.488Z] [2021-03-03T07:06:36.488Z] OOPS: 0 passed, 1 FAILED ``` https://internal.pingcap.net/idc-jenkins/blue/organizations/jenkins/br_ghpr_unit_and_integration_test/detail/br_ghpr_unit_and_integration_test/5150/pipeline#step-876-log-270
1.0
Test testChecksumSuite.TestChecksum fails - 1. What did you do? If possible, provide a recipe for reproducing the error. ``` [2021-03-03T07:06:36.488Z] FAIL: executor_test.go:52: testChecksumSuite.TestChecksum [2021-03-03T07:06:36.488Z] [2021-03-03T07:06:36.488Z] executor_test.go:53: [2021-03-03T07:06:36.488Z] c.Assert(s.mock.Start(), IsNil) [2021-03-03T07:06:36.488Z] ... value *errors.withStack = listen tcp 0.0.0.0:41602: bind: address already in use ("listen tcp 0.0.0.0:41602: bind: address already in use") [2021-03-03T07:06:36.488Z] [2021-03-03T07:06:36.488Z] OOPS: 0 passed, 1 FAILED ``` https://internal.pingcap.net/idc-jenkins/blue/organizations/jenkins/br_ghpr_unit_and_integration_test/detail/br_ghpr_unit_and_integration_test/5150/pipeline#step-876-log-270
non_process
test testchecksumsuite testchecksum fails what did you do if possible provide a recipe for reproducing the error fail executor test go testchecksumsuite testchecksum executor test go c assert s mock start isnil value errors withstack listen tcp bind address already in use listen tcp bind address already in use oops passed failed
0
9,875
12,886,325,043
IssuesEvent
2020-07-13 09:18:36
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Missing source and CI name information from Zeit Now GitHub integration app
CI: Zeit cli process: dependencies stage: awaiting external fix
When recording tests from Zeit Now GitHub application on a pull request, the dashboard is missing all Git commit information and CI name See https://dashboard.cypress.io/#/projects/6v6qq1/runs/4/specs and the blog post https://www.cypress.io/blog/2018/08/28/Record-Test-Artifacts-From-Any-CI/ <img width="1024" alt="screen shot 2018-09-05 at 9 19 00 am" src="https://user-images.githubusercontent.com/2212006/45095815-ff7db900-b0ec-11e8-96d4-f9db27abb78b.png">
1.0
Missing source and CI name information from Zeit Now GitHub integration app - When recording tests from Zeit Now GitHub application on a pull request, the dashboard is missing all Git commit information and CI name See https://dashboard.cypress.io/#/projects/6v6qq1/runs/4/specs and the blog post https://www.cypress.io/blog/2018/08/28/Record-Test-Artifacts-From-Any-CI/ <img width="1024" alt="screen shot 2018-09-05 at 9 19 00 am" src="https://user-images.githubusercontent.com/2212006/45095815-ff7db900-b0ec-11e8-96d4-f9db27abb78b.png">
process
missing source and ci name information from zeit now github integration app when recording tests from zeit now github application on a pull request the dashboard is missing all git commit information and ci name see and the blog post img width alt screen shot at am src
1
182,070
14,101,302,798
IssuesEvent
2020-11-06 06:32:24
JuliaLang/julia
https://api.github.com/repos/JuliaLang/julia
closed
Provide a way to print the reports of the nested tests
testsystem
Hi guys! I have a package with lots of tests. I am using nested test sets to improve organization, something like: ```julia @testset "Test Set" begin @testset "Test Sub Set" ... end ``` However, if nothing fails, then nothing about the subtests is printed. This is not good, because I cannot be sure if every test was executed. Is it possible to provide an option to `@testset` print the reports of the subsets all the time?
1.0
Provide a way to print the reports of the nested tests - Hi guys! I have a package with lots of tests. I am using nested test sets to improve organization, something like: ```julia @testset "Test Set" begin @testset "Test Sub Set" ... end ``` However, if nothing fails, then nothing about the subtests is printed. This is not good, because I cannot be sure if every test was executed. Is it possible to provide an option to `@testset` print the reports of the subsets all the time?
non_process
provide a way to print the reports of the nested tests hi guys i have a package with lots of tests i am using nested test sets to improve organization something like julia testset test set begin testset test sub set end however if nothing fails then nothing about the subtests is printed this is not good because i cannot be sure if every test was executed is it possible to provide an option to testset print the reports of the subsets all the time
0
2,098
4,932,119,019
IssuesEvent
2016-11-28 12:36:04
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
opened
Typeahead suggestions options are cut short on form with attached to start event, if last widget on form
browser: all bug comp: activiti-processList
1. Import process 2. Start a process N.b. does not occur with the dropdown widget **App to import** [typeaheads.zip](https://github.com/Alfresco/alfresco-ng2-components/files/616240/typeaheads.zip) ![screen shot 2016-11-28 at 12 23 51](https://cloud.githubusercontent.com/assets/13200338/20668236/ad72073a-b565-11e6-811e-454311cdc71e.png)
1.0
Typeahead suggestions options are cut short on form with attached to start event, if last widget on form - 1. Import process 2. Start a process N.b. does not occur with the dropdown widget **App to import** [typeaheads.zip](https://github.com/Alfresco/alfresco-ng2-components/files/616240/typeaheads.zip) ![screen shot 2016-11-28 at 12 23 51](https://cloud.githubusercontent.com/assets/13200338/20668236/ad72073a-b565-11e6-811e-454311cdc71e.png)
process
typeahead suggestions options are cut short on form with attached to start event if last widget on form import process start a process n b does not occur with the dropdown widget app to import
1
197,852
22,607,331,083
IssuesEvent
2022-06-29 14:15:27
elikkatzgit/TestingPOM
https://api.github.com/repos/elikkatzgit/TestingPOM
opened
WS-2021-0153 (High) detected in ejs-2.7.4.tgz
security vulnerability
## WS-2021-0153 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ejs-2.7.4.tgz</b></p></summary> <p>Embedded JavaScript templates</p> <p>Library home page: <a href="https://registry.npmjs.org/ejs/-/ejs-2.7.4.tgz">https://registry.npmjs.org/ejs/-/ejs-2.7.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/ejs/package.json</p> <p> Dependency Hierarchy: - webpack-bundle-analyzer-3.3.2.tgz (Root Library) - :x: **ejs-2.7.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/TestingPOM/commit/349fffeed7cf25f2cf5b8b6a05b5e4367130406e">349fffeed7cf25f2cf5b8b6a05b5e4367130406e</a></p> <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Arbitrary Code Injection vulnerability was found in ejs before 3.1.6. Caused by filename which isn't sanitized for display. <p>Publish Date: 2021-01-22 <p>URL: <a href=https://github.com/mde/ejs/commit/abaee2be937236b1b8da9a1f55096c17dda905fd>WS-2021-0153</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-01-22</p> <p>Fix Resolution (ejs): 3.1.6</p> <p>Direct dependency fix Resolution (webpack-bundle-analyzer): 4.0.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
WS-2021-0153 (High) detected in ejs-2.7.4.tgz - ## WS-2021-0153 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ejs-2.7.4.tgz</b></p></summary> <p>Embedded JavaScript templates</p> <p>Library home page: <a href="https://registry.npmjs.org/ejs/-/ejs-2.7.4.tgz">https://registry.npmjs.org/ejs/-/ejs-2.7.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/ejs/package.json</p> <p> Dependency Hierarchy: - webpack-bundle-analyzer-3.3.2.tgz (Root Library) - :x: **ejs-2.7.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/TestingPOM/commit/349fffeed7cf25f2cf5b8b6a05b5e4367130406e">349fffeed7cf25f2cf5b8b6a05b5e4367130406e</a></p> <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Arbitrary Code Injection vulnerability was found in ejs before 3.1.6. Caused by filename which isn't sanitized for display. <p>Publish Date: 2021-01-22 <p>URL: <a href=https://github.com/mde/ejs/commit/abaee2be937236b1b8da9a1f55096c17dda905fd>WS-2021-0153</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-01-22</p> <p>Fix Resolution (ejs): 3.1.6</p> <p>Direct dependency fix Resolution (webpack-bundle-analyzer): 4.0.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_process
ws high detected in ejs tgz ws high severity vulnerability vulnerable library ejs tgz embedded javascript templates library home page a href path to dependency file package json path to vulnerable library node modules ejs package json dependency hierarchy webpack bundle analyzer tgz root library x ejs tgz vulnerable library found in head commit a href found in base branch dev vulnerability details arbitrary code injection vulnerability was found in ejs before caused by filename which isn t sanitized for display publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution ejs direct dependency fix resolution webpack bundle analyzer rescue worker helmet automatic remediation is available for this issue
0
15,313
19,406,013,845
IssuesEvent
2021-12-20 00:48:45
emily-writes-poems/emily-writes-poems-processing
https://api.github.com/repos/emily-writes-poems/emily-writes-poems-processing
opened
migrate: create new feature
script migration processing
build Javascript/Electron functionality from existing Python script mongo_feature.py insert new feature given poem id, poem title, featured text and whether the feature should be the current feature (optional)
1.0
migrate: create new feature - build Javascript/Electron functionality from existing Python script mongo_feature.py insert new feature given poem id, poem title, featured text and whether the feature should be the current feature (optional)
process
migrate create new feature build javascript electron functionality from existing python script mongo feature py insert new feature given poem id poem title featured text and whether the feature should be the current feature optional
1
18,656
24,581,325,103
IssuesEvent
2022-10-13 15:49:38
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[FHIR] Questionnaire resources >JSON > 'Initial' key not getting displayed when admin adds placeholder text in the study builder
Bug P1 Response datastore Process: Fixed Process: Tested dev
Questionnaire resources > 'Initial' key not getting displayed when admin adds placeholder text in the study builder
2.0
[FHIR] Questionnaire resources >JSON > 'Initial' key not getting displayed when admin adds placeholder text in the study builder - Questionnaire resources > 'Initial' key not getting displayed when admin adds placeholder text in the study builder
process
questionnaire resources json initial key not getting displayed when admin adds placeholder text in the study builder questionnaire resources initial key not getting displayed when admin adds placeholder text in the study builder
1
3,340
6,474,890,092
IssuesEvent
2017-08-17 19:06:52
Azure/azure-event-hubs-java
https://api.github.com/repos/Azure/azure-event-hubs-java
closed
log exceptions in one line
EventProcessorHost
This is an issue filed by customer in an internal thread: Logging stack traces – when the EPH logs error messages, it logs every line of the stack trace separately, Instead of logging everything in one log message. Why? This makes is harder to understand what happened, especially when there are multiple partitions and each partition may log errors in parallel. ![stacktrace](https://user-images.githubusercontent.com/11813428/29423444-99125148-8330-11e7-97f6-88dd03a5ef73.jpg)
1.0
log exceptions in one line - This is an issue filed by customer in an internal thread: Logging stack traces – when the EPH logs error messages, it logs every line of the stack trace separately, Instead of logging everything in one log message. Why? This makes is harder to understand what happened, especially when there are multiple partitions and each partition may log errors in parallel. ![stacktrace](https://user-images.githubusercontent.com/11813428/29423444-99125148-8330-11e7-97f6-88dd03a5ef73.jpg)
process
log exceptions in one line this is an issue filed by customer in an internal thread logging stack traces – when the eph logs error messages it logs every line of the stack trace separately instead of logging everything in one log message why this makes is harder to understand what happened especially when there are multiple partitions and each partition may log errors in parallel
1
515,307
14,959,702,120
IssuesEvent
2021-01-27 03:53:22
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
[Coverity CID :216793] Division or modulo by zero in tests/ztest/error_hook/src/main.c
Coverity bug priority: low
Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/f91e9fba51e5da46ee5c6822f8656713d74a6ecf/tests/ztest/error_hook/src/main.c#L68 Category: Integer handling issues Function: `trigger_fault_divide_zero` Component: Tests CID: [216793](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=216793) Details: ``` 62 static void trigger_fault_divide_zero(void) 63 { 64 int a = 1; 65 int b = 0; 66 67 /* divde zero */ >>> CID 216793: Integer handling issues (DIVIDE_BY_ZERO) >>> In expression "a / b", division by expression "b" which may be zero has undefined behavior. 68 a = a / b; 69 printk("a is %d\n", a); 70 } 71 72 static void trigger_fault_oops(void) 73 { ``` Please fix or provide comments in coverity using the link: https://scan9.coverity.com/reports.htm#v32951/p12996. Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
1.0
[Coverity CID :216793] Division or modulo by zero in tests/ztest/error_hook/src/main.c - Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/f91e9fba51e5da46ee5c6822f8656713d74a6ecf/tests/ztest/error_hook/src/main.c#L68 Category: Integer handling issues Function: `trigger_fault_divide_zero` Component: Tests CID: [216793](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=216793) Details: ``` 62 static void trigger_fault_divide_zero(void) 63 { 64 int a = 1; 65 int b = 0; 66 67 /* divde zero */ >>> CID 216793: Integer handling issues (DIVIDE_BY_ZERO) >>> In expression "a / b", division by expression "b" which may be zero has undefined behavior. 68 a = a / b; 69 printk("a is %d\n", a); 70 } 71 72 static void trigger_fault_oops(void) 73 { ``` Please fix or provide comments in coverity using the link: https://scan9.coverity.com/reports.htm#v32951/p12996. Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
non_process
division or modulo by zero in tests ztest error hook src main c static code scan issues found in file category integer handling issues function trigger fault divide zero component tests cid details static void trigger fault divide zero void int a int b divde zero cid integer handling issues divide by zero in expression a b division by expression b which may be zero has undefined behavior a a b printk a is d n a static void trigger fault oops void please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file
0
7,952
11,137,562,422
IssuesEvent
2019-12-20 19:42:17
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Allow for sorting of education data in Open Opps application
Apply Process Approved Requirements Ready State Dept.
Who: Applicants What: ability to sort education data Why: in order to highlight specific education Acceptance Criteria: - Education data will be pulled from USAJOBS and displayed in the sort order that comes over. Allow for applicants to sort that data in Open Opps. - Add arrow sorting per design system. Sample of USAJOBS education sort arrows: ![image.png](https://images.zenhubusercontent.com/59ee08f1a468affe6df7cd6f/809494bb-ca89-4716-ac6a-0f951059edea)
1.0
Allow for sorting of education data in Open Opps application - Who: Applicants What: ability to sort education data Why: in order to highlight specific education Acceptance Criteria: - Education data will be pulled from USAJOBS and displayed in the sort order that comes over. Allow for applicants to sort that data in Open Opps. - Add arrow sorting per design system. Sample of USAJOBS education sort arrows: ![image.png](https://images.zenhubusercontent.com/59ee08f1a468affe6df7cd6f/809494bb-ca89-4716-ac6a-0f951059edea)
process
allow for sorting of education data in open opps application who applicants what ability to sort education data why in order to highlight specific education acceptance criteria education data will be pulled from usajobs and displayed in the sort order that comes over allow for applicants to sort that data in open opps add arrow sorting per design system sample of usajobs education sort arrows
1
5,159
7,933,331,646
IssuesEvent
2018-07-08 04:14:13
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
Speedup for displaying non-permanent data
libs-etherlib status-inprocess type-enhancement
getTrans -v -t 2283440.0 is very slow. Mostly because it creates the trace array (which is huge) in memory by parsing the result of the curl call. If, instead, the curl call was parsed and displayed directly for traces (which we never store on disc anyway), we could (a) start the display immediately, (b) not have to allocate all that memory just to throw it away, (c) not have to build an array
1.0
Speedup for displaying non-permanent data - getTrans -v -t 2283440.0 is very slow. Mostly because it creates the trace array (which is huge) in memory by parsing the result of the curl call. If, instead, the curl call was parsed and displayed directly for traces (which we never store on disc anyway), we could (a) start the display immediately, (b) not have to allocate all that memory just to throw it away, (c) not have to build an array
process
speedup for displaying non permanent data gettrans v t is very slow mostly because it creates the trace array which is huge in memory by parsing the result of the curl call if instead the curl call was parsed and displayed directly for traces which we never store on disc anyway we could a start the display immediately b not have to allocate all that memory just to throw it away c not have to build an array
1
17,954
23,959,917,599
IssuesEvent
2022-09-12 18:07:08
2i2c-org/infrastructure
https://api.github.com/repos/2i2c-org/infrastructure
closed
Design a secure method to receive secret keys
type: enhancement :label: team-process
### Description Sometimes, we need to receive secret info (like OAuth keys!) from community champions / university IT to finish setting up our infrastructure. This needs to be encrypted both at rest and in transit. ### Value / benefit - We can receive secrets in a secret fashion - We should receive it in a way that allows anyone from the 2i2c team to decrypt it, and isn't tied to anyone in particular ### Implementation details _No response_ ### Tasks to complete - [ ] Choose what kinds of transfer options we want to officially recommend - [ ] Work out a process for each and document it ### Updates _No response_
1.0
Design a secure method to receive secret keys - ### Description Sometimes, we need to receive secret info (like OAuth keys!) from community champions / university IT to finish setting up our infrastructure. This needs to be encrypted both at rest and in transit. ### Value / benefit - We can receive secrets in a secret fashion - We should receive it in a way that allows anyone from the 2i2c team to decrypt it, and isn't tied to anyone in particular ### Implementation details _No response_ ### Tasks to complete - [ ] Choose what kinds of transfer options we want to officially recommend - [ ] Work out a process for each and document it ### Updates _No response_
process
design a secure method to receive secret keys description sometimes we need to receive secret info like oauth keys from community champions university it to finish setting up our infrastructure this needs to be encrypted both at rest and in transit value benefit we can receive secrets in a secret fashion we should receive it in a way that allows anyone from the team to decrypt it and isn t tied to anyone in particular implementation details no response tasks to complete choose what kinds of transfer options we want to officially recommend work out a process for each and document it updates no response
1
17,620
23,436,972,351
IssuesEvent
2022-08-15 11:00:17
Tencent/tdesign-miniprogram
https://api.github.com/repos/Tencent/tdesign-miniprogram
closed
[upload] 上传失败无状态
bug in process
### tdesign-miniprogram 版本 latest ### 重现链接 _No response_ ### 重现步骤 ![企业微信截图_1108f326-b247-4dbb-b405-044ac5752e50](https://user-images.githubusercontent.com/64452728/183861044-de435b4d-ae26-40be-ac39-46e8b5ef59d0.png) 上传失败,无法触发重试 ### 期望结果 _No response_ ### 实际结果 上传失败,无法触发重试 ### 框架版本 _No response_ ### 浏览器版本 _No response_ ### 系统版本 _No response_ ### Node版本 _No response_ ### 补充说明 _No response_
1.0
[upload] 上传失败无状态 - ### tdesign-miniprogram 版本 latest ### 重现链接 _No response_ ### 重现步骤 ![企业微信截图_1108f326-b247-4dbb-b405-044ac5752e50](https://user-images.githubusercontent.com/64452728/183861044-de435b4d-ae26-40be-ac39-46e8b5ef59d0.png) 上传失败,无法触发重试 ### 期望结果 _No response_ ### 实际结果 上传失败,无法触发重试 ### 框架版本 _No response_ ### 浏览器版本 _No response_ ### 系统版本 _No response_ ### Node版本 _No response_ ### 补充说明 _No response_
process
上传失败无状态 tdesign miniprogram 版本 latest 重现链接 no response 重现步骤 上传失败,无法触发重试 期望结果 no response 实际结果 上传失败,无法触发重试 框架版本 no response 浏览器版本 no response 系统版本 no response node版本 no response 补充说明 no response
1
283,106
30,889,582,573
IssuesEvent
2023-08-04 02:56:39
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
reopened
CVE-2021-0448 (Medium) detected in linux-stable-rtv4.1.33
Mend: dependency security vulnerability
## CVE-2021-0448 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_conntrack_netlink.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A security vulnerability was found in Linux Kernel before 4.4.239, 4.9.239, 4.14.201, 4.19.150, 5.4.70, and 5.8.13. Missing range check for l3/l4 protonum in netfilter.c <p>Publish Date: 2020-11-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-0448>CVE-2021-0448</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-0448">https://www.linuxkernelcves.com/cves/CVE-2021-0448</a></p> <p>Release Date: 2020-11-07</p> <p>Fix Resolution: v4.4.239,v4.9.239,v4.14.201,v4.19.150,v5.4.70,v5.8.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-0448 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2021-0448 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_conntrack_netlink.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A security vulnerability was found in Linux Kernel before 4.4.239, 4.9.239, 4.14.201, 4.19.150, 5.4.70, and 5.8.13. Missing range check for l3/l4 protonum in netfilter.c <p>Publish Date: 2020-11-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-0448>CVE-2021-0448</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-0448">https://www.linuxkernelcves.com/cves/CVE-2021-0448</a></p> <p>Release Date: 2020-11-07</p> <p>Fix Resolution: v4.4.239,v4.9.239,v4.14.201,v4.19.150,v5.4.70,v5.8.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net netfilter nf conntrack netlink c vulnerability details a security vulnerability was found in linux kernel before and missing range check for protonum in netfilter c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
168,285
14,144,998,162
IssuesEvent
2020-11-10 17:08:43
andreascalzo99/ProgettoBlackFriendly
https://api.github.com/repos/andreascalzo99/ProgettoBlackFriendly
closed
Aggiunta Diagrammi UML
documentation
Sviluppare **diagramma delle classi con prospettiva software** e **diagramma dei package**.
1.0
Aggiunta Diagrammi UML - Sviluppare **diagramma delle classi con prospettiva software** e **diagramma dei package**.
non_process
aggiunta diagrammi uml sviluppare diagramma delle classi con prospettiva software e diagramma dei package
0
255,728
19,323,729,591
IssuesEvent
2021-12-14 09:10:05
kubernetes-sigs/cluster-api-provider-kubevirt
https://api.github.com/repos/kubernetes-sigs/cluster-api-provider-kubevirt
closed
Typo in README.md
good first issue help wanted kind/documentation
The `README.md` file has a simple typo at `line 48`. A go-through of the whole .md file is advised in order to correct any other typos.
1.0
Typo in README.md - The `README.md` file has a simple typo at `line 48`. A go-through of the whole .md file is advised in order to correct any other typos.
non_process
typo in readme md the readme md file has a simple typo at line a go through of the whole md file is advised in order to correct any other typos
0
14,380
17,401,179,228
IssuesEvent
2021-08-02 19:56:23
googleapis/python-storage
https://api.github.com/repos/googleapis/python-storage
opened
New 'page_size' argument breaks unit tests under Python 3.6
type: process
PR #520 added passing `page_size` to the `google.api_core.page_iterator.HTTPIterator` constructor, which is a feature added only in `google-api-core 1.29.0`. `testing/constraints-3.6.txt` pins `google-cloud-core==1.6.0`, the minimum version specified in `setup.py`, but does not pin `google-api-core`, which results in: ```bash $ git remote -v origin git@github.com:googleapis/python-storage (fetch) origin git@github.com:googleapis/python-storage (push) $ git status On branch master Your branch is up to date with 'origin/master'. nothing to commit, working tree clean $ git log -1 commit 4abb40310eca7ec45afc4bc5e4dfafbe083e74d2 (HEAD -> master, origin/master, origin/HEAD) Author: Tres Seaver <tseaver@palladion.com> Date: Tue Jul 27 13:43:42 2021 -0400 fix: make 'requests.exceptions.ChunkedEncodingError retryable by default (#526) Closes #525. $ .nox/unit-3-6/bin/pip list | grep google google-api-core 1.27.0 google-auth 1.24.0 google-cloud-core 1.6.0 google-cloud-storage 1.41.1 /path/to/python-storage google-crc32c 1.1.2 google-resumable-media 1.3.0 googleapis-common-protos 1.53.0 $ nox -re unit-3.6 nox > Running session unit-3.6 nox > Creating virtual environment (virtualenv) using python3.6 in .nox/unit-3-6 nox > python -m pip install mock pytest pytest-cov -c /home/tseaver/projects/agendaless/Google/src/python-storage/testing/constraints-3.6.txt nox > python -m pip install -e . -c /home/tseaver/projects/agendaless/Google/src/python-storage/testing/constraints-3.6.txt nox > py.test --quiet --cov=google.cloud.storage --cov=google.cloud --cov=tests.unit --cov-append --cov-config=.coveragerc --cov-report= --cov-fail-under=0 tests/unit ..............................................s......................... [ 6%] ........................................................................ [ 13%] ........................................................................ [ 20%] ........................................................................ [ 27%] ........................................................................ [ 34%] ........................................................................ [ 41%] ........................................................................ [ 48%] ........................................................................ [ 55%] ........................................................................ [ 62%] ........................................................................ [ 69%] ........................................................................ [ 76%] .........................FF............................................. [ 83%] ........................................................................ [ 90%] ........................................................................ [ 97%] ..................... [100%]Coverage.py warning: Module google.cloud was previously imported, but not measured (module-not-measured) =================================== FAILURES =================================== __________________ TestClient.test__list_resource_w_defaults ___________________ self = <tests.unit.test_client.TestClient testMethod=test__list_resource_w_defaults> def test__list_resource_w_defaults(self): import functools from google.api_core.page_iterator import HTTPIterator from google.api_core.page_iterator import _do_nothing_page_start project = "PROJECT" path = "/path/to/list/resource" item_to_value = mock.Mock(spec=[]) credentials = _make_credentials() client = self._make_one(project=project, credentials=credentials) connection = client._base_connection = _make_connection() > iterator = client._list_resource(path=path, item_to_value=item_to_value,) tests/unit/test_client.py:485: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.cloud.storage.client.Client object at 0x7efcfb9f3b38> path = '/path/to/list/resource', item_to_value = <Mock id='139625018221344'> page_token = None, max_results = None, extra_params = None page_start = <function _do_nothing_page_start at 0x7efcfc69fae8> page_size = None, timeout = 60 retry = <google.api_core.retry.Retry object at 0x7efcfc756c50> def _list_resource( self, path, item_to_value, page_token=None, max_results=None, extra_params=None, page_start=page_iterator._do_nothing_page_start, page_size=None, timeout=_DEFAULT_TIMEOUT, retry=DEFAULT_RETRY, ): api_request = functools.partial( self._connection.api_request, timeout=timeout, retry=retry ) return page_iterator.HTTPIterator( client=self, api_request=api_request, path=path, item_to_value=item_to_value, page_token=page_token, max_results=max_results, extra_params=extra_params, page_start=page_start, > page_size=page_size, ) E TypeError: __init__() got an unexpected keyword argument 'page_size' google/cloud/storage/client.py:403: TypeError __________________ TestClient.test__list_resource_w_explicit ___________________ self = <tests.unit.test_client.TestClient testMethod=test__list_resource_w_explicit> def test__list_resource_w_explicit(self): import functools from google.api_core.page_iterator import HTTPIterator project = "PROJECT" path = "/path/to/list/resource" item_to_value = mock.Mock(spec=[]) page_token = "PAGE-TOKEN" max_results = 47 extra_params = {"foo": "Foo"} page_start = mock.Mock(spec=[]) credentials = _make_credentials() client = self._make_one(project=project, credentials=credentials) connection = client._base_connection = _make_connection() iterator = client._list_resource( path=path, item_to_value=item_to_value, page_token=page_token, max_results=max_results, extra_params=extra_params, > page_start=page_start, ) tests/unit/test_client.py:523: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.cloud.storage.client.Client object at 0x7efcfbdfe470> path = '/path/to/list/resource', item_to_value = <Mock id='139625022614160'> page_token = 'PAGE-TOKEN', max_results = 47, extra_params = {'foo': 'Foo'} page_start = <Mock id='139625022612368'>, page_size = None, timeout = 60 retry = <google.api_core.retry.Retry object at 0x7efcfc756c50> def _list_resource( self, path, item_to_value, page_token=None, max_results=None, extra_params=None, page_start=page_iterator._do_nothing_page_start, page_size=None, timeout=_DEFAULT_TIMEOUT, retry=DEFAULT_RETRY, ): api_request = functools.partial( self._connection.api_request, timeout=timeout, retry=retry ) return page_iterator.HTTPIterator( client=self, api_request=api_request, path=path, item_to_value=item_to_value, page_token=page_token, max_results=max_results, extra_params=extra_params, page_start=page_start, > page_size=page_size, ) E TypeError: __init__() got an unexpected keyword argument 'page_size' google/cloud/storage/client.py:403: TypeError =========================== short test summary info ============================ FAILED tests/unit/test_client.py::TestClient::test__list_resource_w_defaults FAILED tests/unit/test_client.py::TestClient::test__list_resource_w_explicit 2 failed, 1026 passed, 1 skipped in 19.16s nox > Command py.test --quiet --cov=google.cloud.storage --cov=google.cloud --cov=tests.unit --cov-append --cov-config=.coveragerc --cov-report= --cov-fail-under=0 tests/unit failed with exit code 1 nox > Session unit-3.6 failed. ``` The fix would be to add ranges for `google-api-core` and `google-auth` (another transitive dependency) to `setup.py`, and then pin the minimum versions for those ranges in `testing/constraints-3.6.txt`.
1.0
New 'page_size' argument breaks unit tests under Python 3.6 - PR #520 added passing `page_size` to the `google.api_core.page_iterator.HTTPIterator` constructor, which is a feature added only in `google-api-core 1.29.0`. `testing/constraints-3.6.txt` pins `google-cloud-core==1.6.0`, the minimum version specified in `setup.py`, but does not pin `google-api-core`, which results in: ```bash $ git remote -v origin git@github.com:googleapis/python-storage (fetch) origin git@github.com:googleapis/python-storage (push) $ git status On branch master Your branch is up to date with 'origin/master'. nothing to commit, working tree clean $ git log -1 commit 4abb40310eca7ec45afc4bc5e4dfafbe083e74d2 (HEAD -> master, origin/master, origin/HEAD) Author: Tres Seaver <tseaver@palladion.com> Date: Tue Jul 27 13:43:42 2021 -0400 fix: make 'requests.exceptions.ChunkedEncodingError retryable by default (#526) Closes #525. $ .nox/unit-3-6/bin/pip list | grep google google-api-core 1.27.0 google-auth 1.24.0 google-cloud-core 1.6.0 google-cloud-storage 1.41.1 /path/to/python-storage google-crc32c 1.1.2 google-resumable-media 1.3.0 googleapis-common-protos 1.53.0 $ nox -re unit-3.6 nox > Running session unit-3.6 nox > Creating virtual environment (virtualenv) using python3.6 in .nox/unit-3-6 nox > python -m pip install mock pytest pytest-cov -c /home/tseaver/projects/agendaless/Google/src/python-storage/testing/constraints-3.6.txt nox > python -m pip install -e . -c /home/tseaver/projects/agendaless/Google/src/python-storage/testing/constraints-3.6.txt nox > py.test --quiet --cov=google.cloud.storage --cov=google.cloud --cov=tests.unit --cov-append --cov-config=.coveragerc --cov-report= --cov-fail-under=0 tests/unit ..............................................s......................... [ 6%] ........................................................................ [ 13%] ........................................................................ [ 20%] ........................................................................ [ 27%] ........................................................................ [ 34%] ........................................................................ [ 41%] ........................................................................ [ 48%] ........................................................................ [ 55%] ........................................................................ [ 62%] ........................................................................ [ 69%] ........................................................................ [ 76%] .........................FF............................................. [ 83%] ........................................................................ [ 90%] ........................................................................ [ 97%] ..................... [100%]Coverage.py warning: Module google.cloud was previously imported, but not measured (module-not-measured) =================================== FAILURES =================================== __________________ TestClient.test__list_resource_w_defaults ___________________ self = <tests.unit.test_client.TestClient testMethod=test__list_resource_w_defaults> def test__list_resource_w_defaults(self): import functools from google.api_core.page_iterator import HTTPIterator from google.api_core.page_iterator import _do_nothing_page_start project = "PROJECT" path = "/path/to/list/resource" item_to_value = mock.Mock(spec=[]) credentials = _make_credentials() client = self._make_one(project=project, credentials=credentials) connection = client._base_connection = _make_connection() > iterator = client._list_resource(path=path, item_to_value=item_to_value,) tests/unit/test_client.py:485: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.cloud.storage.client.Client object at 0x7efcfb9f3b38> path = '/path/to/list/resource', item_to_value = <Mock id='139625018221344'> page_token = None, max_results = None, extra_params = None page_start = <function _do_nothing_page_start at 0x7efcfc69fae8> page_size = None, timeout = 60 retry = <google.api_core.retry.Retry object at 0x7efcfc756c50> def _list_resource( self, path, item_to_value, page_token=None, max_results=None, extra_params=None, page_start=page_iterator._do_nothing_page_start, page_size=None, timeout=_DEFAULT_TIMEOUT, retry=DEFAULT_RETRY, ): api_request = functools.partial( self._connection.api_request, timeout=timeout, retry=retry ) return page_iterator.HTTPIterator( client=self, api_request=api_request, path=path, item_to_value=item_to_value, page_token=page_token, max_results=max_results, extra_params=extra_params, page_start=page_start, > page_size=page_size, ) E TypeError: __init__() got an unexpected keyword argument 'page_size' google/cloud/storage/client.py:403: TypeError __________________ TestClient.test__list_resource_w_explicit ___________________ self = <tests.unit.test_client.TestClient testMethod=test__list_resource_w_explicit> def test__list_resource_w_explicit(self): import functools from google.api_core.page_iterator import HTTPIterator project = "PROJECT" path = "/path/to/list/resource" item_to_value = mock.Mock(spec=[]) page_token = "PAGE-TOKEN" max_results = 47 extra_params = {"foo": "Foo"} page_start = mock.Mock(spec=[]) credentials = _make_credentials() client = self._make_one(project=project, credentials=credentials) connection = client._base_connection = _make_connection() iterator = client._list_resource( path=path, item_to_value=item_to_value, page_token=page_token, max_results=max_results, extra_params=extra_params, > page_start=page_start, ) tests/unit/test_client.py:523: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.cloud.storage.client.Client object at 0x7efcfbdfe470> path = '/path/to/list/resource', item_to_value = <Mock id='139625022614160'> page_token = 'PAGE-TOKEN', max_results = 47, extra_params = {'foo': 'Foo'} page_start = <Mock id='139625022612368'>, page_size = None, timeout = 60 retry = <google.api_core.retry.Retry object at 0x7efcfc756c50> def _list_resource( self, path, item_to_value, page_token=None, max_results=None, extra_params=None, page_start=page_iterator._do_nothing_page_start, page_size=None, timeout=_DEFAULT_TIMEOUT, retry=DEFAULT_RETRY, ): api_request = functools.partial( self._connection.api_request, timeout=timeout, retry=retry ) return page_iterator.HTTPIterator( client=self, api_request=api_request, path=path, item_to_value=item_to_value, page_token=page_token, max_results=max_results, extra_params=extra_params, page_start=page_start, > page_size=page_size, ) E TypeError: __init__() got an unexpected keyword argument 'page_size' google/cloud/storage/client.py:403: TypeError =========================== short test summary info ============================ FAILED tests/unit/test_client.py::TestClient::test__list_resource_w_defaults FAILED tests/unit/test_client.py::TestClient::test__list_resource_w_explicit 2 failed, 1026 passed, 1 skipped in 19.16s nox > Command py.test --quiet --cov=google.cloud.storage --cov=google.cloud --cov=tests.unit --cov-append --cov-config=.coveragerc --cov-report= --cov-fail-under=0 tests/unit failed with exit code 1 nox > Session unit-3.6 failed. ``` The fix would be to add ranges for `google-api-core` and `google-auth` (another transitive dependency) to `setup.py`, and then pin the minimum versions for those ranges in `testing/constraints-3.6.txt`.
process
new page size argument breaks unit tests under python pr added passing page size to the google api core page iterator httpiterator constructor which is a feature added only in google api core testing constraints txt pins google cloud core the minimum version specified in setup py but does not pin google api core which results in bash git remote v origin git github com googleapis python storage fetch origin git github com googleapis python storage push git status on branch master your branch is up to date with origin master nothing to commit working tree clean git log commit head master origin master origin head author tres seaver date tue jul fix make requests exceptions chunkedencodingerror retryable by default closes nox unit bin pip list grep google google api core google auth google cloud core google cloud storage path to python storage google google resumable media googleapis common protos nox re unit nox running session unit nox creating virtual environment virtualenv using in nox unit nox python m pip install mock pytest pytest cov c home tseaver projects agendaless google src python storage testing constraints txt nox python m pip install e c home tseaver projects agendaless google src python storage testing constraints txt nox py test quiet cov google cloud storage cov google cloud cov tests unit cov append cov config coveragerc cov report cov fail under tests unit s ff coverage py warning module google cloud was previously imported but not measured module not measured failures testclient test list resource w defaults self def test list resource w defaults self import functools from google api core page iterator import httpiterator from google api core page iterator import do nothing page start project project path path to list resource item to value mock mock spec credentials make credentials client self make one project project credentials credentials connection client base connection make connection iterator client list resource path path item to value item to value tests unit test client py self path path to list resource item to value page token none max results none extra params none page start page size none timeout retry def list resource self path item to value page token none max results none extra params none page start page iterator do nothing page start page size none timeout default timeout retry default retry api request functools partial self connection api request timeout timeout retry retry return page iterator httpiterator client self api request api request path path item to value item to value page token page token max results max results extra params extra params page start page start page size page size e typeerror init got an unexpected keyword argument page size google cloud storage client py typeerror testclient test list resource w explicit self def test list resource w explicit self import functools from google api core page iterator import httpiterator project project path path to list resource item to value mock mock spec page token page token max results extra params foo foo page start mock mock spec credentials make credentials client self make one project project credentials credentials connection client base connection make connection iterator client list resource path path item to value item to value page token page token max results max results extra params extra params page start page start tests unit test client py self path path to list resource item to value page token page token max results extra params foo foo page start page size none timeout retry def list resource self path item to value page token none max results none extra params none page start page iterator do nothing page start page size none timeout default timeout retry default retry api request functools partial self connection api request timeout timeout retry retry return page iterator httpiterator client self api request api request path path item to value item to value page token page token max results max results extra params extra params page start page start page size page size e typeerror init got an unexpected keyword argument page size google cloud storage client py typeerror short test summary info failed tests unit test client py testclient test list resource w defaults failed tests unit test client py testclient test list resource w explicit failed passed skipped in nox command py test quiet cov google cloud storage cov google cloud cov tests unit cov append cov config coveragerc cov report cov fail under tests unit failed with exit code nox session unit failed the fix would be to add ranges for google api core and google auth another transitive dependency to setup py and then pin the minimum versions for those ranges in testing constraints txt
1
11,862
14,665,645,043
IssuesEvent
2020-12-29 14:40:53
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
Mobile app (gateway) > Study List > Modifications to filters
Android P1 Process: Enhancement Process: Track 3 iOS
Modify filtering options to not include 'upcoming studies' and 'favorites'. Also, remove the bookmarking action on the study list and overview screens.
2.0
Mobile app (gateway) > Study List > Modifications to filters - Modify filtering options to not include 'upcoming studies' and 'favorites'. Also, remove the bookmarking action on the study list and overview screens.
process
mobile app gateway study list modifications to filters modify filtering options to not include upcoming studies and favorites also remove the bookmarking action on the study list and overview screens
1
469
2,905,797,977
IssuesEvent
2015-06-19 04:05:54
mitchellh/packer
https://api.github.com/repos/mitchellh/packer
closed
Vagrant-Cloud: Retry box upload on failure
bug post-processor/atlas
I currently maintain a Vagrant box for 3 providers. I have set up a workflow where a single command builds the machines and pushes them to Vagrant Cloud (Atlas). Occasionally, one of the boxes fails to upload (for unknown reason) and *it seems* that the whole version is deleted upon this single failure, regardless of whether other boxes are still in transit or not. I then have to create the new version and upload the three boxes manually. **I propose that instead of failing on "first sight of problem", the upload is retried (perhaps a configurable amount of times, maybe 3 times by default).** I tracked down the upload step to [this line](https://github.com/mitchellh/packer/blob/master/post-processor/vagrant-cloud/step_upload.go#L23), but my Go skills are basically non-existent to submit a PR. Also, it seems like it would be better to implement the retrial at the `client.Upload` function level as I suspect that function to be shared on more places, thus having direct benefit to other parts of the program. Thanks for feedback! Let me know if I can provide more details or help anyhow.
1.0
Vagrant-Cloud: Retry box upload on failure - I currently maintain a Vagrant box for 3 providers. I have set up a workflow where a single command builds the machines and pushes them to Vagrant Cloud (Atlas). Occasionally, one of the boxes fails to upload (for unknown reason) and *it seems* that the whole version is deleted upon this single failure, regardless of whether other boxes are still in transit or not. I then have to create the new version and upload the three boxes manually. **I propose that instead of failing on "first sight of problem", the upload is retried (perhaps a configurable amount of times, maybe 3 times by default).** I tracked down the upload step to [this line](https://github.com/mitchellh/packer/blob/master/post-processor/vagrant-cloud/step_upload.go#L23), but my Go skills are basically non-existent to submit a PR. Also, it seems like it would be better to implement the retrial at the `client.Upload` function level as I suspect that function to be shared on more places, thus having direct benefit to other parts of the program. Thanks for feedback! Let me know if I can provide more details or help anyhow.
process
vagrant cloud retry box upload on failure i currently maintain a vagrant box for providers i have set up a workflow where a single command builds the machines and pushes them to vagrant cloud atlas occasionally one of the boxes fails to upload for unknown reason and it seems that the whole version is deleted upon this single failure regardless of whether other boxes are still in transit or not i then have to create the new version and upload the three boxes manually i propose that instead of failing on first sight of problem the upload is retried perhaps a configurable amount of times maybe times by default i tracked down the upload step to but my go skills are basically non existent to submit a pr also it seems like it would be better to implement the retrial at the client upload function level as i suspect that function to be shared on more places thus having direct benefit to other parts of the program thanks for feedback let me know if i can provide more details or help anyhow
1
363
2,797,621,697
IssuesEvent
2015-05-12 14:40:58
DynareTeam/dynare
https://api.github.com/repos/DynareTeam/dynare
closed
Investigate preprocessor crash with external function in model-local variable
bug preprocessor
See ```use_local_variable_for_beta.mod``` at http://www.dynare.org/phpBB3/viewtopic.php?f=1&t=6772
1.0
Investigate preprocessor crash with external function in model-local variable - See ```use_local_variable_for_beta.mod``` at http://www.dynare.org/phpBB3/viewtopic.php?f=1&t=6772
process
investigate preprocessor crash with external function in model local variable see use local variable for beta mod at
1
563,935
16,706,571,443
IssuesEvent
2021-06-09 10:41:17
googleapis/google-api-ruby-client
https://api.github.com/repos/googleapis/google-api-ruby-client
closed
Synthesis failed for gkehub-v1alpha
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate gkehub-v1alpha. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the gkehub-v1alpha API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` 2021-06-08 03:13:21,720 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/google-api-ruby-client 2021-06-08 03:13:22,610 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2021-06-08 03:13:22,612 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2021-06-08 03:13:22,615 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2021-06-08 03:13:22,617 autosynth [DEBUG] > Running: git config push.default simple 2021-06-08 03:13:22,620 autosynth [DEBUG] > Running: git branch -f autosynth-gkehub-v1alpha 2021-06-08 03:13:22,623 autosynth [DEBUG] > Running: git checkout autosynth-gkehub-v1alpha Switched to branch 'autosynth-gkehub-v1alpha' 2021-06-08 03:13:22,837 autosynth [INFO] > Running synthtool 2021-06-08 03:13:22,837 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-gkehub_v1alpha/synth.metadata', 'synth.py', '--'] 2021-06-08 03:13:22,837 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/google-api-ruby-client/gkehub/v1alpha/sponge_log.log 2021-06-08 03:13:22,839 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata generated/google-apis-gkehub_v1alpha/synth.metadata synth.py -- gkehub v1alpha 2021-06-08 03:13:23,034 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py. On branch autosynth-gkehub-v1alpha nothing to commit, working tree clean 2021-06-08 03:13:23,092 synthtool [DEBUG] > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha DEBUG:synthtool:Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha git clean -df bundle install Don't run Bundler as root. Bundler can ask for sudo if it is needed, and installing your bundle as root will break this application for all non-root users on this machine. Fetching source index from https://rubygems.org/ Net::HTTPServiceUnavailable: <?xml version="1.0" encoding="utf-8"?> <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"> <html> <head> <title>503 Service Unavailable</title> </head> <body> <h1>Error 503 Service Unavailable</h1> <p>Service Unavailable</p> <h3>Guru Mediation:</h3> <p>Details: cache-sea4457-SEA 1623147211 1892265244</p> <hr> <p>Varnish cache server</p> </body> </html> chown -R 1000:1000 /workspace/generated 2021-06-08 03:13:31,327 synthtool [ERROR] > Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha: None ERROR:synthtool:Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha: None Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1137, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1062, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 763, in invoke return __callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py", line 41, in <module> shell.run(command, hide_output=False) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-w', '/workspace', '-e', 'USER_GROUP=1000:1000', '--entrypoint', 'script/synth.rb', 'gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth', 'gkehub', 'v1alpha']' returned non-zero exit status 1. 2021-06-08 03:13:31,356 autosynth [ERROR] > Synthesis failed 2021-06-08 03:13:31,357 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 293, in _inner_main ).synthesize(synth_log_path / "sponge_log.log") File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-gkehub_v1alpha/synth.metadata', 'synth.py', '--', 'gkehub', 'v1alpha']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/bdd2ce63-997e-4095-81b3-cd54e5fccbe4/targets/github%2Fsynthtool;config=default/tests;query=google-api-ruby-client;failed=false).
1.0
Synthesis failed for gkehub-v1alpha - Hello! Autosynth couldn't regenerate gkehub-v1alpha. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the gkehub-v1alpha API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` 2021-06-08 03:13:21,720 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/google-api-ruby-client 2021-06-08 03:13:22,610 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2021-06-08 03:13:22,612 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2021-06-08 03:13:22,615 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2021-06-08 03:13:22,617 autosynth [DEBUG] > Running: git config push.default simple 2021-06-08 03:13:22,620 autosynth [DEBUG] > Running: git branch -f autosynth-gkehub-v1alpha 2021-06-08 03:13:22,623 autosynth [DEBUG] > Running: git checkout autosynth-gkehub-v1alpha Switched to branch 'autosynth-gkehub-v1alpha' 2021-06-08 03:13:22,837 autosynth [INFO] > Running synthtool 2021-06-08 03:13:22,837 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-gkehub_v1alpha/synth.metadata', 'synth.py', '--'] 2021-06-08 03:13:22,837 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/google-api-ruby-client/gkehub/v1alpha/sponge_log.log 2021-06-08 03:13:22,839 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata generated/google-apis-gkehub_v1alpha/synth.metadata synth.py -- gkehub v1alpha 2021-06-08 03:13:23,034 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py. On branch autosynth-gkehub-v1alpha nothing to commit, working tree clean 2021-06-08 03:13:23,092 synthtool [DEBUG] > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha DEBUG:synthtool:Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha git clean -df bundle install Don't run Bundler as root. Bundler can ask for sudo if it is needed, and installing your bundle as root will break this application for all non-root users on this machine. Fetching source index from https://rubygems.org/ Net::HTTPServiceUnavailable: <?xml version="1.0" encoding="utf-8"?> <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"> <html> <head> <title>503 Service Unavailable</title> </head> <body> <h1>Error 503 Service Unavailable</h1> <p>Service Unavailable</p> <h3>Guru Mediation:</h3> <p>Details: cache-sea4457-SEA 1623147211 1892265244</p> <hr> <p>Varnish cache server</p> </body> </html> chown -R 1000:1000 /workspace/generated 2021-06-08 03:13:31,327 synthtool [ERROR] > Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha: None ERROR:synthtool:Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gkehub v1alpha: None Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1137, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1062, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 763, in invoke return __callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py", line 41, in <module> shell.run(command, hide_output=False) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-w', '/workspace', '-e', 'USER_GROUP=1000:1000', '--entrypoint', 'script/synth.rb', 'gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth', 'gkehub', 'v1alpha']' returned non-zero exit status 1. 2021-06-08 03:13:31,356 autosynth [ERROR] > Synthesis failed 2021-06-08 03:13:31,357 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 293, in _inner_main ).synthesize(synth_log_path / "sponge_log.log") File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-gkehub_v1alpha/synth.metadata', 'synth.py', '--', 'gkehub', 'v1alpha']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/bdd2ce63-997e-4095-81b3-cd54e5fccbe4/targets/github%2Fsynthtool;config=default/tests;query=google-api-ruby-client;failed=false).
non_process
synthesis failed for gkehub hello autosynth couldn t regenerate gkehub broken heart please investigate and fix this issue within business days while it remains broken this library cannot be updated with changes to the gkehub api and the library grows stale see for trouble shooting tips here s the output from running synth py autosynth logs will be written to tmpfs src logs google api ruby client autosynth running git config global core excludesfile home kbuilder autosynth gitignore autosynth running git config user name yoshi automation autosynth running git config user email yoshi automation google com autosynth running git config push default simple autosynth running git branch f autosynth gkehub autosynth running git checkout autosynth gkehub switched to branch autosynth gkehub autosynth running synthtool autosynth autosynth log file path tmpfs src logs google api ruby client gkehub sponge log log autosynth running tmpfs src github synthtool env bin m synthtool metadata generated google apis gkehub synth metadata synth py gkehub synthtool executing home kbuilder cache synthtool google api ruby client synth py on branch autosynth gkehub nothing to commit working tree clean synthtool running docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth gkehub debug synthtool running docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth gkehub git clean df bundle install don t run bundler as root bundler can ask for sudo if it is needed and installing your bundle as root will break this application for all non root users on this machine fetching source index from net httpserviceunavailable doctype html public dtd xhtml strict en service unavailable error service unavailable service unavailable guru mediation details cache sea varnish cache server chown r workspace generated synthtool failed executing docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth gkehub none error synthtool failed executing docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth gkehub none traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool google api ruby client synth py line in shell run command hide output false file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize synth log path sponge log log file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
17,145
22,692,397,782
IssuesEvent
2022-07-04 23:03:54
sparc4-dev/astropop
https://api.github.com/repos/sparc4-dev/astropop
closed
Imcombine: memory overflow
bug image-processing critical
When combining a large number of images, `imcombine` is using more memory then expected. This is critical and causing errors.
1.0
Imcombine: memory overflow - When combining a large number of images, `imcombine` is using more memory then expected. This is critical and causing errors.
process
imcombine memory overflow when combining a large number of images imcombine is using more memory then expected this is critical and causing errors
1
5,558
2,948,397,394
IssuesEvent
2015-07-06 02:41:21
progrium/dokku
https://api.github.com/repos/progrium/dokku
opened
Document Release Instructions
documentation
Maintainers other than myself should have the ability to make dokku releases. - [ ] Find a way to circulate packagecloud auth tokens. - [ ] Setup a build farm somewhere. DO has been willing to provide an account with credits, maybe something I can look into? - [ ] Add documentation to docs folder.
1.0
Document Release Instructions - Maintainers other than myself should have the ability to make dokku releases. - [ ] Find a way to circulate packagecloud auth tokens. - [ ] Setup a build farm somewhere. DO has been willing to provide an account with credits, maybe something I can look into? - [ ] Add documentation to docs folder.
non_process
document release instructions maintainers other than myself should have the ability to make dokku releases find a way to circulate packagecloud auth tokens setup a build farm somewhere do has been willing to provide an account with credits maybe something i can look into add documentation to docs folder
0
275
2,704,399,535
IssuesEvent
2015-04-07 00:06:36
MultiMC/MultiMC5
https://api.github.com/repos/MultiMC/MultiMC5
closed
Silent Failure with malformed qmod
bug confirmed needs processing quickmods
A malformed qmod causes an error in the log, but no message is shown, so there is no indication of failure
1.0
Silent Failure with malformed qmod - A malformed qmod causes an error in the log, but no message is shown, so there is no indication of failure
process
silent failure with malformed qmod a malformed qmod causes an error in the log but no message is shown so there is no indication of failure
1
5,471
8,337,912,571
IssuesEvent
2018-09-28 12:48:47
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
closed
Do not cache empty resultset - per rule setting
CACHE QUERY PROCESSOR
This is a follow up from #1693. Instead of just having a global variable, we can also have a per-rule setting.
1.0
Do not cache empty resultset - per rule setting - This is a follow up from #1693. Instead of just having a global variable, we can also have a per-rule setting.
process
do not cache empty resultset per rule setting this is a follow up from instead of just having a global variable we can also have a per rule setting
1
20,160
26,713,775,073
IssuesEvent
2023-01-28 08:00:05
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_fs_sharing (__main__.TestMultiprocessing)
module: windows module: multiprocessing triaged skipped
Flaky failures in the last week: https://fburl.com/scuba/opensource_ci_jobs/inmj698k. They only appear to be on windows Platforms: windows, mac, linux, dynamo, rocm cc @peterjc123 @mszhanyi @skyline75489 @nbcsm @VitalyFedyunin
1.0
DISABLED test_fs_sharing (__main__.TestMultiprocessing) - Flaky failures in the last week: https://fburl.com/scuba/opensource_ci_jobs/inmj698k. They only appear to be on windows Platforms: windows, mac, linux, dynamo, rocm cc @peterjc123 @mszhanyi @skyline75489 @nbcsm @VitalyFedyunin
process
disabled test fs sharing main testmultiprocessing flaky failures in the last week they only appear to be on windows platforms windows mac linux dynamo rocm cc mszhanyi nbcsm vitalyfedyunin
1
26,129
6,755,375,993
IssuesEvent
2017-10-24 00:06:40
jascam/CodePlexFoo
https://api.github.com/repos/jascam/CodePlexFoo
closed
Create Example: CSWPFAnimation
bug CodePlexMigrationInitiated impact: Low
Demonstrate animation in WPF #### Migrated CodePlex Work Item Details CodePlex Work Item ID: '2251' Vote count: '2'
1.0
Create Example: CSWPFAnimation - Demonstrate animation in WPF #### Migrated CodePlex Work Item Details CodePlex Work Item ID: '2251' Vote count: '2'
non_process
create example cswpfanimation demonstrate animation in wpf migrated codeplex work item details codeplex work item id vote count
0
10,815
13,609,290,695
IssuesEvent
2020-09-23 04:50:37
googleapis/java-billing
https://api.github.com/repos/googleapis/java-billing
closed
Dependency Dashboard
api: cloudbilling type: process
This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10 --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10 --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any chore deps update dependency com google cloud libraries bom to check this box to trigger a request for renovate to run again on this repository
1
15,162
18,912,589,445
IssuesEvent
2021-11-16 15:27:37
prisma/language-tools
https://api.github.com/repos/prisma/language-tools
opened
Add configuration option to turn of file watcher
kind/feature process/candidate
File watcher is causing problems we can not figure out, so let's build in a switch to turn it off and see if this fixes problems for users.
1.0
Add configuration option to turn of file watcher - File watcher is causing problems we can not figure out, so let's build in a switch to turn it off and see if this fixes problems for users.
process
add configuration option to turn of file watcher file watcher is causing problems we can not figure out so let s build in a switch to turn it off and see if this fixes problems for users
1
8,479
11,643,460,862
IssuesEvent
2020-02-29 13:41:53
SE-Garden/tms-webserver
https://api.github.com/repos/SE-Garden/tms-webserver
closed
DockerイメージをDockerHubへPushできるようにする
kind:アーキ process:検討
## 概要 DockerイメージをDockerHubへPushできるようにする 。 対象のリポジトリは、`SE-Gardenn` 専用のOrganizationとすること。 ## ゴール - DockerイメージをDockerHubにPushして利用できるようにする。 - ローカル環境で動かす際も自分がビルドしたDockerイメージでDockerComposeを利用して立ち上げられるようにする ## 成果物 - docker-compose.yaml - gradle.yaml ## 関連Issue - none
1.0
DockerイメージをDockerHubへPushできるようにする - ## 概要 DockerイメージをDockerHubへPushできるようにする 。 対象のリポジトリは、`SE-Gardenn` 専用のOrganizationとすること。 ## ゴール - DockerイメージをDockerHubにPushして利用できるようにする。 - ローカル環境で動かす際も自分がビルドしたDockerイメージでDockerComposeを利用して立ち上げられるようにする ## 成果物 - docker-compose.yaml - gradle.yaml ## 関連Issue - none
process
dockerイメージをdockerhubへpushできるようにする 概要 dockerイメージをdockerhubへpushできるようにする 。 対象のリポジトリは、 se gardenn 専用のorganizationとすること。 ゴール dockerイメージをdockerhubにpushして利用できるようにする。 ローカル環境で動かす際も自分がビルドしたdockerイメージでdockercomposeを利用して立ち上げられるようにする 成果物 docker compose yaml gradle yaml 関連issue none
1
4,575
7,397,974,802
IssuesEvent
2018-03-19 03:01:11
nokia-wroclaw/innovativeprojects-healthiness-of-data
https://api.github.com/repos/nokia-wroclaw/innovativeprojects-healthiness-of-data
closed
Create a function for unknown kpi names.
backend data processing enhancement help wanted
There are cases when the kpi name extracted from database doesn't match any key in our csv file. We need to create a function that handles these cases: 1. The kpi name is nowhere to be found in the file. No similiar name. These entries will be inserted into a separate database. 2. The kpi name has a similiar name in csv file. Find the newest and best matching one. Example: Extracted from database: sgsn_2012. In our file we can find sgsn_2012a, sgsn_2012b etc.
1.0
Create a function for unknown kpi names. - There are cases when the kpi name extracted from database doesn't match any key in our csv file. We need to create a function that handles these cases: 1. The kpi name is nowhere to be found in the file. No similiar name. These entries will be inserted into a separate database. 2. The kpi name has a similiar name in csv file. Find the newest and best matching one. Example: Extracted from database: sgsn_2012. In our file we can find sgsn_2012a, sgsn_2012b etc.
process
create a function for unknown kpi names there are cases when the kpi name extracted from database doesn t match any key in our csv file we need to create a function that handles these cases the kpi name is nowhere to be found in the file no similiar name these entries will be inserted into a separate database the kpi name has a similiar name in csv file find the newest and best matching one example extracted from database sgsn in our file we can find sgsn sgsn etc
1
559,334
16,556,161,498
IssuesEvent
2021-05-28 14:10:20
sopra-fs21-group-15/server
https://api.github.com/repos/sopra-fs21-group-15/server
closed
OPTIONAL: Make friend-relations symmetric
low priority task
Part of #16 Includes friends-request system --- Time estimate: 6h
1.0
OPTIONAL: Make friend-relations symmetric - Part of #16 Includes friends-request system --- Time estimate: 6h
non_process
optional make friend relations symmetric part of includes friends request system time estimate
0
310,347
9,488,923,171
IssuesEvent
2019-04-22 20:54:50
pulumi/pulumi
https://api.github.com/repos/pulumi/pulumi
closed
Allow changes to property values to be ignored
area/core customer/feedback priority/P1
Currently there is no way to mark that we don't care about changed values of a particular property - similar to Terraform's `ignore_changes` block. Although not required often, there are some use cases which would benefit. For example, enabling DynamoDB AutoScaling alters the read and write capacities at the provider end. Since these are required, and are set to their initial values in the Pulumi program, a refresh will give a diff to reset the values to their initial state.
1.0
Allow changes to property values to be ignored - Currently there is no way to mark that we don't care about changed values of a particular property - similar to Terraform's `ignore_changes` block. Although not required often, there are some use cases which would benefit. For example, enabling DynamoDB AutoScaling alters the read and write capacities at the provider end. Since these are required, and are set to their initial values in the Pulumi program, a refresh will give a diff to reset the values to their initial state.
non_process
allow changes to property values to be ignored currently there is no way to mark that we don t care about changed values of a particular property similar to terraform s ignore changes block although not required often there are some use cases which would benefit for example enabling dynamodb autoscaling alters the read and write capacities at the provider end since these are required and are set to their initial values in the pulumi program a refresh will give a diff to reset the values to their initial state
0
89,386
3,793,372,258
IssuesEvent
2016-03-22 13:44:54
DynamoRIO/dynamorio
https://api.github.com/repos/DynamoRIO/dynamorio
closed
CRASH caused by generated LDR PC from unaligned address
Bug-DRCrash Hotlist-Release OpSys-ARM Priority-High
The ARM architectural manual specifies that the address used by an LDR instruction into PC has to be word aligned: ``` [...] if t == 15 then if address<1:0> == '00' then LoadWritePC(data); else UNPREDICTABLE; [...] ``` However, DynamoRIO seems to sometimes generate loads from unaligned addresses into PC. For example: ``` 0x47e6c318: ldr.w r0, [r10] 0x47e6c31c: ldr r3, [r7, #48] ; 0x30 0x47e6c31e: ldr r3, [r3, #12] 0x47e6c320: cmp r3, #0 0x47e6c322: beq.w 0x47e6c32a 0x47e6c326: b.w 0x47e6c35c => 0x47e6c32a: ldr.w pc, [pc, #14] ; 0x47e6c33a 0x47e6c32e: movw r0, #51396 ; 0xc8c4 0x47e6c332: movt r0, #18411 ; 0x47eb 0x47e6c336: ldr.w pc, [r10, #24] 0x47e6c33a: pop {r0, r3, r5, r7} ``` This is causing a bus error on certain cores, for example APM X-Gene. How to reproduce the crash: run a few Thumb applications via DR on a system using X-Gene cores. Many will crash immediately.
1.0
CRASH caused by generated LDR PC from unaligned address - The ARM architectural manual specifies that the address used by an LDR instruction into PC has to be word aligned: ``` [...] if t == 15 then if address<1:0> == '00' then LoadWritePC(data); else UNPREDICTABLE; [...] ``` However, DynamoRIO seems to sometimes generate loads from unaligned addresses into PC. For example: ``` 0x47e6c318: ldr.w r0, [r10] 0x47e6c31c: ldr r3, [r7, #48] ; 0x30 0x47e6c31e: ldr r3, [r3, #12] 0x47e6c320: cmp r3, #0 0x47e6c322: beq.w 0x47e6c32a 0x47e6c326: b.w 0x47e6c35c => 0x47e6c32a: ldr.w pc, [pc, #14] ; 0x47e6c33a 0x47e6c32e: movw r0, #51396 ; 0xc8c4 0x47e6c332: movt r0, #18411 ; 0x47eb 0x47e6c336: ldr.w pc, [r10, #24] 0x47e6c33a: pop {r0, r3, r5, r7} ``` This is causing a bus error on certain cores, for example APM X-Gene. How to reproduce the crash: run a few Thumb applications via DR on a system using X-Gene cores. Many will crash immediately.
non_process
crash caused by generated ldr pc from unaligned address the arm architectural manual specifies that the address used by an ldr instruction into pc has to be word aligned if t then if address then loadwritepc data else unpredictable however dynamorio seems to sometimes generate loads from unaligned addresses into pc for example ldr w ldr ldr cmp beq w b w ldr w pc movw movt ldr w pc pop this is causing a bus error on certain cores for example apm x gene how to reproduce the crash run a few thumb applications via dr on a system using x gene cores many will crash immediately
0
219,837
16,851,876,054
IssuesEvent
2021-06-20 17:23:42
CFEL-CMI/richmol
https://api.github.com/repos/CFEL-CMI/richmol
closed
Provide user and developer documentation on readthedocs
documentation
Create automatically updated RTD docs for both the develop/ branch as well as the latest release.
1.0
Provide user and developer documentation on readthedocs - Create automatically updated RTD docs for both the develop/ branch as well as the latest release.
non_process
provide user and developer documentation on readthedocs create automatically updated rtd docs for both the develop branch as well as the latest release
0
14,004
16,812,492,838
IssuesEvent
2021-06-17 00:52:42
w3c/webauthn
https://api.github.com/repos/w3c/webauthn
closed
Update README with new Bikeshed build instructions
type:process
Follow-up to #1625. Build instructions in README need to be updated to match.
1.0
Update README with new Bikeshed build instructions - Follow-up to #1625. Build instructions in README need to be updated to match.
process
update readme with new bikeshed build instructions follow up to build instructions in readme need to be updated to match
1
737,734
25,529,307,893
IssuesEvent
2022-11-29 06:50:06
rstudio/gt
https://api.github.com/repos/rstudio/gt
closed
Audit all internal and unsafe uses of `glue::glue()`
Difficulty: [2] Intermediate Effort: [2] Medium Priority: [3] High Type: ٩(͡๏̯͡๏)۶ Refactor
There are several instances where `glue::glue()` is used to generate strings with local variables. The problem is that, by default, the search path for `glue()` includes all environments. There are some steps to make the (convenient) use of `glue()` safer: (1) Use the `base::get()` function as the transformer function (supplying it to the `.transformer` arg). (2) Specify an environment for the `.envir` arg; `new.env()` is a pretty good choice, `emptyenv()` is even better if that's possible. (3) Use named arguments in the `glue()` call for replacements, if possible; these will take precedence over vars in the search path. The dev version of **glue** has the `glue_safe()` function. Once that's available in a CRAN release of **glue**, it should be used.
1.0
Audit all internal and unsafe uses of `glue::glue()` - There are several instances where `glue::glue()` is used to generate strings with local variables. The problem is that, by default, the search path for `glue()` includes all environments. There are some steps to make the (convenient) use of `glue()` safer: (1) Use the `base::get()` function as the transformer function (supplying it to the `.transformer` arg). (2) Specify an environment for the `.envir` arg; `new.env()` is a pretty good choice, `emptyenv()` is even better if that's possible. (3) Use named arguments in the `glue()` call for replacements, if possible; these will take precedence over vars in the search path. The dev version of **glue** has the `glue_safe()` function. Once that's available in a CRAN release of **glue**, it should be used.
non_process
audit all internal and unsafe uses of glue glue there are several instances where glue glue is used to generate strings with local variables the problem is that by default the search path for glue includes all environments there are some steps to make the convenient use of glue safer use the base get function as the transformer function supplying it to the transformer arg specify an environment for the envir arg new env is a pretty good choice emptyenv is even better if that s possible use named arguments in the glue call for replacements if possible these will take precedence over vars in the search path the dev version of glue has the glue safe function once that s available in a cran release of glue it should be used
0
9,465
12,450,510,476
IssuesEvent
2020-05-27 08:55:37
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
Empty cubes after volume_statistics for certain datasets
bug preprocessor
**Describe the bug** CMIP6 EC-Earth ocean 3D data has the whole bottom depth layer set to masked values. Both the variables and the volcello. The volume_statistics preproc computes the weighted mean per layer just fine, and for the bottom layer the results are masked as well. But when it comes to computing the weighted mean over the water volume, the presence of masked values sets the results to NaNs and the data comes out of the preprocessing empty: `result.append(np.average(column, weights=depth_volume))` It looks like the function `np.average` is not handling well the presence of layers with only masked values. Would it be correct check and remove this sort of layers before calling `np.average` ? **Please attach** I'm working with EC-Earth local data and can't seem to find Ecearth data with volcello on /badc, sorry.
1.0
Empty cubes after volume_statistics for certain datasets - **Describe the bug** CMIP6 EC-Earth ocean 3D data has the whole bottom depth layer set to masked values. Both the variables and the volcello. The volume_statistics preproc computes the weighted mean per layer just fine, and for the bottom layer the results are masked as well. But when it comes to computing the weighted mean over the water volume, the presence of masked values sets the results to NaNs and the data comes out of the preprocessing empty: `result.append(np.average(column, weights=depth_volume))` It looks like the function `np.average` is not handling well the presence of layers with only masked values. Would it be correct check and remove this sort of layers before calling `np.average` ? **Please attach** I'm working with EC-Earth local data and can't seem to find Ecearth data with volcello on /badc, sorry.
process
empty cubes after volume statistics for certain datasets describe the bug ec earth ocean data has the whole bottom depth layer set to masked values both the variables and the volcello the volume statistics preproc computes the weighted mean per layer just fine and for the bottom layer the results are masked as well but when it comes to computing the weighted mean over the water volume the presence of masked values sets the results to nans and the data comes out of the preprocessing empty result append np average column weights depth volume it looks like the function np average is not handling well the presence of layers with only masked values would it be correct check and remove this sort of layers before calling np average please attach i m working with ec earth local data and can t seem to find ecearth data with volcello on badc sorry
1