Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
20,438
27,099,615,300
IssuesEvent
2023-02-15 07:28:56
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Force loading the Protobuf rules from the rules_proto repo
P3 type: process team-Rules-Server stale
This big breaking change will force users of the Protobuf rules to use the Starlark rules from the [@rules_proto](https://github.com/bazelbuild/rules_proto) repository. The rules there are currently only wrappers around native rules, but they will be rewritten in the future. Design doc: https://docs.google.com/document/d/1Wai_okIK_NnhNJZai5nkTUvZVHdV5pp1YwNLFtWwpC4/edit?ts=5d2c8077
1.0
Force loading the Protobuf rules from the rules_proto repo - This big breaking change will force users of the Protobuf rules to use the Starlark rules from the [@rules_proto](https://github.com/bazelbuild/rules_proto) repository. The rules there are currently only wrappers around native rules, but they will be rewritten in the future. Design doc: https://docs.google.com/document/d/1Wai_okIK_NnhNJZai5nkTUvZVHdV5pp1YwNLFtWwpC4/edit?ts=5d2c8077
process
force loading the protobuf rules from the rules proto repo this big breaking change will force users of the protobuf rules to use the starlark rules from the repository the rules there are currently only wrappers around native rules but they will be rewritten in the future design doc
1
333,120
29,509,062,817
IssuesEvent
2023-06-03 17:36:43
opensearch-project/sql
https://api.github.com/repos/opensearch-project/sql
opened
[AUTOCUT] Integration Test failed for sql: 2.8.0 deb distribution
untriaged autocut v2.8.0 integ-test-failure
The integration test failed at distribution level for component sql<br>Version: 2.8.0<br>Distribution: deb<br>Architecture: arm64<br>Platform: linux<br><br>Please check the logs: https://build.ci.opensearch.org/job/integ-test/5089/display/redirect<br><br> * Steps to reproduce: See https://github.com/opensearch-project/opensearch-build/tree/main/src/test_workflow#integration-tests<br>* Access components yml file:<br> - [With security](https://ci.opensearch.org/ci/dbc/integ-test/2.8.0/7935/linux/arm64/deb/test-results/5089/integ-test/sql/with-security/sql.yml) (if applicable)<br> - [Without security](https://ci.opensearch.org/ci/dbc/integ-test/2.8.0/7935/linux/arm64/deb/test-results/5089/integ-test/sql/without-security/sql.yml) (if applicable)<br><br> _Note: All in one test report manifest with all the details coming soon. See https://github.com/opensearch-project/opensearch-build/issues/1274_
1.0
[AUTOCUT] Integration Test failed for sql: 2.8.0 deb distribution - The integration test failed at distribution level for component sql<br>Version: 2.8.0<br>Distribution: deb<br>Architecture: arm64<br>Platform: linux<br><br>Please check the logs: https://build.ci.opensearch.org/job/integ-test/5089/display/redirect<br><br> * Steps to reproduce: See https://github.com/opensearch-project/opensearch-build/tree/main/src/test_workflow#integration-tests<br>* Access components yml file:<br> - [With security](https://ci.opensearch.org/ci/dbc/integ-test/2.8.0/7935/linux/arm64/deb/test-results/5089/integ-test/sql/with-security/sql.yml) (if applicable)<br> - [Without security](https://ci.opensearch.org/ci/dbc/integ-test/2.8.0/7935/linux/arm64/deb/test-results/5089/integ-test/sql/without-security/sql.yml) (if applicable)<br><br> _Note: All in one test report manifest with all the details coming soon. See https://github.com/opensearch-project/opensearch-build/issues/1274_
non_process
integration test failed for sql deb distribution the integration test failed at distribution level for component sql version distribution deb architecture platform linux please check the logs steps to reproduce see access components yml file if applicable if applicable note all in one test report manifest with all the details coming soon see
0
139,935
18,869,694,002
IssuesEvent
2021-11-13 01:08:55
samq-ghdemo/ori-kit
https://api.github.com/repos/samq-ghdemo/ori-kit
closed
WS-2021-0039 (Low) detected in core-9.0.0.tgz - autoclosed
security vulnerability
## WS-2021-0039 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>core-9.0.0.tgz</b></p></summary> <p>Angular - the core framework</p> <p>Library home page: <a href="https://registry.npmjs.org/@angular/core/-/core-9.0.0.tgz">https://registry.npmjs.org/@angular/core/-/core-9.0.0.tgz</a></p> <p>Path to dependency file: ori-kit/frontend/package.json</p> <p>Path to vulnerable library: ori-kit/frontend/node_modules/codelyzer/node_modules/@angular/core/package.json</p> <p> Dependency Hierarchy: - codelyzer-6.0.2.tgz (Root Library) - :x: **core-9.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/ori-kit/commit/ba236fd18ec3e6450d68d675bce1609d2e5d3230">ba236fd18ec3e6450d68d675bce1609d2e5d3230</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-Site Scripting (XSS) vulnerability was found in @angular/core before 11.1.1. HTML doesn't specify any way to escape comment end text inside the comment. <p>Publish Date: 2021-01-26 <p>URL: <a href=https://github.com/angular/angular/commit/97ec6e48493bf9418971436d885470a66e71f045>WS-2021-0039</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: High - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/angular/angular/releases/tag/11.1.1">https://github.com/angular/angular/releases/tag/11.1.1</a></p> <p>Release Date: 2021-01-26</p> <p>Fix Resolution: @angular/core - 11.1.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"@angular/core","packageVersion":"9.0.0","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"codelyzer:6.0.2;@angular/core:9.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"@angular/core - 11.1.1"}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2021-0039","vulnerabilityDetails":"Cross-Site Scripting (XSS) vulnerability was found in @angular/core before 11.1.1. HTML doesn\u0027t specify any way to escape comment end text inside the comment.","vulnerabilityUrl":"https://github.com/angular/angular/commit/97ec6e48493bf9418971436d885470a66e71f045","cvss3Severity":"low","cvss3Score":"3.9","cvss3Metrics":{"A":"Low","AC":"High","PR":"High","S":"Unchanged","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
WS-2021-0039 (Low) detected in core-9.0.0.tgz - autoclosed - ## WS-2021-0039 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>core-9.0.0.tgz</b></p></summary> <p>Angular - the core framework</p> <p>Library home page: <a href="https://registry.npmjs.org/@angular/core/-/core-9.0.0.tgz">https://registry.npmjs.org/@angular/core/-/core-9.0.0.tgz</a></p> <p>Path to dependency file: ori-kit/frontend/package.json</p> <p>Path to vulnerable library: ori-kit/frontend/node_modules/codelyzer/node_modules/@angular/core/package.json</p> <p> Dependency Hierarchy: - codelyzer-6.0.2.tgz (Root Library) - :x: **core-9.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/ori-kit/commit/ba236fd18ec3e6450d68d675bce1609d2e5d3230">ba236fd18ec3e6450d68d675bce1609d2e5d3230</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-Site Scripting (XSS) vulnerability was found in @angular/core before 11.1.1. HTML doesn't specify any way to escape comment end text inside the comment. <p>Publish Date: 2021-01-26 <p>URL: <a href=https://github.com/angular/angular/commit/97ec6e48493bf9418971436d885470a66e71f045>WS-2021-0039</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: High - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/angular/angular/releases/tag/11.1.1">https://github.com/angular/angular/releases/tag/11.1.1</a></p> <p>Release Date: 2021-01-26</p> <p>Fix Resolution: @angular/core - 11.1.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"@angular/core","packageVersion":"9.0.0","packageFilePaths":["/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"codelyzer:6.0.2;@angular/core:9.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"@angular/core - 11.1.1"}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2021-0039","vulnerabilityDetails":"Cross-Site Scripting (XSS) vulnerability was found in @angular/core before 11.1.1. HTML doesn\u0027t specify any way to escape comment end text inside the comment.","vulnerabilityUrl":"https://github.com/angular/angular/commit/97ec6e48493bf9418971436d885470a66e71f045","cvss3Severity":"low","cvss3Score":"3.9","cvss3Metrics":{"A":"Low","AC":"High","PR":"High","S":"Unchanged","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
ws low detected in core tgz autoclosed ws low severity vulnerability vulnerable library core tgz angular the core framework library home page a href path to dependency file ori kit frontend package json path to vulnerable library ori kit frontend node modules codelyzer node modules angular core package json dependency hierarchy codelyzer tgz root library x core tgz vulnerable library found in head commit a href found in base branch main vulnerability details cross site scripting xss vulnerability was found in angular core before html doesn t specify any way to escape comment end text inside the comment publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction required scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution angular core isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree codelyzer angular core isminimumfixversionavailable true minimumfixversion angular core basebranches vulnerabilityidentifier ws vulnerabilitydetails cross site scripting xss vulnerability was found in angular core before html doesn specify any way to escape comment end text inside the comment vulnerabilityurl
0
93,384
26,935,461,649
IssuesEvent
2023-02-07 20:14:55
openhwgroup/cva6
https://api.github.com/repos/openhwgroup/cva6
closed
[BUG] Verilator 4.014 error | followed as per travis-ci
Component:Tool-and-build
### Is there an existing CVA6 bug for this? - [X] I have searched the existing bug issues ### Bug Description I have been trying to compile the simulation files of CVA6 using the ```make verilate``` command. I have set everything using the scripts found in travis-ci-emul.sh But execution of make command throws error as: ``` %Error: /home2/jayaraj/Desktop/cva-6-issue/cva6/corev_apu/src/tech_cells_generic/src/rtl/tc_sram.sv:93: Unsupported or unknown PLI call: $urandom %Error: /home2/jayaraj/Desktop/cva-6-issue/cva6/corev_apu/src/tech_cells_generic/src/rtl/tc_clk.sv:87: syntax error, unexpected TIME NUMBER, expecting TYPE-IDENTIFIER %Error: /home2/jayaraj/Desktop/cva-6-issue/cva6/corev_apu/src/tech_cells_generic/src/rtl/tc_clk.sv:90: syntax error, unexpected output, expecting IDENTIFIER or do or final ``` I think it is a verilator version issue. But I have used the same version mentioned in the script. Operating System: Ubuntu 20.04.4 LTS Kernel: Linux 5.11.0-27-generic Architecture: x86-64
1.0
[BUG] Verilator 4.014 error | followed as per travis-ci - ### Is there an existing CVA6 bug for this? - [X] I have searched the existing bug issues ### Bug Description I have been trying to compile the simulation files of CVA6 using the ```make verilate``` command. I have set everything using the scripts found in travis-ci-emul.sh But execution of make command throws error as: ``` %Error: /home2/jayaraj/Desktop/cva-6-issue/cva6/corev_apu/src/tech_cells_generic/src/rtl/tc_sram.sv:93: Unsupported or unknown PLI call: $urandom %Error: /home2/jayaraj/Desktop/cva-6-issue/cva6/corev_apu/src/tech_cells_generic/src/rtl/tc_clk.sv:87: syntax error, unexpected TIME NUMBER, expecting TYPE-IDENTIFIER %Error: /home2/jayaraj/Desktop/cva-6-issue/cva6/corev_apu/src/tech_cells_generic/src/rtl/tc_clk.sv:90: syntax error, unexpected output, expecting IDENTIFIER or do or final ``` I think it is a verilator version issue. But I have used the same version mentioned in the script. Operating System: Ubuntu 20.04.4 LTS Kernel: Linux 5.11.0-27-generic Architecture: x86-64
non_process
verilator error followed as per travis ci is there an existing bug for this i have searched the existing bug issues bug description i have been trying to compile the simulation files of using the make verilate command i have set everything using the scripts found in travis ci emul sh but execution of make command throws error as error jayaraj desktop cva issue corev apu src tech cells generic src rtl tc sram sv unsupported or unknown pli call urandom error jayaraj desktop cva issue corev apu src tech cells generic src rtl tc clk sv syntax error unexpected time number expecting type identifier error jayaraj desktop cva issue corev apu src tech cells generic src rtl tc clk sv syntax error unexpected output expecting identifier or do or final i think it is a verilator version issue but i have used the same version mentioned in the script operating system ubuntu lts kernel linux generic architecture
0
877
3,342,907,509
IssuesEvent
2015-11-15 01:42:25
technofreaky/woocomerce-quick-donation
https://api.github.com/repos/technofreaky/woocomerce-quick-donation
closed
User can not change donate amount after donate button is clicked
Enhancement Processing
Hi, Here is my site http://taafng.org/. Steps 1. enter an amount and click donation 2. plugin takes you to the checkout page 3. Lets say internet connection breaks or you can just click any other link 4. if you go back to the home page, change the amount and Click on the donation button, it now calls the checkout page but the amount showing is still the previous amount entered Regards Tomi
1.0
User can not change donate amount after donate button is clicked - Hi, Here is my site http://taafng.org/. Steps 1. enter an amount and click donation 2. plugin takes you to the checkout page 3. Lets say internet connection breaks or you can just click any other link 4. if you go back to the home page, change the amount and Click on the donation button, it now calls the checkout page but the amount showing is still the previous amount entered Regards Tomi
process
user can not change donate amount after donate button is clicked hi here is my site steps enter an amount and click donation plugin takes you to the checkout page lets say internet connection breaks or you can just click any other link if you go back to the home page change the amount and click on the donation button it now calls the checkout page but the amount showing is still the previous amount entered regards tomi
1
2,780
5,713,504,674
IssuesEvent
2017-04-19 07:57:46
g8os/grid
https://api.github.com/repos/g8os/grid
closed
Deployement of the Grid controller
process_wontfix state_inprogress type_feature
## Goal: Automate the installation of the grid controller ### Definitions: - grid controller: composed of multiple component: - AYS server - AYS Repo of the grid API - Grid API - dnsmasq server - http server The controller will also run G8OS. ### Deployment scenario: 1. boot controller on G8OS from usb. 2. deploy a container with the [grid flist](https://hub.gig.tech/maxux/grid.flist) 2.1 The deployment is driven from another AYS server that can connect to the controller. 2.2 We send a blueprint with the information required to the controller and it get installed and configured 3. Once controller has AYS running. Continue installation by sending blueprint to the controller. 3.1 start a container that runs the dnsmasq server with tftp for ipxe boot and http server to server the kernel 3.2 configure iPXE to boot node with `append zerotier=netid` 3.3 create bootstrap.g8os service 3.4 boot the node and watch them beeing discover
1.0
Deployement of the Grid controller - ## Goal: Automate the installation of the grid controller ### Definitions: - grid controller: composed of multiple component: - AYS server - AYS Repo of the grid API - Grid API - dnsmasq server - http server The controller will also run G8OS. ### Deployment scenario: 1. boot controller on G8OS from usb. 2. deploy a container with the [grid flist](https://hub.gig.tech/maxux/grid.flist) 2.1 The deployment is driven from another AYS server that can connect to the controller. 2.2 We send a blueprint with the information required to the controller and it get installed and configured 3. Once controller has AYS running. Continue installation by sending blueprint to the controller. 3.1 start a container that runs the dnsmasq server with tftp for ipxe boot and http server to server the kernel 3.2 configure iPXE to boot node with `append zerotier=netid` 3.3 create bootstrap.g8os service 3.4 boot the node and watch them beeing discover
process
deployement of the grid controller goal automate the installation of the grid controller definitions grid controller composed of multiple component ays server ays repo of the grid api grid api dnsmasq server http server the controller will also run deployment scenario boot controller on from usb deploy a container with the the deployment is driven from another ays server that can connect to the controller we send a blueprint with the information required to the controller and it get installed and configured once controller has ays running continue installation by sending blueprint to the controller start a container that runs the dnsmasq server with tftp for ipxe boot and http server to server the kernel configure ipxe to boot node with append zerotier netid create bootstrap service boot the node and watch them beeing discover
1
1,414
3,979,539,425
IssuesEvent
2016-05-06 00:19:09
mozilla/schedule-app-core
https://api.github.com/repos/mozilla/schedule-app-core
closed
Use "start time" to sort timeblocks
data processor script spreadsheet
Currently value of the `order` key in [`timeblock` object](https://github.com/mozilla/schedule-app-core/blob/gh-pages/docs/JSON_FORMAT.md#timeblocks-required) is manually entered on the Spreadsheet (e.g., there's a column called "order"). `timeblock` object has a `start time` key that we can use to sort a list of `timeblocks`. Our [Python script](https://github.com/mozilla/schedule-app-data-processor) should use make use of `start time` to automatically generate the `order` key instead. One fewer column in the Spreadsheet to manage. YAY! **Spreadsheet** - [x] remove "order" column from `* Timeblock Values` sheet **Python script** - [x] generate `order` based on `start time` (https://github.com/mozilla/schedule-app-data-processor/issues/20)
1.0
Use "start time" to sort timeblocks - Currently value of the `order` key in [`timeblock` object](https://github.com/mozilla/schedule-app-core/blob/gh-pages/docs/JSON_FORMAT.md#timeblocks-required) is manually entered on the Spreadsheet (e.g., there's a column called "order"). `timeblock` object has a `start time` key that we can use to sort a list of `timeblocks`. Our [Python script](https://github.com/mozilla/schedule-app-data-processor) should use make use of `start time` to automatically generate the `order` key instead. One fewer column in the Spreadsheet to manage. YAY! **Spreadsheet** - [x] remove "order" column from `* Timeblock Values` sheet **Python script** - [x] generate `order` based on `start time` (https://github.com/mozilla/schedule-app-data-processor/issues/20)
process
use start time to sort timeblocks currently value of the order key in is manually entered on the spreadsheet e g there s a column called order timeblock object has a start time key that we can use to sort a list of timeblocks our should use make use of start time to automatically generate the order key instead one fewer column in the spreadsheet to manage yay spreadsheet remove order column from timeblock values sheet python script generate order based on start time
1
4,676
7,517,294,569
IssuesEvent
2018-04-12 02:44:43
UnbFeelings/unb-feelings-GQA
https://api.github.com/repos/UnbFeelings/unb-feelings-GQA
closed
Estimar quantidade de auditorias
document help wanted process wiki
Nesta atividade, é considerada a lista dos objetos a serem auditados, para definir quantas auditorias serão necessárias para atender tal demanda. Contribuindo na criação do documento de descrição das auditorias planejadas.
1.0
Estimar quantidade de auditorias - Nesta atividade, é considerada a lista dos objetos a serem auditados, para definir quantas auditorias serão necessárias para atender tal demanda. Contribuindo na criação do documento de descrição das auditorias planejadas.
process
estimar quantidade de auditorias nesta atividade é considerada a lista dos objetos a serem auditados para definir quantas auditorias serão necessárias para atender tal demanda contribuindo na criação do documento de descrição das auditorias planejadas
1
12,231
14,743,624,715
IssuesEvent
2021-01-07 14:11:13
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Triad - Missing Baserates | parent: 1601
anc-process anp-1 ant-bug ant-child/secondary has attachment
In GitLab by @kdjstudios on Sep 5, 2019, 13:04 **Submitted by:** "Amecia Snelling" <amecia.snelling@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-09-05-73891 **Server:** Internal **Client/Site:** Triad **Account:** Multiple **Issue:** I am trying to process my billing but when finished I looked at the AR report and noticed that I had a lot of baserates missing. I looked at the Revenue Analysis and noticed that a lot of the baserates were not captured. Can someone check to see if some billing codes have been deleted? I will also need to know what can be done to fix this. Please see the spreadsheet that I have attached. All of the highlighted accounts are missing baserates. Please let me know if you have any questions. [revenue_analysis_worksheet_09_03_2019_Master.csv](/uploads/33397d3672bdcecbdc07ed3a1439ce6a/revenue_analysis_worksheet_09_03_2019_Master.csv)
1.0
Triad - Missing Baserates | parent: 1601 - In GitLab by @kdjstudios on Sep 5, 2019, 13:04 **Submitted by:** "Amecia Snelling" <amecia.snelling@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-09-05-73891 **Server:** Internal **Client/Site:** Triad **Account:** Multiple **Issue:** I am trying to process my billing but when finished I looked at the AR report and noticed that I had a lot of baserates missing. I looked at the Revenue Analysis and noticed that a lot of the baserates were not captured. Can someone check to see if some billing codes have been deleted? I will also need to know what can be done to fix this. Please see the spreadsheet that I have attached. All of the highlighted accounts are missing baserates. Please let me know if you have any questions. [revenue_analysis_worksheet_09_03_2019_Master.csv](/uploads/33397d3672bdcecbdc07ed3a1439ce6a/revenue_analysis_worksheet_09_03_2019_Master.csv)
process
triad missing baserates parent in gitlab by kdjstudios on sep submitted by amecia snelling helpdesk server internal client site triad account multiple issue i am trying to process my billing but when finished i looked at the ar report and noticed that i had a lot of baserates missing i looked at the revenue analysis and noticed that a lot of the baserates were not captured can someone check to see if some billing codes have been deleted i will also need to know what can be done to fix this please see the spreadsheet that i have attached all of the highlighted accounts are missing baserates please let me know if you have any questions uploads revenue analysis worksheet master csv
1
4,203
7,164,522,815
IssuesEvent
2018-01-29 11:31:34
LOVDnl/LOVD3
https://api.github.com/repos/LOVDnl/LOVD3
opened
Make protein change field non-mandatory for non-coding transcripts
cat: interface cat: submission process feature request
Transcripts that are non-coding can be identified by their NCBI ID. `NR` and `XR` prefixes indicate a non-coding transcript. For these transcripts, the "Protein change" field should be optional and not mandatory.
1.0
Make protein change field non-mandatory for non-coding transcripts - Transcripts that are non-coding can be identified by their NCBI ID. `NR` and `XR` prefixes indicate a non-coding transcript. For these transcripts, the "Protein change" field should be optional and not mandatory.
process
make protein change field non mandatory for non coding transcripts transcripts that are non coding can be identified by their ncbi id nr and xr prefixes indicate a non coding transcript for these transcripts the protein change field should be optional and not mandatory
1
145,955
11,716,148,820
IssuesEvent
2020-03-09 15:12:54
ICIJ/datashare
https://api.github.com/repos/ICIJ/datashare
closed
Contextualize filters do not work anymore
bug front need testing
To reproduce : - In "File Types", filter on PDFs (by example) - Click on "Contextualize filters" to enable it => Look at the filters, it is not contextualized !
1.0
Contextualize filters do not work anymore - To reproduce : - In "File Types", filter on PDFs (by example) - Click on "Contextualize filters" to enable it => Look at the filters, it is not contextualized !
non_process
contextualize filters do not work anymore to reproduce in file types filter on pdfs by example click on contextualize filters to enable it look at the filters it is not contextualized
0
294,493
25,376,642,166
IssuesEvent
2022-11-21 14:34:37
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
closed
[BUG] CPU mismatch GPU result in test_hash_groupby_collect_with_single_distinct intermittently
bug duplicate test
**Describe the bug** test_hash_groupby_collect_with_single_distinct[[('a', RepeatSeq(Long)), ('b', RepeatSeq(Boolean)), ('c', LongRange(not_null))]][IGNORE_ORDER({'local': True})] - AssertionError: GPU and CPU boolean values are different at [18, 'sort_array(collect_list(b), true)', 3] [2022-11-18T03:41:02.798Z] CPU OUTPUT: [Row(a=-7540734677356764604, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-5831592707909023540, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-5133656973475552689, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-4426181692283497353, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-3917032101531217289, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-3502159106106506455, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-2697073954890740236, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-2123199122092230623, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-1, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=207981845540287738, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=393905103838704542, sort_array(collect_list(b), true)=[False, False, False, False, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=875130347651831881, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=4751953708995107450, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=6084712057446794809, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7198729688045931692, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7528354001793048440, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=7618709293599214015, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7984374766242566542, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=9223372036854775807, sort_array(collect_list(b), true)=[**False, False, False, True**, True, True, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=10, count(c)=10)] [2022-11-18T03:41:02.799Z] GPU OUTPUT: [Row(a=-7540734677356764604, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-5831592707909023540, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-5133656973475552689, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-4426181692283497353, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-3917032101531217289, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-3502159106106506455, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-2697073954890740236, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-2123199122092230623, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-1, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=207981845540287738, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=393905103838704542, sort_array(collect_list(b), true)=[False, False, False, False, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=875130347651831881, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=4751953708995107450, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=6084712057446794809, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7198729688045931692, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7528354001793048440, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=7618709293599214015, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7984374766242566542, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=9223372036854775807, sort_array(collect_list(b), true)=[**False, False, False, False**, True, True, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=10, count(c)=10)] ``` [2022-11-18T03:41:02.797Z] _ test_hash_groupby_collect_with_single_distinct[[('a', RepeatSeq(Long)), ('b', RepeatSeq(Boolean)), ('c', LongRange(not_null))]] _ [2022-11-18T03:41:02.797Z] [gw3] linux -- Python 3.8.15 /usr/bin/python [2022-11-18T03:41:02.797Z] [2022-11-18T03:41:02.797Z] data_gen = [('a', RepeatSeq(Long)), ('b', RepeatSeq(Boolean)), ('c', LongRange(not_null))] [2022-11-18T03:41:02.797Z] [2022-11-18T03:41:02.797Z] @ignore_order(local=True) [2022-11-18T03:41:02.797Z] @pytest.mark.parametrize('data_gen', _full_gen_data_for_collect_op, ids=idfn) [2022-11-18T03:41:02.797Z] def test_hash_groupby_collect_with_single_distinct(data_gen): [2022-11-18T03:41:02.797Z] # test collect_ops with other distinct aggregations [2022-11-18T03:41:02.797Z] > assert_gpu_and_cpu_are_equal_collect( [2022-11-18T03:41:02.797Z] lambda spark: gen_df(spark, data_gen, length=100) [2022-11-18T03:41:02.797Z] .groupby('a') [2022-11-18T03:41:02.797Z] .agg(f.sort_array(f.collect_list('b')), [2022-11-18T03:41:02.797Z] f.sort_array(f.collect_set('b')), [2022-11-18T03:41:02.797Z] f.countDistinct('c'), [2022-11-18T03:41:02.797Z] f.count('c'))) [2022-11-18T03:41:02.797Z] [2022-11-18T03:41:02.797Z] ../../src/main/python/hash_aggregate_test.py:736: [2022-11-18T03:41:02.797Z] _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [2022-11-18T03:41:02.797Z] ../../src/main/python/asserts.py:548: in assert_gpu_and_cpu_are_equal_collect [2022-11-18T03:41:02.797Z] _assert_gpu_and_cpu_are_equal(func, 'COLLECT', conf=conf, is_cpu_first=is_cpu_first) [2022-11-18T03:41:02.797Z] ../../src/main/python/asserts.py:479: in _assert_gpu_and_cpu_are_equal [2022-11-18T03:41:02.797Z] assert_equal(from_cpu, from_gpu) [2022-11-18T03:41:02.797Z] ../../src/main/python/asserts.py:106: in assert_equal [2022-11-18T03:41:02.797Z] _assert_equal(cpu, gpu, float_check=get_float_check(), path=[]) [2022-11-18T03:41:02.797Z] ../../src/main/python/asserts.py:42: in _assert_equal [2022-11-18T03:41:02.797Z] _assert_equal(cpu[index], gpu[index], float_check, path + [index]) [2022-11-18T03:41:02.797Z] ../../src/main/python/asserts.py:35: in _assert_equal [2022-11-18T03:41:02.797Z] _assert_equal(cpu[field], gpu[field], float_check, path + [field]) [2022-11-18T03:41:02.797Z] ../../src/main/python/asserts.py:42: in _assert_equal [2022-11-18T03:41:02.797Z] _assert_equal(cpu[index], gpu[index], float_check, path + [index]) [2022-11-18T03:41:02.797Z] _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [2022-11-18T03:41:02.797Z] [2022-11-18T03:41:02.797Z] cpu = True, gpu = False [2022-11-18T03:41:02.797Z] float_check = <function get_float_check.<locals>.<lambda> at 0x7fcbd6f26ee0> [2022-11-18T03:41:02.797Z] path = [18, 'sort_array(collect_list(b), true)', 3] [2022-11-18T03:41:02.797Z] [2022-11-18T03:41:02.797Z] def _assert_equal(cpu, gpu, float_check, path): [2022-11-18T03:41:02.797Z] t = type(cpu) [2022-11-18T03:41:02.797Z] if (t is Row): [2022-11-18T03:41:02.797Z] assert len(cpu) == len(gpu), "CPU and GPU row have different lengths at {} CPU: {} GPU: {}".format(path, len(cpu), len(gpu)) [2022-11-18T03:41:02.797Z] if hasattr(cpu, "__fields__") and hasattr(gpu, "__fields__"): [2022-11-18T03:41:02.797Z] assert cpu.__fields__ == gpu.__fields__, "CPU and GPU row have different fields at {} CPU: {} GPU: {}".format(path, cpu.__fields__, gpu.__fields__) [2022-11-18T03:41:02.797Z] for field in cpu.__fields__: [2022-11-18T03:41:02.797Z] _assert_equal(cpu[field], gpu[field], float_check, path + [field]) [2022-11-18T03:41:02.797Z] else: [2022-11-18T03:41:02.797Z] for index in range(len(cpu)): [2022-11-18T03:41:02.797Z] _assert_equal(cpu[index], gpu[index], float_check, path + [index]) [2022-11-18T03:41:02.797Z] elif (t is list): [2022-11-18T03:41:02.797Z] assert len(cpu) == len(gpu), "CPU and GPU list have different lengths at {} CPU: {} GPU: {}".format(path, len(cpu), len(gpu)) [2022-11-18T03:41:02.798Z] for index in range(len(cpu)): [2022-11-18T03:41:02.798Z] _assert_equal(cpu[index], gpu[index], float_check, path + [index]) [2022-11-18T03:41:02.798Z] elif (t is tuple): [2022-11-18T03:41:02.798Z] assert len(cpu) == len(gpu), "CPU and GPU list have different lengths at {} CPU: {} GPU: {}".format(path, len(cpu), len(gpu)) [2022-11-18T03:41:02.798Z] for index in range(len(cpu)): [2022-11-18T03:41:02.798Z] _assert_equal(cpu[index], gpu[index], float_check, path + [index]) [2022-11-18T03:41:02.798Z] elif (t is pytypes.GeneratorType): [2022-11-18T03:41:02.798Z] index = 0 [2022-11-18T03:41:02.798Z] # generator has no zip :( so we have to do this the hard way [2022-11-18T03:41:02.798Z] done = False [2022-11-18T03:41:02.798Z] while not done: [2022-11-18T03:41:02.798Z] sub_cpu = None [2022-11-18T03:41:02.798Z] sub_gpu = None [2022-11-18T03:41:02.798Z] try: [2022-11-18T03:41:02.798Z] sub_cpu = next(cpu) [2022-11-18T03:41:02.798Z] except StopIteration: [2022-11-18T03:41:02.798Z] done = True [2022-11-18T03:41:02.798Z] [2022-11-18T03:41:02.798Z] try: [2022-11-18T03:41:02.798Z] sub_gpu = next(gpu) [2022-11-18T03:41:02.798Z] except StopIteration: [2022-11-18T03:41:02.798Z] done = True [2022-11-18T03:41:02.798Z] [2022-11-18T03:41:02.798Z] if done: [2022-11-18T03:41:02.798Z] assert sub_cpu == sub_gpu and sub_cpu == None, "CPU and GPU generators have different lengths at {}".format(path) [2022-11-18T03:41:02.798Z] else: [2022-11-18T03:41:02.798Z] _assert_equal(sub_cpu, sub_gpu, float_check, path + [index]) [2022-11-18T03:41:02.798Z] [2022-11-18T03:41:02.798Z] index = index + 1 [2022-11-18T03:41:02.798Z] elif (t is dict): [2022-11-18T03:41:02.798Z] # The order of key/values is not guaranteed in python dicts, nor are they guaranteed by Spark [2022-11-18T03:41:02.798Z] # so sort the items to do our best with ignoring the order of dicts [2022-11-18T03:41:02.798Z] cpu_items = list(cpu.items()).sort(key=_RowCmp) [2022-11-18T03:41:02.798Z] gpu_items = list(gpu.items()).sort(key=_RowCmp) [2022-11-18T03:41:02.798Z] _assert_equal(cpu_items, gpu_items, float_check, path + ["map"]) [2022-11-18T03:41:02.798Z] elif (t is int): [2022-11-18T03:41:02.798Z] assert cpu == gpu, "GPU and CPU int values are different at {}".format(path) [2022-11-18T03:41:02.798Z] elif (t is float): [2022-11-18T03:41:02.798Z] if (math.isnan(cpu)): [2022-11-18T03:41:02.798Z] assert math.isnan(gpu), "GPU and CPU float values are different at {}".format(path) [2022-11-18T03:41:02.798Z] else: [2022-11-18T03:41:02.798Z] assert float_check(cpu, gpu), "GPU and CPU float values are different {}".format(path) [2022-11-18T03:41:02.798Z] elif isinstance(cpu, str): [2022-11-18T03:41:02.798Z] assert cpu == gpu, "GPU and CPU string values are different at {}".format(path) [2022-11-18T03:41:02.798Z] elif isinstance(cpu, datetime): [2022-11-18T03:41:02.798Z] assert cpu == gpu, "GPU and CPU timestamp values are different at {}".format(path) [2022-11-18T03:41:02.798Z] elif isinstance(cpu, date): [2022-11-18T03:41:02.798Z] assert cpu == gpu, "GPU and CPU date values are different at {}".format(path) [2022-11-18T03:41:02.798Z] elif isinstance(cpu, bool): [2022-11-18T03:41:02.798Z] > assert cpu == gpu, "GPU and CPU boolean values are different at {}".format(path) [2022-11-18T03:41:02.798Z] E AssertionError: GPU and CPU boolean values are different at [18, 'sort_array(collect_list(b), true)', 3] [2022-11-18T03:41:02.798Z] [2022-11-18T03:41:02.798Z] ../../src/main/python/asserts.py:90: AssertionError [2022-11-18T03:41:02.798Z] ----------------------------- Captured stdout call ----------------------------- [2022-11-18T03:41:02.798Z] ### CPU RUN ### [2022-11-18T03:41:02.798Z] ### GPU RUN ### [2022-11-18T03:41:02.798Z] ### COLLECT: GPU TOOK 0.2924313545227051 CPU TOOK 0.29257845878601074 ### [2022-11-18T03:41:02.798Z] CPU OUTPUT: [Row(a=-7540734677356764604, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-5831592707909023540, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-5133656973475552689, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-4426181692283497353, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-3917032101531217289, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-3502159106106506455, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-2697073954890740236, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-2123199122092230623, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-1, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=207981845540287738, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=393905103838704542, sort_array(collect_list(b), true)=[False, False, False, False, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=875130347651831881, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=4751953708995107450, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=6084712057446794809, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7198729688045931692, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7528354001793048440, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=7618709293599214015, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7984374766242566542, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=9223372036854775807, sort_array(collect_list(b), true)=[False, False, False, True, True, True, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=10, count(c)=10)] [2022-11-18T03:41:02.799Z] GPU OUTPUT: [Row(a=-7540734677356764604, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-5831592707909023540, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-5133656973475552689, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-4426181692283497353, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-3917032101531217289, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-3502159106106506455, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-2697073954890740236, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-2123199122092230623, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-1, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=207981845540287738, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=393905103838704542, sort_array(collect_list(b), true)=[False, False, False, False, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=875130347651831881, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=4751953708995107450, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=6084712057446794809, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7198729688045931692, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7528354001793048440, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=7618709293599214015, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7984374766242566542, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=9223372036854775807, sort_array(collect_list(b), true)=[False, False, False, False, True, True, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=10, count(c)=10)] ``` **Steps/Code to reproduce bug** Not always reproducible, just saw one fail right now
1.0
[BUG] CPU mismatch GPU result in test_hash_groupby_collect_with_single_distinct intermittently - **Describe the bug** test_hash_groupby_collect_with_single_distinct[[('a', RepeatSeq(Long)), ('b', RepeatSeq(Boolean)), ('c', LongRange(not_null))]][IGNORE_ORDER({'local': True})] - AssertionError: GPU and CPU boolean values are different at [18, 'sort_array(collect_list(b), true)', 3] [2022-11-18T03:41:02.798Z] CPU OUTPUT: [Row(a=-7540734677356764604, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-5831592707909023540, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-5133656973475552689, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-4426181692283497353, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-3917032101531217289, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-3502159106106506455, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-2697073954890740236, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-2123199122092230623, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-1, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=207981845540287738, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=393905103838704542, sort_array(collect_list(b), true)=[False, False, False, False, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=875130347651831881, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=4751953708995107450, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=6084712057446794809, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7198729688045931692, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7528354001793048440, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=7618709293599214015, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7984374766242566542, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=9223372036854775807, sort_array(collect_list(b), true)=[**False, False, False, True**, True, True, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=10, count(c)=10)] [2022-11-18T03:41:02.799Z] GPU OUTPUT: [Row(a=-7540734677356764604, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-5831592707909023540, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-5133656973475552689, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-4426181692283497353, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-3917032101531217289, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-3502159106106506455, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-2697073954890740236, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-2123199122092230623, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-1, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=207981845540287738, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=393905103838704542, sort_array(collect_list(b), true)=[False, False, False, False, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=875130347651831881, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=4751953708995107450, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=6084712057446794809, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7198729688045931692, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7528354001793048440, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=7618709293599214015, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7984374766242566542, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=9223372036854775807, sort_array(collect_list(b), true)=[**False, False, False, False**, True, True, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=10, count(c)=10)] ``` [2022-11-18T03:41:02.797Z] _ test_hash_groupby_collect_with_single_distinct[[('a', RepeatSeq(Long)), ('b', RepeatSeq(Boolean)), ('c', LongRange(not_null))]] _ [2022-11-18T03:41:02.797Z] [gw3] linux -- Python 3.8.15 /usr/bin/python [2022-11-18T03:41:02.797Z] [2022-11-18T03:41:02.797Z] data_gen = [('a', RepeatSeq(Long)), ('b', RepeatSeq(Boolean)), ('c', LongRange(not_null))] [2022-11-18T03:41:02.797Z] [2022-11-18T03:41:02.797Z] @ignore_order(local=True) [2022-11-18T03:41:02.797Z] @pytest.mark.parametrize('data_gen', _full_gen_data_for_collect_op, ids=idfn) [2022-11-18T03:41:02.797Z] def test_hash_groupby_collect_with_single_distinct(data_gen): [2022-11-18T03:41:02.797Z] # test collect_ops with other distinct aggregations [2022-11-18T03:41:02.797Z] > assert_gpu_and_cpu_are_equal_collect( [2022-11-18T03:41:02.797Z] lambda spark: gen_df(spark, data_gen, length=100) [2022-11-18T03:41:02.797Z] .groupby('a') [2022-11-18T03:41:02.797Z] .agg(f.sort_array(f.collect_list('b')), [2022-11-18T03:41:02.797Z] f.sort_array(f.collect_set('b')), [2022-11-18T03:41:02.797Z] f.countDistinct('c'), [2022-11-18T03:41:02.797Z] f.count('c'))) [2022-11-18T03:41:02.797Z] [2022-11-18T03:41:02.797Z] ../../src/main/python/hash_aggregate_test.py:736: [2022-11-18T03:41:02.797Z] _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [2022-11-18T03:41:02.797Z] ../../src/main/python/asserts.py:548: in assert_gpu_and_cpu_are_equal_collect [2022-11-18T03:41:02.797Z] _assert_gpu_and_cpu_are_equal(func, 'COLLECT', conf=conf, is_cpu_first=is_cpu_first) [2022-11-18T03:41:02.797Z] ../../src/main/python/asserts.py:479: in _assert_gpu_and_cpu_are_equal [2022-11-18T03:41:02.797Z] assert_equal(from_cpu, from_gpu) [2022-11-18T03:41:02.797Z] ../../src/main/python/asserts.py:106: in assert_equal [2022-11-18T03:41:02.797Z] _assert_equal(cpu, gpu, float_check=get_float_check(), path=[]) [2022-11-18T03:41:02.797Z] ../../src/main/python/asserts.py:42: in _assert_equal [2022-11-18T03:41:02.797Z] _assert_equal(cpu[index], gpu[index], float_check, path + [index]) [2022-11-18T03:41:02.797Z] ../../src/main/python/asserts.py:35: in _assert_equal [2022-11-18T03:41:02.797Z] _assert_equal(cpu[field], gpu[field], float_check, path + [field]) [2022-11-18T03:41:02.797Z] ../../src/main/python/asserts.py:42: in _assert_equal [2022-11-18T03:41:02.797Z] _assert_equal(cpu[index], gpu[index], float_check, path + [index]) [2022-11-18T03:41:02.797Z] _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [2022-11-18T03:41:02.797Z] [2022-11-18T03:41:02.797Z] cpu = True, gpu = False [2022-11-18T03:41:02.797Z] float_check = <function get_float_check.<locals>.<lambda> at 0x7fcbd6f26ee0> [2022-11-18T03:41:02.797Z] path = [18, 'sort_array(collect_list(b), true)', 3] [2022-11-18T03:41:02.797Z] [2022-11-18T03:41:02.797Z] def _assert_equal(cpu, gpu, float_check, path): [2022-11-18T03:41:02.797Z] t = type(cpu) [2022-11-18T03:41:02.797Z] if (t is Row): [2022-11-18T03:41:02.797Z] assert len(cpu) == len(gpu), "CPU and GPU row have different lengths at {} CPU: {} GPU: {}".format(path, len(cpu), len(gpu)) [2022-11-18T03:41:02.797Z] if hasattr(cpu, "__fields__") and hasattr(gpu, "__fields__"): [2022-11-18T03:41:02.797Z] assert cpu.__fields__ == gpu.__fields__, "CPU and GPU row have different fields at {} CPU: {} GPU: {}".format(path, cpu.__fields__, gpu.__fields__) [2022-11-18T03:41:02.797Z] for field in cpu.__fields__: [2022-11-18T03:41:02.797Z] _assert_equal(cpu[field], gpu[field], float_check, path + [field]) [2022-11-18T03:41:02.797Z] else: [2022-11-18T03:41:02.797Z] for index in range(len(cpu)): [2022-11-18T03:41:02.797Z] _assert_equal(cpu[index], gpu[index], float_check, path + [index]) [2022-11-18T03:41:02.797Z] elif (t is list): [2022-11-18T03:41:02.797Z] assert len(cpu) == len(gpu), "CPU and GPU list have different lengths at {} CPU: {} GPU: {}".format(path, len(cpu), len(gpu)) [2022-11-18T03:41:02.798Z] for index in range(len(cpu)): [2022-11-18T03:41:02.798Z] _assert_equal(cpu[index], gpu[index], float_check, path + [index]) [2022-11-18T03:41:02.798Z] elif (t is tuple): [2022-11-18T03:41:02.798Z] assert len(cpu) == len(gpu), "CPU and GPU list have different lengths at {} CPU: {} GPU: {}".format(path, len(cpu), len(gpu)) [2022-11-18T03:41:02.798Z] for index in range(len(cpu)): [2022-11-18T03:41:02.798Z] _assert_equal(cpu[index], gpu[index], float_check, path + [index]) [2022-11-18T03:41:02.798Z] elif (t is pytypes.GeneratorType): [2022-11-18T03:41:02.798Z] index = 0 [2022-11-18T03:41:02.798Z] # generator has no zip :( so we have to do this the hard way [2022-11-18T03:41:02.798Z] done = False [2022-11-18T03:41:02.798Z] while not done: [2022-11-18T03:41:02.798Z] sub_cpu = None [2022-11-18T03:41:02.798Z] sub_gpu = None [2022-11-18T03:41:02.798Z] try: [2022-11-18T03:41:02.798Z] sub_cpu = next(cpu) [2022-11-18T03:41:02.798Z] except StopIteration: [2022-11-18T03:41:02.798Z] done = True [2022-11-18T03:41:02.798Z] [2022-11-18T03:41:02.798Z] try: [2022-11-18T03:41:02.798Z] sub_gpu = next(gpu) [2022-11-18T03:41:02.798Z] except StopIteration: [2022-11-18T03:41:02.798Z] done = True [2022-11-18T03:41:02.798Z] [2022-11-18T03:41:02.798Z] if done: [2022-11-18T03:41:02.798Z] assert sub_cpu == sub_gpu and sub_cpu == None, "CPU and GPU generators have different lengths at {}".format(path) [2022-11-18T03:41:02.798Z] else: [2022-11-18T03:41:02.798Z] _assert_equal(sub_cpu, sub_gpu, float_check, path + [index]) [2022-11-18T03:41:02.798Z] [2022-11-18T03:41:02.798Z] index = index + 1 [2022-11-18T03:41:02.798Z] elif (t is dict): [2022-11-18T03:41:02.798Z] # The order of key/values is not guaranteed in python dicts, nor are they guaranteed by Spark [2022-11-18T03:41:02.798Z] # so sort the items to do our best with ignoring the order of dicts [2022-11-18T03:41:02.798Z] cpu_items = list(cpu.items()).sort(key=_RowCmp) [2022-11-18T03:41:02.798Z] gpu_items = list(gpu.items()).sort(key=_RowCmp) [2022-11-18T03:41:02.798Z] _assert_equal(cpu_items, gpu_items, float_check, path + ["map"]) [2022-11-18T03:41:02.798Z] elif (t is int): [2022-11-18T03:41:02.798Z] assert cpu == gpu, "GPU and CPU int values are different at {}".format(path) [2022-11-18T03:41:02.798Z] elif (t is float): [2022-11-18T03:41:02.798Z] if (math.isnan(cpu)): [2022-11-18T03:41:02.798Z] assert math.isnan(gpu), "GPU and CPU float values are different at {}".format(path) [2022-11-18T03:41:02.798Z] else: [2022-11-18T03:41:02.798Z] assert float_check(cpu, gpu), "GPU and CPU float values are different {}".format(path) [2022-11-18T03:41:02.798Z] elif isinstance(cpu, str): [2022-11-18T03:41:02.798Z] assert cpu == gpu, "GPU and CPU string values are different at {}".format(path) [2022-11-18T03:41:02.798Z] elif isinstance(cpu, datetime): [2022-11-18T03:41:02.798Z] assert cpu == gpu, "GPU and CPU timestamp values are different at {}".format(path) [2022-11-18T03:41:02.798Z] elif isinstance(cpu, date): [2022-11-18T03:41:02.798Z] assert cpu == gpu, "GPU and CPU date values are different at {}".format(path) [2022-11-18T03:41:02.798Z] elif isinstance(cpu, bool): [2022-11-18T03:41:02.798Z] > assert cpu == gpu, "GPU and CPU boolean values are different at {}".format(path) [2022-11-18T03:41:02.798Z] E AssertionError: GPU and CPU boolean values are different at [18, 'sort_array(collect_list(b), true)', 3] [2022-11-18T03:41:02.798Z] [2022-11-18T03:41:02.798Z] ../../src/main/python/asserts.py:90: AssertionError [2022-11-18T03:41:02.798Z] ----------------------------- Captured stdout call ----------------------------- [2022-11-18T03:41:02.798Z] ### CPU RUN ### [2022-11-18T03:41:02.798Z] ### GPU RUN ### [2022-11-18T03:41:02.798Z] ### COLLECT: GPU TOOK 0.2924313545227051 CPU TOOK 0.29257845878601074 ### [2022-11-18T03:41:02.798Z] CPU OUTPUT: [Row(a=-7540734677356764604, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-5831592707909023540, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-5133656973475552689, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-4426181692283497353, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-3917032101531217289, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-3502159106106506455, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-2697073954890740236, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-2123199122092230623, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-1, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=207981845540287738, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=393905103838704542, sort_array(collect_list(b), true)=[False, False, False, False, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=875130347651831881, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=4751953708995107450, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=6084712057446794809, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7198729688045931692, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7528354001793048440, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=7618709293599214015, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7984374766242566542, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=9223372036854775807, sort_array(collect_list(b), true)=[False, False, False, True, True, True, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=10, count(c)=10)] [2022-11-18T03:41:02.799Z] GPU OUTPUT: [Row(a=-7540734677356764604, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-5831592707909023540, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-5133656973475552689, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-4426181692283497353, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-3917032101531217289, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-3502159106106506455, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=-2697073954890740236, sort_array(collect_list(b), true)=[False, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-2123199122092230623, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=-1, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=207981845540287738, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=393905103838704542, sort_array(collect_list(b), true)=[False, False, False, False, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=875130347651831881, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=4751953708995107450, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=6084712057446794809, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7198729688045931692, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7528354001793048440, sort_array(collect_list(b), true)=[False, False, False, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=7618709293599214015, sort_array(collect_list(b), true)=[True, True, True, True, True], sort_array(collect_set(b), true)=[True], count(c)=5, count(c)=5), Row(a=7984374766242566542, sort_array(collect_list(b), true)=[False, False, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=5, count(c)=5), Row(a=9223372036854775807, sort_array(collect_list(b), true)=[False, False, False, False, True, True, True, True, True, True], sort_array(collect_set(b), true)=[False, True], count(c)=10, count(c)=10)] ``` **Steps/Code to reproduce bug** Not always reproducible, just saw one fail right now
non_process
cpu mismatch gpu result in test hash groupby collect with single distinct intermittently describe the bug test hash groupby collect with single distinct assertionerror gpu and cpu boolean values are different at cpu output sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c gpu output sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c test hash groupby collect with single distinct linux python usr bin python data gen ignore order local true pytest mark parametrize data gen full gen data for collect op ids idfn def test hash groupby collect with single distinct data gen test collect ops with other distinct aggregations assert gpu and cpu are equal collect lambda spark gen df spark data gen length groupby a agg f sort array f collect list b f sort array f collect set b f countdistinct c f count c src main python hash aggregate test py src main python asserts py in assert gpu and cpu are equal collect assert gpu and cpu are equal func collect conf conf is cpu first is cpu first src main python asserts py in assert gpu and cpu are equal assert equal from cpu from gpu src main python asserts py in assert equal assert equal cpu gpu float check get float check path src main python asserts py in assert equal assert equal cpu gpu float check path src main python asserts py in assert equal assert equal cpu gpu float check path src main python asserts py in assert equal assert equal cpu gpu float check path cpu true gpu false float check at path def assert equal cpu gpu float check path t type cpu if t is row assert len cpu len gpu cpu and gpu row have different lengths at cpu gpu format path len cpu len gpu if hasattr cpu fields and hasattr gpu fields assert cpu fields gpu fields cpu and gpu row have different fields at cpu gpu format path cpu fields gpu fields for field in cpu fields assert equal cpu gpu float check path else for index in range len cpu assert equal cpu gpu float check path elif t is list assert len cpu len gpu cpu and gpu list have different lengths at cpu gpu format path len cpu len gpu for index in range len cpu assert equal cpu gpu float check path elif t is tuple assert len cpu len gpu cpu and gpu list have different lengths at cpu gpu format path len cpu len gpu for index in range len cpu assert equal cpu gpu float check path elif t is pytypes generatortype index generator has no zip so we have to do this the hard way done false while not done sub cpu none sub gpu none try sub cpu next cpu except stopiteration done true try sub gpu next gpu except stopiteration done true if done assert sub cpu sub gpu and sub cpu none cpu and gpu generators have different lengths at format path else assert equal sub cpu sub gpu float check path index index elif t is dict the order of key values is not guaranteed in python dicts nor are they guaranteed by spark so sort the items to do our best with ignoring the order of dicts cpu items list cpu items sort key rowcmp gpu items list gpu items sort key rowcmp assert equal cpu items gpu items float check path elif t is int assert cpu gpu gpu and cpu int values are different at format path elif t is float if math isnan cpu assert math isnan gpu gpu and cpu float values are different at format path else assert float check cpu gpu gpu and cpu float values are different format path elif isinstance cpu str assert cpu gpu gpu and cpu string values are different at format path elif isinstance cpu datetime assert cpu gpu gpu and cpu timestamp values are different at format path elif isinstance cpu date assert cpu gpu gpu and cpu date values are different at format path elif isinstance cpu bool assert cpu gpu gpu and cpu boolean values are different at format path e assertionerror gpu and cpu boolean values are different at src main python asserts py assertionerror captured stdout call cpu run gpu run collect gpu took cpu took cpu output sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c gpu output sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c row a sort array collect list b true sort array collect set b true count c count c steps code to reproduce bug not always reproducible just saw one fail right now
0
15,253
19,190,521,314
IssuesEvent
2021-12-05 22:38:57
km4ack/pi-build
https://api.github.com/repos/km4ack/pi-build
closed
hamlib (rigctl) won't install on Bullseye SOLVED
bug in process
libusb-1.0 not available in bullseye. Need to compile from source with: ``` sudo apt install -y libusb-dev libudev-dev cd ~/Downloads wget https://github.com/libusb/libusb/releases/download/v1.0.23/libusb-1.0.23.tar.bz2 tar -xvf libusb-1.0.23.tar.bz2 cd libusb-1.0.23/ ./configure make sudo make install sudo ldconfig ``` now run hamlib install as before with BAP
1.0
hamlib (rigctl) won't install on Bullseye SOLVED - libusb-1.0 not available in bullseye. Need to compile from source with: ``` sudo apt install -y libusb-dev libudev-dev cd ~/Downloads wget https://github.com/libusb/libusb/releases/download/v1.0.23/libusb-1.0.23.tar.bz2 tar -xvf libusb-1.0.23.tar.bz2 cd libusb-1.0.23/ ./configure make sudo make install sudo ldconfig ``` now run hamlib install as before with BAP
process
hamlib rigctl won t install on bullseye solved libusb not available in bullseye need to compile from source with sudo apt install y libusb dev libudev dev cd downloads wget tar xvf libusb tar cd libusb configure make sudo make install sudo ldconfig now run hamlib install as before with bap
1
109,284
13,758,112,051
IssuesEvent
2020-10-06 23:08:59
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
Provide a way to visually inspect the inheritance hierarchy
Area-IDE Feature Request InternalAsk Need Design Review
View type/member hierarchy - display a hierarchy with links to base and derived types/members. Go To Implementation provides a list of derived types/members, but without (a) base types/members or (b) showing the inheritance relationship between different items
1.0
Provide a way to visually inspect the inheritance hierarchy - View type/member hierarchy - display a hierarchy with links to base and derived types/members. Go To Implementation provides a list of derived types/members, but without (a) base types/members or (b) showing the inheritance relationship between different items
non_process
provide a way to visually inspect the inheritance hierarchy view type member hierarchy display a hierarchy with links to base and derived types members go to implementation provides a list of derived types members but without a base types members or b showing the inheritance relationship between different items
0
342,724
10,320,898,221
IssuesEvent
2019-08-30 22:14:23
HackGT/bolt
https://api.github.com/repos/HackGT/bolt
opened
Track request change history
component / requests priority / medium type / feature
Create a new database table for tracking changes to requests - user, item, quantity, or status (old and new) - and the time it occurred. This will provide better visibility into requests and give us the data to have analytics like average time to approve/fulfill requests, which items are most in demand, etc.
1.0
Track request change history - Create a new database table for tracking changes to requests - user, item, quantity, or status (old and new) - and the time it occurred. This will provide better visibility into requests and give us the data to have analytics like average time to approve/fulfill requests, which items are most in demand, etc.
non_process
track request change history create a new database table for tracking changes to requests user item quantity or status old and new and the time it occurred this will provide better visibility into requests and give us the data to have analytics like average time to approve fulfill requests which items are most in demand etc
0
750
3,223,389,891
IssuesEvent
2015-10-09 09:37:54
MTG/freesound
https://api.github.com/repos/MTG/freesound
opened
Refactor sound processing state handling
Improvement _Processing
The current way in which we handle processing state and processing of sounds does not allow for optimised workflows when we want to reprocess sounds. To improve this, the processing_state field in the sound model should only be used to say that a sound has either failed processing or being successfully processed (similarity to the moderation state). Another property 'processing_ongoing_state' should be used to set the state of the actual processing activity (either queued, processing or finished). This will allow to better handle reprocessing of sounds and optimise other related stuff such as removing sounds from solr and similarity when they fail processing and updating packs, num_sounds on user profiles and other related stuff when sounds change their processing_state.
1.0
Refactor sound processing state handling - The current way in which we handle processing state and processing of sounds does not allow for optimised workflows when we want to reprocess sounds. To improve this, the processing_state field in the sound model should only be used to say that a sound has either failed processing or being successfully processed (similarity to the moderation state). Another property 'processing_ongoing_state' should be used to set the state of the actual processing activity (either queued, processing or finished). This will allow to better handle reprocessing of sounds and optimise other related stuff such as removing sounds from solr and similarity when they fail processing and updating packs, num_sounds on user profiles and other related stuff when sounds change their processing_state.
process
refactor sound processing state handling the current way in which we handle processing state and processing of sounds does not allow for optimised workflows when we want to reprocess sounds to improve this the processing state field in the sound model should only be used to say that a sound has either failed processing or being successfully processed similarity to the moderation state another property processing ongoing state should be used to set the state of the actual processing activity either queued processing or finished this will allow to better handle reprocessing of sounds and optimise other related stuff such as removing sounds from solr and similarity when they fail processing and updating packs num sounds on user profiles and other related stuff when sounds change their processing state
1
506,236
14,660,527,857
IssuesEvent
2020-12-29 00:03:31
bounswe/bounswe2020group8
https://api.github.com/repos/bounswe/bounswe2020group8
closed
Implement Main Product listing for admins
Priority: Medium enhancement web
**Describe the feature** The admins should be able to list and delete main products. **Describe solutions you've considered** The same pattern of normal products will follow here.
1.0
Implement Main Product listing for admins - **Describe the feature** The admins should be able to list and delete main products. **Describe solutions you've considered** The same pattern of normal products will follow here.
non_process
implement main product listing for admins describe the feature the admins should be able to list and delete main products describe solutions you ve considered the same pattern of normal products will follow here
0
2,564
5,316,087,379
IssuesEvent
2017-02-13 18:59:25
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
closed
[subtitles] [fr] «Nous sommes tes fils, Révolution !» - Mélenchon citant Hugo en meeting
Language: French Process: [6] Approved
# Video title «Nous sommes tes fils, Révolution !» - Mélenchon citant Hugo en meeting # URL https://www.youtube.com/watch?v=B8xROYog8KE # Youtube subtitles language French # Duration 2:28 # Subtitles URL https://www.youtube.com/timedtext_editor?tab=captions&v=B8xROYog8KE&action_mde_edit_form=1&ui=hd&ref=player&bl=vmp&lang=fr
1.0
[subtitles] [fr] «Nous sommes tes fils, Révolution !» - Mélenchon citant Hugo en meeting - # Video title «Nous sommes tes fils, Révolution !» - Mélenchon citant Hugo en meeting # URL https://www.youtube.com/watch?v=B8xROYog8KE # Youtube subtitles language French # Duration 2:28 # Subtitles URL https://www.youtube.com/timedtext_editor?tab=captions&v=B8xROYog8KE&action_mde_edit_form=1&ui=hd&ref=player&bl=vmp&lang=fr
process
«nous sommes tes fils révolution » mélenchon citant hugo en meeting video title «nous sommes tes fils révolution » mélenchon citant hugo en meeting url youtube subtitles language french duration subtitles url
1
2,024
4,846,818,235
IssuesEvent
2016-11-10 13:07:57
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
opened
When starting a process should user need to provide a name?
comp: activiti-processList
Comparing starting a process within Activiti to that in the adf components there are some differences, which may be intentional? 1. Unsure what the dropdown 'type' refers to, should this be 'process'? 2. A process can only be started if a name is added ('start' option is greyed out), should renaming a process be avaliable to the user and if so should it be optional? 3. If the name is optional should, when the process is started, include the process name or the name given when starting the task? **Components** ![screen shot 2016-11-10 at 12 45 13](https://cloud.githubusercontent.com/assets/13200338/20177315/1552a3da-a744-11e6-9781-0c53a2d86d70.png) **Activiti** ![screen shot 2016-11-10 at 12 45 28](https://cloud.githubusercontent.com/assets/13200338/20177348/36966be4-a744-11e6-8f45-16ddac0b4eca.png)
1.0
When starting a process should user need to provide a name? - Comparing starting a process within Activiti to that in the adf components there are some differences, which may be intentional? 1. Unsure what the dropdown 'type' refers to, should this be 'process'? 2. A process can only be started if a name is added ('start' option is greyed out), should renaming a process be avaliable to the user and if so should it be optional? 3. If the name is optional should, when the process is started, include the process name or the name given when starting the task? **Components** ![screen shot 2016-11-10 at 12 45 13](https://cloud.githubusercontent.com/assets/13200338/20177315/1552a3da-a744-11e6-9781-0c53a2d86d70.png) **Activiti** ![screen shot 2016-11-10 at 12 45 28](https://cloud.githubusercontent.com/assets/13200338/20177348/36966be4-a744-11e6-8f45-16ddac0b4eca.png)
process
when starting a process should user need to provide a name comparing starting a process within activiti to that in the adf components there are some differences which may be intentional unsure what the dropdown type refers to should this be process a process can only be started if a name is added start option is greyed out should renaming a process be avaliable to the user and if so should it be optional if the name is optional should when the process is started include the process name or the name given when starting the task components activiti
1
280,524
8,683,389,347
IssuesEvent
2018-12-02 17:42:59
sem5640-2018/gatekeeper
https://api.github.com/repos/sem5640-2018/gatekeeper
closed
Batch user data API
enhancement high priority
Some services need to make multiple calls to the API to get user data for indexes. Having a single API call would lower the amount of calls. Plz do.
1.0
Batch user data API - Some services need to make multiple calls to the API to get user data for indexes. Having a single API call would lower the amount of calls. Plz do.
non_process
batch user data api some services need to make multiple calls to the api to get user data for indexes having a single api call would lower the amount of calls plz do
0
9,772
12,760,870,849
IssuesEvent
2020-06-29 08:48:33
ClickHouse/ClickHouse
https://api.github.com/repos/ClickHouse/ClickHouse
opened
UNION ALL issue with processors
bug comp-processors
``` set experimental_use_processors=0; SELECT count() FROM ( SELECT number FROM system.numbers_mt WHERE number <= 100000000 LIMIT 100000000 UNION ALL SELECT number FROM system.numbers_mt WHERE number <= 100000000 LIMIT 100000000 ) ┌───count()─┐ │ 200000000 │ └───────────┘ 1 rows in set. Elapsed: 0.213 sec. Processed 263.52 million rows, 2.11 GB (1.24 billion rows/s., 9.89 GB/s.) ``` ``` set experimental_use_processors=0; SELECT count() FROM ( SELECT number FROM system.numbers_mt WHERE number <= 100000000 LIMIT 100000000 UNION ALL SELECT number FROM system.numbers_mt WHERE number <= 100000000 LIMIT 100000000 ) ↓ Progress: 29.44 billion rows, 235.48 GB (2.58 billion rows/s., 20.62 GB/s.) Cancelling query. ↙ Progress: 29.68 billion rows, 237.41 GB (2.58 billion rows/s., 20.61 GB/s.) Ok. Query was cancelled. 0 rows in set. Elapsed: 11.521 sec. Processed 29.68 billion rows, 237.41 GB (2.58 billion rows/s., 20.61 GB/s.) -- never finishes. ```
1.0
UNION ALL issue with processors - ``` set experimental_use_processors=0; SELECT count() FROM ( SELECT number FROM system.numbers_mt WHERE number <= 100000000 LIMIT 100000000 UNION ALL SELECT number FROM system.numbers_mt WHERE number <= 100000000 LIMIT 100000000 ) ┌───count()─┐ │ 200000000 │ └───────────┘ 1 rows in set. Elapsed: 0.213 sec. Processed 263.52 million rows, 2.11 GB (1.24 billion rows/s., 9.89 GB/s.) ``` ``` set experimental_use_processors=0; SELECT count() FROM ( SELECT number FROM system.numbers_mt WHERE number <= 100000000 LIMIT 100000000 UNION ALL SELECT number FROM system.numbers_mt WHERE number <= 100000000 LIMIT 100000000 ) ↓ Progress: 29.44 billion rows, 235.48 GB (2.58 billion rows/s., 20.62 GB/s.) Cancelling query. ↙ Progress: 29.68 billion rows, 237.41 GB (2.58 billion rows/s., 20.61 GB/s.) Ok. Query was cancelled. 0 rows in set. Elapsed: 11.521 sec. Processed 29.68 billion rows, 237.41 GB (2.58 billion rows/s., 20.61 GB/s.) -- never finishes. ```
process
union all issue with processors set experimental use processors select count from select number from system numbers mt where number limit union all select number from system numbers mt where number limit ┌───count ─┐ │ │ └───────────┘ rows in set elapsed sec processed million rows gb billion rows s gb s set experimental use processors select count from select number from system numbers mt where number limit union all select number from system numbers mt where number limit ↓ progress billion rows gb billion rows s gb s cancelling query ↙ progress billion rows gb billion rows s gb s ok query was cancelled rows in set elapsed sec processed billion rows gb billion rows s gb s never finishes
1
22,672
31,896,194,403
IssuesEvent
2023-09-18 02:05:03
rust-lang/cargo
https://api.github.com/repos/rust-lang/cargo
closed
Tracking Issue for credential-process RFC 2730
T-cargo C-tracking-issue final-comment-period disposition-merge Z-credential-process A-registry-authentication S-waiting-on-feedback
**Summary** RFC: https://github.com/rust-lang/rfcs/pull/2730 Implementation: #8934 Documentation: https://doc.rust-lang.org/nightly/cargo/reference/unstable.html#credential-process Issues: https://github.com/rust-lang/cargo/labels/Z-credential-process This feature provides a configuration option to specify a process to fetch a token to authenticate with a registry. **Unresolved issues** * [X] Is this approach useful enough? Things like macos keychain don't protect against being executed to extract tokens (and I don't see a way to require a password, or force the process to be untrusted). The lack of signatures also cause issues (each update of the toolchain will cause it to become untrusted again). * Making the built-in providers like `cargo:macos-keychain` part of the cargo binary improves security here, since only the Cargo binary is accessing the Keychain. If we get Cargo signed by Apple, then it would be further improved. * [X] Should the `login` API be changed? Providers that need to be interactive need to be able to read from stdin. * This is worked around by having the credential-provider re-open `stdin` from `/dev/tty` or `$CONIN` * [X] Should the storage key off the registry name? [RFC 3139](https://github.com/rust-lang/rfcs/pull/3139) discusses problems with this. RESOLVED: credential helpers are now keying off the index url via #10592. * [x] Should it be possible to pass additional arguments on the command line for `cargo login` that would be sent to the provider? E.g. `cargo login -- --extra-arg-for-provider` * Implemented in #12499 * [X] Figure out the name for `cargo:basic`: `cargo:token-from-stdout` * [X] Remove cargo-1password as a built-in (add the `fn main` so it can be published separately). * [x] Create a new `-Z` flag for `cargo:paseto`. * [x] Publish `cargo-credential-1password` to crates.io * [x] Figure out MSRV support. rust-lang/cargo#12381 is proposing to set MSRV for all packages in the repo. I think it will be important that the credential provider libraries have a relatively relaxed requirement (like stable-2 or something). I think this will be important to figure out before stabilization. This may be complicated with CI costs, so I'm wondering if `cargo-credential` could grow a few independent tests on its own so we could test just the library with an older rust? https://github.com/rust-lang/cargo/pull/12623 * [x] Testing of forwards-compatibility support. It is important that when a user installs a new credential provider that it does not necessarily cause older `cargo`s to fail. There is a version field when we need to do breaking changes, but we need tests that are actually verifying that is working. https://github.com/rust-lang/cargo/pull/12622 * [x] A test with some manually generated JSON with an invalid version file fed into a credential provider should provide the correct error message. * [x] A test with some manually generated JSON with the current V1 format that is fed into a credential provider, and validate that it behaves correctly. This will make sure that if any changes to the credential provider are made that are not compatible with the old format that the test will fail. * [x] Resolve rust-lang/cargo#12642 * [ ] Stabilization PR - #12649 * [ ] Publish `cargo-credential` `v0.4` to [crates.io](https://crates.io/crates/cargo-credential) * [ ] Publish `cargo-credential-1password` `v0.4` to [crates.io](https://crates.io/crates/cargo-credential-1password) **About tracking issues** Tracking issues are used to record the overall progress of implementation. They are also used as hubs connecting to other relevant issues, e.g., bugs or open design questions. A tracking issue is however *not* meant for large scale discussion, questions, or bug reports about a feature. Instead, open a dedicated issue for the specific matter and add the relevant feature gate label.
1.0
Tracking Issue for credential-process RFC 2730 - **Summary** RFC: https://github.com/rust-lang/rfcs/pull/2730 Implementation: #8934 Documentation: https://doc.rust-lang.org/nightly/cargo/reference/unstable.html#credential-process Issues: https://github.com/rust-lang/cargo/labels/Z-credential-process This feature provides a configuration option to specify a process to fetch a token to authenticate with a registry. **Unresolved issues** * [X] Is this approach useful enough? Things like macos keychain don't protect against being executed to extract tokens (and I don't see a way to require a password, or force the process to be untrusted). The lack of signatures also cause issues (each update of the toolchain will cause it to become untrusted again). * Making the built-in providers like `cargo:macos-keychain` part of the cargo binary improves security here, since only the Cargo binary is accessing the Keychain. If we get Cargo signed by Apple, then it would be further improved. * [X] Should the `login` API be changed? Providers that need to be interactive need to be able to read from stdin. * This is worked around by having the credential-provider re-open `stdin` from `/dev/tty` or `$CONIN` * [X] Should the storage key off the registry name? [RFC 3139](https://github.com/rust-lang/rfcs/pull/3139) discusses problems with this. RESOLVED: credential helpers are now keying off the index url via #10592. * [x] Should it be possible to pass additional arguments on the command line for `cargo login` that would be sent to the provider? E.g. `cargo login -- --extra-arg-for-provider` * Implemented in #12499 * [X] Figure out the name for `cargo:basic`: `cargo:token-from-stdout` * [X] Remove cargo-1password as a built-in (add the `fn main` so it can be published separately). * [x] Create a new `-Z` flag for `cargo:paseto`. * [x] Publish `cargo-credential-1password` to crates.io * [x] Figure out MSRV support. rust-lang/cargo#12381 is proposing to set MSRV for all packages in the repo. I think it will be important that the credential provider libraries have a relatively relaxed requirement (like stable-2 or something). I think this will be important to figure out before stabilization. This may be complicated with CI costs, so I'm wondering if `cargo-credential` could grow a few independent tests on its own so we could test just the library with an older rust? https://github.com/rust-lang/cargo/pull/12623 * [x] Testing of forwards-compatibility support. It is important that when a user installs a new credential provider that it does not necessarily cause older `cargo`s to fail. There is a version field when we need to do breaking changes, but we need tests that are actually verifying that is working. https://github.com/rust-lang/cargo/pull/12622 * [x] A test with some manually generated JSON with an invalid version file fed into a credential provider should provide the correct error message. * [x] A test with some manually generated JSON with the current V1 format that is fed into a credential provider, and validate that it behaves correctly. This will make sure that if any changes to the credential provider are made that are not compatible with the old format that the test will fail. * [x] Resolve rust-lang/cargo#12642 * [ ] Stabilization PR - #12649 * [ ] Publish `cargo-credential` `v0.4` to [crates.io](https://crates.io/crates/cargo-credential) * [ ] Publish `cargo-credential-1password` `v0.4` to [crates.io](https://crates.io/crates/cargo-credential-1password) **About tracking issues** Tracking issues are used to record the overall progress of implementation. They are also used as hubs connecting to other relevant issues, e.g., bugs or open design questions. A tracking issue is however *not* meant for large scale discussion, questions, or bug reports about a feature. Instead, open a dedicated issue for the specific matter and add the relevant feature gate label.
process
tracking issue for credential process rfc summary rfc implementation documentation issues this feature provides a configuration option to specify a process to fetch a token to authenticate with a registry unresolved issues is this approach useful enough things like macos keychain don t protect against being executed to extract tokens and i don t see a way to require a password or force the process to be untrusted the lack of signatures also cause issues each update of the toolchain will cause it to become untrusted again making the built in providers like cargo macos keychain part of the cargo binary improves security here since only the cargo binary is accessing the keychain if we get cargo signed by apple then it would be further improved should the login api be changed providers that need to be interactive need to be able to read from stdin this is worked around by having the credential provider re open stdin from dev tty or conin should the storage key off the registry name discusses problems with this resolved credential helpers are now keying off the index url via should it be possible to pass additional arguments on the command line for cargo login that would be sent to the provider e g cargo login extra arg for provider implemented in figure out the name for cargo basic cargo token from stdout remove cargo as a built in add the fn main so it can be published separately create a new z flag for cargo paseto publish cargo credential to crates io figure out msrv support rust lang cargo is proposing to set msrv for all packages in the repo i think it will be important that the credential provider libraries have a relatively relaxed requirement like stable or something i think this will be important to figure out before stabilization this may be complicated with ci costs so i m wondering if cargo credential could grow a few independent tests on its own so we could test just the library with an older rust testing of forwards compatibility support it is important that when a user installs a new credential provider that it does not necessarily cause older cargo s to fail there is a version field when we need to do breaking changes but we need tests that are actually verifying that is working a test with some manually generated json with an invalid version file fed into a credential provider should provide the correct error message a test with some manually generated json with the current format that is fed into a credential provider and validate that it behaves correctly this will make sure that if any changes to the credential provider are made that are not compatible with the old format that the test will fail resolve rust lang cargo stabilization pr publish cargo credential to publish cargo credential to about tracking issues tracking issues are used to record the overall progress of implementation they are also used as hubs connecting to other relevant issues e g bugs or open design questions a tracking issue is however not meant for large scale discussion questions or bug reports about a feature instead open a dedicated issue for the specific matter and add the relevant feature gate label
1
12,217
14,743,029,777
IssuesEvent
2021-01-07 13:17:33
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
recurring charges not on invoice
anc-process anp-important ant-bug ant-support
In GitLab by @kdjstudios on Jul 1, 2019, 14:58 **Submitted by:** Gaylan Garrett <Gaylan.Garrett@Nexa.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/8690702 **Server:** External **Client/Site:** Keener **Account:** Multiple **Issue:** I have four accounts that the recurring did not show up on the 6/23/2019 invoice even though it is an active transaction in the account. 4342360555, 4343940337, 4343337359 and 6106282263. Can you let me know what happened ?
1.0
recurring charges not on invoice - In GitLab by @kdjstudios on Jul 1, 2019, 14:58 **Submitted by:** Gaylan Garrett <Gaylan.Garrett@Nexa.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/8690702 **Server:** External **Client/Site:** Keener **Account:** Multiple **Issue:** I have four accounts that the recurring did not show up on the 6/23/2019 invoice even though it is an active transaction in the account. 4342360555, 4343940337, 4343337359 and 6106282263. Can you let me know what happened ?
process
recurring charges not on invoice in gitlab by kdjstudios on jul submitted by gaylan garrett helpdesk server external client site keener account multiple issue i have four accounts that the recurring did not show up on the invoice even though it is an active transaction in the account and can you let me know what happened
1
22,573
31,799,305,925
IssuesEvent
2023-09-13 10:01:49
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/resourcedetection] system detector sets `host.id` to an empty value on containerized setups
bug Stale priority:p2 processor/resourcedetection
### Component(s) processor/resourcedetection ### What happened? ## Description The `host.id` resource attribute is set by the `system` detector to an empty string value when running on containerized setups. ## Steps to Reproduce Run the Collector contrib Docker image with the configuration provided below. This is also reproducible with custom builds if one has some base images (e.g. `alpine:3.16`). ## Expected Result The detector works in accordance to [the specification](https://github.com/open-telemetry/semantic-conventions/blob/c6efa4b2b44f9950b332a46311834e23b85bf979/docs/resource/host.md) which states that (emphasis mine): > Unique host ID. For Cloud, this must be the instance_id assigned by the cloud provider. For **non-containerized systems,** this should be the machine-id. See the table below for the sources to use to determine the machine-id based on operating system. AIUI, an empty string is not valid since it's not unique, and on containerized environments, this should not be the machine-id since a container is not 'really' a host. My expectation would be either that on containerized environments the `host.id` resource attribute is not set or that it is set to a value that persists across restarts. ## Actual Result On v0.80.0, an empty string is set on `host.id`, which is not a "unique host ID". On v0.72.0 up to v0.79.0, a random UUID that varies on each container restart is set on `host.id`. ### Collector version v0.80.0 ### Environment information ## Environment Running the docker image `otel/opentelemetry-collector-contrib:0.80.0` with the configuration provided below reproduces this. ### OpenTelemetry Collector configuration ```yaml receivers: # Put a dummy receiver just to generate some metrics hostmetrics: collection_interval: 10s scrapers: load: processors: resourcedetection: detectors: [system] exporters: logging: verbosity: detailed service: pipelines: metrics: receivers: [hostmetrics] processors: [resourcedetection] exporters: [logging] ``` ### Log output ```shell 2023-07-12T10:15:03.905Z info service/telemetry.go:81 Setting up own telemetry... 2023-07-12T10:15:03.907Z info service/telemetry.go:104 Serving Prometheus metrics {"address": ":8888", "level": "Basic"} 2023-07-12T10:15:03.907Z info exporter@v0.80.0/exporter.go:275 Development component. May change in the future. {"kind": "exporter", "data_type": "metrics", "name": "logging"} 2023-07-12T10:15:03.910Z info service/service.go:131 Starting otelcol-contrib... {"Version": "0.80.0", "NumCPU": 20} 2023-07-12T10:15:03.910Z info extensions/extensions.go:30 Starting extensions... 2023-07-12T10:15:03.910Z info internal/resourcedetection.go:125 began detecting resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "metrics"} 2023-07-12T10:15:03.912Z info internal/resourcedetection.go:139 detected resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "metrics", "resource": {"host.id":"","host.name":"d5ad29786527","os.type":"linux"}} ``` ### Additional context Prior to v0.80.0 (in particular, in between v0.72.0 that included #18618 and before #18740) this would generate a random UUID on each container restart. An example run of the same configuration on v0.79.0: ``` 2023-07-07T11:04:20.764Z info service/telemetry.go:104 Setting up own telemetry... 2023-07-07T11:04:20.764Z info service/telemetry.go:127 Serving Prometheus metrics {"address": ":8888", "level": "Basic"} 2023-07-07T11:04:20.764Z info exporter@v0.79.0/exporter.go:275 Development component. May change in the future. {"kind": "exporter", "data_type": "metrics", "name": "logging"} 2023-07-07T11:04:20.766Z info service/service.go:131 Starting otelcol-contrib... {"Version": "0.79.0", "NumCPU": 20} 2023-07-07T11:04:20.766Z info extensions/extensions.go:30 Starting extensions... 2023-07-07T11:04:20.766Z info internal/resourcedetection.go:125 began detecting resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "metrics"} 2023-07-07T11:04:20.766Z info internal/resourcedetection.go:139 detected resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "metrics", "resource": {"host.id":"d8aeac44-921a-449b-a1d9-aa1621c50d3d","host.name":"43a4683d77f0","os.type":"linux"}} ``` This was reported on https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/18618#issuecomment-1539413938 and happens because [this code path](https://github.com/panta/machineid/blob/c77705a58e0650035e15e10df9d594c9d169dedf/id_linux.go#L48-L53) is taken on the dependency used by #18618 to run this.
1.0
[processor/resourcedetection] system detector sets `host.id` to an empty value on containerized setups - ### Component(s) processor/resourcedetection ### What happened? ## Description The `host.id` resource attribute is set by the `system` detector to an empty string value when running on containerized setups. ## Steps to Reproduce Run the Collector contrib Docker image with the configuration provided below. This is also reproducible with custom builds if one has some base images (e.g. `alpine:3.16`). ## Expected Result The detector works in accordance to [the specification](https://github.com/open-telemetry/semantic-conventions/blob/c6efa4b2b44f9950b332a46311834e23b85bf979/docs/resource/host.md) which states that (emphasis mine): > Unique host ID. For Cloud, this must be the instance_id assigned by the cloud provider. For **non-containerized systems,** this should be the machine-id. See the table below for the sources to use to determine the machine-id based on operating system. AIUI, an empty string is not valid since it's not unique, and on containerized environments, this should not be the machine-id since a container is not 'really' a host. My expectation would be either that on containerized environments the `host.id` resource attribute is not set or that it is set to a value that persists across restarts. ## Actual Result On v0.80.0, an empty string is set on `host.id`, which is not a "unique host ID". On v0.72.0 up to v0.79.0, a random UUID that varies on each container restart is set on `host.id`. ### Collector version v0.80.0 ### Environment information ## Environment Running the docker image `otel/opentelemetry-collector-contrib:0.80.0` with the configuration provided below reproduces this. ### OpenTelemetry Collector configuration ```yaml receivers: # Put a dummy receiver just to generate some metrics hostmetrics: collection_interval: 10s scrapers: load: processors: resourcedetection: detectors: [system] exporters: logging: verbosity: detailed service: pipelines: metrics: receivers: [hostmetrics] processors: [resourcedetection] exporters: [logging] ``` ### Log output ```shell 2023-07-12T10:15:03.905Z info service/telemetry.go:81 Setting up own telemetry... 2023-07-12T10:15:03.907Z info service/telemetry.go:104 Serving Prometheus metrics {"address": ":8888", "level": "Basic"} 2023-07-12T10:15:03.907Z info exporter@v0.80.0/exporter.go:275 Development component. May change in the future. {"kind": "exporter", "data_type": "metrics", "name": "logging"} 2023-07-12T10:15:03.910Z info service/service.go:131 Starting otelcol-contrib... {"Version": "0.80.0", "NumCPU": 20} 2023-07-12T10:15:03.910Z info extensions/extensions.go:30 Starting extensions... 2023-07-12T10:15:03.910Z info internal/resourcedetection.go:125 began detecting resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "metrics"} 2023-07-12T10:15:03.912Z info internal/resourcedetection.go:139 detected resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "metrics", "resource": {"host.id":"","host.name":"d5ad29786527","os.type":"linux"}} ``` ### Additional context Prior to v0.80.0 (in particular, in between v0.72.0 that included #18618 and before #18740) this would generate a random UUID on each container restart. An example run of the same configuration on v0.79.0: ``` 2023-07-07T11:04:20.764Z info service/telemetry.go:104 Setting up own telemetry... 2023-07-07T11:04:20.764Z info service/telemetry.go:127 Serving Prometheus metrics {"address": ":8888", "level": "Basic"} 2023-07-07T11:04:20.764Z info exporter@v0.79.0/exporter.go:275 Development component. May change in the future. {"kind": "exporter", "data_type": "metrics", "name": "logging"} 2023-07-07T11:04:20.766Z info service/service.go:131 Starting otelcol-contrib... {"Version": "0.79.0", "NumCPU": 20} 2023-07-07T11:04:20.766Z info extensions/extensions.go:30 Starting extensions... 2023-07-07T11:04:20.766Z info internal/resourcedetection.go:125 began detecting resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "metrics"} 2023-07-07T11:04:20.766Z info internal/resourcedetection.go:139 detected resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "metrics", "resource": {"host.id":"d8aeac44-921a-449b-a1d9-aa1621c50d3d","host.name":"43a4683d77f0","os.type":"linux"}} ``` This was reported on https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/18618#issuecomment-1539413938 and happens because [this code path](https://github.com/panta/machineid/blob/c77705a58e0650035e15e10df9d594c9d169dedf/id_linux.go#L48-L53) is taken on the dependency used by #18618 to run this.
process
system detector sets host id to an empty value on containerized setups component s processor resourcedetection what happened description the host id resource attribute is set by the system detector to an empty string value when running on containerized setups steps to reproduce run the collector contrib docker image with the configuration provided below this is also reproducible with custom builds if one has some base images e g alpine expected result the detector works in accordance to which states that emphasis mine unique host id for cloud this must be the instance id assigned by the cloud provider for non containerized systems this should be the machine id see the table below for the sources to use to determine the machine id based on operating system aiui an empty string is not valid since it s not unique and on containerized environments this should not be the machine id since a container is not really a host my expectation would be either that on containerized environments the host id resource attribute is not set or that it is set to a value that persists across restarts actual result on an empty string is set on host id which is not a unique host id on up to a random uuid that varies on each container restart is set on host id collector version environment information environment running the docker image otel opentelemetry collector contrib with the configuration provided below reproduces this opentelemetry collector configuration yaml receivers put a dummy receiver just to generate some metrics hostmetrics collection interval scrapers load processors resourcedetection detectors exporters logging verbosity detailed service pipelines metrics receivers processors exporters log output shell info service telemetry go setting up own telemetry info service telemetry go serving prometheus metrics address level basic info exporter exporter go development component may change in the future kind exporter data type metrics name logging info service service go starting otelcol contrib version numcpu info extensions extensions go starting extensions info internal resourcedetection go began detecting resource information kind processor name resourcedetection pipeline metrics info internal resourcedetection go detected resource information kind processor name resourcedetection pipeline metrics resource host id host name os type linux additional context prior to in particular in between that included and before this would generate a random uuid on each container restart an example run of the same configuration on info service telemetry go setting up own telemetry info service telemetry go serving prometheus metrics address level basic info exporter exporter go development component may change in the future kind exporter data type metrics name logging info service service go starting otelcol contrib version numcpu info extensions extensions go starting extensions info internal resourcedetection go began detecting resource information kind processor name resourcedetection pipeline metrics info internal resourcedetection go detected resource information kind processor name resourcedetection pipeline metrics resource host id host name os type linux this was reported on and happens because is taken on the dependency used by to run this
1
22,328
30,913,675,441
IssuesEvent
2023-08-05 02:34:44
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
pih 1.48036 has 2 GuardDog issues
guarddog typosquatting silent-process-execution
https://pypi.org/project/pih https://inspector.pypi.io/project/pih ```{ "dependency": "pih", "version": "1.48036", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pid, pip", "silent-process-execution": [ { "location": "pih-1.48036/pih/tools.py:781", "code": " result = subprocess.run(command, stdin=subprocess.DEVNULL, stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmp1cvfe72r/pih" } }```
1.0
pih 1.48036 has 2 GuardDog issues - https://pypi.org/project/pih https://inspector.pypi.io/project/pih ```{ "dependency": "pih", "version": "1.48036", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pid, pip", "silent-process-execution": [ { "location": "pih-1.48036/pih/tools.py:781", "code": " result = subprocess.run(command, stdin=subprocess.DEVNULL, stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmp1cvfe72r/pih" } }```
process
pih has guarddog issues dependency pih version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt pid pip silent process execution location pih pih tools py code result subprocess run command stdin subprocess devnull stdout subprocess devnull stderr subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp pih
1
21,234
28,350,152,080
IssuesEvent
2023-04-12 01:34:58
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
--show_task_finish no longer has an implementation
P2 type: process team-OSS stale
### Description of the problem / feature request: I tried using the `--show_task_finish` option with `bazel build --curses=no`, and discovered that it had no effect on the output. I looked in the bazel code, and noticed that while it was recognized as an option, the option no longer had an implementation. The last use of `showTaskFinish` was removed in 3e9eec42f9523bd835d3185f98e5cacc431bb4bd. ### Feature requests: what underlying problem are you trying to solve with this feature? It would be helpful to either implement the option (preferred), or to remove it altogether. ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. Try using `--show_task_finish` ### What operating system are you running Bazel on? Ubuntu 18.04 ### What's the output of `bazel info release`? release 3.1.0
1.0
--show_task_finish no longer has an implementation - ### Description of the problem / feature request: I tried using the `--show_task_finish` option with `bazel build --curses=no`, and discovered that it had no effect on the output. I looked in the bazel code, and noticed that while it was recognized as an option, the option no longer had an implementation. The last use of `showTaskFinish` was removed in 3e9eec42f9523bd835d3185f98e5cacc431bb4bd. ### Feature requests: what underlying problem are you trying to solve with this feature? It would be helpful to either implement the option (preferred), or to remove it altogether. ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. Try using `--show_task_finish` ### What operating system are you running Bazel on? Ubuntu 18.04 ### What's the output of `bazel info release`? release 3.1.0
process
show task finish no longer has an implementation description of the problem feature request i tried using the show task finish option with bazel build curses no and discovered that it had no effect on the output i looked in the bazel code and noticed that while it was recognized as an option the option no longer had an implementation the last use of showtaskfinish was removed in feature requests what underlying problem are you trying to solve with this feature it would be helpful to either implement the option preferred or to remove it altogether bugs what s the simplest easiest way to reproduce this bug please provide a minimal example if possible try using show task finish what operating system are you running bazel on ubuntu what s the output of bazel info release release
1
70,109
18,018,259,207
IssuesEvent
2021-09-16 16:05:21
golang/go
https://api.github.com/repos/golang/go
opened
x/build: ios-arm64-corellium builders have long wait times
Builders NeedsFix
Users have reported long wait times with ios-arm64-corellium builds: ``` • ios-arm64-corellium | running 1h38m44s ``` ``` ios-arm64-corellium rev cfa233d7 (sub-repo mobile rev 855b5ad0) (trybot set for Ib1a2f53); waiting_for_machine; (nil *buildlet.Client), 1h45m36s ago 2021-09-16T14:15:15Z checking_for_snapshot 2021-09-16T14:15:15Z finish_checking_for_snapshot after 35ms 2021-09-16T14:15:15Z get_buildlet +6335.7s (now) ``` `host-ios-arm64-corellium-ios: 2/2 (1 missing)` Perhaps the builders need to be rebooted. @eliasnaur @golang/release
1.0
x/build: ios-arm64-corellium builders have long wait times - Users have reported long wait times with ios-arm64-corellium builds: ``` • ios-arm64-corellium | running 1h38m44s ``` ``` ios-arm64-corellium rev cfa233d7 (sub-repo mobile rev 855b5ad0) (trybot set for Ib1a2f53); waiting_for_machine; (nil *buildlet.Client), 1h45m36s ago 2021-09-16T14:15:15Z checking_for_snapshot 2021-09-16T14:15:15Z finish_checking_for_snapshot after 35ms 2021-09-16T14:15:15Z get_buildlet +6335.7s (now) ``` `host-ios-arm64-corellium-ios: 2/2 (1 missing)` Perhaps the builders need to be rebooted. @eliasnaur @golang/release
non_process
x build ios corellium builders have long wait times users have reported long wait times with ios corellium builds • ios corellium running  ios corellium rev sub repo mobile rev trybot set for waiting for machine nil buildlet client ago checking for snapshot finish checking for snapshot after get buildlet now host ios corellium ios missing perhaps the builders need to be rebooted eliasnaur golang release
0
197,042
14,901,662,478
IssuesEvent
2021-01-21 16:42:45
idaholab/moose
https://api.github.com/repos/idaholab/moose
closed
Add option to show the longest running jobs in TestHarness
C: TestHarness P: normal T: task
## Reason To keep test timing to a minimum, it would be useful to print out the longest running jobs at the end of execution in the test harness. ## Design Add a `--longest-jobs` options to TestHarness that will print the longest running jobs at the end. ## Impact Better management of test timing.
1.0
Add option to show the longest running jobs in TestHarness - ## Reason To keep test timing to a minimum, it would be useful to print out the longest running jobs at the end of execution in the test harness. ## Design Add a `--longest-jobs` options to TestHarness that will print the longest running jobs at the end. ## Impact Better management of test timing.
non_process
add option to show the longest running jobs in testharness reason to keep test timing to a minimum it would be useful to print out the longest running jobs at the end of execution in the test harness design add a longest jobs options to testharness that will print the longest running jobs at the end impact better management of test timing
0
222,584
24,709,177,726
IssuesEvent
2022-10-19 22:08:18
lukebrogan-mend/Umbraco-CMS
https://api.github.com/repos/lukebrogan-mend/Umbraco-CMS
closed
CVE-2020-28481 (Medium) detected in socket.io-2.1.1.tgz - autoclosed
security vulnerability
## CVE-2020-28481 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>socket.io-2.1.1.tgz</b></p></summary> <p>node.js realtime framework server</p> <p>Library home page: <a href="https://registry.npmjs.org/socket.io/-/socket.io-2.1.1.tgz">https://registry.npmjs.org/socket.io/-/socket.io-2.1.1.tgz</a></p> <p>Path to dependency file: /src/Umbraco.Web.UI.Client/package.json</p> <p>Path to vulnerable library: /src/Umbraco.Web.UI.Client/node_modules/socket.io/package.json</p> <p> Dependency Hierarchy: - karma-4.4.1.tgz (Root Library) - :x: **socket.io-2.1.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/Umbraco-CMS/commit/24bd18757bbbe3324d85424fdabf1d6bdaf1695e">24bd18757bbbe3324d85424fdabf1d6bdaf1695e</a></p> <p>Found in base branch: <b>v8/contrib</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package socket.io before 2.4.0 are vulnerable to Insecure Defaults due to CORS Misconfiguration. All domains are whitelisted by default. <p>Publish Date: 2021-01-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28481>CVE-2020-28481</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28481">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28481</a></p> <p>Release Date: 2021-01-19</p> <p>Fix Resolution (socket.io): 2.4.0</p> <p>Direct dependency fix Resolution (karma): 5.0.8</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2020-28481 (Medium) detected in socket.io-2.1.1.tgz - autoclosed - ## CVE-2020-28481 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>socket.io-2.1.1.tgz</b></p></summary> <p>node.js realtime framework server</p> <p>Library home page: <a href="https://registry.npmjs.org/socket.io/-/socket.io-2.1.1.tgz">https://registry.npmjs.org/socket.io/-/socket.io-2.1.1.tgz</a></p> <p>Path to dependency file: /src/Umbraco.Web.UI.Client/package.json</p> <p>Path to vulnerable library: /src/Umbraco.Web.UI.Client/node_modules/socket.io/package.json</p> <p> Dependency Hierarchy: - karma-4.4.1.tgz (Root Library) - :x: **socket.io-2.1.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/Umbraco-CMS/commit/24bd18757bbbe3324d85424fdabf1d6bdaf1695e">24bd18757bbbe3324d85424fdabf1d6bdaf1695e</a></p> <p>Found in base branch: <b>v8/contrib</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package socket.io before 2.4.0 are vulnerable to Insecure Defaults due to CORS Misconfiguration. All domains are whitelisted by default. <p>Publish Date: 2021-01-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28481>CVE-2020-28481</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28481">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28481</a></p> <p>Release Date: 2021-01-19</p> <p>Fix Resolution (socket.io): 2.4.0</p> <p>Direct dependency fix Resolution (karma): 5.0.8</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_process
cve medium detected in socket io tgz autoclosed cve medium severity vulnerability vulnerable library socket io tgz node js realtime framework server library home page a href path to dependency file src umbraco web ui client package json path to vulnerable library src umbraco web ui client node modules socket io package json dependency hierarchy karma tgz root library x socket io tgz vulnerable library found in head commit a href found in base branch contrib vulnerability details the package socket io before are vulnerable to insecure defaults due to cors misconfiguration all domains are whitelisted by default publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution socket io direct dependency fix resolution karma rescue worker helmet automatic remediation is available for this issue
0
4,291
7,191,972,477
IssuesEvent
2018-02-02 23:27:37
amaster507/ifbmt
https://api.github.com/repos/amaster507/ifbmt
closed
Call Log with Reminders
enhancement idea process
@jrwilliams27 A feature that would be very useful would be a way to record a call/contact log for calls/emails/contact to a church - apart from the notes or integrated into the notes. The date/time could be automatically recorded (or changed), a note made, and a scheduled time to contact the church again could be made. A reminder option could give options to notify at a certain date/time, and then when that time is due, the reminder notifies you of the contact(s) needed to be made, filter out the contact(s), that list would be generated, and then one can work through the contact(s).
1.0
Call Log with Reminders - @jrwilliams27 A feature that would be very useful would be a way to record a call/contact log for calls/emails/contact to a church - apart from the notes or integrated into the notes. The date/time could be automatically recorded (or changed), a note made, and a scheduled time to contact the church again could be made. A reminder option could give options to notify at a certain date/time, and then when that time is due, the reminder notifies you of the contact(s) needed to be made, filter out the contact(s), that list would be generated, and then one can work through the contact(s).
process
call log with reminders a feature that would be very useful would be a way to record a call contact log for calls emails contact to a church apart from the notes or integrated into the notes the date time could be automatically recorded or changed a note made and a scheduled time to contact the church again could be made a reminder option could give options to notify at a certain date time and then when that time is due the reminder notifies you of the contact s needed to be made filter out the contact s that list would be generated and then one can work through the contact s
1
15,759
27,848,810,888
IssuesEvent
2023-03-20 17:11:39
dgidb/dgidb-v5
https://api.github.com/repos/dgidb/dgidb-v5
closed
Interaction Record Page
requirement frontend priority:high
We have individual record pages for genes and drugs. We need to implement this for interactions. Prod example: https://www.dgidb.org/interactions/b8e5098d-4915-4c3f-974a-8a0cf9b97bd6#_summary
1.0
Interaction Record Page - We have individual record pages for genes and drugs. We need to implement this for interactions. Prod example: https://www.dgidb.org/interactions/b8e5098d-4915-4c3f-974a-8a0cf9b97bd6#_summary
non_process
interaction record page we have individual record pages for genes and drugs we need to implement this for interactions prod example
0
22,315
30,870,180,952
IssuesEvent
2023-08-03 10:45:40
benthosdev/benthos
https://api.github.com/repos/benthosdev/benthos
closed
Bloblang interpolation in http jwt claims
enhancement question processors
Hi, I was trying to use http processor with jwt enabled and I need to set the "iat" claim with current timestamp using this code: ``` jwt: enabled: true private_key_file: private.pem signing_method: RS256 claims: iss: iotbuilder sub: benthos iat: ${! timestamp_unix() * 1000 } ``` but on the server the decoded JWT is: ``` { iat: '${! timestamp_unix() * 1000 }', iss: 'iotbuilder', sub: 'benthos' } ``` The iat claim is not interpolated.
1.0
Bloblang interpolation in http jwt claims - Hi, I was trying to use http processor with jwt enabled and I need to set the "iat" claim with current timestamp using this code: ``` jwt: enabled: true private_key_file: private.pem signing_method: RS256 claims: iss: iotbuilder sub: benthos iat: ${! timestamp_unix() * 1000 } ``` but on the server the decoded JWT is: ``` { iat: '${! timestamp_unix() * 1000 }', iss: 'iotbuilder', sub: 'benthos' } ``` The iat claim is not interpolated.
process
bloblang interpolation in http jwt claims hi i was trying to use http processor with jwt enabled and i need to set the iat claim with current timestamp using this code jwt enabled true private key file private pem signing method claims iss iotbuilder sub benthos iat timestamp unix but on the server the decoded jwt is iat timestamp unix iss iotbuilder sub benthos the iat claim is not interpolated
1
31,663
6,583,086,205
IssuesEvent
2017-09-13 03:07:00
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
[client] networkconfig address validation missing
Team: Client Type: Defect
The following program runs fine even though no valid address has been added. ``` public class Main { public static void main(String[] args){ HazelcastInstance server = Hazelcast.newHazelcastInstance(); ClientConfig clientConfig = new ClientConfig(); clientConfig.getNetworkConfig().addAddress(""); HazelcastInstance client = HazelcastClient.newHazelcastClient(clientConfig); System.out.println(client.getAtomicLong("foo").incrementAndGet()); } } ``` So we violate fail-fast. We should check if the passed address is a valid ip address or hostname.
1.0
[client] networkconfig address validation missing - The following program runs fine even though no valid address has been added. ``` public class Main { public static void main(String[] args){ HazelcastInstance server = Hazelcast.newHazelcastInstance(); ClientConfig clientConfig = new ClientConfig(); clientConfig.getNetworkConfig().addAddress(""); HazelcastInstance client = HazelcastClient.newHazelcastClient(clientConfig); System.out.println(client.getAtomicLong("foo").incrementAndGet()); } } ``` So we violate fail-fast. We should check if the passed address is a valid ip address or hostname.
non_process
networkconfig address validation missing the following program runs fine even though no valid address has been added public class main public static void main string args hazelcastinstance server hazelcast newhazelcastinstance clientconfig clientconfig new clientconfig clientconfig getnetworkconfig addaddress hazelcastinstance client hazelcastclient newhazelcastclient clientconfig system out println client getatomiclong foo incrementandget so we violate fail fast we should check if the passed address is a valid ip address or hostname
0
230,822
17,647,444,025
IssuesEvent
2021-08-20 08:24:47
sertit/eoreader
https://api.github.com/repos/sertit/eoreader
closed
DOC: Updates
documentation
It could be cool to create one page for every supported sensor. This could be useful to: - Show what product type is not used for a specific sensor - Warn the user about some specificity - ... [EDIT]: Not really useful until a user needs it
1.0
DOC: Updates - It could be cool to create one page for every supported sensor. This could be useful to: - Show what product type is not used for a specific sensor - Warn the user about some specificity - ... [EDIT]: Not really useful until a user needs it
non_process
doc updates it could be cool to create one page for every supported sensor this could be useful to show what product type is not used for a specific sensor warn the user about some specificity not really useful until a user needs it
0
245,666
7,889,254,256
IssuesEvent
2018-06-28 02:55:33
ThePoptartCrpr/DawnOfMan
https://api.github.com/repos/ThePoptartCrpr/DawnOfMan
closed
Inventory icon overlap with Nutrition
category: compatibility priority: high type: bug
# Description <!-- Please describe your issue in detail below. --> Nutrition's icon in the inventory overlaps the sort icon added by Quark, as seen below. This can be fixed by modifying the position of Nutrition's icon in the inventory via configs. # Evidence <!-- Please attach any useful evidence below, i.e. logs or screenshots --> ![2018-06-27_17 09 47](https://user-images.githubusercontent.com/25708063/42006358-2c151008-7a2e-11e8-83c9-7cf7a3de62c5.png)
1.0
Inventory icon overlap with Nutrition - # Description <!-- Please describe your issue in detail below. --> Nutrition's icon in the inventory overlaps the sort icon added by Quark, as seen below. This can be fixed by modifying the position of Nutrition's icon in the inventory via configs. # Evidence <!-- Please attach any useful evidence below, i.e. logs or screenshots --> ![2018-06-27_17 09 47](https://user-images.githubusercontent.com/25708063/42006358-2c151008-7a2e-11e8-83c9-7cf7a3de62c5.png)
non_process
inventory icon overlap with nutrition description nutrition s icon in the inventory overlaps the sort icon added by quark as seen below this can be fixed by modifying the position of nutrition s icon in the inventory via configs evidence
0
12,928
15,296,184,558
IssuesEvent
2021-02-24 06:23:45
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
closed
Let NaNs pass through in OrdinalEncoder
Hard New Feature module:preprocessing
We should allow NaNs to pass-through in `OrdinalEncoder`. One reason is for supporting categorical features with NaNs in the `HistGradientBoosting ` estimators: For the native categorical support, we require categories to be encoded as done by `OrdinalEncoder`. Yet, `OrdinalEncoder` is unable to be fitted with NaNs. So we can't use the native categorical support of the Hist-GBDT if the categorical features have NaNs... Which is ironic because we do support NaNs as a native category in the Hist-GBDT code. In terms of API, we'd need a new `handle_missing={'error', 'passthrough'}` parameter which would default to 'error'. CC @ogrisel @lorentzenchr @thomasjpfan (Note: This is a subset of the feature request from https://github.com/scikit-learn/scikit-learn/issues/17123) (Marking as "Hard" because anything encoding-related is usually a headache ;) )
1.0
Let NaNs pass through in OrdinalEncoder - We should allow NaNs to pass-through in `OrdinalEncoder`. One reason is for supporting categorical features with NaNs in the `HistGradientBoosting ` estimators: For the native categorical support, we require categories to be encoded as done by `OrdinalEncoder`. Yet, `OrdinalEncoder` is unable to be fitted with NaNs. So we can't use the native categorical support of the Hist-GBDT if the categorical features have NaNs... Which is ironic because we do support NaNs as a native category in the Hist-GBDT code. In terms of API, we'd need a new `handle_missing={'error', 'passthrough'}` parameter which would default to 'error'. CC @ogrisel @lorentzenchr @thomasjpfan (Note: This is a subset of the feature request from https://github.com/scikit-learn/scikit-learn/issues/17123) (Marking as "Hard" because anything encoding-related is usually a headache ;) )
process
let nans pass through in ordinalencoder we should allow nans to pass through in ordinalencoder one reason is for supporting categorical features with nans in the histgradientboosting estimators for the native categorical support we require categories to be encoded as done by ordinalencoder yet ordinalencoder is unable to be fitted with nans so we can t use the native categorical support of the hist gbdt if the categorical features have nans which is ironic because we do support nans as a native category in the hist gbdt code in terms of api we d need a new handle missing error passthrough parameter which would default to error cc ogrisel lorentzenchr thomasjpfan note this is a subset of the feature request from marking as hard because anything encoding related is usually a headache
1
3,684
6,714,762,424
IssuesEvent
2017-10-13 18:14:42
syndesisio/syndesis
https://api.github.com/repos/syndesisio/syndesis
opened
Feature toggle
process/retro
Discussion: Should we use feature toggles? Actually, I think that's a very good thing as it allows to avoid branching (which I really would love to avoid as long as possible). This means we add new features guarded by toggles which can be enabled / disabled via configuration. That way we can deliver path release directly from `master`. My only concern here is, how to tackle heavy refactorings like changing our domain model or fundamental API changes (until we get to a proper versioned API ?)
1.0
Feature toggle - Discussion: Should we use feature toggles? Actually, I think that's a very good thing as it allows to avoid branching (which I really would love to avoid as long as possible). This means we add new features guarded by toggles which can be enabled / disabled via configuration. That way we can deliver path release directly from `master`. My only concern here is, how to tackle heavy refactorings like changing our domain model or fundamental API changes (until we get to a proper versioned API ?)
process
feature toggle discussion should we use feature toggles actually i think that s a very good thing as it allows to avoid branching which i really would love to avoid as long as possible this means we add new features guarded by toggles which can be enabled disabled via configuration that way we can deliver path release directly from master my only concern here is how to tackle heavy refactorings like changing our domain model or fundamental api changes until we get to a proper versioned api
1
17,912
23,899,873,068
IssuesEvent
2022-09-08 17:45:09
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Bump local node for services 0.30.0-alpha.0 and mirror 0.64.0-beta1
enhancement P2 process
### Problem The services 0.30.0-alpha.0 and mirror 0.64.0-beta1 versions are available ### Solution Update the index.spec.ts with ``` process.env['NETWORK_NODE_IMAGE_TAG'] = '0.30.0-alpha.0'; process.env['HAVEGED_IMAGE_TAG'] = '0.30.0-alpha.0'; process.env['MIRROR_IMAGE_TAG'] = '0.64.0-beta1'; ``` ### Alternatives _No response_
1.0
Bump local node for services 0.30.0-alpha.0 and mirror 0.64.0-beta1 - ### Problem The services 0.30.0-alpha.0 and mirror 0.64.0-beta1 versions are available ### Solution Update the index.spec.ts with ``` process.env['NETWORK_NODE_IMAGE_TAG'] = '0.30.0-alpha.0'; process.env['HAVEGED_IMAGE_TAG'] = '0.30.0-alpha.0'; process.env['MIRROR_IMAGE_TAG'] = '0.64.0-beta1'; ``` ### Alternatives _No response_
process
bump local node for services alpha and mirror problem the services alpha and mirror versions are available solution update the index spec ts with process env alpha process env alpha process env alternatives no response
1
290,726
25,090,428,863
IssuesEvent
2022-11-08 05:31:34
crispindeity/issue-tracker
https://api.github.com/repos/crispindeity/issue-tracker
closed
Add Label Integration Test
📬 API BE ✅ Test
# Description - Label Integration Test 코드 작성 - [x] Long save() - [x] ResponseLablesDto findAll() - [x] void delete() - [x] Long update()
1.0
Add Label Integration Test - # Description - Label Integration Test 코드 작성 - [x] Long save() - [x] ResponseLablesDto findAll() - [x] void delete() - [x] Long update()
non_process
add label integration test description label integration test 코드 작성 long save responselablesdto findall void delete long update
0
21,781
30,294,584,378
IssuesEvent
2023-07-09 17:45:27
The-Data-Alchemists-Manipal/MindWave
https://api.github.com/repos/The-Data-Alchemists-Manipal/MindWave
closed
Counting People in a frame using OpenCV
image-processing
### Is your feature request related to a problem? Please describe. There are various occasions in which we need to study an image or video. In doing so, first thing we capture is people. What if we can count them automatically coming in and out of a video. It doesn't matter if they are completely in the frame, but any part of there body tells they are present in the frame. ### Describe the solution you'd like To do so, i would use OpenCV to count the number of people in a frame, using face, body, legs analysis. ### Describe alternatives you've considered _No response_ ### Additional context Under GSSOC'23 please assign me this task. ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
Counting People in a frame using OpenCV - ### Is your feature request related to a problem? Please describe. There are various occasions in which we need to study an image or video. In doing so, first thing we capture is people. What if we can count them automatically coming in and out of a video. It doesn't matter if they are completely in the frame, but any part of there body tells they are present in the frame. ### Describe the solution you'd like To do so, i would use OpenCV to count the number of people in a frame, using face, body, legs analysis. ### Describe alternatives you've considered _No response_ ### Additional context Under GSSOC'23 please assign me this task. ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
process
counting people in a frame using opencv is your feature request related to a problem please describe there are various occasions in which we need to study an image or video in doing so first thing we capture is people what if we can count them automatically coming in and out of a video it doesn t matter if they are completely in the frame but any part of there body tells they are present in the frame describe the solution you d like to do so i would use opencv to count the number of people in a frame using face body legs analysis describe alternatives you ve considered no response additional context under gssoc please assign me this task code of conduct i agree to follow this project s code of conduct
1
276,688
30,519,959,142
IssuesEvent
2023-07-19 07:22:16
elastic/integrations
https://api.github.com/repos/elastic/integrations
closed
[O365] Migrate to CEL input
Team:Security-External Integrations Integration:o365 8.8 candidate 8.9 candidate
Our O365 integration relies on the[ O365 Management API input](https://www.elastic.co/guide/en/beats/filebeat/current/filebeat-input-o365audit.html), which is still under a Beta flag. O365 is the only integration leveraging this input and we should look at migrating to the httpjson or cel input as a more suitable alternative. This would avoid us having to maintain and support the o365 input for one integration, and bring the integration inline with our other API based integration. **Questions** 1. Are there other integrations using the O365 Management API input? 2. Does moving to httpjson allos us to address https://github.com/elastic/integrations/issues/1868
True
[O365] Migrate to CEL input - Our O365 integration relies on the[ O365 Management API input](https://www.elastic.co/guide/en/beats/filebeat/current/filebeat-input-o365audit.html), which is still under a Beta flag. O365 is the only integration leveraging this input and we should look at migrating to the httpjson or cel input as a more suitable alternative. This would avoid us having to maintain and support the o365 input for one integration, and bring the integration inline with our other API based integration. **Questions** 1. Are there other integrations using the O365 Management API input? 2. Does moving to httpjson allos us to address https://github.com/elastic/integrations/issues/1868
non_process
migrate to cel input our integration relies on the which is still under a beta flag is the only integration leveraging this input and we should look at migrating to the httpjson or cel input as a more suitable alternative this would avoid us having to maintain and support the input for one integration and bring the integration inline with our other api based integration questions are there other integrations using the management api input does moving to httpjson allos us to address
0
213,682
24,016,302,174
IssuesEvent
2022-09-15 01:16:22
Baneeishaque/locations
https://api.github.com/repos/Baneeishaque/locations
opened
WS-2020-0042 (High) detected in acorn-5.7.4.tgz
security vulnerability
## WS-2020-0042 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.4.tgz</b></p></summary> <p>ECMAScript parser</p> <p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/acorn/package.json</p> <p> Dependency Hierarchy: - jest-expo-27.0.1.tgz (Root Library) - jest-22.4.4.tgz - jest-cli-22.4.4.tgz - jest-environment-jsdom-22.4.3.tgz - jsdom-11.12.0.tgz - :x: **acorn-5.7.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Baneeishaque/locations/commit/b8d5a4143a4e0caa970dcb555e1daf11e1ebbc43">b8d5a4143a4e0caa970dcb555e1daf11e1ebbc43</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser. <p>Publish Date: 2020-03-01 <p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p> <p>Release Date: 2020-03-01</p> <p>Fix Resolution (acorn): 6.4.1</p> <p>Direct dependency fix Resolution (jest-expo): 38.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2020-0042 (High) detected in acorn-5.7.4.tgz - ## WS-2020-0042 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.4.tgz</b></p></summary> <p>ECMAScript parser</p> <p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/acorn/package.json</p> <p> Dependency Hierarchy: - jest-expo-27.0.1.tgz (Root Library) - jest-22.4.4.tgz - jest-cli-22.4.4.tgz - jest-environment-jsdom-22.4.3.tgz - jsdom-11.12.0.tgz - :x: **acorn-5.7.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Baneeishaque/locations/commit/b8d5a4143a4e0caa970dcb555e1daf11e1ebbc43">b8d5a4143a4e0caa970dcb555e1daf11e1ebbc43</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser. <p>Publish Date: 2020-03-01 <p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p> <p>Release Date: 2020-03-01</p> <p>Fix Resolution (acorn): 6.4.1</p> <p>Direct dependency fix Resolution (jest-expo): 38.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws high detected in acorn tgz ws high severity vulnerability vulnerable library acorn tgz ecmascript parser library home page a href path to dependency file package json path to vulnerable library node modules acorn package json dependency hierarchy jest expo tgz root library jest tgz jest cli tgz jest environment jsdom tgz jsdom tgz x acorn tgz vulnerable library found in head commit a href vulnerability details acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution acorn direct dependency fix resolution jest expo step up your open source security game with mend
0
1,135
3,619,369,370
IssuesEvent
2016-02-08 15:47:53
NuCivic/dkan
https://api.github.com/repos/NuCivic/dkan
closed
Invalid argument supplied for foreach() entity_rdf.module:49
bug Component: Viz Entity Processed/JIRA
Getting that notice from Viz Entity module operations.
1.0
Invalid argument supplied for foreach() entity_rdf.module:49 - Getting that notice from Viz Entity module operations.
process
invalid argument supplied for foreach entity rdf module getting that notice from viz entity module operations
1
259,684
8,198,925,032
IssuesEvent
2018-08-31 18:10:21
johannbrehmer/madminer
https://api.github.com/repos/johannbrehmer/madminer
closed
Load MadMiner settings to GoldMine
enhancement high priority
Add a function that loads the parameter, benchmark, and morphing setup from an HDF5 file to the GoldMine class.
1.0
Load MadMiner settings to GoldMine - Add a function that loads the parameter, benchmark, and morphing setup from an HDF5 file to the GoldMine class.
non_process
load madminer settings to goldmine add a function that loads the parameter benchmark and morphing setup from an file to the goldmine class
0
5,715
8,567,917,329
IssuesEvent
2018-11-10 16:33:55
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
Extend getTokenInfo to accept other queries
libs-tokenlib status-inprocess type-enhancement
Currently, getTokenInfo call in the file getTokenBal.cpp only accepts a value of "balance". It would be easy to extend that to accept many other queries on standard tokens.
1.0
Extend getTokenInfo to accept other queries - Currently, getTokenInfo call in the file getTokenBal.cpp only accepts a value of "balance". It would be easy to extend that to accept many other queries on standard tokens.
process
extend gettokeninfo to accept other queries currently gettokeninfo call in the file gettokenbal cpp only accepts a value of balance it would be easy to extend that to accept many other queries on standard tokens
1
3,552
6,587,491,179
IssuesEvent
2017-09-13 21:17:36
cliffparnitzky/ApparelManager
https://api.github.com/repos/cliffparnitzky/ApparelManager
closed
Add notification tokens for BE user and order creator
Comp - notification Improvement ⚙ - Processed
Add notification tokens for - [ ] BE user - [ ] order creator
1.0
Add notification tokens for BE user and order creator - Add notification tokens for - [ ] BE user - [ ] order creator
process
add notification tokens for be user and order creator add notification tokens for be user order creator
1
226,471
18,020,514,707
IssuesEvent
2021-09-16 18:47:28
arfc/moltres
https://api.github.com/repos/arfc/moltres
closed
Fix ManuHX Jacobian term, allow heat transfer coeff to vary with time, & make a test for ManuHX
Comp:Core Difficulty:2-Challenging Priority:1-Critical Status:4-In Progress Type:Bug Type:Feature Type:Test
This issue can be closed when the following three objectives involving `ManuHX` are met: 1. Fix the Jacobian term. It is currently missing a `_phi[_j][_qp]` term. 2. Allow the heat transfer coeff `htc` to vary with time by making it `controllable` under MOOSE's `Control` system. This allows us to model a loss of heat sink accident when using `ManuHX` as a heat exchanger 3. Make a test for `ManuHX`. It is currently not covered under any existing tests. 4. (optional) Come up with a more intuitive/mnemonic name for `ManuHX` e.g. `VolumetricHeatSink` or `ConvectiveHeatSink`.
1.0
Fix ManuHX Jacobian term, allow heat transfer coeff to vary with time, & make a test for ManuHX - This issue can be closed when the following three objectives involving `ManuHX` are met: 1. Fix the Jacobian term. It is currently missing a `_phi[_j][_qp]` term. 2. Allow the heat transfer coeff `htc` to vary with time by making it `controllable` under MOOSE's `Control` system. This allows us to model a loss of heat sink accident when using `ManuHX` as a heat exchanger 3. Make a test for `ManuHX`. It is currently not covered under any existing tests. 4. (optional) Come up with a more intuitive/mnemonic name for `ManuHX` e.g. `VolumetricHeatSink` or `ConvectiveHeatSink`.
non_process
fix manuhx jacobian term allow heat transfer coeff to vary with time make a test for manuhx this issue can be closed when the following three objectives involving manuhx are met fix the jacobian term it is currently missing a phi term allow the heat transfer coeff htc to vary with time by making it controllable under moose s control system this allows us to model a loss of heat sink accident when using manuhx as a heat exchanger make a test for manuhx it is currently not covered under any existing tests optional come up with a more intuitive mnemonic name for manuhx e g volumetricheatsink or convectiveheatsink
0
1,636
4,258,072,805
IssuesEvent
2016-07-11 03:54:44
triplea-game/triplea
https://api.github.com/repos/triplea-game/triplea
opened
1.9.0.0 release!
High Priority Process
Let's begin the release process! Is everything in that we would want for the 1.9.0.0 release? I started this source forge thread to let folks know about the prerelease: http://tripleadev.1671093.n2.nabble.com/TripleA-1-9-0-0-Prerelease-td7592899.html For now, I would like to nominate version 1.9.0.0.2461 for the 1.9.0.0 release. My thoughts are we will test that specific version and see if we find any issues that we would want to fix. If so, we'll roll forward to the next latest version and repeat until we are happy. Ideally this will not involve any delays or pauses in normal development. After a week of testing or so, we will then call the release official and begin to migrate folks over. As is, we need to determine how to migrate Derby DB data as well to a new lobby instance, I'm not sure right now how to do that. Otherwise let's use this thread to discuss the next release, process and progress.
1.0
1.9.0.0 release! - Let's begin the release process! Is everything in that we would want for the 1.9.0.0 release? I started this source forge thread to let folks know about the prerelease: http://tripleadev.1671093.n2.nabble.com/TripleA-1-9-0-0-Prerelease-td7592899.html For now, I would like to nominate version 1.9.0.0.2461 for the 1.9.0.0 release. My thoughts are we will test that specific version and see if we find any issues that we would want to fix. If so, we'll roll forward to the next latest version and repeat until we are happy. Ideally this will not involve any delays or pauses in normal development. After a week of testing or so, we will then call the release official and begin to migrate folks over. As is, we need to determine how to migrate Derby DB data as well to a new lobby instance, I'm not sure right now how to do that. Otherwise let's use this thread to discuss the next release, process and progress.
process
release let s begin the release process is everything in that we would want for the release i started this source forge thread to let folks know about the prerelease for now i would like to nominate version for the release my thoughts are we will test that specific version and see if we find any issues that we would want to fix if so we ll roll forward to the next latest version and repeat until we are happy ideally this will not involve any delays or pauses in normal development after a week of testing or so we will then call the release official and begin to migrate folks over as is we need to determine how to migrate derby db data as well to a new lobby instance i m not sure right now how to do that otherwise let s use this thread to discuss the next release process and progress
1
8,552
2,611,517,501
IssuesEvent
2015-02-27 05:52:12
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Fort mode does not work
auto-migrated Priority-Medium Type-Defect
``` When starting a fort game, it does not load and returns to game options menu(team select, map and mode select etc.) ``` Original issue reported on code.google.com by `pako.a...@gmail.com` on 12 Apr 2013 at 5:28
1.0
Fort mode does not work - ``` When starting a fort game, it does not load and returns to game options menu(team select, map and mode select etc.) ``` Original issue reported on code.google.com by `pako.a...@gmail.com` on 12 Apr 2013 at 5:28
non_process
fort mode does not work when starting a fort game it does not load and returns to game options menu team select map and mode select etc original issue reported on code google com by pako a gmail com on apr at
0
15,508
19,703,265,794
IssuesEvent
2022-01-12 18:52:15
googleapis/java-channel
https://api.github.com/repos/googleapis/java-channel
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'channel' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'channel' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname channel invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
1
19,549
25,867,519,081
IssuesEvent
2022-12-13 22:21:16
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
X86 disassembler fails to recognize FWAIT and FDISI
Feature: Processor/x86 Status: Internal
**Describe the bug** In some situations the ghidra disassembler fails to recognize certain floating point instructions. **To Reproduce** Steps to reproduce the behavior: 1. If you disassemble the following binary blob (from the OpenWatcom C library - function __FPE2Handler) [50 53 51 52 56 57 55 1e 06 83 ec 1c 89 e5 d9 75 00 9b 9b db e2 9b db e1 fb e8 92 4b fb] you will get this: 00109ac8 50 PUSH EAX 00109ac9 53 PUSH EBX 00109aca 51 PUSH ECX 00109acb 52 PUSH EDX 00109acc 56 PUSH ESI 00109acd 57 PUSH EDI 00109ace 55 PUSH EBP 00109acf 1e PUSH DS 00109ad0 06 PUSH ES 00109ad1 83 ec 1c SUB ESP,0x1c 00109ad4 89 e5 MOV EBP,ESP 00109ad6 d9 75 00 FNSTENV [EBP]=>local_40 00109ad9 9b WAIT 00109ada 9b db e2 FCLEX 00109add 9b WAIT 00109ade db ?? DBh 00109adf e1 ?? E1h 00109ae0 fb STI with an error bookmark at 00109ade saying "Error [Bad Instruction]: Unable to resolve constructor at 00109ade (flow from 00109add)" **Expected behavior** 00109ad9 through 00109adf are not disassembled correctly here. If you look at the original assembly listing for the function in the OpenWatcom sources (https://github.com/open-watcom/open-watcom-v2/blob/6b9cb4438959d845a2b4dd4bf19f7908ecd608e9/bld/clib/math/a/fpeinth.asm) you'll see that they should be disassembled as: FWAIT FDISI **Screenshots** ![image](https://user-images.githubusercontent.com/33301226/169338962-f4ff5c4f-4f52-46c0-be2d-2b470f4fea8b.png) ![image](https://user-images.githubusercontent.com/33301226/169339117-2b3b2300-0879-42d4-8d9a-dc1b63f03dd2.png) **Attachments** **Environment (please complete the following information):** - OS: Windows 10 Pro 20H2 - Java Version: 12.0.1 - Ghidra Version: 10.1 - Ghidra Origin: GitHub releases **Additional context**
1.0
X86 disassembler fails to recognize FWAIT and FDISI - **Describe the bug** In some situations the ghidra disassembler fails to recognize certain floating point instructions. **To Reproduce** Steps to reproduce the behavior: 1. If you disassemble the following binary blob (from the OpenWatcom C library - function __FPE2Handler) [50 53 51 52 56 57 55 1e 06 83 ec 1c 89 e5 d9 75 00 9b 9b db e2 9b db e1 fb e8 92 4b fb] you will get this: 00109ac8 50 PUSH EAX 00109ac9 53 PUSH EBX 00109aca 51 PUSH ECX 00109acb 52 PUSH EDX 00109acc 56 PUSH ESI 00109acd 57 PUSH EDI 00109ace 55 PUSH EBP 00109acf 1e PUSH DS 00109ad0 06 PUSH ES 00109ad1 83 ec 1c SUB ESP,0x1c 00109ad4 89 e5 MOV EBP,ESP 00109ad6 d9 75 00 FNSTENV [EBP]=>local_40 00109ad9 9b WAIT 00109ada 9b db e2 FCLEX 00109add 9b WAIT 00109ade db ?? DBh 00109adf e1 ?? E1h 00109ae0 fb STI with an error bookmark at 00109ade saying "Error [Bad Instruction]: Unable to resolve constructor at 00109ade (flow from 00109add)" **Expected behavior** 00109ad9 through 00109adf are not disassembled correctly here. If you look at the original assembly listing for the function in the OpenWatcom sources (https://github.com/open-watcom/open-watcom-v2/blob/6b9cb4438959d845a2b4dd4bf19f7908ecd608e9/bld/clib/math/a/fpeinth.asm) you'll see that they should be disassembled as: FWAIT FDISI **Screenshots** ![image](https://user-images.githubusercontent.com/33301226/169338962-f4ff5c4f-4f52-46c0-be2d-2b470f4fea8b.png) ![image](https://user-images.githubusercontent.com/33301226/169339117-2b3b2300-0879-42d4-8d9a-dc1b63f03dd2.png) **Attachments** **Environment (please complete the following information):** - OS: Windows 10 Pro 20H2 - Java Version: 12.0.1 - Ghidra Version: 10.1 - Ghidra Origin: GitHub releases **Additional context**
process
disassembler fails to recognize fwait and fdisi describe the bug in some situations the ghidra disassembler fails to recognize certain floating point instructions to reproduce steps to reproduce the behavior if you disassemble the following binary blob from the openwatcom c library function you will get this push eax push ebx push ecx push edx push esi push edi push ebp push ds push es ec sub esp mov ebp esp fnstenv local wait db fclex wait db dbh fb sti with an error bookmark at saying error unable to resolve constructor at flow from expected behavior through are not disassembled correctly here if you look at the original assembly listing for the function in the openwatcom sources you ll see that they should be disassembled as fwait fdisi screenshots attachments environment please complete the following information os windows pro java version ghidra version ghidra origin github releases additional context
1
324,007
27,769,205,724
IssuesEvent
2023-03-16 13:29:21
matrix-org/synapse
https://api.github.com/repos/matrix-org/synapse
closed
Complement `TestPartialStateJoin/Outgoing_device_list_updates/Device_list_updates_reach_newly_joined_servers_in_partial_state_rooms` is flakey
A-Federated-Join A-Device-List-Tracking S-Major T-Task Z-Flake Z-Dev-Wishlist O-Frequent A-Testing
https://github.com/matrix-org/synapse/actions/runs/3172224388/jobs/5166469780#step:4:2964 ``` client.go:604: [CSAPI] GET hs1/_matrix/client/v3/sync => 200 OK (125.434036ms) federation_room_join_partial_state_test.go:1674: @t24alice:hs1's partial state join to !0:host.docker.internal:38237 completed. client.go:604: [CSAPI] PUT hs1/_matrix/client/v3/devices/PBJFBBTZIN => 200 OK (18.65378ms) federation_room_join_partial_state_test.go:1677: @t24alice:hs1 sent device list update. federation_room_join_partial_state_test.go:1534: Complement server received m.device_list_update: {"device_display_name":"A new device name 3","device_id":"PBJFBBTZIN","org.matrix.opentracing_context":"{}","prev_id":[35],"stream_id":38,"user_id":"@t24alice:hs1"} federation_room_join_partial_state_test.go:1534: Complement server received m.device_list_update: {"device_display_name":"A new device name 3","device_id":"PBJFBBTZIN","org.matrix.opentracing_context":"{}","prev_id":[36],"stream_id":40,"user_id":"@t24alice:hs1"} federation_room_join_partial_state_test.go:1680: @charlie, @derek and @elsie received device list update. federation_room_join_partial_state_test.go:1534: Complement server received m.device_list_update: {"device_display_name":"A new device name 3","device_id":"PBJFBBTZIN","org.matrix.opentracing_context":"{}","prev_id":[38],"stream_id":39,"user_id":"@t24alice:hs1"} panic: test timed out after 1h0m0s goroutine 3428 [running]: testing.(*M).startAlarm.func1() /opt/hostedtoolcache/go/1.17.13/x64/src/testing/testing.go:1788 +0x8e created by time.goFunc /opt/hostedtoolcache/go/1.17.13/x64/src/time/sleep.go:180 +0x31 ``` Synapse sent an extra device update, which the test wasn't interested in. I think putting the device update in a channel blocked, and then the test timed out after an hour.
1.0
Complement `TestPartialStateJoin/Outgoing_device_list_updates/Device_list_updates_reach_newly_joined_servers_in_partial_state_rooms` is flakey - https://github.com/matrix-org/synapse/actions/runs/3172224388/jobs/5166469780#step:4:2964 ``` client.go:604: [CSAPI] GET hs1/_matrix/client/v3/sync => 200 OK (125.434036ms) federation_room_join_partial_state_test.go:1674: @t24alice:hs1's partial state join to !0:host.docker.internal:38237 completed. client.go:604: [CSAPI] PUT hs1/_matrix/client/v3/devices/PBJFBBTZIN => 200 OK (18.65378ms) federation_room_join_partial_state_test.go:1677: @t24alice:hs1 sent device list update. federation_room_join_partial_state_test.go:1534: Complement server received m.device_list_update: {"device_display_name":"A new device name 3","device_id":"PBJFBBTZIN","org.matrix.opentracing_context":"{}","prev_id":[35],"stream_id":38,"user_id":"@t24alice:hs1"} federation_room_join_partial_state_test.go:1534: Complement server received m.device_list_update: {"device_display_name":"A new device name 3","device_id":"PBJFBBTZIN","org.matrix.opentracing_context":"{}","prev_id":[36],"stream_id":40,"user_id":"@t24alice:hs1"} federation_room_join_partial_state_test.go:1680: @charlie, @derek and @elsie received device list update. federation_room_join_partial_state_test.go:1534: Complement server received m.device_list_update: {"device_display_name":"A new device name 3","device_id":"PBJFBBTZIN","org.matrix.opentracing_context":"{}","prev_id":[38],"stream_id":39,"user_id":"@t24alice:hs1"} panic: test timed out after 1h0m0s goroutine 3428 [running]: testing.(*M).startAlarm.func1() /opt/hostedtoolcache/go/1.17.13/x64/src/testing/testing.go:1788 +0x8e created by time.goFunc /opt/hostedtoolcache/go/1.17.13/x64/src/time/sleep.go:180 +0x31 ``` Synapse sent an extra device update, which the test wasn't interested in. I think putting the device update in a channel blocked, and then the test timed out after an hour.
non_process
complement testpartialstatejoin outgoing device list updates device list updates reach newly joined servers in partial state rooms is flakey client go get matrix client sync ok federation room join partial state test go s partial state join to host docker internal completed client go put matrix client devices pbjfbbtzin ok federation room join partial state test go sent device list update federation room join partial state test go complement server received m device list update device display name a new device name device id pbjfbbtzin org matrix opentracing context prev id stream id user id federation room join partial state test go complement server received m device list update device display name a new device name device id pbjfbbtzin org matrix opentracing context prev id stream id user id federation room join partial state test go charlie derek and elsie received device list update federation room join partial state test go complement server received m device list update device display name a new device name device id pbjfbbtzin org matrix opentracing context prev id stream id user id panic test timed out after goroutine testing m startalarm opt hostedtoolcache go src testing testing go created by time gofunc opt hostedtoolcache go src time sleep go synapse sent an extra device update which the test wasn t interested in i think putting the device update in a channel blocked and then the test timed out after an hour
0
231,782
7,643,286,581
IssuesEvent
2018-05-08 12:12:35
robotology/wb-toolbox
https://api.github.com/repos/robotology/wb-toolbox
closed
Implement the new parameter parsing method for Simulink
complexity:medium component:library-mex priority:normal status:in-progress type:task
Implement in `SimulinkBlockInformation` the new logic for parsing parameters from the masks. Many of the methods currently declared in `BlockInformation` will be ported in the `SimulinkBlockInformation` implementation.
1.0
Implement the new parameter parsing method for Simulink - Implement in `SimulinkBlockInformation` the new logic for parsing parameters from the masks. Many of the methods currently declared in `BlockInformation` will be ported in the `SimulinkBlockInformation` implementation.
non_process
implement the new parameter parsing method for simulink implement in simulinkblockinformation the new logic for parsing parameters from the masks many of the methods currently declared in blockinformation will be ported in the simulinkblockinformation implementation
0
17,121
22,638,792,586
IssuesEvent
2022-06-30 22:12:08
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/transform] Refactor newPathGetSetters to share similar cases
help wanted good first issue priority:p3 comp: transformprocessor
**Is your feature request related to a problem? Please describe.** Right now in the transform processor each signal's newPathGetSetter function has the same logic for `resource` and `instrumentation_scope`, resulting in duplicate code and unit tests. **Describe the solution you'd like** The access logic for `resource` and `instrumentation_scope` should be moved to the `common` package and then reused in each signal. Unit tests should be moved as well.
1.0
[processor/transform] Refactor newPathGetSetters to share similar cases - **Is your feature request related to a problem? Please describe.** Right now in the transform processor each signal's newPathGetSetter function has the same logic for `resource` and `instrumentation_scope`, resulting in duplicate code and unit tests. **Describe the solution you'd like** The access logic for `resource` and `instrumentation_scope` should be moved to the `common` package and then reused in each signal. Unit tests should be moved as well.
process
refactor newpathgetsetters to share similar cases is your feature request related to a problem please describe right now in the transform processor each signal s newpathgetsetter function has the same logic for resource and instrumentation scope resulting in duplicate code and unit tests describe the solution you d like the access logic for resource and instrumentation scope should be moved to the common package and then reused in each signal unit tests should be moved as well
1
3,693
6,718,474,277
IssuesEvent
2017-10-15 13:20:52
PaddlePaddle/models
https://api.github.com/repos/PaddlePaddle/models
closed
add text classification demo for nested sequence data
model in process
text classification with nested sequence data in paddle.v2
1.0
add text classification demo for nested sequence data - text classification with nested sequence data in paddle.v2
process
add text classification demo for nested sequence data text classification with nested sequence data in paddle
1
16,807
22,053,710,781
IssuesEvent
2022-05-30 10:56:49
0xffset/rOSt
https://api.github.com/repos/0xffset/rOSt
closed
Implement syscall return for kernel-mode processes
syscalls processes
Currently the syscall is returning using the `sysretq` instruction, which is fine and all, but if we want to support kernel processes (for drivers etc.) we need to use `iretq` instead, because `sysretq` changes the privilege mode to Ring3.
1.0
Implement syscall return for kernel-mode processes - Currently the syscall is returning using the `sysretq` instruction, which is fine and all, but if we want to support kernel processes (for drivers etc.) we need to use `iretq` instead, because `sysretq` changes the privilege mode to Ring3.
process
implement syscall return for kernel mode processes currently the syscall is returning using the sysretq instruction which is fine and all but if we want to support kernel processes for drivers etc we need to use iretq instead because sysretq changes the privilege mode to
1
17,774
5,513,112,795
IssuesEvent
2017-03-17 11:29:18
akvo/akvo-flow
https://api.github.com/repos/akvo/akvo-flow
closed
Remove duplicated elements across dashboard
Legacy clean-up & Code Refactoring Ready for release
Currently there are a few elements across the dashboard that do not have any purpose, provide no new information to the user or guidance, and are duplicated. With this issue, we will identify these elements and remove them all at once. Simplify. Declutter. Simplify. ![screen shot 2016-11-28 at 16 51 09](https://cloud.githubusercontent.com/assets/12456965/20674980/14650cb0-b58b-11e6-9135-042ab6020834.png) *(example of a duplicated submenu and page title)
1.0
Remove duplicated elements across dashboard - Currently there are a few elements across the dashboard that do not have any purpose, provide no new information to the user or guidance, and are duplicated. With this issue, we will identify these elements and remove them all at once. Simplify. Declutter. Simplify. ![screen shot 2016-11-28 at 16 51 09](https://cloud.githubusercontent.com/assets/12456965/20674980/14650cb0-b58b-11e6-9135-042ab6020834.png) *(example of a duplicated submenu and page title)
non_process
remove duplicated elements across dashboard currently there are a few elements across the dashboard that do not have any purpose provide no new information to the user or guidance and are duplicated with this issue we will identify these elements and remove them all at once simplify declutter simplify example of a duplicated submenu and page title
0
7,046
10,207,861,060
IssuesEvent
2019-08-14 08:45:39
endlessm/azafea
https://api.github.com/repos/endlessm/azafea
closed
Activation records don't have their vendors normalized
bug endless event processors
It seems I only did it for ping records (where this is essential, to deduplicate the ping configurations), but not for activation records.
1.0
Activation records don't have their vendors normalized - It seems I only did it for ping records (where this is essential, to deduplicate the ping configurations), but not for activation records.
process
activation records don t have their vendors normalized it seems i only did it for ping records where this is essential to deduplicate the ping configurations but not for activation records
1
43,661
7,057,766,697
IssuesEvent
2018-01-04 17:38:19
aeternity/elixir-wallet
https://api.github.com/repos/aeternity/elixir-wallet
closed
Add packages that the user needs to install in README
documentation
Add the packages/libraries that the user must install prior to running the elixir-wallet project. This is related to issue #65
1.0
Add packages that the user needs to install in README - Add the packages/libraries that the user must install prior to running the elixir-wallet project. This is related to issue #65
non_process
add packages that the user needs to install in readme add the packages libraries that the user must install prior to running the elixir wallet project this is related to issue
0
62,019
6,773,501,248
IssuesEvent
2017-10-27 06:21:37
kubernetes/test-infra
https://api.github.com/repos/kubernetes/test-infra
closed
Should check GOPATH in kubetest/util.go
area/kubetest
If there is no `GOPATH` env configed, `k8s` func will return `src/k8s.io/...`, so `Build` func will construct cmd such as: ```shell $ make -C src/k8s.io/kubernetes quick-release ``` and `kubetest` will throw error `make[1]: *** src/k8s.io/kubernetes: No such file or directory. Stop.`
1.0
Should check GOPATH in kubetest/util.go - If there is no `GOPATH` env configed, `k8s` func will return `src/k8s.io/...`, so `Build` func will construct cmd such as: ```shell $ make -C src/k8s.io/kubernetes quick-release ``` and `kubetest` will throw error `make[1]: *** src/k8s.io/kubernetes: No such file or directory. Stop.`
non_process
should check gopath in kubetest util go if there is no gopath env configed func will return src io so build func will construct cmd such as shell make c src io kubernetes quick release and kubetest will throw error make src io kubernetes no such file or directory stop
0
10,066
13,044,161,802
IssuesEvent
2020-07-29 03:47:26
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `StrToDateDatetime` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `StrToDateDatetime` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @andylokandy ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `StrToDateDatetime` from TiDB - ## Description Port the scalar function `StrToDateDatetime` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @andylokandy ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function strtodatedatetime from tidb description port the scalar function strtodatedatetime from tidb to coprocessor score mentor s andylokandy recommended skills rust programming learning materials already implemented expressions ported from tidb
1
13,279
15,759,848,671
IssuesEvent
2021-03-31 08:22:08
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
(migrate/MySQL) Print visible warnings when a user-edited migration contains explicit database names
process/candidate team/migrations
Explicit database names will not work nicely with the shadow database — the inferred schemas will be wrong and drift will always be detected. See https://github.com/prisma/prisma/issues/5987 for an example. We can detect prefixed identifiers in migration scripts, so we should warn about them.
1.0
(migrate/MySQL) Print visible warnings when a user-edited migration contains explicit database names - Explicit database names will not work nicely with the shadow database — the inferred schemas will be wrong and drift will always be detected. See https://github.com/prisma/prisma/issues/5987 for an example. We can detect prefixed identifiers in migration scripts, so we should warn about them.
process
migrate mysql print visible warnings when a user edited migration contains explicit database names explicit database names will not work nicely with the shadow database — the inferred schemas will be wrong and drift will always be detected see for an example we can detect prefixed identifiers in migration scripts so we should warn about them
1
179,767
30,296,465,963
IssuesEvent
2023-07-09 22:42:11
Esri/calcite-design-system
https://api.github.com/repos/Esri/calcite-design-system
reopened
Audit components `open`/`closed` props for consistency
enhancement p - low Calcite (design) needs milestone design-complete
### Description Perform an audit across components with the `open`/`expanded`/`active` prop for consistency. The audit results can lead us to add deprecations and/or update values across components. cc @geospatialem @ashetland @macandcheese @jcfranco <img width="1427" alt="image" src="https://github.com/Esri/calcite-components/assets/29716057/5d221faa-3994-4c9c-826e-bc3d0d6c0deb"> ### Acceptance Criteria `open` (Reveals a floating element) Applies to: Combobox, Input Date Picker, Input Time Picker, Modal, Dropdown, Popover, & Tooltip `expanded` (Reveals extended content) Applies to: Accordion, Block, Tree Item, List Item, & Action Bar Deprecated `closed` prop on **Block** and **List Item** and added `expanded` prop ### Which Component Applies across multiple components using a `open`/`expanded`/`active` prop: Block, Combobox, Input Date Picker, Input Time Picker, List Item, Modal, Dropdown, Popover, Tooltip, Accordion, Tree Item, & Action Bar ### Esri team Calcite (design)
2.0
Audit components `open`/`closed` props for consistency - ### Description Perform an audit across components with the `open`/`expanded`/`active` prop for consistency. The audit results can lead us to add deprecations and/or update values across components. cc @geospatialem @ashetland @macandcheese @jcfranco <img width="1427" alt="image" src="https://github.com/Esri/calcite-components/assets/29716057/5d221faa-3994-4c9c-826e-bc3d0d6c0deb"> ### Acceptance Criteria `open` (Reveals a floating element) Applies to: Combobox, Input Date Picker, Input Time Picker, Modal, Dropdown, Popover, & Tooltip `expanded` (Reveals extended content) Applies to: Accordion, Block, Tree Item, List Item, & Action Bar Deprecated `closed` prop on **Block** and **List Item** and added `expanded` prop ### Which Component Applies across multiple components using a `open`/`expanded`/`active` prop: Block, Combobox, Input Date Picker, Input Time Picker, List Item, Modal, Dropdown, Popover, Tooltip, Accordion, Tree Item, & Action Bar ### Esri team Calcite (design)
non_process
audit components open closed props for consistency description perform an audit across components with the open expanded active prop for consistency the audit results can lead us to add deprecations and or update values across components cc geospatialem ashetland macandcheese jcfranco img width alt image src acceptance criteria open reveals a floating element applies to combobox input date picker input time picker modal dropdown popover tooltip expanded reveals extended content applies to accordion block tree item list item action bar deprecated closed prop on block and list item and added expanded prop which component applies across multiple components using a open expanded active prop block combobox input date picker input time picker list item modal dropdown popover tooltip accordion tree item action bar esri team calcite design
0
1,960
4,777,570,933
IssuesEvent
2016-10-27 16:39:34
paulkornikov/Pragonas
https://api.github.com/repos/paulkornikov/Pragonas
closed
Provision existe mais le processus provision recrée la provision
a-bug financement - provisions processus workload III
pour le compte ibis, duplication des provisions existantes depuis la date de création de la famille.
1.0
Provision existe mais le processus provision recrée la provision - pour le compte ibis, duplication des provisions existantes depuis la date de création de la famille.
process
provision existe mais le processus provision recrée la provision pour le compte ibis duplication des provisions existantes depuis la date de création de la famille
1
35,502
2,789,932,261
IssuesEvent
2015-05-08 22:30:47
google/google-visualization-api-issues
https://api.github.com/repos/google/google-visualization-api-issues
opened
Feature Request: Add option to have vertical grid lines
Priority-Low Type-Enhancement
Original [issue 556](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=556) created by orwant on 2011-03-20T19:12:41.000Z: <b>What would you like to see us add to this API?</b> Add option to have vertical grid lines. <b>What component is this issue related to (PieChart, LineChart, DataTable,</b> <b>Query, etc)?</b> LineChart. <b>*********************************************************</b> <b>For developers viewing this issue: please click the 'star' icon to be</b> <b>notified of future changes, and to let us know how many of you are</b> <b>interested in seeing it resolved.</b> <b>*********************************************************</b>
1.0
Feature Request: Add option to have vertical grid lines - Original [issue 556](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=556) created by orwant on 2011-03-20T19:12:41.000Z: <b>What would you like to see us add to this API?</b> Add option to have vertical grid lines. <b>What component is this issue related to (PieChart, LineChart, DataTable,</b> <b>Query, etc)?</b> LineChart. <b>*********************************************************</b> <b>For developers viewing this issue: please click the 'star' icon to be</b> <b>notified of future changes, and to let us know how many of you are</b> <b>interested in seeing it resolved.</b> <b>*********************************************************</b>
non_process
feature request add option to have vertical grid lines original created by orwant on what would you like to see us add to this api add option to have vertical grid lines what component is this issue related to piechart linechart datatable query etc linechart for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved
0
17,318
23,138,300,003
IssuesEvent
2022-07-28 15:59:30
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
MySQL JSON fields' filters don't work for heterogeneous type fields
Type:Bug Database/MySQL Querying/Processor .Correctness .Regression/master
MySQL JSON fields, but not Postgres JSON fields, will coerce to the [JSON type automagically when given comparisons](https://dev.mysql.com/doc/refman/8.0/en/json.html#json-comparison) like we do in the filters, which is highly unintuitive in our specific situation. We want to coerce to the ordinary types upon querying and have the hierarchical field type behavior like we do in the Postgres JSON fields. Consider this under #22174, given the epic isn't closed out yet. **To Reproduce** 1. Insert a JSON column into a setup MySQL database table with heterogeneous members, like something with two rows `{"blah": 123"}` and `{"blah": "some stuff"}` on column `some_json`. 2. Go to the table and filter on the values of `some_json -> blah` 3. You will always get null results, if the JSON column is heterogeneous in type **Expected behavior** Filtering happens, we get the rows corresponding to that value or what-have-you **Screenshot** from Nick: https://files.slack.com/files-pri/T078VCLCR-F03QN355UEM/chrome_lpjzn5qnuk.gif **Information about your Metabase Installation:** `{ "browser-info": { "language": "en-US", "platform": "MacIntel", "userAgent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36", "vendor": "Google Inc." }, "system-info": { "file.encoding": "UTF-8", "java.runtime.name": "OpenJDK Runtime Environment", "java.runtime.version": "16.0.1+9", "java.vendor": "AdoptOpenJDK", "java.vendor.url": "https://adoptopenjdk.net/", "java.version": "16.0.1", "java.vm.name": "OpenJDK 64-Bit Server VM", "java.vm.version": "16.0.1+9", "os.name": "Mac OS X", "os.version": "11.2", "user.language": "en", "user.timezone": "UTC" }, "metabase-info": { "databases": [ "h2", "mysql" ], "hosting-env": "unknown", "application-database": "mysql", "application-database-details": { "database": { "name": "MySQL", "version": "8.0.29" }, "jdbc-driver": { "name": "MariaDB Connector/J", "version": "2.7.5" } }, "run-mode": "dev", "version": { "date": "2022-01-19", "src_hash": "71e53e611d2b59048e8c28fc64b297b472c8920a", "tag": "v0.42.1-SNAPSHOT", "branch": "cleanup-sv-2", "hash": "1e8dc79" }, "settings": { "report-timezone": null } } }` **Severity** We haven't released MySQL yet, but I am going to put a ticket in here so I can refer in PR **Additional context** Internal slack https://metaboat.slack.com/archives/C03E8MQJZBM/p1658327516672559
1.0
MySQL JSON fields' filters don't work for heterogeneous type fields - MySQL JSON fields, but not Postgres JSON fields, will coerce to the [JSON type automagically when given comparisons](https://dev.mysql.com/doc/refman/8.0/en/json.html#json-comparison) like we do in the filters, which is highly unintuitive in our specific situation. We want to coerce to the ordinary types upon querying and have the hierarchical field type behavior like we do in the Postgres JSON fields. Consider this under #22174, given the epic isn't closed out yet. **To Reproduce** 1. Insert a JSON column into a setup MySQL database table with heterogeneous members, like something with two rows `{"blah": 123"}` and `{"blah": "some stuff"}` on column `some_json`. 2. Go to the table and filter on the values of `some_json -> blah` 3. You will always get null results, if the JSON column is heterogeneous in type **Expected behavior** Filtering happens, we get the rows corresponding to that value or what-have-you **Screenshot** from Nick: https://files.slack.com/files-pri/T078VCLCR-F03QN355UEM/chrome_lpjzn5qnuk.gif **Information about your Metabase Installation:** `{ "browser-info": { "language": "en-US", "platform": "MacIntel", "userAgent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36", "vendor": "Google Inc." }, "system-info": { "file.encoding": "UTF-8", "java.runtime.name": "OpenJDK Runtime Environment", "java.runtime.version": "16.0.1+9", "java.vendor": "AdoptOpenJDK", "java.vendor.url": "https://adoptopenjdk.net/", "java.version": "16.0.1", "java.vm.name": "OpenJDK 64-Bit Server VM", "java.vm.version": "16.0.1+9", "os.name": "Mac OS X", "os.version": "11.2", "user.language": "en", "user.timezone": "UTC" }, "metabase-info": { "databases": [ "h2", "mysql" ], "hosting-env": "unknown", "application-database": "mysql", "application-database-details": { "database": { "name": "MySQL", "version": "8.0.29" }, "jdbc-driver": { "name": "MariaDB Connector/J", "version": "2.7.5" } }, "run-mode": "dev", "version": { "date": "2022-01-19", "src_hash": "71e53e611d2b59048e8c28fc64b297b472c8920a", "tag": "v0.42.1-SNAPSHOT", "branch": "cleanup-sv-2", "hash": "1e8dc79" }, "settings": { "report-timezone": null } } }` **Severity** We haven't released MySQL yet, but I am going to put a ticket in here so I can refer in PR **Additional context** Internal slack https://metaboat.slack.com/archives/C03E8MQJZBM/p1658327516672559
process
mysql json fields filters don t work for heterogeneous type fields mysql json fields but not postgres json fields will coerce to the like we do in the filters which is highly unintuitive in our specific situation we want to coerce to the ordinary types upon querying and have the hierarchical field type behavior like we do in the postgres json fields consider this under given the epic isn t closed out yet to reproduce insert a json column into a setup mysql database table with heterogeneous members like something with two rows blah and blah some stuff on column some json go to the table and filter on the values of some json blah you will always get null results if the json column is heterogeneous in type expected behavior filtering happens we get the rows corresponding to that value or what have you screenshot from nick information about your metabase installation browser info language en us platform macintel useragent mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari vendor google inc system info file encoding utf java runtime name openjdk runtime environment java runtime version java vendor adoptopenjdk java vendor url java version java vm name openjdk bit server vm java vm version os name mac os x os version user language en user timezone utc metabase info databases mysql hosting env unknown application database mysql application database details database name mysql version jdbc driver name mariadb connector j version run mode dev version date src hash tag snapshot branch cleanup sv hash settings report timezone null severity we haven t released mysql yet but i am going to put a ticket in here so i can refer in pr additional context internal slack
1
21,224
28,310,618,662
IssuesEvent
2023-04-10 15:05:02
cse442-at-ub/project_s23-iweatherify
https://api.github.com/repos/cse442-at-ub/project_s23-iweatherify
closed
Connect Unit Settings to the Logged-in Homepage reflecting different weather data based on unit preferences selected
Processing Task Sprint 3
**Tests** Test Setup: Visit https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442a/#/homepage as a logged-in user with a new valid registered account *Test 1* 1) Check to confirm your newly registered account has the defaults unit of Fahrenheit, mph for wind speed, and Hg for pressure ![Screenshot 2023-04-07 at 4.40.14 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/bed30b69-b545-45b7-b12a-1b26c6a46b59) 2) As a logged-in user visit the Unit Settings page https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442a/#/unitsSettings ![Screenshot 2023-04-07 at 4.38.56 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/5facff28-6882-4643-907b-61941c97f580) 3) Change the default Temperature from °F to °C 4) Change The Wind option from mph to km/h 5) Change for Pressure Hg to mb 6) Click on Save ![Screenshot 2023-04-07 at 4.42.50 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/6d8a451e-f567-47f3-b406-94b90d4ddad3) 7) Visit the logged-in homepage https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442a/#/homepage 8) Confirm that the units have changed to C for temperature, km/h for Wind Speed and mb for Pressure ![Screenshot 2023-04-07 at 5.10.01 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/cfd21813-feaa-4df2-8a72-a4dc0aba5187) *Test 2* 1) After completing test 1 go back to the unit settings page https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442a/#/unitsSettings 2) Check that the Temperature has °C, Wind on km/h, and pressure on MB ![Screenshot 2023-04-07 at 5.07.25 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/5c362a17-d4a7-49b3-9e92-f658419b002d) 3) Change the default Temperature from °C to °F 4) Change The Wind option from km/h to mph 5) Change for Pressure mb to to Hg 6) Click on Save ![Screenshot 2023-04-07 at 5.08.17 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/ea5565f9-3fed-4a9a-a242-0420b1b69d0a) 7) Visit the logged-in homepage https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442a/#/homepage 8) Confirm that the units have changed to F for temperature, mph for Wind Speed and Hg for Pressure ![Screenshot 2023-04-07 at 5.09.12 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/00675b7f-4436-4ca1-9638-8001ebf8c8fa)
1.0
Connect Unit Settings to the Logged-in Homepage reflecting different weather data based on unit preferences selected - **Tests** Test Setup: Visit https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442a/#/homepage as a logged-in user with a new valid registered account *Test 1* 1) Check to confirm your newly registered account has the defaults unit of Fahrenheit, mph for wind speed, and Hg for pressure ![Screenshot 2023-04-07 at 4.40.14 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/bed30b69-b545-45b7-b12a-1b26c6a46b59) 2) As a logged-in user visit the Unit Settings page https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442a/#/unitsSettings ![Screenshot 2023-04-07 at 4.38.56 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/5facff28-6882-4643-907b-61941c97f580) 3) Change the default Temperature from °F to °C 4) Change The Wind option from mph to km/h 5) Change for Pressure Hg to mb 6) Click on Save ![Screenshot 2023-04-07 at 4.42.50 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/6d8a451e-f567-47f3-b406-94b90d4ddad3) 7) Visit the logged-in homepage https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442a/#/homepage 8) Confirm that the units have changed to C for temperature, km/h for Wind Speed and mb for Pressure ![Screenshot 2023-04-07 at 5.10.01 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/cfd21813-feaa-4df2-8a72-a4dc0aba5187) *Test 2* 1) After completing test 1 go back to the unit settings page https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442a/#/unitsSettings 2) Check that the Temperature has °C, Wind on km/h, and pressure on MB ![Screenshot 2023-04-07 at 5.07.25 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/5c362a17-d4a7-49b3-9e92-f658419b002d) 3) Change the default Temperature from °C to °F 4) Change The Wind option from km/h to mph 5) Change for Pressure mb to to Hg 6) Click on Save ![Screenshot 2023-04-07 at 5.08.17 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/ea5565f9-3fed-4a9a-a242-0420b1b69d0a) 7) Visit the logged-in homepage https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442a/#/homepage 8) Confirm that the units have changed to F for temperature, mph for Wind Speed and Hg for Pressure ![Screenshot 2023-04-07 at 5.09.12 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/00675b7f-4436-4ca1-9638-8001ebf8c8fa)
process
connect unit settings to the logged in homepage reflecting different weather data based on unit preferences selected tests test setup visit as a logged in user with a new valid registered account test check to confirm your newly registered account has the defaults unit of fahrenheit mph for wind speed and hg for pressure as a logged in user visit the unit settings page change the default temperature from °f to °c change the wind option from mph to km h change for pressure hg to mb click on save visit the logged in homepage confirm that the units have changed to c for temperature km h for wind speed and mb for pressure test after completing test go back to the unit settings page check that the temperature has °c wind on km h and pressure on mb change the default temperature from °c to °f change the wind option from km h to mph change for pressure mb to to hg click on save visit the logged in homepage confirm that the units have changed to f for temperature mph for wind speed and hg for pressure
1
670,167
22,678,387,432
IssuesEvent
2022-07-04 07:39:38
nakhll-company/nakhll_frontend
https://api.github.com/repos/nakhll-company/nakhll_frontend
opened
لطفا ابتدا وارد سایت شوید
bug Priority 1
اگر وارد سایت نشده باشیم در صفحه لندینگ ارور "لطفا ابتدا وارد شوید" نمایش داده می شود ولی اگر در سایت باشیم و Session اکسپایر شود، این ارور در همه صفحات نمایش داده می شود و با کلیک روی پروفایل برای ورود دوباره، ارور 500 می دهد. باید حتما یک بار سایت را رفرش کرد تا بتوانیم دوباره وارد بشویم.
1.0
لطفا ابتدا وارد سایت شوید - اگر وارد سایت نشده باشیم در صفحه لندینگ ارور "لطفا ابتدا وارد شوید" نمایش داده می شود ولی اگر در سایت باشیم و Session اکسپایر شود، این ارور در همه صفحات نمایش داده می شود و با کلیک روی پروفایل برای ورود دوباره، ارور 500 می دهد. باید حتما یک بار سایت را رفرش کرد تا بتوانیم دوباره وارد بشویم.
non_process
لطفا ابتدا وارد سایت شوید اگر وارد سایت نشده باشیم در صفحه لندینگ ارور لطفا ابتدا وارد شوید نمایش داده می شود ولی اگر در سایت باشیم و session اکسپایر شود، این ارور در همه صفحات نمایش داده می شود و با کلیک روی پروفایل برای ورود دوباره، ارور می دهد باید حتما یک بار سایت را رفرش کرد تا بتوانیم دوباره وارد بشویم
0
42,804
5,538,157,784
IssuesEvent
2017-03-22 00:28:31
archesproject/arches
https://api.github.com/repos/archesproject/arches
closed
Remove or Implement Tile Name in Card Configuration
bug Card Designer
### User Story Currently a user is able to enter a "Tile Name" in a Card's settings, when multiple values are enabled for that card. ![image](https://cloud.githubusercontent.com/assets/10427268/22891749/ef0599f8-f1d5-11e6-826d-817e577e0dcd.png) This feature is not actually implemented in the rest of the app. ### Definition of Done Before full documentation of the Card Designer can be finished, the Tile Name entry should either be removed (from the screenshot above as well as the card preview), or it should be fully implemented.
1.0
Remove or Implement Tile Name in Card Configuration - ### User Story Currently a user is able to enter a "Tile Name" in a Card's settings, when multiple values are enabled for that card. ![image](https://cloud.githubusercontent.com/assets/10427268/22891749/ef0599f8-f1d5-11e6-826d-817e577e0dcd.png) This feature is not actually implemented in the rest of the app. ### Definition of Done Before full documentation of the Card Designer can be finished, the Tile Name entry should either be removed (from the screenshot above as well as the card preview), or it should be fully implemented.
non_process
remove or implement tile name in card configuration user story currently a user is able to enter a tile name in a card s settings when multiple values are enabled for that card this feature is not actually implemented in the rest of the app definition of done before full documentation of the card designer can be finished the tile name entry should either be removed from the screenshot above as well as the card preview or it should be fully implemented
0
457,453
13,156,810,674
IssuesEvent
2020-08-10 11:32:52
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.instagram.com - see bug description
browser-firefox-mobile engine-gecko ml-needsdiagnosis-false priority-critical
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 6.0; Mobile; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/56375 --> **URL**: https://www.instagram.com/ **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 6.0 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: can't login **Steps to Reproduce**: Login says it cant connect to instagram i should get online <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/8/8ed13eec-b136-49df-8a4a-ffede8b2c0e5.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200501050101</li><li>channel: alpha</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/8/28f22538-6474-4562-8309-afdb0f86604a) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.instagram.com - see bug description - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 6.0; Mobile; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/56375 --> **URL**: https://www.instagram.com/ **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 6.0 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: can't login **Steps to Reproduce**: Login says it cant connect to instagram i should get online <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/8/8ed13eec-b136-49df-8a4a-ffede8b2c0e5.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200501050101</li><li>channel: alpha</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/8/28f22538-6474-4562-8309-afdb0f86604a) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description can t login steps to reproduce login says it cant connect to instagram i should get online view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel alpha hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
159,693
20,085,893,620
IssuesEvent
2022-02-05 01:08:09
AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches
https://api.github.com/repos/AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches
opened
CVE-2022-21727 (High) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl
security vulnerability
## CVE-2022-21727 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /FinalProject/requirements.txt</p> <p>Path to vulnerable library: /teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Tensorflow is an Open Source Machine Learning Framework. The implementation of shape inference for `Dequantize` is vulnerable to an integer overflow weakness. The `axis` argument can be `-1` (the default value for the optional argument) or any other positive value at most the number of dimensions of the input. Unfortunately, the upper bound is not checked, and, since the code computes `axis + 1`, an attacker can trigger an integer overflow. The fix will be included in TensorFlow 2.8.0. We will also cherrypick this commit on TensorFlow 2.7.1, TensorFlow 2.6.3, and TensorFlow 2.5.3, as these are also affected and still in supported range. <p>Publish Date: 2022-02-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21727>CVE-2022-21727</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-c6fh-56w7-fvjw">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-c6fh-56w7-fvjw</a></p> <p>Release Date: 2022-02-03</p> <p>Fix Resolution: tensorflow - 2.5.3,2.6.3,2.7.1;tensorflow-cpu - 2.5.3,2.6.3,2.7.1;tensorflow-gpu - 2.5.3,2.6.3,2.7.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-21727 (High) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2022-21727 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /FinalProject/requirements.txt</p> <p>Path to vulnerable library: /teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Tensorflow is an Open Source Machine Learning Framework. The implementation of shape inference for `Dequantize` is vulnerable to an integer overflow weakness. The `axis` argument can be `-1` (the default value for the optional argument) or any other positive value at most the number of dimensions of the input. Unfortunately, the upper bound is not checked, and, since the code computes `axis + 1`, an attacker can trigger an integer overflow. The fix will be included in TensorFlow 2.8.0. We will also cherrypick this commit on TensorFlow 2.7.1, TensorFlow 2.6.3, and TensorFlow 2.5.3, as these are also affected and still in supported range. <p>Publish Date: 2022-02-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21727>CVE-2022-21727</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-c6fh-56w7-fvjw">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-c6fh-56w7-fvjw</a></p> <p>Release Date: 2022-02-03</p> <p>Fix Resolution: tensorflow - 2.5.3,2.6.3,2.7.1;tensorflow-cpu - 2.5.3,2.6.3,2.7.1;tensorflow-gpu - 2.5.3,2.6.3,2.7.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file finalproject requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an open source machine learning framework the implementation of shape inference for dequantize is vulnerable to an integer overflow weakness the axis argument can be the default value for the optional argument or any other positive value at most the number of dimensions of the input unfortunately the upper bound is not checked and since the code computes axis an attacker can trigger an integer overflow the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
0
286,779
24,784,584,754
IssuesEvent
2022-10-24 08:44:51
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
Test uart async failed on Nucleo F429ZI
bug priority: low area: UART area: DMA platform: STM32 area: Tests
**Describe the bug** The test `tests/drivers/uart/uart_async_api/drivers.uart.async_api` fails on Nucleo F429ZI with the following error `Assertion failed at WEST_TOPDIR/zephyr/tests/drivers/uart/uart_async_api/src/test_uart_async.c:526: uart_async_write_abort_test_write_abort: (sent not equal to received) Sent is not equal to received.`
1.0
Test uart async failed on Nucleo F429ZI - **Describe the bug** The test `tests/drivers/uart/uart_async_api/drivers.uart.async_api` fails on Nucleo F429ZI with the following error `Assertion failed at WEST_TOPDIR/zephyr/tests/drivers/uart/uart_async_api/src/test_uart_async.c:526: uart_async_write_abort_test_write_abort: (sent not equal to received) Sent is not equal to received.`
non_process
test uart async failed on nucleo describe the bug the test tests drivers uart uart async api drivers uart async api fails on nucleo with the following error assertion failed at west topdir zephyr tests drivers uart uart async api src test uart async c uart async write abort test write abort sent not equal to received sent is not equal to received
0
680,190
23,261,735,048
IssuesEvent
2022-08-04 14:01:21
Earthcomputer/multiconnect
https://api.github.com/repos/Earthcomputer/multiconnect
closed
[1.19] -> [1.18.2] Biome types not translated (correctly)
bug priority: low
Biome types are not correctly translated in the newest [1.6-beta.16](https://github.com/Earthcomputer/multiconnect/commit/837b2821d4bcf6cd6fbd4926a650ed69bd788a74) 1.19 protocol (correct biome): ![image](https://user-images.githubusercontent.com/102829507/175566516-d7abf040-6d20-4fcb-9238-7eb18e617457.png) forced 1.18 protocol (wrong biome): ![image](https://user-images.githubusercontent.com/102829507/175566743-8960b8bb-c1f8-47e4-bc24-717eab4e2d38.png) Tested on server `gommehd.net` (lobby).
1.0
[1.19] -> [1.18.2] Biome types not translated (correctly) - Biome types are not correctly translated in the newest [1.6-beta.16](https://github.com/Earthcomputer/multiconnect/commit/837b2821d4bcf6cd6fbd4926a650ed69bd788a74) 1.19 protocol (correct biome): ![image](https://user-images.githubusercontent.com/102829507/175566516-d7abf040-6d20-4fcb-9238-7eb18e617457.png) forced 1.18 protocol (wrong biome): ![image](https://user-images.githubusercontent.com/102829507/175566743-8960b8bb-c1f8-47e4-bc24-717eab4e2d38.png) Tested on server `gommehd.net` (lobby).
non_process
biome types not translated correctly biome types are not correctly translated in the newest protocol correct biome forced protocol wrong biome tested on server gommehd net lobby
0
155,702
5,959,334,470
IssuesEvent
2017-05-29 10:42:05
siteorigin/siteorigin-panels
https://api.github.com/repos/siteorigin/siteorigin-panels
closed
Compatibility with new WordPress 4.8 widgets
priority-1
WordPress 4.8 (beta) introduces a few widgets with more dynamic forms. These forms all come up blank in Page Builder. We need to try trigger actions and enqueue scripts to get these new widgets working. The widgets are * Video * Audio * Image * Text (which replaces the old text widget with a rich text TinyMCE widget). We need to make sure that we introduce compatibility in a backwards compatible way for the text widget.
1.0
Compatibility with new WordPress 4.8 widgets - WordPress 4.8 (beta) introduces a few widgets with more dynamic forms. These forms all come up blank in Page Builder. We need to try trigger actions and enqueue scripts to get these new widgets working. The widgets are * Video * Audio * Image * Text (which replaces the old text widget with a rich text TinyMCE widget). We need to make sure that we introduce compatibility in a backwards compatible way for the text widget.
non_process
compatibility with new wordpress widgets wordpress beta introduces a few widgets with more dynamic forms these forms all come up blank in page builder we need to try trigger actions and enqueue scripts to get these new widgets working the widgets are video audio image text which replaces the old text widget with a rich text tinymce widget we need to make sure that we introduce compatibility in a backwards compatible way for the text widget
0
810,262
30,233,673,365
IssuesEvent
2023-07-06 08:45:55
AdguardTeam/AdguardBrowserExtension
https://api.github.com/repos/AdguardTeam/AdguardBrowserExtension
closed
add screen warning to Adguard extension for Firefox Mobile (Android)
Feature Request Priority: P4
### Issue Details When i visit malicious website Adguard extension doesn't trigger warning message.The link is in the list of Online malicious URLs ### Proposed solution _No response_ ### Alternative solution The screen warning message should be appear for Adguard Firefox mobile. User can decide to continue or go back from the site
1.0
add screen warning to Adguard extension for Firefox Mobile (Android) - ### Issue Details When i visit malicious website Adguard extension doesn't trigger warning message.The link is in the list of Online malicious URLs ### Proposed solution _No response_ ### Alternative solution The screen warning message should be appear for Adguard Firefox mobile. User can decide to continue or go back from the site
non_process
add screen warning to adguard extension for firefox mobile android issue details when i visit malicious website adguard extension doesn t trigger warning message the link is in the list of online malicious urls proposed solution no response alternative solution the screen warning message should be appear for adguard firefox mobile user can decide to continue or go back from the site
0
18,325
24,444,030,971
IssuesEvent
2022-10-06 16:26:02
MPMG-DCC-UFMG/C01
https://api.github.com/repos/MPMG-DCC-UFMG/C01
closed
Passo Opções
[1] Bug [2] Alta Prioridade [0] Desenvolvimento [3] Processamento Dinâmico
## Comportamento Esperado Espera-se que a configuração do passo Opções não apresente problemas na branch `dev`, assim como na branch `master`. ## Comportamento Atual Durante a execução de coletores utilizando o passo Opções, temos exceções semelhantes a essa (log bruto): > `"message": "2022-08-25 06:42:33 [scrapy.core.scraper] ERROR: Spider error processing <GET http://api.conectbr.com.br/Licitacao/Busca/?token=XEOk3i50lCa5IcpiNG7vXQ==> (referer: https://use.fontawesome.com/a61dc1ee64.css)\nTraceback (most recent call last):\n File \"/usr/local/lib/python3.8/dist-packages/twisted/internet/defer.py\", line 824, in adapt\n extracted = result.result()\n File \"/usr/local/lib/python3.8/dist-packages/scrapy/utils/py36.py\", line 8, in collect_asyncgen\n async for x in result:\n File \"/usr/src/app/crawling/spiders/static_page.py\", line 372, in parse\n responses = await self.dynamic_processing(response)\n File \"/usr/src/app/crawling/spiders/static_page.py\", line 328, in dynamic_processing\n page_dict = await steps.execute_steps(pagina=page)\n File \"<string>\", line 11, in execute_steps\n File \"/usr/local/lib/python3.8/dist-packages/step_crawler/functions_file.py\", line 140, in opcoes\n options.append(value.toString().split(\":\")[-1])\nAttributeError: 'NoneType' object has no attribute 'toString'"` Na interface: ![image](https://user-images.githubusercontent.com/48096245/186594350-8214a72c-6f82-4bef-8b96-d26716c20b13.png) ## Passos para reproduzir o erro Executar um coletor usando o passo de Opções (exemplo abaixo) e verificar problemas em seu funcionamento. ## Especificações da Coleta ``` { "source_name": "Licita\u00e7\u00f5es Itabirito", "base_url": "http:\/\/api.conectbr.com.br\/Licitacao\/Busca\/?token=XEOk3i50lCa5IcpiNG7vXQ==", "obey_robots": false, "crawler_description": "Teste passo op\u00e7\u00f5es.", "crawler_type_desc": "Outro", "crawler_issue": 5850, "data_path": "5850IT", "sc_scheduler_persist": true, "sc_scheduler_queue_refresh": 10, "sc_queue_hits": 10, "sc_queue_window": 60, "sc_queue_moderated": true, "sc_dupefilter_timeout": 600, "sc_global_page_per_domain_limit": null, "sc_global_page_per_domain_limit_timeout": 600, "sc_domain_max_page_timeout": 600, "sc_scheduler_ip_refresh": 60, "sc_scheduler_backlog_blacklist": true, "sc_scheduler_type_enabled": true, "sc_scheduler_ip_enabled": true, "sc_scheduler_item_retries": 3, "sc_scheduler_queue_timeout": 3600, "sc_httperror_allow_all": true, "sc_retry_times": 3, "sc_download_timeout": 10, "antiblock_download_delay": 2, "antiblock_autothrottle_enabled": false, "antiblock_autothrottle_start_delay": 2, "antiblock_autothrottle_max_delay": 10, "antiblock_ip_rotation_enabled": false, "antiblock_ip_rotation_type": "tor", "antiblock_max_reqs_per_ip": 10, "antiblock_max_reuse_rounds": 10, "antiblock_proxy_list": "", "antiblock_user_agent_rotation_enabled": false, "antiblock_reqs_per_user_agent": 100, "antiblock_user_agents_list": "", "antiblock_insert_cookies_enabled": false, "antiblock_cookies_list": "", "captcha": "none", "has_webdriver": false, "webdriver_path": "", "img_xpath": "", "sound_xpath": "", "dynamic_processing": true, "browser_type": "chromium", "skip_iter_errors": false, "browser_resolution_width": 1280, "browser_resolution_height": 720, "explore_links": true, "link_extractor_max_depth": 1, "link_extractor_allow_url": "^(http:\\\/\\\/api.conectbr.com.br\\\/Licitacao\\\/AbreArquivo\\\/.*)", "link_extractor_allow_domains": "", "link_extractor_tags": "", "link_extractor_attrs": "", "link_extractor_check_type": false, "link_extractor_process_value": "", "download_files": true, "download_files_allow_url": "^(http:\\\/\\\/api.conectbr.com.br\\\/Licitacao\\\/AbreArquivo\\\/.*)", "download_files_allow_extensions": "pdf,doc", "download_files_allow_domains": "", "download_files_tags": "", "download_files_attrs": "", "download_files_process_value": "", "download_files_check_large_content": true, "download_imgs": false, "steps": "{\"step\":\"root\",\"depth\":0,\"children\":[{\"step\":\"espere\",\"depth\":1,\"arguments\":{\"segundos\":\"5\"}},{\"step\":\"para_cada\",\"depth\":1,\"iterator\":\"ano\",\"children\":[{\"step\":\"selecione\",\"depth\":2,\"arguments\":{\"xpath\":\"'\/\/*[@id=\\\"NUM_ANO_REFERENCIA\\\"]'\",\"opcao\":\"ano\"}},{\"step\":\"espere\",\"depth\":2,\"arguments\":{\"segundos\":\"5\"}},{\"step\":\"clique\",\"depth\":2,\"arguments\":{\"elemento\":\"'\/\/*[@id=\\\"formDados\\\"]\/fieldset\/div\/div[7]\/input'\"}},{\"step\":\"espere\",\"depth\":2,\"arguments\":{\"segundos\":\"5\"}},{\"step\":\"screenshot\",\"depth\":2,\"arguments\":{}}],\"iterable\":{\"call\":{\"step\":\"opcoes\",\"arguments\":{\"xpath\":\"'\/html\/body\/div[1]\/div\/form\/fieldset\/div\/div[1]\/select'\", "encoding_detection_method": 1, "expected_runtime_category": "medium", "templated_url_parameter_handlers": [], "templated_url_response_handlers": [], "instance_id": "166198811654876", "crawler_id": 2 } ``` ## Sistema (caso necessário) - MP ou local: local. - Branch específica: `dev`. - Sistema diferente: não. ## Screenshots (caso necessário) ![image](https://user-images.githubusercontent.com/48096245/186594350-8214a72c-6f82-4bef-8b96-d26716c20b13.png) EDIT: troquei a coleta de exemplo para uma equivalente, porém mais sucinta, para que a testagem da funcionalidade seja mais objetiva.
1.0
Passo Opções - ## Comportamento Esperado Espera-se que a configuração do passo Opções não apresente problemas na branch `dev`, assim como na branch `master`. ## Comportamento Atual Durante a execução de coletores utilizando o passo Opções, temos exceções semelhantes a essa (log bruto): > `"message": "2022-08-25 06:42:33 [scrapy.core.scraper] ERROR: Spider error processing <GET http://api.conectbr.com.br/Licitacao/Busca/?token=XEOk3i50lCa5IcpiNG7vXQ==> (referer: https://use.fontawesome.com/a61dc1ee64.css)\nTraceback (most recent call last):\n File \"/usr/local/lib/python3.8/dist-packages/twisted/internet/defer.py\", line 824, in adapt\n extracted = result.result()\n File \"/usr/local/lib/python3.8/dist-packages/scrapy/utils/py36.py\", line 8, in collect_asyncgen\n async for x in result:\n File \"/usr/src/app/crawling/spiders/static_page.py\", line 372, in parse\n responses = await self.dynamic_processing(response)\n File \"/usr/src/app/crawling/spiders/static_page.py\", line 328, in dynamic_processing\n page_dict = await steps.execute_steps(pagina=page)\n File \"<string>\", line 11, in execute_steps\n File \"/usr/local/lib/python3.8/dist-packages/step_crawler/functions_file.py\", line 140, in opcoes\n options.append(value.toString().split(\":\")[-1])\nAttributeError: 'NoneType' object has no attribute 'toString'"` Na interface: ![image](https://user-images.githubusercontent.com/48096245/186594350-8214a72c-6f82-4bef-8b96-d26716c20b13.png) ## Passos para reproduzir o erro Executar um coletor usando o passo de Opções (exemplo abaixo) e verificar problemas em seu funcionamento. ## Especificações da Coleta ``` { "source_name": "Licita\u00e7\u00f5es Itabirito", "base_url": "http:\/\/api.conectbr.com.br\/Licitacao\/Busca\/?token=XEOk3i50lCa5IcpiNG7vXQ==", "obey_robots": false, "crawler_description": "Teste passo op\u00e7\u00f5es.", "crawler_type_desc": "Outro", "crawler_issue": 5850, "data_path": "5850IT", "sc_scheduler_persist": true, "sc_scheduler_queue_refresh": 10, "sc_queue_hits": 10, "sc_queue_window": 60, "sc_queue_moderated": true, "sc_dupefilter_timeout": 600, "sc_global_page_per_domain_limit": null, "sc_global_page_per_domain_limit_timeout": 600, "sc_domain_max_page_timeout": 600, "sc_scheduler_ip_refresh": 60, "sc_scheduler_backlog_blacklist": true, "sc_scheduler_type_enabled": true, "sc_scheduler_ip_enabled": true, "sc_scheduler_item_retries": 3, "sc_scheduler_queue_timeout": 3600, "sc_httperror_allow_all": true, "sc_retry_times": 3, "sc_download_timeout": 10, "antiblock_download_delay": 2, "antiblock_autothrottle_enabled": false, "antiblock_autothrottle_start_delay": 2, "antiblock_autothrottle_max_delay": 10, "antiblock_ip_rotation_enabled": false, "antiblock_ip_rotation_type": "tor", "antiblock_max_reqs_per_ip": 10, "antiblock_max_reuse_rounds": 10, "antiblock_proxy_list": "", "antiblock_user_agent_rotation_enabled": false, "antiblock_reqs_per_user_agent": 100, "antiblock_user_agents_list": "", "antiblock_insert_cookies_enabled": false, "antiblock_cookies_list": "", "captcha": "none", "has_webdriver": false, "webdriver_path": "", "img_xpath": "", "sound_xpath": "", "dynamic_processing": true, "browser_type": "chromium", "skip_iter_errors": false, "browser_resolution_width": 1280, "browser_resolution_height": 720, "explore_links": true, "link_extractor_max_depth": 1, "link_extractor_allow_url": "^(http:\\\/\\\/api.conectbr.com.br\\\/Licitacao\\\/AbreArquivo\\\/.*)", "link_extractor_allow_domains": "", "link_extractor_tags": "", "link_extractor_attrs": "", "link_extractor_check_type": false, "link_extractor_process_value": "", "download_files": true, "download_files_allow_url": "^(http:\\\/\\\/api.conectbr.com.br\\\/Licitacao\\\/AbreArquivo\\\/.*)", "download_files_allow_extensions": "pdf,doc", "download_files_allow_domains": "", "download_files_tags": "", "download_files_attrs": "", "download_files_process_value": "", "download_files_check_large_content": true, "download_imgs": false, "steps": "{\"step\":\"root\",\"depth\":0,\"children\":[{\"step\":\"espere\",\"depth\":1,\"arguments\":{\"segundos\":\"5\"}},{\"step\":\"para_cada\",\"depth\":1,\"iterator\":\"ano\",\"children\":[{\"step\":\"selecione\",\"depth\":2,\"arguments\":{\"xpath\":\"'\/\/*[@id=\\\"NUM_ANO_REFERENCIA\\\"]'\",\"opcao\":\"ano\"}},{\"step\":\"espere\",\"depth\":2,\"arguments\":{\"segundos\":\"5\"}},{\"step\":\"clique\",\"depth\":2,\"arguments\":{\"elemento\":\"'\/\/*[@id=\\\"formDados\\\"]\/fieldset\/div\/div[7]\/input'\"}},{\"step\":\"espere\",\"depth\":2,\"arguments\":{\"segundos\":\"5\"}},{\"step\":\"screenshot\",\"depth\":2,\"arguments\":{}}],\"iterable\":{\"call\":{\"step\":\"opcoes\",\"arguments\":{\"xpath\":\"'\/html\/body\/div[1]\/div\/form\/fieldset\/div\/div[1]\/select'\", "encoding_detection_method": 1, "expected_runtime_category": "medium", "templated_url_parameter_handlers": [], "templated_url_response_handlers": [], "instance_id": "166198811654876", "crawler_id": 2 } ``` ## Sistema (caso necessário) - MP ou local: local. - Branch específica: `dev`. - Sistema diferente: não. ## Screenshots (caso necessário) ![image](https://user-images.githubusercontent.com/48096245/186594350-8214a72c-6f82-4bef-8b96-d26716c20b13.png) EDIT: troquei a coleta de exemplo para uma equivalente, porém mais sucinta, para que a testagem da funcionalidade seja mais objetiva.
process
passo opções comportamento esperado espera se que a configuração do passo opções não apresente problemas na branch dev assim como na branch master comportamento atual durante a execução de coletores utilizando o passo opções temos exceções semelhantes a essa log bruto message error spider error processing line in execute steps n file usr local lib dist packages step crawler functions file py line in opcoes n options append value tostring split nattributeerror nonetype object has no attribute tostring na interface passos para reproduzir o erro executar um coletor usando o passo de opções exemplo abaixo e verificar problemas em seu funcionamento especificações da coleta source name licita itabirito base url http api conectbr com br licitacao busca token obey robots false crawler description teste passo op crawler type desc outro crawler issue data path sc scheduler persist true sc scheduler queue refresh sc queue hits sc queue window sc queue moderated true sc dupefilter timeout sc global page per domain limit null sc global page per domain limit timeout sc domain max page timeout sc scheduler ip refresh sc scheduler backlog blacklist true sc scheduler type enabled true sc scheduler ip enabled true sc scheduler item retries sc scheduler queue timeout sc httperror allow all true sc retry times sc download timeout antiblock download delay antiblock autothrottle enabled false antiblock autothrottle start delay antiblock autothrottle max delay antiblock ip rotation enabled false antiblock ip rotation type tor antiblock max reqs per ip antiblock max reuse rounds antiblock proxy list antiblock user agent rotation enabled false antiblock reqs per user agent antiblock user agents list antiblock insert cookies enabled false antiblock cookies list captcha none has webdriver false webdriver path img xpath sound xpath dynamic processing true browser type chromium skip iter errors false browser resolution width browser resolution height explore links true link extractor max depth link extractor allow url http api conectbr com br licitacao abrearquivo link extractor allow domains link extractor tags link extractor attrs link extractor check type false link extractor process value download files true download files allow url http api conectbr com br licitacao abrearquivo download files allow extensions pdf doc download files allow domains download files tags download files attrs download files process value download files check large content true download imgs false steps step root depth children opcao ano step espere depth arguments segundos step clique depth arguments elemento fieldset div div input step espere depth arguments segundos step screenshot depth arguments iterable call step opcoes arguments xpath html body div div form fieldset div div select encoding detection method expected runtime category medium templated url parameter handlers templated url response handlers instance id crawler id sistema caso necessário mp ou local local branch específica dev sistema diferente não screenshots caso necessário edit troquei a coleta de exemplo para uma equivalente porém mais sucinta para que a testagem da funcionalidade seja mais objetiva
1
17,311
9,695,271,516
IssuesEvent
2019-05-24 21:46:44
PowerShell/PowerShellEditorServices
https://api.github.com/repos/PowerShell/PowerShellEditorServices
closed
Consider moving to netstandard2.0
Area-General Issue-Enhancement Issue-Performance Waiting-v2
It seems this isn't necessary just yet for PowerShell 6.0.0-beta1, the netstandard1.6 binaries load just fine there. We should evaluate what we'd gain by moving to netstandard2.0 in terms of APIs, though. This work was started here but not merged yet: https://github.com/PowerShell/PowerShellEditorServices/pull/448
True
Consider moving to netstandard2.0 - It seems this isn't necessary just yet for PowerShell 6.0.0-beta1, the netstandard1.6 binaries load just fine there. We should evaluate what we'd gain by moving to netstandard2.0 in terms of APIs, though. This work was started here but not merged yet: https://github.com/PowerShell/PowerShellEditorServices/pull/448
non_process
consider moving to it seems this isn t necessary just yet for powershell the binaries load just fine there we should evaluate what we d gain by moving to in terms of apis though this work was started here but not merged yet
0
450,939
32,000,184,529
IssuesEvent
2023-09-21 11:47:38
Bisaloo/cran-task-view-analysis
https://api.github.com/repos/Bisaloo/cran-task-view-analysis
opened
Discussion
documentation help wanted question
from https://github.com/WHO-Collaboratory/collaboratory-epipipeline-community/discussions/6 # Comments from reviewers ## General comments The team has done a lot of great work flagging issues and making pull requests for existing packages to improve the overall health of the epi pipeline ecosystem. The finalised report is likely most useful for the package developer community. It is not clear how unaware developers would be of the fact that their packages fail the presented checks. For package users, this likely would not influence decision making on which packages to use - this is primarily determined by functionality. We were confused about the deprecated deps check - would this cause the package to become archived in CRAN? ## Future directions Integration of checks into the existing task view or into a platform that lists functionality and categorises packages by into subtopics would be helpful It would be helpful to understand what proportion of epi packages are not captured through the epidemiology CRAN tag (github or improperly tagged CRAN packages)
1.0
Discussion - from https://github.com/WHO-Collaboratory/collaboratory-epipipeline-community/discussions/6 # Comments from reviewers ## General comments The team has done a lot of great work flagging issues and making pull requests for existing packages to improve the overall health of the epi pipeline ecosystem. The finalised report is likely most useful for the package developer community. It is not clear how unaware developers would be of the fact that their packages fail the presented checks. For package users, this likely would not influence decision making on which packages to use - this is primarily determined by functionality. We were confused about the deprecated deps check - would this cause the package to become archived in CRAN? ## Future directions Integration of checks into the existing task view or into a platform that lists functionality and categorises packages by into subtopics would be helpful It would be helpful to understand what proportion of epi packages are not captured through the epidemiology CRAN tag (github or improperly tagged CRAN packages)
non_process
discussion from comments from reviewers general comments the team has done a lot of great work flagging issues and making pull requests for existing packages to improve the overall health of the epi pipeline ecosystem the finalised report is likely most useful for the package developer community it is not clear how unaware developers would be of the fact that their packages fail the presented checks for package users this likely would not influence decision making on which packages to use this is primarily determined by functionality we were confused about the deprecated deps check would this cause the package to become archived in cran future directions integration of checks into the existing task view or into a platform that lists functionality and categorises packages by into subtopics would be helpful it would be helpful to understand what proportion of epi packages are not captured through the epidemiology cran tag github or improperly tagged cran packages
0
11,893
14,688,939,277
IssuesEvent
2021-01-02 06:20:49
wangbjun/blog
https://api.github.com/repos/wangbjun/blog
opened
PHP多进程编程应用 | JWang的博客
/2017/coding/php/php-multi-process.html Gitalk
https://wangbjun.site/2017/coding/php/php-multi-process.html 在日常开发中,我们经常会遇到需要使用脚本处理一些数据,在数据量比较大的情况下,我们可以采用并行的方式处理,比如说: 1.启动多个实例这种方式简单实用,推荐,比如说使用下面的shell脚本我们就可以轻松的启动多个进程去处理 12345678#!/bin/bash for((i=1;i<=8;i++))do /usr/bi
1.0
PHP多进程编程应用 | JWang的博客 - https://wangbjun.site/2017/coding/php/php-multi-process.html 在日常开发中,我们经常会遇到需要使用脚本处理一些数据,在数据量比较大的情况下,我们可以采用并行的方式处理,比如说: 1.启动多个实例这种方式简单实用,推荐,比如说使用下面的shell脚本我们就可以轻松的启动多个进程去处理 12345678#!/bin/bash for((i=1;i<=8;i++))do /usr/bi
process
php多进程编程应用 jwang的博客 在日常开发中,我们经常会遇到需要使用脚本处理一些数据,在数据量比较大的情况下,我们可以采用并行的方式处理,比如说: 启动多个实例这种方式简单实用,推荐,比如说使用下面的shell脚本我们就可以轻松的启动多个进程去处理 bin bash for i i i do usr bi
1
21,418
29,359,591,047
IssuesEvent
2023-05-28 00:36:45
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Brasília] Fullstack Developer (Bolsa CNPq) na Coodesh
SALVADOR INFRAESTRUTURA JAVA MYSQL JAVASCRIPT FULL-STACK HTML POSTGRESQL REQUISITOS PROCESSOS GITHUB UMA DOCUMENTAÇÃO SUBVERSION TREINAMENTO HARDWARE ALOCADO Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/bolsa-cnpq-145745456?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Alelo Recursos Genéticos</strong> está em busca de <strong><ins>Fullstack Developer (Bolsa CNPq)</ins></strong> para compor seu time!<br></p> <p>A Plataforma Alelo Recursos Genéticos é uma plataforma de serviços de tecnologia da informação (TI), de base Web, dedicada às atividades de pesquisa em conservação e enriquecimento de recursos genéticos (RG) desenvolvida pela Empresa Brasileira de Pesquisa Agropecuária (Embrapa). A Plataforma Alelo é formada pela infraestrutura de informática (hardware), bases de dados e sistemas de tecnologia da informação (softwares), processos e serviços dedicados à gestão, treinamento e apoio à documentação de dados e informações (peopleware) das atividades desenvolvidas em núcleos de criação (animal), coleções biológicas (microrganismos), coleções e bancos de germoplasma (vegetal) com foco no tema RG.</p> ## Alelo Recursos Genéticos: <p>A Plataforma Alelo Recursos Genéticos é uma plataforma de serviços de tecnologia da informação (TI), de base Web, dedicada às atividades de pesquisa em conservação e enriquecimento de recursos genéticos (RG) desenvolvida pela Empresa Brasileira de Pesquisa Agropecuária (Embrapa). A Plataforma Alelo é formada pela infraestrutura de informática (hardware), bases de dados e sistemas de tecnologia da informação (softwares), processos e serviços dedicados à gestão, treinamento e apoio à documentação de dados e informações (peopleware) das atividades desenvolvidas em núcleos de criação (animal), coleções biológicas (microrganismos), coleções e bancos de germoplasma (vegetal) com foco no tema RG.</p> </p> ## Habilidades: - Java - Javascript - PostgreSQL - MySQL - HTML 5 - CSS 3 ## Local: Brasília ## Requisitos: - Experiência em projetos de P&D; - Currículo Lattes preenchido com a experiência; - Conhecimentos em Java; - Conhecimentos em Subversion; - Conhecimentos em JavaScript; - Conhecimentos em PostgresSQL e MySQL; - Conhecimentos em HTML/CSS. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer (Bolsa CNPq) na Alelo Recursos Genéticos](https://coodesh.com/vagas/bolsa-cnpq-145745456?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Categoria Full-Stack
1.0
[Brasília] Fullstack Developer (Bolsa CNPq) na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/bolsa-cnpq-145745456?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Alelo Recursos Genéticos</strong> está em busca de <strong><ins>Fullstack Developer (Bolsa CNPq)</ins></strong> para compor seu time!<br></p> <p>A Plataforma Alelo Recursos Genéticos é uma plataforma de serviços de tecnologia da informação (TI), de base Web, dedicada às atividades de pesquisa em conservação e enriquecimento de recursos genéticos (RG) desenvolvida pela Empresa Brasileira de Pesquisa Agropecuária (Embrapa). A Plataforma Alelo é formada pela infraestrutura de informática (hardware), bases de dados e sistemas de tecnologia da informação (softwares), processos e serviços dedicados à gestão, treinamento e apoio à documentação de dados e informações (peopleware) das atividades desenvolvidas em núcleos de criação (animal), coleções biológicas (microrganismos), coleções e bancos de germoplasma (vegetal) com foco no tema RG.</p> ## Alelo Recursos Genéticos: <p>A Plataforma Alelo Recursos Genéticos é uma plataforma de serviços de tecnologia da informação (TI), de base Web, dedicada às atividades de pesquisa em conservação e enriquecimento de recursos genéticos (RG) desenvolvida pela Empresa Brasileira de Pesquisa Agropecuária (Embrapa). A Plataforma Alelo é formada pela infraestrutura de informática (hardware), bases de dados e sistemas de tecnologia da informação (softwares), processos e serviços dedicados à gestão, treinamento e apoio à documentação de dados e informações (peopleware) das atividades desenvolvidas em núcleos de criação (animal), coleções biológicas (microrganismos), coleções e bancos de germoplasma (vegetal) com foco no tema RG.</p> </p> ## Habilidades: - Java - Javascript - PostgreSQL - MySQL - HTML 5 - CSS 3 ## Local: Brasília ## Requisitos: - Experiência em projetos de P&D; - Currículo Lattes preenchido com a experiência; - Conhecimentos em Java; - Conhecimentos em Subversion; - Conhecimentos em JavaScript; - Conhecimentos em PostgresSQL e MySQL; - Conhecimentos em HTML/CSS. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer (Bolsa CNPq) na Alelo Recursos Genéticos](https://coodesh.com/vagas/bolsa-cnpq-145745456?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Categoria Full-Stack
process
fullstack developer bolsa cnpq na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a alelo recursos genéticos está em busca de fullstack developer bolsa cnpq para compor seu time a plataforma alelo recursos genéticos é uma plataforma de serviços de tecnologia da informação ti de base web dedicada às atividades de pesquisa em conservação e enriquecimento de recursos genéticos rg desenvolvida pela empresa brasileira de pesquisa agropecuária embrapa a plataforma alelo é formada pela infraestrutura de informática hardware bases de dados e sistemas de tecnologia da informação softwares processos e serviços dedicados à gestão treinamento e apoio à documentação de dados e informações peopleware das atividades desenvolvidas em núcleos de criação animal coleções biológicas microrganismos coleções e bancos de germoplasma vegetal com foco no tema rg alelo recursos genéticos a plataforma alelo recursos genéticos é uma plataforma de serviços de tecnologia da informação ti de base web dedicada às atividades de pesquisa em conservação e enriquecimento de recursos genéticos rg desenvolvida pela empresa brasileira de pesquisa agropecuária embrapa a plataforma alelo é formada pela infraestrutura de informática hardware bases de dados e sistemas de tecnologia da informação softwares processos e serviços dedicados à gestão treinamento e apoio à documentação de dados e informações peopleware das atividades desenvolvidas em núcleos de criação animal coleções biológicas microrganismos coleções e bancos de germoplasma vegetal com foco no tema rg habilidades java javascript postgresql mysql html css local brasília requisitos experiência em projetos de p d currículo lattes preenchido com a experiência conhecimentos em java conhecimentos em subversion conhecimentos em javascript conhecimentos em postgressql e mysql conhecimentos em html css como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado categoria full stack
1
15,278
19,269,459,057
IssuesEvent
2021-12-10 02:12:20
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
closed
Add container test for bazel-bot Docker image
type: process
We should have an easy way to ensure the bazel-bot image contains what we want. See https://github.com/googleapis/repo-automation-bots/blob/main/packages/release-trigger/cloudbuild-test.yaml as an example
1.0
Add container test for bazel-bot Docker image - We should have an easy way to ensure the bazel-bot image contains what we want. See https://github.com/googleapis/repo-automation-bots/blob/main/packages/release-trigger/cloudbuild-test.yaml as an example
process
add container test for bazel bot docker image we should have an easy way to ensure the bazel bot image contains what we want see as an example
1
6,754
9,881,537,875
IssuesEvent
2019-06-24 14:53:22
usgpo/bill-status
https://api.github.com/repos/usgpo/bill-status
closed
To Include Bill Version in Bill Status Document
reprocessing files working on it
Legislation may be changed or amended as it makes its way through the legislative process. You will sometimes see different bill text versions in the Text tab of a bill record. A list of the different versions are in the [link](https://www.gpo.gov/help/about_congressional_bills.htm). Versions are needed for accessing bills the following xml pages: [Bill Summary (In URL)](https://www.gpo.gov/fdsys/pkg/BILLS-115hconres1eh/html/BILLS-115hconres1eh.htm) [CONGRESSIONAL BILLS - BULK DATA](https://www.gpo.gov/fdsys/bulkdata/BILLS/115/1/hr)
1.0
To Include Bill Version in Bill Status Document - Legislation may be changed or amended as it makes its way through the legislative process. You will sometimes see different bill text versions in the Text tab of a bill record. A list of the different versions are in the [link](https://www.gpo.gov/help/about_congressional_bills.htm). Versions are needed for accessing bills the following xml pages: [Bill Summary (In URL)](https://www.gpo.gov/fdsys/pkg/BILLS-115hconres1eh/html/BILLS-115hconres1eh.htm) [CONGRESSIONAL BILLS - BULK DATA](https://www.gpo.gov/fdsys/bulkdata/BILLS/115/1/hr)
process
to include bill version in bill status document legislation may be changed or amended as it makes its way through the legislative process you will sometimes see different bill text versions in the text tab of a bill record a list of the different versions are in the versions are needed for accessing bills the following xml pages
1
12,321
14,879,490,839
IssuesEvent
2021-01-20 07:46:39
lutraconsulting/qgis-crayfish-plugin
https://api.github.com/repos/lutraconsulting/qgis-crayfish-plugin
closed
Issues with polygon contours
critical bug processing
Polygons generated from contour processing algorithm are not very usable. They contain lots of very tight angles, sliver polygons, self-intersecting, duplicate nodes, polygons with less than 3 nodes and small areas: ![image](https://user-images.githubusercontent.com/2649047/77290193-8eecbc00-6cd3-11ea-9b87-da7e2459ecbd.png)
1.0
Issues with polygon contours - Polygons generated from contour processing algorithm are not very usable. They contain lots of very tight angles, sliver polygons, self-intersecting, duplicate nodes, polygons with less than 3 nodes and small areas: ![image](https://user-images.githubusercontent.com/2649047/77290193-8eecbc00-6cd3-11ea-9b87-da7e2459ecbd.png)
process
issues with polygon contours polygons generated from contour processing algorithm are not very usable they contain lots of very tight angles sliver polygons self intersecting duplicate nodes polygons with less than nodes and small areas
1
203,981
15,396,353,620
IssuesEvent
2021-03-03 20:31:07
mozilla/foundation.mozilla.org
https://api.github.com/repos/mozilla/foundation.mozilla.org
closed
Add a new CI task that actually tests pages
engineering testing
We currently have a Cypress CI task set up that exists purely to take snapshots for Percy to then look at. However, we should create a separate Cypress CI task to actually test our end points: do the pages exist? Do all their CSS and JS requirements load? etc. Not having this caught us out because a build rewrite changed the build for "mozfest.compiled.js" to "index.compiled.js" and this caused the mozfest site to be unable to load its JS bundle, which on prod (with DEBUG=False) means users were getting a server error 500. - [ ] start with copying all Percy tests, but removing the snapshot instructions and instead checking whether the pages even load
1.0
Add a new CI task that actually tests pages - We currently have a Cypress CI task set up that exists purely to take snapshots for Percy to then look at. However, we should create a separate Cypress CI task to actually test our end points: do the pages exist? Do all their CSS and JS requirements load? etc. Not having this caught us out because a build rewrite changed the build for "mozfest.compiled.js" to "index.compiled.js" and this caused the mozfest site to be unable to load its JS bundle, which on prod (with DEBUG=False) means users were getting a server error 500. - [ ] start with copying all Percy tests, but removing the snapshot instructions and instead checking whether the pages even load
non_process
add a new ci task that actually tests pages we currently have a cypress ci task set up that exists purely to take snapshots for percy to then look at however we should create a separate cypress ci task to actually test our end points do the pages exist do all their css and js requirements load etc not having this caught us out because a build rewrite changed the build for mozfest compiled js to index compiled js and this caused the mozfest site to be unable to load its js bundle which on prod with debug false means users were getting a server error start with copying all percy tests but removing the snapshot instructions and instead checking whether the pages even load
0
2,752
3,146,325,638
IssuesEvent
2015-09-14 22:21:46
cortoproject/corto
https://api.github.com/repos/cortoproject/corto
closed
Introduce hidden and alias members
Corto:TypeSystem Corto:Usability enhancement Style
In certain cases, a class will want to control the order in which members appear, to increase usability when providing an initializer. This can introduce problems, when the class inherits from another class, and the base class exposes members that should appear after the members of the subclass. A workaround has been to make the baseclass visibility PRIVATE, add a member to the subclass (usually with the same name) and copy the value of that member manually to the base class (usually in the constructor). An example: ```c++ class Foo :: b: int32 class Bar: Foo, private :: a: int32 b: int32 int32 construct(): super.a = this.a Bar o: 10, 20 // a = 10, b = 20 ``` That is not a very nice solution since it introduces a redundant member with the same value, which needs to be kept in sync manually. The following proposal introduces two extra features to the type system that address this problem in a cleaner way. The first new feature is `alias`. An alias is a member that occupies the same memory space as another member. An alias may not be declared in a context where the member it points to is also visible, and may also not point to a member it doesn't have access to (private). The above example with an alias: ```c++ class Foo :: b: int32 class Bar: Foo, local :: a: int32 alias b: Foo::b Bar o: 10, 20 // a = 10, b = 20 ``` Notice the usage of `local` here: a local member won't show up in an initializer, but doesn't protect the member from outside access. This approach has another drawback however, and that is that `Foo` in this context now needs to be local. Local has semantics associated with whether data is relevant outside the context of the application (does it need to be persistently stored) and this wouldn't be a very orthogonal solution. Hence the introduction of hidden: ```c++ class Foo :: b: int32 class Bar: Foo, hidden :: a: int32 alias b: Foo::b Bar o: 10, 20 // a = 10, b = 20 ``` Hidden simply hides the members of a and doesn't affect the way in which members can be accessed or should be serialized. Therefore, the combination of `hidden` and `alias` can be a powerful tool in controlling the order of members in initializers, and thus increase the usability of types. An alias shall always mirror the attributes of the original member (modifiers, state and weak) in order to not break the model. When the inheritance modifier(s) of the base class are more restrictive than the modifiers of the aliassed member itself, the most restrictive applies. In practice, that will mean that alias members are only useful for hidden members, or hidden base classes. For example: ```c++ class Foo :: b: int32 // global, public class Bar: Foo, private :: a: int32 alias b: Foo::b // NOT OK: can't access Foo::b ``` ```c++ class Foo :: b: int32 // global, public class Bar: Foo, local :: a: int32 alias b: Foo::b // NOT OK: useless, since a local members don't show up in initializers ``` ```c++ class Foo :: b: int32 // global, public class Bar: Foo, hidden :: a: int32 alias b: Foo::b // OK: member is hidden by inheritance modifier ``` ```c++ class Foo :: b: int32, hidden class Bar: Foo :: a: int32 alias b: Foo::b // OK: member is hidden itself ```
True
Introduce hidden and alias members - In certain cases, a class will want to control the order in which members appear, to increase usability when providing an initializer. This can introduce problems, when the class inherits from another class, and the base class exposes members that should appear after the members of the subclass. A workaround has been to make the baseclass visibility PRIVATE, add a member to the subclass (usually with the same name) and copy the value of that member manually to the base class (usually in the constructor). An example: ```c++ class Foo :: b: int32 class Bar: Foo, private :: a: int32 b: int32 int32 construct(): super.a = this.a Bar o: 10, 20 // a = 10, b = 20 ``` That is not a very nice solution since it introduces a redundant member with the same value, which needs to be kept in sync manually. The following proposal introduces two extra features to the type system that address this problem in a cleaner way. The first new feature is `alias`. An alias is a member that occupies the same memory space as another member. An alias may not be declared in a context where the member it points to is also visible, and may also not point to a member it doesn't have access to (private). The above example with an alias: ```c++ class Foo :: b: int32 class Bar: Foo, local :: a: int32 alias b: Foo::b Bar o: 10, 20 // a = 10, b = 20 ``` Notice the usage of `local` here: a local member won't show up in an initializer, but doesn't protect the member from outside access. This approach has another drawback however, and that is that `Foo` in this context now needs to be local. Local has semantics associated with whether data is relevant outside the context of the application (does it need to be persistently stored) and this wouldn't be a very orthogonal solution. Hence the introduction of hidden: ```c++ class Foo :: b: int32 class Bar: Foo, hidden :: a: int32 alias b: Foo::b Bar o: 10, 20 // a = 10, b = 20 ``` Hidden simply hides the members of a and doesn't affect the way in which members can be accessed or should be serialized. Therefore, the combination of `hidden` and `alias` can be a powerful tool in controlling the order of members in initializers, and thus increase the usability of types. An alias shall always mirror the attributes of the original member (modifiers, state and weak) in order to not break the model. When the inheritance modifier(s) of the base class are more restrictive than the modifiers of the aliassed member itself, the most restrictive applies. In practice, that will mean that alias members are only useful for hidden members, or hidden base classes. For example: ```c++ class Foo :: b: int32 // global, public class Bar: Foo, private :: a: int32 alias b: Foo::b // NOT OK: can't access Foo::b ``` ```c++ class Foo :: b: int32 // global, public class Bar: Foo, local :: a: int32 alias b: Foo::b // NOT OK: useless, since a local members don't show up in initializers ``` ```c++ class Foo :: b: int32 // global, public class Bar: Foo, hidden :: a: int32 alias b: Foo::b // OK: member is hidden by inheritance modifier ``` ```c++ class Foo :: b: int32, hidden class Bar: Foo :: a: int32 alias b: Foo::b // OK: member is hidden itself ```
non_process
introduce hidden and alias members in certain cases a class will want to control the order in which members appear to increase usability when providing an initializer this can introduce problems when the class inherits from another class and the base class exposes members that should appear after the members of the subclass a workaround has been to make the baseclass visibility private add a member to the subclass usually with the same name and copy the value of that member manually to the base class usually in the constructor an example c class foo b class bar foo private a b construct super a this a bar o a b that is not a very nice solution since it introduces a redundant member with the same value which needs to be kept in sync manually the following proposal introduces two extra features to the type system that address this problem in a cleaner way the first new feature is alias an alias is a member that occupies the same memory space as another member an alias may not be declared in a context where the member it points to is also visible and may also not point to a member it doesn t have access to private the above example with an alias c class foo b class bar foo local a alias b foo b bar o a b notice the usage of local here a local member won t show up in an initializer but doesn t protect the member from outside access this approach has another drawback however and that is that foo in this context now needs to be local local has semantics associated with whether data is relevant outside the context of the application does it need to be persistently stored and this wouldn t be a very orthogonal solution hence the introduction of hidden c class foo b class bar foo hidden a alias b foo b bar o a b hidden simply hides the members of a and doesn t affect the way in which members can be accessed or should be serialized therefore the combination of hidden and alias can be a powerful tool in controlling the order of members in initializers and thus increase the usability of types an alias shall always mirror the attributes of the original member modifiers state and weak in order to not break the model when the inheritance modifier s of the base class are more restrictive than the modifiers of the aliassed member itself the most restrictive applies in practice that will mean that alias members are only useful for hidden members or hidden base classes for example c class foo b global public class bar foo private a alias b foo b not ok can t access foo b c class foo b global public class bar foo local a alias b foo b not ok useless since a local members don t show up in initializers c class foo b global public class bar foo hidden a alias b foo b ok member is hidden by inheritance modifier c class foo b hidden class bar foo a alias b foo b ok member is hidden itself
0
9,165
12,219,631,321
IssuesEvent
2020-05-01 22:17:15
jyn514/rcc
https://api.github.com/repos/jyn514/rcc
opened
\n is a valid preprocessing token
bug preprocessor
### Expected behavior <!-- A description of what you expected to happen. You could also paste the output of another compiler, I like `clang -x c - -Wall -Wextra -pedantic` --> `\n` should be a valid preprocessor token so that it can be used in token pasting. Found in `/usr/include/sys/sysmacros.h`. glibc is cursed. ### Code <!-- The code that was not interpreted correctly goes here. This should also include the error message you got. --> ```c #define __SYSMACROS_DM(symbol) __SYSMACROS_DM1 (a b \n c) #define __SYSMACROS_DM1(...) __glibc_macro_warning (#__VA_ARGS__) __SYSMACROS_DM(a) <stdin>:1:53 error: invalid token: unknown token: '\' #define __SYSMACROS_DM(symbol) __SYSMACROS_DM1 (a b \n c) ^ ``` Output from `clang -E`: ```c __glibc_macro_warning ("a b \n c") ``` <!-- If you know where to find it, include the relevant part of the C standard There's a copy at http://port70.net/~nsz/c/c11/n1570.html --> http://port70.net/~nsz/c/c11/n1570.html#6.4p3: > The categories of preprocessing tokens are: header names, identifiers, preprocessing numbers, character constants, string literals, punctuators, and single non-white-space characters that do not lexically match the other preprocessing token categories. "If it's not whitespace, it's valid!" :facepalm:
1.0
\n is a valid preprocessing token - ### Expected behavior <!-- A description of what you expected to happen. You could also paste the output of another compiler, I like `clang -x c - -Wall -Wextra -pedantic` --> `\n` should be a valid preprocessor token so that it can be used in token pasting. Found in `/usr/include/sys/sysmacros.h`. glibc is cursed. ### Code <!-- The code that was not interpreted correctly goes here. This should also include the error message you got. --> ```c #define __SYSMACROS_DM(symbol) __SYSMACROS_DM1 (a b \n c) #define __SYSMACROS_DM1(...) __glibc_macro_warning (#__VA_ARGS__) __SYSMACROS_DM(a) <stdin>:1:53 error: invalid token: unknown token: '\' #define __SYSMACROS_DM(symbol) __SYSMACROS_DM1 (a b \n c) ^ ``` Output from `clang -E`: ```c __glibc_macro_warning ("a b \n c") ``` <!-- If you know where to find it, include the relevant part of the C standard There's a copy at http://port70.net/~nsz/c/c11/n1570.html --> http://port70.net/~nsz/c/c11/n1570.html#6.4p3: > The categories of preprocessing tokens are: header names, identifiers, preprocessing numbers, character constants, string literals, punctuators, and single non-white-space characters that do not lexically match the other preprocessing token categories. "If it's not whitespace, it's valid!" :facepalm:
process
n is a valid preprocessing token expected behavior a description of what you expected to happen you could also paste the output of another compiler i like clang x c wall wextra pedantic n should be a valid preprocessor token so that it can be used in token pasting found in usr include sys sysmacros h glibc is cursed code the code that was not interpreted correctly goes here this should also include the error message you got c define sysmacros dm symbol sysmacros a b n c define sysmacros glibc macro warning va args sysmacros dm a error invalid token unknown token define sysmacros dm symbol sysmacros a b n c output from clang e c glibc macro warning a b n c if you know where to find it include the relevant part of the c standard there s a copy at the categories of preprocessing tokens are header names identifiers preprocessing numbers character constants string literals punctuators and single non white space characters that do not lexically match the other preprocessing token categories if it s not whitespace it s valid facepalm
1
10,747
13,541,450,986
IssuesEvent
2020-09-16 15:54:02
department-of-veterans-affairs/notification-api
https://api.github.com/repos/department-of-veterans-affairs/notification-api
closed
Procure SSL certificates
Process Task Release
**Notes:** - Need to setup the domain - We need the commercial certs - We need the VA certs
1.0
Procure SSL certificates - **Notes:** - Need to setup the domain - We need the commercial certs - We need the VA certs
process
procure ssl certificates notes need to setup the domain we need the commercial certs we need the va certs
1
15,327
19,433,139,472
IssuesEvent
2021-12-21 14:16:00
threefoldtech/tfchain
https://api.github.com/repos/threefoldtech/tfchain
closed
npos: Test and document hard fork scenario in case grandpa is stalled
process_wontfix
Given the testing in #122 > If more than 1/3 of the nodes die, GRANDPA finalization dies. At this point, new validators will be elected (as described above). However, the GRANDPA validator set only updates if the block in which the update is triggered is finalized. Since GRANDPA is currently dead in this scenario, the block will never finalize, and GRANDPA will stall forever. To recover from this scenario, a hard fork scenario is a solution.
1.0
npos: Test and document hard fork scenario in case grandpa is stalled - Given the testing in #122 > If more than 1/3 of the nodes die, GRANDPA finalization dies. At this point, new validators will be elected (as described above). However, the GRANDPA validator set only updates if the block in which the update is triggered is finalized. Since GRANDPA is currently dead in this scenario, the block will never finalize, and GRANDPA will stall forever. To recover from this scenario, a hard fork scenario is a solution.
process
npos test and document hard fork scenario in case grandpa is stalled given the testing in if more than of the nodes die grandpa finalization dies at this point new validators will be elected as described above however the grandpa validator set only updates if the block in which the update is triggered is finalized since grandpa is currently dead in this scenario the block will never finalize and grandpa will stall forever to recover from this scenario a hard fork scenario is a solution
1
285,796
21,554,979,460
IssuesEvent
2022-04-30 08:50:38
twistedsistem/OCTGN-Marvel-Champions
https://api.github.com/repos/twistedsistem/OCTGN-Marvel-Champions
closed
Keyboard Shortcuts
documentation enhancement
Definitely one of the most useful things that can be added, in my opinion. Also, I just want to say that you've nailed the double-click actions. Definitely very intuitive for each card type. You've also nailed the add/remove for threat/damage! Based on 12-15 playthroughs, I've found these to be the most important actions to have a shortcut. As always, these are merely suggestions, and if nothing else possibly just a place to have discussion. ![image](https://user-images.githubusercontent.com/56445225/72705295-b6cb7200-3b20-11ea-9351-d908ceffa22d.png)
1.0
Keyboard Shortcuts - Definitely one of the most useful things that can be added, in my opinion. Also, I just want to say that you've nailed the double-click actions. Definitely very intuitive for each card type. You've also nailed the add/remove for threat/damage! Based on 12-15 playthroughs, I've found these to be the most important actions to have a shortcut. As always, these are merely suggestions, and if nothing else possibly just a place to have discussion. ![image](https://user-images.githubusercontent.com/56445225/72705295-b6cb7200-3b20-11ea-9351-d908ceffa22d.png)
non_process
keyboard shortcuts definitely one of the most useful things that can be added in my opinion also i just want to say that you ve nailed the double click actions definitely very intuitive for each card type you ve also nailed the add remove for threat damage based on playthroughs i ve found these to be the most important actions to have a shortcut as always these are merely suggestions and if nothing else possibly just a place to have discussion
0
27,231
4,939,525,681
IssuesEvent
2016-11-29 14:37:49
jfabry/LiveRobotProgramming
https://api.github.com/repos/jfabry/LiveRobotProgramming
closed
Import of a program that does not parse crashes the interpreter
Component-Parser Priority-High Type-Defect
We should check in LRPProgram>>exportedMachines and loadImports if the current ast is not a failure
1.0
Import of a program that does not parse crashes the interpreter - We should check in LRPProgram>>exportedMachines and loadImports if the current ast is not a failure
non_process
import of a program that does not parse crashes the interpreter we should check in lrpprogram exportedmachines and loadimports if the current ast is not a failure
0
19,304
10,367,593,689
IssuesEvent
2019-09-07 09:41:17
emscripten-core/emscripten
https://api.github.com/repos/emscripten-core/emscripten
closed
Unnecessary label variables
fastcomp performance wontfix
I looked for small testcases where OdinMonkey has more phis (even after EliminatePhis and other optimizations) than emscripten's LLVM IR. One of the interesting cases I found was emscripten/tests/emscripten_get_now.cpp (compiled with emcc -O2). One thing that's happening is that the Relooper is using a label variable to share control paths that aren't shared in the LLVM IR. To an optimizer that doesn't understand label variables, this makes it look like more blocks are reachable from more places, which means that some blocks no longer dominate some blocks they used to, which means that the optimizer inserts more phis. The generated JS code looks roughly like this: ``` o = 0; while (1) { ... if (...) { o = 6; break; } if (...) { o = 10; break; } if (...) { break; } } if (o == 6) { ... } else if (o == 10) { ... } else { ... } ``` In cases like this, it would be preferable to avoid the label variable by moving the ending code into the if statements inside the loop. I manually edited the JS file to try this, and it eliminated several of the interesting phis in the resulting MIR. This would intermix loop and non-loop blocks even more, but that's something we should fix in the JIT itself anyway [0]. [0] https://bugzilla.mozilla.org/show_bug.cgi?id=844779
True
Unnecessary label variables - I looked for small testcases where OdinMonkey has more phis (even after EliminatePhis and other optimizations) than emscripten's LLVM IR. One of the interesting cases I found was emscripten/tests/emscripten_get_now.cpp (compiled with emcc -O2). One thing that's happening is that the Relooper is using a label variable to share control paths that aren't shared in the LLVM IR. To an optimizer that doesn't understand label variables, this makes it look like more blocks are reachable from more places, which means that some blocks no longer dominate some blocks they used to, which means that the optimizer inserts more phis. The generated JS code looks roughly like this: ``` o = 0; while (1) { ... if (...) { o = 6; break; } if (...) { o = 10; break; } if (...) { break; } } if (o == 6) { ... } else if (o == 10) { ... } else { ... } ``` In cases like this, it would be preferable to avoid the label variable by moving the ending code into the if statements inside the loop. I manually edited the JS file to try this, and it eliminated several of the interesting phis in the resulting MIR. This would intermix loop and non-loop blocks even more, but that's something we should fix in the JIT itself anyway [0]. [0] https://bugzilla.mozilla.org/show_bug.cgi?id=844779
non_process
unnecessary label variables i looked for small testcases where odinmonkey has more phis even after eliminatephis and other optimizations than emscripten s llvm ir one of the interesting cases i found was emscripten tests emscripten get now cpp compiled with emcc one thing that s happening is that the relooper is using a label variable to share control paths that aren t shared in the llvm ir to an optimizer that doesn t understand label variables this makes it look like more blocks are reachable from more places which means that some blocks no longer dominate some blocks they used to which means that the optimizer inserts more phis the generated js code looks roughly like this o while if o break if o break if break if o else if o else in cases like this it would be preferable to avoid the label variable by moving the ending code into the if statements inside the loop i manually edited the js file to try this and it eliminated several of the interesting phis in the resulting mir this would intermix loop and non loop blocks even more but that s something we should fix in the jit itself anyway
0
19,525
25,835,760,502
IssuesEvent
2022-12-12 19:30:09
AcademySoftwareFoundation/OpenCue
https://api.github.com/repos/AcademySoftwareFoundation/OpenCue
closed
Upgrade log4j
process
**Describe the process** Upgrade to a version of log4j which is not affected by the Log4Shell vulnerability.
1.0
Upgrade log4j - **Describe the process** Upgrade to a version of log4j which is not affected by the Log4Shell vulnerability.
process
upgrade describe the process upgrade to a version of which is not affected by the vulnerability
1
5,022
7,845,729,742
IssuesEvent
2018-06-19 13:44:51
openvstorage/framework
https://api.github.com/repos/openvstorage/framework
closed
Saving a vdisk builds a huge sequence
process_wontfix type_bug
### Problem description When saving a vdisk, a huge sequence is generated. (56MB!) ``` ArakoonException: Unknown error code 0x26, server said: update Sequence([Assert ;"ovs_index_vdisk|volume_id|faf9b4e98b061d61f43807bc846f426971f4cc84";0; Set ;"ovs_index_vdisk|volume_id|faf9b4e98b061d61f43807bc846f426971f4cc84";55;"..."; Set ;"ovs_reverseindex_vpool_b0e66d3d-bd12-45e0-ba6a-224a510d485b|vdisks|e23014b6-da6d-4992-b818-5999c214ee5f";1;"..."; Replace ;"ovs_listcache_vdisk|ovs_list_a42d01310e7f8d8f1a0523a68714c362c12753fefc1de5dee9a9e22393bc52f5|__all";0; Replace ;"ovs_listcache_vdisk|ovs_list_a42d01310e7f8d8f1a0523a68714c362c12753fefc1de5dee9a9e22393bc52f5|parentsnapshot";0; Replace ;"ovs_listcache_vdisk|ovs_list_d *SNIP* ;"ovs_listcache_vdisk|ovs_list_e80bbf9602e3746a33d8b5ba378b0be2fa4f0fae6745b3f271c788b087afba3d|__all";0; Replace ;"ovs_listcache_vdisk|ovs_list_e80bbf9602e3746a33d8b5ba378b0be2fa4f0fae6745b3f271c788b087afba3d|parentsnapshot";0; Replace ;"ovs_listcache_vdisk|ovs_list_cde7e94cdb25a1a20fd708d33b2d8ebb96a99f69fd527dfb5cadf0f75844e0cd|__all";0; ... (475334 more) ]) has size 53484496 > 33554432 ``` ``` /opt/OpenvStorage/ovs/dal/dataobject.pyc in save(self, recursive, skip, _hook) 693 self._mutex_version.acquire(30) 694 self._persistent.set(self._key, self._data, transaction=transaction) --> 695 self._persistent.apply_transaction(transaction) 696 self._volatile.delete(self._key) 697 successful = True /usr/lib/python2.7/dist-packages/ovs_extensions/storage/persistent/pyrakoonstore.pyc in apply_transaction(self, transaction) 140 """ 141 try: --> 142 return self._client.apply_transaction(transaction) 143 except ArakoonAssertionFailed as assertion: 144 raise AssertException(assertion) /usr/lib/python2.7/dist-packages/ovs_extensions/db/arakoon/pyrakoon/client.pyc in apply_transaction(self, transaction) 198 Applies a transaction 199 """ --> 200 return PyrakoonClient._try(self._identifier, self._client.sequence, self._sequences[transaction], max_duration=1) 201 202 @staticmethod /usr/lib/python2.7/dist-packages/ovs_extensions/db/arakoon/pyrakoon/client.pyc in _try(identifier, method, *args, **kwargs) 212 start = time.time() 213 try: --> 214 return_value = method(*args, **kwargs) 215 except (ArakoonSockNotReadable, ArakoonSockReadNoBytes, ArakoonSockSendError): 216 logger.debug('Error during arakoon call {0}, retry'.format(method.__name__)) <update_argspec> in sequence(self, seq, sync) /usr/lib/python2.7/dist-packages/ovs_extensions/db/arakoon/pyrakoon/pyrakoon/compat.pyc in wrapped(*args, **kwargs) 165 raise 166 --> 167 raise new_exception 168 169 return wrapped ``` This error was given back by Arakoon because the sequence was too big. This should be looked into Statistics of the environment: - 76MB ovs-db size - +-500 vdisks
1.0
Saving a vdisk builds a huge sequence - ### Problem description When saving a vdisk, a huge sequence is generated. (56MB!) ``` ArakoonException: Unknown error code 0x26, server said: update Sequence([Assert ;"ovs_index_vdisk|volume_id|faf9b4e98b061d61f43807bc846f426971f4cc84";0; Set ;"ovs_index_vdisk|volume_id|faf9b4e98b061d61f43807bc846f426971f4cc84";55;"..."; Set ;"ovs_reverseindex_vpool_b0e66d3d-bd12-45e0-ba6a-224a510d485b|vdisks|e23014b6-da6d-4992-b818-5999c214ee5f";1;"..."; Replace ;"ovs_listcache_vdisk|ovs_list_a42d01310e7f8d8f1a0523a68714c362c12753fefc1de5dee9a9e22393bc52f5|__all";0; Replace ;"ovs_listcache_vdisk|ovs_list_a42d01310e7f8d8f1a0523a68714c362c12753fefc1de5dee9a9e22393bc52f5|parentsnapshot";0; Replace ;"ovs_listcache_vdisk|ovs_list_d *SNIP* ;"ovs_listcache_vdisk|ovs_list_e80bbf9602e3746a33d8b5ba378b0be2fa4f0fae6745b3f271c788b087afba3d|__all";0; Replace ;"ovs_listcache_vdisk|ovs_list_e80bbf9602e3746a33d8b5ba378b0be2fa4f0fae6745b3f271c788b087afba3d|parentsnapshot";0; Replace ;"ovs_listcache_vdisk|ovs_list_cde7e94cdb25a1a20fd708d33b2d8ebb96a99f69fd527dfb5cadf0f75844e0cd|__all";0; ... (475334 more) ]) has size 53484496 > 33554432 ``` ``` /opt/OpenvStorage/ovs/dal/dataobject.pyc in save(self, recursive, skip, _hook) 693 self._mutex_version.acquire(30) 694 self._persistent.set(self._key, self._data, transaction=transaction) --> 695 self._persistent.apply_transaction(transaction) 696 self._volatile.delete(self._key) 697 successful = True /usr/lib/python2.7/dist-packages/ovs_extensions/storage/persistent/pyrakoonstore.pyc in apply_transaction(self, transaction) 140 """ 141 try: --> 142 return self._client.apply_transaction(transaction) 143 except ArakoonAssertionFailed as assertion: 144 raise AssertException(assertion) /usr/lib/python2.7/dist-packages/ovs_extensions/db/arakoon/pyrakoon/client.pyc in apply_transaction(self, transaction) 198 Applies a transaction 199 """ --> 200 return PyrakoonClient._try(self._identifier, self._client.sequence, self._sequences[transaction], max_duration=1) 201 202 @staticmethod /usr/lib/python2.7/dist-packages/ovs_extensions/db/arakoon/pyrakoon/client.pyc in _try(identifier, method, *args, **kwargs) 212 start = time.time() 213 try: --> 214 return_value = method(*args, **kwargs) 215 except (ArakoonSockNotReadable, ArakoonSockReadNoBytes, ArakoonSockSendError): 216 logger.debug('Error during arakoon call {0}, retry'.format(method.__name__)) <update_argspec> in sequence(self, seq, sync) /usr/lib/python2.7/dist-packages/ovs_extensions/db/arakoon/pyrakoon/pyrakoon/compat.pyc in wrapped(*args, **kwargs) 165 raise 166 --> 167 raise new_exception 168 169 return wrapped ``` This error was given back by Arakoon because the sequence was too big. This should be looked into Statistics of the environment: - 76MB ovs-db size - +-500 vdisks
process
saving a vdisk builds a huge sequence problem description when saving a vdisk a huge sequence is generated arakoonexception unknown error code server said update sequence assert ovs index vdisk volume id set ovs index vdisk volume id set ovs reverseindex vpool vdisks replace ovs listcache vdisk ovs list all replace ovs listcache vdisk ovs list parentsnapshot replace ovs listcache vdisk ovs list d snip ovs listcache vdisk ovs list all replace ovs listcache vdisk ovs list parentsnapshot replace ovs listcache vdisk ovs list all more has size opt openvstorage ovs dal dataobject pyc in save self recursive skip hook self mutex version acquire self persistent set self key self data transaction transaction self persistent apply transaction transaction self volatile delete self key successful true usr lib dist packages ovs extensions storage persistent pyrakoonstore pyc in apply transaction self transaction try return self client apply transaction transaction except arakoonassertionfailed as assertion raise assertexception assertion usr lib dist packages ovs extensions db arakoon pyrakoon client pyc in apply transaction self transaction applies a transaction return pyrakoonclient try self identifier self client sequence self sequences max duration staticmethod usr lib dist packages ovs extensions db arakoon pyrakoon client pyc in try identifier method args kwargs start time time try return value method args kwargs except arakoonsocknotreadable arakoonsockreadnobytes arakoonsocksenderror logger debug error during arakoon call retry format method name in sequence self seq sync usr lib dist packages ovs extensions db arakoon pyrakoon pyrakoon compat pyc in wrapped args kwargs raise raise new exception return wrapped this error was given back by arakoon because the sequence was too big this should be looked into statistics of the environment ovs db size vdisks
1
290,245
32,045,667,578
IssuesEvent
2023-09-23 01:34:27
Chiencc/asuswrt-gt-ac5300
https://api.github.com/repos/Chiencc/asuswrt-gt-ac5300
reopened
jquery-ui-1.10.3.min.js: 5 vulnerabilities (highest severity is: 6.1)
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-ui-1.10.3.min.js</b></p></summary> <p>A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js">https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js</a></p> <p>Path to dependency file: /release/src/router/www/AdaptiveQoS_Bandwidth_Monitor.asp</p> <p>Path to vulnerable library: /release/src/router/www/calendar/jquery-ui.js,/release/src/router/www/device-map/../calendar/jquery-ui.js,/release/src/router/www/calendar/jquery-ui.js</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jquery-ui version) | Remediation Possible** | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2021-41184](https://www.mend.io/vulnerability-database/CVE-2021-41184) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | jquery-ui-1.10.3.min.js | Direct | jquery-ui - 1.13.0 | &#10060; | | [CVE-2021-41183](https://www.mend.io/vulnerability-database/CVE-2021-41183) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | jquery-ui-1.10.3.min.js | Direct | jquery-ui - 1.13.0 | &#10060; | | [CVE-2021-41182](https://www.mend.io/vulnerability-database/CVE-2021-41182) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | jquery-ui-1.10.3.min.js | Direct | jquery-ui - 1.13.0 | &#10060; | | [CVE-2022-31160](https://www.mend.io/vulnerability-database/CVE-2022-31160) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | jquery-ui-1.10.3.min.js | Direct | jquery-ui - 1.13.2 | &#10060; | | [CVE-2016-7103](https://www.mend.io/vulnerability-database/CVE-2016-7103) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | jquery-ui-1.10.3.min.js | Direct | katello - 4.7.2 | &#10060; | <p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-41184</summary> ### Vulnerable Library - <b>jquery-ui-1.10.3.min.js</b></p> <p>A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js">https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js</a></p> <p>Path to dependency file: /release/src/router/www/AdaptiveQoS_Bandwidth_Monitor.asp</p> <p>Path to vulnerable library: /release/src/router/www/calendar/jquery-ui.js,/release/src/router/www/device-map/../calendar/jquery-ui.js,/release/src/router/www/calendar/jquery-ui.js</p> <p> Dependency Hierarchy: - :x: **jquery-ui-1.10.3.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery-UI is the official jQuery user interface library. Prior to version 1.13.0, accepting the value of the `of` option of the `.position()` util from untrusted sources may execute untrusted code. The issue is fixed in jQuery UI 1.13.0. Any string value passed to the `of` option is now treated as a CSS selector. A workaround is to not accept the value of the `of` option from untrusted sources. <p>Publish Date: 2021-10-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-41184>CVE-2021-41184</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-41184">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-41184</a></p> <p>Release Date: 2021-10-26</p> <p>Fix Resolution: jquery-ui - 1.13.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-41183</summary> ### Vulnerable Library - <b>jquery-ui-1.10.3.min.js</b></p> <p>A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js">https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js</a></p> <p>Path to dependency file: /release/src/router/www/AdaptiveQoS_Bandwidth_Monitor.asp</p> <p>Path to vulnerable library: /release/src/router/www/calendar/jquery-ui.js,/release/src/router/www/device-map/../calendar/jquery-ui.js,/release/src/router/www/calendar/jquery-ui.js</p> <p> Dependency Hierarchy: - :x: **jquery-ui-1.10.3.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery-UI is the official jQuery user interface library. Prior to version 1.13.0, accepting the value of various `*Text` options of the Datepicker widget from untrusted sources may execute untrusted code. The issue is fixed in jQuery UI 1.13.0. The values passed to various `*Text` options are now always treated as pure text, not HTML. A workaround is to not accept the value of the `*Text` options from untrusted sources. <p>Publish Date: 2021-10-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-41183>CVE-2021-41183</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-41183">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-41183</a></p> <p>Release Date: 2021-10-26</p> <p>Fix Resolution: jquery-ui - 1.13.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-41182</summary> ### Vulnerable Library - <b>jquery-ui-1.10.3.min.js</b></p> <p>A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js">https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js</a></p> <p>Path to dependency file: /release/src/router/www/AdaptiveQoS_Bandwidth_Monitor.asp</p> <p>Path to vulnerable library: /release/src/router/www/calendar/jquery-ui.js,/release/src/router/www/device-map/../calendar/jquery-ui.js,/release/src/router/www/calendar/jquery-ui.js</p> <p> Dependency Hierarchy: - :x: **jquery-ui-1.10.3.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery-UI is the official jQuery user interface library. Prior to version 1.13.0, accepting the value of the `altField` option of the Datepicker widget from untrusted sources may execute untrusted code. The issue is fixed in jQuery UI 1.13.0. Any string value passed to the `altField` option is now treated as a CSS selector. A workaround is to not accept the value of the `altField` option from untrusted sources. <p>Publish Date: 2021-10-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-41182>CVE-2021-41182</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-41182">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-41182</a></p> <p>Release Date: 2021-10-26</p> <p>Fix Resolution: jquery-ui - 1.13.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2022-31160</summary> ### Vulnerable Library - <b>jquery-ui-1.10.3.min.js</b></p> <p>A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js">https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js</a></p> <p>Path to dependency file: /release/src/router/www/AdaptiveQoS_Bandwidth_Monitor.asp</p> <p>Path to vulnerable library: /release/src/router/www/calendar/jquery-ui.js,/release/src/router/www/device-map/../calendar/jquery-ui.js,/release/src/router/www/calendar/jquery-ui.js</p> <p> Dependency Hierarchy: - :x: **jquery-ui-1.10.3.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery UI is a curated set of user interface interactions, effects, widgets, and themes built on top of jQuery. Versions prior to 1.13.2 are potentially vulnerable to cross-site scripting. Initializing a checkboxradio widget on an input enclosed within a label makes that parent label contents considered as the input label. Calling `.checkboxradio( "refresh" )` on such a widget and the initial HTML contained encoded HTML entities will make them erroneously get decoded. This can lead to potentially executing JavaScript code. The bug has been patched in jQuery UI 1.13.2. To remediate the issue, someone who can change the initial HTML can wrap all the non-input contents of the `label` in a `span`. <p>Publish Date: 2022-07-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-31160>CVE-2022-31160</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-31160">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-31160</a></p> <p>Release Date: 2022-07-20</p> <p>Fix Resolution: jquery-ui - 1.13.2</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2016-7103</summary> ### Vulnerable Library - <b>jquery-ui-1.10.3.min.js</b></p> <p>A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js">https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js</a></p> <p>Path to dependency file: /release/src/router/www/AdaptiveQoS_Bandwidth_Monitor.asp</p> <p>Path to vulnerable library: /release/src/router/www/calendar/jquery-ui.js,/release/src/router/www/device-map/../calendar/jquery-ui.js,/release/src/router/www/calendar/jquery-ui.js</p> <p> Dependency Hierarchy: - :x: **jquery-ui-1.10.3.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Cross-site scripting (XSS) vulnerability in jQuery UI before 1.12.0 might allow remote attackers to inject arbitrary web script or HTML via the closeText parameter of the dialog function. <p>Publish Date: 2017-03-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-7103>CVE-2016-7103</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2017-03-15</p> <p>Fix Resolution: katello - 4.7.2</p> </p> <p></p> </details>
True
jquery-ui-1.10.3.min.js: 5 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-ui-1.10.3.min.js</b></p></summary> <p>A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js">https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js</a></p> <p>Path to dependency file: /release/src/router/www/AdaptiveQoS_Bandwidth_Monitor.asp</p> <p>Path to vulnerable library: /release/src/router/www/calendar/jquery-ui.js,/release/src/router/www/device-map/../calendar/jquery-ui.js,/release/src/router/www/calendar/jquery-ui.js</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jquery-ui version) | Remediation Possible** | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2021-41184](https://www.mend.io/vulnerability-database/CVE-2021-41184) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | jquery-ui-1.10.3.min.js | Direct | jquery-ui - 1.13.0 | &#10060; | | [CVE-2021-41183](https://www.mend.io/vulnerability-database/CVE-2021-41183) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | jquery-ui-1.10.3.min.js | Direct | jquery-ui - 1.13.0 | &#10060; | | [CVE-2021-41182](https://www.mend.io/vulnerability-database/CVE-2021-41182) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | jquery-ui-1.10.3.min.js | Direct | jquery-ui - 1.13.0 | &#10060; | | [CVE-2022-31160](https://www.mend.io/vulnerability-database/CVE-2022-31160) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | jquery-ui-1.10.3.min.js | Direct | jquery-ui - 1.13.2 | &#10060; | | [CVE-2016-7103](https://www.mend.io/vulnerability-database/CVE-2016-7103) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | jquery-ui-1.10.3.min.js | Direct | katello - 4.7.2 | &#10060; | <p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-41184</summary> ### Vulnerable Library - <b>jquery-ui-1.10.3.min.js</b></p> <p>A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js">https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js</a></p> <p>Path to dependency file: /release/src/router/www/AdaptiveQoS_Bandwidth_Monitor.asp</p> <p>Path to vulnerable library: /release/src/router/www/calendar/jquery-ui.js,/release/src/router/www/device-map/../calendar/jquery-ui.js,/release/src/router/www/calendar/jquery-ui.js</p> <p> Dependency Hierarchy: - :x: **jquery-ui-1.10.3.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery-UI is the official jQuery user interface library. Prior to version 1.13.0, accepting the value of the `of` option of the `.position()` util from untrusted sources may execute untrusted code. The issue is fixed in jQuery UI 1.13.0. Any string value passed to the `of` option is now treated as a CSS selector. A workaround is to not accept the value of the `of` option from untrusted sources. <p>Publish Date: 2021-10-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-41184>CVE-2021-41184</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-41184">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-41184</a></p> <p>Release Date: 2021-10-26</p> <p>Fix Resolution: jquery-ui - 1.13.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-41183</summary> ### Vulnerable Library - <b>jquery-ui-1.10.3.min.js</b></p> <p>A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js">https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js</a></p> <p>Path to dependency file: /release/src/router/www/AdaptiveQoS_Bandwidth_Monitor.asp</p> <p>Path to vulnerable library: /release/src/router/www/calendar/jquery-ui.js,/release/src/router/www/device-map/../calendar/jquery-ui.js,/release/src/router/www/calendar/jquery-ui.js</p> <p> Dependency Hierarchy: - :x: **jquery-ui-1.10.3.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery-UI is the official jQuery user interface library. Prior to version 1.13.0, accepting the value of various `*Text` options of the Datepicker widget from untrusted sources may execute untrusted code. The issue is fixed in jQuery UI 1.13.0. The values passed to various `*Text` options are now always treated as pure text, not HTML. A workaround is to not accept the value of the `*Text` options from untrusted sources. <p>Publish Date: 2021-10-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-41183>CVE-2021-41183</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-41183">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-41183</a></p> <p>Release Date: 2021-10-26</p> <p>Fix Resolution: jquery-ui - 1.13.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-41182</summary> ### Vulnerable Library - <b>jquery-ui-1.10.3.min.js</b></p> <p>A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js">https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js</a></p> <p>Path to dependency file: /release/src/router/www/AdaptiveQoS_Bandwidth_Monitor.asp</p> <p>Path to vulnerable library: /release/src/router/www/calendar/jquery-ui.js,/release/src/router/www/device-map/../calendar/jquery-ui.js,/release/src/router/www/calendar/jquery-ui.js</p> <p> Dependency Hierarchy: - :x: **jquery-ui-1.10.3.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery-UI is the official jQuery user interface library. Prior to version 1.13.0, accepting the value of the `altField` option of the Datepicker widget from untrusted sources may execute untrusted code. The issue is fixed in jQuery UI 1.13.0. Any string value passed to the `altField` option is now treated as a CSS selector. A workaround is to not accept the value of the `altField` option from untrusted sources. <p>Publish Date: 2021-10-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-41182>CVE-2021-41182</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-41182">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-41182</a></p> <p>Release Date: 2021-10-26</p> <p>Fix Resolution: jquery-ui - 1.13.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2022-31160</summary> ### Vulnerable Library - <b>jquery-ui-1.10.3.min.js</b></p> <p>A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js">https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js</a></p> <p>Path to dependency file: /release/src/router/www/AdaptiveQoS_Bandwidth_Monitor.asp</p> <p>Path to vulnerable library: /release/src/router/www/calendar/jquery-ui.js,/release/src/router/www/device-map/../calendar/jquery-ui.js,/release/src/router/www/calendar/jquery-ui.js</p> <p> Dependency Hierarchy: - :x: **jquery-ui-1.10.3.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery UI is a curated set of user interface interactions, effects, widgets, and themes built on top of jQuery. Versions prior to 1.13.2 are potentially vulnerable to cross-site scripting. Initializing a checkboxradio widget on an input enclosed within a label makes that parent label contents considered as the input label. Calling `.checkboxradio( "refresh" )` on such a widget and the initial HTML contained encoded HTML entities will make them erroneously get decoded. This can lead to potentially executing JavaScript code. The bug has been patched in jQuery UI 1.13.2. To remediate the issue, someone who can change the initial HTML can wrap all the non-input contents of the `label` in a `span`. <p>Publish Date: 2022-07-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-31160>CVE-2022-31160</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-31160">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-31160</a></p> <p>Release Date: 2022-07-20</p> <p>Fix Resolution: jquery-ui - 1.13.2</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2016-7103</summary> ### Vulnerable Library - <b>jquery-ui-1.10.3.min.js</b></p> <p>A curated set of user interface interactions, effects, widgets, and themes built on top of the jQuery JavaScript Library.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js">https://cdnjs.cloudflare.com/ajax/libs/jqueryui/1.10.3/jquery-ui.min.js</a></p> <p>Path to dependency file: /release/src/router/www/AdaptiveQoS_Bandwidth_Monitor.asp</p> <p>Path to vulnerable library: /release/src/router/www/calendar/jquery-ui.js,/release/src/router/www/device-map/../calendar/jquery-ui.js,/release/src/router/www/calendar/jquery-ui.js</p> <p> Dependency Hierarchy: - :x: **jquery-ui-1.10.3.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Cross-site scripting (XSS) vulnerability in jQuery UI before 1.12.0 might allow remote attackers to inject arbitrary web script or HTML via the closeText parameter of the dialog function. <p>Publish Date: 2017-03-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-7103>CVE-2016-7103</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2017-03-15</p> <p>Fix Resolution: katello - 4.7.2</p> </p> <p></p> </details>
non_process
jquery ui min js vulnerabilities highest severity is vulnerable library jquery ui min js a curated set of user interface interactions effects widgets and themes built on top of the jquery javascript library library home page a href path to dependency file release src router www adaptiveqos bandwidth monitor asp path to vulnerable library release src router www calendar jquery ui js release src router www device map calendar jquery ui js release src router www calendar jquery ui js found in head commit a href vulnerabilities cve severity cvss dependency type fixed in jquery ui version remediation possible medium jquery ui min js direct jquery ui medium jquery ui min js direct jquery ui medium jquery ui min js direct jquery ui medium jquery ui min js direct jquery ui medium jquery ui min js direct katello in some cases remediation pr cannot be created automatically for a vulnerability despite the availability of remediation details cve vulnerable library jquery ui min js a curated set of user interface interactions effects widgets and themes built on top of the jquery javascript library library home page a href path to dependency file release src router www adaptiveqos bandwidth monitor asp path to vulnerable library release src router www calendar jquery ui js release src router www device map calendar jquery ui js release src router www calendar jquery ui js dependency hierarchy x jquery ui min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery ui is the official jquery user interface library prior to version accepting the value of the of option of the position util from untrusted sources may execute untrusted code the issue is fixed in jquery ui any string value passed to the of option is now treated as a css selector a workaround is to not accept the value of the of option from untrusted sources publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery ui cve vulnerable library jquery ui min js a curated set of user interface interactions effects widgets and themes built on top of the jquery javascript library library home page a href path to dependency file release src router www adaptiveqos bandwidth monitor asp path to vulnerable library release src router www calendar jquery ui js release src router www device map calendar jquery ui js release src router www calendar jquery ui js dependency hierarchy x jquery ui min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery ui is the official jquery user interface library prior to version accepting the value of various text options of the datepicker widget from untrusted sources may execute untrusted code the issue is fixed in jquery ui the values passed to various text options are now always treated as pure text not html a workaround is to not accept the value of the text options from untrusted sources publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery ui cve vulnerable library jquery ui min js a curated set of user interface interactions effects widgets and themes built on top of the jquery javascript library library home page a href path to dependency file release src router www adaptiveqos bandwidth monitor asp path to vulnerable library release src router www calendar jquery ui js release src router www device map calendar jquery ui js release src router www calendar jquery ui js dependency hierarchy x jquery ui min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery ui is the official jquery user interface library prior to version accepting the value of the altfield option of the datepicker widget from untrusted sources may execute untrusted code the issue is fixed in jquery ui any string value passed to the altfield option is now treated as a css selector a workaround is to not accept the value of the altfield option from untrusted sources publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery ui cve vulnerable library jquery ui min js a curated set of user interface interactions effects widgets and themes built on top of the jquery javascript library library home page a href path to dependency file release src router www adaptiveqos bandwidth monitor asp path to vulnerable library release src router www calendar jquery ui js release src router www device map calendar jquery ui js release src router www calendar jquery ui js dependency hierarchy x jquery ui min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery ui is a curated set of user interface interactions effects widgets and themes built on top of jquery versions prior to are potentially vulnerable to cross site scripting initializing a checkboxradio widget on an input enclosed within a label makes that parent label contents considered as the input label calling checkboxradio refresh on such a widget and the initial html contained encoded html entities will make them erroneously get decoded this can lead to potentially executing javascript code the bug has been patched in jquery ui to remediate the issue someone who can change the initial html can wrap all the non input contents of the label in a span publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery ui cve vulnerable library jquery ui min js a curated set of user interface interactions effects widgets and themes built on top of the jquery javascript library library home page a href path to dependency file release src router www adaptiveqos bandwidth monitor asp path to vulnerable library release src router www calendar jquery ui js release src router www device map calendar jquery ui js release src router www calendar jquery ui js dependency hierarchy x jquery ui min js vulnerable library found in head commit a href found in base branch master vulnerability details cross site scripting xss vulnerability in jquery ui before might allow remote attackers to inject arbitrary web script or html via the closetext parameter of the dialog function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution katello
0
21,222
28,307,230,279
IssuesEvent
2023-04-10 12:15:58
ASWE-PDA/Personal-Digital-Agent
https://api.github.com/repos/ASWE-PDA/Personal-Digital-Agent
reopened
Speech-to-text
enhancement adapter speech processing
As a user, I want the agent to understand what I'm saying so that the agent can work according to my statements. Acceptance criteria: - [x] API for speech-to-text has been selected - [x] API can be called by an adapter module in the backend
1.0
Speech-to-text - As a user, I want the agent to understand what I'm saying so that the agent can work according to my statements. Acceptance criteria: - [x] API for speech-to-text has been selected - [x] API can be called by an adapter module in the backend
process
speech to text as a user i want the agent to understand what i m saying so that the agent can work according to my statements acceptance criteria api for speech to text has been selected api can be called by an adapter module in the backend
1
246,954
20,946,810,105
IssuesEvent
2022-03-26 02:10:54
Uuvana-Studios/longvinter-windows-client
https://api.github.com/repos/Uuvana-Studios/longvinter-windows-client
closed
Duplicate item bug
bug High Priority Tested
**Describe the bug** When placing an item down such as a lamp post or a turret if you left click mutliple times quickly it will duplicate the item. **To Reproduce** Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** If applicable, add screenshots to help explain your problem. **Desktop (please complete the following information):** - OS: [e.g. Windows] - Game Version [e.g. 1.0] - Steam Version [e.g. 1.0] **Additional context** Add any other context about the problem here.
1.0
Duplicate item bug - **Describe the bug** When placing an item down such as a lamp post or a turret if you left click mutliple times quickly it will duplicate the item. **To Reproduce** Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** If applicable, add screenshots to help explain your problem. **Desktop (please complete the following information):** - OS: [e.g. Windows] - Game Version [e.g. 1.0] - Steam Version [e.g. 1.0] **Additional context** Add any other context about the problem here.
non_process
duplicate item bug describe the bug when placing an item down such as a lamp post or a turret if you left click mutliple times quickly it will duplicate the item to reproduce steps to reproduce the behavior go to click on scroll down to see error expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem desktop please complete the following information os game version steam version additional context add any other context about the problem here
0
210
2,636,664,195
IssuesEvent
2015-03-10 06:42:47
fabric8io/fabric8
https://api.github.com/repos/fabric8io/fabric8
closed
[Process Controller] Create file system IO API for process controller
process manager
It will be nice to have the API (somewhere near the `ProcessController` level) so process controller could read and write files within its sandbox.
1.0
[Process Controller] Create file system IO API for process controller - It will be nice to have the API (somewhere near the `ProcessController` level) so process controller could read and write files within its sandbox.
process
create file system io api for process controller it will be nice to have the api somewhere near the processcontroller level so process controller could read and write files within its sandbox
1
21,181
28,149,441,071
IssuesEvent
2023-04-02 21:28:58
zotero/zotero
https://api.github.com/repos/zotero/zotero
opened
Prevent selection of child items in classic citation dialog
Papercuts Word Processor Integration
https://forums.zotero.org/discussion/104070/one-bibliography-position-deviates-cause-not-found and many others Until we get rid of the classic citation dialog, we should prevent attachments and notes from being selected. Not sure if we can hide those easily in the tree — even if we only showed top-level items, there could be standalone attachments and notes — so maybe there just needs to be a check for item type with an error dialog that prevents accepting the dialog?
1.0
Prevent selection of child items in classic citation dialog - https://forums.zotero.org/discussion/104070/one-bibliography-position-deviates-cause-not-found and many others Until we get rid of the classic citation dialog, we should prevent attachments and notes from being selected. Not sure if we can hide those easily in the tree — even if we only showed top-level items, there could be standalone attachments and notes — so maybe there just needs to be a check for item type with an error dialog that prevents accepting the dialog?
process
prevent selection of child items in classic citation dialog and many others until we get rid of the classic citation dialog we should prevent attachments and notes from being selected not sure if we can hide those easily in the tree — even if we only showed top level items there could be standalone attachments and notes — so maybe there just needs to be a check for item type with an error dialog that prevents accepting the dialog
1
3,825
6,802,323,753
IssuesEvent
2017-11-02 19:47:34
WikiWatershed/model-my-watershed
https://api.github.com/repos/WikiWatershed/model-my-watershed
closed
Geoprocessing API: Validate analyze AOI's size
Geoprocessing API tested/verified WPF
### AOI Area Check that shapes submitted to the analyze endpoints have an area less than the [`MMW_MAX_AREA`](https://github.com/WikiWatershed/model-my-watershed/blob/develop/src/mmw/mmw/settings/base.py#L384). ### Payload size This check should already be performed by nginx, and [limit the request body size to 5mb](https://github.com/WikiWatershed/model-my-watershed/blob/develop/deployment/ansible/roles/model-my-watershed.app/templates/nginx-app.conf.j2#L11)
1.0
Geoprocessing API: Validate analyze AOI's size - ### AOI Area Check that shapes submitted to the analyze endpoints have an area less than the [`MMW_MAX_AREA`](https://github.com/WikiWatershed/model-my-watershed/blob/develop/src/mmw/mmw/settings/base.py#L384). ### Payload size This check should already be performed by nginx, and [limit the request body size to 5mb](https://github.com/WikiWatershed/model-my-watershed/blob/develop/deployment/ansible/roles/model-my-watershed.app/templates/nginx-app.conf.j2#L11)
process
geoprocessing api validate analyze aoi s size aoi area check that shapes submitted to the analyze endpoints have an area less than the payload size this check should already be performed by nginx and
1
12,069
8,582,909,160
IssuesEvent
2018-11-13 18:14:15
PowerShell/Announcements
https://api.github.com/repos/PowerShell/Announcements
opened
Microsoft Security Advisory CVE-2018-8415: Microsoft PowerShell Tampering Vulnerability
PowerShell Security
# Microsoft Security Advisory CVE-2018-8415: Microsoft PowerShell Tampering Vulnerability ## Executive Summary A tampering vulnerability exists in PowerShell that could allow an attacker to execute unlogged code. To exploit this vulnerability, an attacker would need to log on to the affected system and run a specially crafted application. The security update addresses the vulnerability by correcting log management of special characters.System administrators are advised to update PowerShell Core to version `6.0.5` or `6.1.1`. ## Discussion Please use PowerShell/PowerShell#8254 for discussion of this advisory. ## Affected Software The vulnerability affects PowerShell Core prior to the following versions: | PowerShell Core Version | Fixed in | |-------------------------|-------------------| | 6.0 | 6.0.5 | | 6.1 | 6.1.1 | | 6.2 | https://github.com/PowerShell/PowerShell/pull/8253 | ## Advisory FAQ ### How do I know if I am affected? If all of the following are true: 1. Run `pwsh -v`, then, check the version in the table in [Affected Software](#affected-software) to see if your version of PowerShell Core is affected. 1. If you are running a version of PowerShell Core where the executable is not `pwsh` or `pwsh.exe`, then you are affected. This only existed for preview version of `6.0`. ### How do I update to an unaffected version? Follow the instructions at [Installing PowerShell Core](https://docs.microsoft.com/en-us/powershell/scripting/setup/installing-powershell?view=powershell-6) to install the latest version of PowerShell Core. ## Other Information ### Commit IDs [f8f3774d](https://github.com/PowerShell/PowerShell/commit/f8f3774dfca94557ff05446c87192ea1edbb356d) ### Reporting Security Issues If you have found a potential security issue in PowerShell Core, please email details to secure@microsoft.com. ### Support You can ask questions about this issue on GitHub in the PowerShell organization. This is located at https://github.com/PowerShell/. The Announcements repo (https://github.com/PowerShell/Announcements) will contain this bulletin as an issue and will include a link to a discussion issue where you can ask questions. ### What if the update breaks my script or module? You can uninstall the newer version of PowerShell Core and install the previous version of PowerShell Core. This should be treated as a temporary measure. Therefore, the script or module should be updated to work with the patched version of PowerShell Core. ### Acknowledgments Microsoft recognizes the efforts of those in the security community who help us protect customers through coordinated vulnerability disclosure. See [acknowledgments](https://portal.msrc.microsoft.com/en-us/security-guidance/acknowledgments) for more information. ### External Links [CVE-2018-8415](https://portal.msrc.microsoft.com/en-US/security-guidance/advisory/CVE-2018-8415) ### Revisions V1.0 (November 13, 2018): Advisory published. *Version 1.0* *Last Updated 2018-11-13*
True
Microsoft Security Advisory CVE-2018-8415: Microsoft PowerShell Tampering Vulnerability - # Microsoft Security Advisory CVE-2018-8415: Microsoft PowerShell Tampering Vulnerability ## Executive Summary A tampering vulnerability exists in PowerShell that could allow an attacker to execute unlogged code. To exploit this vulnerability, an attacker would need to log on to the affected system and run a specially crafted application. The security update addresses the vulnerability by correcting log management of special characters.System administrators are advised to update PowerShell Core to version `6.0.5` or `6.1.1`. ## Discussion Please use PowerShell/PowerShell#8254 for discussion of this advisory. ## Affected Software The vulnerability affects PowerShell Core prior to the following versions: | PowerShell Core Version | Fixed in | |-------------------------|-------------------| | 6.0 | 6.0.5 | | 6.1 | 6.1.1 | | 6.2 | https://github.com/PowerShell/PowerShell/pull/8253 | ## Advisory FAQ ### How do I know if I am affected? If all of the following are true: 1. Run `pwsh -v`, then, check the version in the table in [Affected Software](#affected-software) to see if your version of PowerShell Core is affected. 1. If you are running a version of PowerShell Core where the executable is not `pwsh` or `pwsh.exe`, then you are affected. This only existed for preview version of `6.0`. ### How do I update to an unaffected version? Follow the instructions at [Installing PowerShell Core](https://docs.microsoft.com/en-us/powershell/scripting/setup/installing-powershell?view=powershell-6) to install the latest version of PowerShell Core. ## Other Information ### Commit IDs [f8f3774d](https://github.com/PowerShell/PowerShell/commit/f8f3774dfca94557ff05446c87192ea1edbb356d) ### Reporting Security Issues If you have found a potential security issue in PowerShell Core, please email details to secure@microsoft.com. ### Support You can ask questions about this issue on GitHub in the PowerShell organization. This is located at https://github.com/PowerShell/. The Announcements repo (https://github.com/PowerShell/Announcements) will contain this bulletin as an issue and will include a link to a discussion issue where you can ask questions. ### What if the update breaks my script or module? You can uninstall the newer version of PowerShell Core and install the previous version of PowerShell Core. This should be treated as a temporary measure. Therefore, the script or module should be updated to work with the patched version of PowerShell Core. ### Acknowledgments Microsoft recognizes the efforts of those in the security community who help us protect customers through coordinated vulnerability disclosure. See [acknowledgments](https://portal.msrc.microsoft.com/en-us/security-guidance/acknowledgments) for more information. ### External Links [CVE-2018-8415](https://portal.msrc.microsoft.com/en-US/security-guidance/advisory/CVE-2018-8415) ### Revisions V1.0 (November 13, 2018): Advisory published. *Version 1.0* *Last Updated 2018-11-13*
non_process
microsoft security advisory cve microsoft powershell tampering vulnerability microsoft security advisory cve microsoft powershell tampering vulnerability executive summary a tampering vulnerability exists in powershell that could allow an attacker to execute unlogged code to exploit this vulnerability an attacker would need to log on to the affected system and run a specially crafted application the security update addresses the vulnerability by correcting log management of special characters system administrators are advised to update powershell core to version or discussion please use powershell powershell for discussion of this advisory affected software the vulnerability affects powershell core prior to the following versions powershell core version fixed in advisory faq how do i know if i am affected if all of the following are true run pwsh v then check the version in the table in affected software to see if your version of powershell core is affected if you are running a version of powershell core where the executable is not pwsh or pwsh exe then you are affected this only existed for preview version of how do i update to an unaffected version follow the instructions at to install the latest version of powershell core other information commit ids reporting security issues if you have found a potential security issue in powershell core please email details to secure microsoft com support you can ask questions about this issue on github in the powershell organization this is located at the announcements repo will contain this bulletin as an issue and will include a link to a discussion issue where you can ask questions what if the update breaks my script or module you can uninstall the newer version of powershell core and install the previous version of powershell core this should be treated as a temporary measure therefore the script or module should be updated to work with the patched version of powershell core acknowledgments microsoft recognizes the efforts of those in the security community who help us protect customers through coordinated vulnerability disclosure see for more information external links revisions november advisory published version last updated
0