Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
58,469
6,599,523,733
IssuesEvent
2017-09-16 20:53:27
nskins/goby
https://api.github.com/repos/nskins/goby
closed
Refactor tests to use `let` instead of instance variables
better test suite
I noticed that most of the tests use a before action to assign instance variables that are used throughout the tests. This is not great for a couple of reasons: 1. Instance variables are created wherever they are referenced, so you are more prone to typos. 2. Oftentimes objects are being instantiated that are not being used in all of the tests, which is unnecessary. I suggest refactoring to use `let`, this will clean up the code, protect against inadvertent typos, and speed up run time (`let` is lazy-loaded so you won't be instantiating every object for every test). For example, the Escape Spec would now look like this: ```ruby #spec/goby/battle/escape_spec.rb RSpec.describe Goby::Escape do let(:player) { Player.new } let(:monster) { Monster.new } let(:escape) { Escape.new } context "constructor" do it "has an appropriate default name" do expect(escape.name).to eq "Escape" end end context "run" do # The purpose of this test is to run the code without error. it "should return a usable result" do # Exercise both branches of this function w/ high probability. 20.times do escape.run(player, monster) expect(player.escaped).to_not be nil end end end end ``` This is a pretty big refactor, I'm happy to take it on if you think it's worth it.
1.0
Refactor tests to use `let` instead of instance variables - I noticed that most of the tests use a before action to assign instance variables that are used throughout the tests. This is not great for a couple of reasons: 1. Instance variables are created wherever they are referenced, so you are more prone to typos. 2. Oftentimes objects are being instantiated that are not being used in all of the tests, which is unnecessary. I suggest refactoring to use `let`, this will clean up the code, protect against inadvertent typos, and speed up run time (`let` is lazy-loaded so you won't be instantiating every object for every test). For example, the Escape Spec would now look like this: ```ruby #spec/goby/battle/escape_spec.rb RSpec.describe Goby::Escape do let(:player) { Player.new } let(:monster) { Monster.new } let(:escape) { Escape.new } context "constructor" do it "has an appropriate default name" do expect(escape.name).to eq "Escape" end end context "run" do # The purpose of this test is to run the code without error. it "should return a usable result" do # Exercise both branches of this function w/ high probability. 20.times do escape.run(player, monster) expect(player.escaped).to_not be nil end end end end ``` This is a pretty big refactor, I'm happy to take it on if you think it's worth it.
test
refactor tests to use let instead of instance variables i noticed that most of the tests use a before action to assign instance variables that are used throughout the tests this is not great for a couple of reasons instance variables are created wherever they are referenced so you are more prone to typos oftentimes objects are being instantiated that are not being used in all of the tests which is unnecessary i suggest refactoring to use let this will clean up the code protect against inadvertent typos and speed up run time let is lazy loaded so you won t be instantiating every object for every test for example the escape spec would now look like this ruby spec goby battle escape spec rb rspec describe goby escape do let player player new let monster monster new let escape escape new context constructor do it has an appropriate default name do expect escape name to eq escape end end context run do the purpose of this test is to run the code without error it should return a usable result do exercise both branches of this function w high probability times do escape run player monster expect player escaped to not be nil end end end end this is a pretty big refactor i m happy to take it on if you think it s worth it
1
245,888
20,809,849,629
IssuesEvent
2022-03-18 00:27:40
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Blocked count highlight doesn't cover entire number
bug feature/shields priority/P3 needs-discussion QA/Yes QA/Test-Plan-Specified feature/shields/panel OS/Desktop
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Blocked count highlight doesn't cover entire number ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. install `1.38.35` 2. launch Brave 3. enable `Shields V2` via `brave://flags` 4. restart Brave 5. sit on an XHR-happy page (any Facebook one will do) 6. check the Shields-blocked count via the icon-tip in the URL bar 7. once it reaches `99+`, click on it to expand the flyout panel 8. click to open its `Advanced controls` sub-panel 9. hover over the number for `Trackers & ads blocked (standard)` 10. note the highlight paint region ## Actual result: <!--Please add screenshots if needed--> <img width="1312" alt="Screen Shot 2022-03-11 at 12 41 01 AM" src="https://user-images.githubusercontent.com/387249/157833842-8933eca0-785e-4dae-afca-da0b4e72ab21.png"> ## Expected result: Should paint the entire number with the hover highlight ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> 100% ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.38.35 Chromium: 99.0.4844.51 (Official Build) nightly (arm64) -- | -- Revision | d537ec02474b5afe23684e7963d538896c63ac77-refs/branch-heads/4844@{#875} OS | macOS Version 11.6.4 (Build 20G417) cc @nullhook @rebron @sri
1.0
Blocked count highlight doesn't cover entire number - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Blocked count highlight doesn't cover entire number ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. install `1.38.35` 2. launch Brave 3. enable `Shields V2` via `brave://flags` 4. restart Brave 5. sit on an XHR-happy page (any Facebook one will do) 6. check the Shields-blocked count via the icon-tip in the URL bar 7. once it reaches `99+`, click on it to expand the flyout panel 8. click to open its `Advanced controls` sub-panel 9. hover over the number for `Trackers & ads blocked (standard)` 10. note the highlight paint region ## Actual result: <!--Please add screenshots if needed--> <img width="1312" alt="Screen Shot 2022-03-11 at 12 41 01 AM" src="https://user-images.githubusercontent.com/387249/157833842-8933eca0-785e-4dae-afca-da0b4e72ab21.png"> ## Expected result: Should paint the entire number with the hover highlight ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> 100% ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.38.35 Chromium: 99.0.4844.51 (Official Build) nightly (arm64) -- | -- Revision | d537ec02474b5afe23684e7963d538896c63ac77-refs/branch-heads/4844@{#875} OS | macOS Version 11.6.4 (Build 20G417) cc @nullhook @rebron @sri
test
blocked count highlight doesn t cover entire number have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description blocked count highlight doesn t cover entire number steps to reproduce install launch brave enable shields via brave flags restart brave sit on an xhr happy page any facebook one will do check the shields blocked count via the icon tip in the url bar once it reaches click on it to expand the flyout panel click to open its advanced controls sub panel hover over the number for trackers ads blocked standard note the highlight paint region actual result img width alt screen shot at am src expected result should paint the entire number with the hover highlight reproduces how often brave version brave version info brave chromium   official build  nightly  revision refs branch heads os macos version build cc nullhook rebron sri
1
60,154
6,672,967,037
IssuesEvent
2017-10-04 13:39:55
w3c/web-platform-tests
https://api.github.com/repos/w3c/web-platform-tests
closed
sequential_async_test
infra testharness.js
Originally posted as https://github.com/w3c/testharness.js/issues/96 by @mvano on 09 Dec 2014, 14:16 UTC: > I'd like to run multiple async tests from a single page, but execute them sequentially, not in parallel. The idea would be for later tests to delay execution until the previous ones are done. > > The goal would be to make tests that share state (across the tests, or in the object under test) be deterministic. Here's a contrived example: > > ``` javascript > var i = 0; > > sequential_async_test(function(test) { > setTimeout(function() { > assert_equals(i, 0); > i++; > test.done(); > }, Math.round(Math.random() * 100)); > }, 'Async test 0'); > > sequential_async_test(function(test) { > setTimeout(function() { > assert_equals(i, 1); > i++; > test.done(); > }, Math.round(Math.random() * 100)); > }, 'Async test 1'); > ``` > > I guess the alternative is to make separate pages, each with a single `async_test`. I'll probably do that for the time being.
1.0
sequential_async_test - Originally posted as https://github.com/w3c/testharness.js/issues/96 by @mvano on 09 Dec 2014, 14:16 UTC: > I'd like to run multiple async tests from a single page, but execute them sequentially, not in parallel. The idea would be for later tests to delay execution until the previous ones are done. > > The goal would be to make tests that share state (across the tests, or in the object under test) be deterministic. Here's a contrived example: > > ``` javascript > var i = 0; > > sequential_async_test(function(test) { > setTimeout(function() { > assert_equals(i, 0); > i++; > test.done(); > }, Math.round(Math.random() * 100)); > }, 'Async test 0'); > > sequential_async_test(function(test) { > setTimeout(function() { > assert_equals(i, 1); > i++; > test.done(); > }, Math.round(Math.random() * 100)); > }, 'Async test 1'); > ``` > > I guess the alternative is to make separate pages, each with a single `async_test`. I'll probably do that for the time being.
test
sequential async test originally posted as by mvano on dec utc i d like to run multiple async tests from a single page but execute them sequentially not in parallel the idea would be for later tests to delay execution until the previous ones are done the goal would be to make tests that share state across the tests or in the object under test be deterministic here s a contrived example javascript var i sequential async test function test settimeout function assert equals i i test done math round math random async test sequential async test function test settimeout function assert equals i i test done math round math random async test i guess the alternative is to make separate pages each with a single async test i ll probably do that for the time being
1
80,537
15,586,293,928
IssuesEvent
2021-03-18 01:36:51
saurockSaurav/weather-information-api
https://api.github.com/repos/saurockSaurav/weather-information-api
opened
CVE-2020-14061 (High) detected in jackson-databind-2.8.11.3.jar
security vulnerability
## CVE-2020-14061 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.11.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /weather-information-api/weather-rest-api-service/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.11.3/jackson-databind-2.8.11.3.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.5.20.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.11.3.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to oracle.jms.AQjmsQueueConnectionFactory, oracle.jms.AQjmsXATopicConnectionFactory, oracle.jms.AQjmsTopicConnectionFactory, oracle.jms.AQjmsXAQueueConnectionFactory, and oracle.jms.AQjmsXAConnectionFactory (aka weblogic/oracle-aqjms). <p>Publish Date: 2020-06-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14061>CVE-2020-14061</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061</a></p> <p>Release Date: 2020-06-14</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-14061 (High) detected in jackson-databind-2.8.11.3.jar - ## CVE-2020-14061 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.11.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /weather-information-api/weather-rest-api-service/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.11.3/jackson-databind-2.8.11.3.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.5.20.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.11.3.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to oracle.jms.AQjmsQueueConnectionFactory, oracle.jms.AQjmsXATopicConnectionFactory, oracle.jms.AQjmsTopicConnectionFactory, oracle.jms.AQjmsXAQueueConnectionFactory, and oracle.jms.AQjmsXAConnectionFactory (aka weblogic/oracle-aqjms). <p>Publish Date: 2020-06-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14061>CVE-2020-14061</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061</a></p> <p>Release Date: 2020-06-14</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file weather information api weather rest api service pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to oracle jms aqjmsqueueconnectionfactory oracle jms aqjmsxatopicconnectionfactory oracle jms aqjmstopicconnectionfactory oracle jms aqjmsxaqueueconnectionfactory and oracle jms aqjmsxaconnectionfactory aka weblogic oracle aqjms publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
153,343
12,141,093,418
IssuesEvent
2020-04-23 21:45:07
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
Emit a warning when a codeblock is using "compile-fail" instead of "compile_fail"
A-doctests C-enhancement T-rustdoc
I just found out that a lot of error code explanations were using "compile-fail" instead of "compile_fail". I'm fixing this issue as part of something a bit bigger, however I assume this error is pretty common in rust and that rustdoc should warn people about it. The big issue here is that since it's not a known tag, rustdoc doesn't recognize the codeblock as a rust one and therefore doesn't test it, which is pretty bad.
1.0
Emit a warning when a codeblock is using "compile-fail" instead of "compile_fail" - I just found out that a lot of error code explanations were using "compile-fail" instead of "compile_fail". I'm fixing this issue as part of something a bit bigger, however I assume this error is pretty common in rust and that rustdoc should warn people about it. The big issue here is that since it's not a known tag, rustdoc doesn't recognize the codeblock as a rust one and therefore doesn't test it, which is pretty bad.
test
emit a warning when a codeblock is using compile fail instead of compile fail i just found out that a lot of error code explanations were using compile fail instead of compile fail i m fixing this issue as part of something a bit bigger however i assume this error is pretty common in rust and that rustdoc should warn people about it the big issue here is that since it s not a known tag rustdoc doesn t recognize the codeblock as a rust one and therefore doesn t test it which is pretty bad
1
189,901
14,527,387,969
IssuesEvent
2020-12-14 15:17:49
navarrotheus/caramelo-tec-2-CK0236
https://api.github.com/repos/navarrotheus/caramelo-tec-2-CK0236
opened
Testar SolicitationService
BACK TESTES
## Método create - [ ] Teste de sucesso: Cria a solicitação com sucesso - [ ] Teste de erro: Usuário com tal id não existe - [ ] Teste de erro: Pet com tal id não existe ## Método update - [ ] Teste de sucesso: Atualiza a solicitação com sucesso - [ ] Teste de sucesso: Caso a solicitação seja aceita, cria uma adoção com o solicitante e o pet e seta a disponibilidade do pet como falsa - [ ] Teste de erro: Usuário com tal id não existe ## Método delete - [ ] Teste de sucesso: Deleta a solicitação com sucesso - [ ] Teste de erro: Usuário com tal id não existe ## Método search - [ ] Teste de sucesso: Busca as solicitações do usuário com sucesso - [ ] Teste de erro: Usuário com tal id não existe ## Método searchPetSolicitations - [ ] Teste de sucesso: Busca as solicitações do usuário com sucesso - [ ] Teste de erro: Usuário com tal id não existe - [ ] Teste de erro: Pet com tal id não existe - [ ] Teste de erro: Usuário tenta buscar as solicitações de um Pet de outro usuário
1.0
Testar SolicitationService - ## Método create - [ ] Teste de sucesso: Cria a solicitação com sucesso - [ ] Teste de erro: Usuário com tal id não existe - [ ] Teste de erro: Pet com tal id não existe ## Método update - [ ] Teste de sucesso: Atualiza a solicitação com sucesso - [ ] Teste de sucesso: Caso a solicitação seja aceita, cria uma adoção com o solicitante e o pet e seta a disponibilidade do pet como falsa - [ ] Teste de erro: Usuário com tal id não existe ## Método delete - [ ] Teste de sucesso: Deleta a solicitação com sucesso - [ ] Teste de erro: Usuário com tal id não existe ## Método search - [ ] Teste de sucesso: Busca as solicitações do usuário com sucesso - [ ] Teste de erro: Usuário com tal id não existe ## Método searchPetSolicitations - [ ] Teste de sucesso: Busca as solicitações do usuário com sucesso - [ ] Teste de erro: Usuário com tal id não existe - [ ] Teste de erro: Pet com tal id não existe - [ ] Teste de erro: Usuário tenta buscar as solicitações de um Pet de outro usuário
test
testar solicitationservice método create teste de sucesso cria a solicitação com sucesso teste de erro usuário com tal id não existe teste de erro pet com tal id não existe método update teste de sucesso atualiza a solicitação com sucesso teste de sucesso caso a solicitação seja aceita cria uma adoção com o solicitante e o pet e seta a disponibilidade do pet como falsa teste de erro usuário com tal id não existe método delete teste de sucesso deleta a solicitação com sucesso teste de erro usuário com tal id não existe método search teste de sucesso busca as solicitações do usuário com sucesso teste de erro usuário com tal id não existe método searchpetsolicitations teste de sucesso busca as solicitações do usuário com sucesso teste de erro usuário com tal id não existe teste de erro pet com tal id não existe teste de erro usuário tenta buscar as solicitações de um pet de outro usuário
1
605,103
18,724,969,978
IssuesEvent
2021-11-03 15:26:27
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Details section is empty on Reject/Approve transaction screen
priority/P3 QA/No release-notes/exclude feature/wallet OS/Android
We should fill tx details there if there are any
1.0
Details section is empty on Reject/Approve transaction screen - We should fill tx details there if there are any
non_test
details section is empty on reject approve transaction screen we should fill tx details there if there are any
0
209,653
16,048,047,764
IssuesEvent
2021-04-22 15:41:35
input-output-hk/ouroboros-network
https://api.github.com/repos/input-output-hk/ouroboros-network
closed
Property test which checks that codecs produce a valid CBOR encoding
testing
[validFlatTerm](http://hackage.haskell.org/package/cborg-0.2.5.0/docs/Codec-CBOR-FlatTerm.html#v:validFlatTerm) ``` validFlatTerm . toFlatTerm :: CBOR.Encoding -> Bool ```
1.0
Property test which checks that codecs produce a valid CBOR encoding - [validFlatTerm](http://hackage.haskell.org/package/cborg-0.2.5.0/docs/Codec-CBOR-FlatTerm.html#v:validFlatTerm) ``` validFlatTerm . toFlatTerm :: CBOR.Encoding -> Bool ```
test
property test which checks that codecs produce a valid cbor encoding validflatterm toflatterm cbor encoding bool
1
623,348
19,665,663,607
IssuesEvent
2022-01-10 22:10:23
aesimpson/sama-sanity
https://api.github.com/repos/aesimpson/sama-sanity
closed
CSS Fix for Multiple Modules within a section
Priority:Moderate
At xxxxlg wide screens, the stacked modules don't wrap:
1.0
CSS Fix for Multiple Modules within a section - At xxxxlg wide screens, the stacked modules don't wrap:
non_test
css fix for multiple modules within a section at xxxxlg wide screens the stacked modules don t wrap
0
186,817
14,409,561,604
IssuesEvent
2020-12-04 02:33:39
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/maps/embeddable/tooltip_filter_actions·js - maps app embeddable tooltip filter actions apply filter to current view "before all" hook for "should display create filter button when tooltip is locked"
Team:Geo failed-test test-cloud
**Version: 7.10.0** **Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/maps/embeddable/tooltip_filter_actions·js** **Stack Trace:** ``` Error: retry.try timeout: TypeError: Cannot read property 'clearValue' of undefined at retry.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp1/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/test/functional/services/listing_table.ts:107:28) at process._tickCallback (internal/process/next_tick.js:68:7) at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp1/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:28:9) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp1/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:68:13) ``` **Other test failures:** - maps app embeddable tooltip filter actions panel actions "before all" hook for "should display more actions button when tooltip is locked" _Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/845/testReport/_
2.0
[test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/maps/embeddable/tooltip_filter_actions·js - maps app embeddable tooltip filter actions apply filter to current view "before all" hook for "should display create filter button when tooltip is locked" - **Version: 7.10.0** **Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/maps/embeddable/tooltip_filter_actions·js** **Stack Trace:** ``` Error: retry.try timeout: TypeError: Cannot read property 'clearValue' of undefined at retry.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp1/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/test/functional/services/listing_table.ts:107:28) at process._tickCallback (internal/process/next_tick.js:68:7) at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp1/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:28:9) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp1/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:68:13) ``` **Other test failures:** - maps app embeddable tooltip filter actions panel actions "before all" hook for "should display more actions button when tooltip is locked" _Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/845/testReport/_
test
chrome x pack ui functional x pack test functional apps maps embeddable tooltip filter actions·js maps app embeddable tooltip filter actions apply filter to current view before all hook for should display create filter button when tooltip is locked version class chrome x pack ui functional x pack test functional apps maps embeddable tooltip filter actions·js stack trace error retry try timeout typeerror cannot read property clearvalue of undefined at retry try var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node linux immutable ci cloud common build kibana test functional services listing table ts at process tickcallback internal process next tick js at onfailure var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node linux immutable ci cloud common build kibana test common services retry retry for success ts at retryforsuccess var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node linux immutable ci cloud common build kibana test common services retry retry for success ts other test failures maps app embeddable tooltip filter actions panel actions before all hook for should display more actions button when tooltip is locked test report
1
142,542
11,484,798,098
IssuesEvent
2020-02-11 05:16:42
proarc/proarc
https://api.github.com/repos/proarc/proarc
closed
RDflow - čárový kod se nepropíše
6 k testování RDFlow Release-3.5.16
Čárový kód je v metadatech <mods:identifier type="barcode">26001600877</mods:identifier> Ale nezobrazí se v tabulce Bibliografického záznamu.
1.0
RDflow - čárový kod se nepropíše - Čárový kód je v metadatech <mods:identifier type="barcode">26001600877</mods:identifier> Ale nezobrazí se v tabulce Bibliografického záznamu.
test
rdflow čárový kod se nepropíše čárový kód je v metadatech ale nezobrazí se v tabulce bibliografického záznamu
1
24,697
4,106,194,761
IssuesEvent
2016-06-06 07:38:34
Mr-Kumar-Abhishek/zuzeelik
https://api.github.com/repos/Mr-Kumar-Abhishek/zuzeelik
opened
Test builds from zuzeelik's pre-alpha version v0.0.0-0.5.0 in windows OS
testing
Test builds from zuzeelik's pre-alpha version` v0.0.0-0.5.0` in windows OS. Probably there could be some bugs with built-in functions that handles *quotes*. In *nix systems, all built-in functions are working fine.
1.0
Test builds from zuzeelik's pre-alpha version v0.0.0-0.5.0 in windows OS - Test builds from zuzeelik's pre-alpha version` v0.0.0-0.5.0` in windows OS. Probably there could be some bugs with built-in functions that handles *quotes*. In *nix systems, all built-in functions are working fine.
test
test builds from zuzeelik s pre alpha version in windows os test builds from zuzeelik s pre alpha version in windows os probably there could be some bugs with built in functions that handles quotes in nix systems all built in functions are working fine
1
2,057
2,873,078,819
IssuesEvent
2015-06-08 15:18:24
meumobi/sitebuilder
https://api.github.com/repos/meumobi/sitebuilder
closed
new user can't accept invites if another user is already logged in
bug sitebuilder
The new user can't validade, and the error message is: You need to choose a valid language
1.0
new user can't accept invites if another user is already logged in - The new user can't validade, and the error message is: You need to choose a valid language
non_test
new user can t accept invites if another user is already logged in the new user can t validade and the error message is you need to choose a valid language
0
152,600
12,121,608,652
IssuesEvent
2020-04-22 09:34:43
Students-of-the-city-of-Kostroma/Ray-of-hope
https://api.github.com/repos/Students-of-the-city-of-Kostroma/Ray-of-hope
closed
Протестировать регистрацию и авторизацию организации на новом сервере
AppServer LoginOrg O3 PR5 RegOrg Sprint 14 Testing
Epic #286 Task #287 #288 Протестировать функции, принимающие данные от клиента, обрабатывающие их и возвращающие ответ, на соответствие спецификациям [о регистрации](https://docs.google.com/document/d/1QkQMIYAaNvvknFlBldHHwvcq2iuWHP7E4QYHCFUZdy4) и [авторизации](https://docs.google.com/document/d/1tc8xZATtXaF6GUWpDQYYTfyn8zvWJbQeJYwcmKPkdW4) организации. Поднимать баги по мере нахождения.
1.0
Протестировать регистрацию и авторизацию организации на новом сервере - Epic #286 Task #287 #288 Протестировать функции, принимающие данные от клиента, обрабатывающие их и возвращающие ответ, на соответствие спецификациям [о регистрации](https://docs.google.com/document/d/1QkQMIYAaNvvknFlBldHHwvcq2iuWHP7E4QYHCFUZdy4) и [авторизации](https://docs.google.com/document/d/1tc8xZATtXaF6GUWpDQYYTfyn8zvWJbQeJYwcmKPkdW4) организации. Поднимать баги по мере нахождения.
test
протестировать регистрацию и авторизацию организации на новом сервере epic task протестировать функции принимающие данные от клиента обрабатывающие их и возвращающие ответ на соответствие спецификациям и организации поднимать баги по мере нахождения
1
269,843
23,471,396,035
IssuesEvent
2022-08-16 22:22:31
red/red
https://api.github.com/repos/red/red
closed
Deceptive error message from `set-quiet`
status.built status.tested type.bug
**Describe the bug** ``` >> o: object [] == make object! [] >> set-quiet in o 'x 1 *** Script Error: set-quiet does not allow word! for its word argument ;) what??? *** Where: set-quiet *** Near : 1 *** Stack: ``` **To reproduce** `set-quiet none 1` **Expected behavior** "Doesn't accept none for it's word argument" **Platform version** ``` red-view-14aug22-4eb8ad83f.exe ```
1.0
Deceptive error message from `set-quiet` - **Describe the bug** ``` >> o: object [] == make object! [] >> set-quiet in o 'x 1 *** Script Error: set-quiet does not allow word! for its word argument ;) what??? *** Where: set-quiet *** Near : 1 *** Stack: ``` **To reproduce** `set-quiet none 1` **Expected behavior** "Doesn't accept none for it's word argument" **Platform version** ``` red-view-14aug22-4eb8ad83f.exe ```
test
deceptive error message from set quiet describe the bug o object make object set quiet in o x script error set quiet does not allow word for its word argument what where set quiet near stack to reproduce set quiet none expected behavior doesn t accept none for it s word argument platform version red view exe
1
256,392
22,048,396,783
IssuesEvent
2022-05-30 06:04:16
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
ccl/changefeedccl: TestChangefeedBackfillCheckpoint failed
C-test-failure O-robot branch-release-22.1
ccl/changefeedccl.TestChangefeedBackfillCheckpoint [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=5312529&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=5312529&tab=artifacts#/) on release-22.1 @ [34e6fcfdc8f3155831305ab4a78f960aaad3e7bc](https://github.com/cockroachdb/cockroach/commits/34e6fcfdc8f3155831305ab4a78f960aaad3e7bc): ``` === RUN TestChangefeedBackfillCheckpoint test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/a77002d7c9453d7cd2d382f907780e13/logTestChangefeedBackfillCheckpoint2134013902 test_log_scope.go:80: use -show-logs to present logs inline === CONT TestChangefeedBackfillCheckpoint changefeed_test.go:5440: -- test log scope end -- --- FAIL: TestChangefeedBackfillCheckpoint (541.52s) === RUN TestChangefeedBackfillCheckpoint/enterprise-limit=100_B changefeed_test.go:5420: Error Trace: changefeed_test.go:5420 helpers_test.go:554 Error: Received unexpected error: retrying txn failed after 50 attempts. original error: pq: restart transaction: TransactionRetryWithProtoRefreshError: TransactionRetryError: retry txn (RETRY_SERIALIZABLE - failed preemptive refresh due to a conflict: committed value on key /Table/109/1/"foo"/0/766181192725659648/0): "sql txn" meta={id=1ef45b68 key=/Table/109/1/"foo"/0/766179641595625472/0 pri=0.00000000 epo=50 ts=1653890353.444952019,2 min=1653890008.216986178,0 seq=8087} lock=true stat=PENDING rts=1653890344.066420091,0 wto=false gul=1653890008.716986178,0. Test: TestChangefeedBackfillCheckpoint/enterprise-limit=100_B --- FAIL: TestChangefeedBackfillCheckpoint/enterprise-limit=100_B (486.62s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) Parameters in this failure: - TAGS=bazel,gss,deadlock </p> </details> /cc @cockroachdb/cdc <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestChangefeedBackfillCheckpoint.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
ccl/changefeedccl: TestChangefeedBackfillCheckpoint failed - ccl/changefeedccl.TestChangefeedBackfillCheckpoint [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=5312529&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=5312529&tab=artifacts#/) on release-22.1 @ [34e6fcfdc8f3155831305ab4a78f960aaad3e7bc](https://github.com/cockroachdb/cockroach/commits/34e6fcfdc8f3155831305ab4a78f960aaad3e7bc): ``` === RUN TestChangefeedBackfillCheckpoint test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/a77002d7c9453d7cd2d382f907780e13/logTestChangefeedBackfillCheckpoint2134013902 test_log_scope.go:80: use -show-logs to present logs inline === CONT TestChangefeedBackfillCheckpoint changefeed_test.go:5440: -- test log scope end -- --- FAIL: TestChangefeedBackfillCheckpoint (541.52s) === RUN TestChangefeedBackfillCheckpoint/enterprise-limit=100_B changefeed_test.go:5420: Error Trace: changefeed_test.go:5420 helpers_test.go:554 Error: Received unexpected error: retrying txn failed after 50 attempts. original error: pq: restart transaction: TransactionRetryWithProtoRefreshError: TransactionRetryError: retry txn (RETRY_SERIALIZABLE - failed preemptive refresh due to a conflict: committed value on key /Table/109/1/"foo"/0/766181192725659648/0): "sql txn" meta={id=1ef45b68 key=/Table/109/1/"foo"/0/766179641595625472/0 pri=0.00000000 epo=50 ts=1653890353.444952019,2 min=1653890008.216986178,0 seq=8087} lock=true stat=PENDING rts=1653890344.066420091,0 wto=false gul=1653890008.716986178,0. Test: TestChangefeedBackfillCheckpoint/enterprise-limit=100_B --- FAIL: TestChangefeedBackfillCheckpoint/enterprise-limit=100_B (486.62s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) Parameters in this failure: - TAGS=bazel,gss,deadlock </p> </details> /cc @cockroachdb/cdc <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestChangefeedBackfillCheckpoint.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
ccl changefeedccl testchangefeedbackfillcheckpoint failed ccl changefeedccl testchangefeedbackfillcheckpoint with on release run testchangefeedbackfillcheckpoint test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline cont testchangefeedbackfillcheckpoint changefeed test go test log scope end fail testchangefeedbackfillcheckpoint run testchangefeedbackfillcheckpoint enterprise limit b changefeed test go error trace changefeed test go helpers test go error received unexpected error retrying txn failed after attempts original error pq restart transaction transactionretrywithprotorefresherror transactionretryerror retry txn retry serializable failed preemptive refresh due to a conflict committed value on key table foo sql txn meta id key table foo pri epo ts min seq lock true stat pending rts wto false gul test testchangefeedbackfillcheckpoint enterprise limit b fail testchangefeedbackfillcheckpoint enterprise limit b help see also parameters in this failure tags bazel gss deadlock cc cockroachdb cdc
1
65,588
14,740,878,706
IssuesEvent
2021-01-07 09:45:58
hiptest/ember-easy-datatable
https://api.github.com/repos/hiptest/ember-easy-datatable
opened
CVE-2018-16487 (Medium) detected in lodash-3.10.1.tgz
security vulnerability
## CVE-2018-16487 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: ember-easy-datatable/package.json</p> <p>Path to vulnerable library: ember-easy-datatable/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - ember-cli-2.12.2.tgz (Root Library) - broccoli-babel-transpiler-5.7.4.tgz - babel-core-5.8.38.tgz - :x: **lodash-3.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/hiptest/ember-easy-datatable/commit/174fc2ea19b9aaaa080440d9cb938d1e9a2d6120">174fc2ea19b9aaaa080440d9cb938d1e9a2d6120</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.10.1","isTransitiveDependency":true,"dependencyTree":"ember-cli:2.12.2;broccoli-babel-transpiler:5.7.4;babel-core:5.8.38;lodash:3.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"}],"vulnerabilityIdentifier":"CVE-2018-16487","vulnerabilityDetails":"A prototype pollution vulnerability was found in lodash \u003c4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487","cvss3Severity":"medium","cvss3Score":"5.6","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-16487 (Medium) detected in lodash-3.10.1.tgz - ## CVE-2018-16487 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: ember-easy-datatable/package.json</p> <p>Path to vulnerable library: ember-easy-datatable/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - ember-cli-2.12.2.tgz (Root Library) - broccoli-babel-transpiler-5.7.4.tgz - babel-core-5.8.38.tgz - :x: **lodash-3.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/hiptest/ember-easy-datatable/commit/174fc2ea19b9aaaa080440d9cb938d1e9a2d6120">174fc2ea19b9aaaa080440d9cb938d1e9a2d6120</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.10.1","isTransitiveDependency":true,"dependencyTree":"ember-cli:2.12.2;broccoli-babel-transpiler:5.7.4;babel-core:5.8.38;lodash:3.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"}],"vulnerabilityIdentifier":"CVE-2018-16487","vulnerabilityDetails":"A prototype pollution vulnerability was found in lodash \u003c4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487","cvss3Severity":"medium","cvss3Score":"5.6","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file ember easy datatable package json path to vulnerable library ember easy datatable node modules lodash package json dependency hierarchy ember cli tgz root library broccoli babel transpiler tgz babel core tgz x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype vulnerabilityurl
0
249,856
21,195,374,409
IssuesEvent
2022-04-08 23:31:16
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
ci-cadvisor-e2e failed
priority/important-soon area/cadvisor sig/node kind/failing-test triage/accepted
### Which jobs are failing? ci-cadvisor-e2e ### Which tests are failing? ``` W0322 22:57:59.666] 2022/03/22 22:57:59 main.go:331: Something went wrong: encountered 1 errors: [error during go run /go/src/k8s.io/kubernetes/test/e2e_node/runner/remote/run_remote.go --cleanup --logtostderr --vmodule=*=4 --ssh-env=gce --results-dir=/workspace/_artifacts --project=ci-cadvisor-e2e --zone=us-central1-f --ssh-user=prow --ssh-key=/workspace/.ssh/google_compute_engine --ginkgo-flags=--nodes=1 --test_args= --test-timeout=10m0s --image-config-file=/workspace/test-infra/jobs/e2e_node/containerd/image-config.yaml --test-suite=cadvisor: exit status 1] W0322 22:57:59.673] Traceback (most recent call last): W0322 22:57:59.674] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 723, in <module> W0322 22:57:59.674] main(parse_args()) W0322 22:57:59.674] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 569, in main W0322 22:57:59.674] mode.start(runner_args) W0322 22:57:59.674] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 228, in start W0322 22:57:59.675] check_env(env, self.command, *args) W0322 22:57:59.675] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 111, in check_env W0322 22:57:59.675] subprocess.check_call(cmd, env=env) W0322 22:57:59.675] File "/usr/lib/python2.7/subprocess.py", line 190, in check_call W0322 22:57:59.675] raise CalledProcessError(retcode, cmd) W0322 22:57:59.676] subprocess.CalledProcessError: Command '('kubetest', '--dump=/workspace/_artifacts', '--gcp-service-account=/etc/service-account/service-account.json', '--up', '--down', '--test', '--deployment=node', '--provider=gce', '--cluster=bootstrap-e2e', '--gcp-network=bootstrap-e2e', '--gcp-project=ci-cadvisor-e2e', '--gcp-zone=us-central1-f', '--node-args=--image-config-file=/workspace/test-infra/jobs/e2e_node/containerd/image-config.yaml --test-suite=cadvisor', '--node-tests=true', '--test_args=--nodes=1', '--timeout=10m')' returned non-zero exit status 1 E0322 22:57:59.677] Command failed I0322 22:57:59.678] process 339 exited with code 1 after 2.2m E0322 22:57:59.678] FAIL: ci-cadvisor-e2e ``` ### Since when has it been failing? Since 03-22. Changes list that may - https://github.com/kubernetes/kubernetes/compare/dd604a0f9a2e015a869d01957297e33f2c0c7025...95e30f66c? - - Candidate1: https://github.com/kubernetes/kubernetes/pull/108704 (but for windows) - https://github.com/kubernetes/test-infra/compare/ce1169a66...a6fe00c22 - - Candidate1: https://github.com/kubernetes/test-infra/pull/25727 - - Candidate2: https://github.com/kubernetes/test-infra/pull/25405 ### Testgrid link https://testgrid.k8s.io/sig-node-cadvisor#cadvisor-e2e ### Reason for failure (if possible) _No response_ ### Anything else we need to know? _No response_ ### Relevant SIG(s) /sig node /area cadvisor
1.0
ci-cadvisor-e2e failed - ### Which jobs are failing? ci-cadvisor-e2e ### Which tests are failing? ``` W0322 22:57:59.666] 2022/03/22 22:57:59 main.go:331: Something went wrong: encountered 1 errors: [error during go run /go/src/k8s.io/kubernetes/test/e2e_node/runner/remote/run_remote.go --cleanup --logtostderr --vmodule=*=4 --ssh-env=gce --results-dir=/workspace/_artifacts --project=ci-cadvisor-e2e --zone=us-central1-f --ssh-user=prow --ssh-key=/workspace/.ssh/google_compute_engine --ginkgo-flags=--nodes=1 --test_args= --test-timeout=10m0s --image-config-file=/workspace/test-infra/jobs/e2e_node/containerd/image-config.yaml --test-suite=cadvisor: exit status 1] W0322 22:57:59.673] Traceback (most recent call last): W0322 22:57:59.674] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 723, in <module> W0322 22:57:59.674] main(parse_args()) W0322 22:57:59.674] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 569, in main W0322 22:57:59.674] mode.start(runner_args) W0322 22:57:59.674] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 228, in start W0322 22:57:59.675] check_env(env, self.command, *args) W0322 22:57:59.675] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 111, in check_env W0322 22:57:59.675] subprocess.check_call(cmd, env=env) W0322 22:57:59.675] File "/usr/lib/python2.7/subprocess.py", line 190, in check_call W0322 22:57:59.675] raise CalledProcessError(retcode, cmd) W0322 22:57:59.676] subprocess.CalledProcessError: Command '('kubetest', '--dump=/workspace/_artifacts', '--gcp-service-account=/etc/service-account/service-account.json', '--up', '--down', '--test', '--deployment=node', '--provider=gce', '--cluster=bootstrap-e2e', '--gcp-network=bootstrap-e2e', '--gcp-project=ci-cadvisor-e2e', '--gcp-zone=us-central1-f', '--node-args=--image-config-file=/workspace/test-infra/jobs/e2e_node/containerd/image-config.yaml --test-suite=cadvisor', '--node-tests=true', '--test_args=--nodes=1', '--timeout=10m')' returned non-zero exit status 1 E0322 22:57:59.677] Command failed I0322 22:57:59.678] process 339 exited with code 1 after 2.2m E0322 22:57:59.678] FAIL: ci-cadvisor-e2e ``` ### Since when has it been failing? Since 03-22. Changes list that may - https://github.com/kubernetes/kubernetes/compare/dd604a0f9a2e015a869d01957297e33f2c0c7025...95e30f66c? - - Candidate1: https://github.com/kubernetes/kubernetes/pull/108704 (but for windows) - https://github.com/kubernetes/test-infra/compare/ce1169a66...a6fe00c22 - - Candidate1: https://github.com/kubernetes/test-infra/pull/25727 - - Candidate2: https://github.com/kubernetes/test-infra/pull/25405 ### Testgrid link https://testgrid.k8s.io/sig-node-cadvisor#cadvisor-e2e ### Reason for failure (if possible) _No response_ ### Anything else we need to know? _No response_ ### Relevant SIG(s) /sig node /area cadvisor
test
ci cadvisor failed which jobs are failing ci cadvisor which tests are failing main go something went wrong encountered errors traceback most recent call last file workspace test infra jenkins scenarios kubernetes py line in main parse args file workspace test infra jenkins scenarios kubernetes py line in main mode start runner args file workspace test infra jenkins scenarios kubernetes py line in start check env env self command args file workspace test infra jenkins scenarios kubernetes py line in check env subprocess check call cmd env env file usr lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command kubetest dump workspace artifacts gcp service account etc service account service account json up down test deployment node provider gce cluster bootstrap gcp network bootstrap gcp project ci cadvisor gcp zone us f node args image config file workspace test infra jobs node containerd image config yaml test suite cadvisor node tests true test args nodes timeout returned non zero exit status command failed process exited with code after fail ci cadvisor since when has it been failing since changes list that may (but for windows) testgrid link reason for failure if possible no response anything else we need to know no response relevant sig s sig node area cadvisor
1
51,145
13,190,289,230
IssuesEvent
2020-08-13 09:55:47
ESA-VirES/WebClient-Framework
https://api.github.com/repos/ESA-VirES/WebClient-Framework
opened
Broken server-side interpolation of the EEF data.
defect
When selecting MAG and EEF data the server responds with following error: ``` Error: Problem retrieving data: 'Interp1D' object has no attribute 'indices_nearest' ``` ![image](https://user-images.githubusercontent.com/4050019/90120566-2b832980-dd5b-11ea-939d-700b3761cef6.png) This is a regression introduces in v3.3.0. Observed on the production instance. Already fixed on staging. FAO @lmar76
1.0
Broken server-side interpolation of the EEF data. - When selecting MAG and EEF data the server responds with following error: ``` Error: Problem retrieving data: 'Interp1D' object has no attribute 'indices_nearest' ``` ![image](https://user-images.githubusercontent.com/4050019/90120566-2b832980-dd5b-11ea-939d-700b3761cef6.png) This is a regression introduces in v3.3.0. Observed on the production instance. Already fixed on staging. FAO @lmar76
non_test
broken server side interpolation of the eef data when selecting mag and eef data the server responds with following error error problem retrieving data object has no attribute indices nearest this is a regression introduces in observed on the production instance already fixed on staging fao
0
21,665
3,911,689,155
IssuesEvent
2016-04-20 07:21:57
Legion-Expansion/Legion-Expansion
https://api.github.com/repos/Legion-Expansion/Legion-Expansion
reopened
Remove Icon Extensions and Icon Reloader as dependencies
needs testing pte
This will be needed for PTE release, but it will break icons on 89755.
1.0
Remove Icon Extensions and Icon Reloader as dependencies - This will be needed for PTE release, but it will break icons on 89755.
test
remove icon extensions and icon reloader as dependencies this will be needed for pte release but it will break icons on
1
275,534
23,921,258,929
IssuesEvent
2022-09-09 17:07:25
ECP-WarpX/WarpX
https://api.github.com/repos/ECP-WarpX/WarpX
opened
Invalid memory access when moving window and timers-based load-balancing is used
bug bug: affects latest release component: load balancing
I am opening this issue because I have observed an invalid memory access when moving window and load-balancing based on timers are used in combination. Here I provide a small reproducer: ``` ################################# ####### GENERAL PARAMETERS ###### ################################# max_step = 10 amr.n_cell = 64 64 64 amr.max_grid_size = 32 amr.blocking_factor = 32 amr.max_level = 0 geometry.dims = 3 geometry.prob_lo = -10.e-6 -10.e-6 -10.e-6 # physical domain geometry.prob_hi = 10.e-6 10.e-6 10.e-6 algo.load_balance_intervals = 3::100 algo.load_balance_with_sfc = 0 algo.load_balance_costs_update = timers warpx.do_moving_window = 1 warpx.moving_window_dir = z warpx.moving_window_v = 1.0 warpx.start_moving_window_step = 2 ################################# ####### Boundary condition ###### ################################# boundary.field_lo = pml pml pml boundary.field_hi = pml pml pml ################################# ############ NUMERICS ########### ################################# warpx.verbose = 1 warpx.cfl = 0.99 # Order of particle shape factors algo.particle_shape = 3 ################################# ############ PLASMA ############# ################################# particles.species_names = electrons electrons.species_type = electron electrons.injection_style = "NUniformPerCell" electrons.num_particles_per_cell_each_dim = 1 1 2 electrons.profile = constant electrons.density = 1.e25 # number of electrons per m^3 electrons.momentum_distribution_type = "gaussian" electrons.ux_th = 0.01 # uth the std of the (unitless) momentum electrons.uy_th = 0.01 # uth the std of the (unitless) momentum electrons.uz_th = 0.01 # uth the std of the (unitless) momentum ``` When WarpX runs this inputfile (even without GPUs or OMP support), `valgrind` detects the following issue: ``` STEP 3 starts ... ==41155== Invalid read of size 4 ==41155== at 0x55CFBD: Add<float> (AMReX_GpuAtomic.H:584) ==41155== by 0x55CFBD: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435) ==41155== by 0x55F8EF: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:192) ==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269) ==41155== by 0x1BB863: main (main.cpp:67) ==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd ==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422) ==41155== by 0x1F088A: allocate (new_allocator.h:127) ==41155== by 0x1F088A: allocate (alloc_traits.h:464) ==41155== by 0x1F088A: _M_allocate (stl_vector.h:346) ==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635) ==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31) ==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22) ==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962) ==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170) ==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680) ==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548) ==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779) ==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472) ==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378) ==41155== by 0x1BB856: main (main.cpp:65) ==41155== ==41155== Invalid write of size 4 ==41155== at 0x55CFC1: Add<float> (AMReX_GpuAtomic.H:584) ==41155== by 0x55CFC1: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435) ==41155== by 0x55F8EF: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:192) ==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269) ==41155== by 0x1BB863: main (main.cpp:67) ==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd ==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422) ==41155== by 0x1F088A: allocate (new_allocator.h:127) ==41155== by 0x1F088A: allocate (alloc_traits.h:464) ==41155== by 0x1F088A: _M_allocate (stl_vector.h:346) ==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635) ==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31) ==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22) ==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962) ==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170) ==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680) ==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548) ==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779) ==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472) ==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378) ==41155== by 0x1BB856: main (main.cpp:65) ==41155== ==41155== Invalid read of size 4 ==41155== at 0x55CFBD: Add<float> (AMReX_GpuAtomic.H:584) ==41155== by 0x55CFBD: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435) ==41155== by 0x55F92F: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:193) ==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269) ==41155== by 0x1BB863: main (main.cpp:67) ==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd ==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422) ==41155== by 0x1F088A: allocate (new_allocator.h:127) ==41155== by 0x1F088A: allocate (alloc_traits.h:464) ==41155== by 0x1F088A: _M_allocate (stl_vector.h:346) ==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635) ==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31) ==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22) ==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962) ==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170) ==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680) ==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548) ==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779) ==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472) ==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378) ==41155== by 0x1BB856: main (main.cpp:65) ==41155== ==41155== Invalid write of size 4 ==41155== at 0x55CFC1: Add<float> (AMReX_GpuAtomic.H:584) ==41155== by 0x55CFC1: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435) ==41155== by 0x55F92F: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:193) ==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269) ==41155== by 0x1BB863: main (main.cpp:67) ==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd ==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422) ==41155== by 0x1F088A: allocate (new_allocator.h:127) ==41155== by 0x1F088A: allocate (alloc_traits.h:464) ==41155== by 0x1F088A: _M_allocate (stl_vector.h:346) ==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635) ==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31) ==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22) ==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962) ==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170) ==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680) ==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548) ==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779) ==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472) ==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378) ==41155== by 0x1BB856: main (main.cpp:65) ==41155== STEP 3 ends. TIME = 1.787413796e-15 DT = 5.958046162e-16 Evolve time = 44.5962677 s; This step = 13.71976852 s; Avg. per step = 14.86542225 s ```
1.0
Invalid memory access when moving window and timers-based load-balancing is used - I am opening this issue because I have observed an invalid memory access when moving window and load-balancing based on timers are used in combination. Here I provide a small reproducer: ``` ################################# ####### GENERAL PARAMETERS ###### ################################# max_step = 10 amr.n_cell = 64 64 64 amr.max_grid_size = 32 amr.blocking_factor = 32 amr.max_level = 0 geometry.dims = 3 geometry.prob_lo = -10.e-6 -10.e-6 -10.e-6 # physical domain geometry.prob_hi = 10.e-6 10.e-6 10.e-6 algo.load_balance_intervals = 3::100 algo.load_balance_with_sfc = 0 algo.load_balance_costs_update = timers warpx.do_moving_window = 1 warpx.moving_window_dir = z warpx.moving_window_v = 1.0 warpx.start_moving_window_step = 2 ################################# ####### Boundary condition ###### ################################# boundary.field_lo = pml pml pml boundary.field_hi = pml pml pml ################################# ############ NUMERICS ########### ################################# warpx.verbose = 1 warpx.cfl = 0.99 # Order of particle shape factors algo.particle_shape = 3 ################################# ############ PLASMA ############# ################################# particles.species_names = electrons electrons.species_type = electron electrons.injection_style = "NUniformPerCell" electrons.num_particles_per_cell_each_dim = 1 1 2 electrons.profile = constant electrons.density = 1.e25 # number of electrons per m^3 electrons.momentum_distribution_type = "gaussian" electrons.ux_th = 0.01 # uth the std of the (unitless) momentum electrons.uy_th = 0.01 # uth the std of the (unitless) momentum electrons.uz_th = 0.01 # uth the std of the (unitless) momentum ``` When WarpX runs this inputfile (even without GPUs or OMP support), `valgrind` detects the following issue: ``` STEP 3 starts ... ==41155== Invalid read of size 4 ==41155== at 0x55CFBD: Add<float> (AMReX_GpuAtomic.H:584) ==41155== by 0x55CFBD: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435) ==41155== by 0x55F8EF: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:192) ==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269) ==41155== by 0x1BB863: main (main.cpp:67) ==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd ==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422) ==41155== by 0x1F088A: allocate (new_allocator.h:127) ==41155== by 0x1F088A: allocate (alloc_traits.h:464) ==41155== by 0x1F088A: _M_allocate (stl_vector.h:346) ==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635) ==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31) ==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22) ==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962) ==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170) ==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680) ==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548) ==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779) ==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472) ==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378) ==41155== by 0x1BB856: main (main.cpp:65) ==41155== ==41155== Invalid write of size 4 ==41155== at 0x55CFC1: Add<float> (AMReX_GpuAtomic.H:584) ==41155== by 0x55CFC1: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435) ==41155== by 0x55F8EF: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:192) ==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269) ==41155== by 0x1BB863: main (main.cpp:67) ==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd ==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422) ==41155== by 0x1F088A: allocate (new_allocator.h:127) ==41155== by 0x1F088A: allocate (alloc_traits.h:464) ==41155== by 0x1F088A: _M_allocate (stl_vector.h:346) ==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635) ==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31) ==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22) ==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962) ==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170) ==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680) ==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548) ==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779) ==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472) ==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378) ==41155== by 0x1BB856: main (main.cpp:65) ==41155== ==41155== Invalid read of size 4 ==41155== at 0x55CFBD: Add<float> (AMReX_GpuAtomic.H:584) ==41155== by 0x55CFBD: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435) ==41155== by 0x55F92F: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:193) ==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269) ==41155== by 0x1BB863: main (main.cpp:67) ==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd ==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422) ==41155== by 0x1F088A: allocate (new_allocator.h:127) ==41155== by 0x1F088A: allocate (alloc_traits.h:464) ==41155== by 0x1F088A: _M_allocate (stl_vector.h:346) ==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635) ==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31) ==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22) ==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962) ==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170) ==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680) ==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548) ==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779) ==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472) ==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378) ==41155== by 0x1BB856: main (main.cpp:65) ==41155== ==41155== Invalid write of size 4 ==41155== at 0x55CFC1: Add<float> (AMReX_GpuAtomic.H:584) ==41155== by 0x55CFC1: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435) ==41155== by 0x55F92F: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:193) ==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269) ==41155== by 0x1BB863: main (main.cpp:67) ==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd ==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422) ==41155== by 0x1F088A: allocate (new_allocator.h:127) ==41155== by 0x1F088A: allocate (alloc_traits.h:464) ==41155== by 0x1F088A: _M_allocate (stl_vector.h:346) ==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635) ==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31) ==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22) ==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962) ==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170) ==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680) ==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548) ==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779) ==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472) ==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378) ==41155== by 0x1BB856: main (main.cpp:65) ==41155== STEP 3 ends. TIME = 1.787413796e-15 DT = 5.958046162e-16 Evolve time = 44.5962677 s; This step = 13.71976852 s; Avg. per step = 14.86542225 s ```
test
invalid memory access when moving window and timers based load balancing is used i am opening this issue because i have observed an invalid memory access when moving window and load balancing based on timers are used in combination here i provide a small reproducer general parameters max step amr n cell amr max grid size amr blocking factor amr max level geometry dims geometry prob lo e e e physical domain geometry prob hi e e e algo load balance intervals algo load balance with sfc algo load balance costs update timers warpx do moving window warpx moving window dir z warpx moving window v warpx start moving window step boundary condition boundary field lo pml pml pml boundary field hi pml pml pml numerics warpx verbose warpx cfl order of particle shape factors algo particle shape plasma particles species names electrons electrons species type electron electrons injection style nuniformpercell electrons num particles per cell each dim electrons profile constant electrons density number of electrons per m electrons momentum distribution type gaussian electrons ux th uth the std of the unitless momentum electrons uy th uth the std of the unitless momentum electrons uz th uth the std of the unitless momentum when warpx runs this inputfile even without gpus or omp support valgrind detects the following issue step starts invalid read of size at add amrex gpuatomic h by warpx shiftmf amrex multifab amrex geometry const int int int float bool amrex parserexecutor const warpxmovingwindow cpp by warpx movewindow int bool warpxmovingwindow cpp by warpx evolve int warpxevolve cpp by main main cpp address is bytes before a block of size alloc d at operator new unsigned long vg replace malloc c by allocate new allocator h by allocate alloc traits h by m allocate stl vector h by std vector m default append unsigned long vector tcc by define amrex layoutdata h by layoutdata amrex layoutdata h by make unique const amrex boxarray const amrex distributionmapping unique ptr h by warpx alloclevelmfs int amrex boxarray const amrex distributionmapping const amrex intvect const amrex intvect const amrex intvect const amrex intvect const amrex intvect const bool warpx cpp by warpx allocleveldata int amrex boxarray const amrex distributionmapping const warpx cpp by warpx makenewlevelfromscratch int float amrex boxarray const amrex distributionmapping const warpx cpp by amrex amrmesh makenewgrids float amrex amrmesh cpp by initfromscratch warpxinitdata cpp by warpx initdata warpxinitdata cpp by main main cpp invalid write of size at add amrex gpuatomic h by warpx shiftmf amrex multifab amrex geometry const int int int float bool amrex parserexecutor const warpxmovingwindow cpp by warpx movewindow int bool warpxmovingwindow cpp by warpx evolve int warpxevolve cpp by main main cpp address is bytes before a block of size alloc d at operator new unsigned long vg replace malloc c by allocate new allocator h by allocate alloc traits h by m allocate stl vector h by std vector m default append unsigned long vector tcc by define amrex layoutdata h by layoutdata amrex layoutdata h by make unique const amrex boxarray const amrex distributionmapping unique ptr h by warpx alloclevelmfs int amrex boxarray const amrex distributionmapping const amrex intvect const amrex intvect const amrex intvect const amrex intvect const amrex intvect const bool warpx cpp by warpx allocleveldata int amrex boxarray const amrex distributionmapping const warpx cpp by warpx makenewlevelfromscratch int float amrex boxarray const amrex distributionmapping const warpx cpp by amrex amrmesh makenewgrids float amrex amrmesh cpp by initfromscratch warpxinitdata cpp by warpx initdata warpxinitdata cpp by main main cpp invalid read of size at add amrex gpuatomic h by warpx shiftmf amrex multifab amrex geometry const int int int float bool amrex parserexecutor const warpxmovingwindow cpp by warpx movewindow int bool warpxmovingwindow cpp by warpx evolve int warpxevolve cpp by main main cpp address is bytes before a block of size alloc d at operator new unsigned long vg replace malloc c by allocate new allocator h by allocate alloc traits h by m allocate stl vector h by std vector m default append unsigned long vector tcc by define amrex layoutdata h by layoutdata amrex layoutdata h by make unique const amrex boxarray const amrex distributionmapping unique ptr h by warpx alloclevelmfs int amrex boxarray const amrex distributionmapping const amrex intvect const amrex intvect const amrex intvect const amrex intvect const amrex intvect const bool warpx cpp by warpx allocleveldata int amrex boxarray const amrex distributionmapping const warpx cpp by warpx makenewlevelfromscratch int float amrex boxarray const amrex distributionmapping const warpx cpp by amrex amrmesh makenewgrids float amrex amrmesh cpp by initfromscratch warpxinitdata cpp by warpx initdata warpxinitdata cpp by main main cpp invalid write of size at add amrex gpuatomic h by warpx shiftmf amrex multifab amrex geometry const int int int float bool amrex parserexecutor const warpxmovingwindow cpp by warpx movewindow int bool warpxmovingwindow cpp by warpx evolve int warpxevolve cpp by main main cpp address is bytes before a block of size alloc d at operator new unsigned long vg replace malloc c by allocate new allocator h by allocate alloc traits h by m allocate stl vector h by std vector m default append unsigned long vector tcc by define amrex layoutdata h by layoutdata amrex layoutdata h by make unique const amrex boxarray const amrex distributionmapping unique ptr h by warpx alloclevelmfs int amrex boxarray const amrex distributionmapping const amrex intvect const amrex intvect const amrex intvect const amrex intvect const amrex intvect const bool warpx cpp by warpx allocleveldata int amrex boxarray const amrex distributionmapping const warpx cpp by warpx makenewlevelfromscratch int float amrex boxarray const amrex distributionmapping const warpx cpp by amrex amrmesh makenewgrids float amrex amrmesh cpp by initfromscratch warpxinitdata cpp by warpx initdata warpxinitdata cpp by main main cpp step ends time dt evolve time s this step s avg per step s
1
111,332
9,526,672,346
IssuesEvent
2019-04-28 21:56:13
quinoacomputing/quinoa
https://api.github.com/repos/quinoacomputing/quinoa
opened
Port to C++17
Charm++ fileconv inciter meshconv parser rngtest unittest walker
- [x] use `this` in lambdas - [x] use `[[maybe_unused]]` instead of `IGNORE` - [x] replace `std::enable_if` with `if constexpr` - [x] replace brigand `for_each` with a fold calling lambdas - [x] simplify `Scheme` with variadic generic lambdas passed to `std::visit` - [ ] use `std::void_t` to simplify `Has_*` types - [ ] Replace the overloads in Control.h with variadic ones. Besides C++17 features, the implementation of the Field template may give a clue on how to do the variadic deep-access for types in tagged tuple, see Alexandrescu, Modern C++ Design: Generic Programming and Design Patterns Applied, p69.
2.0
Port to C++17 - - [x] use `this` in lambdas - [x] use `[[maybe_unused]]` instead of `IGNORE` - [x] replace `std::enable_if` with `if constexpr` - [x] replace brigand `for_each` with a fold calling lambdas - [x] simplify `Scheme` with variadic generic lambdas passed to `std::visit` - [ ] use `std::void_t` to simplify `Has_*` types - [ ] Replace the overloads in Control.h with variadic ones. Besides C++17 features, the implementation of the Field template may give a clue on how to do the variadic deep-access for types in tagged tuple, see Alexandrescu, Modern C++ Design: Generic Programming and Design Patterns Applied, p69.
test
port to c use this in lambdas use instead of ignore replace std enable if with if constexpr replace brigand for each with a fold calling lambdas simplify scheme with variadic generic lambdas passed to std visit use std void t to simplify has types replace the overloads in control h with variadic ones besides c features the implementation of the field template may give a clue on how to do the variadic deep access for types in tagged tuple see alexandrescu modern c design generic programming and design patterns applied
1
204,373
7,087,353,417
IssuesEvent
2018-01-11 17:31:06
salesagility/SuiteCRM
https://api.github.com/repos/salesagility/SuiteCRM
closed
In an existing meeting, changing the Accounts to another Company then saving causes SuiteCRM run out of memory.
Fix Proposed High Priority Resolved: Next Release bug
#### Issue If you have a meeting and change the Account it is associated with and save. SuiteCRM stops for a while and eventually comes back with: Fatal error: Allowed memory size of 134217728 bytes exhausted (tried to allocate 8208 bytes) in /suitecrm/code/website/include/HTMLPurifier/HTMLPurifier.standalone.php on line 15136 going back to the meeting, the Account is changed. However, there are several Reminder boxes now filling up the meeting. #### Expected Behavior Saves successfully and returns to the DetailView #### Actual Behavior SuiteCRM stops for a while before eventually showing: Fatal error: Allowed memory size of 134217728 bytes exhausted (tried to allocate 8208 bytes) in /suitecrm/code/website/include/HTMLPurifier/HTMLPurifier.standalone.php on line 15136 nothing appears in SuiteCRM log. #### Possible Fix #### Steps to Reproduce 1. Edit a pre-existing meeting which has an associated Account 2. Change the meeting's associated Account to another Account 3. Save the Meeting 4. SuiteCRM stops responding for a period of time until memory limit is hit. 5. Go back to SuiteCRM and view the meeting, several Reminders have appeared. #### Context Customer raised issue. Some Accounts have been changing their name/certain sections of the business splitting into own entity. #### Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * SuiteCRM Version used: Version 7.9.7 This issue appears on https://demo.suiteondemand.com I've attached a screenshot which shows multiple reminders appearing. ![reminders](https://user-images.githubusercontent.com/236256/34068968-4dbcb37e-e23e-11e7-8837-917d60280dab.png)
1.0
In an existing meeting, changing the Accounts to another Company then saving causes SuiteCRM run out of memory. - #### Issue If you have a meeting and change the Account it is associated with and save. SuiteCRM stops for a while and eventually comes back with: Fatal error: Allowed memory size of 134217728 bytes exhausted (tried to allocate 8208 bytes) in /suitecrm/code/website/include/HTMLPurifier/HTMLPurifier.standalone.php on line 15136 going back to the meeting, the Account is changed. However, there are several Reminder boxes now filling up the meeting. #### Expected Behavior Saves successfully and returns to the DetailView #### Actual Behavior SuiteCRM stops for a while before eventually showing: Fatal error: Allowed memory size of 134217728 bytes exhausted (tried to allocate 8208 bytes) in /suitecrm/code/website/include/HTMLPurifier/HTMLPurifier.standalone.php on line 15136 nothing appears in SuiteCRM log. #### Possible Fix #### Steps to Reproduce 1. Edit a pre-existing meeting which has an associated Account 2. Change the meeting's associated Account to another Account 3. Save the Meeting 4. SuiteCRM stops responding for a period of time until memory limit is hit. 5. Go back to SuiteCRM and view the meeting, several Reminders have appeared. #### Context Customer raised issue. Some Accounts have been changing their name/certain sections of the business splitting into own entity. #### Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * SuiteCRM Version used: Version 7.9.7 This issue appears on https://demo.suiteondemand.com I've attached a screenshot which shows multiple reminders appearing. ![reminders](https://user-images.githubusercontent.com/236256/34068968-4dbcb37e-e23e-11e7-8837-917d60280dab.png)
non_test
in an existing meeting changing the accounts to another company then saving causes suitecrm run out of memory issue if you have a meeting and change the account it is associated with and save suitecrm stops for a while and eventually comes back with fatal error allowed memory size of bytes exhausted tried to allocate bytes in suitecrm code website include htmlpurifier htmlpurifier standalone php on line going back to the meeting the account is changed however there are several reminder boxes now filling up the meeting expected behavior saves successfully and returns to the detailview actual behavior suitecrm stops for a while before eventually showing fatal error allowed memory size of bytes exhausted tried to allocate bytes in suitecrm code website include htmlpurifier htmlpurifier standalone php on line nothing appears in suitecrm log possible fix steps to reproduce edit a pre existing meeting which has an associated account change the meeting s associated account to another account save the meeting suitecrm stops responding for a period of time until memory limit is hit go back to suitecrm and view the meeting several reminders have appeared context customer raised issue some accounts have been changing their name certain sections of the business splitting into own entity your environment suitecrm version used version this issue appears on i ve attached a screenshot which shows multiple reminders appearing
0
264,883
23,145,072,200
IssuesEvent
2022-07-28 23:13:22
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Seridores - Registro por lotação - Coração de Jesus
generalization test development template-Síntese tecnologia informatica tag-Servidores subtag-Registro por lotação
DoD: Realizar o teste de Generalização do validador da tag Seridores - Registro por lotação para o Município de Coração de Jesus.
1.0
Teste de generalizacao para a tag Seridores - Registro por lotação - Coração de Jesus - DoD: Realizar o teste de Generalização do validador da tag Seridores - Registro por lotação para o Município de Coração de Jesus.
test
teste de generalizacao para a tag seridores registro por lotação coração de jesus dod realizar o teste de generalização do validador da tag seridores registro por lotação para o município de coração de jesus
1
349,897
24,960,894,335
IssuesEvent
2022-11-01 15:26:34
AY2223S1-CS2103T-T09-3/tp
https://api.github.com/repos/AY2223S1-CS2103T-T09-3/tp
closed
[PE-D][Tester D] Misleading lines
documentation
![Screenshot 2022-10-28 at 16.25.30.png](https://raw.githubusercontent.com/optionalemon/ped/main/files/547182f8-b314-46b0-b1fe-4d5ce15986ad.png) addproj command requires more than one necessary field to be added, and this line is misleading as it may make people think that the only necessary field is the project name (Small visual issue that addproj and addProj does not match in this line, even though I understand that both works, maybe pulling the line I have screenshot below to the top would be clearer to the user in the UG) ![Screenshot 2022-10-28 at 16.28.34.png](https://raw.githubusercontent.com/optionalemon/ped/main/files/1ffeb3f5-009a-4189-be6e-69e450cd0d09.png) <!--session: 1666944041950-12b678a2-9d92-444c-b20f-e586ee74af4e--> <!--Version: Web v3.4.4--> ------------- Labels: `severity.VeryLow` `type.DocumentationBug` original: optionalemon/ped#3
1.0
[PE-D][Tester D] Misleading lines - ![Screenshot 2022-10-28 at 16.25.30.png](https://raw.githubusercontent.com/optionalemon/ped/main/files/547182f8-b314-46b0-b1fe-4d5ce15986ad.png) addproj command requires more than one necessary field to be added, and this line is misleading as it may make people think that the only necessary field is the project name (Small visual issue that addproj and addProj does not match in this line, even though I understand that both works, maybe pulling the line I have screenshot below to the top would be clearer to the user in the UG) ![Screenshot 2022-10-28 at 16.28.34.png](https://raw.githubusercontent.com/optionalemon/ped/main/files/1ffeb3f5-009a-4189-be6e-69e450cd0d09.png) <!--session: 1666944041950-12b678a2-9d92-444c-b20f-e586ee74af4e--> <!--Version: Web v3.4.4--> ------------- Labels: `severity.VeryLow` `type.DocumentationBug` original: optionalemon/ped#3
non_test
misleading lines addproj command requires more than one necessary field to be added and this line is misleading as it may make people think that the only necessary field is the project name small visual issue that addproj and addproj does not match in this line even though i understand that both works maybe pulling the line i have screenshot below to the top would be clearer to the user in the ug labels severity verylow type documentationbug original optionalemon ped
0
2,497
2,736,487,420
IssuesEvent
2015-04-19 13:53:08
tgstation/-tg-station
https://api.github.com/repos/tgstation/-tg-station
opened
newscaster blares, "attach_spans(input, spans)"
Bug say() code
self-explanatory. (This)[https://github.com/tgstation/-tg-station/blob/master/code/game/say.dm#L71] seems to be at fault, although I have no idea how.
1.0
newscaster blares, "attach_spans(input, spans)" - self-explanatory. (This)[https://github.com/tgstation/-tg-station/blob/master/code/game/say.dm#L71] seems to be at fault, although I have no idea how.
non_test
newscaster blares attach spans input spans self explanatory this seems to be at fault although i have no idea how
0
109,100
9,368,333,041
IssuesEvent
2019-04-03 08:27:46
Microsoft/AzureStorageExplorer
https://api.github.com/repos/Microsoft/AzureStorageExplorer
opened
Support to upload multiple folders at a time using 'Upload Folder' dialog
🧪 testing
**Storage Explorer Version:** 1.7.0_20190401.1 **Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10 **Architecture:** ia32/x64 **Regression From:** Not a regression **Actually:** We can drag multiple folders at the same time to one blob container/file share. And the selected folders can be uploaded successfully. **Suggestion:** Can we support to upload multiple folders using 'Upload Folder' dialog?
1.0
Support to upload multiple folders at a time using 'Upload Folder' dialog - **Storage Explorer Version:** 1.7.0_20190401.1 **Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10 **Architecture:** ia32/x64 **Regression From:** Not a regression **Actually:** We can drag multiple folders at the same time to one blob container/file share. And the selected folders can be uploaded successfully. **Suggestion:** Can we support to upload multiple folders using 'Upload Folder' dialog?
test
support to upload multiple folders at a time using upload folder dialog storage explorer version platform os linux ubuntu macos high sierra windows architecture regression from not a regression actually we can drag multiple folders at the same time to one blob container file share and the selected folders can be uploaded successfully suggestion can we support to upload multiple folders using upload folder dialog
1
172,123
6,499,309,802
IssuesEvent
2017-08-22 20:57:49
DCLP/dclpxsltbox
https://api.github.com/repos/DCLP/dclpxsltbox
closed
Displaying the tag “inverse”
bug component: XSLT priority: medium review
It is accepted by the system, but not displayed: in P.Oxy. 75.5023 (TM 128952) part. C, 1 (and ff.), I would expect something like “(turned 180°)” https://github.com/DCLP/idp.data/blob/master/DCLP/129/128952.xml#L164 Leiden+: ```(1, inverse) (ἦ\χ/(ος)) <#β=2#> ``` XML: ```xml <lb n="1" rend="inverse"/> <expan>ἦ<add place="above">χ</add><ex>ος</ex></expan> <num value="2">β</num> ``` [Displayed](http://dclp.github.io/dclpxsltbox/output/dclp/129/128952.html): ```ἦ\χ/(ος) β``` (see now #251 for incorrect placement of hyphen; comments previously here have been moved there).
1.0
Displaying the tag “inverse” - It is accepted by the system, but not displayed: in P.Oxy. 75.5023 (TM 128952) part. C, 1 (and ff.), I would expect something like “(turned 180°)” https://github.com/DCLP/idp.data/blob/master/DCLP/129/128952.xml#L164 Leiden+: ```(1, inverse) (ἦ\χ/(ος)) <#β=2#> ``` XML: ```xml <lb n="1" rend="inverse"/> <expan>ἦ<add place="above">χ</add><ex>ος</ex></expan> <num value="2">β</num> ``` [Displayed](http://dclp.github.io/dclpxsltbox/output/dclp/129/128952.html): ```ἦ\χ/(ος) β``` (see now #251 for incorrect placement of hyphen; comments previously here have been moved there).
non_test
displaying the tag “inverse” it is accepted by the system but not displayed in p oxy tm part c and ff i would expect something like “ turned ° ” leiden inverse ἦ χ ος xml xml ἦ χ ος β ἦ χ ος β see now for incorrect placement of hyphen comments previously here have been moved there
0
329,350
28,236,897,592
IssuesEvent
2023-04-06 01:54:46
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
The error message doesn't disappear after clearing invalid strings in text editor
:heavy_check_mark: merged 🧪 testing :gear: tables :beetle: regression
**Storage Explorer Version**: 1.28.0-dev **Build Number**: 20230330.2 **Branch**: feature/table-explorer-react **Platform/OS**: Windows 10/Linux Ubuntu 22.04/MacOS Ventura 13.0.1 (Apple M1 Pro) **Architecture**: ia32/x64 **How Found**: From running test cases **Regression From**: Previous release (1.28.1) ## Steps to Reproduce ## 1. Expand one storage account -> Tables. 2. Right click one table -> Click '[2] Open'. 3. Click 'Query' -> Open the text editor. 4. Type '1' into the text editor -> Execute query. 5. There is an error message -> Click 'Clear query'. 6. Check whether the error message disappears. ## Expected Experience ## The error message disappears. ![image](https://user-images.githubusercontent.com/41351993/228744281-9ac08446-f704-4f4e-8397-010212d7b732.png) ## Actual Experience ## The error message doesn't disappear. ![image](https://user-images.githubusercontent.com/41351993/228744266-9e0336bf-2304-45f0-8fbd-7f770bf852c9.png)
1.0
The error message doesn't disappear after clearing invalid strings in text editor - **Storage Explorer Version**: 1.28.0-dev **Build Number**: 20230330.2 **Branch**: feature/table-explorer-react **Platform/OS**: Windows 10/Linux Ubuntu 22.04/MacOS Ventura 13.0.1 (Apple M1 Pro) **Architecture**: ia32/x64 **How Found**: From running test cases **Regression From**: Previous release (1.28.1) ## Steps to Reproduce ## 1. Expand one storage account -> Tables. 2. Right click one table -> Click '[2] Open'. 3. Click 'Query' -> Open the text editor. 4. Type '1' into the text editor -> Execute query. 5. There is an error message -> Click 'Clear query'. 6. Check whether the error message disappears. ## Expected Experience ## The error message disappears. ![image](https://user-images.githubusercontent.com/41351993/228744281-9ac08446-f704-4f4e-8397-010212d7b732.png) ## Actual Experience ## The error message doesn't disappear. ![image](https://user-images.githubusercontent.com/41351993/228744266-9e0336bf-2304-45f0-8fbd-7f770bf852c9.png)
test
the error message doesn t disappear after clearing invalid strings in text editor storage explorer version dev build number branch feature table explorer react platform os windows linux ubuntu macos ventura apple pro architecture how found from running test cases regression from previous release steps to reproduce expand one storage account tables right click one table click open click query open the text editor type into the text editor execute query there is an error message click clear query check whether the error message disappears expected experience the error message disappears actual experience the error message doesn t disappear
1
227,416
18,062,324,757
IssuesEvent
2021-09-20 15:09:40
trinodb/trino
https://api.github.com/repos/trinodb/trino
closed
randomTableSuffix is not random enough leading to test flakiness
bug test
example ``` Error: io.trino.plugin.oracle.TestOracleTypeMapping.testDate Time elapsed: 0.284 s <<< FAILURE! java.lang.RuntimeException: Error executing sql: CREATE TABLE test_date_4nznv AS SELECT CAST(DATE '1952-04-03' AS DATE) col_0, CAST(DATE '1970-01-01' AS DATE) col_1, CAST(DATE '1970-02-03' AS DATE) col_2, CAST(DATE '2017-07-01' AS DATE) col_3, CAST(DATE '2017-01-01' AS DATE) col_4, CAST(DATE '1983-04-01' AS DATE) col_5, CAST(DATE '1983-10-01' AS DATE) col_6 at io.trino.testing.sql.TrinoSqlExecutor.execute(TrinoSqlExecutor.java:45) at io.trino.testing.sql.TestTable.<init>(TestTable.java:48) at io.trino.testing.sql.TestTable.<init>(TestTable.java:41) at io.trino.testing.datatype.CreateAsSelectDataSetup.setupTestTable(CreateAsSelectDataSetup.java:46) at io.trino.testing.datatype.SqlDataTypeTest.execute(SqlDataTypeTest.java:78) at io.trino.plugin.oracle.AbstractTestOracleTypeMapping.testDate(AbstractTestOracleTypeMapping.java:696) ```
1.0
randomTableSuffix is not random enough leading to test flakiness - example ``` Error: io.trino.plugin.oracle.TestOracleTypeMapping.testDate Time elapsed: 0.284 s <<< FAILURE! java.lang.RuntimeException: Error executing sql: CREATE TABLE test_date_4nznv AS SELECT CAST(DATE '1952-04-03' AS DATE) col_0, CAST(DATE '1970-01-01' AS DATE) col_1, CAST(DATE '1970-02-03' AS DATE) col_2, CAST(DATE '2017-07-01' AS DATE) col_3, CAST(DATE '2017-01-01' AS DATE) col_4, CAST(DATE '1983-04-01' AS DATE) col_5, CAST(DATE '1983-10-01' AS DATE) col_6 at io.trino.testing.sql.TrinoSqlExecutor.execute(TrinoSqlExecutor.java:45) at io.trino.testing.sql.TestTable.<init>(TestTable.java:48) at io.trino.testing.sql.TestTable.<init>(TestTable.java:41) at io.trino.testing.datatype.CreateAsSelectDataSetup.setupTestTable(CreateAsSelectDataSetup.java:46) at io.trino.testing.datatype.SqlDataTypeTest.execute(SqlDataTypeTest.java:78) at io.trino.plugin.oracle.AbstractTestOracleTypeMapping.testDate(AbstractTestOracleTypeMapping.java:696) ```
test
randomtablesuffix is not random enough leading to test flakiness example error io trino plugin oracle testoracletypemapping testdate time elapsed s failure java lang runtimeexception error executing sql create table test date as select cast date as date col cast date as date col cast date as date col cast date as date col cast date as date col cast date as date col cast date as date col at io trino testing sql trinosqlexecutor execute trinosqlexecutor java at io trino testing sql testtable testtable java at io trino testing sql testtable testtable java at io trino testing datatype createasselectdatasetup setuptesttable createasselectdatasetup java at io trino testing datatype sqldatatypetest execute sqldatatypetest java at io trino plugin oracle abstracttestoracletypemapping testdate abstracttestoracletypemapping java
1
174,597
6,541,452,911
IssuesEvent
2017-09-01 20:02:10
wlandau-lilly/drake
https://api.github.com/repos/wlandau-lilly/drake
closed
Crop hover text if too long in `dataframes_graph()`
TOP PRIORITY wlandau-lilly has a patch and is waiting for permission to release it
Long hover text could slow down performance.
1.0
Crop hover text if too long in `dataframes_graph()` - Long hover text could slow down performance.
non_test
crop hover text if too long in dataframes graph long hover text could slow down performance
0
556,223
16,478,335,277
IssuesEvent
2021-05-24 08:33:23
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
mobile.twitter.com - see bug description
browser-firefox-tablet engine-gecko ml-needsdiagnosis-false priority-critical
<!-- @browser: Firefox Mobile (Tablet) 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/74030 --> **URL**: https://mobile.twitter.com/home **Browser / Version**: Firefox Mobile (Tablet) 68.0 **Operating System**: Android 8.0.0 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: I'm requesting desktop but it keeps giving me mobile version **Steps to Reproduce**: I requested desktop but am redirected to mobile no matter what <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/5/13f8751e-5466-4d50-9285-b35eff0b9a41.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827194101</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/5/f66f53ac-20d0-41bd-b173-e5f6b5495642) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
mobile.twitter.com - see bug description - <!-- @browser: Firefox Mobile (Tablet) 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/74030 --> **URL**: https://mobile.twitter.com/home **Browser / Version**: Firefox Mobile (Tablet) 68.0 **Operating System**: Android 8.0.0 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: I'm requesting desktop but it keeps giving me mobile version **Steps to Reproduce**: I requested desktop but am redirected to mobile no matter what <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/5/13f8751e-5466-4d50-9285-b35eff0b9a41.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827194101</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/5/f66f53ac-20d0-41bd-b173-e5f6b5495642) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
mobile twitter com see bug description url browser version firefox mobile tablet operating system android tested another browser yes chrome problem type something else description i m requesting desktop but it keeps giving me mobile version steps to reproduce i requested desktop but am redirected to mobile no matter what view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
172,921
13,357,637,088
IssuesEvent
2020-08-31 10:10:20
apache/shardingsphere
https://api.github.com/repos/apache/shardingsphere
closed
PostgreSQL dedicated column missed in t_order cause GeneralDQLIT.assertExecuteQuery failed
test
## Bug Report ### Which version of ShardingSphere did you use? master branch, 5.0.0-RC1-SNAPSHOT ### Which project did you use? ShardingSphere-JDBC or ShardingSphere-Proxy? ShardingSphere-JDBC ### Expected behavior Related dql-test-case pass. ### Actual behavior Test failed, exception thrown: ``` org.postgresql.util.PSQLException: ERROR: column "rule" does not exist at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2440) at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2183) at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:308) at org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:441) at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:365) at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:307) at org.postgresql.jdbc.PgStatement.executeCachedSql(PgStatement.java:293) at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:270) at org.postgresql.jdbc.PgStatement.executeQuery(PgStatement.java:224) at com.zaxxer.hikari.pool.ProxyStatement.executeQuery(ProxyStatement.java:111) at com.zaxxer.hikari.pool.HikariProxyStatement.executeQuery(HikariProxyStatement.java) at org.apache.shardingsphere.driver.executor.StatementExecutor$1.createQueryResult(StatementExecutor.java:81) at org.apache.shardingsphere.driver.executor.StatementExecutor$1.executeSQL(StatementExecutor.java:77) at org.apache.shardingsphere.driver.executor.StatementExecutor$1.executeSQL(StatementExecutor.java:73) at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.impl.DefaultSQLExecutorCallback.execute0(DefaultSQLExecutorCallback.java:75) at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.impl.DefaultSQLExecutorCallback.execute(DefaultSQLExecutorCallback.java:57) at org.apache.shardingsphere.infra.executor.kernel.ExecutorKernel.syncExecute(ExecutorKernel.java:99) at org.apache.shardingsphere.infra.executor.kernel.ExecutorKernel.parallelExecute(ExecutorKernel.java:95) at org.apache.shardingsphere.infra.executor.kernel.ExecutorKernel.execute(ExecutorKernel.java:78) at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.SQLExecutor.execute(SQLExecutor.java:66) at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.SQLExecutor.execute(SQLExecutor.java:50) at org.apache.shardingsphere.driver.executor.StatementExecutor.executeQuery(StatementExecutor.java:85) at org.apache.shardingsphere.driver.jdbc.core.statement.ShardingSphereStatement.executeQuery(ShardingSphereStatement.java:127) at org.apache.shardingsphere.dbtest.engine.dql.GeneralDQLIT.assertExecuteQueryForStatement(GeneralDQLIT.java:70) at org.apache.shardingsphere.dbtest.engine.dql.GeneralDQLIT.assertExecuteQuery(GeneralDQLIT.java:60) ``` ### Reason analyze (If you can) `rule` and `start_point` columns doesn't exist in `t_order` table. Since they're special column type, we can't add these columns for all types of database. Possible solution: - Make `schema.xml > table-create > sql` support `db-types` definition, e.g. `db-types=PostgreSQL`, then the same name tables that have different columns will be created for different database. ### Steps to reproduce the behavior, such as: SQL to execute, sharding rule configuration, when exception occur etc. ### Example codes for reproduce this issue (such as a github link).
1.0
PostgreSQL dedicated column missed in t_order cause GeneralDQLIT.assertExecuteQuery failed - ## Bug Report ### Which version of ShardingSphere did you use? master branch, 5.0.0-RC1-SNAPSHOT ### Which project did you use? ShardingSphere-JDBC or ShardingSphere-Proxy? ShardingSphere-JDBC ### Expected behavior Related dql-test-case pass. ### Actual behavior Test failed, exception thrown: ``` org.postgresql.util.PSQLException: ERROR: column "rule" does not exist at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2440) at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2183) at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:308) at org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:441) at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:365) at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:307) at org.postgresql.jdbc.PgStatement.executeCachedSql(PgStatement.java:293) at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:270) at org.postgresql.jdbc.PgStatement.executeQuery(PgStatement.java:224) at com.zaxxer.hikari.pool.ProxyStatement.executeQuery(ProxyStatement.java:111) at com.zaxxer.hikari.pool.HikariProxyStatement.executeQuery(HikariProxyStatement.java) at org.apache.shardingsphere.driver.executor.StatementExecutor$1.createQueryResult(StatementExecutor.java:81) at org.apache.shardingsphere.driver.executor.StatementExecutor$1.executeSQL(StatementExecutor.java:77) at org.apache.shardingsphere.driver.executor.StatementExecutor$1.executeSQL(StatementExecutor.java:73) at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.impl.DefaultSQLExecutorCallback.execute0(DefaultSQLExecutorCallback.java:75) at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.impl.DefaultSQLExecutorCallback.execute(DefaultSQLExecutorCallback.java:57) at org.apache.shardingsphere.infra.executor.kernel.ExecutorKernel.syncExecute(ExecutorKernel.java:99) at org.apache.shardingsphere.infra.executor.kernel.ExecutorKernel.parallelExecute(ExecutorKernel.java:95) at org.apache.shardingsphere.infra.executor.kernel.ExecutorKernel.execute(ExecutorKernel.java:78) at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.SQLExecutor.execute(SQLExecutor.java:66) at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.SQLExecutor.execute(SQLExecutor.java:50) at org.apache.shardingsphere.driver.executor.StatementExecutor.executeQuery(StatementExecutor.java:85) at org.apache.shardingsphere.driver.jdbc.core.statement.ShardingSphereStatement.executeQuery(ShardingSphereStatement.java:127) at org.apache.shardingsphere.dbtest.engine.dql.GeneralDQLIT.assertExecuteQueryForStatement(GeneralDQLIT.java:70) at org.apache.shardingsphere.dbtest.engine.dql.GeneralDQLIT.assertExecuteQuery(GeneralDQLIT.java:60) ``` ### Reason analyze (If you can) `rule` and `start_point` columns doesn't exist in `t_order` table. Since they're special column type, we can't add these columns for all types of database. Possible solution: - Make `schema.xml > table-create > sql` support `db-types` definition, e.g. `db-types=PostgreSQL`, then the same name tables that have different columns will be created for different database. ### Steps to reproduce the behavior, such as: SQL to execute, sharding rule configuration, when exception occur etc. ### Example codes for reproduce this issue (such as a github link).
test
postgresql dedicated column missed in t order cause generaldqlit assertexecutequery failed bug report which version of shardingsphere did you use master branch snapshot which project did you use shardingsphere jdbc or shardingsphere proxy shardingsphere jdbc expected behavior related dql test case pass actual behavior test failed exception thrown org postgresql util psqlexception error column rule does not exist at org postgresql core queryexecutorimpl receiveerrorresponse queryexecutorimpl java at org postgresql core queryexecutorimpl processresults queryexecutorimpl java at org postgresql core queryexecutorimpl execute queryexecutorimpl java at org postgresql jdbc pgstatement executeinternal pgstatement java at org postgresql jdbc pgstatement execute pgstatement java at org postgresql jdbc pgstatement executewithflags pgstatement java at org postgresql jdbc pgstatement executecachedsql pgstatement java at org postgresql jdbc pgstatement executewithflags pgstatement java at org postgresql jdbc pgstatement executequery pgstatement java at com zaxxer hikari pool proxystatement executequery proxystatement java at com zaxxer hikari pool hikariproxystatement executequery hikariproxystatement java at org apache shardingsphere driver executor statementexecutor createqueryresult statementexecutor java at org apache shardingsphere driver executor statementexecutor executesql statementexecutor java at org apache shardingsphere driver executor statementexecutor executesql statementexecutor java at org apache shardingsphere infra executor sql resourced jdbc executor impl defaultsqlexecutorcallback defaultsqlexecutorcallback java at org apache shardingsphere infra executor sql resourced jdbc executor impl defaultsqlexecutorcallback execute defaultsqlexecutorcallback java at org apache shardingsphere infra executor kernel executorkernel syncexecute executorkernel java at org apache shardingsphere infra executor kernel executorkernel parallelexecute executorkernel java at org apache shardingsphere infra executor kernel executorkernel execute executorkernel java at org apache shardingsphere infra executor sql resourced jdbc executor sqlexecutor execute sqlexecutor java at org apache shardingsphere infra executor sql resourced jdbc executor sqlexecutor execute sqlexecutor java at org apache shardingsphere driver executor statementexecutor executequery statementexecutor java at org apache shardingsphere driver jdbc core statement shardingspherestatement executequery shardingspherestatement java at org apache shardingsphere dbtest engine dql generaldqlit assertexecutequeryforstatement generaldqlit java at org apache shardingsphere dbtest engine dql generaldqlit assertexecutequery generaldqlit java reason analyze if you can rule and start point columns doesn t exist in t order table since they re special column type we can t add these columns for all types of database possible solution make schema xml table create sql support db types definition e g db types postgresql then the same name tables that have different columns will be created for different database steps to reproduce the behavior such as sql to execute sharding rule configuration when exception occur etc example codes for reproduce this issue such as a github link
1
87,519
15,779,925,860
IssuesEvent
2021-04-01 09:18:35
AlexRogalskiy/gradle-java-sample
https://api.github.com/repos/AlexRogalskiy/gradle-java-sample
opened
CVE-2021-21346 (High) detected in xstream-1.4.10.jar
security vulnerability
## CVE-2021-21346 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.10.jar</b></p></summary> <p>XStream is a serialization library from Java objects to XML and back.</p> <p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p> <p>Path to dependency file: gradle-java-sample/buildSrc/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.thoughtworks.xstream/xstream/1.4.10/dfecae23647abc9d9fd0416629a4213a3882b101/xstream-1.4.10.jar</p> <p> Dependency Hierarchy: - gradle-versions-plugin-0.28.0.jar (Root Library) - :x: **xstream-1.4.10.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/gradle-java-sample/commit/e537d2e240e1b0b48107d38039b89c5b5d6fd977">e537d2e240e1b0b48107d38039b89c5b5d6fd977</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to load and execute arbitrary code from a remote host only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21346>CVE-2021-21346</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-4hrm-m67v-5cxr">https://github.com/x-stream/xstream/security/advisories/GHSA-4hrm-m67v-5cxr</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-21346 (High) detected in xstream-1.4.10.jar - ## CVE-2021-21346 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.10.jar</b></p></summary> <p>XStream is a serialization library from Java objects to XML and back.</p> <p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p> <p>Path to dependency file: gradle-java-sample/buildSrc/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.thoughtworks.xstream/xstream/1.4.10/dfecae23647abc9d9fd0416629a4213a3882b101/xstream-1.4.10.jar</p> <p> Dependency Hierarchy: - gradle-versions-plugin-0.28.0.jar (Root Library) - :x: **xstream-1.4.10.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/gradle-java-sample/commit/e537d2e240e1b0b48107d38039b89c5b5d6fd977">e537d2e240e1b0b48107d38039b89c5b5d6fd977</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to load and execute arbitrary code from a remote host only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21346>CVE-2021-21346</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-4hrm-m67v-5cxr">https://github.com/x-stream/xstream/security/advisories/GHSA-4hrm-m67v-5cxr</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in xstream jar cve high severity vulnerability vulnerable library xstream jar xstream is a serialization library from java objects to xml and back library home page a href path to dependency file gradle java sample buildsrc build gradle path to vulnerable library home wss scanner gradle caches modules files com thoughtworks xstream xstream xstream jar dependency hierarchy gradle versions plugin jar root library x xstream jar vulnerable library found in head commit a href vulnerability details xstream is a java library to serialize objects to xml and back again in xstream before version there is a vulnerability which may allow a remote attacker to load and execute arbitrary code from a remote host only by manipulating the processed input stream no user is affected who followed the recommendation to setup xstream s security framework with a whitelist limited to the minimal required types if you rely on xstream s default blacklist of the security framework you will have to use at least version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream step up your open source security game with whitesource
0
432,446
30,284,203,165
IssuesEvent
2023-07-08 13:07:19
Danfall369/Catalog-of-my-things
https://api.github.com/repos/Danfall369/Catalog-of-my-things
closed
[2.5pt] (Group Task) General Structure
documentation
# Group Task ## Create `Item` class: Following the [UML Diagram](https://github.com/microverseinc/curriculum-ruby/blob/main/group-capstone/images/catalog_of_my_things.png) - [x] Create Item class in a separate .rb file. - [x] All Item class properties visible in the diagram should be defined and set up in the constructor method. Exception: properties for the 1-to-many relationships should NOT be set in the constructor method. Instead, they should have a custom setter method created. - [x] Add all methods visible in the diagram. ## Implement methods #### _can_be_archived?() in the Item class_ - - [x] Should return true if published_date is older than 10 years. - - [x] Otherwise, it should return false. #### _Move_to_archive() in the Item class_ - - [x] Should reuse can_be_archived?() method. - - [x] Should change the archived property to true if the result of the can_be_archived?() method is true. - - [x] Should do nothing if the result of the can_be_archived?() method is false. ## Create Basic UI - [x] Create a main.rb file that will serve as your console app entry-point. - [x] Present the user with a list of options to perform. - [x] Let users choose an option. - [x] If needed, ask for parameters for the option. - [x] Have a way to quit the app.
1.0
[2.5pt] (Group Task) General Structure - # Group Task ## Create `Item` class: Following the [UML Diagram](https://github.com/microverseinc/curriculum-ruby/blob/main/group-capstone/images/catalog_of_my_things.png) - [x] Create Item class in a separate .rb file. - [x] All Item class properties visible in the diagram should be defined and set up in the constructor method. Exception: properties for the 1-to-many relationships should NOT be set in the constructor method. Instead, they should have a custom setter method created. - [x] Add all methods visible in the diagram. ## Implement methods #### _can_be_archived?() in the Item class_ - - [x] Should return true if published_date is older than 10 years. - - [x] Otherwise, it should return false. #### _Move_to_archive() in the Item class_ - - [x] Should reuse can_be_archived?() method. - - [x] Should change the archived property to true if the result of the can_be_archived?() method is true. - - [x] Should do nothing if the result of the can_be_archived?() method is false. ## Create Basic UI - [x] Create a main.rb file that will serve as your console app entry-point. - [x] Present the user with a list of options to perform. - [x] Let users choose an option. - [x] If needed, ask for parameters for the option. - [x] Have a way to quit the app.
non_test
group task general structure group task create item class following the create item class in a separate rb file all item class properties visible in the diagram should be defined and set up in the constructor method exception properties for the to many relationships should not be set in the constructor method instead they should have a custom setter method created add all methods visible in the diagram implement methods can be archived in the item class should return true if published date is older than years otherwise it should return false move to archive in the item class should reuse can be archived method should change the archived property to true if the result of the can be archived method is true should do nothing if the result of the can be archived method is false create basic ui create a main rb file that will serve as your console app entry point present the user with a list of options to perform let users choose an option if needed ask for parameters for the option have a way to quit the app
0
321,243
27,517,311,502
IssuesEvent
2023-03-06 12:53:10
IntellectualSites/FastAsyncWorldEdit
https://api.github.com/repos/IntellectualSites/FastAsyncWorldEdit
closed
Off by one error for negative coordinates when using -r with //deform
Requires Testing
### Server Implementation Paper ### Server Version 1.19.2 ### Describe the bug When using `//deform -r` negative coordinates are off by one. ### To Reproduce Select a region around 0,0 Do `//deform -r 0` Observe how parts of the region that are in the negatives within the game's coordinate system move around. ### Expected behaviour Nothing should change as we do not modify the x,y,z variables. ### Screenshots / Videos Input and also expected output ![2023-02-07_15 11 14](https://user-images.githubusercontent.com/16972002/217268335-4f39dba8-d689-4127-b7fc-21cb6a595fb4.png) Current output ![2023-02-07_15 11 20](https://user-images.githubusercontent.com/16972002/217268391-5fc3b1a0-62fa-4b0e-86ec-99740c75b3b7.png) ### Error log (if applicable) _No response_ ### Fawe Debugpaste https://athion.net/ISPaster/paste/view/4e4bf7430f1b4ac6847ad782e6904782 ### Fawe Version FastAsyncWorldEdit version 2.5.2-SNAPSHOT-349 ### Checklist - [X] I have included a Fawe debugpaste. - [X] I am using the newest build from https://ci.athion.net/job/FastAsyncWorldEdit/ and the issue still persists. ### Anything else? When I do `//deform -r x=0` it fetches block at x=0. When I do `//deform -r x=-1` it fetches block at x=0. When I do `//deform -r x=-2` it fetches block at x=-1. When I do `//deform -r x=-3` it fetches block at x=-2. The same thing happens an all three axes.
1.0
Off by one error for negative coordinates when using -r with //deform - ### Server Implementation Paper ### Server Version 1.19.2 ### Describe the bug When using `//deform -r` negative coordinates are off by one. ### To Reproduce Select a region around 0,0 Do `//deform -r 0` Observe how parts of the region that are in the negatives within the game's coordinate system move around. ### Expected behaviour Nothing should change as we do not modify the x,y,z variables. ### Screenshots / Videos Input and also expected output ![2023-02-07_15 11 14](https://user-images.githubusercontent.com/16972002/217268335-4f39dba8-d689-4127-b7fc-21cb6a595fb4.png) Current output ![2023-02-07_15 11 20](https://user-images.githubusercontent.com/16972002/217268391-5fc3b1a0-62fa-4b0e-86ec-99740c75b3b7.png) ### Error log (if applicable) _No response_ ### Fawe Debugpaste https://athion.net/ISPaster/paste/view/4e4bf7430f1b4ac6847ad782e6904782 ### Fawe Version FastAsyncWorldEdit version 2.5.2-SNAPSHOT-349 ### Checklist - [X] I have included a Fawe debugpaste. - [X] I am using the newest build from https://ci.athion.net/job/FastAsyncWorldEdit/ and the issue still persists. ### Anything else? When I do `//deform -r x=0` it fetches block at x=0. When I do `//deform -r x=-1` it fetches block at x=0. When I do `//deform -r x=-2` it fetches block at x=-1. When I do `//deform -r x=-3` it fetches block at x=-2. The same thing happens an all three axes.
test
off by one error for negative coordinates when using r with deform server implementation paper server version describe the bug when using deform r negative coordinates are off by one to reproduce select a region around do deform r observe how parts of the region that are in the negatives within the game s coordinate system move around expected behaviour nothing should change as we do not modify the x y z variables screenshots videos input and also expected output current output error log if applicable no response fawe debugpaste fawe version fastasyncworldedit version snapshot checklist i have included a fawe debugpaste i am using the newest build from and the issue still persists anything else when i do deform r x it fetches block at x when i do deform r x it fetches block at x when i do deform r x it fetches block at x when i do deform r x it fetches block at x the same thing happens an all three axes
1
63,646
6,877,413,596
IssuesEvent
2017-11-20 07:51:34
alibaba/pouch
https://api.github.com/repos/alibaba/pouch
closed
[bug]The swagger spec at "swagger.yml" is invalid against swagger specification 2.0
areas/test kind/bug
**Issue Description** ``` The swagger spec at "swagger.yml" is invalid against swagger specification 2.0 - definitions.ContainerConfig.properties.Cmd in body must be of type array - definitions.ContainerConfig.properties.Entrypoint in body must be of type array ``` **How to reproduce it (as minimally and precisely as possible)**: ``` cd apis swagger validate swagger.yml ```
1.0
[bug]The swagger spec at "swagger.yml" is invalid against swagger specification 2.0 - **Issue Description** ``` The swagger spec at "swagger.yml" is invalid against swagger specification 2.0 - definitions.ContainerConfig.properties.Cmd in body must be of type array - definitions.ContainerConfig.properties.Entrypoint in body must be of type array ``` **How to reproduce it (as minimally and precisely as possible)**: ``` cd apis swagger validate swagger.yml ```
test
the swagger spec at swagger yml is invalid against swagger specification issue description the swagger spec at swagger yml is invalid against swagger specification definitions containerconfig properties cmd in body must be of type array definitions containerconfig properties entrypoint in body must be of type array how to reproduce it as minimally and precisely as possible cd apis swagger validate swagger yml
1
701,998
24,118,661,778
IssuesEvent
2022-09-20 16:40:04
IslasGECI/dimorfismo
https://api.github.com/repos/IslasGECI/dimorfismo
reopened
Los nombres de las propiedades en `logistic_model_parameters.json` están en _spanglish_
Status: Available Priority: Low Type: Maintenance wontfix
Abajo se puede ver que nombres de las propiedades están en _spanglish_: ``` { "parametrosNormalizacion": { "valorMinimo": { "longitudCraneo": [3.7037], "altoPico": [165.58], "longitudPico": [29.67], "tarso": [101.78] }, "valorMaximo": { "longitudCraneo": [83.18], "altoPico": [46.5], "longitudPico": [193.22], "tarso": [35.44] } }, "parametrosModelo": [ { "Variables": "(Intercept)", "Estimate": -18.948, "_row": "(Intercept)" }, { "Variables": "longitudCraneo", "Estimate": 6.576, "_row": "longitudCraneo" }, { "Variables": "altoPico", "Estimate": 8.816, "_row": "altoPico" }, { "Variables": "longitudPico", "Estimate": 7.172, "_row": "longitudPico" }, { "Variables": "tarso", "Estimate": 5.726, "_row": "tarso" } ] } ```
1.0
Los nombres de las propiedades en `logistic_model_parameters.json` están en _spanglish_ - Abajo se puede ver que nombres de las propiedades están en _spanglish_: ``` { "parametrosNormalizacion": { "valorMinimo": { "longitudCraneo": [3.7037], "altoPico": [165.58], "longitudPico": [29.67], "tarso": [101.78] }, "valorMaximo": { "longitudCraneo": [83.18], "altoPico": [46.5], "longitudPico": [193.22], "tarso": [35.44] } }, "parametrosModelo": [ { "Variables": "(Intercept)", "Estimate": -18.948, "_row": "(Intercept)" }, { "Variables": "longitudCraneo", "Estimate": 6.576, "_row": "longitudCraneo" }, { "Variables": "altoPico", "Estimate": 8.816, "_row": "altoPico" }, { "Variables": "longitudPico", "Estimate": 7.172, "_row": "longitudPico" }, { "Variables": "tarso", "Estimate": 5.726, "_row": "tarso" } ] } ```
non_test
los nombres de las propiedades en logistic model parameters json están en spanglish abajo se puede ver que nombres de las propiedades están en spanglish parametrosnormalizacion valorminimo longitudcraneo altopico longitudpico tarso valormaximo longitudcraneo altopico longitudpico tarso parametrosmodelo variables intercept estimate row intercept variables longitudcraneo estimate row longitudcraneo variables altopico estimate row altopico variables longitudpico estimate row longitudpico variables tarso estimate row tarso
0
330,550
28,439,086,148
IssuesEvent
2023-04-15 17:26:59
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed
C-test-failure O-robot branch-release-23.1
pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test.TestSQLStatsCompactor [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9624052?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9624052?buildTab=artifacts#/) on release-23.1 @ [84905b8396e7cd7ef999b264faaafe13110d631a](https://github.com/cockroachdb/cockroach/commits/84905b8396e7cd7ef999b264faaafe13110d631a): ``` === RUN TestSQLStatsCompactor test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/9a30d5271a65be6b3f2aa1433be9011f/logTestSQLStatsCompactor803069076 test_log_scope.go:79: use -show-logs to present logs inline === CONT TestSQLStatsCompactor compaction_test.go:262: -- test log scope end -- test logs left over in: /artifacts/tmp/_tmp/9a30d5271a65be6b3f2aa1433be9011f/logTestSQLStatsCompactor803069076 --- FAIL: TestSQLStatsCompactor (137.60s) === RUN TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=205/rowsDeletePerTxn=0 compaction_test.go:227: Error Trace: pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:227 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:234 Error: Not equal: expected: 25 actual : 13 Test: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=205/rowsDeletePerTxn=0 Messages: expected 25 number of wide scans issued, but 13 number of wide scan issued --- FAIL: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=205/rowsDeletePerTxn=0 (16.75s) === RUN TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=1024 compaction_test.go:227: Error Trace: pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:227 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:234 Error: Not equal: expected: 32 actual : 16 Test: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=1024 Messages: expected 32 number of wide scans issued, but 16 number of wide scan issued --- FAIL: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=1024 (20.04s) === RUN TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=2 compaction_test.go:227: Error Trace: pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:227 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:234 Error: Not equal: expected: 32 actual : 24 Test: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=2 Messages: expected 32 number of wide scans issued, but 24 number of wide scan issued --- FAIL: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=2 (32.50s) ``` <p>Parameters: <code>TAGS=bazel,gss,deadlock</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #101361 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed [C-test-failure O-robot T-cluster-observability branch-release-23.1.0] - #94880 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed [C-test-failure O-robot branch-release-22.2] - #80442 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed [C-test-failure O-robot branch-master] </p> </details> /cc @cockroachdb/cluster-observability <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSQLStatsCompactor.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed - pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test.TestSQLStatsCompactor [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9624052?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9624052?buildTab=artifacts#/) on release-23.1 @ [84905b8396e7cd7ef999b264faaafe13110d631a](https://github.com/cockroachdb/cockroach/commits/84905b8396e7cd7ef999b264faaafe13110d631a): ``` === RUN TestSQLStatsCompactor test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/9a30d5271a65be6b3f2aa1433be9011f/logTestSQLStatsCompactor803069076 test_log_scope.go:79: use -show-logs to present logs inline === CONT TestSQLStatsCompactor compaction_test.go:262: -- test log scope end -- test logs left over in: /artifacts/tmp/_tmp/9a30d5271a65be6b3f2aa1433be9011f/logTestSQLStatsCompactor803069076 --- FAIL: TestSQLStatsCompactor (137.60s) === RUN TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=205/rowsDeletePerTxn=0 compaction_test.go:227: Error Trace: pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:227 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:234 Error: Not equal: expected: 25 actual : 13 Test: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=205/rowsDeletePerTxn=0 Messages: expected 25 number of wide scans issued, but 13 number of wide scan issued --- FAIL: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=205/rowsDeletePerTxn=0 (16.75s) === RUN TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=1024 compaction_test.go:227: Error Trace: pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:227 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:234 Error: Not equal: expected: 32 actual : 16 Test: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=1024 Messages: expected 32 number of wide scans issued, but 16 number of wide scan issued --- FAIL: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=1024 (20.04s) === RUN TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=2 compaction_test.go:227: Error Trace: pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:227 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:234 Error: Not equal: expected: 32 actual : 24 Test: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=2 Messages: expected 32 number of wide scans issued, but 24 number of wide scan issued --- FAIL: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=2 (32.50s) ``` <p>Parameters: <code>TAGS=bazel,gss,deadlock</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #101361 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed [C-test-failure O-robot T-cluster-observability branch-release-23.1.0] - #94880 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed [C-test-failure O-robot branch-release-22.2] - #80442 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed [C-test-failure O-robot branch-master] </p> </details> /cc @cockroachdb/cluster-observability <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSQLStatsCompactor.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
pkg sql sqlstats persistedsqlstats persistedsqlstats test testsqlstatscompactor failed pkg sql sqlstats persistedsqlstats persistedsqlstats test testsqlstatscompactor with on release run testsqlstatscompactor test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline cont testsqlstatscompactor compaction test go test log scope end test logs left over in artifacts tmp tmp fail testsqlstatscompactor run testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn compaction test go error trace pkg sql sqlstats persistedsqlstats persistedsqlstats test test pkg sql sqlstats persistedsqlstats compaction test go pkg sql sqlstats persistedsqlstats persistedsqlstats test test pkg sql sqlstats persistedsqlstats compaction test go error not equal expected actual test testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn messages expected number of wide scans issued but number of wide scan issued fail testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn run testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn compaction test go error trace pkg sql sqlstats persistedsqlstats persistedsqlstats test test pkg sql sqlstats persistedsqlstats compaction test go pkg sql sqlstats persistedsqlstats persistedsqlstats test test pkg sql sqlstats persistedsqlstats compaction test go error not equal expected actual test testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn messages expected number of wide scans issued but number of wide scan issued fail testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn run testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn compaction test go error trace pkg sql sqlstats persistedsqlstats persistedsqlstats test test pkg sql sqlstats persistedsqlstats compaction test go pkg sql sqlstats persistedsqlstats persistedsqlstats test test pkg sql sqlstats persistedsqlstats compaction test go error not equal expected actual test testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn messages expected number of wide scans issued but number of wide scan issued fail testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn parameters tags bazel gss deadlock help see also same failure on other branches pkg sql sqlstats persistedsqlstats persistedsqlstats test testsqlstatscompactor failed pkg sql sqlstats persistedsqlstats persistedsqlstats test testsqlstatscompactor failed pkg sql sqlstats persistedsqlstats persistedsqlstats test testsqlstatscompactor failed cc cockroachdb cluster observability
1
98,003
4,015,861,783
IssuesEvent
2016-05-15 07:04:30
oshri551/angular2Project
https://api.github.com/repos/oshri551/angular2Project
closed
Add Angular material 2
effort1: easy (2 hours) priority: normal Task
## Todos: * Add dependncy to package.json * Add angular material to vendors.ts
1.0
Add Angular material 2 - ## Todos: * Add dependncy to package.json * Add angular material to vendors.ts
non_test
add angular material todos add dependncy to package json add angular material to vendors ts
0
270,091
23,490,515,867
IssuesEvent
2022-08-17 18:14:31
tijlleenders/ZinZen
https://api.github.com/repos/tijlleenders/ZinZen
opened
Update the cypress test for Goals header
UI test
We are creating a change in the header by replacing the icon at the top right of the header. Cypress test has to be modified if there is PR that is targeting to resolve #591
1.0
Update the cypress test for Goals header - We are creating a change in the header by replacing the icon at the top right of the header. Cypress test has to be modified if there is PR that is targeting to resolve #591
test
update the cypress test for goals header we are creating a change in the header by replacing the icon at the top right of the header cypress test has to be modified if there is pr that is targeting to resolve
1
158,513
12,417,829,741
IssuesEvent
2020-05-22 21:48:20
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
closed
Design Issues - Dropdown
[zube]: To Test
When there is no label for a dropdown, the height of the dropdown is different than with a label. Should we make all dropdowns the same height regardless of label? Or will we require a label for every dropdown? This is from the Environment Variables from a Resource on the workloads page inside the Command tab. <img width="1072" alt="Screen Shot 2020-05-04 at 4 53 54 PM" src="https://user-images.githubusercontent.com/11410997/81024476-64406800-8e28-11ea-8c45-d03d052b59ed.png">
1.0
Design Issues - Dropdown - When there is no label for a dropdown, the height of the dropdown is different than with a label. Should we make all dropdowns the same height regardless of label? Or will we require a label for every dropdown? This is from the Environment Variables from a Resource on the workloads page inside the Command tab. <img width="1072" alt="Screen Shot 2020-05-04 at 4 53 54 PM" src="https://user-images.githubusercontent.com/11410997/81024476-64406800-8e28-11ea-8c45-d03d052b59ed.png">
test
design issues dropdown when there is no label for a dropdown the height of the dropdown is different than with a label should we make all dropdowns the same height regardless of label or will we require a label for every dropdown this is from the environment variables from a resource on the workloads page inside the command tab img width alt screen shot at pm src
1
5,181
2,572,399,558
IssuesEvent
2015-02-10 22:17:16
jasonhall/jasonhall
https://api.github.com/repos/jasonhall/jasonhall
opened
has labels
auto-migrated Priority-Low Type-Enhancement
``` this issue has labels that change with comments ``` ----- Original issue reported on code.google.com by jasonhall@google.com on 11 Dec 2014 at 8:07 * Blocked on: #8
1.0
has labels - ``` this issue has labels that change with comments ``` ----- Original issue reported on code.google.com by jasonhall@google.com on 11 Dec 2014 at 8:07 * Blocked on: #8
non_test
has labels this issue has labels that change with comments original issue reported on code google com by jasonhall google com on dec at blocked on
0
222,797
24,711,330,791
IssuesEvent
2022-10-20 01:14:05
theWhiteFox/ion-vue-app
https://api.github.com/repos/theWhiteFox/ion-vue-app
closed
WS-2020-0042 (High) detected in acorn-5.7.4.tgz - autoclosed
security vulnerability
## WS-2020-0042 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.4.tgz</b></p></summary> <p>ECMAScript parser</p> <p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/jsdom/node_modules/acorn/package.json</p> <p> Dependency Hierarchy: - cli-plugin-unit-jest-4.5.13.tgz (Root Library) - jest-24.9.0.tgz - jest-cli-24.9.0.tgz - jest-config-24.9.0.tgz - jest-environment-jsdom-24.9.0.tgz - jsdom-11.12.0.tgz - :x: **acorn-5.7.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/theWhiteFox/ion-vue-app/commit/5bed41ba25cd539697eec7f8a2456a9190a81333">5bed41ba25cd539697eec7f8a2456a9190a81333</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser. <p>Publish Date: 2020-03-01 <p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p> <p>Release Date: 2020-03-01</p> <p>Fix Resolution (acorn): 6.4.1</p> <p>Direct dependency fix Resolution (@vue/cli-plugin-unit-jest): 5.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2020-0042 (High) detected in acorn-5.7.4.tgz - autoclosed - ## WS-2020-0042 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.4.tgz</b></p></summary> <p>ECMAScript parser</p> <p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/jsdom/node_modules/acorn/package.json</p> <p> Dependency Hierarchy: - cli-plugin-unit-jest-4.5.13.tgz (Root Library) - jest-24.9.0.tgz - jest-cli-24.9.0.tgz - jest-config-24.9.0.tgz - jest-environment-jsdom-24.9.0.tgz - jsdom-11.12.0.tgz - :x: **acorn-5.7.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/theWhiteFox/ion-vue-app/commit/5bed41ba25cd539697eec7f8a2456a9190a81333">5bed41ba25cd539697eec7f8a2456a9190a81333</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser. <p>Publish Date: 2020-03-01 <p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p> <p>Release Date: 2020-03-01</p> <p>Fix Resolution (acorn): 6.4.1</p> <p>Direct dependency fix Resolution (@vue/cli-plugin-unit-jest): 5.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
ws high detected in acorn tgz autoclosed ws high severity vulnerability vulnerable library acorn tgz ecmascript parser library home page a href path to dependency file package json path to vulnerable library node modules jsdom node modules acorn package json dependency hierarchy cli plugin unit jest tgz root library jest tgz jest cli tgz jest config tgz jest environment jsdom tgz jsdom tgz x acorn tgz vulnerable library found in head commit a href found in base branch main vulnerability details acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution acorn direct dependency fix resolution vue cli plugin unit jest step up your open source security game with mend
0
162,242
12,640,436,819
IssuesEvent
2020-06-16 03:01:00
microsoft/azure-tools-for-java
https://api.github.com/repos/microsoft/azure-tools-for-java
closed
[intelliJ][Spark on Cosmos Serverless]Remotely run failed sometimes.
External Dependency HDInsight IntelliJ Internal Test Security Rating -- LOW
Build: develop 1247 Repro Steps: 1. Create a Spark on Cosmos Serverless configuration. 2. Remotely run it. Result: Job run failed. ![image](https://user-images.githubusercontent.com/35950097/64755294-3d61b880-d55d-11e9-8a66-35e48e7a88f3.png) Frequency: Sometimes.
1.0
[intelliJ][Spark on Cosmos Serverless]Remotely run failed sometimes. - Build: develop 1247 Repro Steps: 1. Create a Spark on Cosmos Serverless configuration. 2. Remotely run it. Result: Job run failed. ![image](https://user-images.githubusercontent.com/35950097/64755294-3d61b880-d55d-11e9-8a66-35e48e7a88f3.png) Frequency: Sometimes.
test
remotely run failed sometimes build develop repro steps create a spark on cosmos serverless configuration remotely run it result job run failed frequency sometimes
1
216,420
16,761,160,001
IssuesEvent
2021-06-13 20:19:53
snowpackjs/astro
https://api.github.com/repos/snowpackjs/astro
closed
Integration tests
testing
We've had issues in the past where some bugs (especially those related to module resolution) are not caught by integration tests due to have we have the monorepo set up and the use of snowpack's `workspaceRoot`. We would like to have some integration tests that run to smoke test released versions. They might do: 1. Run `npm init astro` and step through the questions. 2. Start up `astro dev` and make sure it doesn't crash. Fetch the HTML from a few routes and verify its content. 3. Run the build `astro build` and make sure it doesn't crash. Check that the correct output was created.
1.0
Integration tests - We've had issues in the past where some bugs (especially those related to module resolution) are not caught by integration tests due to have we have the monorepo set up and the use of snowpack's `workspaceRoot`. We would like to have some integration tests that run to smoke test released versions. They might do: 1. Run `npm init astro` and step through the questions. 2. Start up `astro dev` and make sure it doesn't crash. Fetch the HTML from a few routes and verify its content. 3. Run the build `astro build` and make sure it doesn't crash. Check that the correct output was created.
test
integration tests we ve had issues in the past where some bugs especially those related to module resolution are not caught by integration tests due to have we have the monorepo set up and the use of snowpack s workspaceroot we would like to have some integration tests that run to smoke test released versions they might do run npm init astro and step through the questions start up astro dev and make sure it doesn t crash fetch the html from a few routes and verify its content run the build astro build and make sure it doesn t crash check that the correct output was created
1
33,100
4,807,065,395
IssuesEvent
2016-11-02 20:21:41
moment/moment
https://api.github.com/repos/moment/moment
closed
3 tests failed. diff:diff across DST (2336.17) diff:diff across DST (2336.19) relative time:custom rounding (2630.3)
Unit Test Failed
### Client info ``` Date String : Tue Nov 01 2016 13:58:48 GMT-0600 (CST) Locale String : 11/1/2016, 1:58:48 PM Offset : 360 User Agent : Mozilla/5.0 (Macintosh; Intel Mac OS X 10_12_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/54.0.2840.71 Safari/537.36 Moment Version: 2.15.2 ``` ==== ### diff:diff across DST (2336.17) month diff across DST, lower bound ```javascript // Expected true // Actual false false === true ``` ==== ### diff:diff across DST (2336.19) year diff across DST, lower bound ```javascript // Expected true // Actual false false === true ``` ==== ### relative time:custom rounding (2630.3) Round down towards the nearest day ```javascript // Expected in 30 days // Actual in a month "in a month" === "in 30 days" ```
1.0
3 tests failed. diff:diff across DST (2336.17) diff:diff across DST (2336.19) relative time:custom rounding (2630.3) - ### Client info ``` Date String : Tue Nov 01 2016 13:58:48 GMT-0600 (CST) Locale String : 11/1/2016, 1:58:48 PM Offset : 360 User Agent : Mozilla/5.0 (Macintosh; Intel Mac OS X 10_12_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/54.0.2840.71 Safari/537.36 Moment Version: 2.15.2 ``` ==== ### diff:diff across DST (2336.17) month diff across DST, lower bound ```javascript // Expected true // Actual false false === true ``` ==== ### diff:diff across DST (2336.19) year diff across DST, lower bound ```javascript // Expected true // Actual false false === true ``` ==== ### relative time:custom rounding (2630.3) Round down towards the nearest day ```javascript // Expected in 30 days // Actual in a month "in a month" === "in 30 days" ```
test
tests failed diff diff across dst diff diff across dst relative time custom rounding client info date string tue nov gmt cst locale string pm offset user agent mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari moment version diff diff across dst month diff across dst lower bound javascript expected true actual false false true diff diff across dst year diff across dst lower bound javascript expected true actual false false true relative time custom rounding round down towards the nearest day javascript expected in days actual in a month in a month in days
1
373,987
26,094,895,141
IssuesEvent
2022-12-26 17:40:08
StraykerPL/RockPaperScissors
https://api.github.com/repos/StraykerPL/RockPaperScissors
closed
Feature: Add Testing label to repo
documentation
**Is your feature request related to a problem? Please describe.** There's no `testing` label for checking general parts of product. **Describe the solution you'd like** Add `testing` label in GitHub, with description `This needs to be tested`. **Describe alternatives you've considered** **Additional context**
1.0
Feature: Add Testing label to repo - **Is your feature request related to a problem? Please describe.** There's no `testing` label for checking general parts of product. **Describe the solution you'd like** Add `testing` label in GitHub, with description `This needs to be tested`. **Describe alternatives you've considered** **Additional context**
non_test
feature add testing label to repo is your feature request related to a problem please describe there s no testing label for checking general parts of product describe the solution you d like add testing label in github with description this needs to be tested describe alternatives you ve considered additional context
0
45,395
13,110,398,729
IssuesEvent
2020-08-04 20:35:04
mwilliams7197/jest-environment-serverless
https://api.github.com/repos/mwilliams7197/jest-environment-serverless
opened
CVE-2019-20149 (High) detected in kind-of-6.0.2.tgz
security vulnerability
## CVE-2019-20149 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kind-of-6.0.2.tgz</b></p></summary> <p>Get the native type of a value.</p> <p>Library home page: <a href="https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz">https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/jest-environment-serverless/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/jest-environment-serverless/node_modules/kind-of/package.json</p> <p> Dependency Hierarchy: - jest-config-24.8.0.tgz (Root Library) - micromatch-3.1.10.tgz - :x: **kind-of-6.0.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mwilliams7197/jest-environment-serverless/commit/02682832cf2379d2ecc74a2ddc62a35db5341137">02682832cf2379d2ecc74a2ddc62a35db5341137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by 'constructor': {'name':'Symbol'}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result. <p>Publish Date: 2019-12-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149>CVE-2019-20149</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149</a></p> <p>Release Date: 2019-12-30</p> <p>Fix Resolution: 6.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"kind-of","packageVersion":"6.0.2","isTransitiveDependency":true,"dependencyTree":"jest-config:24.8.0;micromatch:3.1.10;kind-of:6.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.3"}],"vulnerabilityIdentifier":"CVE-2019-20149","vulnerabilityDetails":"ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by \u0027constructor\u0027: {\u0027name\u0027:\u0027Symbol\u0027}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-20149 (High) detected in kind-of-6.0.2.tgz - ## CVE-2019-20149 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kind-of-6.0.2.tgz</b></p></summary> <p>Get the native type of a value.</p> <p>Library home page: <a href="https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz">https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/jest-environment-serverless/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/jest-environment-serverless/node_modules/kind-of/package.json</p> <p> Dependency Hierarchy: - jest-config-24.8.0.tgz (Root Library) - micromatch-3.1.10.tgz - :x: **kind-of-6.0.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mwilliams7197/jest-environment-serverless/commit/02682832cf2379d2ecc74a2ddc62a35db5341137">02682832cf2379d2ecc74a2ddc62a35db5341137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by 'constructor': {'name':'Symbol'}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result. <p>Publish Date: 2019-12-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149>CVE-2019-20149</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149</a></p> <p>Release Date: 2019-12-30</p> <p>Fix Resolution: 6.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"kind-of","packageVersion":"6.0.2","isTransitiveDependency":true,"dependencyTree":"jest-config:24.8.0;micromatch:3.1.10;kind-of:6.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.3"}],"vulnerabilityIdentifier":"CVE-2019-20149","vulnerabilityDetails":"ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by \u0027constructor\u0027: {\u0027name\u0027:\u0027Symbol\u0027}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_test
cve high detected in kind of tgz cve high severity vulnerability vulnerable library kind of tgz get the native type of a value library home page a href path to dependency file tmp ws scm jest environment serverless package json path to vulnerable library tmp ws scm jest environment serverless node modules kind of package json dependency hierarchy jest config tgz root library micromatch tgz x kind of tgz vulnerable library found in head commit a href vulnerability details ctorname in index js in kind of allows external user input to overwrite certain internal attributes via a conflicting name as demonstrated by constructor name symbol hence a crafted payload can overwrite this builtin attribute to manipulate the type detection result publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails ctorname in index js in kind of allows external user input to overwrite certain internal attributes via a conflicting name as demonstrated by hence a crafted payload can overwrite this builtin attribute to manipulate the type detection result vulnerabilityurl
0
279,582
24,236,252,614
IssuesEvent
2022-09-26 23:44:44
microsoft/vscode-python
https://api.github.com/repos/microsoft/vscode-python
opened
Test finalized API for python environments
testplan-item
Refs: https://github.com/microsoft/vscode-python/issues/19101 - [ ] anyOS: @rchiodo - [ ] anyOS: @DonJayamanne Complexity: 5 --- ### Requirements 1. Use Insiders build of the python extension ### Test APIs exposed via `ProposedExtensionAPI` See `ProposedExtensionAPI` for APIs to test: https://github.com/microsoft/vscode-python/blob/0f045780e58e190e57411f4ac1227cda97c4fde2/src/client/proposedApiTypes.ts#L8 Example usage: Copy over contents of https://github.com/microsoft/vscode-python/blob/main/src/client/proposedApiTypes.ts as needed. ```typescript const extension = extensions.getExtension('ms-python.python'); if (extension) { if (!extension.isActive) { await extension.activate(); } const api: IExtensionApi & IProposedExtensionAPI = extension.exports as IExtensionApi & IProposedExtensionAPI; if (api.environment) { const envID = api.environment.getActiveEnvironmentId(); } } ```
1.0
Test finalized API for python environments - Refs: https://github.com/microsoft/vscode-python/issues/19101 - [ ] anyOS: @rchiodo - [ ] anyOS: @DonJayamanne Complexity: 5 --- ### Requirements 1. Use Insiders build of the python extension ### Test APIs exposed via `ProposedExtensionAPI` See `ProposedExtensionAPI` for APIs to test: https://github.com/microsoft/vscode-python/blob/0f045780e58e190e57411f4ac1227cda97c4fde2/src/client/proposedApiTypes.ts#L8 Example usage: Copy over contents of https://github.com/microsoft/vscode-python/blob/main/src/client/proposedApiTypes.ts as needed. ```typescript const extension = extensions.getExtension('ms-python.python'); if (extension) { if (!extension.isActive) { await extension.activate(); } const api: IExtensionApi & IProposedExtensionAPI = extension.exports as IExtensionApi & IProposedExtensionAPI; if (api.environment) { const envID = api.environment.getActiveEnvironmentId(); } } ```
test
test finalized api for python environments refs anyos rchiodo anyos donjayamanne complexity requirements use insiders build of the python extension test apis exposed via proposedextensionapi see proposedextensionapi for apis to test example usage copy over contents of as needed typescript const extension extensions getextension ms python python if extension if extension isactive await extension activate const api iextensionapi iproposedextensionapi extension exports as iextensionapi iproposedextensionapi if api environment const envid api environment getactiveenvironmentid
1
65,130
8,788,950,967
IssuesEvent
2018-12-21 00:43:11
mozilla/dinobuildr
https://api.github.com/repos/mozilla/dinobuildr
opened
Epic: Clean up documentation
documentation enhancement
The documentation needs some TLC. #53 #59 #145 #68 All of these issues are docs related. If we add more before I finish the above list let's add them so there's an easy place to track all of the things we need to update.
1.0
Epic: Clean up documentation - The documentation needs some TLC. #53 #59 #145 #68 All of these issues are docs related. If we add more before I finish the above list let's add them so there's an easy place to track all of the things we need to update.
non_test
epic clean up documentation the documentation needs some tlc all of these issues are docs related if we add more before i finish the above list let s add them so there s an easy place to track all of the things we need to update
0
181,537
6,661,863,948
IssuesEvent
2017-10-02 10:32:05
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
mail.google.com - site is not usable
browser-firefox priority-critical status-needsinfo status-needstriage
<!-- @browser: Firefox 58.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.12; rv:58.0) Gecko/20100101 Firefox/58.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://mail.google.com/mail/u/0/ **Browser / Version**: Firefox 58.0 **Operating System**: Mac OS X 10.12 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Gmail never completes loading unless you activate legacy basic HTML mode. Tested it in Safari and it loads fine. **Steps to Reproduce**: layout.css.servo.enabled: true [![Screenshot Description](https://webcompat.com/uploads/2017/9/d7b49bbd-7f9c-4e4e-bd33-dbed5e8823e8-thumb.jpg)](https://webcompat.com/uploads/2017/9/d7b49bbd-7f9c-4e4e-bd33-dbed5e8823e8.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
mail.google.com - site is not usable - <!-- @browser: Firefox 58.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.12; rv:58.0) Gecko/20100101 Firefox/58.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://mail.google.com/mail/u/0/ **Browser / Version**: Firefox 58.0 **Operating System**: Mac OS X 10.12 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Gmail never completes loading unless you activate legacy basic HTML mode. Tested it in Safari and it loads fine. **Steps to Reproduce**: layout.css.servo.enabled: true [![Screenshot Description](https://webcompat.com/uploads/2017/9/d7b49bbd-7f9c-4e4e-bd33-dbed5e8823e8-thumb.jpg)](https://webcompat.com/uploads/2017/9/d7b49bbd-7f9c-4e4e-bd33-dbed5e8823e8.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
mail google com site is not usable url browser version firefox operating system mac os x tested another browser yes problem type site is not usable description gmail never completes loading unless you activate legacy basic html mode tested it in safari and it loads fine steps to reproduce layout css servo enabled true from with ❤️
0
537,302
15,726,840,891
IssuesEvent
2021-03-29 11:53:24
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
xt-xcc does not support deprecated attribute
area: Toolchains bug priority: medium
**Describe the bug** file:gcc.h ``` #ifndef __deprecated #define __deprecated __attribute__((deprecated)) #endif ``` file:bluetooth.h ``` BT_LE_SCAN_FILTER_DUPLICATE __deprecated = BT_LE_SCAN_OPT_FILTER_DUPLICATE, ``` xt-xcc does not support deprecated attribute. **Logs and console output** ``` [17/152] Building C object zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj FAILED: zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj C:\usr\xtensa\XtDevTools\install\tools\RI-2020.5-win32\XtensaTools\bin\xt-xcc.exe -DBUILD_VERSION=zephyr-v2.4.0-1113-g0bcd66ac36d8 -DKERNEL -D_FORTIFY_SOURCE=2 -D__ZEPHYR__=1 -I../include -Izephyr/include/generated -I../soc/xtensa/ZC3827 -I../soc/xtensa/ZC3827/include -I../subsys/settings/include -I../subsys/bluetooth -IE:/work/rtos_work/zephyrproject/modules/crypto/tinycrypt/lib/include -IE:/work/rtos_work/zephyrproject/modules/hal/xtensa/include -IE:/work/rtos_work/zephyrproject/modules/hal/xtensa/zephyr/soc/ZC3827 -isystem ../lib/libc/minimal/include -Os -imacros E:/work/rtos_work/zephyrproject/zephyr/build/zephyr/include/generated/autoconf.h -ffreestanding -fno-common -g -imacrosE:/work/rtos_work/zephyrproject/zephyr/include/toolchain/xcc_missing_defs.h -fms-extensions -imacros E:/work/rtos_work/zephyrproject/zephyr/include/toolchain/zephyr_stdint.h -Wall -Wformat -Wformat-security -Wno-format-zero-length -Wno-main -Wno-pointer-sign -Wpointer-arith -Werror=implicit-int -fno-pic -fno-strict-overflow -ffunction-sections -fdata-sections -mlongcalls -std=c99 -MD -MT zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj -MF zephyr\subsys\bluetooth\host\CMakeFiles\subsys__bluetooth__host.dir\conn.c.obj.d -o zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj -c E:/work/rtos_work/zephyrproject/zephyr/subsys/bluetooth/host/conn.c In file included from ../include/bluetooth/conn.h:23, from ../include/bluetooth/hci.h:20, from E:/work/rtos_work/zephyrproject/zephyr/subsys/bluetooth/host/conn.c:21: ../include/bluetooth/bluetooth.h:1240: error: expected ',' or '}' before '__attribute__' In file included from ../include/bluetooth/hci.h:20, from E:/work/rtos_work/zephyrproject/zephyr/subsys/bluetooth/host/conn.c:21: ../include/bluetooth/conn.h:685: error: expected ',' or '}' before '__attribute__' ``` **Environment (please complete the following information):** - OS: ( Windows) - Toolchain (xt-cc.)
1.0
xt-xcc does not support deprecated attribute - **Describe the bug** file:gcc.h ``` #ifndef __deprecated #define __deprecated __attribute__((deprecated)) #endif ``` file:bluetooth.h ``` BT_LE_SCAN_FILTER_DUPLICATE __deprecated = BT_LE_SCAN_OPT_FILTER_DUPLICATE, ``` xt-xcc does not support deprecated attribute. **Logs and console output** ``` [17/152] Building C object zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj FAILED: zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj C:\usr\xtensa\XtDevTools\install\tools\RI-2020.5-win32\XtensaTools\bin\xt-xcc.exe -DBUILD_VERSION=zephyr-v2.4.0-1113-g0bcd66ac36d8 -DKERNEL -D_FORTIFY_SOURCE=2 -D__ZEPHYR__=1 -I../include -Izephyr/include/generated -I../soc/xtensa/ZC3827 -I../soc/xtensa/ZC3827/include -I../subsys/settings/include -I../subsys/bluetooth -IE:/work/rtos_work/zephyrproject/modules/crypto/tinycrypt/lib/include -IE:/work/rtos_work/zephyrproject/modules/hal/xtensa/include -IE:/work/rtos_work/zephyrproject/modules/hal/xtensa/zephyr/soc/ZC3827 -isystem ../lib/libc/minimal/include -Os -imacros E:/work/rtos_work/zephyrproject/zephyr/build/zephyr/include/generated/autoconf.h -ffreestanding -fno-common -g -imacrosE:/work/rtos_work/zephyrproject/zephyr/include/toolchain/xcc_missing_defs.h -fms-extensions -imacros E:/work/rtos_work/zephyrproject/zephyr/include/toolchain/zephyr_stdint.h -Wall -Wformat -Wformat-security -Wno-format-zero-length -Wno-main -Wno-pointer-sign -Wpointer-arith -Werror=implicit-int -fno-pic -fno-strict-overflow -ffunction-sections -fdata-sections -mlongcalls -std=c99 -MD -MT zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj -MF zephyr\subsys\bluetooth\host\CMakeFiles\subsys__bluetooth__host.dir\conn.c.obj.d -o zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj -c E:/work/rtos_work/zephyrproject/zephyr/subsys/bluetooth/host/conn.c In file included from ../include/bluetooth/conn.h:23, from ../include/bluetooth/hci.h:20, from E:/work/rtos_work/zephyrproject/zephyr/subsys/bluetooth/host/conn.c:21: ../include/bluetooth/bluetooth.h:1240: error: expected ',' or '}' before '__attribute__' In file included from ../include/bluetooth/hci.h:20, from E:/work/rtos_work/zephyrproject/zephyr/subsys/bluetooth/host/conn.c:21: ../include/bluetooth/conn.h:685: error: expected ',' or '}' before '__attribute__' ``` **Environment (please complete the following information):** - OS: ( Windows) - Toolchain (xt-cc.)
non_test
xt xcc does not support deprecated attribute describe the bug file gcc h ifndef deprecated define deprecated attribute deprecated endif file bluetooth h bt le scan filter duplicate deprecated bt le scan opt filter duplicate xt xcc does not support deprecated attribute logs and console output building c object zephyr subsys bluetooth host cmakefiles subsys bluetooth host dir conn c obj failed zephyr subsys bluetooth host cmakefiles subsys bluetooth host dir conn c obj c usr xtensa xtdevtools install tools ri xtensatools bin xt xcc exe dbuild version zephyr dkernel d fortify source d zephyr i include izephyr include generated i soc xtensa i soc xtensa include i subsys settings include i subsys bluetooth ie work rtos work zephyrproject modules crypto tinycrypt lib include ie work rtos work zephyrproject modules hal xtensa include ie work rtos work zephyrproject modules hal xtensa zephyr soc isystem lib libc minimal include os imacros e work rtos work zephyrproject zephyr build zephyr include generated autoconf h ffreestanding fno common g imacrose work rtos work zephyrproject zephyr include toolchain xcc missing defs h fms extensions imacros e work rtos work zephyrproject zephyr include toolchain zephyr stdint h wall wformat wformat security wno format zero length wno main wno pointer sign wpointer arith werror implicit int fno pic fno strict overflow ffunction sections fdata sections mlongcalls std md mt zephyr subsys bluetooth host cmakefiles subsys bluetooth host dir conn c obj mf zephyr subsys bluetooth host cmakefiles subsys bluetooth host dir conn c obj d o zephyr subsys bluetooth host cmakefiles subsys bluetooth host dir conn c obj c e work rtos work zephyrproject zephyr subsys bluetooth host conn c in file included from include bluetooth conn h from include bluetooth hci h from e work rtos work zephyrproject zephyr subsys bluetooth host conn c include bluetooth bluetooth h error expected or before attribute in file included from include bluetooth hci h from e work rtos work zephyrproject zephyr subsys bluetooth host conn c include bluetooth conn h error expected or before attribute environment please complete the following information os windows toolchain xt cc
0
214,029
16,554,644,016
IssuesEvent
2021-05-28 12:38:57
Realm667/WolfenDoom
https://api.github.com/repos/Realm667/WolfenDoom
closed
C2M5_A - Stahlhimmel - moving train
mapping playtesting suggestion
C2M5_A ... While on the moving train, and outside, Things scroll. This area has sector type 46 - Wind south weak. The effect is not weak at all. I suggest removing this sector type from this area. The wind is gale force ! Not a gook look IMO. Also, somehow, Dirty Douglas fell from the train. To "retrieve" him, I had to walk back to the rear carriage of the train. UPDATE ... I reproduced this. He can fall at line 4903. This line has the block monster flag. Would also adding the "block players" flag help ? There needs to be a door or something at sector 1335. Player can got back and forth between sectors 1359 and 1377 and the train starts/stops. This is not wanted. Once the player reaches sector 1359, he should not be able to go back. Train has stopped.
1.0
C2M5_A - Stahlhimmel - moving train - C2M5_A ... While on the moving train, and outside, Things scroll. This area has sector type 46 - Wind south weak. The effect is not weak at all. I suggest removing this sector type from this area. The wind is gale force ! Not a gook look IMO. Also, somehow, Dirty Douglas fell from the train. To "retrieve" him, I had to walk back to the rear carriage of the train. UPDATE ... I reproduced this. He can fall at line 4903. This line has the block monster flag. Would also adding the "block players" flag help ? There needs to be a door or something at sector 1335. Player can got back and forth between sectors 1359 and 1377 and the train starts/stops. This is not wanted. Once the player reaches sector 1359, he should not be able to go back. Train has stopped.
test
a stahlhimmel moving train a while on the moving train and outside things scroll this area has sector type wind south weak the effect is not weak at all i suggest removing this sector type from this area the wind is gale force not a gook look imo also somehow dirty douglas fell from the train to retrieve him i had to walk back to the rear carriage of the train update i reproduced this he can fall at line this line has the block monster flag would also adding the block players flag help there needs to be a door or something at sector player can got back and forth between sectors and and the train starts stops this is not wanted once the player reaches sector he should not be able to go back train has stopped
1
130,614
10,618,153,911
IssuesEvent
2019-10-13 01:31:42
magento/graphql-ce
https://api.github.com/repos/magento/graphql-ce
closed
[Test coverage] Cover CartAddressTypeResolver
Award: test coverage Component: QuoteGraphQl Progress: PR created good first issue
### Description: Cover with api-functional tests ``` QuoteGraphQl\Model\Resolver\CartAddressTypeResolver ```
1.0
[Test coverage] Cover CartAddressTypeResolver - ### Description: Cover with api-functional tests ``` QuoteGraphQl\Model\Resolver\CartAddressTypeResolver ```
test
cover cartaddresstyperesolver description cover with api functional tests quotegraphql model resolver cartaddresstyperesolver
1
84,407
15,720,901,073
IssuesEvent
2021-03-29 01:33:55
benchmarkdebricked/generator-jhipster
https://api.github.com/repos/benchmarkdebricked/generator-jhipster
closed
CVE-2020-5398 (High) detected in spring-web-5.0.6.RELEASE.jar - autoclosed
security vulnerability
## CVE-2020-5398 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.0.6.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /tmp/ws-scm/generator-jhipster/test/templates/ci-cd/maven-ngx-yarn/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-web/5.0.6.RELEASE/spring-web-5.0.6.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.2.RELEASE.jar (Root Library) - :x: **spring-web-5.0.6.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/benchmarkdebricked/generator-jhipster/commit/56521ff393b8726bd648cc9bf6e2fd9552b7cc38">56521ff393b8726bd648cc9bf6e2fd9552b7cc38</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework, versions 5.2.x prior to 5.2.3, versions 5.1.x prior to 5.1.13, and versions 5.0.x prior to 5.0.16, an application is vulnerable to a reflected file download (RFD) attack when it sets a "Content-Disposition" header in the response where the filename attribute is derived from user supplied input. <p>Publish Date: 2020-01-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-5398>CVE-2020-5398</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://pivotal.io/security/cve-2020-5398">https://pivotal.io/security/cve-2020-5398</a></p> <p>Release Date: 2020-01-17</p> <p>Fix Resolution: org.springframework:spring-web:5.0.16.RELEASE,org.springframework:spring-web:5.1.13.RELEASE,org.springframework:spring-web:5.2.3.RELEASE</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-5398 (High) detected in spring-web-5.0.6.RELEASE.jar - autoclosed - ## CVE-2020-5398 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.0.6.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /tmp/ws-scm/generator-jhipster/test/templates/ci-cd/maven-ngx-yarn/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-web/5.0.6.RELEASE/spring-web-5.0.6.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.2.RELEASE.jar (Root Library) - :x: **spring-web-5.0.6.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/benchmarkdebricked/generator-jhipster/commit/56521ff393b8726bd648cc9bf6e2fd9552b7cc38">56521ff393b8726bd648cc9bf6e2fd9552b7cc38</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework, versions 5.2.x prior to 5.2.3, versions 5.1.x prior to 5.1.13, and versions 5.0.x prior to 5.0.16, an application is vulnerable to a reflected file download (RFD) attack when it sets a "Content-Disposition" header in the response where the filename attribute is derived from user supplied input. <p>Publish Date: 2020-01-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-5398>CVE-2020-5398</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://pivotal.io/security/cve-2020-5398">https://pivotal.io/security/cve-2020-5398</a></p> <p>Release Date: 2020-01-17</p> <p>Fix Resolution: org.springframework:spring-web:5.0.16.RELEASE,org.springframework:spring-web:5.1.13.RELEASE,org.springframework:spring-web:5.2.3.RELEASE</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in spring web release jar autoclosed cve high severity vulnerability vulnerable library spring web release jar spring web library home page a href path to dependency file tmp ws scm generator jhipster test templates ci cd maven ngx yarn pom xml path to vulnerable library root repository org springframework spring web release spring web release jar dependency hierarchy spring boot starter web release jar root library x spring web release jar vulnerable library found in head commit a href vulnerability details in spring framework versions x prior to versions x prior to and versions x prior to an application is vulnerable to a reflected file download rfd attack when it sets a content disposition header in the response where the filename attribute is derived from user supplied input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring web release org springframework spring web release org springframework spring web release step up your open source security game with whitesource
0
287,176
24,814,680,744
IssuesEvent
2022-10-25 12:16:54
guardicore/monkey
https://api.github.com/repos/guardicore/monkey
closed
Add automated integration tests for post-breach actions
Impact: High Complexity: Medium Testing
Our exploiters are tested nightly in our blackbox test suite. In order to expedite the release process, as well as catch bugs early, all post-breach actions need automated tests.
1.0
Add automated integration tests for post-breach actions - Our exploiters are tested nightly in our blackbox test suite. In order to expedite the release process, as well as catch bugs early, all post-breach actions need automated tests.
test
add automated integration tests for post breach actions our exploiters are tested nightly in our blackbox test suite in order to expedite the release process as well as catch bugs early all post breach actions need automated tests
1
142,748
11,491,959,967
IssuesEvent
2020-02-11 20:00:49
fica99/42sh
https://api.github.com/repos/fica99/42sh
closed
Lexer
Lexer Testing
Lexer ===================== Описание функций лексера ----------------------------------- **Прототип функции:** ```t_lex_tkn *lex_get_next_tkn(char **str, size_t pos)``` **Описание параметров функции:** * `char **str` - входная строка подлежащая анализу; * `size_t pos` - индекс позиции с которой необходимо провести анализ; Функция возвращает токен одного из типов, в случае конца строки возвращается токен `"END"`, в случае ошибки возвращается `"NULL"`. *** **Прототип функции:** ```t_lex_tkn *lex_del_tkn(t_lex_tkn *token)``` **Описание параметров функции:** * `t_lex_tkn *token` - токен; Функция производит удаление токена. *** **Прототип функции:** ```t_lex_tkn **lex_get_tkns(char **str)``` **Описание параметров функции:** * `char **str` - входная строка подлежащая анализу; Функция возвращает массив токенов, заканчивающийся значением `"NULL"`, в случае ошибки возвращается `"NULL"`. *** **Прототип функции:** ```t_lex_tkn **lex_del_tkns(t_lex_tkn **tokens)``` **Описание параметров функции:** * `t_lex_tkn **tokens` - массив токенов; Функция производит удаление массива токенов. *** Описание структуры данных ----------------------------------- **Структура токена:** ``` typedef struct s_lex_tkn { char *value; t_lex_tkn_type type; t_lex_tkn_class class; size_t start_pos; size_t end_pos; } t_lex_tkn; ``` *** **Поля структуры токена:** * `value` - строковое значение токена; * `type` - хранит тип токена; * `class` - хранит класс токена; * `start_pos` - индекс начала позиции; * `end_pos` - индекс конца позиции; *** **Типы возвращаемых токенов:** * `T_NULL` - пустое значение; * `T_ERR` - ошибка; * `T_END` - `"\0"`; * `T_WORD` - `"ls -l ~/"`; * `T_PIPE` - `"|"`; * `T_GREATER` - `">"`; * `T_GREATER_GREATER` - `">>"`; * `T_LESS` - `"<"`; * `T_LESS_LESS` - `"<<"`; * `T_LESS_AND` - `"<&"`; * `T_GREATER_AND` - `">&"`; * `T_SEP` - `";"`; * `T_CONTROL_SUB` - `"$()"`; * `T_AND_AND` - `"&&"`; * `T_OR_OR` - `"||"`; * `T_AND` - `"&"`; * `T_OPEN_FIG_BRACE` - `"{"`; * `T_CLOSE_FIG_BRACE` - `"}"`; * `T_ARITH_OPERS` - `"$(())"`. *** **Классы возвращаемых токенов:** * `C_NULL` - `"T_NULL"`; * `C_END` - `"\0"`; * `C_WORD` - `"T_WORD"`; * `C_PIPE` - `"|"`; * `C_REDIR` - `"<", "<<", ">", ">>", "<&", ">&"`; * `C_SEP` - `";"`; * `C_CONTROL_SUB` - `"$()"`; * `C_LOG_OPERS` - `"&&", "||"`; * `C_AND` - `"&"`; * `C_FIG_BRACE` - `"{". "}"`; * `C_ARITH_OPERS` - `"$(())"`;
1.0
Lexer - Lexer ===================== Описание функций лексера ----------------------------------- **Прототип функции:** ```t_lex_tkn *lex_get_next_tkn(char **str, size_t pos)``` **Описание параметров функции:** * `char **str` - входная строка подлежащая анализу; * `size_t pos` - индекс позиции с которой необходимо провести анализ; Функция возвращает токен одного из типов, в случае конца строки возвращается токен `"END"`, в случае ошибки возвращается `"NULL"`. *** **Прототип функции:** ```t_lex_tkn *lex_del_tkn(t_lex_tkn *token)``` **Описание параметров функции:** * `t_lex_tkn *token` - токен; Функция производит удаление токена. *** **Прототип функции:** ```t_lex_tkn **lex_get_tkns(char **str)``` **Описание параметров функции:** * `char **str` - входная строка подлежащая анализу; Функция возвращает массив токенов, заканчивающийся значением `"NULL"`, в случае ошибки возвращается `"NULL"`. *** **Прототип функции:** ```t_lex_tkn **lex_del_tkns(t_lex_tkn **tokens)``` **Описание параметров функции:** * `t_lex_tkn **tokens` - массив токенов; Функция производит удаление массива токенов. *** Описание структуры данных ----------------------------------- **Структура токена:** ``` typedef struct s_lex_tkn { char *value; t_lex_tkn_type type; t_lex_tkn_class class; size_t start_pos; size_t end_pos; } t_lex_tkn; ``` *** **Поля структуры токена:** * `value` - строковое значение токена; * `type` - хранит тип токена; * `class` - хранит класс токена; * `start_pos` - индекс начала позиции; * `end_pos` - индекс конца позиции; *** **Типы возвращаемых токенов:** * `T_NULL` - пустое значение; * `T_ERR` - ошибка; * `T_END` - `"\0"`; * `T_WORD` - `"ls -l ~/"`; * `T_PIPE` - `"|"`; * `T_GREATER` - `">"`; * `T_GREATER_GREATER` - `">>"`; * `T_LESS` - `"<"`; * `T_LESS_LESS` - `"<<"`; * `T_LESS_AND` - `"<&"`; * `T_GREATER_AND` - `">&"`; * `T_SEP` - `";"`; * `T_CONTROL_SUB` - `"$()"`; * `T_AND_AND` - `"&&"`; * `T_OR_OR` - `"||"`; * `T_AND` - `"&"`; * `T_OPEN_FIG_BRACE` - `"{"`; * `T_CLOSE_FIG_BRACE` - `"}"`; * `T_ARITH_OPERS` - `"$(())"`. *** **Классы возвращаемых токенов:** * `C_NULL` - `"T_NULL"`; * `C_END` - `"\0"`; * `C_WORD` - `"T_WORD"`; * `C_PIPE` - `"|"`; * `C_REDIR` - `"<", "<<", ">", ">>", "<&", ">&"`; * `C_SEP` - `";"`; * `C_CONTROL_SUB` - `"$()"`; * `C_LOG_OPERS` - `"&&", "||"`; * `C_AND` - `"&"`; * `C_FIG_BRACE` - `"{". "}"`; * `C_ARITH_OPERS` - `"$(())"`;
test
lexer lexer описание функций лексера прототип функции t lex tkn lex get next tkn char str size t pos описание параметров функции char str входная строка подлежащая анализу size t pos индекс позиции с которой необходимо провести анализ функция возвращает токен одного из типов в случае конца строки возвращается токен end в случае ошибки возвращается null прототип функции t lex tkn lex del tkn t lex tkn token описание параметров функции t lex tkn token токен функция производит удаление токена прототип функции t lex tkn lex get tkns char str описание параметров функции char str входная строка подлежащая анализу функция возвращает массив токенов заканчивающийся значением null в случае ошибки возвращается null прототип функции t lex tkn lex del tkns t lex tkn tokens описание параметров функции t lex tkn tokens массив токенов функция производит удаление массива токенов описание структуры данных структура токена typedef struct s lex tkn char value t lex tkn type type t lex tkn class class size t start pos size t end pos t lex tkn поля структуры токена value строковое значение токена type хранит тип токена class хранит класс токена start pos индекс начала позиции end pos индекс конца позиции типы возвращаемых токенов t null пустое значение t err ошибка t end t word ls l t pipe t greater t greater greater t less t less less t less and t greater and t sep t control sub t and and t or or t and t open fig brace t close fig brace t arith opers классы возвращаемых токенов c null t null c end c word t word c pipe c redir c sep c control sub c log opers c and c fig brace c arith opers
1
7,217
2,610,358,233
IssuesEvent
2015-02-26 19:55:59
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
opened
Can't figure this out.
auto-migrated Priority-Medium Type-Defect
``` What's the problem? Can't figure out how to set this up to post on my wordpress blog and I can't find any documentation. The setup requires an API URL and even my best programmer has no idea what you're talking about. The old version worked fine, but when it updated it seems to have lost all my presets. What browser are you using? Firefox What version of ScribeFire are you running? 4 ``` ----- Original issue reported on code.google.com by `cpasites...@marshallhomeweb.com` on 16 Nov 2011 at 1:06
1.0
Can't figure this out. - ``` What's the problem? Can't figure out how to set this up to post on my wordpress blog and I can't find any documentation. The setup requires an API URL and even my best programmer has no idea what you're talking about. The old version worked fine, but when it updated it seems to have lost all my presets. What browser are you using? Firefox What version of ScribeFire are you running? 4 ``` ----- Original issue reported on code.google.com by `cpasites...@marshallhomeweb.com` on 16 Nov 2011 at 1:06
non_test
can t figure this out what s the problem can t figure out how to set this up to post on my wordpress blog and i can t find any documentation the setup requires an api url and even my best programmer has no idea what you re talking about the old version worked fine but when it updated it seems to have lost all my presets what browser are you using firefox what version of scribefire are you running original issue reported on code google com by cpasites marshallhomeweb com on nov at
0
142,566
21,785,754,881
IssuesEvent
2022-05-14 05:10:44
depromeet/antoon_web
https://api.github.com/repos/depromeet/antoon_web
closed
[회원가입] 스크린 디자인 반영
v0.1 component design
## 💡 개요 - 사용자 회원 가입 페이지의 스크린 디자인 변경 사항 반영하기 ## 📑 작업 사항 - [x] 회원가입 페이지의 css를 수정했습니다. ## 🔎 기타
1.0
[회원가입] 스크린 디자인 반영 - ## 💡 개요 - 사용자 회원 가입 페이지의 스크린 디자인 변경 사항 반영하기 ## 📑 작업 사항 - [x] 회원가입 페이지의 css를 수정했습니다. ## 🔎 기타
non_test
스크린 디자인 반영 💡 개요 사용자 회원 가입 페이지의 스크린 디자인 변경 사항 반영하기 📑 작업 사항 회원가입 페이지의 css를 수정했습니다 🔎 기타
0
128,081
18,025,744,977
IssuesEvent
2021-09-17 04:07:52
scriptex/atanas.info
https://api.github.com/repos/scriptex/atanas.info
closed
CVE-2021-3801 (Medium) detected in prismjs-1.24.1.tgz
security vulnerability
## CVE-2021-3801 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prismjs-1.24.1.tgz</b></p></summary> <p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p> <p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.24.1.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.24.1.tgz</a></p> <p>Path to dependency file: atanas.info/package.json</p> <p>Path to vulnerable library: atanas.info/node_modules/prismjs/package.json</p> <p> Dependency Hierarchy: - vuepress-1.8.2.tgz (Root Library) - core-1.8.2.tgz - markdown-1.8.2.tgz - :x: **prismjs-1.24.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/scriptex/atanas.info/commit/8571f344549c5a72fcd59f7a81cc3ede049ae9c0">8571f344549c5a72fcd59f7a81cc3ede049ae9c0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> prism is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3801>CVE-2021-3801</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-3801 (Medium) detected in prismjs-1.24.1.tgz - ## CVE-2021-3801 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prismjs-1.24.1.tgz</b></p></summary> <p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p> <p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.24.1.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.24.1.tgz</a></p> <p>Path to dependency file: atanas.info/package.json</p> <p>Path to vulnerable library: atanas.info/node_modules/prismjs/package.json</p> <p> Dependency Hierarchy: - vuepress-1.8.2.tgz (Root Library) - core-1.8.2.tgz - markdown-1.8.2.tgz - :x: **prismjs-1.24.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/scriptex/atanas.info/commit/8571f344549c5a72fcd59f7a81cc3ede049ae9c0">8571f344549c5a72fcd59f7a81cc3ede049ae9c0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> prism is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3801>CVE-2021-3801</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in prismjs tgz cve medium severity vulnerability vulnerable library prismjs tgz lightweight robust elegant syntax highlighting a spin off project from dabblet library home page a href path to dependency file atanas info package json path to vulnerable library atanas info node modules prismjs package json dependency hierarchy vuepress tgz root library core tgz markdown tgz x prismjs tgz vulnerable library found in head commit a href vulnerability details prism is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href step up your open source security game with whitesource
0
452,552
32,062,087,067
IssuesEvent
2023-09-24 19:23:40
vapoursynth/vapoursynth
https://api.github.com/repos/vapoursynth/vapoursynth
closed
Documentation issues
Area: Documentation
There should be nicer link to the website in the docs, for example in the right hand navigation bar
1.0
Documentation issues - There should be nicer link to the website in the docs, for example in the right hand navigation bar
non_test
documentation issues there should be nicer link to the website in the docs for example in the right hand navigation bar
0
72,267
7,293,212,898
IssuesEvent
2018-02-25 11:46:58
haaspors/rlib
https://api.github.com/repos/haaspors/rlib
closed
tests: all tests don't run on windows debug
bug tests
MSC adds more padding between test data so that the magic test finder code desn't work.
1.0
tests: all tests don't run on windows debug - MSC adds more padding between test data so that the magic test finder code desn't work.
test
tests all tests don t run on windows debug msc adds more padding between test data so that the magic test finder code desn t work
1
81,574
10,150,040,901
IssuesEvent
2019-08-05 16:36:15
universelabs/universe-design
https://api.github.com/repos/universelabs/universe-design
closed
Design instruction card for Universe Feather packaging
design
**Description** Design instruction card for Universe Feather packaging. **Tasks** - [x] ~~Create document.~~ - [x] ~~Design packaging card for Universe Feather device.~~ - [x] ~~Design packaging card dielines, folds, etc.~~
1.0
Design instruction card for Universe Feather packaging - **Description** Design instruction card for Universe Feather packaging. **Tasks** - [x] ~~Create document.~~ - [x] ~~Design packaging card for Universe Feather device.~~ - [x] ~~Design packaging card dielines, folds, etc.~~
non_test
design instruction card for universe feather packaging description design instruction card for universe feather packaging tasks create document design packaging card for universe feather device design packaging card dielines folds etc
0
302,172
26,129,899,292
IssuesEvent
2022-12-29 02:16:51
apache/helix
https://api.github.com/repos/apache/helix
opened
[Failed CI Test] testGetAllMetadataStoreRealms(org.apache.helix.rest.server.TestMetadataStoreDirectoryAccessor)
FailedTestTracking
This issue is created for tracking unstable test: testGetAllMetadataStoreRealms(org.apache.helix.rest.server.TestMetadataStoreDirectoryAccessor)
1.0
[Failed CI Test] testGetAllMetadataStoreRealms(org.apache.helix.rest.server.TestMetadataStoreDirectoryAccessor) - This issue is created for tracking unstable test: testGetAllMetadataStoreRealms(org.apache.helix.rest.server.TestMetadataStoreDirectoryAccessor)
test
testgetallmetadatastorerealms org apache helix rest server testmetadatastoredirectoryaccessor this issue is created for tracking unstable test testgetallmetadatastorerealms org apache helix rest server testmetadatastoredirectoryaccessor
1
292,420
25,210,968,176
IssuesEvent
2022-11-14 03:50:17
pingcap/tidb
https://api.github.com/repos/pingcap/tidb
closed
DATA RACE in the (*Chunk).resetForReuse()
type/bug component/test severity/major affects-6.4
## Bug Report Please answer these questions before submitting your issue. Thanks! ### 1. Minimal reproduce step (Required) ``` ================== WARNING: DATA RACE Write at 0x00c02a98e450 by goroutine 702063: github.com/pingcap/tidb/util/chunk.(*Chunk).resetForReuse() util/chunk/chunk.go:137 +0x7b8 github.com/pingcap/tidb/util/chunk.(*allocator).Reset() util/chunk/alloc.go:123 +0x12a github.com/pingcap/tidb/testkit.(*TestKit).MustExec.func1() testkit/testkit.go:120 +0xde runtime.deferreturn() GOROOT/src/runtime/panic.go:476 +0x32 github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1561 +0x6aa github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute() executor/projection.go:223 +0x651 github.com/pingcap/tidb/executor.(*ProjectionExec).Next() executor/projection.go:181 +0xcc github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute() executor/projection.go:223 +0x651 github.com/pingcap/tidb/executor.(*ProjectionExec).Next() executor/projection.go:181 +0xcc github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute() executor/projection.go:223 +0x651 github.com/pingcap/tidb/executor.(*ProjectionExec).Next() executor/projection.go:181 +0xcc github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute() executor/projection.go:223 +0x651 github.com/pingcap/tidb/executor.(*ProjectionExec).Next() executor/projection.go:181 +0xcc github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows() executor/aggregate.go:1382 +0x124 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1361 +0xad1 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk() executor/aggregate.go:1429 +0x276 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1366 +0xb24 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows() executor/aggregate.go:1382 +0x124 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1361 +0xad1 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk() executor/aggregate.go:1429 +0x276 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1366 +0xb24 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows() executor/aggregate.go:1382 +0x124 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1361 +0xad1 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows() executor/aggregate.go:1382 +0x124 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1361 +0xad1 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows() executor/aggregate.go:1382 +0x124 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeCurGroupRowsAndFetchChild() executor/aggregate.go:1390 +0x64 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1343 +0x70c github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk() executor/aggregate.go:1429 +0x276 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1366 +0xb24 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows() executor/aggregate.go:1382 +0x124 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1361 +0xad1 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk() executor/aggregate.go:1429 +0x276 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1366 +0xb24 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/tidb/executor.(*HashAggExec).Open() executor/aggregate.go:311 +0x3ba github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor() executor/adapter.go:1111 +0xe9 github.com/pingcap/tidb/executor.(*ExecStmt).Exec() executor/adapter.go:494 +0x994 github.com/pingcap/tidb/session.runStmt() session/session.go:2333 +0x761 github.com/pingcap/tidb/session.(*session).ExecuteStmt() session/session.go:2197 +0x1025 github.com/pingcap/tidb/testkit.(*TestKit).ExecWithContext() testkit/testkit.go:296 +0x7c7 github.com/pingcap/tidb/testkit.(*TestKit).Exec() testkit/testkit.go:270 +0x866 github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1568 +0x829 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/tidb/executor.(*HashAggExec).Open() executor/aggregate.go:311 +0x3ba github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor() executor/adapter.go:1111 +0xe9 github.com/pingcap/tidb/executor.(*ExecStmt).Exec() executor/adapter.go:494 +0x994 github.com/pingcap/tidb/session.runStmt() session/session.go:2333 +0x761 github.com/pingcap/tidb/session.(*session).ExecuteStmt() session/session.go:2197 +0x1025 github.com/pingcap/tidb/testkit.(*TestKit).ExecWithContext() testkit/testkit.go:296 +0x7c7 github.com/pingcap/tidb/testkit.(*TestKit).Exec() testkit/testkit.go:270 +0x866 github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1568 +0x829 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/tidb/executor.(*HashAggExec).Open() executor/aggregate.go:311 +0x3ba github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor() executor/adapter.go:1111 +0xe9 github.com/pingcap/tidb/executor.(*ExecStmt).Exec() executor/adapter.go:494 +0x994 github.com/pingcap/tidb/session.runStmt() session/session.go:2333 +0x761 github.com/pingcap/tidb/session.(*session).ExecuteStmt() session/session.go:2197 +0x1025 github.com/pingcap/tidb/testkit.(*TestKit).ExecWithContext() testkit/testkit.go:296 +0x7c7 github.com/pingcap/tidb/testkit.(*TestKit).Exec() testkit/testkit.go:270 +0x866 github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1568 +0x829 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1004 +0x269 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/tidb/executor.(*HashAggExec).Open() executor/aggregate.go:314 +0x3f1 github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor() executor/adapter.go:1111 +0xe9 github.com/pingcap/tidb/executor.(*ExecStmt).Exec() executor/adapter.go:494 +0x994 github.com/pingcap/tidb/session.runStmt() session/session.go:2333 +0x761 github.com/pingcap/tidb/session.(*session).ExecuteStmt() session/session.go:2197 +0x1025 Previous read at 0x00c02a98e450 by goroutine 716207: github.com/pingcap/tidb/util/chunk.(*Chunk).NumRows() util/chunk/chunk.go:352 +0x552 github.com/pingcap/tidb/executor.(*HashJoinExec).fetchBuildSideRows() executor/join.go:308 +0x4d8 github.com/pingcap/tidb/executor.(*HashJoinExec).fetchAndBuildHashTable.func2() executor/join.go:1170 +0xea github.com/pingcap/tidb/util.WithRecovery() util/misc.go:96 +0x6d github.com/pingcap/tidb/executor.(*HashJoinExec).fetchAndBuildHashTable.func4() executor/join.go:1167 +0x47 Goroutine 702063 (running) created at: testing.(*T).Run() GOROOT/src/testing/testing.go:1493 +0x75d testing.runTests.func1() GOROOT/src/testing/testing.go:1846 +0x99 testing.tRunner() GOROOT/src/testing/testing.go:1446 +0x216 testing.runTests() GOROOT/src/testing/testing.go:1844 +0x7ec testing.(*M).Run() GOROOT/src/testing/testing.go:1726 +0xa84 github.com/pingcap/tidb/testkit/testmain.(*testingM).Run() testkit/testmain/wrapper.go:27 +0x42 go.uber.org/goleak.VerifyTestMain() external/org_uber_go_goleak/testmain.go:53 +0x70 github.com/pingcap/tidb/executor_test.TestMain() executor/main_test.go:70 +0xbc6 main.main() bazel-out/k8-fastbuild/bin/executor/executor_test_/testmain.go:2288 +0x5e8 Goroutine 716207 (finished) created at: github.com/pingcap/tidb/executor.(*HashJoinExec).fetchAndBuildHashTable() executor/join.go:1167 +0x3d6 github.com/pingcap/tidb/executor.(*HashJoinExec).Next.func1() executor/join.go:1126 +0xbc github.com/pingcap/tidb/util.WithRecovery() util/misc.go:96 +0x6d github.com/pingcap/tidb/executor.(*HashJoinExec).Next.func2() executor/join.go:1124 +0x47 ================== ``` <!-- a step by step guide for reproducing the bug. --> ### 2. What did you expect to see? (Required) ### 3. What did you see instead (Required) ### 4. What is your TiDB version? (Required) <!-- Paste the output of SELECT tidb_version() -->
1.0
DATA RACE in the (*Chunk).resetForReuse() - ## Bug Report Please answer these questions before submitting your issue. Thanks! ### 1. Minimal reproduce step (Required) ``` ================== WARNING: DATA RACE Write at 0x00c02a98e450 by goroutine 702063: github.com/pingcap/tidb/util/chunk.(*Chunk).resetForReuse() util/chunk/chunk.go:137 +0x7b8 github.com/pingcap/tidb/util/chunk.(*allocator).Reset() util/chunk/alloc.go:123 +0x12a github.com/pingcap/tidb/testkit.(*TestKit).MustExec.func1() testkit/testkit.go:120 +0xde runtime.deferreturn() GOROOT/src/runtime/panic.go:476 +0x32 github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1561 +0x6aa github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute() executor/projection.go:223 +0x651 github.com/pingcap/tidb/executor.(*ProjectionExec).Next() executor/projection.go:181 +0xcc github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute() executor/projection.go:223 +0x651 github.com/pingcap/tidb/executor.(*ProjectionExec).Next() executor/projection.go:181 +0xcc github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute() executor/projection.go:223 +0x651 github.com/pingcap/tidb/executor.(*ProjectionExec).Next() executor/projection.go:181 +0xcc github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute() executor/projection.go:223 +0x651 github.com/pingcap/tidb/executor.(*ProjectionExec).Next() executor/projection.go:181 +0xcc github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows() executor/aggregate.go:1382 +0x124 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1361 +0xad1 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk() executor/aggregate.go:1429 +0x276 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1366 +0xb24 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows() executor/aggregate.go:1382 +0x124 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1361 +0xad1 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk() executor/aggregate.go:1429 +0x276 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1366 +0xb24 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows() executor/aggregate.go:1382 +0x124 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1361 +0xad1 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows() executor/aggregate.go:1382 +0x124 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1361 +0xad1 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows() executor/aggregate.go:1382 +0x124 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeCurGroupRowsAndFetchChild() executor/aggregate.go:1390 +0x64 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1343 +0x70c github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk() executor/aggregate.go:1429 +0x276 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1366 +0xb24 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows() executor/aggregate.go:1382 +0x124 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1361 +0xad1 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk() executor/aggregate.go:1429 +0x276 github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup() executor/aggregate.go:1366 +0xb24 github.com/pingcap/tidb/executor.(*StreamAggExec).Next() executor/aggregate.go:1313 +0x215 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/tidb/executor.(*HashAggExec).Open() executor/aggregate.go:311 +0x3ba github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor() executor/adapter.go:1111 +0xe9 github.com/pingcap/tidb/executor.(*ExecStmt).Exec() executor/adapter.go:494 +0x994 github.com/pingcap/tidb/session.runStmt() session/session.go:2333 +0x761 github.com/pingcap/tidb/session.(*session).ExecuteStmt() session/session.go:2197 +0x1025 github.com/pingcap/tidb/testkit.(*TestKit).ExecWithContext() testkit/testkit.go:296 +0x7c7 github.com/pingcap/tidb/testkit.(*TestKit).Exec() testkit/testkit.go:270 +0x866 github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1568 +0x829 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/tidb/executor.(*HashAggExec).Open() executor/aggregate.go:311 +0x3ba github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor() executor/adapter.go:1111 +0xe9 github.com/pingcap/tidb/executor.(*ExecStmt).Exec() executor/adapter.go:494 +0x994 github.com/pingcap/tidb/session.runStmt() session/session.go:2333 +0x761 github.com/pingcap/tidb/session.(*session).ExecuteStmt() session/session.go:2197 +0x1025 github.com/pingcap/tidb/testkit.(*TestKit).ExecWithContext() testkit/testkit.go:296 +0x7c7 github.com/pingcap/tidb/testkit.(*TestKit).Exec() testkit/testkit.go:270 +0x866 github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1568 +0x829 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/tidb/executor.(*HashAggExec).Open() executor/aggregate.go:311 +0x3ba github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor() executor/adapter.go:1111 +0xe9 github.com/pingcap/tidb/executor.(*ExecStmt).Exec() executor/adapter.go:494 +0x994 github.com/pingcap/tidb/session.runStmt() session/session.go:2333 +0x761 github.com/pingcap/tidb/session.(*session).ExecuteStmt() session/session.go:2197 +0x1025 github.com/pingcap/tidb/testkit.(*TestKit).ExecWithContext() testkit/testkit.go:296 +0x7c7 github.com/pingcap/tidb/testkit.(*TestKit).Exec() testkit/testkit.go:270 +0x866 github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1568 +0x829 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1004 +0x269 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/failpoint.Eval() external/com_github_pingcap_failpoint/failpoints.go:271 +0x44 github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults() executor/aggregate.go:1120 +0x1de github.com/pingcap/tidb/executor.(*HashAggExec).execute() executor/aggregate.go:1038 +0xee9 github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec() executor/aggregate.go:963 +0x17a github.com/pingcap/tidb/executor.(*HashAggExec).Next() executor/aggregate.go:782 +0x106 github.com/pingcap/tidb/executor.Next() executor/executor.go:325 +0x5c3 github.com/pingcap/tidb/executor.(*ExecStmt).next() executor/adapter.go:1118 +0x89 github.com/pingcap/tidb/executor.(*recordSet).Next() executor/adapter.go:153 +0x164 github.com/pingcap/tidb/session.(*execStmtResult).Next() <autogenerated>:1 +0x76 github.com/pingcap/tidb/session.GetRows4Test() session/tidb.go:336 +0xea github.com/pingcap/tidb/executor_test.TestRandomPanicConsume() executor/aggregate_test.go:1570 +0x904 github.com/pingcap/tidb/executor.(*HashAggExec).Open() executor/aggregate.go:314 +0x3f1 github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor() executor/adapter.go:1111 +0xe9 github.com/pingcap/tidb/executor.(*ExecStmt).Exec() executor/adapter.go:494 +0x994 github.com/pingcap/tidb/session.runStmt() session/session.go:2333 +0x761 github.com/pingcap/tidb/session.(*session).ExecuteStmt() session/session.go:2197 +0x1025 Previous read at 0x00c02a98e450 by goroutine 716207: github.com/pingcap/tidb/util/chunk.(*Chunk).NumRows() util/chunk/chunk.go:352 +0x552 github.com/pingcap/tidb/executor.(*HashJoinExec).fetchBuildSideRows() executor/join.go:308 +0x4d8 github.com/pingcap/tidb/executor.(*HashJoinExec).fetchAndBuildHashTable.func2() executor/join.go:1170 +0xea github.com/pingcap/tidb/util.WithRecovery() util/misc.go:96 +0x6d github.com/pingcap/tidb/executor.(*HashJoinExec).fetchAndBuildHashTable.func4() executor/join.go:1167 +0x47 Goroutine 702063 (running) created at: testing.(*T).Run() GOROOT/src/testing/testing.go:1493 +0x75d testing.runTests.func1() GOROOT/src/testing/testing.go:1846 +0x99 testing.tRunner() GOROOT/src/testing/testing.go:1446 +0x216 testing.runTests() GOROOT/src/testing/testing.go:1844 +0x7ec testing.(*M).Run() GOROOT/src/testing/testing.go:1726 +0xa84 github.com/pingcap/tidb/testkit/testmain.(*testingM).Run() testkit/testmain/wrapper.go:27 +0x42 go.uber.org/goleak.VerifyTestMain() external/org_uber_go_goleak/testmain.go:53 +0x70 github.com/pingcap/tidb/executor_test.TestMain() executor/main_test.go:70 +0xbc6 main.main() bazel-out/k8-fastbuild/bin/executor/executor_test_/testmain.go:2288 +0x5e8 Goroutine 716207 (finished) created at: github.com/pingcap/tidb/executor.(*HashJoinExec).fetchAndBuildHashTable() executor/join.go:1167 +0x3d6 github.com/pingcap/tidb/executor.(*HashJoinExec).Next.func1() executor/join.go:1126 +0xbc github.com/pingcap/tidb/util.WithRecovery() util/misc.go:96 +0x6d github.com/pingcap/tidb/executor.(*HashJoinExec).Next.func2() executor/join.go:1124 +0x47 ================== ``` <!-- a step by step guide for reproducing the bug. --> ### 2. What did you expect to see? (Required) ### 3. What did you see instead (Required) ### 4. What is your TiDB version? (Required) <!-- Paste the output of SELECT tidb_version() -->
test
data race in the chunk resetforreuse bug report please answer these questions before submitting your issue thanks minimal reproduce step required warning data race write at by goroutine github com pingcap tidb util chunk chunk resetforreuse util chunk chunk go github com pingcap tidb util chunk allocator reset util chunk alloc go github com pingcap tidb testkit testkit mustexec testkit testkit go runtime deferreturn goroot src runtime panic go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor projectionexec parallelexecute executor projection go github com pingcap tidb executor projectionexec next executor projection go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor projectionexec parallelexecute executor projection go github com pingcap tidb executor projectionexec next executor projection go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor projectionexec parallelexecute executor projection go github com pingcap tidb executor projectionexec next executor projection go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor projectionexec parallelexecute executor projection go github com pingcap tidb executor projectionexec next executor projection go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec consumegrouprows executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec consumegrouprows executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec consumegrouprows executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec consumegrouprows executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec consumegrouprows executor aggregate go github com pingcap tidb executor streamaggexec consumecurgrouprowsandfetchchild executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec consumegrouprows executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap tidb executor hashaggexec open executor aggregate go github com pingcap tidb executor execstmt openexecutor executor adapter go github com pingcap tidb executor execstmt exec executor adapter go github com pingcap tidb session runstmt session session go github com pingcap tidb session session executestmt session session go github com pingcap tidb testkit testkit execwithcontext testkit testkit go github com pingcap tidb testkit testkit exec testkit testkit go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap tidb executor hashaggexec open executor aggregate go github com pingcap tidb executor execstmt openexecutor executor adapter go github com pingcap tidb executor execstmt exec executor adapter go github com pingcap tidb session runstmt session session go github com pingcap tidb session session executestmt session session go github com pingcap tidb testkit testkit execwithcontext testkit testkit go github com pingcap tidb testkit testkit exec testkit testkit go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap tidb executor hashaggexec open executor aggregate go github com pingcap tidb executor execstmt openexecutor executor adapter go github com pingcap tidb executor execstmt exec executor adapter go github com pingcap tidb session runstmt session session go github com pingcap tidb session session executestmt session session go github com pingcap tidb testkit testkit execwithcontext testkit testkit go github com pingcap tidb testkit testkit exec testkit testkit go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap tidb executor hashaggexec open executor aggregate go github com pingcap tidb executor execstmt openexecutor executor adapter go github com pingcap tidb executor execstmt exec executor adapter go github com pingcap tidb session runstmt session session go github com pingcap tidb session session executestmt session session go previous read at by goroutine github com pingcap tidb util chunk chunk numrows util chunk chunk go github com pingcap tidb executor hashjoinexec fetchbuildsiderows executor join go github com pingcap tidb executor hashjoinexec fetchandbuildhashtable executor join go github com pingcap tidb util withrecovery util misc go github com pingcap tidb executor hashjoinexec fetchandbuildhashtable executor join go goroutine running created at testing t run goroot src testing testing go testing runtests goroot src testing testing go testing trunner goroot src testing testing go testing runtests goroot src testing testing go testing m run goroot src testing testing go github com pingcap tidb testkit testmain testingm run testkit testmain wrapper go go uber org goleak verifytestmain external org uber go goleak testmain go github com pingcap tidb executor test testmain executor main test go main main bazel out fastbuild bin executor executor test testmain go goroutine finished created at github com pingcap tidb executor hashjoinexec fetchandbuildhashtable executor join go github com pingcap tidb executor hashjoinexec next executor join go github com pingcap tidb util withrecovery util misc go github com pingcap tidb executor hashjoinexec next executor join go what did you expect to see required what did you see instead required what is your tidb version required
1
90,531
11,415,000,646
IssuesEvent
2020-02-02 07:52:41
microsoft/ApplicationInspector
https://api.github.com/repos/microsoft/ApplicationInspector
closed
Report Show Skipped Code as Analyzed
working as designed
**Describe the bug** The final report output.html show analyzed 0.99% but detects several kinds of code, I think the result is wrong and this could be the skipped %. **Desktop (please complete the following information):** - OS: Windows 10 - Linux - Browser FireFox <img width="271" alt="Captura de Pantalla 2020-01-27 a la(s) 08 55 57" src="https://user-images.githubusercontent.com/60227319/73173091-4eadfa80-40e3-11ea-98df-08319f96e7e0.png"> <img width="834" alt="Captura de Pantalla 2020-01-27 a la(s) 08 58 09" src="https://user-images.githubusercontent.com/60227319/73173092-4eadfa80-40e3-11ea-97f6-84b6f99bb62b.png">
1.0
Report Show Skipped Code as Analyzed - **Describe the bug** The final report output.html show analyzed 0.99% but detects several kinds of code, I think the result is wrong and this could be the skipped %. **Desktop (please complete the following information):** - OS: Windows 10 - Linux - Browser FireFox <img width="271" alt="Captura de Pantalla 2020-01-27 a la(s) 08 55 57" src="https://user-images.githubusercontent.com/60227319/73173091-4eadfa80-40e3-11ea-98df-08319f96e7e0.png"> <img width="834" alt="Captura de Pantalla 2020-01-27 a la(s) 08 58 09" src="https://user-images.githubusercontent.com/60227319/73173092-4eadfa80-40e3-11ea-97f6-84b6f99bb62b.png">
non_test
report show skipped code as analyzed describe the bug the final report output html show analyzed but detects several kinds of code i think the result is wrong and this could be the skipped desktop please complete the following information os windows linux browser firefox img width alt captura de pantalla a la s src img width alt captura de pantalla a la s src
0
157,338
12,370,789,574
IssuesEvent
2020-05-18 17:24:22
hashgraph/hedera-wallet-ios
https://api.github.com/repos/hashgraph/hedera-wallet-ios
closed
Wrap hedera-cli CryptoGetAccountBalance
testing
Write a wrapper script for hedera-cli. The first command should be to get_account_balance and should return the balance in a script-friendly form.
1.0
Wrap hedera-cli CryptoGetAccountBalance - Write a wrapper script for hedera-cli. The first command should be to get_account_balance and should return the balance in a script-friendly form.
test
wrap hedera cli cryptogetaccountbalance write a wrapper script for hedera cli the first command should be to get account balance and should return the balance in a script friendly form
1
294,736
25,399,499,216
IssuesEvent
2022-11-22 11:00:02
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
sql/colexec: TestMergeJoinerMultiBatchRuns failed
C-test-failure O-robot branch-master
sql/colexec.TestMergeJoinerMultiBatchRuns [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7641761?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7641761?buildTab=artifacts#/) on master @ [dca415eddac0d659ae6d76b4e3dfdf4076adbd34](https://github.com/cockroachdb/cockroach/commits/dca415eddac0d659ae6d76b4e3dfdf4076adbd34): Fatal error: ``` panic: test timed out after 14m55s ``` Stack: ``` goroutine 36988 [running]: testing.(*M).startAlarm.func1() GOROOT/src/testing/testing.go:2036 +0xbb created by time.goFunc GOROOT/src/time/sleep.go:176 +0x48 ``` <details><summary>Log preceding fatal error</summary> <p> ``` === RUN TestMergeJoinerMultiBatchRuns test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/1727f600d839fa94e6186075a07a436e/logTestMergeJoinerMultiBatchRuns2968884249 test_log_scope.go:79: use -show-logs to present logs inline === RUN TestMergeJoinerMultiBatchRuns/groupSize=511/numInputBatches=2 === RUN TestMergeJoinerMultiBatchRuns/groupSize=1023/numInputBatches=1 === RUN TestMergeJoinerMultiBatchRuns/groupSize=2046/numInputBatches=1 === RUN TestMergeJoinerMultiBatchRuns/groupSize=2046/numInputBatches=16 ``` </p> </details> <p>Parameters: <code>TAGS=bazel,gss</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestMergeJoinerMultiBatchRuns.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
sql/colexec: TestMergeJoinerMultiBatchRuns failed - sql/colexec.TestMergeJoinerMultiBatchRuns [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7641761?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7641761?buildTab=artifacts#/) on master @ [dca415eddac0d659ae6d76b4e3dfdf4076adbd34](https://github.com/cockroachdb/cockroach/commits/dca415eddac0d659ae6d76b4e3dfdf4076adbd34): Fatal error: ``` panic: test timed out after 14m55s ``` Stack: ``` goroutine 36988 [running]: testing.(*M).startAlarm.func1() GOROOT/src/testing/testing.go:2036 +0xbb created by time.goFunc GOROOT/src/time/sleep.go:176 +0x48 ``` <details><summary>Log preceding fatal error</summary> <p> ``` === RUN TestMergeJoinerMultiBatchRuns test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/1727f600d839fa94e6186075a07a436e/logTestMergeJoinerMultiBatchRuns2968884249 test_log_scope.go:79: use -show-logs to present logs inline === RUN TestMergeJoinerMultiBatchRuns/groupSize=511/numInputBatches=2 === RUN TestMergeJoinerMultiBatchRuns/groupSize=1023/numInputBatches=1 === RUN TestMergeJoinerMultiBatchRuns/groupSize=2046/numInputBatches=1 === RUN TestMergeJoinerMultiBatchRuns/groupSize=2046/numInputBatches=16 ``` </p> </details> <p>Parameters: <code>TAGS=bazel,gss</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestMergeJoinerMultiBatchRuns.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
sql colexec testmergejoinermultibatchruns failed sql colexec testmergejoinermultibatchruns with on master fatal error panic test timed out after stack goroutine testing m startalarm goroot src testing testing go created by time gofunc goroot src time sleep go log preceding fatal error run testmergejoinermultibatchruns test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline run testmergejoinermultibatchruns groupsize numinputbatches run testmergejoinermultibatchruns groupsize numinputbatches run testmergejoinermultibatchruns groupsize numinputbatches run testmergejoinermultibatchruns groupsize numinputbatches parameters tags bazel gss help see also cc cockroachdb sql queries
1
340,146
30,495,814,143
IssuesEvent
2023-07-18 10:43:55
MohistMC/Mohist
https://api.github.com/repos/MohistMC/Mohist
reopened
[1.20.1]在服务器内合成物品时有小概率崩溃,且存档会受到部分回档
Wait Needs Testing
<!-- ISSUE_TEMPLATE_3 -> IMPORTANT: DO NOT DELETE THIS LINE.--> <!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.--> <!-- If you don't know where to upload your logs and crash reports, you can use these websites : --> <!-- https://gist.github.com (recommended) --> <!-- https://mclo.gs --> <!-- https://haste.mohistmc.com --> <!-- https://pastebin.com --> <!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT --> **Minecraft Version :** {1.20.1} **Mohist Version :** {177} **Operating System :** {windows 10 pro} **Logs :** {> # # A fatal error has been detected by the Java Runtime Environment: # # Internal Error (macroAssembler_x86.cpp:864), pid=7428, tid=10556 # fatal error: DEBUG MESSAGE: duplicated predicate failed which is impossible # # JRE version: Java(TM) SE Runtime Environment (17.0.7+8) (build 17.0.7+8-LTS-224) # Java VM: Java HotSpot(TM) 64-Bit Server VM (17.0.7+8-LTS-224, mixed mode, sharing, tiered, compressed class ptrs, g1 gc, windows-amd64) # No core dump will be written. Minidumps are not enabled by default on client versions of Windows # # An error report file with more information is saved as: # D:\MC Server 1.20.1\hs_err_pid7428.log # # If you would like to submit a bug report, please visit: # https://bugreport.java.com/bugreport/crash.jsp # 服务器已关闭,将于20秒后重启} **Mod list :** {jei-1.20.1-forge-15.1.0.19,create-1.20.1-0.5.1.d,curios-forge-5.2.0-beta.3+1.20.1,Jade-1.20-forge-11.1.4} **Description of issue :** {在服务器内合成物品时有小概率崩溃,且存档会受到部分回档}
1.0
[1.20.1]在服务器内合成物品时有小概率崩溃,且存档会受到部分回档 - <!-- ISSUE_TEMPLATE_3 -> IMPORTANT: DO NOT DELETE THIS LINE.--> <!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.--> <!-- If you don't know where to upload your logs and crash reports, you can use these websites : --> <!-- https://gist.github.com (recommended) --> <!-- https://mclo.gs --> <!-- https://haste.mohistmc.com --> <!-- https://pastebin.com --> <!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT --> **Minecraft Version :** {1.20.1} **Mohist Version :** {177} **Operating System :** {windows 10 pro} **Logs :** {> # # A fatal error has been detected by the Java Runtime Environment: # # Internal Error (macroAssembler_x86.cpp:864), pid=7428, tid=10556 # fatal error: DEBUG MESSAGE: duplicated predicate failed which is impossible # # JRE version: Java(TM) SE Runtime Environment (17.0.7+8) (build 17.0.7+8-LTS-224) # Java VM: Java HotSpot(TM) 64-Bit Server VM (17.0.7+8-LTS-224, mixed mode, sharing, tiered, compressed class ptrs, g1 gc, windows-amd64) # No core dump will be written. Minidumps are not enabled by default on client versions of Windows # # An error report file with more information is saved as: # D:\MC Server 1.20.1\hs_err_pid7428.log # # If you would like to submit a bug report, please visit: # https://bugreport.java.com/bugreport/crash.jsp # 服务器已关闭,将于20秒后重启} **Mod list :** {jei-1.20.1-forge-15.1.0.19,create-1.20.1-0.5.1.d,curios-forge-5.2.0-beta.3+1.20.1,Jade-1.20-forge-11.1.4} **Description of issue :** {在服务器内合成物品时有小概率崩溃,且存档会受到部分回档}
test
在服务器内合成物品时有小概率崩溃,且存档会受到部分回档 important do not delete this line minecraft version mohist version operating system windows pro logs a fatal error has been detected by the java runtime environment internal error macroassembler cpp pid tid fatal error debug message duplicated predicate failed which is impossible jre version java tm se runtime environment build lts java vm java hotspot tm bit server vm lts mixed mode sharing tiered compressed class ptrs gc windows no core dump will be written minidumps are not enabled by default on client versions of windows an error report file with more information is saved as d mc server hs err log if you would like to submit a bug report please visit 服务器已关闭, mod list jei forge ,create d,curios forge beta ,jade forge description of issue 在服务器内合成物品时有小概率崩溃,且存档会受到部分回档
1
349,005
31,766,072,644
IssuesEvent
2023-09-12 08:51:33
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Add a test restore app class
area/test kind/task
**Description** > **Note** > Blocked by #13966 Based off of #13966, add a new class which allows launching a test instance of `RestoreApp`. This class should behave like any Spring application, with the added capability of specifying the `backupId` command line argument, and specifying the `RESTORE` profile. See the hack day PR for an example: https://github.com/camunda/zeebe/blob/99892d55658536ed2715b0790ac2257cbf7293c6/qa/util/src/main/java/io/camunda/zeebe/qa/util/cluster/TestRestoreApp.java
1.0
Add a test restore app class - **Description** > **Note** > Blocked by #13966 Based off of #13966, add a new class which allows launching a test instance of `RestoreApp`. This class should behave like any Spring application, with the added capability of specifying the `backupId` command line argument, and specifying the `RESTORE` profile. See the hack day PR for an example: https://github.com/camunda/zeebe/blob/99892d55658536ed2715b0790ac2257cbf7293c6/qa/util/src/main/java/io/camunda/zeebe/qa/util/cluster/TestRestoreApp.java
test
add a test restore app class description note blocked by based off of add a new class which allows launching a test instance of restoreapp this class should behave like any spring application with the added capability of specifying the backupid command line argument and specifying the restore profile see the hack day pr for an example
1
127,566
10,475,271,624
IssuesEvent
2019-09-23 15:59:38
kcigeospatial/Fred_Co_Land-Management
https://api.github.com/repos/kcigeospatial/Fred_Co_Land-Management
closed
Use Permit - Awaiting Fee Payment Notification - Standard Format
Ready for Test Env. Retest
Did not get an Awaiting Fee Payment notification for the Home Occupation permit.
2.0
Use Permit - Awaiting Fee Payment Notification - Standard Format - Did not get an Awaiting Fee Payment notification for the Home Occupation permit.
test
use permit awaiting fee payment notification standard format did not get an awaiting fee payment notification for the home occupation permit
1
347,371
31,160,491,061
IssuesEvent
2023-08-16 15:39:58
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
reopened
DISABLED test_mm_sparse_first_T_cuda_bfloat16 (__main__.TestSparseSemiStructuredCUDA)
module: sparse triaged module: flaky-tests skipped
Platforms: linux, slow This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_mm_sparse_first_T_cuda_bfloat16&suite=TestSparseSemiStructuredCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/15355252942). Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_mm_sparse_first_T_cuda_bfloat16` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `test_sparse_semi_structured.py` cc @alexsamardzic @nikitaved @pearu @cpuhrsch @amjames @bhosmer
1.0
DISABLED test_mm_sparse_first_T_cuda_bfloat16 (__main__.TestSparseSemiStructuredCUDA) - Platforms: linux, slow This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_mm_sparse_first_T_cuda_bfloat16&suite=TestSparseSemiStructuredCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/15355252942). Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_mm_sparse_first_T_cuda_bfloat16` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `test_sparse_semi_structured.py` cc @alexsamardzic @nikitaved @pearu @cpuhrsch @amjames @bhosmer
test
disabled test mm sparse first t cuda main testsparsesemistructuredcuda platforms linux slow this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test mm sparse first t cuda there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path test sparse semi structured py cc alexsamardzic nikitaved pearu cpuhrsch amjames bhosmer
1
61,782
14,640,710,321
IssuesEvent
2020-12-25 03:21:45
fu1771695yongxie/pm
https://api.github.com/repos/fu1771695yongxie/pm
opened
CVE-2018-3721 (Medium) detected in lodash-3.10.1.tgz
security vulnerability
## CVE-2018-3721 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: pm/package.json</p> <p>Path to vulnerable library: pm/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - babel-core-5.8.38.tgz (Root Library) - :x: **lodash-3.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/pm/commit/1c06cbe4c354bfe6922fec380958337d18de7e44">1c06cbe4c354bfe6922fec380958337d18de7e44</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> lodash node module before 4.17.5 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via defaultsDeep, merge, and mergeWith functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3721>CVE-2018-3721</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3721">https://nvd.nist.gov/vuln/detail/CVE-2018-3721</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 4.17.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-3721 (Medium) detected in lodash-3.10.1.tgz - ## CVE-2018-3721 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: pm/package.json</p> <p>Path to vulnerable library: pm/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - babel-core-5.8.38.tgz (Root Library) - :x: **lodash-3.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/pm/commit/1c06cbe4c354bfe6922fec380958337d18de7e44">1c06cbe4c354bfe6922fec380958337d18de7e44</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> lodash node module before 4.17.5 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via defaultsDeep, merge, and mergeWith functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3721>CVE-2018-3721</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3721">https://nvd.nist.gov/vuln/detail/CVE-2018-3721</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 4.17.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file pm package json path to vulnerable library pm node modules lodash package json dependency hierarchy babel core tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details lodash node module before suffers from a modification of assumed immutable data maid vulnerability via defaultsdeep merge and mergewith functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
1,847
2,576,389,011
IssuesEvent
2015-02-12 09:42:37
w3c/csswg-test
https://api.github.com/repos/w3c/csswg-test
closed
ime-mode
spec:ui testtwf
Might be particularly relevant here in Shenzhen, since most people presumably use an IME.
1.0
ime-mode - Might be particularly relevant here in Shenzhen, since most people presumably use an IME.
test
ime mode might be particularly relevant here in shenzhen since most people presumably use an ime
1
225,379
24,828,515,171
IssuesEvent
2022-10-25 23:39:26
snowdensb/jpo-ode
https://api.github.com/repos/snowdensb/jpo-ode
reopened
CVE-2022-40156 (High) detected in woodstox-core-6.2.4.jar
security vulnerability
## CVE-2022-40156 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>woodstox-core-6.2.4.jar</b></p></summary> <p>Woodstox is a high-performance XML processor that implements Stax (JSR-173), SAX2 and Stax2 APIs</p> <p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p> <p>Path to dependency file: /jpo-ode-svcs/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar</p> <p> Dependency Hierarchy: - jackson-dataformat-xml-2.12.3.jar (Root Library) - :x: **woodstox-core-6.2.4.jar** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Those using Xstream to seralize XML data may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. This effect may support a denial of service attack. <p>Publish Date: 2022-09-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40156>CVE-2022-40156</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-09-16</p> <p>Fix Resolution: woodstox-core-5.4.0,woodstox-core-6.4.0</p> </p> </details> <p></p>
True
CVE-2022-40156 (High) detected in woodstox-core-6.2.4.jar - ## CVE-2022-40156 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>woodstox-core-6.2.4.jar</b></p></summary> <p>Woodstox is a high-performance XML processor that implements Stax (JSR-173), SAX2 and Stax2 APIs</p> <p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p> <p>Path to dependency file: /jpo-ode-svcs/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar</p> <p> Dependency Hierarchy: - jackson-dataformat-xml-2.12.3.jar (Root Library) - :x: **woodstox-core-6.2.4.jar** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Those using Xstream to seralize XML data may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. This effect may support a denial of service attack. <p>Publish Date: 2022-09-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40156>CVE-2022-40156</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-09-16</p> <p>Fix Resolution: woodstox-core-5.4.0,woodstox-core-6.4.0</p> </p> </details> <p></p>
non_test
cve high detected in woodstox core jar cve high severity vulnerability vulnerable library woodstox core jar woodstox is a high performance xml processor that implements stax jsr and apis library home page a href path to dependency file jpo ode svcs pom xml path to vulnerable library home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar dependency hierarchy jackson dataformat xml jar root library x woodstox core jar vulnerable library found in base branch dev vulnerability details those using xstream to seralize xml data may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow this effect may support a denial of service attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution woodstox core woodstox core
0
780,057
27,377,960,077
IssuesEvent
2023-02-28 07:56:14
ballerina-platform/ballerina-dev-website
https://api.github.com/repos/ballerina-platform/ballerina-dev-website
closed
Update `Style guide` page documentation to use our documenting practices
Priority/High Type/Task Points/0.5 Area/LearnPages Category/Content
## Description $subject. ## Related website/documentation area [1] https://ballerina.io/learn/style-guide/annotations-documentation-and-comments/#comments:~:text=%23%20%2B%20value%20%2D%20value%20input%20parameter Our practice in documentation is, - The description starts with a capital letter and ends with a period. - Parameter doc lines start with a simple letter and no period at the end. e.g. ```bal # Adds two integers. # # + x - an integer # + y - another integer # + return - the sum of `x` and `y` public function add(int x, int y) returns int { return x + y; } ``` We need to update the last example in [1] to adhere above practice.
1.0
Update `Style guide` page documentation to use our documenting practices - ## Description $subject. ## Related website/documentation area [1] https://ballerina.io/learn/style-guide/annotations-documentation-and-comments/#comments:~:text=%23%20%2B%20value%20%2D%20value%20input%20parameter Our practice in documentation is, - The description starts with a capital letter and ends with a period. - Parameter doc lines start with a simple letter and no period at the end. e.g. ```bal # Adds two integers. # # + x - an integer # + y - another integer # + return - the sum of `x` and `y` public function add(int x, int y) returns int { return x + y; } ``` We need to update the last example in [1] to adhere above practice.
non_test
update style guide page documentation to use our documenting practices description subject related website documentation area our practice in documentation is the description starts with a capital letter and ends with a period parameter doc lines start with a simple letter and no period at the end e g bal adds two integers x an integer y another integer return the sum of x and y public function add int x int y returns int return x y we need to update the last example in to adhere above practice
0
189,945
6,803,281,551
IssuesEvent
2017-11-02 23:55:50
sul-dlss/preservation_catalog
https://api.github.com/repos/sul-dlss/preservation_catalog
closed
(MV) Directories ONLY contain files
high priority ready
- content and metadata directories each contain only files. With RSpec tests.
1.0
(MV) Directories ONLY contain files - - content and metadata directories each contain only files. With RSpec tests.
non_test
mv directories only contain files content and metadata directories each contain only files with rspec tests
0
275,284
8,575,547,918
IssuesEvent
2018-11-12 17:34:56
aowen87/TicketTester
https://api.github.com/repos/aowen87/TicketTester
closed
Can we remove Sim V1 files?
Expected Use: 3 - Occasional Feature Impact: 3 - Medium Priority: Normal
I checked with Brad, and he thinks we can remove the Sim version reader/write and library files. He suggest we keep the V2 naming, in case a V3 comes along. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1910 Status: Resolved Project: VisIt Tracker: Feature Priority: Normal Subject: Can we remove Sim V1 files? Assigned to: Kathleen Biagas Category: Target version: 2.8 Author: Kathleen Biagas Start: 07/11/2014 Due date: % Done: 0 Estimated time: Created: 07/11/2014 06:41 pm Updated: 08/19/2014 08:00 pm Likelihood: Severity: Found in version: Impact: 3 - Medium Expected Use: 3 - Occasional OS: All Support Group: Any Description: I checked with Brad, and he thinks we can remove the Sim version reader/write and library files. He suggest we keep the V2 naming, in case a V3 comes along. Comments: I removed V1 from sim and removed the SimV1 reader and writer.
1.0
Can we remove Sim V1 files? - I checked with Brad, and he thinks we can remove the Sim version reader/write and library files. He suggest we keep the V2 naming, in case a V3 comes along. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1910 Status: Resolved Project: VisIt Tracker: Feature Priority: Normal Subject: Can we remove Sim V1 files? Assigned to: Kathleen Biagas Category: Target version: 2.8 Author: Kathleen Biagas Start: 07/11/2014 Due date: % Done: 0 Estimated time: Created: 07/11/2014 06:41 pm Updated: 08/19/2014 08:00 pm Likelihood: Severity: Found in version: Impact: 3 - Medium Expected Use: 3 - Occasional OS: All Support Group: Any Description: I checked with Brad, and he thinks we can remove the Sim version reader/write and library files. He suggest we keep the V2 naming, in case a V3 comes along. Comments: I removed V1 from sim and removed the SimV1 reader and writer.
non_test
can we remove sim files i checked with brad and he thinks we can remove the sim version reader write and library files he suggest we keep the naming in case a comes along redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker feature priority normal subject can we remove sim files assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood severity found in version impact medium expected use occasional os all support group any description i checked with brad and he thinks we can remove the sim version reader write and library files he suggest we keep the naming in case a comes along comments i removed from sim and removed the reader and writer
0
328,596
9,997,129,421
IssuesEvent
2019-07-12 02:54:40
momentum-mod/website
https://api.github.com/repos/momentum-mod/website
closed
Allow embedding YouTube videos for maps
Enhancement Priority: Medium Size: Medium
In accordance with #282 , we need to be able to support youtube videos in the map info table to be able to actually show them. [youtubesurfvideos JSON data from BorkChops](https://github.com/momentum-mod/website/files/3364441/youtubesurfvideos.txt)
1.0
Allow embedding YouTube videos for maps - In accordance with #282 , we need to be able to support youtube videos in the map info table to be able to actually show them. [youtubesurfvideos JSON data from BorkChops](https://github.com/momentum-mod/website/files/3364441/youtubesurfvideos.txt)
non_test
allow embedding youtube videos for maps in accordance with we need to be able to support youtube videos in the map info table to be able to actually show them
0
59,862
8,381,207,251
IssuesEvent
2018-10-07 22:30:17
damienbod/angular-auth-oidc-client
https://api.github.com/repos/damienbod/angular-auth-oidc-client
closed
Refreshing the page wipes all storage data
enhancement enhancement documentation
Hi, i'm doing a POC using your library and i got this issue. So basically i'm able to login, get token, send the token using http interceptors, all work fine. My api check the token, all good! But if i refresh the page, all auth storage data is wiped! On my debug, i found out that when calling authorizedCallback, it then calls resetAuthorizationData straight away wiping the data. https://github.com/damienbod/angular-auth-oidc-client/blob/40790b8909f5218536c6f4f92d7f8329ac5b225d/src/services/oidc.security.service.ts#L194. Is this the expected behaviour? i'm doing whats in the read.me (configure on the app.module, setting up in app.component... any thoughts ? Thanks in advance
1.0
Refreshing the page wipes all storage data - Hi, i'm doing a POC using your library and i got this issue. So basically i'm able to login, get token, send the token using http interceptors, all work fine. My api check the token, all good! But if i refresh the page, all auth storage data is wiped! On my debug, i found out that when calling authorizedCallback, it then calls resetAuthorizationData straight away wiping the data. https://github.com/damienbod/angular-auth-oidc-client/blob/40790b8909f5218536c6f4f92d7f8329ac5b225d/src/services/oidc.security.service.ts#L194. Is this the expected behaviour? i'm doing whats in the read.me (configure on the app.module, setting up in app.component... any thoughts ? Thanks in advance
non_test
refreshing the page wipes all storage data hi i m doing a poc using your library and i got this issue so basically i m able to login get token send the token using http interceptors all work fine my api check the token all good but if i refresh the page all auth storage data is wiped on my debug i found out that when calling authorizedcallback it then calls resetauthorizationdata straight away wiping the data is this the expected behaviour i m doing whats in the read me configure on the app module setting up in app component any thoughts thanks in advance
0
144,562
11,624,176,032
IssuesEvent
2020-02-27 10:17:41
SPW-DIG/metawal-core-geonetwork
https://api.github.com/repos/SPW-DIG/metawal-core-geonetwork
closed
Ressource en ligne / Protocole DB & FILE
Env prod - OK Env test - OK Env valid - OK
- [x] Editeur / Ajout de choix DB et FILE:RASTER/VECTOR dans la liste de choix * https://github.com/geonetwork/core-geonetwork/pull/4383 * https://github.com/metadata101/iso19115-3.2018/pull/23 ```xml <mrd:onLine> <cit:CI_OnlineResource> <cit:linkage> <gco:CharacterString>dbora:schema</gco:CharacterString> </cit:linkage> <cit:protocol> <gco:CharacterString>DB:ORACLE</gco:CharacterString> </cit:protocol> <cit:name> <gco:CharacterString>tablename</gco:CharacterString> </cit:name> <cit:function> <cit:CI_OnLineFunctionCode codeList="http://standards.iso.org/iso/19139/resources/gmxCodelists.xml#CI_OnLineFunctionCode" codeListValue="fileAccess"/> </cit:function> </cit:CI_OnlineResource> </mrd:onLine> <mrd:onLine> <cit:CI_OnlineResource> <cit:linkage gco:nilReason="withheld"> <gco:CharacterString>/geo/data/dgo3_dada.ecw</gco:CharacterString> </cit:linkage> <cit:protocol> <gco:CharacterString>FILE:RASTER</gco:CharacterString> </cit:protocol> <cit:name> <gco:CharacterString>Données au format ECW</gco:CharacterString> </cit:name> <cit:function> <cit:CI_OnLineFunctionCode codeList="http://standards.iso.org/iso/19139/resources/gmxCodelists.xml#CI_OnLineFunctionCode" codeListValue="fileAccess"/> </cit:function> </cit:CI_OnlineResource> </mrd:onLine> ``` - [x] UFO / Pas de withheld - [x] Publish / Do not check download by default. Intranet / Check all. GN ref https://github.com/geonetwork/core-geonetwork/pull/4405 - [ ] Tester que quand il y a download pour un groupe, les infos ne sont pas affichées
1.0
Ressource en ligne / Protocole DB & FILE - - [x] Editeur / Ajout de choix DB et FILE:RASTER/VECTOR dans la liste de choix * https://github.com/geonetwork/core-geonetwork/pull/4383 * https://github.com/metadata101/iso19115-3.2018/pull/23 ```xml <mrd:onLine> <cit:CI_OnlineResource> <cit:linkage> <gco:CharacterString>dbora:schema</gco:CharacterString> </cit:linkage> <cit:protocol> <gco:CharacterString>DB:ORACLE</gco:CharacterString> </cit:protocol> <cit:name> <gco:CharacterString>tablename</gco:CharacterString> </cit:name> <cit:function> <cit:CI_OnLineFunctionCode codeList="http://standards.iso.org/iso/19139/resources/gmxCodelists.xml#CI_OnLineFunctionCode" codeListValue="fileAccess"/> </cit:function> </cit:CI_OnlineResource> </mrd:onLine> <mrd:onLine> <cit:CI_OnlineResource> <cit:linkage gco:nilReason="withheld"> <gco:CharacterString>/geo/data/dgo3_dada.ecw</gco:CharacterString> </cit:linkage> <cit:protocol> <gco:CharacterString>FILE:RASTER</gco:CharacterString> </cit:protocol> <cit:name> <gco:CharacterString>Données au format ECW</gco:CharacterString> </cit:name> <cit:function> <cit:CI_OnLineFunctionCode codeList="http://standards.iso.org/iso/19139/resources/gmxCodelists.xml#CI_OnLineFunctionCode" codeListValue="fileAccess"/> </cit:function> </cit:CI_OnlineResource> </mrd:onLine> ``` - [x] UFO / Pas de withheld - [x] Publish / Do not check download by default. Intranet / Check all. GN ref https://github.com/geonetwork/core-geonetwork/pull/4405 - [ ] Tester que quand il y a download pour un groupe, les infos ne sont pas affichées
test
ressource en ligne protocole db file editeur ajout de choix db et file raster vector dans la liste de choix xml dbora schema db oracle tablename cit ci onlinefunctioncode codelist codelistvalue fileaccess geo data dada ecw file raster données au format ecw cit ci onlinefunctioncode codelist codelistvalue fileaccess ufo pas de withheld publish do not check download by default intranet check all gn ref tester que quand il y a download pour un groupe les infos ne sont pas affichées
1
219,908
17,119,793,843
IssuesEvent
2021-07-12 02:25:46
aimakerspace/PeekingDuck
https://api.github.com/repos/aimakerspace/PeekingDuck
closed
In the unit tests, windows appends double backslash instead of forward slash in os.path.join
testing
In windows, the os.path.join use double backslash, thus this affect the assertion checks in the unit test that assumed forward slash in the output. (As seen below) Unit Tests affected: - test_declarativeloader - test_configloader - test_runner Suggestion: To do a string replacement before assert statement to change double backslash into 1 forward slash ![image](https://user-images.githubusercontent.com/61336482/124552779-8d162000-de66-11eb-9d02-8b8e65ad7f7b.png)
1.0
In the unit tests, windows appends double backslash instead of forward slash in os.path.join - In windows, the os.path.join use double backslash, thus this affect the assertion checks in the unit test that assumed forward slash in the output. (As seen below) Unit Tests affected: - test_declarativeloader - test_configloader - test_runner Suggestion: To do a string replacement before assert statement to change double backslash into 1 forward slash ![image](https://user-images.githubusercontent.com/61336482/124552779-8d162000-de66-11eb-9d02-8b8e65ad7f7b.png)
test
in the unit tests windows appends double backslash instead of forward slash in os path join in windows the os path join use double backslash thus this affect the assertion checks in the unit test that assumed forward slash in the output as seen below unit tests affected test declarativeloader test configloader test runner suggestion to do a string replacement before assert statement to change double backslash into forward slash
1
380,050
26,399,494,053
IssuesEvent
2023-01-12 23:07:05
aptos-labs/aptos-core
https://api.github.com/repos/aptos-labs/aptos-core
opened
Add documentation for running fullnodes on AWS and Azure
documentation
# Add documentation for running fullnodes on AWS and Azure <!-- A clear and concise description of the feature you are requesting --> ## Motivation We have documentation for running fullnodes on source/docker and GCP but not for AWS and Azure We should have multiple cloud provider options similar to our validator node docs ## Pitch **Describe the solution you'd like** Replicate this doc https://aptos.dev/nodes/full-node/run-a-fullnode-on-gcp/ for AWS and Azure ## Additional context Refer to validator + VFN instructions: https://aptos.dev/nodes/validator-node/operator/running-validator-node/run-validator-node-using-azure https://aptos.dev/nodes/validator-node/operator/running-validator-node/run-validator-node-using-aws Initial steps are the same but replace 'aptos-node' module with 'fullnode' module?
1.0
Add documentation for running fullnodes on AWS and Azure - # Add documentation for running fullnodes on AWS and Azure <!-- A clear and concise description of the feature you are requesting --> ## Motivation We have documentation for running fullnodes on source/docker and GCP but not for AWS and Azure We should have multiple cloud provider options similar to our validator node docs ## Pitch **Describe the solution you'd like** Replicate this doc https://aptos.dev/nodes/full-node/run-a-fullnode-on-gcp/ for AWS and Azure ## Additional context Refer to validator + VFN instructions: https://aptos.dev/nodes/validator-node/operator/running-validator-node/run-validator-node-using-azure https://aptos.dev/nodes/validator-node/operator/running-validator-node/run-validator-node-using-aws Initial steps are the same but replace 'aptos-node' module with 'fullnode' module?
non_test
add documentation for running fullnodes on aws and azure add documentation for running fullnodes on aws and azure motivation we have documentation for running fullnodes on source docker and gcp but not for aws and azure we should have multiple cloud provider options similar to our validator node docs pitch describe the solution you d like replicate this doc for aws and azure additional context refer to validator vfn instructions initial steps are the same but replace aptos node module with fullnode module
0
344,511
30,749,907,916
IssuesEvent
2023-07-28 18:13:16
saltstack/salt
https://api.github.com/repos/saltstack/salt
opened
[Increase Test Coverage] Batch 9
Tests
Increase the code coverage percent on the following files to at least 80%. File | Percent salt/modules/win_pkg.py | 36 salt/auth/__init__.py | 62 salt/payload.py | 68 salt/states/service.py | 58 salt/modules/win_file.py | 61 Please be aware that currently the percentage might be inaccurate if the module uses __salt__ due to https://github.com/saltstack/salt/issues/64696
1.0
[Increase Test Coverage] Batch 9 - Increase the code coverage percent on the following files to at least 80%. File | Percent salt/modules/win_pkg.py | 36 salt/auth/__init__.py | 62 salt/payload.py | 68 salt/states/service.py | 58 salt/modules/win_file.py | 61 Please be aware that currently the percentage might be inaccurate if the module uses __salt__ due to https://github.com/saltstack/salt/issues/64696
test
batch increase the code coverage percent on the following files to at least file percent salt modules win pkg py salt auth init py salt payload py salt states service py salt modules win file py please be aware that currently the percentage might be inaccurate if the module uses salt due to
1
226,959
18,045,975,745
IssuesEvent
2021-09-18 22:42:44
logicmoo/logicmoo_workspace
https://api.github.com/repos/logicmoo/logicmoo_workspace
opened
logicmoo.pfc.test.sanity_base.MT_03 JUnit
Test_9999 logicmoo.pfc.test.sanity_base unit_test MT_03
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif mt_03.pl) GH_MASTER_ISSUE_FINFO= ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AMT_03 GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/1629eba4a2a1da0e1b731d156198a7168dafae44 https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/68/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/1629eba4a2a1da0e1b731d156198a7168dafae44 https://github.com/logicmoo/logicmoo_workspace/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl ``` % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/mt_03.pl'), %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) %~ this_test_might_need( :-( expects_dialect(pfc))) %:- add_import_module(header_sane,baseKB,end). :- set_defaultAssertMt(myMt). %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/mt_03.pl:17 %~ pfc_iri : include_module_file(myMt:library('pfclib/system_each_module.pfc'),myMt). /*~ %~ pfc_iri:include_module_file(myMt:library('pfclib/system_each_module.pfc'),myMt) ~*/ :- expects_dialect(pfc). :- mpred_trace_exec. mtProlog(modA). No source location!? %~ message_hook_type(error) %~ message_hook( %~ error( %~ permission_error(redefine,imported_procedure,baseKB:mtProlog/1), %~ context(system:'$record_clause'/3,Context_Kw)), %~ error, %~ [ 'No permission to ~w ~w `~p\'' - [ redefine, %~ imported_procedure, %~ baseKB : mtProlog/1]]) /*~ No permission to redefine imported_procedure `baseKB:(mtProlog/1)' ERROR: No permission to redefine imported_procedure `baseKB:(mtProlog/1)' ~*/ mtProlog(modB). No source location!? %~ message_hook_type(error) %~ message_hook( %~ error( %~ permission_error(redefine,imported_procedure,baseKB:mtProlog/1), %~ context(system:'$record_clause'/3,Context_Kw)), %~ error, %~ [ 'No permission to ~w ~w `~p\'' - [ redefine, %~ imported_procedure, %~ baseKB : mtProlog/1]]) /*~ No permission to redefine imported_procedure `baseKB:(mtProlog/1)' ERROR: No permission to redefine imported_procedure `baseKB:(mtProlog/1)' ~*/ modA: (codeA:- printAll('$current_source_module'(_M)),codeB). No source location!? modB: (codeB). %:- \+ modA:codeA. No source location!? %:- \+ modA:codeA. genlMt(modA,modB). % before test, to make sure codeA was not accdently defined in modB % before test, to make sure codeA was not accdently defined in modB :- sanity(\+ module_clause(modB:codeA,_)). :- sanity(\+ module_clause(modA:codeB,_)). :- sanity( module_clause(modA:codeA,_)). :- sanity( module_clause(modB:codeB,_)). % before test, genlMt makes the rule available and should not corrupt the modA module % before test, genlMt makes the rule available and should not corrupt the modA module :- warn_fail_TODO(clause_u(modA:codeB,_)). % make sure genlMt didnt unassert %~ :-( warn_fail_TODO( clause_u(modA:codeB,Kw))). % make sure genlMt didnt unassert :- sanity(clause_u(modB:codeB,_)). % run the test % run the test modA: (:- codeA). % to make codeB sure is available in modA No source location!? % to make codeB sure is available in modA :- mpred_must( clause_u(modA:codeB,_)). % to make sure codeA does not get accdently defined in modB %~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56 %~ failed_mpred_test( clause_u(modA:codeB,Kw)) %~ FILE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56 %~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56 %~ FILE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56 %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/mt_03.pl:56 %~ DUMP_BREAK/0 %~ message_hook_type(error) %~ message_hook( initialization_exception(abort), %~ error, %~ [ 'Prolog initialisation failed:', nl,'Unknown message: ~p'-[abort]]) %~ unused(save_junit_results) ``` totalTime=3 ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AMT_03 GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/1629eba4a2a1da0e1b731d156198a7168dafae44 https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/68/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/1629eba4a2a1da0e1b731d156198a7168dafae44 https://github.com/logicmoo/logicmoo_workspace/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k mt_03.pl (returned 1)
3.0
logicmoo.pfc.test.sanity_base.MT_03 JUnit - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif mt_03.pl) GH_MASTER_ISSUE_FINFO= ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AMT_03 GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/1629eba4a2a1da0e1b731d156198a7168dafae44 https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/68/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/1629eba4a2a1da0e1b731d156198a7168dafae44 https://github.com/logicmoo/logicmoo_workspace/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl ``` % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/mt_03.pl'), %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) %~ this_test_might_need( :-( expects_dialect(pfc))) %:- add_import_module(header_sane,baseKB,end). :- set_defaultAssertMt(myMt). %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/mt_03.pl:17 %~ pfc_iri : include_module_file(myMt:library('pfclib/system_each_module.pfc'),myMt). /*~ %~ pfc_iri:include_module_file(myMt:library('pfclib/system_each_module.pfc'),myMt) ~*/ :- expects_dialect(pfc). :- mpred_trace_exec. mtProlog(modA). No source location!? %~ message_hook_type(error) %~ message_hook( %~ error( %~ permission_error(redefine,imported_procedure,baseKB:mtProlog/1), %~ context(system:'$record_clause'/3,Context_Kw)), %~ error, %~ [ 'No permission to ~w ~w `~p\'' - [ redefine, %~ imported_procedure, %~ baseKB : mtProlog/1]]) /*~ No permission to redefine imported_procedure `baseKB:(mtProlog/1)' ERROR: No permission to redefine imported_procedure `baseKB:(mtProlog/1)' ~*/ mtProlog(modB). No source location!? %~ message_hook_type(error) %~ message_hook( %~ error( %~ permission_error(redefine,imported_procedure,baseKB:mtProlog/1), %~ context(system:'$record_clause'/3,Context_Kw)), %~ error, %~ [ 'No permission to ~w ~w `~p\'' - [ redefine, %~ imported_procedure, %~ baseKB : mtProlog/1]]) /*~ No permission to redefine imported_procedure `baseKB:(mtProlog/1)' ERROR: No permission to redefine imported_procedure `baseKB:(mtProlog/1)' ~*/ modA: (codeA:- printAll('$current_source_module'(_M)),codeB). No source location!? modB: (codeB). %:- \+ modA:codeA. No source location!? %:- \+ modA:codeA. genlMt(modA,modB). % before test, to make sure codeA was not accdently defined in modB % before test, to make sure codeA was not accdently defined in modB :- sanity(\+ module_clause(modB:codeA,_)). :- sanity(\+ module_clause(modA:codeB,_)). :- sanity( module_clause(modA:codeA,_)). :- sanity( module_clause(modB:codeB,_)). % before test, genlMt makes the rule available and should not corrupt the modA module % before test, genlMt makes the rule available and should not corrupt the modA module :- warn_fail_TODO(clause_u(modA:codeB,_)). % make sure genlMt didnt unassert %~ :-( warn_fail_TODO( clause_u(modA:codeB,Kw))). % make sure genlMt didnt unassert :- sanity(clause_u(modB:codeB,_)). % run the test % run the test modA: (:- codeA). % to make codeB sure is available in modA No source location!? % to make codeB sure is available in modA :- mpred_must( clause_u(modA:codeB,_)). % to make sure codeA does not get accdently defined in modB %~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56 %~ failed_mpred_test( clause_u(modA:codeB,Kw)) %~ FILE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56 %~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56 %~ FILE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56 %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/mt_03.pl:56 %~ DUMP_BREAK/0 %~ message_hook_type(error) %~ message_hook( initialization_exception(abort), %~ error, %~ [ 'Prolog initialisation failed:', nl,'Unknown message: ~p'-[abort]]) %~ unused(save_junit_results) ``` totalTime=3 ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AMT_03 GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/1629eba4a2a1da0e1b731d156198a7168dafae44 https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/68/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/1629eba4a2a1da0e1b731d156198a7168dafae44 https://github.com/logicmoo/logicmoo_workspace/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k mt_03.pl (returned 1)
test
logicmoo pfc test sanity base mt junit cd var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base timeout foreground preserve status s sigkill k lmoo clif mt pl gh master issue finfo issue search gitlab latest this build github running var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base mt pl this test might need use module library logicmoo plarkc this test might need expects dialect pfc add import module header sane basekb end set defaultassertmt mymt var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base mt pl pfc iri include module file mymt library pfclib system each module pfc mymt pfc iri include module file mymt library pfclib system each module pfc mymt expects dialect pfc mpred trace exec mtprolog moda no source location message hook type error message hook error permission error redefine imported procedure basekb mtprolog context system record clause context kw error no permission to w w p redefine imported procedure basekb mtprolog no permission to redefine imported procedure basekb mtprolog error no permission to redefine imported procedure basekb mtprolog mtprolog modb no source location message hook type error message hook error permission error redefine imported procedure basekb mtprolog context system record clause context kw error no permission to w w p redefine imported procedure basekb mtprolog no permission to redefine imported procedure basekb mtprolog error no permission to redefine imported procedure basekb mtprolog moda codea printall current source module m codeb no source location modb codeb moda codea no source location moda codea genlmt moda modb before test to make sure codea was not accdently defined in modb before test to make sure codea was not accdently defined in modb sanity module clause modb codea sanity module clause moda codeb sanity module clause moda codea sanity module clause modb codeb before test genlmt makes the rule available and should not corrupt the moda module before test genlmt makes the rule available and should not corrupt the moda module warn fail todo clause u moda codeb make sure genlmt didnt unassert warn fail todo clause u moda codeb kw make sure genlmt didnt unassert sanity clause u modb codeb run the test run the test moda codea to make codeb sure is available in moda no source location to make codeb sure is available in moda mpred must clause u moda codeb to make sure codea does not get accdently defined in modb file failed mpred test clause u moda codeb kw file file file var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base mt pl dump break message hook type error message hook initialization exception abort error unused save junit results totaltime issue search gitlab latest this build github failed var lib jenkins workspace logicmoo workspace bin lmoo junit minor k mt pl returned
1
124,621
17,772,676,455
IssuesEvent
2021-08-30 15:18:44
kapseliboi/html2canvas
https://api.github.com/repos/kapseliboi/html2canvas
opened
CVE-2020-15168 (Medium) detected in node-fetch-1.7.3.tgz
security vulnerability
## CVE-2020-15168 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-fetch-1.7.3.tgz</b></p></summary> <p>A light-weight module that brings window.fetch to node.js and io.js</p> <p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-1.7.3.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-1.7.3.tgz</a></p> <p>Path to dependency file: html2canvas/www/package.json</p> <p>Path to vulnerable library: html2canvas/www/node_modules/node-fetch/package.json</p> <p> Dependency Hierarchy: - glamor-2.20.40.tgz (Root Library) - fbjs-0.8.17.tgz - isomorphic-fetch-2.2.1.tgz - :x: **node-fetch-1.7.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/html2canvas/commit/0ae2bdc652fe2e15c2adc0e9e9d841a564f7053d">0ae2bdc652fe2e15c2adc0e9e9d841a564f7053d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> node-fetch before versions 2.6.1 and 3.0.0-beta.9 did not honor the size option after following a redirect, which means that when a content size was over the limit, a FetchError would never get thrown and the process would end without failure. For most people, this fix will have a little or no impact. However, if you are relying on node-fetch to gate files above a size, the impact could be significant, for example: If you don't double-check the size of the data after fetch() has completed, your JS thread could get tied up doing work on a large file (DoS) and/or cost you money in computing. <p>Publish Date: 2020-09-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15168>CVE-2020-15168</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/node-fetch/node-fetch/security/advisories/GHSA-w7rc-rwvf-8q5r">https://github.com/node-fetch/node-fetch/security/advisories/GHSA-w7rc-rwvf-8q5r</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 2.6.1,3.0.0-beta.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-15168 (Medium) detected in node-fetch-1.7.3.tgz - ## CVE-2020-15168 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-fetch-1.7.3.tgz</b></p></summary> <p>A light-weight module that brings window.fetch to node.js and io.js</p> <p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-1.7.3.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-1.7.3.tgz</a></p> <p>Path to dependency file: html2canvas/www/package.json</p> <p>Path to vulnerable library: html2canvas/www/node_modules/node-fetch/package.json</p> <p> Dependency Hierarchy: - glamor-2.20.40.tgz (Root Library) - fbjs-0.8.17.tgz - isomorphic-fetch-2.2.1.tgz - :x: **node-fetch-1.7.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/html2canvas/commit/0ae2bdc652fe2e15c2adc0e9e9d841a564f7053d">0ae2bdc652fe2e15c2adc0e9e9d841a564f7053d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> node-fetch before versions 2.6.1 and 3.0.0-beta.9 did not honor the size option after following a redirect, which means that when a content size was over the limit, a FetchError would never get thrown and the process would end without failure. For most people, this fix will have a little or no impact. However, if you are relying on node-fetch to gate files above a size, the impact could be significant, for example: If you don't double-check the size of the data after fetch() has completed, your JS thread could get tied up doing work on a large file (DoS) and/or cost you money in computing. <p>Publish Date: 2020-09-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15168>CVE-2020-15168</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/node-fetch/node-fetch/security/advisories/GHSA-w7rc-rwvf-8q5r">https://github.com/node-fetch/node-fetch/security/advisories/GHSA-w7rc-rwvf-8q5r</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 2.6.1,3.0.0-beta.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in node fetch tgz cve medium severity vulnerability vulnerable library node fetch tgz a light weight module that brings window fetch to node js and io js library home page a href path to dependency file www package json path to vulnerable library www node modules node fetch package json dependency hierarchy glamor tgz root library fbjs tgz isomorphic fetch tgz x node fetch tgz vulnerable library found in head commit a href found in base branch master vulnerability details node fetch before versions and beta did not honor the size option after following a redirect which means that when a content size was over the limit a fetcherror would never get thrown and the process would end without failure for most people this fix will have a little or no impact however if you are relying on node fetch to gate files above a size the impact could be significant for example if you don t double check the size of the data after fetch has completed your js thread could get tied up doing work on a large file dos and or cost you money in computing publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution beta step up your open source security game with whitesource
0
144,608
19,292,292,920
IssuesEvent
2021-12-12 01:27:32
rvvergara/next-js-basic
https://api.github.com/repos/rvvergara/next-js-basic
opened
CVE-2021-43803 (High) detected in next-9.3.2.tgz
security vulnerability
## CVE-2021-43803 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>next-9.3.2.tgz</b></p></summary> <p>The React Framework</p> <p>Library home page: <a href="https://registry.npmjs.org/next/-/next-9.3.2.tgz">https://registry.npmjs.org/next/-/next-9.3.2.tgz</a></p> <p>Path to dependency file: next-js-basic/package.json</p> <p>Path to vulnerable library: next-js-basic/node_modules/next/package.json</p> <p> Dependency Hierarchy: - :x: **next-9.3.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Next.js is a React framework. In versions of Next.js prior to 12.0.5 or 11.1.3, invalid or malformed URLs could lead to a server crash. In order to be affected by this issue, the deployment must use Next.js versions above 11.1.0 and below 12.0.5, Node.js above 15.0.0, and next start or a custom server. Deployments on Vercel are not affected, along with similar environments where invalid requests are filtered before reaching Next.js. Versions 12.0.5 and 11.1.3 contain patches for this issue. <p>Publish Date: 2021-12-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43803>CVE-2021-43803</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/vercel/next.js/security/advisories/GHSA-25mp-g6fv-mqxx">https://github.com/vercel/next.js/security/advisories/GHSA-25mp-g6fv-mqxx</a></p> <p>Release Date: 2021-12-10</p> <p>Fix Resolution: next - 11.1.3,12.0.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-43803 (High) detected in next-9.3.2.tgz - ## CVE-2021-43803 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>next-9.3.2.tgz</b></p></summary> <p>The React Framework</p> <p>Library home page: <a href="https://registry.npmjs.org/next/-/next-9.3.2.tgz">https://registry.npmjs.org/next/-/next-9.3.2.tgz</a></p> <p>Path to dependency file: next-js-basic/package.json</p> <p>Path to vulnerable library: next-js-basic/node_modules/next/package.json</p> <p> Dependency Hierarchy: - :x: **next-9.3.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Next.js is a React framework. In versions of Next.js prior to 12.0.5 or 11.1.3, invalid or malformed URLs could lead to a server crash. In order to be affected by this issue, the deployment must use Next.js versions above 11.1.0 and below 12.0.5, Node.js above 15.0.0, and next start or a custom server. Deployments on Vercel are not affected, along with similar environments where invalid requests are filtered before reaching Next.js. Versions 12.0.5 and 11.1.3 contain patches for this issue. <p>Publish Date: 2021-12-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43803>CVE-2021-43803</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/vercel/next.js/security/advisories/GHSA-25mp-g6fv-mqxx">https://github.com/vercel/next.js/security/advisories/GHSA-25mp-g6fv-mqxx</a></p> <p>Release Date: 2021-12-10</p> <p>Fix Resolution: next - 11.1.3,12.0.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in next tgz cve high severity vulnerability vulnerable library next tgz the react framework library home page a href path to dependency file next js basic package json path to vulnerable library next js basic node modules next package json dependency hierarchy x next tgz vulnerable library found in base branch master vulnerability details next js is a react framework in versions of next js prior to or invalid or malformed urls could lead to a server crash in order to be affected by this issue the deployment must use next js versions above and below node js above and next start or a custom server deployments on vercel are not affected along with similar environments where invalid requests are filtered before reaching next js versions and contain patches for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution next step up your open source security game with whitesource
0
86,680
8,042,556,289
IssuesEvent
2018-07-31 08:32:48
SuperuserLabs/thankful
https://api.github.com/repos/SuperuserLabs/thankful
opened
Add ability to run as plain webapp
improves: testing/qa priority: low
Continues work started in #70 - [ ] Fix all the errors from missing `browser` etc - [ ] Fix Metamask support when not in a WebExtension - [ ] Polyfill `browser.storage.local` to use `window.localStorage` when in webapp-mode
1.0
Add ability to run as plain webapp - Continues work started in #70 - [ ] Fix all the errors from missing `browser` etc - [ ] Fix Metamask support when not in a WebExtension - [ ] Polyfill `browser.storage.local` to use `window.localStorage` when in webapp-mode
test
add ability to run as plain webapp continues work started in fix all the errors from missing browser etc fix metamask support when not in a webextension polyfill browser storage local to use window localstorage when in webapp mode
1
301,467
26,050,575,761
IssuesEvent
2022-12-22 18:13:31
vegaprotocol/vega
https://api.github.com/repos/vegaprotocol/vega
closed
Implement test coverage for 0038-OLIQ-008
feature tests
IN order to get test coverage for 0038-OLIQ-liquidity_provision_order_type.md we need to cover the following ACs - [ ] 0038-OLIQ-008
1.0
Implement test coverage for 0038-OLIQ-008 - IN order to get test coverage for 0038-OLIQ-liquidity_provision_order_type.md we need to cover the following ACs - [ ] 0038-OLIQ-008
test
implement test coverage for oliq in order to get test coverage for oliq liquidity provision order type md we need to cover the following acs oliq
1
183,261
14,219,402,452
IssuesEvent
2020-11-17 13:14:19
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
util/stop: TestStopperIsStopped failed
C-test-failure O-robot branch-master
[(util/stop).TestStopperIsStopped failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2448413&tab=buildLog) on [master@807755c5473da7f3ad384fe8bb04b804ab56f911](https://github.com/cockroachdb/cockroach/commits/807755c5473da7f3ad384fe8bb04b804ab56f911): ``` === RUN TestStopperIsStopped I201116 23:11:04.028342 83 util/stop/stopper.go:564 quiescing stopper.go:98: leaked stopper, created at: goroutine 82 [running]: runtime/debug.Stack(0x3306240, 0xc000610000, 0x0) /usr/local/go/src/runtime/debug/stack.go:24 +0xab github.com/cockroachdb/cockroach/pkg/util/stop.register(0xc000158960) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:52 +0x45 github.com/cockroachdb/cockroach/pkg/util/stop.NewStopper(0x0, 0x0, 0x0, 0x5c7d51) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0x39c github.com/cockroachdb/cockroach/pkg/util/stop_test.TestStopperIsStopped(0xc000092900) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper_test.go:88 +0xa5 testing.tRunner(0xc000092900, 0x217b400) /usr/local/go/src/testing/testing.go:1123 +0x203 created by testing.(*T).Run /usr/local/go/src/testing/testing.go:1168 +0x5bc --- FAIL: TestStopperIsStopped (0.10s) ``` <details><summary>More</summary><p> Parameters: - GOFLAGS=-json ``` make stressrace TESTS=TestStopperIsStopped PKG=./pkg/util/stop TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestStopperIsStopped.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
1.0
util/stop: TestStopperIsStopped failed - [(util/stop).TestStopperIsStopped failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2448413&tab=buildLog) on [master@807755c5473da7f3ad384fe8bb04b804ab56f911](https://github.com/cockroachdb/cockroach/commits/807755c5473da7f3ad384fe8bb04b804ab56f911): ``` === RUN TestStopperIsStopped I201116 23:11:04.028342 83 util/stop/stopper.go:564 quiescing stopper.go:98: leaked stopper, created at: goroutine 82 [running]: runtime/debug.Stack(0x3306240, 0xc000610000, 0x0) /usr/local/go/src/runtime/debug/stack.go:24 +0xab github.com/cockroachdb/cockroach/pkg/util/stop.register(0xc000158960) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:52 +0x45 github.com/cockroachdb/cockroach/pkg/util/stop.NewStopper(0x0, 0x0, 0x0, 0x5c7d51) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0x39c github.com/cockroachdb/cockroach/pkg/util/stop_test.TestStopperIsStopped(0xc000092900) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper_test.go:88 +0xa5 testing.tRunner(0xc000092900, 0x217b400) /usr/local/go/src/testing/testing.go:1123 +0x203 created by testing.(*T).Run /usr/local/go/src/testing/testing.go:1168 +0x5bc --- FAIL: TestStopperIsStopped (0.10s) ``` <details><summary>More</summary><p> Parameters: - GOFLAGS=-json ``` make stressrace TESTS=TestStopperIsStopped PKG=./pkg/util/stop TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestStopperIsStopped.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
util stop teststopperisstopped failed on run teststopperisstopped util stop stopper go quiescing stopper go leaked stopper created at goroutine runtime debug stack usr local go src runtime debug stack go github com cockroachdb cockroach pkg util stop register go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg util stop newstopper go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg util stop test teststopperisstopped go src github com cockroachdb cockroach pkg util stop stopper test go testing trunner usr local go src testing testing go created by testing t run usr local go src testing testing go fail teststopperisstopped more parameters goflags json make stressrace tests teststopperisstopped pkg pkg util stop testtimeout stressflags timeout powered by
1
31,252
7,333,057,416
IssuesEvent
2018-03-05 18:10:28
dotnet/roslyn-analyzers
https://api.github.com/repos/dotnet/roslyn-analyzers
closed
CA2100 does not fire for msdn/docs sample
Area-Microsoft.CodeQuality.Analyzers Bug Dependency-DataFlow
#### Analyzer package Example: Microsoft.CodeQuality.Analyzers.Exp, version [2.6.1-beta1-62702-01](https://dotnet.myget.org/feed/roslyn-analyzers/package/nuget/Microsoft.CodeQuality.Analyzers.Exp/2.6.1-beta1-62702-01) #### Analyzer ReviewSqlQueriesForSecurityVulnerabilities #### Repro steps 1. Copy the C#or VB sample from https://docs.microsoft.com/en-us/visualstudio/code-quality/ca2100-review-sql-queries-for-security-vulnerabilities #### Expected behavior CA2100 fires. #### Actual behavior No CA2100 This seems to be due to the fact that `SqlCommand.CommandText` overrides `DbCommand.CommandText`, and the latter implements `IDbCommand.CommandText`, while the analyzer only checks if the invoked property (i.e. `SqlCommand.CommandText`) directly implements `IDbCommand.CommandText`.
1.0
CA2100 does not fire for msdn/docs sample - #### Analyzer package Example: Microsoft.CodeQuality.Analyzers.Exp, version [2.6.1-beta1-62702-01](https://dotnet.myget.org/feed/roslyn-analyzers/package/nuget/Microsoft.CodeQuality.Analyzers.Exp/2.6.1-beta1-62702-01) #### Analyzer ReviewSqlQueriesForSecurityVulnerabilities #### Repro steps 1. Copy the C#or VB sample from https://docs.microsoft.com/en-us/visualstudio/code-quality/ca2100-review-sql-queries-for-security-vulnerabilities #### Expected behavior CA2100 fires. #### Actual behavior No CA2100 This seems to be due to the fact that `SqlCommand.CommandText` overrides `DbCommand.CommandText`, and the latter implements `IDbCommand.CommandText`, while the analyzer only checks if the invoked property (i.e. `SqlCommand.CommandText`) directly implements `IDbCommand.CommandText`.
non_test
does not fire for msdn docs sample analyzer package example microsoft codequality analyzers exp version analyzer reviewsqlqueriesforsecurityvulnerabilities repro steps copy the c or vb sample from expected behavior fires actual behavior no this seems to be due to the fact that sqlcommand commandtext overrides dbcommand commandtext and the latter implements idbcommand commandtext while the analyzer only checks if the invoked property i e sqlcommand commandtext directly implements idbcommand commandtext
0
482,901
13,915,543,286
IssuesEvent
2020-10-21 00:56:03
visit-dav/visit
https://api.github.com/repos/visit-dav/visit
closed
len_string is no longer part of exodus format
bug impact medium likelihood medium priority reviewed
https://github.com/visit-dav/visit/blob/629922a7ac2219d9f312aacbd2486e5f2a1c99f0/src/databases/Exodus/avtExodusFileFormat.C#L141 I think that "maximum_name_length" is the parameter needed here. I asked Greg Sjaardema and it looks like len_str is only added when QA records are added to a file and was mistakenly added to files by ex_copy until some point between v 7.08 and v 7.18. With the current logic, if len_string is not defined, VisIt does not see variables in a file. I can provide example exodus files upon request
1.0
len_string is no longer part of exodus format - https://github.com/visit-dav/visit/blob/629922a7ac2219d9f312aacbd2486e5f2a1c99f0/src/databases/Exodus/avtExodusFileFormat.C#L141 I think that "maximum_name_length" is the parameter needed here. I asked Greg Sjaardema and it looks like len_str is only added when QA records are added to a file and was mistakenly added to files by ex_copy until some point between v 7.08 and v 7.18. With the current logic, if len_string is not defined, VisIt does not see variables in a file. I can provide example exodus files upon request
non_test
len string is no longer part of exodus format i think that maximum name length is the parameter needed here i asked greg sjaardema and it looks like len str is only added when qa records are added to a file and was mistakenly added to files by ex copy until some point between v and v with the current logic if len string is not defined visit does not see variables in a file i can provide example exodus files upon request
0
71,161
7,236,035,860
IssuesEvent
2018-02-13 04:23:50
mono/mono
https://api.github.com/repos/mono/mono
closed
dim-diamondshape.exe and dim-sharedgenerics.exe tests fail on FullAOT
area-Runtime: AOT test enhancement
They were reenabled recently by https://github.com/mono/mono/commit/5c4510a8a7f50ebaa9ad1c19711df995dca15661 and now fail e.g. on https://jenkins.mono-project.com/job/test-mono-pull-request-amd64-fullaot/7814/testReport/ ``` MonoTests.runtime.dim-diamondshape.exe MESSAGE: Calling IFoo.Foo on Foo - expecting exception. Exception caught: System.NotSupportedException: Interface method 'IFoo:Foo (int)' in class 'FooClass' has multiple candidate implementations. at Program.Negative () [0x00015] in <ae5be20b4348427f93f5a12d54425b6c>:0 Calling I1.Func on I47Class - expecting exception Exception caught: System.NotSupportedException: Interface method 'I1:Func (int)' in class 'I47Class' has multiple candidate implementations. at Program.Negative () [0x0005e] in <ae5be20b4348427f93f5a12d54425b6c>:0 Calling GI1<T>.Func on GI23Class<S> - expecting exception Exception caught: System.NotSupportedException: Interface method 'GI1`1:Func<S> (System.Type[]&)' in class 'GI23Class`1<object>' has multiple candidate implementations. at Program.Negative () [0x000aa] in <ae5be20b4348427f93f5a12d54425b6c>:0 Calling I1.Func on I4Class - expecting I4.Func At I4.Func PASS Calling I1.Func on I8Class - expecting I8.Func At I8.Func PASS Calling GI1.Func on GI4Class<object> - expecting GI4.Func<S> +++++++++++++++++++ STACK TRACE: Unhandled Exception: System.ExecutionEngineException: Attempting to JIT compile method 'GI4`1<object>:GI1<T>.Func<string> (System.Type[]&)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information. at Program.Positive () [0x00069] in <ae5be20b4348427f93f5a12d54425b6c>:0 at Program.Main () [0x00007] in <ae5be20b4348427f93f5a12d54425b6c>:0 [ERROR] FATAL UNHANDLED EXCEPTION: System.ExecutionEngineException: Attempting to JIT compile method 'GI4`1<object>:GI1<T>.Func<string> (System.Type[]&)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information. at Program.Positive () [0x00069] in <ae5be20b4348427f93f5a12d54425b6c>:0 at Program.Main () [0x00007] in <ae5be20b4348427f93f5a12d54425b6c>:0 | 0.12 sec | 8 ``` ``` MonoTests.runtime.dim-sharedgenerics.exe MESSAGE: Calling IFoo<string>.Foo on FooBar<string, object> - expecting default method IFoo<string>.Foo At IFoo.Foo:Arg=ABC, TypeOf(T)=System.String PASS Calling IBar<string[]>.Foo on FooBar<string, object> - expecting default method IBar<object>.Foo +++++++++++++++++++ STACK TRACE: Unhandled Exception: System.ExecutionEngineException: Attempting to JIT compile method 'IBar`1<object>:Bar (object)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information. at Program.Main () [0x00046] in <e0066f8e0a474276ad30263edbd086bc>:0 [ERROR] FATAL UNHANDLED EXCEPTION: System.ExecutionEngineException: Attempting to JIT compile method 'IBar`1<object>:Bar (object)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information. at Program.Main () [0x00046] in <e0066f8e0a474276ad30263edbd086bc>:0 ```
1.0
dim-diamondshape.exe and dim-sharedgenerics.exe tests fail on FullAOT - They were reenabled recently by https://github.com/mono/mono/commit/5c4510a8a7f50ebaa9ad1c19711df995dca15661 and now fail e.g. on https://jenkins.mono-project.com/job/test-mono-pull-request-amd64-fullaot/7814/testReport/ ``` MonoTests.runtime.dim-diamondshape.exe MESSAGE: Calling IFoo.Foo on Foo - expecting exception. Exception caught: System.NotSupportedException: Interface method 'IFoo:Foo (int)' in class 'FooClass' has multiple candidate implementations. at Program.Negative () [0x00015] in <ae5be20b4348427f93f5a12d54425b6c>:0 Calling I1.Func on I47Class - expecting exception Exception caught: System.NotSupportedException: Interface method 'I1:Func (int)' in class 'I47Class' has multiple candidate implementations. at Program.Negative () [0x0005e] in <ae5be20b4348427f93f5a12d54425b6c>:0 Calling GI1<T>.Func on GI23Class<S> - expecting exception Exception caught: System.NotSupportedException: Interface method 'GI1`1:Func<S> (System.Type[]&)' in class 'GI23Class`1<object>' has multiple candidate implementations. at Program.Negative () [0x000aa] in <ae5be20b4348427f93f5a12d54425b6c>:0 Calling I1.Func on I4Class - expecting I4.Func At I4.Func PASS Calling I1.Func on I8Class - expecting I8.Func At I8.Func PASS Calling GI1.Func on GI4Class<object> - expecting GI4.Func<S> +++++++++++++++++++ STACK TRACE: Unhandled Exception: System.ExecutionEngineException: Attempting to JIT compile method 'GI4`1<object>:GI1<T>.Func<string> (System.Type[]&)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information. at Program.Positive () [0x00069] in <ae5be20b4348427f93f5a12d54425b6c>:0 at Program.Main () [0x00007] in <ae5be20b4348427f93f5a12d54425b6c>:0 [ERROR] FATAL UNHANDLED EXCEPTION: System.ExecutionEngineException: Attempting to JIT compile method 'GI4`1<object>:GI1<T>.Func<string> (System.Type[]&)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information. at Program.Positive () [0x00069] in <ae5be20b4348427f93f5a12d54425b6c>:0 at Program.Main () [0x00007] in <ae5be20b4348427f93f5a12d54425b6c>:0 | 0.12 sec | 8 ``` ``` MonoTests.runtime.dim-sharedgenerics.exe MESSAGE: Calling IFoo<string>.Foo on FooBar<string, object> - expecting default method IFoo<string>.Foo At IFoo.Foo:Arg=ABC, TypeOf(T)=System.String PASS Calling IBar<string[]>.Foo on FooBar<string, object> - expecting default method IBar<object>.Foo +++++++++++++++++++ STACK TRACE: Unhandled Exception: System.ExecutionEngineException: Attempting to JIT compile method 'IBar`1<object>:Bar (object)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information. at Program.Main () [0x00046] in <e0066f8e0a474276ad30263edbd086bc>:0 [ERROR] FATAL UNHANDLED EXCEPTION: System.ExecutionEngineException: Attempting to JIT compile method 'IBar`1<object>:Bar (object)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information. at Program.Main () [0x00046] in <e0066f8e0a474276ad30263edbd086bc>:0 ```
test
dim diamondshape exe and dim sharedgenerics exe tests fail on fullaot they were reenabled recently by and now fail e g on monotests runtime dim diamondshape exe message calling ifoo foo on foo expecting exception exception caught system notsupportedexception interface method ifoo foo int in class fooclass has multiple candidate implementations at program negative in calling func on expecting exception exception caught system notsupportedexception interface method func int in class has multiple candidate implementations at program negative in calling func on expecting exception exception caught system notsupportedexception interface method func system type in class has multiple candidate implementations at program negative in calling func on expecting func at func pass calling func on expecting func at func pass calling func on expecting func stack trace unhandled exception system executionengineexception attempting to jit compile method func system type while running in aot only mode see for more information at program positive in at program main in fatal unhandled exception system executionengineexception attempting to jit compile method func system type while running in aot only mode see for more information at program positive in at program main in sec monotests runtime dim sharedgenerics exe message calling ifoo foo on foobar expecting default method ifoo foo at ifoo foo arg abc typeof t system string pass calling ibar foo on foobar expecting default method ibar foo stack trace unhandled exception system executionengineexception attempting to jit compile method ibar bar object while running in aot only mode see for more information at program main in fatal unhandled exception system executionengineexception attempting to jit compile method ibar bar object while running in aot only mode see for more information at program main in
1
43,406
5,537,627,344
IssuesEvent
2017-03-21 22:38:29
fossology/fossology
https://api.github.com/repos/fossology/fossology
closed
run Stress Testing weekly with latest code
bug Category: Testing Priority: Normal Status: New Tracker: Bug
--- Author Name: **larry shi** Original Redmine Issue: 6987, http://www.fossology.org/issues/6987 Original Date: 2014/05/07 Original Assignee: Dong Ma --- manually or automatically.
1.0
run Stress Testing weekly with latest code - --- Author Name: **larry shi** Original Redmine Issue: 6987, http://www.fossology.org/issues/6987 Original Date: 2014/05/07 Original Assignee: Dong Ma --- manually or automatically.
test
run stress testing weekly with latest code author name larry shi original redmine issue original date original assignee dong ma manually or automatically
1
163,270
12,710,354,294
IssuesEvent
2020-06-23 13:46:21
pywbem/pywbemtools
https://api.github.com/repos/pywbem/pywbemtools
closed
Test: CIMInstanceName.to_wbem_uri() sorts by default in pywbem 1.0.0
area: test resolution: fixed type: bug
CIMInstanceName.to_wbem_uri() sorts by default since pywbem 1.0.0. This affects some of the pywbemcli tests. Since pywbemcli needs to run with both pywbem before and after 1.0.0, the tests need to be adjusted so that accomodate both.
1.0
Test: CIMInstanceName.to_wbem_uri() sorts by default in pywbem 1.0.0 - CIMInstanceName.to_wbem_uri() sorts by default since pywbem 1.0.0. This affects some of the pywbemcli tests. Since pywbemcli needs to run with both pywbem before and after 1.0.0, the tests need to be adjusted so that accomodate both.
test
test ciminstancename to wbem uri sorts by default in pywbem ciminstancename to wbem uri sorts by default since pywbem this affects some of the pywbemcli tests since pywbemcli needs to run with both pywbem before and after the tests need to be adjusted so that accomodate both
1
2,237
2,524,972,710
IssuesEvent
2015-01-20 21:20:29
graybeal/ont
https://api.github.com/repos/graybeal/ont
closed
Please set up periodic "harvest" for updated version of GCOOS ontology
1 star content enhancement imported Milestone-Beta1 ooici Priority-Medium
_From [steph_wa...@consolidated.net](https://code.google.com/u/112846428158176258467/) on September 14, 2009 19:27:24_ What capability do you want added or improved? Felimon at GCOOS requests a periodic "harvest" of the updated version of GCOOS ontology for the MMI repository. Harvest from: http://gcoos.rsmas.miami.edu/dp/srv_gcoos_generateOWL.php Where do you want this capability to be accessible? to be automatic What sort of input/command mechanism do you want? What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies, text contents of vocabulary (voc2rdf), operating system, browser/version (Firefox, Safari, Chrome, IE, etc.), screenshot, etc.) _Original issue: http://code.google.com/p/mmisw/issues/detail?id=178_
1.0
Please set up periodic "harvest" for updated version of GCOOS ontology - _From [steph_wa...@consolidated.net](https://code.google.com/u/112846428158176258467/) on September 14, 2009 19:27:24_ What capability do you want added or improved? Felimon at GCOOS requests a periodic "harvest" of the updated version of GCOOS ontology for the MMI repository. Harvest from: http://gcoos.rsmas.miami.edu/dp/srv_gcoos_generateOWL.php Where do you want this capability to be accessible? to be automatic What sort of input/command mechanism do you want? What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies, text contents of vocabulary (voc2rdf), operating system, browser/version (Firefox, Safari, Chrome, IE, etc.), screenshot, etc.) _Original issue: http://code.google.com/p/mmisw/issues/detail?id=178_
non_test
please set up periodic harvest for updated version of gcoos ontology from on september what capability do you want added or improved felimon at gcoos requests a periodic harvest of the updated version of gcoos ontology for the mmi repository harvest from where do you want this capability to be accessible to be automatic what sort of input command mechanism do you want what is the desired output content format location other details of your desired capability what version of the product are you using please provide any additional information below particular ontology ies text contents of vocabulary operating system browser version firefox safari chrome ie etc screenshot etc original issue
0
73,729
7,353,421,953
IssuesEvent
2018-03-09 00:37:10
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Add a common.log for tests
discuss test
- **Subsystem**: Test It's come up a couple of times so far, so it's probably time for a proper discussion on the use of `console.log` in tests (i.e. should tests have any `console.log` output, and if so, when).
1.0
Add a common.log for tests - - **Subsystem**: Test It's come up a couple of times so far, so it's probably time for a proper discussion on the use of `console.log` in tests (i.e. should tests have any `console.log` output, and if so, when).
test
add a common log for tests subsystem test it s come up a couple of times so far so it s probably time for a proper discussion on the use of console log in tests i e should tests have any console log output and if so when
1
153,174
5,886,738,916
IssuesEvent
2017-05-17 04:20:39
FDPA/fdpa
https://api.github.com/repos/FDPA/fdpa
closed
Local Resolution PDFs are returning 404
Priority ready for work
Craft thinks the PDF lives here http://107.170.57.118/assets/uploads/resolutions/Bethlehem-Township.pdf But that link is 404. This is true of all PDFs.
1.0
Local Resolution PDFs are returning 404 - Craft thinks the PDF lives here http://107.170.57.118/assets/uploads/resolutions/Bethlehem-Township.pdf But that link is 404. This is true of all PDFs.
non_test
local resolution pdfs are returning craft thinks the pdf lives here but that link is this is true of all pdfs
0
118,629
15,342,905,984
IssuesEvent
2021-02-27 18:05:03
plotn/coolreader
https://api.github.com/repos/plotn/coolreader
closed
Перенести настройки "Опции рендеринга" и "Уровень совместимости DOM"
design
Пользователь [написал](https://4pda.ru/forum/index.php?s=&showtopic=995536&view=findpost&p=104618639): > "Опции рендеринга" и "Уровень совместимости DOM" на мой взгляд стоит убрать в "Редкие и экспериментальные". Понять что это такое у простого юзверя нет никаких шансов (даже невзирая на разъяснения), да и умолчальное их значение подойдёт подавляющему большинству. А то маячит, пугает, справки не даёт, дублируется в "Шрифтах" и "CSS" Дальше ещё немного обсуждения: > хорошая мысль. Но лучше тогда в "тонкие настройки шрифта", дополнительно переименовав их в "... рендеринга и шрифта"
1.0
Перенести настройки "Опции рендеринга" и "Уровень совместимости DOM" - Пользователь [написал](https://4pda.ru/forum/index.php?s=&showtopic=995536&view=findpost&p=104618639): > "Опции рендеринга" и "Уровень совместимости DOM" на мой взгляд стоит убрать в "Редкие и экспериментальные". Понять что это такое у простого юзверя нет никаких шансов (даже невзирая на разъяснения), да и умолчальное их значение подойдёт подавляющему большинству. А то маячит, пугает, справки не даёт, дублируется в "Шрифтах" и "CSS" Дальше ещё немного обсуждения: > хорошая мысль. Но лучше тогда в "тонкие настройки шрифта", дополнительно переименовав их в "... рендеринга и шрифта"
non_test
перенести настройки опции рендеринга и уровень совместимости dom пользователь опции рендеринга и уровень совместимости dom на мой взгляд стоит убрать в редкие и экспериментальные понять что это такое у простого юзверя нет никаких шансов даже невзирая на разъяснения да и умолчальное их значение подойдёт подавляющему большинству а то маячит пугает справки не даёт дублируется в шрифтах и css дальше ещё немного обсуждения хорошая мысль но лучше тогда в тонкие настройки шрифта дополнительно переименовав их в рендеринга и шрифта
0