Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
58,469
| 6,599,523,733
|
IssuesEvent
|
2017-09-16 20:53:27
|
nskins/goby
|
https://api.github.com/repos/nskins/goby
|
closed
|
Refactor tests to use `let` instead of instance variables
|
better test suite
|
I noticed that most of the tests use a before action to assign instance variables that are used throughout the tests. This is not great for a couple of reasons:
1. Instance variables are created wherever they are referenced, so you are more prone to typos.
2. Oftentimes objects are being instantiated that are not being used in all of the tests, which is unnecessary.
I suggest refactoring to use `let`, this will clean up the code, protect against inadvertent typos, and speed up run time (`let` is lazy-loaded so you won't be instantiating every object for every test).
For example, the Escape Spec would now look like this:
```ruby
#spec/goby/battle/escape_spec.rb
RSpec.describe Goby::Escape do
let(:player) { Player.new }
let(:monster) { Monster.new }
let(:escape) { Escape.new }
context "constructor" do
it "has an appropriate default name" do
expect(escape.name).to eq "Escape"
end
end
context "run" do
# The purpose of this test is to run the code without error.
it "should return a usable result" do
# Exercise both branches of this function w/ high probability.
20.times do
escape.run(player, monster)
expect(player.escaped).to_not be nil
end
end
end
end
```
This is a pretty big refactor, I'm happy to take it on if you think it's worth it.
|
1.0
|
Refactor tests to use `let` instead of instance variables - I noticed that most of the tests use a before action to assign instance variables that are used throughout the tests. This is not great for a couple of reasons:
1. Instance variables are created wherever they are referenced, so you are more prone to typos.
2. Oftentimes objects are being instantiated that are not being used in all of the tests, which is unnecessary.
I suggest refactoring to use `let`, this will clean up the code, protect against inadvertent typos, and speed up run time (`let` is lazy-loaded so you won't be instantiating every object for every test).
For example, the Escape Spec would now look like this:
```ruby
#spec/goby/battle/escape_spec.rb
RSpec.describe Goby::Escape do
let(:player) { Player.new }
let(:monster) { Monster.new }
let(:escape) { Escape.new }
context "constructor" do
it "has an appropriate default name" do
expect(escape.name).to eq "Escape"
end
end
context "run" do
# The purpose of this test is to run the code without error.
it "should return a usable result" do
# Exercise both branches of this function w/ high probability.
20.times do
escape.run(player, monster)
expect(player.escaped).to_not be nil
end
end
end
end
```
This is a pretty big refactor, I'm happy to take it on if you think it's worth it.
|
test
|
refactor tests to use let instead of instance variables i noticed that most of the tests use a before action to assign instance variables that are used throughout the tests this is not great for a couple of reasons instance variables are created wherever they are referenced so you are more prone to typos oftentimes objects are being instantiated that are not being used in all of the tests which is unnecessary i suggest refactoring to use let this will clean up the code protect against inadvertent typos and speed up run time let is lazy loaded so you won t be instantiating every object for every test for example the escape spec would now look like this ruby spec goby battle escape spec rb rspec describe goby escape do let player player new let monster monster new let escape escape new context constructor do it has an appropriate default name do expect escape name to eq escape end end context run do the purpose of this test is to run the code without error it should return a usable result do exercise both branches of this function w high probability times do escape run player monster expect player escaped to not be nil end end end end this is a pretty big refactor i m happy to take it on if you think it s worth it
| 1
|
245,888
| 20,809,849,629
|
IssuesEvent
|
2022-03-18 00:27:40
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
Blocked count highlight doesn't cover entire number
|
bug feature/shields priority/P3 needs-discussion QA/Yes QA/Test-Plan-Specified feature/shields/panel OS/Desktop
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
Blocked count highlight doesn't cover entire number
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. install `1.38.35`
2. launch Brave
3. enable `Shields V2` via `brave://flags`
4. restart Brave
5. sit on an XHR-happy page (any Facebook one will do)
6. check the Shields-blocked count via the icon-tip in the URL bar
7. once it reaches `99+`, click on it to expand the flyout panel
8. click to open its `Advanced controls` sub-panel
9. hover over the number for `Trackers & ads blocked (standard)`
10. note the highlight paint region
## Actual result:
<!--Please add screenshots if needed-->
<img width="1312" alt="Screen Shot 2022-03-11 at 12 41 01 AM" src="https://user-images.githubusercontent.com/387249/157833842-8933eca0-785e-4dae-afca-da0b4e72ab21.png">
## Expected result:
Should paint the entire number with the hover highlight
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
100%
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.38.35 Chromium: 99.0.4844.51 (Official Build) nightly (arm64)
-- | --
Revision | d537ec02474b5afe23684e7963d538896c63ac77-refs/branch-heads/4844@{#875}
OS | macOS Version 11.6.4 (Build 20G417)
cc @nullhook @rebron @sri
|
1.0
|
Blocked count highlight doesn't cover entire number - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
Blocked count highlight doesn't cover entire number
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. install `1.38.35`
2. launch Brave
3. enable `Shields V2` via `brave://flags`
4. restart Brave
5. sit on an XHR-happy page (any Facebook one will do)
6. check the Shields-blocked count via the icon-tip in the URL bar
7. once it reaches `99+`, click on it to expand the flyout panel
8. click to open its `Advanced controls` sub-panel
9. hover over the number for `Trackers & ads blocked (standard)`
10. note the highlight paint region
## Actual result:
<!--Please add screenshots if needed-->
<img width="1312" alt="Screen Shot 2022-03-11 at 12 41 01 AM" src="https://user-images.githubusercontent.com/387249/157833842-8933eca0-785e-4dae-afca-da0b4e72ab21.png">
## Expected result:
Should paint the entire number with the hover highlight
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
100%
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.38.35 Chromium: 99.0.4844.51 (Official Build) nightly (arm64)
-- | --
Revision | d537ec02474b5afe23684e7963d538896c63ac77-refs/branch-heads/4844@{#875}
OS | macOS Version 11.6.4 (Build 20G417)
cc @nullhook @rebron @sri
|
test
|
blocked count highlight doesn t cover entire number have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description blocked count highlight doesn t cover entire number steps to reproduce install launch brave enable shields via brave flags restart brave sit on an xhr happy page any facebook one will do check the shields blocked count via the icon tip in the url bar once it reaches click on it to expand the flyout panel click to open its advanced controls sub panel hover over the number for trackers ads blocked standard note the highlight paint region actual result img width alt screen shot at am src expected result should paint the entire number with the hover highlight reproduces how often brave version brave version info brave chromium official build nightly revision refs branch heads os macos version build cc nullhook rebron sri
| 1
|
60,154
| 6,672,967,037
|
IssuesEvent
|
2017-10-04 13:39:55
|
w3c/web-platform-tests
|
https://api.github.com/repos/w3c/web-platform-tests
|
closed
|
sequential_async_test
|
infra testharness.js
|
Originally posted as https://github.com/w3c/testharness.js/issues/96 by @mvano on 09 Dec 2014, 14:16 UTC:
> I'd like to run multiple async tests from a single page, but execute them sequentially, not in parallel. The idea would be for later tests to delay execution until the previous ones are done.
>
> The goal would be to make tests that share state (across the tests, or in the object under test) be deterministic. Here's a contrived example:
>
> ``` javascript
> var i = 0;
>
> sequential_async_test(function(test) {
> setTimeout(function() {
> assert_equals(i, 0);
> i++;
> test.done();
> }, Math.round(Math.random() * 100));
> }, 'Async test 0');
>
> sequential_async_test(function(test) {
> setTimeout(function() {
> assert_equals(i, 1);
> i++;
> test.done();
> }, Math.round(Math.random() * 100));
> }, 'Async test 1');
> ```
>
> I guess the alternative is to make separate pages, each with a single `async_test`. I'll probably do that for the time being.
|
1.0
|
sequential_async_test - Originally posted as https://github.com/w3c/testharness.js/issues/96 by @mvano on 09 Dec 2014, 14:16 UTC:
> I'd like to run multiple async tests from a single page, but execute them sequentially, not in parallel. The idea would be for later tests to delay execution until the previous ones are done.
>
> The goal would be to make tests that share state (across the tests, or in the object under test) be deterministic. Here's a contrived example:
>
> ``` javascript
> var i = 0;
>
> sequential_async_test(function(test) {
> setTimeout(function() {
> assert_equals(i, 0);
> i++;
> test.done();
> }, Math.round(Math.random() * 100));
> }, 'Async test 0');
>
> sequential_async_test(function(test) {
> setTimeout(function() {
> assert_equals(i, 1);
> i++;
> test.done();
> }, Math.round(Math.random() * 100));
> }, 'Async test 1');
> ```
>
> I guess the alternative is to make separate pages, each with a single `async_test`. I'll probably do that for the time being.
|
test
|
sequential async test originally posted as by mvano on dec utc i d like to run multiple async tests from a single page but execute them sequentially not in parallel the idea would be for later tests to delay execution until the previous ones are done the goal would be to make tests that share state across the tests or in the object under test be deterministic here s a contrived example javascript var i sequential async test function test settimeout function assert equals i i test done math round math random async test sequential async test function test settimeout function assert equals i i test done math round math random async test i guess the alternative is to make separate pages each with a single async test i ll probably do that for the time being
| 1
|
80,537
| 15,586,293,928
|
IssuesEvent
|
2021-03-18 01:36:51
|
saurockSaurav/weather-information-api
|
https://api.github.com/repos/saurockSaurav/weather-information-api
|
opened
|
CVE-2020-14061 (High) detected in jackson-databind-2.8.11.3.jar
|
security vulnerability
|
## CVE-2020-14061 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.11.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /weather-information-api/weather-rest-api-service/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.11.3/jackson-databind-2.8.11.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.5.20.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.11.3.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to oracle.jms.AQjmsQueueConnectionFactory, oracle.jms.AQjmsXATopicConnectionFactory, oracle.jms.AQjmsTopicConnectionFactory, oracle.jms.AQjmsXAQueueConnectionFactory, and oracle.jms.AQjmsXAConnectionFactory (aka weblogic/oracle-aqjms).
<p>Publish Date: 2020-06-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14061>CVE-2020-14061</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061</a></p>
<p>Release Date: 2020-06-14</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-14061 (High) detected in jackson-databind-2.8.11.3.jar - ## CVE-2020-14061 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.11.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /weather-information-api/weather-rest-api-service/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.11.3/jackson-databind-2.8.11.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.5.20.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.11.3.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to oracle.jms.AQjmsQueueConnectionFactory, oracle.jms.AQjmsXATopicConnectionFactory, oracle.jms.AQjmsTopicConnectionFactory, oracle.jms.AQjmsXAQueueConnectionFactory, and oracle.jms.AQjmsXAConnectionFactory (aka weblogic/oracle-aqjms).
<p>Publish Date: 2020-06-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14061>CVE-2020-14061</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061</a></p>
<p>Release Date: 2020-06-14</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file weather information api weather rest api service pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to oracle jms aqjmsqueueconnectionfactory oracle jms aqjmsxatopicconnectionfactory oracle jms aqjmstopicconnectionfactory oracle jms aqjmsxaqueueconnectionfactory and oracle jms aqjmsxaconnectionfactory aka weblogic oracle aqjms publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
| 0
|
153,343
| 12,141,093,418
|
IssuesEvent
|
2020-04-23 21:45:07
|
rust-lang/rust
|
https://api.github.com/repos/rust-lang/rust
|
closed
|
Emit a warning when a codeblock is using "compile-fail" instead of "compile_fail"
|
A-doctests C-enhancement T-rustdoc
|
I just found out that a lot of error code explanations were using "compile-fail" instead of "compile_fail". I'm fixing this issue as part of something a bit bigger, however I assume this error is pretty common in rust and that rustdoc should warn people about it.
The big issue here is that since it's not a known tag, rustdoc doesn't recognize the codeblock as a rust one and therefore doesn't test it, which is pretty bad.
|
1.0
|
Emit a warning when a codeblock is using "compile-fail" instead of "compile_fail" - I just found out that a lot of error code explanations were using "compile-fail" instead of "compile_fail". I'm fixing this issue as part of something a bit bigger, however I assume this error is pretty common in rust and that rustdoc should warn people about it.
The big issue here is that since it's not a known tag, rustdoc doesn't recognize the codeblock as a rust one and therefore doesn't test it, which is pretty bad.
|
test
|
emit a warning when a codeblock is using compile fail instead of compile fail i just found out that a lot of error code explanations were using compile fail instead of compile fail i m fixing this issue as part of something a bit bigger however i assume this error is pretty common in rust and that rustdoc should warn people about it the big issue here is that since it s not a known tag rustdoc doesn t recognize the codeblock as a rust one and therefore doesn t test it which is pretty bad
| 1
|
189,901
| 14,527,387,969
|
IssuesEvent
|
2020-12-14 15:17:49
|
navarrotheus/caramelo-tec-2-CK0236
|
https://api.github.com/repos/navarrotheus/caramelo-tec-2-CK0236
|
opened
|
Testar SolicitationService
|
BACK TESTES
|
## Método create
- [ ] Teste de sucesso: Cria a solicitação com sucesso
- [ ] Teste de erro: Usuário com tal id não existe
- [ ] Teste de erro: Pet com tal id não existe
## Método update
- [ ] Teste de sucesso: Atualiza a solicitação com sucesso
- [ ] Teste de sucesso: Caso a solicitação seja aceita, cria uma adoção com o solicitante e o pet e seta a disponibilidade do pet como falsa
- [ ] Teste de erro: Usuário com tal id não existe
## Método delete
- [ ] Teste de sucesso: Deleta a solicitação com sucesso
- [ ] Teste de erro: Usuário com tal id não existe
## Método search
- [ ] Teste de sucesso: Busca as solicitações do usuário com sucesso
- [ ] Teste de erro: Usuário com tal id não existe
## Método searchPetSolicitations
- [ ] Teste de sucesso: Busca as solicitações do usuário com sucesso
- [ ] Teste de erro: Usuário com tal id não existe
- [ ] Teste de erro: Pet com tal id não existe
- [ ] Teste de erro: Usuário tenta buscar as solicitações de um Pet de outro usuário
|
1.0
|
Testar SolicitationService - ## Método create
- [ ] Teste de sucesso: Cria a solicitação com sucesso
- [ ] Teste de erro: Usuário com tal id não existe
- [ ] Teste de erro: Pet com tal id não existe
## Método update
- [ ] Teste de sucesso: Atualiza a solicitação com sucesso
- [ ] Teste de sucesso: Caso a solicitação seja aceita, cria uma adoção com o solicitante e o pet e seta a disponibilidade do pet como falsa
- [ ] Teste de erro: Usuário com tal id não existe
## Método delete
- [ ] Teste de sucesso: Deleta a solicitação com sucesso
- [ ] Teste de erro: Usuário com tal id não existe
## Método search
- [ ] Teste de sucesso: Busca as solicitações do usuário com sucesso
- [ ] Teste de erro: Usuário com tal id não existe
## Método searchPetSolicitations
- [ ] Teste de sucesso: Busca as solicitações do usuário com sucesso
- [ ] Teste de erro: Usuário com tal id não existe
- [ ] Teste de erro: Pet com tal id não existe
- [ ] Teste de erro: Usuário tenta buscar as solicitações de um Pet de outro usuário
|
test
|
testar solicitationservice método create teste de sucesso cria a solicitação com sucesso teste de erro usuário com tal id não existe teste de erro pet com tal id não existe método update teste de sucesso atualiza a solicitação com sucesso teste de sucesso caso a solicitação seja aceita cria uma adoção com o solicitante e o pet e seta a disponibilidade do pet como falsa teste de erro usuário com tal id não existe método delete teste de sucesso deleta a solicitação com sucesso teste de erro usuário com tal id não existe método search teste de sucesso busca as solicitações do usuário com sucesso teste de erro usuário com tal id não existe método searchpetsolicitations teste de sucesso busca as solicitações do usuário com sucesso teste de erro usuário com tal id não existe teste de erro pet com tal id não existe teste de erro usuário tenta buscar as solicitações de um pet de outro usuário
| 1
|
605,103
| 18,724,969,978
|
IssuesEvent
|
2021-11-03 15:26:27
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
Details section is empty on Reject/Approve transaction screen
|
priority/P3 QA/No release-notes/exclude feature/wallet OS/Android
|
We should fill tx details there if there are any
|
1.0
|
Details section is empty on Reject/Approve transaction screen - We should fill tx details there if there are any
|
non_test
|
details section is empty on reject approve transaction screen we should fill tx details there if there are any
| 0
|
209,653
| 16,048,047,764
|
IssuesEvent
|
2021-04-22 15:41:35
|
input-output-hk/ouroboros-network
|
https://api.github.com/repos/input-output-hk/ouroboros-network
|
closed
|
Property test which checks that codecs produce a valid CBOR encoding
|
testing
|
[validFlatTerm](http://hackage.haskell.org/package/cborg-0.2.5.0/docs/Codec-CBOR-FlatTerm.html#v:validFlatTerm)
```
validFlatTerm . toFlatTerm :: CBOR.Encoding -> Bool
```
|
1.0
|
Property test which checks that codecs produce a valid CBOR encoding - [validFlatTerm](http://hackage.haskell.org/package/cborg-0.2.5.0/docs/Codec-CBOR-FlatTerm.html#v:validFlatTerm)
```
validFlatTerm . toFlatTerm :: CBOR.Encoding -> Bool
```
|
test
|
property test which checks that codecs produce a valid cbor encoding validflatterm toflatterm cbor encoding bool
| 1
|
623,348
| 19,665,663,607
|
IssuesEvent
|
2022-01-10 22:10:23
|
aesimpson/sama-sanity
|
https://api.github.com/repos/aesimpson/sama-sanity
|
closed
|
CSS Fix for Multiple Modules within a section
|
Priority:Moderate
|
At xxxxlg wide screens, the stacked modules don't wrap:
|
1.0
|
CSS Fix for Multiple Modules within a section - At xxxxlg wide screens, the stacked modules don't wrap:
|
non_test
|
css fix for multiple modules within a section at xxxxlg wide screens the stacked modules don t wrap
| 0
|
186,817
| 14,409,561,604
|
IssuesEvent
|
2020-12-04 02:33:39
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
[test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/maps/embeddable/tooltip_filter_actions·js - maps app embeddable tooltip filter actions apply filter to current view "before all" hook for "should display create filter button when tooltip is locked"
|
Team:Geo failed-test test-cloud
|
**Version: 7.10.0**
**Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/maps/embeddable/tooltip_filter_actions·js**
**Stack Trace:**
```
Error: retry.try timeout: TypeError: Cannot read property 'clearValue' of undefined
at retry.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp1/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/test/functional/services/listing_table.ts:107:28)
at process._tickCallback (internal/process/next_tick.js:68:7)
at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp1/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:28:9)
at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp1/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:68:13)
```
**Other test failures:**
- maps app embeddable tooltip filter actions panel actions "before all" hook for "should display more actions button when tooltip is locked"
_Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/845/testReport/_
|
2.0
|
[test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/maps/embeddable/tooltip_filter_actions·js - maps app embeddable tooltip filter actions apply filter to current view "before all" hook for "should display create filter button when tooltip is locked" - **Version: 7.10.0**
**Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/maps/embeddable/tooltip_filter_actions·js**
**Stack Trace:**
```
Error: retry.try timeout: TypeError: Cannot read property 'clearValue' of undefined
at retry.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp1/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/test/functional/services/listing_table.ts:107:28)
at process._tickCallback (internal/process/next_tick.js:68:7)
at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp1/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:28:9)
at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp1/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:68:13)
```
**Other test failures:**
- maps app embeddable tooltip filter actions panel actions "before all" hook for "should display more actions button when tooltip is locked"
_Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/845/testReport/_
|
test
|
chrome x pack ui functional x pack test functional apps maps embeddable tooltip filter actions·js maps app embeddable tooltip filter actions apply filter to current view before all hook for should display create filter button when tooltip is locked version class chrome x pack ui functional x pack test functional apps maps embeddable tooltip filter actions·js stack trace error retry try timeout typeerror cannot read property clearvalue of undefined at retry try var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node linux immutable ci cloud common build kibana test functional services listing table ts at process tickcallback internal process next tick js at onfailure var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node linux immutable ci cloud common build kibana test common services retry retry for success ts at retryforsuccess var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node linux immutable ci cloud common build kibana test common services retry retry for success ts other test failures maps app embeddable tooltip filter actions panel actions before all hook for should display more actions button when tooltip is locked test report
| 1
|
142,542
| 11,484,798,098
|
IssuesEvent
|
2020-02-11 05:16:42
|
proarc/proarc
|
https://api.github.com/repos/proarc/proarc
|
closed
|
RDflow - čárový kod se nepropíše
|
6 k testování RDFlow Release-3.5.16
|
Čárový kód je v metadatech
<mods:identifier type="barcode">26001600877</mods:identifier>
Ale nezobrazí se v tabulce Bibliografického záznamu.
|
1.0
|
RDflow - čárový kod se nepropíše - Čárový kód je v metadatech
<mods:identifier type="barcode">26001600877</mods:identifier>
Ale nezobrazí se v tabulce Bibliografického záznamu.
|
test
|
rdflow čárový kod se nepropíše čárový kód je v metadatech ale nezobrazí se v tabulce bibliografického záznamu
| 1
|
24,697
| 4,106,194,761
|
IssuesEvent
|
2016-06-06 07:38:34
|
Mr-Kumar-Abhishek/zuzeelik
|
https://api.github.com/repos/Mr-Kumar-Abhishek/zuzeelik
|
opened
|
Test builds from zuzeelik's pre-alpha version v0.0.0-0.5.0 in windows OS
|
testing
|
Test builds from zuzeelik's pre-alpha version` v0.0.0-0.5.0` in windows OS. Probably there could be some bugs with built-in functions that handles *quotes*. In *nix systems, all built-in functions are working fine.
|
1.0
|
Test builds from zuzeelik's pre-alpha version v0.0.0-0.5.0 in windows OS - Test builds from zuzeelik's pre-alpha version` v0.0.0-0.5.0` in windows OS. Probably there could be some bugs with built-in functions that handles *quotes*. In *nix systems, all built-in functions are working fine.
|
test
|
test builds from zuzeelik s pre alpha version in windows os test builds from zuzeelik s pre alpha version in windows os probably there could be some bugs with built in functions that handles quotes in nix systems all built in functions are working fine
| 1
|
2,057
| 2,873,078,819
|
IssuesEvent
|
2015-06-08 15:18:24
|
meumobi/sitebuilder
|
https://api.github.com/repos/meumobi/sitebuilder
|
closed
|
new user can't accept invites if another user is already logged in
|
bug sitebuilder
|
The new user can't validade, and the error message is: You need to choose a valid language
|
1.0
|
new user can't accept invites if another user is already logged in - The new user can't validade, and the error message is: You need to choose a valid language
|
non_test
|
new user can t accept invites if another user is already logged in the new user can t validade and the error message is you need to choose a valid language
| 0
|
152,600
| 12,121,608,652
|
IssuesEvent
|
2020-04-22 09:34:43
|
Students-of-the-city-of-Kostroma/Ray-of-hope
|
https://api.github.com/repos/Students-of-the-city-of-Kostroma/Ray-of-hope
|
closed
|
Протестировать регистрацию и авторизацию организации на новом сервере
|
AppServer LoginOrg O3 PR5 RegOrg Sprint 14 Testing
|
Epic #286 Task #287 #288
Протестировать функции, принимающие данные от клиента, обрабатывающие их и возвращающие ответ, на соответствие спецификациям [о регистрации](https://docs.google.com/document/d/1QkQMIYAaNvvknFlBldHHwvcq2iuWHP7E4QYHCFUZdy4) и [авторизации](https://docs.google.com/document/d/1tc8xZATtXaF6GUWpDQYYTfyn8zvWJbQeJYwcmKPkdW4) организации. Поднимать баги по мере нахождения.
|
1.0
|
Протестировать регистрацию и авторизацию организации на новом сервере - Epic #286 Task #287 #288
Протестировать функции, принимающие данные от клиента, обрабатывающие их и возвращающие ответ, на соответствие спецификациям [о регистрации](https://docs.google.com/document/d/1QkQMIYAaNvvknFlBldHHwvcq2iuWHP7E4QYHCFUZdy4) и [авторизации](https://docs.google.com/document/d/1tc8xZATtXaF6GUWpDQYYTfyn8zvWJbQeJYwcmKPkdW4) организации. Поднимать баги по мере нахождения.
|
test
|
протестировать регистрацию и авторизацию организации на новом сервере epic task протестировать функции принимающие данные от клиента обрабатывающие их и возвращающие ответ на соответствие спецификациям и организации поднимать баги по мере нахождения
| 1
|
269,843
| 23,471,396,035
|
IssuesEvent
|
2022-08-16 22:22:31
|
red/red
|
https://api.github.com/repos/red/red
|
closed
|
Deceptive error message from `set-quiet`
|
status.built status.tested type.bug
|
**Describe the bug**
```
>> o: object []
== make object! []
>> set-quiet in o 'x 1
*** Script Error: set-quiet does not allow word! for its word argument ;) what???
*** Where: set-quiet
*** Near : 1
*** Stack:
```
**To reproduce**
`set-quiet none 1`
**Expected behavior**
"Doesn't accept none for it's word argument"
**Platform version**
```
red-view-14aug22-4eb8ad83f.exe
```
|
1.0
|
Deceptive error message from `set-quiet` - **Describe the bug**
```
>> o: object []
== make object! []
>> set-quiet in o 'x 1
*** Script Error: set-quiet does not allow word! for its word argument ;) what???
*** Where: set-quiet
*** Near : 1
*** Stack:
```
**To reproduce**
`set-quiet none 1`
**Expected behavior**
"Doesn't accept none for it's word argument"
**Platform version**
```
red-view-14aug22-4eb8ad83f.exe
```
|
test
|
deceptive error message from set quiet describe the bug o object make object set quiet in o x script error set quiet does not allow word for its word argument what where set quiet near stack to reproduce set quiet none expected behavior doesn t accept none for it s word argument platform version red view exe
| 1
|
256,392
| 22,048,396,783
|
IssuesEvent
|
2022-05-30 06:04:16
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
ccl/changefeedccl: TestChangefeedBackfillCheckpoint failed
|
C-test-failure O-robot branch-release-22.1
|
ccl/changefeedccl.TestChangefeedBackfillCheckpoint [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=5312529&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=5312529&tab=artifacts#/) on release-22.1 @ [34e6fcfdc8f3155831305ab4a78f960aaad3e7bc](https://github.com/cockroachdb/cockroach/commits/34e6fcfdc8f3155831305ab4a78f960aaad3e7bc):
```
=== RUN TestChangefeedBackfillCheckpoint
test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/a77002d7c9453d7cd2d382f907780e13/logTestChangefeedBackfillCheckpoint2134013902
test_log_scope.go:80: use -show-logs to present logs inline
=== CONT TestChangefeedBackfillCheckpoint
changefeed_test.go:5440: -- test log scope end --
--- FAIL: TestChangefeedBackfillCheckpoint (541.52s)
=== RUN TestChangefeedBackfillCheckpoint/enterprise-limit=100_B
changefeed_test.go:5420:
Error Trace: changefeed_test.go:5420
helpers_test.go:554
Error: Received unexpected error:
retrying txn failed after 50 attempts. original error: pq: restart transaction: TransactionRetryWithProtoRefreshError: TransactionRetryError: retry txn (RETRY_SERIALIZABLE - failed preemptive refresh due to a conflict: committed value on key /Table/109/1/"foo"/0/766181192725659648/0): "sql txn" meta={id=1ef45b68 key=/Table/109/1/"foo"/0/766179641595625472/0 pri=0.00000000 epo=50 ts=1653890353.444952019,2 min=1653890008.216986178,0 seq=8087} lock=true stat=PENDING rts=1653890344.066420091,0 wto=false gul=1653890008.716986178,0.
Test: TestChangefeedBackfillCheckpoint/enterprise-limit=100_B
--- FAIL: TestChangefeedBackfillCheckpoint/enterprise-limit=100_B (486.62s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
Parameters in this failure:
- TAGS=bazel,gss,deadlock
</p>
</details>
/cc @cockroachdb/cdc
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestChangefeedBackfillCheckpoint.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
ccl/changefeedccl: TestChangefeedBackfillCheckpoint failed - ccl/changefeedccl.TestChangefeedBackfillCheckpoint [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=5312529&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=5312529&tab=artifacts#/) on release-22.1 @ [34e6fcfdc8f3155831305ab4a78f960aaad3e7bc](https://github.com/cockroachdb/cockroach/commits/34e6fcfdc8f3155831305ab4a78f960aaad3e7bc):
```
=== RUN TestChangefeedBackfillCheckpoint
test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/a77002d7c9453d7cd2d382f907780e13/logTestChangefeedBackfillCheckpoint2134013902
test_log_scope.go:80: use -show-logs to present logs inline
=== CONT TestChangefeedBackfillCheckpoint
changefeed_test.go:5440: -- test log scope end --
--- FAIL: TestChangefeedBackfillCheckpoint (541.52s)
=== RUN TestChangefeedBackfillCheckpoint/enterprise-limit=100_B
changefeed_test.go:5420:
Error Trace: changefeed_test.go:5420
helpers_test.go:554
Error: Received unexpected error:
retrying txn failed after 50 attempts. original error: pq: restart transaction: TransactionRetryWithProtoRefreshError: TransactionRetryError: retry txn (RETRY_SERIALIZABLE - failed preemptive refresh due to a conflict: committed value on key /Table/109/1/"foo"/0/766181192725659648/0): "sql txn" meta={id=1ef45b68 key=/Table/109/1/"foo"/0/766179641595625472/0 pri=0.00000000 epo=50 ts=1653890353.444952019,2 min=1653890008.216986178,0 seq=8087} lock=true stat=PENDING rts=1653890344.066420091,0 wto=false gul=1653890008.716986178,0.
Test: TestChangefeedBackfillCheckpoint/enterprise-limit=100_B
--- FAIL: TestChangefeedBackfillCheckpoint/enterprise-limit=100_B (486.62s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
Parameters in this failure:
- TAGS=bazel,gss,deadlock
</p>
</details>
/cc @cockroachdb/cdc
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestChangefeedBackfillCheckpoint.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
ccl changefeedccl testchangefeedbackfillcheckpoint failed ccl changefeedccl testchangefeedbackfillcheckpoint with on release run testchangefeedbackfillcheckpoint test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline cont testchangefeedbackfillcheckpoint changefeed test go test log scope end fail testchangefeedbackfillcheckpoint run testchangefeedbackfillcheckpoint enterprise limit b changefeed test go error trace changefeed test go helpers test go error received unexpected error retrying txn failed after attempts original error pq restart transaction transactionretrywithprotorefresherror transactionretryerror retry txn retry serializable failed preemptive refresh due to a conflict committed value on key table foo sql txn meta id key table foo pri epo ts min seq lock true stat pending rts wto false gul test testchangefeedbackfillcheckpoint enterprise limit b fail testchangefeedbackfillcheckpoint enterprise limit b help see also parameters in this failure tags bazel gss deadlock cc cockroachdb cdc
| 1
|
65,588
| 14,740,878,706
|
IssuesEvent
|
2021-01-07 09:45:58
|
hiptest/ember-easy-datatable
|
https://api.github.com/repos/hiptest/ember-easy-datatable
|
opened
|
CVE-2018-16487 (Medium) detected in lodash-3.10.1.tgz
|
security vulnerability
|
## CVE-2018-16487 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary>
<p>The modern build of lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p>
<p>Path to dependency file: ember-easy-datatable/package.json</p>
<p>Path to vulnerable library: ember-easy-datatable/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- ember-cli-2.12.2.tgz (Root Library)
- broccoli-babel-transpiler-5.7.4.tgz
- babel-core-5.8.38.tgz
- :x: **lodash-3.10.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/hiptest/ember-easy-datatable/commit/174fc2ea19b9aaaa080440d9cb938d1e9a2d6120">174fc2ea19b9aaaa080440d9cb938d1e9a2d6120</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.
<p>Publish Date: 2019-02-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p>
<p>Release Date: 2019-02-01</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.10.1","isTransitiveDependency":true,"dependencyTree":"ember-cli:2.12.2;broccoli-babel-transpiler:5.7.4;babel-core:5.8.38;lodash:3.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"}],"vulnerabilityIdentifier":"CVE-2018-16487","vulnerabilityDetails":"A prototype pollution vulnerability was found in lodash \u003c4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487","cvss3Severity":"medium","cvss3Score":"5.6","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2018-16487 (Medium) detected in lodash-3.10.1.tgz - ## CVE-2018-16487 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary>
<p>The modern build of lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p>
<p>Path to dependency file: ember-easy-datatable/package.json</p>
<p>Path to vulnerable library: ember-easy-datatable/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- ember-cli-2.12.2.tgz (Root Library)
- broccoli-babel-transpiler-5.7.4.tgz
- babel-core-5.8.38.tgz
- :x: **lodash-3.10.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/hiptest/ember-easy-datatable/commit/174fc2ea19b9aaaa080440d9cb938d1e9a2d6120">174fc2ea19b9aaaa080440d9cb938d1e9a2d6120</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.
<p>Publish Date: 2019-02-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p>
<p>Release Date: 2019-02-01</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.10.1","isTransitiveDependency":true,"dependencyTree":"ember-cli:2.12.2;broccoli-babel-transpiler:5.7.4;babel-core:5.8.38;lodash:3.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"}],"vulnerabilityIdentifier":"CVE-2018-16487","vulnerabilityDetails":"A prototype pollution vulnerability was found in lodash \u003c4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487","cvss3Severity":"medium","cvss3Score":"5.6","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file ember easy datatable package json path to vulnerable library ember easy datatable node modules lodash package json dependency hierarchy ember cli tgz root library broccoli babel transpiler tgz babel core tgz x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype vulnerabilityurl
| 0
|
249,856
| 21,195,374,409
|
IssuesEvent
|
2022-04-08 23:31:16
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
ci-cadvisor-e2e failed
|
priority/important-soon area/cadvisor sig/node kind/failing-test triage/accepted
|
### Which jobs are failing?
ci-cadvisor-e2e
### Which tests are failing?
```
W0322 22:57:59.666] 2022/03/22 22:57:59 main.go:331: Something went wrong: encountered 1 errors: [error during go run /go/src/k8s.io/kubernetes/test/e2e_node/runner/remote/run_remote.go --cleanup --logtostderr --vmodule=*=4 --ssh-env=gce --results-dir=/workspace/_artifacts --project=ci-cadvisor-e2e --zone=us-central1-f --ssh-user=prow --ssh-key=/workspace/.ssh/google_compute_engine --ginkgo-flags=--nodes=1 --test_args= --test-timeout=10m0s --image-config-file=/workspace/test-infra/jobs/e2e_node/containerd/image-config.yaml --test-suite=cadvisor: exit status 1]
W0322 22:57:59.673] Traceback (most recent call last):
W0322 22:57:59.674] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 723, in <module>
W0322 22:57:59.674] main(parse_args())
W0322 22:57:59.674] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 569, in main
W0322 22:57:59.674] mode.start(runner_args)
W0322 22:57:59.674] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 228, in start
W0322 22:57:59.675] check_env(env, self.command, *args)
W0322 22:57:59.675] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 111, in check_env
W0322 22:57:59.675] subprocess.check_call(cmd, env=env)
W0322 22:57:59.675] File "/usr/lib/python2.7/subprocess.py", line 190, in check_call
W0322 22:57:59.675] raise CalledProcessError(retcode, cmd)
W0322 22:57:59.676] subprocess.CalledProcessError: Command '('kubetest', '--dump=/workspace/_artifacts', '--gcp-service-account=/etc/service-account/service-account.json', '--up', '--down', '--test', '--deployment=node', '--provider=gce', '--cluster=bootstrap-e2e', '--gcp-network=bootstrap-e2e', '--gcp-project=ci-cadvisor-e2e', '--gcp-zone=us-central1-f', '--node-args=--image-config-file=/workspace/test-infra/jobs/e2e_node/containerd/image-config.yaml --test-suite=cadvisor', '--node-tests=true', '--test_args=--nodes=1', '--timeout=10m')' returned non-zero exit status 1
E0322 22:57:59.677] Command failed
I0322 22:57:59.678] process 339 exited with code 1 after 2.2m
E0322 22:57:59.678] FAIL: ci-cadvisor-e2e
```
### Since when has it been failing?
Since 03-22.
Changes list that may
- https://github.com/kubernetes/kubernetes/compare/dd604a0f9a2e015a869d01957297e33f2c0c7025...95e30f66c?
- - Candidate1: https://github.com/kubernetes/kubernetes/pull/108704 (but for windows)
- https://github.com/kubernetes/test-infra/compare/ce1169a66...a6fe00c22
- - Candidate1: https://github.com/kubernetes/test-infra/pull/25727
- - Candidate2: https://github.com/kubernetes/test-infra/pull/25405
### Testgrid link
https://testgrid.k8s.io/sig-node-cadvisor#cadvisor-e2e
### Reason for failure (if possible)
_No response_
### Anything else we need to know?
_No response_
### Relevant SIG(s)
/sig node
/area cadvisor
|
1.0
|
ci-cadvisor-e2e failed - ### Which jobs are failing?
ci-cadvisor-e2e
### Which tests are failing?
```
W0322 22:57:59.666] 2022/03/22 22:57:59 main.go:331: Something went wrong: encountered 1 errors: [error during go run /go/src/k8s.io/kubernetes/test/e2e_node/runner/remote/run_remote.go --cleanup --logtostderr --vmodule=*=4 --ssh-env=gce --results-dir=/workspace/_artifacts --project=ci-cadvisor-e2e --zone=us-central1-f --ssh-user=prow --ssh-key=/workspace/.ssh/google_compute_engine --ginkgo-flags=--nodes=1 --test_args= --test-timeout=10m0s --image-config-file=/workspace/test-infra/jobs/e2e_node/containerd/image-config.yaml --test-suite=cadvisor: exit status 1]
W0322 22:57:59.673] Traceback (most recent call last):
W0322 22:57:59.674] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 723, in <module>
W0322 22:57:59.674] main(parse_args())
W0322 22:57:59.674] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 569, in main
W0322 22:57:59.674] mode.start(runner_args)
W0322 22:57:59.674] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 228, in start
W0322 22:57:59.675] check_env(env, self.command, *args)
W0322 22:57:59.675] File "/workspace/./test-infra/jenkins/../scenarios/kubernetes_e2e.py", line 111, in check_env
W0322 22:57:59.675] subprocess.check_call(cmd, env=env)
W0322 22:57:59.675] File "/usr/lib/python2.7/subprocess.py", line 190, in check_call
W0322 22:57:59.675] raise CalledProcessError(retcode, cmd)
W0322 22:57:59.676] subprocess.CalledProcessError: Command '('kubetest', '--dump=/workspace/_artifacts', '--gcp-service-account=/etc/service-account/service-account.json', '--up', '--down', '--test', '--deployment=node', '--provider=gce', '--cluster=bootstrap-e2e', '--gcp-network=bootstrap-e2e', '--gcp-project=ci-cadvisor-e2e', '--gcp-zone=us-central1-f', '--node-args=--image-config-file=/workspace/test-infra/jobs/e2e_node/containerd/image-config.yaml --test-suite=cadvisor', '--node-tests=true', '--test_args=--nodes=1', '--timeout=10m')' returned non-zero exit status 1
E0322 22:57:59.677] Command failed
I0322 22:57:59.678] process 339 exited with code 1 after 2.2m
E0322 22:57:59.678] FAIL: ci-cadvisor-e2e
```
### Since when has it been failing?
Since 03-22.
Changes list that may
- https://github.com/kubernetes/kubernetes/compare/dd604a0f9a2e015a869d01957297e33f2c0c7025...95e30f66c?
- - Candidate1: https://github.com/kubernetes/kubernetes/pull/108704 (but for windows)
- https://github.com/kubernetes/test-infra/compare/ce1169a66...a6fe00c22
- - Candidate1: https://github.com/kubernetes/test-infra/pull/25727
- - Candidate2: https://github.com/kubernetes/test-infra/pull/25405
### Testgrid link
https://testgrid.k8s.io/sig-node-cadvisor#cadvisor-e2e
### Reason for failure (if possible)
_No response_
### Anything else we need to know?
_No response_
### Relevant SIG(s)
/sig node
/area cadvisor
|
test
|
ci cadvisor failed which jobs are failing ci cadvisor which tests are failing main go something went wrong encountered errors traceback most recent call last file workspace test infra jenkins scenarios kubernetes py line in main parse args file workspace test infra jenkins scenarios kubernetes py line in main mode start runner args file workspace test infra jenkins scenarios kubernetes py line in start check env env self command args file workspace test infra jenkins scenarios kubernetes py line in check env subprocess check call cmd env env file usr lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command kubetest dump workspace artifacts gcp service account etc service account service account json up down test deployment node provider gce cluster bootstrap gcp network bootstrap gcp project ci cadvisor gcp zone us f node args image config file workspace test infra jobs node containerd image config yaml test suite cadvisor node tests true test args nodes timeout returned non zero exit status command failed process exited with code after fail ci cadvisor since when has it been failing since changes list that may (but for windows) testgrid link reason for failure if possible no response anything else we need to know no response relevant sig s sig node area cadvisor
| 1
|
51,145
| 13,190,289,230
|
IssuesEvent
|
2020-08-13 09:55:47
|
ESA-VirES/WebClient-Framework
|
https://api.github.com/repos/ESA-VirES/WebClient-Framework
|
opened
|
Broken server-side interpolation of the EEF data.
|
defect
|
When selecting MAG and EEF data the server responds with following error:
```
Error: Problem retrieving data: 'Interp1D' object has no attribute 'indices_nearest'
```

This is a regression introduces in v3.3.0.
Observed on the production instance.
Already fixed on staging.
FAO @lmar76
|
1.0
|
Broken server-side interpolation of the EEF data. - When selecting MAG and EEF data the server responds with following error:
```
Error: Problem retrieving data: 'Interp1D' object has no attribute 'indices_nearest'
```

This is a regression introduces in v3.3.0.
Observed on the production instance.
Already fixed on staging.
FAO @lmar76
|
non_test
|
broken server side interpolation of the eef data when selecting mag and eef data the server responds with following error error problem retrieving data object has no attribute indices nearest this is a regression introduces in observed on the production instance already fixed on staging fao
| 0
|
21,665
| 3,911,689,155
|
IssuesEvent
|
2016-04-20 07:21:57
|
Legion-Expansion/Legion-Expansion
|
https://api.github.com/repos/Legion-Expansion/Legion-Expansion
|
reopened
|
Remove Icon Extensions and Icon Reloader as dependencies
|
needs testing pte
|
This will be needed for PTE release, but it will break icons on 89755.
|
1.0
|
Remove Icon Extensions and Icon Reloader as dependencies - This will be needed for PTE release, but it will break icons on 89755.
|
test
|
remove icon extensions and icon reloader as dependencies this will be needed for pte release but it will break icons on
| 1
|
275,534
| 23,921,258,929
|
IssuesEvent
|
2022-09-09 17:07:25
|
ECP-WarpX/WarpX
|
https://api.github.com/repos/ECP-WarpX/WarpX
|
opened
|
Invalid memory access when moving window and timers-based load-balancing is used
|
bug bug: affects latest release component: load balancing
|
I am opening this issue because I have observed an invalid memory access when moving window and load-balancing based on timers are used in combination.
Here I provide a small reproducer:
```
#################################
####### GENERAL PARAMETERS ######
#################################
max_step = 10
amr.n_cell = 64 64 64
amr.max_grid_size = 32
amr.blocking_factor = 32
amr.max_level = 0
geometry.dims = 3
geometry.prob_lo = -10.e-6 -10.e-6 -10.e-6 # physical domain
geometry.prob_hi = 10.e-6 10.e-6 10.e-6
algo.load_balance_intervals = 3::100
algo.load_balance_with_sfc = 0
algo.load_balance_costs_update = timers
warpx.do_moving_window = 1
warpx.moving_window_dir = z
warpx.moving_window_v = 1.0
warpx.start_moving_window_step = 2
#################################
####### Boundary condition ######
#################################
boundary.field_lo = pml pml pml
boundary.field_hi = pml pml pml
#################################
############ NUMERICS ###########
#################################
warpx.verbose = 1
warpx.cfl = 0.99
# Order of particle shape factors
algo.particle_shape = 3
#################################
############ PLASMA #############
#################################
particles.species_names = electrons
electrons.species_type = electron
electrons.injection_style = "NUniformPerCell"
electrons.num_particles_per_cell_each_dim = 1 1 2
electrons.profile = constant
electrons.density = 1.e25 # number of electrons per m^3
electrons.momentum_distribution_type = "gaussian"
electrons.ux_th = 0.01 # uth the std of the (unitless) momentum
electrons.uy_th = 0.01 # uth the std of the (unitless) momentum
electrons.uz_th = 0.01 # uth the std of the (unitless) momentum
```
When WarpX runs this inputfile (even without GPUs or OMP support), `valgrind` detects the following issue:
```
STEP 3 starts ...
==41155== Invalid read of size 4
==41155== at 0x55CFBD: Add<float> (AMReX_GpuAtomic.H:584)
==41155== by 0x55CFBD: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435)
==41155== by 0x55F8EF: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:192)
==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269)
==41155== by 0x1BB863: main (main.cpp:67)
==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd
==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422)
==41155== by 0x1F088A: allocate (new_allocator.h:127)
==41155== by 0x1F088A: allocate (alloc_traits.h:464)
==41155== by 0x1F088A: _M_allocate (stl_vector.h:346)
==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635)
==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31)
==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22)
==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962)
==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170)
==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680)
==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548)
==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779)
==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472)
==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378)
==41155== by 0x1BB856: main (main.cpp:65)
==41155==
==41155== Invalid write of size 4
==41155== at 0x55CFC1: Add<float> (AMReX_GpuAtomic.H:584)
==41155== by 0x55CFC1: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435)
==41155== by 0x55F8EF: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:192)
==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269)
==41155== by 0x1BB863: main (main.cpp:67)
==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd
==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422)
==41155== by 0x1F088A: allocate (new_allocator.h:127)
==41155== by 0x1F088A: allocate (alloc_traits.h:464)
==41155== by 0x1F088A: _M_allocate (stl_vector.h:346)
==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635)
==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31)
==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22)
==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962)
==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170)
==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680)
==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548)
==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779)
==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472)
==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378)
==41155== by 0x1BB856: main (main.cpp:65)
==41155==
==41155== Invalid read of size 4
==41155== at 0x55CFBD: Add<float> (AMReX_GpuAtomic.H:584)
==41155== by 0x55CFBD: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435)
==41155== by 0x55F92F: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:193)
==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269)
==41155== by 0x1BB863: main (main.cpp:67)
==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd
==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422)
==41155== by 0x1F088A: allocate (new_allocator.h:127)
==41155== by 0x1F088A: allocate (alloc_traits.h:464)
==41155== by 0x1F088A: _M_allocate (stl_vector.h:346)
==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635)
==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31)
==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22)
==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962)
==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170)
==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680)
==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548)
==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779)
==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472)
==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378)
==41155== by 0x1BB856: main (main.cpp:65)
==41155==
==41155== Invalid write of size 4
==41155== at 0x55CFC1: Add<float> (AMReX_GpuAtomic.H:584)
==41155== by 0x55CFC1: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435)
==41155== by 0x55F92F: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:193)
==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269)
==41155== by 0x1BB863: main (main.cpp:67)
==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd
==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422)
==41155== by 0x1F088A: allocate (new_allocator.h:127)
==41155== by 0x1F088A: allocate (alloc_traits.h:464)
==41155== by 0x1F088A: _M_allocate (stl_vector.h:346)
==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635)
==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31)
==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22)
==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962)
==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170)
==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680)
==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548)
==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779)
==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472)
==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378)
==41155== by 0x1BB856: main (main.cpp:65)
==41155==
STEP 3 ends. TIME = 1.787413796e-15 DT = 5.958046162e-16
Evolve time = 44.5962677 s; This step = 13.71976852 s; Avg. per step = 14.86542225 s
```
|
1.0
|
Invalid memory access when moving window and timers-based load-balancing is used - I am opening this issue because I have observed an invalid memory access when moving window and load-balancing based on timers are used in combination.
Here I provide a small reproducer:
```
#################################
####### GENERAL PARAMETERS ######
#################################
max_step = 10
amr.n_cell = 64 64 64
amr.max_grid_size = 32
amr.blocking_factor = 32
amr.max_level = 0
geometry.dims = 3
geometry.prob_lo = -10.e-6 -10.e-6 -10.e-6 # physical domain
geometry.prob_hi = 10.e-6 10.e-6 10.e-6
algo.load_balance_intervals = 3::100
algo.load_balance_with_sfc = 0
algo.load_balance_costs_update = timers
warpx.do_moving_window = 1
warpx.moving_window_dir = z
warpx.moving_window_v = 1.0
warpx.start_moving_window_step = 2
#################################
####### Boundary condition ######
#################################
boundary.field_lo = pml pml pml
boundary.field_hi = pml pml pml
#################################
############ NUMERICS ###########
#################################
warpx.verbose = 1
warpx.cfl = 0.99
# Order of particle shape factors
algo.particle_shape = 3
#################################
############ PLASMA #############
#################################
particles.species_names = electrons
electrons.species_type = electron
electrons.injection_style = "NUniformPerCell"
electrons.num_particles_per_cell_each_dim = 1 1 2
electrons.profile = constant
electrons.density = 1.e25 # number of electrons per m^3
electrons.momentum_distribution_type = "gaussian"
electrons.ux_th = 0.01 # uth the std of the (unitless) momentum
electrons.uy_th = 0.01 # uth the std of the (unitless) momentum
electrons.uz_th = 0.01 # uth the std of the (unitless) momentum
```
When WarpX runs this inputfile (even without GPUs or OMP support), `valgrind` detects the following issue:
```
STEP 3 starts ...
==41155== Invalid read of size 4
==41155== at 0x55CFBD: Add<float> (AMReX_GpuAtomic.H:584)
==41155== by 0x55CFBD: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435)
==41155== by 0x55F8EF: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:192)
==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269)
==41155== by 0x1BB863: main (main.cpp:67)
==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd
==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422)
==41155== by 0x1F088A: allocate (new_allocator.h:127)
==41155== by 0x1F088A: allocate (alloc_traits.h:464)
==41155== by 0x1F088A: _M_allocate (stl_vector.h:346)
==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635)
==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31)
==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22)
==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962)
==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170)
==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680)
==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548)
==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779)
==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472)
==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378)
==41155== by 0x1BB856: main (main.cpp:65)
==41155==
==41155== Invalid write of size 4
==41155== at 0x55CFC1: Add<float> (AMReX_GpuAtomic.H:584)
==41155== by 0x55CFC1: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435)
==41155== by 0x55F8EF: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:192)
==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269)
==41155== by 0x1BB863: main (main.cpp:67)
==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd
==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422)
==41155== by 0x1F088A: allocate (new_allocator.h:127)
==41155== by 0x1F088A: allocate (alloc_traits.h:464)
==41155== by 0x1F088A: _M_allocate (stl_vector.h:346)
==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635)
==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31)
==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22)
==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962)
==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170)
==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680)
==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548)
==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779)
==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472)
==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378)
==41155== by 0x1BB856: main (main.cpp:65)
==41155==
==41155== Invalid read of size 4
==41155== at 0x55CFBD: Add<float> (AMReX_GpuAtomic.H:584)
==41155== by 0x55CFBD: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435)
==41155== by 0x55F92F: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:193)
==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269)
==41155== by 0x1BB863: main (main.cpp:67)
==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd
==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422)
==41155== by 0x1F088A: allocate (new_allocator.h:127)
==41155== by 0x1F088A: allocate (alloc_traits.h:464)
==41155== by 0x1F088A: _M_allocate (stl_vector.h:346)
==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635)
==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31)
==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22)
==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962)
==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170)
==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680)
==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548)
==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779)
==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472)
==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378)
==41155== by 0x1BB856: main (main.cpp:65)
==41155==
==41155== Invalid write of size 4
==41155== at 0x55CFC1: Add<float> (AMReX_GpuAtomic.H:584)
==41155== by 0x55CFC1: WarpX::shiftMF(amrex::MultiFab&, amrex::Geometry const&, int, int, int, float, bool, amrex::ParserExecutor<3> const&) (WarpXMovingWindow.cpp:435)
==41155== by 0x55F92F: WarpX::MoveWindow(int, bool) (WarpXMovingWindow.cpp:193)
==41155== by 0x372D78: WarpX::Evolve(int) (WarpXEvolve.cpp:269)
==41155== by 0x1BB863: main (main.cpp:67)
==41155== Address 0xb8c925c is 4 bytes before a block of size 32 alloc'd
==41155== at 0x4840F2F: operator new(unsigned long) (vg_replace_malloc.c:422)
==41155== by 0x1F088A: allocate (new_allocator.h:127)
==41155== by 0x1F088A: allocate (alloc_traits.h:464)
==41155== by 0x1F088A: _M_allocate (stl_vector.h:346)
==41155== by 0x1F088A: std::vector<float, std::allocator<float> >::_M_default_append(unsigned long) (vector.tcc:635)
==41155== by 0x1D8EEB: define (AMReX_LayoutData.H:31)
==41155== by 0x1D8EEB: LayoutData (AMReX_LayoutData.H:22)
==41155== by 0x1D8EEB: make_unique<amrex::LayoutData<float>, const amrex::BoxArray&, const amrex::DistributionMapping&> (unique_ptr.h:962)
==41155== by 0x1D8EEB: WarpX::AllocLevelMFs(int, amrex::BoxArray const&, amrex::DistributionMapping const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, amrex::IntVect const&, bool) (WarpX.cpp:2170)
==41155== by 0x1DCEAB: WarpX::AllocLevelData(int, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1680)
==41155== by 0x1DCFC7: WarpX::MakeNewLevelFromScratch(int, float, amrex::BoxArray const&, amrex::DistributionMapping const&) (WarpX.cpp:1548)
==41155== by 0x6D620D: amrex::AmrMesh::MakeNewGrids(float) (AMReX_AmrMesh.cpp:779)
==41155== by 0x3DDC2F: InitFromScratch (WarpXInitData.cpp:472)
==41155== by 0x3DDC2F: WarpX::InitData() (WarpXInitData.cpp:378)
==41155== by 0x1BB856: main (main.cpp:65)
==41155==
STEP 3 ends. TIME = 1.787413796e-15 DT = 5.958046162e-16
Evolve time = 44.5962677 s; This step = 13.71976852 s; Avg. per step = 14.86542225 s
```
|
test
|
invalid memory access when moving window and timers based load balancing is used i am opening this issue because i have observed an invalid memory access when moving window and load balancing based on timers are used in combination here i provide a small reproducer general parameters max step amr n cell amr max grid size amr blocking factor amr max level geometry dims geometry prob lo e e e physical domain geometry prob hi e e e algo load balance intervals algo load balance with sfc algo load balance costs update timers warpx do moving window warpx moving window dir z warpx moving window v warpx start moving window step boundary condition boundary field lo pml pml pml boundary field hi pml pml pml numerics warpx verbose warpx cfl order of particle shape factors algo particle shape plasma particles species names electrons electrons species type electron electrons injection style nuniformpercell electrons num particles per cell each dim electrons profile constant electrons density number of electrons per m electrons momentum distribution type gaussian electrons ux th uth the std of the unitless momentum electrons uy th uth the std of the unitless momentum electrons uz th uth the std of the unitless momentum when warpx runs this inputfile even without gpus or omp support valgrind detects the following issue step starts invalid read of size at add amrex gpuatomic h by warpx shiftmf amrex multifab amrex geometry const int int int float bool amrex parserexecutor const warpxmovingwindow cpp by warpx movewindow int bool warpxmovingwindow cpp by warpx evolve int warpxevolve cpp by main main cpp address is bytes before a block of size alloc d at operator new unsigned long vg replace malloc c by allocate new allocator h by allocate alloc traits h by m allocate stl vector h by std vector m default append unsigned long vector tcc by define amrex layoutdata h by layoutdata amrex layoutdata h by make unique const amrex boxarray const amrex distributionmapping unique ptr h by warpx alloclevelmfs int amrex boxarray const amrex distributionmapping const amrex intvect const amrex intvect const amrex intvect const amrex intvect const amrex intvect const bool warpx cpp by warpx allocleveldata int amrex boxarray const amrex distributionmapping const warpx cpp by warpx makenewlevelfromscratch int float amrex boxarray const amrex distributionmapping const warpx cpp by amrex amrmesh makenewgrids float amrex amrmesh cpp by initfromscratch warpxinitdata cpp by warpx initdata warpxinitdata cpp by main main cpp invalid write of size at add amrex gpuatomic h by warpx shiftmf amrex multifab amrex geometry const int int int float bool amrex parserexecutor const warpxmovingwindow cpp by warpx movewindow int bool warpxmovingwindow cpp by warpx evolve int warpxevolve cpp by main main cpp address is bytes before a block of size alloc d at operator new unsigned long vg replace malloc c by allocate new allocator h by allocate alloc traits h by m allocate stl vector h by std vector m default append unsigned long vector tcc by define amrex layoutdata h by layoutdata amrex layoutdata h by make unique const amrex boxarray const amrex distributionmapping unique ptr h by warpx alloclevelmfs int amrex boxarray const amrex distributionmapping const amrex intvect const amrex intvect const amrex intvect const amrex intvect const amrex intvect const bool warpx cpp by warpx allocleveldata int amrex boxarray const amrex distributionmapping const warpx cpp by warpx makenewlevelfromscratch int float amrex boxarray const amrex distributionmapping const warpx cpp by amrex amrmesh makenewgrids float amrex amrmesh cpp by initfromscratch warpxinitdata cpp by warpx initdata warpxinitdata cpp by main main cpp invalid read of size at add amrex gpuatomic h by warpx shiftmf amrex multifab amrex geometry const int int int float bool amrex parserexecutor const warpxmovingwindow cpp by warpx movewindow int bool warpxmovingwindow cpp by warpx evolve int warpxevolve cpp by main main cpp address is bytes before a block of size alloc d at operator new unsigned long vg replace malloc c by allocate new allocator h by allocate alloc traits h by m allocate stl vector h by std vector m default append unsigned long vector tcc by define amrex layoutdata h by layoutdata amrex layoutdata h by make unique const amrex boxarray const amrex distributionmapping unique ptr h by warpx alloclevelmfs int amrex boxarray const amrex distributionmapping const amrex intvect const amrex intvect const amrex intvect const amrex intvect const amrex intvect const bool warpx cpp by warpx allocleveldata int amrex boxarray const amrex distributionmapping const warpx cpp by warpx makenewlevelfromscratch int float amrex boxarray const amrex distributionmapping const warpx cpp by amrex amrmesh makenewgrids float amrex amrmesh cpp by initfromscratch warpxinitdata cpp by warpx initdata warpxinitdata cpp by main main cpp invalid write of size at add amrex gpuatomic h by warpx shiftmf amrex multifab amrex geometry const int int int float bool amrex parserexecutor const warpxmovingwindow cpp by warpx movewindow int bool warpxmovingwindow cpp by warpx evolve int warpxevolve cpp by main main cpp address is bytes before a block of size alloc d at operator new unsigned long vg replace malloc c by allocate new allocator h by allocate alloc traits h by m allocate stl vector h by std vector m default append unsigned long vector tcc by define amrex layoutdata h by layoutdata amrex layoutdata h by make unique const amrex boxarray const amrex distributionmapping unique ptr h by warpx alloclevelmfs int amrex boxarray const amrex distributionmapping const amrex intvect const amrex intvect const amrex intvect const amrex intvect const amrex intvect const bool warpx cpp by warpx allocleveldata int amrex boxarray const amrex distributionmapping const warpx cpp by warpx makenewlevelfromscratch int float amrex boxarray const amrex distributionmapping const warpx cpp by amrex amrmesh makenewgrids float amrex amrmesh cpp by initfromscratch warpxinitdata cpp by warpx initdata warpxinitdata cpp by main main cpp step ends time dt evolve time s this step s avg per step s
| 1
|
111,332
| 9,526,672,346
|
IssuesEvent
|
2019-04-28 21:56:13
|
quinoacomputing/quinoa
|
https://api.github.com/repos/quinoacomputing/quinoa
|
opened
|
Port to C++17
|
Charm++ fileconv inciter meshconv parser rngtest unittest walker
|
- [x] use `this` in lambdas
- [x] use `[[maybe_unused]]` instead of `IGNORE`
- [x] replace `std::enable_if` with `if constexpr`
- [x] replace brigand `for_each` with a fold calling lambdas
- [x] simplify `Scheme` with variadic generic lambdas passed to `std::visit`
- [ ] use `std::void_t` to simplify `Has_*` types
- [ ] Replace the overloads in Control.h with variadic ones. Besides C++17 features, the implementation of the Field template may give a clue on how to do the variadic deep-access for types in tagged tuple, see Alexandrescu, Modern C++ Design: Generic Programming and Design Patterns Applied, p69.
|
2.0
|
Port to C++17 - - [x] use `this` in lambdas
- [x] use `[[maybe_unused]]` instead of `IGNORE`
- [x] replace `std::enable_if` with `if constexpr`
- [x] replace brigand `for_each` with a fold calling lambdas
- [x] simplify `Scheme` with variadic generic lambdas passed to `std::visit`
- [ ] use `std::void_t` to simplify `Has_*` types
- [ ] Replace the overloads in Control.h with variadic ones. Besides C++17 features, the implementation of the Field template may give a clue on how to do the variadic deep-access for types in tagged tuple, see Alexandrescu, Modern C++ Design: Generic Programming and Design Patterns Applied, p69.
|
test
|
port to c use this in lambdas use instead of ignore replace std enable if with if constexpr replace brigand for each with a fold calling lambdas simplify scheme with variadic generic lambdas passed to std visit use std void t to simplify has types replace the overloads in control h with variadic ones besides c features the implementation of the field template may give a clue on how to do the variadic deep access for types in tagged tuple see alexandrescu modern c design generic programming and design patterns applied
| 1
|
204,373
| 7,087,353,417
|
IssuesEvent
|
2018-01-11 17:31:06
|
salesagility/SuiteCRM
|
https://api.github.com/repos/salesagility/SuiteCRM
|
closed
|
In an existing meeting, changing the Accounts to another Company then saving causes SuiteCRM run out of memory.
|
Fix Proposed High Priority Resolved: Next Release bug
|
#### Issue
If you have a meeting and change the Account it is associated with and save. SuiteCRM stops for a while and eventually comes back with:
Fatal error: Allowed memory size of 134217728 bytes exhausted (tried to allocate 8208 bytes) in /suitecrm/code/website/include/HTMLPurifier/HTMLPurifier.standalone.php on line 15136
going back to the meeting, the Account is changed. However, there are several Reminder boxes now filling up the meeting.
#### Expected Behavior
Saves successfully and returns to the DetailView
#### Actual Behavior
SuiteCRM stops for a while before eventually showing:
Fatal error: Allowed memory size of 134217728 bytes exhausted (tried to allocate 8208 bytes) in /suitecrm/code/website/include/HTMLPurifier/HTMLPurifier.standalone.php on line 15136
nothing appears in SuiteCRM log.
#### Possible Fix
#### Steps to Reproduce
1. Edit a pre-existing meeting which has an associated Account
2. Change the meeting's associated Account to another Account
3. Save the Meeting
4. SuiteCRM stops responding for a period of time until memory limit is hit.
5. Go back to SuiteCRM and view the meeting, several Reminders have appeared.
#### Context
Customer raised issue. Some Accounts have been changing their name/certain sections of the business splitting into own entity.
#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* SuiteCRM Version used: Version 7.9.7
This issue appears on https://demo.suiteondemand.com
I've attached a screenshot which shows multiple reminders appearing.

|
1.0
|
In an existing meeting, changing the Accounts to another Company then saving causes SuiteCRM run out of memory. -
#### Issue
If you have a meeting and change the Account it is associated with and save. SuiteCRM stops for a while and eventually comes back with:
Fatal error: Allowed memory size of 134217728 bytes exhausted (tried to allocate 8208 bytes) in /suitecrm/code/website/include/HTMLPurifier/HTMLPurifier.standalone.php on line 15136
going back to the meeting, the Account is changed. However, there are several Reminder boxes now filling up the meeting.
#### Expected Behavior
Saves successfully and returns to the DetailView
#### Actual Behavior
SuiteCRM stops for a while before eventually showing:
Fatal error: Allowed memory size of 134217728 bytes exhausted (tried to allocate 8208 bytes) in /suitecrm/code/website/include/HTMLPurifier/HTMLPurifier.standalone.php on line 15136
nothing appears in SuiteCRM log.
#### Possible Fix
#### Steps to Reproduce
1. Edit a pre-existing meeting which has an associated Account
2. Change the meeting's associated Account to another Account
3. Save the Meeting
4. SuiteCRM stops responding for a period of time until memory limit is hit.
5. Go back to SuiteCRM and view the meeting, several Reminders have appeared.
#### Context
Customer raised issue. Some Accounts have been changing their name/certain sections of the business splitting into own entity.
#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* SuiteCRM Version used: Version 7.9.7
This issue appears on https://demo.suiteondemand.com
I've attached a screenshot which shows multiple reminders appearing.

|
non_test
|
in an existing meeting changing the accounts to another company then saving causes suitecrm run out of memory issue if you have a meeting and change the account it is associated with and save suitecrm stops for a while and eventually comes back with fatal error allowed memory size of bytes exhausted tried to allocate bytes in suitecrm code website include htmlpurifier htmlpurifier standalone php on line going back to the meeting the account is changed however there are several reminder boxes now filling up the meeting expected behavior saves successfully and returns to the detailview actual behavior suitecrm stops for a while before eventually showing fatal error allowed memory size of bytes exhausted tried to allocate bytes in suitecrm code website include htmlpurifier htmlpurifier standalone php on line nothing appears in suitecrm log possible fix steps to reproduce edit a pre existing meeting which has an associated account change the meeting s associated account to another account save the meeting suitecrm stops responding for a period of time until memory limit is hit go back to suitecrm and view the meeting several reminders have appeared context customer raised issue some accounts have been changing their name certain sections of the business splitting into own entity your environment suitecrm version used version this issue appears on i ve attached a screenshot which shows multiple reminders appearing
| 0
|
264,883
| 23,145,072,200
|
IssuesEvent
|
2022-07-28 23:13:22
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Seridores - Registro por lotação - Coração de Jesus
|
generalization test development template-Síntese tecnologia informatica tag-Servidores subtag-Registro por lotação
|
DoD: Realizar o teste de Generalização do validador da tag Seridores - Registro por lotação para o Município de Coração de Jesus.
|
1.0
|
Teste de generalizacao para a tag Seridores - Registro por lotação - Coração de Jesus - DoD: Realizar o teste de Generalização do validador da tag Seridores - Registro por lotação para o Município de Coração de Jesus.
|
test
|
teste de generalizacao para a tag seridores registro por lotação coração de jesus dod realizar o teste de generalização do validador da tag seridores registro por lotação para o município de coração de jesus
| 1
|
349,897
| 24,960,894,335
|
IssuesEvent
|
2022-11-01 15:26:34
|
AY2223S1-CS2103T-T09-3/tp
|
https://api.github.com/repos/AY2223S1-CS2103T-T09-3/tp
|
closed
|
[PE-D][Tester D] Misleading lines
|
documentation
|

addproj command requires more than one necessary field to be added, and this line is misleading as it may make people think that the only necessary field is the project name
(Small visual issue that addproj and addProj does not match in this line, even though I understand that both works, maybe pulling the line I have screenshot below to the top would be clearer to the user in the UG)

<!--session: 1666944041950-12b678a2-9d92-444c-b20f-e586ee74af4e-->
<!--Version: Web v3.4.4-->
-------------
Labels: `severity.VeryLow` `type.DocumentationBug`
original: optionalemon/ped#3
|
1.0
|
[PE-D][Tester D] Misleading lines - 
addproj command requires more than one necessary field to be added, and this line is misleading as it may make people think that the only necessary field is the project name
(Small visual issue that addproj and addProj does not match in this line, even though I understand that both works, maybe pulling the line I have screenshot below to the top would be clearer to the user in the UG)

<!--session: 1666944041950-12b678a2-9d92-444c-b20f-e586ee74af4e-->
<!--Version: Web v3.4.4-->
-------------
Labels: `severity.VeryLow` `type.DocumentationBug`
original: optionalemon/ped#3
|
non_test
|
misleading lines addproj command requires more than one necessary field to be added and this line is misleading as it may make people think that the only necessary field is the project name small visual issue that addproj and addproj does not match in this line even though i understand that both works maybe pulling the line i have screenshot below to the top would be clearer to the user in the ug labels severity verylow type documentationbug original optionalemon ped
| 0
|
2,497
| 2,736,487,420
|
IssuesEvent
|
2015-04-19 13:53:08
|
tgstation/-tg-station
|
https://api.github.com/repos/tgstation/-tg-station
|
opened
|
newscaster blares, "attach_spans(input, spans)"
|
Bug say() code
|
self-explanatory.
(This)[https://github.com/tgstation/-tg-station/blob/master/code/game/say.dm#L71] seems to be at fault, although I have no idea how.
|
1.0
|
newscaster blares, "attach_spans(input, spans)" - self-explanatory.
(This)[https://github.com/tgstation/-tg-station/blob/master/code/game/say.dm#L71] seems to be at fault, although I have no idea how.
|
non_test
|
newscaster blares attach spans input spans self explanatory this seems to be at fault although i have no idea how
| 0
|
109,100
| 9,368,333,041
|
IssuesEvent
|
2019-04-03 08:27:46
|
Microsoft/AzureStorageExplorer
|
https://api.github.com/repos/Microsoft/AzureStorageExplorer
|
opened
|
Support to upload multiple folders at a time using 'Upload Folder' dialog
|
🧪 testing
|
**Storage Explorer Version:** 1.7.0_20190401.1
**Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10
**Architecture:** ia32/x64
**Regression From:** Not a regression
**Actually:**
We can drag multiple folders at the same time to one blob container/file share. And the selected folders can be uploaded successfully.
**Suggestion:**
Can we support to upload multiple folders using 'Upload Folder' dialog?
|
1.0
|
Support to upload multiple folders at a time using 'Upload Folder' dialog - **Storage Explorer Version:** 1.7.0_20190401.1
**Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10
**Architecture:** ia32/x64
**Regression From:** Not a regression
**Actually:**
We can drag multiple folders at the same time to one blob container/file share. And the selected folders can be uploaded successfully.
**Suggestion:**
Can we support to upload multiple folders using 'Upload Folder' dialog?
|
test
|
support to upload multiple folders at a time using upload folder dialog storage explorer version platform os linux ubuntu macos high sierra windows architecture regression from not a regression actually we can drag multiple folders at the same time to one blob container file share and the selected folders can be uploaded successfully suggestion can we support to upload multiple folders using upload folder dialog
| 1
|
172,123
| 6,499,309,802
|
IssuesEvent
|
2017-08-22 20:57:49
|
DCLP/dclpxsltbox
|
https://api.github.com/repos/DCLP/dclpxsltbox
|
closed
|
Displaying the tag “inverse”
|
bug component: XSLT priority: medium review
|
It is accepted by the system, but not displayed: in P.Oxy. 75.5023 (TM 128952) part. C, 1 (and ff.), I would expect something like “(turned 180°)”
https://github.com/DCLP/idp.data/blob/master/DCLP/129/128952.xml#L164
Leiden+:
```(1, inverse) (ἦ\χ/(ος)) <#β=2#> ```
XML:
```xml
<lb n="1" rend="inverse"/> <expan>ἦ<add place="above">χ</add><ex>ος</ex></expan> <num value="2">β</num>
```
[Displayed](http://dclp.github.io/dclpxsltbox/output/dclp/129/128952.html):
```ἦ\χ/(ος) β```
(see now #251 for incorrect placement of hyphen; comments previously here have been moved there).
|
1.0
|
Displaying the tag “inverse” - It is accepted by the system, but not displayed: in P.Oxy. 75.5023 (TM 128952) part. C, 1 (and ff.), I would expect something like “(turned 180°)”
https://github.com/DCLP/idp.data/blob/master/DCLP/129/128952.xml#L164
Leiden+:
```(1, inverse) (ἦ\χ/(ος)) <#β=2#> ```
XML:
```xml
<lb n="1" rend="inverse"/> <expan>ἦ<add place="above">χ</add><ex>ος</ex></expan> <num value="2">β</num>
```
[Displayed](http://dclp.github.io/dclpxsltbox/output/dclp/129/128952.html):
```ἦ\χ/(ος) β```
(see now #251 for incorrect placement of hyphen; comments previously here have been moved there).
|
non_test
|
displaying the tag “inverse” it is accepted by the system but not displayed in p oxy tm part c and ff i would expect something like “ turned ° ” leiden inverse ἦ χ ος xml xml ἦ χ ος β ἦ χ ος β see now for incorrect placement of hyphen comments previously here have been moved there
| 0
|
329,350
| 28,236,897,592
|
IssuesEvent
|
2023-04-06 01:54:46
|
microsoft/AzureStorageExplorer
|
https://api.github.com/repos/microsoft/AzureStorageExplorer
|
closed
|
The error message doesn't disappear after clearing invalid strings in text editor
|
:heavy_check_mark: merged 🧪 testing :gear: tables :beetle: regression
|
**Storage Explorer Version**: 1.28.0-dev
**Build Number**: 20230330.2
**Branch**: feature/table-explorer-react
**Platform/OS**: Windows 10/Linux Ubuntu 22.04/MacOS Ventura 13.0.1 (Apple M1 Pro)
**Architecture**: ia32/x64
**How Found**: From running test cases
**Regression From**: Previous release (1.28.1)
## Steps to Reproduce ##
1. Expand one storage account -> Tables.
2. Right click one table -> Click '[2] Open'.
3. Click 'Query' -> Open the text editor.
4. Type '1' into the text editor -> Execute query.
5. There is an error message -> Click 'Clear query'.
6. Check whether the error message disappears.
## Expected Experience ##
The error message disappears.

## Actual Experience ##
The error message doesn't disappear.

|
1.0
|
The error message doesn't disappear after clearing invalid strings in text editor - **Storage Explorer Version**: 1.28.0-dev
**Build Number**: 20230330.2
**Branch**: feature/table-explorer-react
**Platform/OS**: Windows 10/Linux Ubuntu 22.04/MacOS Ventura 13.0.1 (Apple M1 Pro)
**Architecture**: ia32/x64
**How Found**: From running test cases
**Regression From**: Previous release (1.28.1)
## Steps to Reproduce ##
1. Expand one storage account -> Tables.
2. Right click one table -> Click '[2] Open'.
3. Click 'Query' -> Open the text editor.
4. Type '1' into the text editor -> Execute query.
5. There is an error message -> Click 'Clear query'.
6. Check whether the error message disappears.
## Expected Experience ##
The error message disappears.

## Actual Experience ##
The error message doesn't disappear.

|
test
|
the error message doesn t disappear after clearing invalid strings in text editor storage explorer version dev build number branch feature table explorer react platform os windows linux ubuntu macos ventura apple pro architecture how found from running test cases regression from previous release steps to reproduce expand one storage account tables right click one table click open click query open the text editor type into the text editor execute query there is an error message click clear query check whether the error message disappears expected experience the error message disappears actual experience the error message doesn t disappear
| 1
|
227,416
| 18,062,324,757
|
IssuesEvent
|
2021-09-20 15:09:40
|
trinodb/trino
|
https://api.github.com/repos/trinodb/trino
|
closed
|
randomTableSuffix is not random enough leading to test flakiness
|
bug test
|
example
```
Error: io.trino.plugin.oracle.TestOracleTypeMapping.testDate Time elapsed: 0.284 s <<< FAILURE!
java.lang.RuntimeException:
Error executing sql:
CREATE TABLE test_date_4nznv AS SELECT CAST(DATE '1952-04-03' AS DATE) col_0,
CAST(DATE '1970-01-01' AS DATE) col_1,
CAST(DATE '1970-02-03' AS DATE) col_2,
CAST(DATE '2017-07-01' AS DATE) col_3,
CAST(DATE '2017-01-01' AS DATE) col_4,
CAST(DATE '1983-04-01' AS DATE) col_5,
CAST(DATE '1983-10-01' AS DATE) col_6
at io.trino.testing.sql.TrinoSqlExecutor.execute(TrinoSqlExecutor.java:45)
at io.trino.testing.sql.TestTable.<init>(TestTable.java:48)
at io.trino.testing.sql.TestTable.<init>(TestTable.java:41)
at io.trino.testing.datatype.CreateAsSelectDataSetup.setupTestTable(CreateAsSelectDataSetup.java:46)
at io.trino.testing.datatype.SqlDataTypeTest.execute(SqlDataTypeTest.java:78)
at io.trino.plugin.oracle.AbstractTestOracleTypeMapping.testDate(AbstractTestOracleTypeMapping.java:696)
```
|
1.0
|
randomTableSuffix is not random enough leading to test flakiness - example
```
Error: io.trino.plugin.oracle.TestOracleTypeMapping.testDate Time elapsed: 0.284 s <<< FAILURE!
java.lang.RuntimeException:
Error executing sql:
CREATE TABLE test_date_4nznv AS SELECT CAST(DATE '1952-04-03' AS DATE) col_0,
CAST(DATE '1970-01-01' AS DATE) col_1,
CAST(DATE '1970-02-03' AS DATE) col_2,
CAST(DATE '2017-07-01' AS DATE) col_3,
CAST(DATE '2017-01-01' AS DATE) col_4,
CAST(DATE '1983-04-01' AS DATE) col_5,
CAST(DATE '1983-10-01' AS DATE) col_6
at io.trino.testing.sql.TrinoSqlExecutor.execute(TrinoSqlExecutor.java:45)
at io.trino.testing.sql.TestTable.<init>(TestTable.java:48)
at io.trino.testing.sql.TestTable.<init>(TestTable.java:41)
at io.trino.testing.datatype.CreateAsSelectDataSetup.setupTestTable(CreateAsSelectDataSetup.java:46)
at io.trino.testing.datatype.SqlDataTypeTest.execute(SqlDataTypeTest.java:78)
at io.trino.plugin.oracle.AbstractTestOracleTypeMapping.testDate(AbstractTestOracleTypeMapping.java:696)
```
|
test
|
randomtablesuffix is not random enough leading to test flakiness example error io trino plugin oracle testoracletypemapping testdate time elapsed s failure java lang runtimeexception error executing sql create table test date as select cast date as date col cast date as date col cast date as date col cast date as date col cast date as date col cast date as date col cast date as date col at io trino testing sql trinosqlexecutor execute trinosqlexecutor java at io trino testing sql testtable testtable java at io trino testing sql testtable testtable java at io trino testing datatype createasselectdatasetup setuptesttable createasselectdatasetup java at io trino testing datatype sqldatatypetest execute sqldatatypetest java at io trino plugin oracle abstracttestoracletypemapping testdate abstracttestoracletypemapping java
| 1
|
174,597
| 6,541,452,911
|
IssuesEvent
|
2017-09-01 20:02:10
|
wlandau-lilly/drake
|
https://api.github.com/repos/wlandau-lilly/drake
|
closed
|
Crop hover text if too long in `dataframes_graph()`
|
TOP PRIORITY wlandau-lilly has a patch and is waiting for permission to release it
|
Long hover text could slow down performance.
|
1.0
|
Crop hover text if too long in `dataframes_graph()` - Long hover text could slow down performance.
|
non_test
|
crop hover text if too long in dataframes graph long hover text could slow down performance
| 0
|
556,223
| 16,478,335,277
|
IssuesEvent
|
2021-05-24 08:33:23
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
mobile.twitter.com - see bug description
|
browser-firefox-tablet engine-gecko ml-needsdiagnosis-false priority-critical
|
<!-- @browser: Firefox Mobile (Tablet) 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/74030 -->
**URL**: https://mobile.twitter.com/home
**Browser / Version**: Firefox Mobile (Tablet) 68.0
**Operating System**: Android 8.0.0
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: I'm requesting desktop but it keeps giving me mobile version
**Steps to Reproduce**:
I requested desktop but am redirected to mobile no matter what
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/5/13f8751e-5466-4d50-9285-b35eff0b9a41.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827194101</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/5/f66f53ac-20d0-41bd-b173-e5f6b5495642)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
mobile.twitter.com - see bug description - <!-- @browser: Firefox Mobile (Tablet) 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/74030 -->
**URL**: https://mobile.twitter.com/home
**Browser / Version**: Firefox Mobile (Tablet) 68.0
**Operating System**: Android 8.0.0
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: I'm requesting desktop but it keeps giving me mobile version
**Steps to Reproduce**:
I requested desktop but am redirected to mobile no matter what
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/5/13f8751e-5466-4d50-9285-b35eff0b9a41.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827194101</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/5/f66f53ac-20d0-41bd-b173-e5f6b5495642)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
mobile twitter com see bug description url browser version firefox mobile tablet operating system android tested another browser yes chrome problem type something else description i m requesting desktop but it keeps giving me mobile version steps to reproduce i requested desktop but am redirected to mobile no matter what view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
172,921
| 13,357,637,088
|
IssuesEvent
|
2020-08-31 10:10:20
|
apache/shardingsphere
|
https://api.github.com/repos/apache/shardingsphere
|
closed
|
PostgreSQL dedicated column missed in t_order cause GeneralDQLIT.assertExecuteQuery failed
|
test
|
## Bug Report
### Which version of ShardingSphere did you use?
master branch, 5.0.0-RC1-SNAPSHOT
### Which project did you use? ShardingSphere-JDBC or ShardingSphere-Proxy?
ShardingSphere-JDBC
### Expected behavior
Related dql-test-case pass.
### Actual behavior
Test failed, exception thrown:
```
org.postgresql.util.PSQLException: ERROR: column "rule" does not exist
at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2440)
at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2183)
at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:308)
at org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:441)
at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:365)
at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:307)
at org.postgresql.jdbc.PgStatement.executeCachedSql(PgStatement.java:293)
at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:270)
at org.postgresql.jdbc.PgStatement.executeQuery(PgStatement.java:224)
at com.zaxxer.hikari.pool.ProxyStatement.executeQuery(ProxyStatement.java:111)
at com.zaxxer.hikari.pool.HikariProxyStatement.executeQuery(HikariProxyStatement.java)
at org.apache.shardingsphere.driver.executor.StatementExecutor$1.createQueryResult(StatementExecutor.java:81)
at org.apache.shardingsphere.driver.executor.StatementExecutor$1.executeSQL(StatementExecutor.java:77)
at org.apache.shardingsphere.driver.executor.StatementExecutor$1.executeSQL(StatementExecutor.java:73)
at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.impl.DefaultSQLExecutorCallback.execute0(DefaultSQLExecutorCallback.java:75)
at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.impl.DefaultSQLExecutorCallback.execute(DefaultSQLExecutorCallback.java:57)
at org.apache.shardingsphere.infra.executor.kernel.ExecutorKernel.syncExecute(ExecutorKernel.java:99)
at org.apache.shardingsphere.infra.executor.kernel.ExecutorKernel.parallelExecute(ExecutorKernel.java:95)
at org.apache.shardingsphere.infra.executor.kernel.ExecutorKernel.execute(ExecutorKernel.java:78)
at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.SQLExecutor.execute(SQLExecutor.java:66)
at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.SQLExecutor.execute(SQLExecutor.java:50)
at org.apache.shardingsphere.driver.executor.StatementExecutor.executeQuery(StatementExecutor.java:85)
at org.apache.shardingsphere.driver.jdbc.core.statement.ShardingSphereStatement.executeQuery(ShardingSphereStatement.java:127)
at org.apache.shardingsphere.dbtest.engine.dql.GeneralDQLIT.assertExecuteQueryForStatement(GeneralDQLIT.java:70)
at org.apache.shardingsphere.dbtest.engine.dql.GeneralDQLIT.assertExecuteQuery(GeneralDQLIT.java:60)
```
### Reason analyze (If you can)
`rule` and `start_point` columns doesn't exist in `t_order` table.
Since they're special column type, we can't add these columns for all types of database. Possible solution:
- Make `schema.xml > table-create > sql` support `db-types` definition, e.g. `db-types=PostgreSQL`, then the same name tables that have different columns will be created for different database.
### Steps to reproduce the behavior, such as: SQL to execute, sharding rule configuration, when exception occur etc.
### Example codes for reproduce this issue (such as a github link).
|
1.0
|
PostgreSQL dedicated column missed in t_order cause GeneralDQLIT.assertExecuteQuery failed - ## Bug Report
### Which version of ShardingSphere did you use?
master branch, 5.0.0-RC1-SNAPSHOT
### Which project did you use? ShardingSphere-JDBC or ShardingSphere-Proxy?
ShardingSphere-JDBC
### Expected behavior
Related dql-test-case pass.
### Actual behavior
Test failed, exception thrown:
```
org.postgresql.util.PSQLException: ERROR: column "rule" does not exist
at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2440)
at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2183)
at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:308)
at org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:441)
at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:365)
at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:307)
at org.postgresql.jdbc.PgStatement.executeCachedSql(PgStatement.java:293)
at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:270)
at org.postgresql.jdbc.PgStatement.executeQuery(PgStatement.java:224)
at com.zaxxer.hikari.pool.ProxyStatement.executeQuery(ProxyStatement.java:111)
at com.zaxxer.hikari.pool.HikariProxyStatement.executeQuery(HikariProxyStatement.java)
at org.apache.shardingsphere.driver.executor.StatementExecutor$1.createQueryResult(StatementExecutor.java:81)
at org.apache.shardingsphere.driver.executor.StatementExecutor$1.executeSQL(StatementExecutor.java:77)
at org.apache.shardingsphere.driver.executor.StatementExecutor$1.executeSQL(StatementExecutor.java:73)
at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.impl.DefaultSQLExecutorCallback.execute0(DefaultSQLExecutorCallback.java:75)
at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.impl.DefaultSQLExecutorCallback.execute(DefaultSQLExecutorCallback.java:57)
at org.apache.shardingsphere.infra.executor.kernel.ExecutorKernel.syncExecute(ExecutorKernel.java:99)
at org.apache.shardingsphere.infra.executor.kernel.ExecutorKernel.parallelExecute(ExecutorKernel.java:95)
at org.apache.shardingsphere.infra.executor.kernel.ExecutorKernel.execute(ExecutorKernel.java:78)
at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.SQLExecutor.execute(SQLExecutor.java:66)
at org.apache.shardingsphere.infra.executor.sql.resourced.jdbc.executor.SQLExecutor.execute(SQLExecutor.java:50)
at org.apache.shardingsphere.driver.executor.StatementExecutor.executeQuery(StatementExecutor.java:85)
at org.apache.shardingsphere.driver.jdbc.core.statement.ShardingSphereStatement.executeQuery(ShardingSphereStatement.java:127)
at org.apache.shardingsphere.dbtest.engine.dql.GeneralDQLIT.assertExecuteQueryForStatement(GeneralDQLIT.java:70)
at org.apache.shardingsphere.dbtest.engine.dql.GeneralDQLIT.assertExecuteQuery(GeneralDQLIT.java:60)
```
### Reason analyze (If you can)
`rule` and `start_point` columns doesn't exist in `t_order` table.
Since they're special column type, we can't add these columns for all types of database. Possible solution:
- Make `schema.xml > table-create > sql` support `db-types` definition, e.g. `db-types=PostgreSQL`, then the same name tables that have different columns will be created for different database.
### Steps to reproduce the behavior, such as: SQL to execute, sharding rule configuration, when exception occur etc.
### Example codes for reproduce this issue (such as a github link).
|
test
|
postgresql dedicated column missed in t order cause generaldqlit assertexecutequery failed bug report which version of shardingsphere did you use master branch snapshot which project did you use shardingsphere jdbc or shardingsphere proxy shardingsphere jdbc expected behavior related dql test case pass actual behavior test failed exception thrown org postgresql util psqlexception error column rule does not exist at org postgresql core queryexecutorimpl receiveerrorresponse queryexecutorimpl java at org postgresql core queryexecutorimpl processresults queryexecutorimpl java at org postgresql core queryexecutorimpl execute queryexecutorimpl java at org postgresql jdbc pgstatement executeinternal pgstatement java at org postgresql jdbc pgstatement execute pgstatement java at org postgresql jdbc pgstatement executewithflags pgstatement java at org postgresql jdbc pgstatement executecachedsql pgstatement java at org postgresql jdbc pgstatement executewithflags pgstatement java at org postgresql jdbc pgstatement executequery pgstatement java at com zaxxer hikari pool proxystatement executequery proxystatement java at com zaxxer hikari pool hikariproxystatement executequery hikariproxystatement java at org apache shardingsphere driver executor statementexecutor createqueryresult statementexecutor java at org apache shardingsphere driver executor statementexecutor executesql statementexecutor java at org apache shardingsphere driver executor statementexecutor executesql statementexecutor java at org apache shardingsphere infra executor sql resourced jdbc executor impl defaultsqlexecutorcallback defaultsqlexecutorcallback java at org apache shardingsphere infra executor sql resourced jdbc executor impl defaultsqlexecutorcallback execute defaultsqlexecutorcallback java at org apache shardingsphere infra executor kernel executorkernel syncexecute executorkernel java at org apache shardingsphere infra executor kernel executorkernel parallelexecute executorkernel java at org apache shardingsphere infra executor kernel executorkernel execute executorkernel java at org apache shardingsphere infra executor sql resourced jdbc executor sqlexecutor execute sqlexecutor java at org apache shardingsphere infra executor sql resourced jdbc executor sqlexecutor execute sqlexecutor java at org apache shardingsphere driver executor statementexecutor executequery statementexecutor java at org apache shardingsphere driver jdbc core statement shardingspherestatement executequery shardingspherestatement java at org apache shardingsphere dbtest engine dql generaldqlit assertexecutequeryforstatement generaldqlit java at org apache shardingsphere dbtest engine dql generaldqlit assertexecutequery generaldqlit java reason analyze if you can rule and start point columns doesn t exist in t order table since they re special column type we can t add these columns for all types of database possible solution make schema xml table create sql support db types definition e g db types postgresql then the same name tables that have different columns will be created for different database steps to reproduce the behavior such as sql to execute sharding rule configuration when exception occur etc example codes for reproduce this issue such as a github link
| 1
|
87,519
| 15,779,925,860
|
IssuesEvent
|
2021-04-01 09:18:35
|
AlexRogalskiy/gradle-java-sample
|
https://api.github.com/repos/AlexRogalskiy/gradle-java-sample
|
opened
|
CVE-2021-21346 (High) detected in xstream-1.4.10.jar
|
security vulnerability
|
## CVE-2021-21346 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.10.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p>
<p>Path to dependency file: gradle-java-sample/buildSrc/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.thoughtworks.xstream/xstream/1.4.10/dfecae23647abc9d9fd0416629a4213a3882b101/xstream-1.4.10.jar</p>
<p>
Dependency Hierarchy:
- gradle-versions-plugin-0.28.0.jar (Root Library)
- :x: **xstream-1.4.10.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/gradle-java-sample/commit/e537d2e240e1b0b48107d38039b89c5b5d6fd977">e537d2e240e1b0b48107d38039b89c5b5d6fd977</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to load and execute arbitrary code from a remote host only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21346>CVE-2021-21346</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-4hrm-m67v-5cxr">https://github.com/x-stream/xstream/security/advisories/GHSA-4hrm-m67v-5cxr</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-21346 (High) detected in xstream-1.4.10.jar - ## CVE-2021-21346 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.10.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p>
<p>Path to dependency file: gradle-java-sample/buildSrc/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.thoughtworks.xstream/xstream/1.4.10/dfecae23647abc9d9fd0416629a4213a3882b101/xstream-1.4.10.jar</p>
<p>
Dependency Hierarchy:
- gradle-versions-plugin-0.28.0.jar (Root Library)
- :x: **xstream-1.4.10.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/gradle-java-sample/commit/e537d2e240e1b0b48107d38039b89c5b5d6fd977">e537d2e240e1b0b48107d38039b89c5b5d6fd977</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to load and execute arbitrary code from a remote host only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21346>CVE-2021-21346</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-4hrm-m67v-5cxr">https://github.com/x-stream/xstream/security/advisories/GHSA-4hrm-m67v-5cxr</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in xstream jar cve high severity vulnerability vulnerable library xstream jar xstream is a serialization library from java objects to xml and back library home page a href path to dependency file gradle java sample buildsrc build gradle path to vulnerable library home wss scanner gradle caches modules files com thoughtworks xstream xstream xstream jar dependency hierarchy gradle versions plugin jar root library x xstream jar vulnerable library found in head commit a href vulnerability details xstream is a java library to serialize objects to xml and back again in xstream before version there is a vulnerability which may allow a remote attacker to load and execute arbitrary code from a remote host only by manipulating the processed input stream no user is affected who followed the recommendation to setup xstream s security framework with a whitelist limited to the minimal required types if you rely on xstream s default blacklist of the security framework you will have to use at least version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream step up your open source security game with whitesource
| 0
|
432,446
| 30,284,203,165
|
IssuesEvent
|
2023-07-08 13:07:19
|
Danfall369/Catalog-of-my-things
|
https://api.github.com/repos/Danfall369/Catalog-of-my-things
|
closed
|
[2.5pt] (Group Task) General Structure
|
documentation
|
# Group Task
## Create `Item` class: Following the [UML Diagram](https://github.com/microverseinc/curriculum-ruby/blob/main/group-capstone/images/catalog_of_my_things.png)
- [x] Create Item class in a separate .rb file.
- [x] All Item class properties visible in the diagram should be defined and set up in the constructor method. Exception: properties for the 1-to-many relationships should NOT be set in the constructor method. Instead, they should have a custom setter method created.
- [x] Add all methods visible in the diagram.
## Implement methods
#### _can_be_archived?() in the Item class_
- - [x] Should return true if published_date is older than 10 years.
- - [x] Otherwise, it should return false.
#### _Move_to_archive() in the Item class_
- - [x] Should reuse can_be_archived?() method.
- - [x] Should change the archived property to true if the result of the can_be_archived?() method is true.
- - [x] Should do nothing if the result of the can_be_archived?() method is false.
## Create Basic UI
- [x] Create a main.rb file that will serve as your console app entry-point.
- [x] Present the user with a list of options to perform.
- [x] Let users choose an option.
- [x] If needed, ask for parameters for the option.
- [x] Have a way to quit the app.
|
1.0
|
[2.5pt] (Group Task) General Structure - # Group Task
## Create `Item` class: Following the [UML Diagram](https://github.com/microverseinc/curriculum-ruby/blob/main/group-capstone/images/catalog_of_my_things.png)
- [x] Create Item class in a separate .rb file.
- [x] All Item class properties visible in the diagram should be defined and set up in the constructor method. Exception: properties for the 1-to-many relationships should NOT be set in the constructor method. Instead, they should have a custom setter method created.
- [x] Add all methods visible in the diagram.
## Implement methods
#### _can_be_archived?() in the Item class_
- - [x] Should return true if published_date is older than 10 years.
- - [x] Otherwise, it should return false.
#### _Move_to_archive() in the Item class_
- - [x] Should reuse can_be_archived?() method.
- - [x] Should change the archived property to true if the result of the can_be_archived?() method is true.
- - [x] Should do nothing if the result of the can_be_archived?() method is false.
## Create Basic UI
- [x] Create a main.rb file that will serve as your console app entry-point.
- [x] Present the user with a list of options to perform.
- [x] Let users choose an option.
- [x] If needed, ask for parameters for the option.
- [x] Have a way to quit the app.
|
non_test
|
group task general structure group task create item class following the create item class in a separate rb file all item class properties visible in the diagram should be defined and set up in the constructor method exception properties for the to many relationships should not be set in the constructor method instead they should have a custom setter method created add all methods visible in the diagram implement methods can be archived in the item class should return true if published date is older than years otherwise it should return false move to archive in the item class should reuse can be archived method should change the archived property to true if the result of the can be archived method is true should do nothing if the result of the can be archived method is false create basic ui create a main rb file that will serve as your console app entry point present the user with a list of options to perform let users choose an option if needed ask for parameters for the option have a way to quit the app
| 0
|
321,243
| 27,517,311,502
|
IssuesEvent
|
2023-03-06 12:53:10
|
IntellectualSites/FastAsyncWorldEdit
|
https://api.github.com/repos/IntellectualSites/FastAsyncWorldEdit
|
closed
|
Off by one error for negative coordinates when using -r with //deform
|
Requires Testing
|
### Server Implementation
Paper
### Server Version
1.19.2
### Describe the bug
When using `//deform -r` negative coordinates are off by one.
### To Reproduce
Select a region around 0,0
Do `//deform -r 0`
Observe how parts of the region that are in the negatives within the game's coordinate system move around.
### Expected behaviour
Nothing should change as we do not modify the x,y,z variables.
### Screenshots / Videos
Input and also expected output

Current output

### Error log (if applicable)
_No response_
### Fawe Debugpaste
https://athion.net/ISPaster/paste/view/4e4bf7430f1b4ac6847ad782e6904782
### Fawe Version
FastAsyncWorldEdit version 2.5.2-SNAPSHOT-349
### Checklist
- [X] I have included a Fawe debugpaste.
- [X] I am using the newest build from https://ci.athion.net/job/FastAsyncWorldEdit/ and the issue still persists.
### Anything else?
When I do `//deform -r x=0` it fetches block at x=0.
When I do `//deform -r x=-1` it fetches block at x=0.
When I do `//deform -r x=-2` it fetches block at x=-1.
When I do `//deform -r x=-3` it fetches block at x=-2.
The same thing happens an all three axes.
|
1.0
|
Off by one error for negative coordinates when using -r with //deform - ### Server Implementation
Paper
### Server Version
1.19.2
### Describe the bug
When using `//deform -r` negative coordinates are off by one.
### To Reproduce
Select a region around 0,0
Do `//deform -r 0`
Observe how parts of the region that are in the negatives within the game's coordinate system move around.
### Expected behaviour
Nothing should change as we do not modify the x,y,z variables.
### Screenshots / Videos
Input and also expected output

Current output

### Error log (if applicable)
_No response_
### Fawe Debugpaste
https://athion.net/ISPaster/paste/view/4e4bf7430f1b4ac6847ad782e6904782
### Fawe Version
FastAsyncWorldEdit version 2.5.2-SNAPSHOT-349
### Checklist
- [X] I have included a Fawe debugpaste.
- [X] I am using the newest build from https://ci.athion.net/job/FastAsyncWorldEdit/ and the issue still persists.
### Anything else?
When I do `//deform -r x=0` it fetches block at x=0.
When I do `//deform -r x=-1` it fetches block at x=0.
When I do `//deform -r x=-2` it fetches block at x=-1.
When I do `//deform -r x=-3` it fetches block at x=-2.
The same thing happens an all three axes.
|
test
|
off by one error for negative coordinates when using r with deform server implementation paper server version describe the bug when using deform r negative coordinates are off by one to reproduce select a region around do deform r observe how parts of the region that are in the negatives within the game s coordinate system move around expected behaviour nothing should change as we do not modify the x y z variables screenshots videos input and also expected output current output error log if applicable no response fawe debugpaste fawe version fastasyncworldedit version snapshot checklist i have included a fawe debugpaste i am using the newest build from and the issue still persists anything else when i do deform r x it fetches block at x when i do deform r x it fetches block at x when i do deform r x it fetches block at x when i do deform r x it fetches block at x the same thing happens an all three axes
| 1
|
63,646
| 6,877,413,596
|
IssuesEvent
|
2017-11-20 07:51:34
|
alibaba/pouch
|
https://api.github.com/repos/alibaba/pouch
|
closed
|
[bug]The swagger spec at "swagger.yml" is invalid against swagger specification 2.0
|
areas/test kind/bug
|
**Issue Description**
```
The swagger spec at "swagger.yml" is invalid against swagger specification 2.0
- definitions.ContainerConfig.properties.Cmd in body must be of type array
- definitions.ContainerConfig.properties.Entrypoint in body must be of type array
```
**How to reproduce it (as minimally and precisely as possible)**:
```
cd apis
swagger validate swagger.yml
```
|
1.0
|
[bug]The swagger spec at "swagger.yml" is invalid against swagger specification 2.0 - **Issue Description**
```
The swagger spec at "swagger.yml" is invalid against swagger specification 2.0
- definitions.ContainerConfig.properties.Cmd in body must be of type array
- definitions.ContainerConfig.properties.Entrypoint in body must be of type array
```
**How to reproduce it (as minimally and precisely as possible)**:
```
cd apis
swagger validate swagger.yml
```
|
test
|
the swagger spec at swagger yml is invalid against swagger specification issue description the swagger spec at swagger yml is invalid against swagger specification definitions containerconfig properties cmd in body must be of type array definitions containerconfig properties entrypoint in body must be of type array how to reproduce it as minimally and precisely as possible cd apis swagger validate swagger yml
| 1
|
701,998
| 24,118,661,778
|
IssuesEvent
|
2022-09-20 16:40:04
|
IslasGECI/dimorfismo
|
https://api.github.com/repos/IslasGECI/dimorfismo
|
reopened
|
Los nombres de las propiedades en `logistic_model_parameters.json` están en _spanglish_
|
Status: Available Priority: Low Type: Maintenance wontfix
|
Abajo se puede ver que nombres de las propiedades están en _spanglish_:
```
{
"parametrosNormalizacion": {
"valorMinimo": {
"longitudCraneo": [3.7037],
"altoPico": [165.58],
"longitudPico": [29.67],
"tarso": [101.78]
},
"valorMaximo": {
"longitudCraneo": [83.18],
"altoPico": [46.5],
"longitudPico": [193.22],
"tarso": [35.44]
}
},
"parametrosModelo": [
{
"Variables": "(Intercept)",
"Estimate": -18.948,
"_row": "(Intercept)"
},
{
"Variables": "longitudCraneo",
"Estimate": 6.576,
"_row": "longitudCraneo"
},
{
"Variables": "altoPico",
"Estimate": 8.816,
"_row": "altoPico"
},
{
"Variables": "longitudPico",
"Estimate": 7.172,
"_row": "longitudPico"
},
{
"Variables": "tarso",
"Estimate": 5.726,
"_row": "tarso"
}
]
}
```
|
1.0
|
Los nombres de las propiedades en `logistic_model_parameters.json` están en _spanglish_ - Abajo se puede ver que nombres de las propiedades están en _spanglish_:
```
{
"parametrosNormalizacion": {
"valorMinimo": {
"longitudCraneo": [3.7037],
"altoPico": [165.58],
"longitudPico": [29.67],
"tarso": [101.78]
},
"valorMaximo": {
"longitudCraneo": [83.18],
"altoPico": [46.5],
"longitudPico": [193.22],
"tarso": [35.44]
}
},
"parametrosModelo": [
{
"Variables": "(Intercept)",
"Estimate": -18.948,
"_row": "(Intercept)"
},
{
"Variables": "longitudCraneo",
"Estimate": 6.576,
"_row": "longitudCraneo"
},
{
"Variables": "altoPico",
"Estimate": 8.816,
"_row": "altoPico"
},
{
"Variables": "longitudPico",
"Estimate": 7.172,
"_row": "longitudPico"
},
{
"Variables": "tarso",
"Estimate": 5.726,
"_row": "tarso"
}
]
}
```
|
non_test
|
los nombres de las propiedades en logistic model parameters json están en spanglish abajo se puede ver que nombres de las propiedades están en spanglish parametrosnormalizacion valorminimo longitudcraneo altopico longitudpico tarso valormaximo longitudcraneo altopico longitudpico tarso parametrosmodelo variables intercept estimate row intercept variables longitudcraneo estimate row longitudcraneo variables altopico estimate row altopico variables longitudpico estimate row longitudpico variables tarso estimate row tarso
| 0
|
330,550
| 28,439,086,148
|
IssuesEvent
|
2023-04-15 17:26:59
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed
|
C-test-failure O-robot branch-release-23.1
|
pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test.TestSQLStatsCompactor [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9624052?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9624052?buildTab=artifacts#/) on release-23.1 @ [84905b8396e7cd7ef999b264faaafe13110d631a](https://github.com/cockroachdb/cockroach/commits/84905b8396e7cd7ef999b264faaafe13110d631a):
```
=== RUN TestSQLStatsCompactor
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/9a30d5271a65be6b3f2aa1433be9011f/logTestSQLStatsCompactor803069076
test_log_scope.go:79: use -show-logs to present logs inline
=== CONT TestSQLStatsCompactor
compaction_test.go:262: -- test log scope end --
test logs left over in: /artifacts/tmp/_tmp/9a30d5271a65be6b3f2aa1433be9011f/logTestSQLStatsCompactor803069076
--- FAIL: TestSQLStatsCompactor (137.60s)
=== RUN TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=205/rowsDeletePerTxn=0
compaction_test.go:227:
Error Trace: pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:227
pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:234
Error: Not equal:
expected: 25
actual : 13
Test: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=205/rowsDeletePerTxn=0
Messages: expected 25 number of wide scans issued, but 13 number of wide scan issued
--- FAIL: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=205/rowsDeletePerTxn=0 (16.75s)
=== RUN TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=1024
compaction_test.go:227:
Error Trace: pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:227
pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:234
Error: Not equal:
expected: 32
actual : 16
Test: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=1024
Messages: expected 32 number of wide scans issued, but 16 number of wide scan issued
--- FAIL: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=1024 (20.04s)
=== RUN TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=2
compaction_test.go:227:
Error Trace: pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:227
pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:234
Error: Not equal:
expected: 32
actual : 24
Test: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=2
Messages: expected 32 number of wide scans issued, but 24 number of wide scan issued
--- FAIL: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=2 (32.50s)
```
<p>Parameters: <code>TAGS=bazel,gss,deadlock</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #101361 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed [C-test-failure O-robot T-cluster-observability branch-release-23.1.0]
- #94880 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed [C-test-failure O-robot branch-release-22.2]
- #80442 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed [C-test-failure O-robot branch-master]
</p>
</details>
/cc @cockroachdb/cluster-observability
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSQLStatsCompactor.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed - pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test.TestSQLStatsCompactor [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9624052?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9624052?buildTab=artifacts#/) on release-23.1 @ [84905b8396e7cd7ef999b264faaafe13110d631a](https://github.com/cockroachdb/cockroach/commits/84905b8396e7cd7ef999b264faaafe13110d631a):
```
=== RUN TestSQLStatsCompactor
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/9a30d5271a65be6b3f2aa1433be9011f/logTestSQLStatsCompactor803069076
test_log_scope.go:79: use -show-logs to present logs inline
=== CONT TestSQLStatsCompactor
compaction_test.go:262: -- test log scope end --
test logs left over in: /artifacts/tmp/_tmp/9a30d5271a65be6b3f2aa1433be9011f/logTestSQLStatsCompactor803069076
--- FAIL: TestSQLStatsCompactor (137.60s)
=== RUN TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=205/rowsDeletePerTxn=0
compaction_test.go:227:
Error Trace: pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:227
pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:234
Error: Not equal:
expected: 25
actual : 13
Test: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=205/rowsDeletePerTxn=0
Messages: expected 25 number of wide scans issued, but 13 number of wide scan issued
--- FAIL: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=205/rowsDeletePerTxn=0 (16.75s)
=== RUN TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=1024
compaction_test.go:227:
Error Trace: pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:227
pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:234
Error: Not equal:
expected: 32
actual : 16
Test: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=1024
Messages: expected 32 number of wide scans issued, but 16 number of wide scan issued
--- FAIL: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=1024 (20.04s)
=== RUN TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=2
compaction_test.go:227:
Error Trace: pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:227
pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test_test/pkg/sql/sqlstats/persistedsqlstats/compaction_test.go:234
Error: Not equal:
expected: 32
actual : 24
Test: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=2
Messages: expected 32 number of wide scans issued, but 24 number of wide scan issued
--- FAIL: TestSQLStatsCompactor/stmtCount=200/maxPersistedRowLimit=40/rowsDeletePerTxn=2 (32.50s)
```
<p>Parameters: <code>TAGS=bazel,gss,deadlock</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #101361 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed [C-test-failure O-robot T-cluster-observability branch-release-23.1.0]
- #94880 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed [C-test-failure O-robot branch-release-22.2]
- #80442 pkg/sql/sqlstats/persistedsqlstats/persistedsqlstats_test: TestSQLStatsCompactor failed [C-test-failure O-robot branch-master]
</p>
</details>
/cc @cockroachdb/cluster-observability
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSQLStatsCompactor.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
pkg sql sqlstats persistedsqlstats persistedsqlstats test testsqlstatscompactor failed pkg sql sqlstats persistedsqlstats persistedsqlstats test testsqlstatscompactor with on release run testsqlstatscompactor test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline cont testsqlstatscompactor compaction test go test log scope end test logs left over in artifacts tmp tmp fail testsqlstatscompactor run testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn compaction test go error trace pkg sql sqlstats persistedsqlstats persistedsqlstats test test pkg sql sqlstats persistedsqlstats compaction test go pkg sql sqlstats persistedsqlstats persistedsqlstats test test pkg sql sqlstats persistedsqlstats compaction test go error not equal expected actual test testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn messages expected number of wide scans issued but number of wide scan issued fail testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn run testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn compaction test go error trace pkg sql sqlstats persistedsqlstats persistedsqlstats test test pkg sql sqlstats persistedsqlstats compaction test go pkg sql sqlstats persistedsqlstats persistedsqlstats test test pkg sql sqlstats persistedsqlstats compaction test go error not equal expected actual test testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn messages expected number of wide scans issued but number of wide scan issued fail testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn run testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn compaction test go error trace pkg sql sqlstats persistedsqlstats persistedsqlstats test test pkg sql sqlstats persistedsqlstats compaction test go pkg sql sqlstats persistedsqlstats persistedsqlstats test test pkg sql sqlstats persistedsqlstats compaction test go error not equal expected actual test testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn messages expected number of wide scans issued but number of wide scan issued fail testsqlstatscompactor stmtcount maxpersistedrowlimit rowsdeletepertxn parameters tags bazel gss deadlock help see also same failure on other branches pkg sql sqlstats persistedsqlstats persistedsqlstats test testsqlstatscompactor failed pkg sql sqlstats persistedsqlstats persistedsqlstats test testsqlstatscompactor failed pkg sql sqlstats persistedsqlstats persistedsqlstats test testsqlstatscompactor failed cc cockroachdb cluster observability
| 1
|
98,003
| 4,015,861,783
|
IssuesEvent
|
2016-05-15 07:04:30
|
oshri551/angular2Project
|
https://api.github.com/repos/oshri551/angular2Project
|
closed
|
Add Angular material 2
|
effort1: easy (2 hours) priority: normal Task
|
## Todos:
* Add dependncy to package.json
* Add angular material to vendors.ts
|
1.0
|
Add Angular material 2 - ## Todos:
* Add dependncy to package.json
* Add angular material to vendors.ts
|
non_test
|
add angular material todos add dependncy to package json add angular material to vendors ts
| 0
|
270,091
| 23,490,515,867
|
IssuesEvent
|
2022-08-17 18:14:31
|
tijlleenders/ZinZen
|
https://api.github.com/repos/tijlleenders/ZinZen
|
opened
|
Update the cypress test for Goals header
|
UI test
|
We are creating a change in the header by replacing the icon at the top right of the header. Cypress test has to be modified if there is PR that is targeting to resolve #591
|
1.0
|
Update the cypress test for Goals header - We are creating a change in the header by replacing the icon at the top right of the header. Cypress test has to be modified if there is PR that is targeting to resolve #591
|
test
|
update the cypress test for goals header we are creating a change in the header by replacing the icon at the top right of the header cypress test has to be modified if there is pr that is targeting to resolve
| 1
|
158,513
| 12,417,829,741
|
IssuesEvent
|
2020-05-22 21:48:20
|
rancher/dashboard
|
https://api.github.com/repos/rancher/dashboard
|
closed
|
Design Issues - Dropdown
|
[zube]: To Test
|
When there is no label for a dropdown, the height of the dropdown is different than with a label. Should we make all dropdowns the same height regardless of label? Or will we require a label for every dropdown?
This is from the Environment Variables from a Resource on the workloads page inside the Command tab.
<img width="1072" alt="Screen Shot 2020-05-04 at 4 53 54 PM" src="https://user-images.githubusercontent.com/11410997/81024476-64406800-8e28-11ea-8c45-d03d052b59ed.png">
|
1.0
|
Design Issues - Dropdown - When there is no label for a dropdown, the height of the dropdown is different than with a label. Should we make all dropdowns the same height regardless of label? Or will we require a label for every dropdown?
This is from the Environment Variables from a Resource on the workloads page inside the Command tab.
<img width="1072" alt="Screen Shot 2020-05-04 at 4 53 54 PM" src="https://user-images.githubusercontent.com/11410997/81024476-64406800-8e28-11ea-8c45-d03d052b59ed.png">
|
test
|
design issues dropdown when there is no label for a dropdown the height of the dropdown is different than with a label should we make all dropdowns the same height regardless of label or will we require a label for every dropdown this is from the environment variables from a resource on the workloads page inside the command tab img width alt screen shot at pm src
| 1
|
5,181
| 2,572,399,558
|
IssuesEvent
|
2015-02-10 22:17:16
|
jasonhall/jasonhall
|
https://api.github.com/repos/jasonhall/jasonhall
|
opened
|
has labels
|
auto-migrated Priority-Low Type-Enhancement
|
```
this issue has labels that change with comments
```
-----
Original issue reported on code.google.com by jasonhall@google.com on 11 Dec 2014 at 8:07
* Blocked on: #8
|
1.0
|
has labels - ```
this issue has labels that change with comments
```
-----
Original issue reported on code.google.com by jasonhall@google.com on 11 Dec 2014 at 8:07
* Blocked on: #8
|
non_test
|
has labels this issue has labels that change with comments original issue reported on code google com by jasonhall google com on dec at blocked on
| 0
|
222,797
| 24,711,330,791
|
IssuesEvent
|
2022-10-20 01:14:05
|
theWhiteFox/ion-vue-app
|
https://api.github.com/repos/theWhiteFox/ion-vue-app
|
closed
|
WS-2020-0042 (High) detected in acorn-5.7.4.tgz - autoclosed
|
security vulnerability
|
## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.4.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/jsdom/node_modules/acorn/package.json</p>
<p>
Dependency Hierarchy:
- cli-plugin-unit-jest-4.5.13.tgz (Root Library)
- jest-24.9.0.tgz
- jest-cli-24.9.0.tgz
- jest-config-24.9.0.tgz
- jest-environment-jsdom-24.9.0.tgz
- jsdom-11.12.0.tgz
- :x: **acorn-5.7.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/theWhiteFox/ion-vue-app/commit/5bed41ba25cd539697eec7f8a2456a9190a81333">5bed41ba25cd539697eec7f8a2456a9190a81333</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-01
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-01</p>
<p>Fix Resolution (acorn): 6.4.1</p>
<p>Direct dependency fix Resolution (@vue/cli-plugin-unit-jest): 5.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2020-0042 (High) detected in acorn-5.7.4.tgz - autoclosed - ## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.4.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/jsdom/node_modules/acorn/package.json</p>
<p>
Dependency Hierarchy:
- cli-plugin-unit-jest-4.5.13.tgz (Root Library)
- jest-24.9.0.tgz
- jest-cli-24.9.0.tgz
- jest-config-24.9.0.tgz
- jest-environment-jsdom-24.9.0.tgz
- jsdom-11.12.0.tgz
- :x: **acorn-5.7.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/theWhiteFox/ion-vue-app/commit/5bed41ba25cd539697eec7f8a2456a9190a81333">5bed41ba25cd539697eec7f8a2456a9190a81333</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-01
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-01</p>
<p>Fix Resolution (acorn): 6.4.1</p>
<p>Direct dependency fix Resolution (@vue/cli-plugin-unit-jest): 5.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
ws high detected in acorn tgz autoclosed ws high severity vulnerability vulnerable library acorn tgz ecmascript parser library home page a href path to dependency file package json path to vulnerable library node modules jsdom node modules acorn package json dependency hierarchy cli plugin unit jest tgz root library jest tgz jest cli tgz jest config tgz jest environment jsdom tgz jsdom tgz x acorn tgz vulnerable library found in head commit a href found in base branch main vulnerability details acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution acorn direct dependency fix resolution vue cli plugin unit jest step up your open source security game with mend
| 0
|
162,242
| 12,640,436,819
|
IssuesEvent
|
2020-06-16 03:01:00
|
microsoft/azure-tools-for-java
|
https://api.github.com/repos/microsoft/azure-tools-for-java
|
closed
|
[intelliJ][Spark on Cosmos Serverless]Remotely run failed sometimes.
|
External Dependency HDInsight IntelliJ Internal Test Security Rating -- LOW
|
Build:
develop 1247
Repro Steps:
1. Create a Spark on Cosmos Serverless configuration.
2. Remotely run it.
Result:
Job run failed.

Frequency:
Sometimes.
|
1.0
|
[intelliJ][Spark on Cosmos Serverless]Remotely run failed sometimes. - Build:
develop 1247
Repro Steps:
1. Create a Spark on Cosmos Serverless configuration.
2. Remotely run it.
Result:
Job run failed.

Frequency:
Sometimes.
|
test
|
remotely run failed sometimes build develop repro steps create a spark on cosmos serverless configuration remotely run it result job run failed frequency sometimes
| 1
|
216,420
| 16,761,160,001
|
IssuesEvent
|
2021-06-13 20:19:53
|
snowpackjs/astro
|
https://api.github.com/repos/snowpackjs/astro
|
closed
|
Integration tests
|
testing
|
We've had issues in the past where some bugs (especially those related to module resolution) are not caught by integration tests due to have we have the monorepo set up and the use of snowpack's `workspaceRoot`.
We would like to have some integration tests that run to smoke test released versions. They might do:
1. Run `npm init astro` and step through the questions.
2. Start up `astro dev` and make sure it doesn't crash. Fetch the HTML from a few routes and verify its content.
3. Run the build `astro build` and make sure it doesn't crash. Check that the correct output was created.
|
1.0
|
Integration tests - We've had issues in the past where some bugs (especially those related to module resolution) are not caught by integration tests due to have we have the monorepo set up and the use of snowpack's `workspaceRoot`.
We would like to have some integration tests that run to smoke test released versions. They might do:
1. Run `npm init astro` and step through the questions.
2. Start up `astro dev` and make sure it doesn't crash. Fetch the HTML from a few routes and verify its content.
3. Run the build `astro build` and make sure it doesn't crash. Check that the correct output was created.
|
test
|
integration tests we ve had issues in the past where some bugs especially those related to module resolution are not caught by integration tests due to have we have the monorepo set up and the use of snowpack s workspaceroot we would like to have some integration tests that run to smoke test released versions they might do run npm init astro and step through the questions start up astro dev and make sure it doesn t crash fetch the html from a few routes and verify its content run the build astro build and make sure it doesn t crash check that the correct output was created
| 1
|
33,100
| 4,807,065,395
|
IssuesEvent
|
2016-11-02 20:21:41
|
moment/moment
|
https://api.github.com/repos/moment/moment
|
closed
|
3 tests failed. diff:diff across DST (2336.17) diff:diff across DST (2336.19) relative time:custom rounding (2630.3)
|
Unit Test Failed
|
### Client info
```
Date String : Tue Nov 01 2016 13:58:48 GMT-0600 (CST)
Locale String : 11/1/2016, 1:58:48 PM
Offset : 360
User Agent : Mozilla/5.0 (Macintosh; Intel Mac OS X 10_12_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/54.0.2840.71 Safari/537.36
Moment Version: 2.15.2
```
====
### diff:diff across DST (2336.17)
month diff across DST, lower bound
```javascript
// Expected true
// Actual false
false === true
```
====
### diff:diff across DST (2336.19)
year diff across DST, lower bound
```javascript
// Expected true
// Actual false
false === true
```
====
### relative time:custom rounding (2630.3)
Round down towards the nearest day
```javascript
// Expected in 30 days
// Actual in a month
"in a month" === "in 30 days"
```
|
1.0
|
3 tests failed. diff:diff across DST (2336.17) diff:diff across DST (2336.19) relative time:custom rounding (2630.3) - ### Client info
```
Date String : Tue Nov 01 2016 13:58:48 GMT-0600 (CST)
Locale String : 11/1/2016, 1:58:48 PM
Offset : 360
User Agent : Mozilla/5.0 (Macintosh; Intel Mac OS X 10_12_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/54.0.2840.71 Safari/537.36
Moment Version: 2.15.2
```
====
### diff:diff across DST (2336.17)
month diff across DST, lower bound
```javascript
// Expected true
// Actual false
false === true
```
====
### diff:diff across DST (2336.19)
year diff across DST, lower bound
```javascript
// Expected true
// Actual false
false === true
```
====
### relative time:custom rounding (2630.3)
Round down towards the nearest day
```javascript
// Expected in 30 days
// Actual in a month
"in a month" === "in 30 days"
```
|
test
|
tests failed diff diff across dst diff diff across dst relative time custom rounding client info date string tue nov gmt cst locale string pm offset user agent mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari moment version diff diff across dst month diff across dst lower bound javascript expected true actual false false true diff diff across dst year diff across dst lower bound javascript expected true actual false false true relative time custom rounding round down towards the nearest day javascript expected in days actual in a month in a month in days
| 1
|
373,987
| 26,094,895,141
|
IssuesEvent
|
2022-12-26 17:40:08
|
StraykerPL/RockPaperScissors
|
https://api.github.com/repos/StraykerPL/RockPaperScissors
|
closed
|
Feature: Add Testing label to repo
|
documentation
|
**Is your feature request related to a problem? Please describe.**
There's no `testing` label for checking general parts of product.
**Describe the solution you'd like**
Add `testing` label in GitHub, with description `This needs to be tested`.
**Describe alternatives you've considered**
**Additional context**
|
1.0
|
Feature: Add Testing label to repo - **Is your feature request related to a problem? Please describe.**
There's no `testing` label for checking general parts of product.
**Describe the solution you'd like**
Add `testing` label in GitHub, with description `This needs to be tested`.
**Describe alternatives you've considered**
**Additional context**
|
non_test
|
feature add testing label to repo is your feature request related to a problem please describe there s no testing label for checking general parts of product describe the solution you d like add testing label in github with description this needs to be tested describe alternatives you ve considered additional context
| 0
|
45,395
| 13,110,398,729
|
IssuesEvent
|
2020-08-04 20:35:04
|
mwilliams7197/jest-environment-serverless
|
https://api.github.com/repos/mwilliams7197/jest-environment-serverless
|
opened
|
CVE-2019-20149 (High) detected in kind-of-6.0.2.tgz
|
security vulnerability
|
## CVE-2019-20149 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kind-of-6.0.2.tgz</b></p></summary>
<p>Get the native type of a value.</p>
<p>Library home page: <a href="https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz">https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/jest-environment-serverless/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/jest-environment-serverless/node_modules/kind-of/package.json</p>
<p>
Dependency Hierarchy:
- jest-config-24.8.0.tgz (Root Library)
- micromatch-3.1.10.tgz
- :x: **kind-of-6.0.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mwilliams7197/jest-environment-serverless/commit/02682832cf2379d2ecc74a2ddc62a35db5341137">02682832cf2379d2ecc74a2ddc62a35db5341137</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by 'constructor': {'name':'Symbol'}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result.
<p>Publish Date: 2019-12-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149>CVE-2019-20149</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149</a></p>
<p>Release Date: 2019-12-30</p>
<p>Fix Resolution: 6.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"kind-of","packageVersion":"6.0.2","isTransitiveDependency":true,"dependencyTree":"jest-config:24.8.0;micromatch:3.1.10;kind-of:6.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.3"}],"vulnerabilityIdentifier":"CVE-2019-20149","vulnerabilityDetails":"ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by \u0027constructor\u0027: {\u0027name\u0027:\u0027Symbol\u0027}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-20149 (High) detected in kind-of-6.0.2.tgz - ## CVE-2019-20149 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kind-of-6.0.2.tgz</b></p></summary>
<p>Get the native type of a value.</p>
<p>Library home page: <a href="https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz">https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/jest-environment-serverless/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/jest-environment-serverless/node_modules/kind-of/package.json</p>
<p>
Dependency Hierarchy:
- jest-config-24.8.0.tgz (Root Library)
- micromatch-3.1.10.tgz
- :x: **kind-of-6.0.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mwilliams7197/jest-environment-serverless/commit/02682832cf2379d2ecc74a2ddc62a35db5341137">02682832cf2379d2ecc74a2ddc62a35db5341137</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by 'constructor': {'name':'Symbol'}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result.
<p>Publish Date: 2019-12-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149>CVE-2019-20149</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149</a></p>
<p>Release Date: 2019-12-30</p>
<p>Fix Resolution: 6.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"kind-of","packageVersion":"6.0.2","isTransitiveDependency":true,"dependencyTree":"jest-config:24.8.0;micromatch:3.1.10;kind-of:6.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.3"}],"vulnerabilityIdentifier":"CVE-2019-20149","vulnerabilityDetails":"ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by \u0027constructor\u0027: {\u0027name\u0027:\u0027Symbol\u0027}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve high detected in kind of tgz cve high severity vulnerability vulnerable library kind of tgz get the native type of a value library home page a href path to dependency file tmp ws scm jest environment serverless package json path to vulnerable library tmp ws scm jest environment serverless node modules kind of package json dependency hierarchy jest config tgz root library micromatch tgz x kind of tgz vulnerable library found in head commit a href vulnerability details ctorname in index js in kind of allows external user input to overwrite certain internal attributes via a conflicting name as demonstrated by constructor name symbol hence a crafted payload can overwrite this builtin attribute to manipulate the type detection result publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails ctorname in index js in kind of allows external user input to overwrite certain internal attributes via a conflicting name as demonstrated by hence a crafted payload can overwrite this builtin attribute to manipulate the type detection result vulnerabilityurl
| 0
|
279,582
| 24,236,252,614
|
IssuesEvent
|
2022-09-26 23:44:44
|
microsoft/vscode-python
|
https://api.github.com/repos/microsoft/vscode-python
|
opened
|
Test finalized API for python environments
|
testplan-item
|
Refs: https://github.com/microsoft/vscode-python/issues/19101
- [ ] anyOS: @rchiodo
- [ ] anyOS: @DonJayamanne
Complexity: 5
---
### Requirements
1. Use Insiders build of the python extension
### Test APIs exposed via `ProposedExtensionAPI`
See `ProposedExtensionAPI` for APIs to test: https://github.com/microsoft/vscode-python/blob/0f045780e58e190e57411f4ac1227cda97c4fde2/src/client/proposedApiTypes.ts#L8
Example usage:
Copy over contents of https://github.com/microsoft/vscode-python/blob/main/src/client/proposedApiTypes.ts as needed.
```typescript
const extension = extensions.getExtension('ms-python.python');
if (extension) {
if (!extension.isActive) {
await extension.activate();
}
const api: IExtensionApi & IProposedExtensionAPI = extension.exports as IExtensionApi & IProposedExtensionAPI;
if (api.environment) {
const envID = api.environment.getActiveEnvironmentId();
}
}
```
|
1.0
|
Test finalized API for python environments - Refs: https://github.com/microsoft/vscode-python/issues/19101
- [ ] anyOS: @rchiodo
- [ ] anyOS: @DonJayamanne
Complexity: 5
---
### Requirements
1. Use Insiders build of the python extension
### Test APIs exposed via `ProposedExtensionAPI`
See `ProposedExtensionAPI` for APIs to test: https://github.com/microsoft/vscode-python/blob/0f045780e58e190e57411f4ac1227cda97c4fde2/src/client/proposedApiTypes.ts#L8
Example usage:
Copy over contents of https://github.com/microsoft/vscode-python/blob/main/src/client/proposedApiTypes.ts as needed.
```typescript
const extension = extensions.getExtension('ms-python.python');
if (extension) {
if (!extension.isActive) {
await extension.activate();
}
const api: IExtensionApi & IProposedExtensionAPI = extension.exports as IExtensionApi & IProposedExtensionAPI;
if (api.environment) {
const envID = api.environment.getActiveEnvironmentId();
}
}
```
|
test
|
test finalized api for python environments refs anyos rchiodo anyos donjayamanne complexity requirements use insiders build of the python extension test apis exposed via proposedextensionapi see proposedextensionapi for apis to test example usage copy over contents of as needed typescript const extension extensions getextension ms python python if extension if extension isactive await extension activate const api iextensionapi iproposedextensionapi extension exports as iextensionapi iproposedextensionapi if api environment const envid api environment getactiveenvironmentid
| 1
|
65,130
| 8,788,950,967
|
IssuesEvent
|
2018-12-21 00:43:11
|
mozilla/dinobuildr
|
https://api.github.com/repos/mozilla/dinobuildr
|
opened
|
Epic: Clean up documentation
|
documentation enhancement
|
The documentation needs some TLC.
#53
#59
#145
#68
All of these issues are docs related. If we add more before I finish the above list let's add them so there's an easy place to track all of the things we need to update.
|
1.0
|
Epic: Clean up documentation - The documentation needs some TLC.
#53
#59
#145
#68
All of these issues are docs related. If we add more before I finish the above list let's add them so there's an easy place to track all of the things we need to update.
|
non_test
|
epic clean up documentation the documentation needs some tlc all of these issues are docs related if we add more before i finish the above list let s add them so there s an easy place to track all of the things we need to update
| 0
|
181,537
| 6,661,863,948
|
IssuesEvent
|
2017-10-02 10:32:05
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
mail.google.com - site is not usable
|
browser-firefox priority-critical status-needsinfo status-needstriage
|
<!-- @browser: Firefox 58.0 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.12; rv:58.0) Gecko/20100101 Firefox/58.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://mail.google.com/mail/u/0/
**Browser / Version**: Firefox 58.0
**Operating System**: Mac OS X 10.12
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Gmail never completes loading unless you activate legacy basic HTML mode. Tested it in Safari and it loads fine.
**Steps to Reproduce**:
layout.css.servo.enabled: true
[](https://webcompat.com/uploads/2017/9/d7b49bbd-7f9c-4e4e-bd33-dbed5e8823e8.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
mail.google.com - site is not usable - <!-- @browser: Firefox 58.0 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.12; rv:58.0) Gecko/20100101 Firefox/58.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://mail.google.com/mail/u/0/
**Browser / Version**: Firefox 58.0
**Operating System**: Mac OS X 10.12
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Gmail never completes loading unless you activate legacy basic HTML mode. Tested it in Safari and it loads fine.
**Steps to Reproduce**:
layout.css.servo.enabled: true
[](https://webcompat.com/uploads/2017/9/d7b49bbd-7f9c-4e4e-bd33-dbed5e8823e8.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
mail google com site is not usable url browser version firefox operating system mac os x tested another browser yes problem type site is not usable description gmail never completes loading unless you activate legacy basic html mode tested it in safari and it loads fine steps to reproduce layout css servo enabled true from with ❤️
| 0
|
537,302
| 15,726,840,891
|
IssuesEvent
|
2021-03-29 11:53:24
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
xt-xcc does not support deprecated attribute
|
area: Toolchains bug priority: medium
|
**Describe the bug**
file:gcc.h
```
#ifndef __deprecated
#define __deprecated __attribute__((deprecated))
#endif
```
file:bluetooth.h
```
BT_LE_SCAN_FILTER_DUPLICATE __deprecated =
BT_LE_SCAN_OPT_FILTER_DUPLICATE,
```
xt-xcc does not support deprecated attribute.
**Logs and console output**
```
[17/152] Building C object zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj
FAILED: zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj
C:\usr\xtensa\XtDevTools\install\tools\RI-2020.5-win32\XtensaTools\bin\xt-xcc.exe -DBUILD_VERSION=zephyr-v2.4.0-1113-g0bcd66ac36d8 -DKERNEL -D_FORTIFY_SOURCE=2 -D__ZEPHYR__=1 -I../include -Izephyr/include/generated -I../soc/xtensa/ZC3827 -I../soc/xtensa/ZC3827/include -I../subsys/settings/include -I../subsys/bluetooth -IE:/work/rtos_work/zephyrproject/modules/crypto/tinycrypt/lib/include -IE:/work/rtos_work/zephyrproject/modules/hal/xtensa/include -IE:/work/rtos_work/zephyrproject/modules/hal/xtensa/zephyr/soc/ZC3827 -isystem ../lib/libc/minimal/include -Os -imacros E:/work/rtos_work/zephyrproject/zephyr/build/zephyr/include/generated/autoconf.h -ffreestanding -fno-common -g -imacrosE:/work/rtos_work/zephyrproject/zephyr/include/toolchain/xcc_missing_defs.h -fms-extensions -imacros E:/work/rtos_work/zephyrproject/zephyr/include/toolchain/zephyr_stdint.h -Wall -Wformat -Wformat-security -Wno-format-zero-length -Wno-main -Wno-pointer-sign -Wpointer-arith -Werror=implicit-int -fno-pic -fno-strict-overflow -ffunction-sections -fdata-sections -mlongcalls -std=c99 -MD -MT zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj -MF zephyr\subsys\bluetooth\host\CMakeFiles\subsys__bluetooth__host.dir\conn.c.obj.d -o zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj -c E:/work/rtos_work/zephyrproject/zephyr/subsys/bluetooth/host/conn.c
In file included from ../include/bluetooth/conn.h:23,
from ../include/bluetooth/hci.h:20,
from E:/work/rtos_work/zephyrproject/zephyr/subsys/bluetooth/host/conn.c:21:
../include/bluetooth/bluetooth.h:1240: error: expected ',' or '}' before '__attribute__'
In file included from ../include/bluetooth/hci.h:20,
from E:/work/rtos_work/zephyrproject/zephyr/subsys/bluetooth/host/conn.c:21:
../include/bluetooth/conn.h:685: error: expected ',' or '}' before '__attribute__'
```
**Environment (please complete the following information):**
- OS: ( Windows)
- Toolchain (xt-cc.)
|
1.0
|
xt-xcc does not support deprecated attribute - **Describe the bug**
file:gcc.h
```
#ifndef __deprecated
#define __deprecated __attribute__((deprecated))
#endif
```
file:bluetooth.h
```
BT_LE_SCAN_FILTER_DUPLICATE __deprecated =
BT_LE_SCAN_OPT_FILTER_DUPLICATE,
```
xt-xcc does not support deprecated attribute.
**Logs and console output**
```
[17/152] Building C object zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj
FAILED: zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj
C:\usr\xtensa\XtDevTools\install\tools\RI-2020.5-win32\XtensaTools\bin\xt-xcc.exe -DBUILD_VERSION=zephyr-v2.4.0-1113-g0bcd66ac36d8 -DKERNEL -D_FORTIFY_SOURCE=2 -D__ZEPHYR__=1 -I../include -Izephyr/include/generated -I../soc/xtensa/ZC3827 -I../soc/xtensa/ZC3827/include -I../subsys/settings/include -I../subsys/bluetooth -IE:/work/rtos_work/zephyrproject/modules/crypto/tinycrypt/lib/include -IE:/work/rtos_work/zephyrproject/modules/hal/xtensa/include -IE:/work/rtos_work/zephyrproject/modules/hal/xtensa/zephyr/soc/ZC3827 -isystem ../lib/libc/minimal/include -Os -imacros E:/work/rtos_work/zephyrproject/zephyr/build/zephyr/include/generated/autoconf.h -ffreestanding -fno-common -g -imacrosE:/work/rtos_work/zephyrproject/zephyr/include/toolchain/xcc_missing_defs.h -fms-extensions -imacros E:/work/rtos_work/zephyrproject/zephyr/include/toolchain/zephyr_stdint.h -Wall -Wformat -Wformat-security -Wno-format-zero-length -Wno-main -Wno-pointer-sign -Wpointer-arith -Werror=implicit-int -fno-pic -fno-strict-overflow -ffunction-sections -fdata-sections -mlongcalls -std=c99 -MD -MT zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj -MF zephyr\subsys\bluetooth\host\CMakeFiles\subsys__bluetooth__host.dir\conn.c.obj.d -o zephyr/subsys/bluetooth/host/CMakeFiles/subsys__bluetooth__host.dir/conn.c.obj -c E:/work/rtos_work/zephyrproject/zephyr/subsys/bluetooth/host/conn.c
In file included from ../include/bluetooth/conn.h:23,
from ../include/bluetooth/hci.h:20,
from E:/work/rtos_work/zephyrproject/zephyr/subsys/bluetooth/host/conn.c:21:
../include/bluetooth/bluetooth.h:1240: error: expected ',' or '}' before '__attribute__'
In file included from ../include/bluetooth/hci.h:20,
from E:/work/rtos_work/zephyrproject/zephyr/subsys/bluetooth/host/conn.c:21:
../include/bluetooth/conn.h:685: error: expected ',' or '}' before '__attribute__'
```
**Environment (please complete the following information):**
- OS: ( Windows)
- Toolchain (xt-cc.)
|
non_test
|
xt xcc does not support deprecated attribute describe the bug file gcc h ifndef deprecated define deprecated attribute deprecated endif file bluetooth h bt le scan filter duplicate deprecated bt le scan opt filter duplicate xt xcc does not support deprecated attribute logs and console output building c object zephyr subsys bluetooth host cmakefiles subsys bluetooth host dir conn c obj failed zephyr subsys bluetooth host cmakefiles subsys bluetooth host dir conn c obj c usr xtensa xtdevtools install tools ri xtensatools bin xt xcc exe dbuild version zephyr dkernel d fortify source d zephyr i include izephyr include generated i soc xtensa i soc xtensa include i subsys settings include i subsys bluetooth ie work rtos work zephyrproject modules crypto tinycrypt lib include ie work rtos work zephyrproject modules hal xtensa include ie work rtos work zephyrproject modules hal xtensa zephyr soc isystem lib libc minimal include os imacros e work rtos work zephyrproject zephyr build zephyr include generated autoconf h ffreestanding fno common g imacrose work rtos work zephyrproject zephyr include toolchain xcc missing defs h fms extensions imacros e work rtos work zephyrproject zephyr include toolchain zephyr stdint h wall wformat wformat security wno format zero length wno main wno pointer sign wpointer arith werror implicit int fno pic fno strict overflow ffunction sections fdata sections mlongcalls std md mt zephyr subsys bluetooth host cmakefiles subsys bluetooth host dir conn c obj mf zephyr subsys bluetooth host cmakefiles subsys bluetooth host dir conn c obj d o zephyr subsys bluetooth host cmakefiles subsys bluetooth host dir conn c obj c e work rtos work zephyrproject zephyr subsys bluetooth host conn c in file included from include bluetooth conn h from include bluetooth hci h from e work rtos work zephyrproject zephyr subsys bluetooth host conn c include bluetooth bluetooth h error expected or before attribute in file included from include bluetooth hci h from e work rtos work zephyrproject zephyr subsys bluetooth host conn c include bluetooth conn h error expected or before attribute environment please complete the following information os windows toolchain xt cc
| 0
|
214,029
| 16,554,644,016
|
IssuesEvent
|
2021-05-28 12:38:57
|
Realm667/WolfenDoom
|
https://api.github.com/repos/Realm667/WolfenDoom
|
closed
|
C2M5_A - Stahlhimmel - moving train
|
mapping playtesting suggestion
|
C2M5_A ...
While on the moving train, and outside, Things scroll. This area has sector type 46 - Wind south weak.
The effect is not weak at all. I suggest removing this sector type from this area. The wind is gale force ! Not a gook look IMO.
Also, somehow, Dirty Douglas fell from the train. To "retrieve" him, I had to walk back to the rear carriage of the train.
UPDATE ... I reproduced this. He can fall at line 4903. This line has the block monster flag. Would also adding the "block players" flag help ?
There needs to be a door or something at sector 1335. Player can got back and forth between sectors 1359 and 1377 and the train starts/stops. This is not wanted. Once the player reaches sector 1359, he should not be able to go back. Train has stopped.
|
1.0
|
C2M5_A - Stahlhimmel - moving train - C2M5_A ...
While on the moving train, and outside, Things scroll. This area has sector type 46 - Wind south weak.
The effect is not weak at all. I suggest removing this sector type from this area. The wind is gale force ! Not a gook look IMO.
Also, somehow, Dirty Douglas fell from the train. To "retrieve" him, I had to walk back to the rear carriage of the train.
UPDATE ... I reproduced this. He can fall at line 4903. This line has the block monster flag. Would also adding the "block players" flag help ?
There needs to be a door or something at sector 1335. Player can got back and forth between sectors 1359 and 1377 and the train starts/stops. This is not wanted. Once the player reaches sector 1359, he should not be able to go back. Train has stopped.
|
test
|
a stahlhimmel moving train a while on the moving train and outside things scroll this area has sector type wind south weak the effect is not weak at all i suggest removing this sector type from this area the wind is gale force not a gook look imo also somehow dirty douglas fell from the train to retrieve him i had to walk back to the rear carriage of the train update i reproduced this he can fall at line this line has the block monster flag would also adding the block players flag help there needs to be a door or something at sector player can got back and forth between sectors and and the train starts stops this is not wanted once the player reaches sector he should not be able to go back train has stopped
| 1
|
130,614
| 10,618,153,911
|
IssuesEvent
|
2019-10-13 01:31:42
|
magento/graphql-ce
|
https://api.github.com/repos/magento/graphql-ce
|
closed
|
[Test coverage] Cover CartAddressTypeResolver
|
Award: test coverage Component: QuoteGraphQl Progress: PR created good first issue
|
### Description:
Cover with api-functional tests
```
QuoteGraphQl\Model\Resolver\CartAddressTypeResolver
```
|
1.0
|
[Test coverage] Cover CartAddressTypeResolver - ### Description:
Cover with api-functional tests
```
QuoteGraphQl\Model\Resolver\CartAddressTypeResolver
```
|
test
|
cover cartaddresstyperesolver description cover with api functional tests quotegraphql model resolver cartaddresstyperesolver
| 1
|
84,407
| 15,720,901,073
|
IssuesEvent
|
2021-03-29 01:33:55
|
benchmarkdebricked/generator-jhipster
|
https://api.github.com/repos/benchmarkdebricked/generator-jhipster
|
closed
|
CVE-2020-5398 (High) detected in spring-web-5.0.6.RELEASE.jar - autoclosed
|
security vulnerability
|
## CVE-2020-5398 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.0.6.RELEASE.jar</b></p></summary>
<p>Spring Web</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /tmp/ws-scm/generator-jhipster/test/templates/ci-cd/maven-ngx-yarn/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-web/5.0.6.RELEASE/spring-web-5.0.6.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.2.RELEASE.jar (Root Library)
- :x: **spring-web-5.0.6.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/benchmarkdebricked/generator-jhipster/commit/56521ff393b8726bd648cc9bf6e2fd9552b7cc38">56521ff393b8726bd648cc9bf6e2fd9552b7cc38</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework, versions 5.2.x prior to 5.2.3, versions 5.1.x prior to 5.1.13, and versions 5.0.x prior to 5.0.16, an application is vulnerable to a reflected file download (RFD) attack when it sets a "Content-Disposition" header in the response where the filename attribute is derived from user supplied input.
<p>Publish Date: 2020-01-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-5398>CVE-2020-5398</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://pivotal.io/security/cve-2020-5398">https://pivotal.io/security/cve-2020-5398</a></p>
<p>Release Date: 2020-01-17</p>
<p>Fix Resolution: org.springframework:spring-web:5.0.16.RELEASE,org.springframework:spring-web:5.1.13.RELEASE,org.springframework:spring-web:5.2.3.RELEASE</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-5398 (High) detected in spring-web-5.0.6.RELEASE.jar - autoclosed - ## CVE-2020-5398 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.0.6.RELEASE.jar</b></p></summary>
<p>Spring Web</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /tmp/ws-scm/generator-jhipster/test/templates/ci-cd/maven-ngx-yarn/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-web/5.0.6.RELEASE/spring-web-5.0.6.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.2.RELEASE.jar (Root Library)
- :x: **spring-web-5.0.6.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/benchmarkdebricked/generator-jhipster/commit/56521ff393b8726bd648cc9bf6e2fd9552b7cc38">56521ff393b8726bd648cc9bf6e2fd9552b7cc38</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework, versions 5.2.x prior to 5.2.3, versions 5.1.x prior to 5.1.13, and versions 5.0.x prior to 5.0.16, an application is vulnerable to a reflected file download (RFD) attack when it sets a "Content-Disposition" header in the response where the filename attribute is derived from user supplied input.
<p>Publish Date: 2020-01-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-5398>CVE-2020-5398</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://pivotal.io/security/cve-2020-5398">https://pivotal.io/security/cve-2020-5398</a></p>
<p>Release Date: 2020-01-17</p>
<p>Fix Resolution: org.springframework:spring-web:5.0.16.RELEASE,org.springframework:spring-web:5.1.13.RELEASE,org.springframework:spring-web:5.2.3.RELEASE</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in spring web release jar autoclosed cve high severity vulnerability vulnerable library spring web release jar spring web library home page a href path to dependency file tmp ws scm generator jhipster test templates ci cd maven ngx yarn pom xml path to vulnerable library root repository org springframework spring web release spring web release jar dependency hierarchy spring boot starter web release jar root library x spring web release jar vulnerable library found in head commit a href vulnerability details in spring framework versions x prior to versions x prior to and versions x prior to an application is vulnerable to a reflected file download rfd attack when it sets a content disposition header in the response where the filename attribute is derived from user supplied input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring web release org springframework spring web release org springframework spring web release step up your open source security game with whitesource
| 0
|
287,176
| 24,814,680,744
|
IssuesEvent
|
2022-10-25 12:16:54
|
guardicore/monkey
|
https://api.github.com/repos/guardicore/monkey
|
closed
|
Add automated integration tests for post-breach actions
|
Impact: High Complexity: Medium Testing
|
Our exploiters are tested nightly in our blackbox test suite. In order to expedite the release process, as well as catch bugs early, all post-breach actions need automated tests.
|
1.0
|
Add automated integration tests for post-breach actions - Our exploiters are tested nightly in our blackbox test suite. In order to expedite the release process, as well as catch bugs early, all post-breach actions need automated tests.
|
test
|
add automated integration tests for post breach actions our exploiters are tested nightly in our blackbox test suite in order to expedite the release process as well as catch bugs early all post breach actions need automated tests
| 1
|
142,748
| 11,491,959,967
|
IssuesEvent
|
2020-02-11 20:00:49
|
fica99/42sh
|
https://api.github.com/repos/fica99/42sh
|
closed
|
Lexer
|
Lexer Testing
|
Lexer
=====================
Описание функций лексера
-----------------------------------
**Прототип функции:** ```t_lex_tkn *lex_get_next_tkn(char **str, size_t pos)```
**Описание параметров функции:**
* `char **str` - входная строка подлежащая анализу;
* `size_t pos` - индекс позиции с которой необходимо провести анализ;
Функция возвращает токен одного из типов, в случае конца строки возвращается токен `"END"`,
в случае ошибки возвращается `"NULL"`.
***
**Прототип функции:** ```t_lex_tkn *lex_del_tkn(t_lex_tkn *token)```
**Описание параметров функции:**
* `t_lex_tkn *token` - токен;
Функция производит удаление токена.
***
**Прототип функции:** ```t_lex_tkn **lex_get_tkns(char **str)```
**Описание параметров функции:**
* `char **str` - входная строка подлежащая анализу;
Функция возвращает массив токенов, заканчивающийся значением `"NULL"`, в случае ошибки возвращается `"NULL"`.
***
**Прототип функции:** ```t_lex_tkn **lex_del_tkns(t_lex_tkn **tokens)```
**Описание параметров функции:**
* `t_lex_tkn **tokens` - массив токенов;
Функция производит удаление массива токенов.
***
Описание структуры данных
-----------------------------------
**Структура токена:**
```
typedef struct s_lex_tkn
{
char *value;
t_lex_tkn_type type;
t_lex_tkn_class class;
size_t start_pos;
size_t end_pos;
} t_lex_tkn;
```
***
**Поля структуры токена:**
* `value` - строковое значение токена;
* `type` - хранит тип токена;
* `class` - хранит класс токена;
* `start_pos` - индекс начала позиции;
* `end_pos` - индекс конца позиции;
***
**Типы возвращаемых токенов:**
* `T_NULL` - пустое значение;
* `T_ERR` - ошибка;
* `T_END` - `"\0"`;
* `T_WORD` - `"ls -l ~/"`;
* `T_PIPE` - `"|"`;
* `T_GREATER` - `">"`;
* `T_GREATER_GREATER` - `">>"`;
* `T_LESS` - `"<"`;
* `T_LESS_LESS` - `"<<"`;
* `T_LESS_AND` - `"<&"`;
* `T_GREATER_AND` - `">&"`;
* `T_SEP` - `";"`;
* `T_CONTROL_SUB` - `"$()"`;
* `T_AND_AND` - `"&&"`;
* `T_OR_OR` - `"||"`;
* `T_AND` - `"&"`;
* `T_OPEN_FIG_BRACE` - `"{"`;
* `T_CLOSE_FIG_BRACE` - `"}"`;
* `T_ARITH_OPERS` - `"$(())"`.
***
**Классы возвращаемых токенов:**
* `C_NULL` - `"T_NULL"`;
* `C_END` - `"\0"`;
* `C_WORD` - `"T_WORD"`;
* `C_PIPE` - `"|"`;
* `C_REDIR` - `"<", "<<", ">", ">>", "<&", ">&"`;
* `C_SEP` - `";"`;
* `C_CONTROL_SUB` - `"$()"`;
* `C_LOG_OPERS` - `"&&", "||"`;
* `C_AND` - `"&"`;
* `C_FIG_BRACE` - `"{". "}"`;
* `C_ARITH_OPERS` - `"$(())"`;
|
1.0
|
Lexer - Lexer
=====================
Описание функций лексера
-----------------------------------
**Прототип функции:** ```t_lex_tkn *lex_get_next_tkn(char **str, size_t pos)```
**Описание параметров функции:**
* `char **str` - входная строка подлежащая анализу;
* `size_t pos` - индекс позиции с которой необходимо провести анализ;
Функция возвращает токен одного из типов, в случае конца строки возвращается токен `"END"`,
в случае ошибки возвращается `"NULL"`.
***
**Прототип функции:** ```t_lex_tkn *lex_del_tkn(t_lex_tkn *token)```
**Описание параметров функции:**
* `t_lex_tkn *token` - токен;
Функция производит удаление токена.
***
**Прототип функции:** ```t_lex_tkn **lex_get_tkns(char **str)```
**Описание параметров функции:**
* `char **str` - входная строка подлежащая анализу;
Функция возвращает массив токенов, заканчивающийся значением `"NULL"`, в случае ошибки возвращается `"NULL"`.
***
**Прототип функции:** ```t_lex_tkn **lex_del_tkns(t_lex_tkn **tokens)```
**Описание параметров функции:**
* `t_lex_tkn **tokens` - массив токенов;
Функция производит удаление массива токенов.
***
Описание структуры данных
-----------------------------------
**Структура токена:**
```
typedef struct s_lex_tkn
{
char *value;
t_lex_tkn_type type;
t_lex_tkn_class class;
size_t start_pos;
size_t end_pos;
} t_lex_tkn;
```
***
**Поля структуры токена:**
* `value` - строковое значение токена;
* `type` - хранит тип токена;
* `class` - хранит класс токена;
* `start_pos` - индекс начала позиции;
* `end_pos` - индекс конца позиции;
***
**Типы возвращаемых токенов:**
* `T_NULL` - пустое значение;
* `T_ERR` - ошибка;
* `T_END` - `"\0"`;
* `T_WORD` - `"ls -l ~/"`;
* `T_PIPE` - `"|"`;
* `T_GREATER` - `">"`;
* `T_GREATER_GREATER` - `">>"`;
* `T_LESS` - `"<"`;
* `T_LESS_LESS` - `"<<"`;
* `T_LESS_AND` - `"<&"`;
* `T_GREATER_AND` - `">&"`;
* `T_SEP` - `";"`;
* `T_CONTROL_SUB` - `"$()"`;
* `T_AND_AND` - `"&&"`;
* `T_OR_OR` - `"||"`;
* `T_AND` - `"&"`;
* `T_OPEN_FIG_BRACE` - `"{"`;
* `T_CLOSE_FIG_BRACE` - `"}"`;
* `T_ARITH_OPERS` - `"$(())"`.
***
**Классы возвращаемых токенов:**
* `C_NULL` - `"T_NULL"`;
* `C_END` - `"\0"`;
* `C_WORD` - `"T_WORD"`;
* `C_PIPE` - `"|"`;
* `C_REDIR` - `"<", "<<", ">", ">>", "<&", ">&"`;
* `C_SEP` - `";"`;
* `C_CONTROL_SUB` - `"$()"`;
* `C_LOG_OPERS` - `"&&", "||"`;
* `C_AND` - `"&"`;
* `C_FIG_BRACE` - `"{". "}"`;
* `C_ARITH_OPERS` - `"$(())"`;
|
test
|
lexer lexer описание функций лексера прототип функции t lex tkn lex get next tkn char str size t pos описание параметров функции char str входная строка подлежащая анализу size t pos индекс позиции с которой необходимо провести анализ функция возвращает токен одного из типов в случае конца строки возвращается токен end в случае ошибки возвращается null прототип функции t lex tkn lex del tkn t lex tkn token описание параметров функции t lex tkn token токен функция производит удаление токена прототип функции t lex tkn lex get tkns char str описание параметров функции char str входная строка подлежащая анализу функция возвращает массив токенов заканчивающийся значением null в случае ошибки возвращается null прототип функции t lex tkn lex del tkns t lex tkn tokens описание параметров функции t lex tkn tokens массив токенов функция производит удаление массива токенов описание структуры данных структура токена typedef struct s lex tkn char value t lex tkn type type t lex tkn class class size t start pos size t end pos t lex tkn поля структуры токена value строковое значение токена type хранит тип токена class хранит класс токена start pos индекс начала позиции end pos индекс конца позиции типы возвращаемых токенов t null пустое значение t err ошибка t end t word ls l t pipe t greater t greater greater t less t less less t less and t greater and t sep t control sub t and and t or or t and t open fig brace t close fig brace t arith opers классы возвращаемых токенов c null t null c end c word t word c pipe c redir c sep c control sub c log opers c and c fig brace c arith opers
| 1
|
7,217
| 2,610,358,233
|
IssuesEvent
|
2015-02-26 19:55:59
|
chrsmith/scribefire-chrome
|
https://api.github.com/repos/chrsmith/scribefire-chrome
|
opened
|
Can't figure this out.
|
auto-migrated Priority-Medium Type-Defect
|
```
What's the problem?
Can't figure out how to set this up to post on my wordpress blog and I can't
find any documentation. The setup requires an API URL and even my best
programmer has no idea what you're talking about.
The old version worked fine, but when it updated it seems to have lost all my
presets.
What browser are you using?
Firefox
What version of ScribeFire are you running?
4
```
-----
Original issue reported on code.google.com by `cpasites...@marshallhomeweb.com` on 16 Nov 2011 at 1:06
|
1.0
|
Can't figure this out. - ```
What's the problem?
Can't figure out how to set this up to post on my wordpress blog and I can't
find any documentation. The setup requires an API URL and even my best
programmer has no idea what you're talking about.
The old version worked fine, but when it updated it seems to have lost all my
presets.
What browser are you using?
Firefox
What version of ScribeFire are you running?
4
```
-----
Original issue reported on code.google.com by `cpasites...@marshallhomeweb.com` on 16 Nov 2011 at 1:06
|
non_test
|
can t figure this out what s the problem can t figure out how to set this up to post on my wordpress blog and i can t find any documentation the setup requires an api url and even my best programmer has no idea what you re talking about the old version worked fine but when it updated it seems to have lost all my presets what browser are you using firefox what version of scribefire are you running original issue reported on code google com by cpasites marshallhomeweb com on nov at
| 0
|
142,566
| 21,785,754,881
|
IssuesEvent
|
2022-05-14 05:10:44
|
depromeet/antoon_web
|
https://api.github.com/repos/depromeet/antoon_web
|
closed
|
[회원가입] 스크린 디자인 반영
|
v0.1 component design
|
## 💡 개요
- 사용자 회원 가입 페이지의 스크린 디자인 변경 사항 반영하기
## 📑 작업 사항
- [x] 회원가입 페이지의 css를 수정했습니다.
## 🔎 기타
|
1.0
|
[회원가입] 스크린 디자인 반영 - ## 💡 개요
- 사용자 회원 가입 페이지의 스크린 디자인 변경 사항 반영하기
## 📑 작업 사항
- [x] 회원가입 페이지의 css를 수정했습니다.
## 🔎 기타
|
non_test
|
스크린 디자인 반영 💡 개요 사용자 회원 가입 페이지의 스크린 디자인 변경 사항 반영하기 📑 작업 사항 회원가입 페이지의 css를 수정했습니다 🔎 기타
| 0
|
128,081
| 18,025,744,977
|
IssuesEvent
|
2021-09-17 04:07:52
|
scriptex/atanas.info
|
https://api.github.com/repos/scriptex/atanas.info
|
closed
|
CVE-2021-3801 (Medium) detected in prismjs-1.24.1.tgz
|
security vulnerability
|
## CVE-2021-3801 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prismjs-1.24.1.tgz</b></p></summary>
<p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p>
<p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.24.1.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.24.1.tgz</a></p>
<p>Path to dependency file: atanas.info/package.json</p>
<p>Path to vulnerable library: atanas.info/node_modules/prismjs/package.json</p>
<p>
Dependency Hierarchy:
- vuepress-1.8.2.tgz (Root Library)
- core-1.8.2.tgz
- markdown-1.8.2.tgz
- :x: **prismjs-1.24.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/scriptex/atanas.info/commit/8571f344549c5a72fcd59f7a81cc3ede049ae9c0">8571f344549c5a72fcd59f7a81cc3ede049ae9c0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
prism is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3801>CVE-2021-3801</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-3801 (Medium) detected in prismjs-1.24.1.tgz - ## CVE-2021-3801 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prismjs-1.24.1.tgz</b></p></summary>
<p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p>
<p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.24.1.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.24.1.tgz</a></p>
<p>Path to dependency file: atanas.info/package.json</p>
<p>Path to vulnerable library: atanas.info/node_modules/prismjs/package.json</p>
<p>
Dependency Hierarchy:
- vuepress-1.8.2.tgz (Root Library)
- core-1.8.2.tgz
- markdown-1.8.2.tgz
- :x: **prismjs-1.24.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/scriptex/atanas.info/commit/8571f344549c5a72fcd59f7a81cc3ede049ae9c0">8571f344549c5a72fcd59f7a81cc3ede049ae9c0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
prism is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3801>CVE-2021-3801</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in prismjs tgz cve medium severity vulnerability vulnerable library prismjs tgz lightweight robust elegant syntax highlighting a spin off project from dabblet library home page a href path to dependency file atanas info package json path to vulnerable library atanas info node modules prismjs package json dependency hierarchy vuepress tgz root library core tgz markdown tgz x prismjs tgz vulnerable library found in head commit a href vulnerability details prism is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href step up your open source security game with whitesource
| 0
|
452,552
| 32,062,087,067
|
IssuesEvent
|
2023-09-24 19:23:40
|
vapoursynth/vapoursynth
|
https://api.github.com/repos/vapoursynth/vapoursynth
|
closed
|
Documentation issues
|
Area: Documentation
|
There should be nicer link to the website in the docs, for example in the right hand navigation bar
|
1.0
|
Documentation issues - There should be nicer link to the website in the docs, for example in the right hand navigation bar
|
non_test
|
documentation issues there should be nicer link to the website in the docs for example in the right hand navigation bar
| 0
|
72,267
| 7,293,212,898
|
IssuesEvent
|
2018-02-25 11:46:58
|
haaspors/rlib
|
https://api.github.com/repos/haaspors/rlib
|
closed
|
tests: all tests don't run on windows debug
|
bug tests
|
MSC adds more padding between test data so that the magic test finder code desn't work.
|
1.0
|
tests: all tests don't run on windows debug - MSC adds more padding between test data so that the magic test finder code desn't work.
|
test
|
tests all tests don t run on windows debug msc adds more padding between test data so that the magic test finder code desn t work
| 1
|
81,574
| 10,150,040,901
|
IssuesEvent
|
2019-08-05 16:36:15
|
universelabs/universe-design
|
https://api.github.com/repos/universelabs/universe-design
|
closed
|
Design instruction card for Universe Feather packaging
|
design
|
**Description**
Design instruction card for Universe Feather packaging.
**Tasks**
- [x] ~~Create document.~~
- [x] ~~Design packaging card for Universe Feather device.~~
- [x] ~~Design packaging card dielines, folds, etc.~~
|
1.0
|
Design instruction card for Universe Feather packaging - **Description**
Design instruction card for Universe Feather packaging.
**Tasks**
- [x] ~~Create document.~~
- [x] ~~Design packaging card for Universe Feather device.~~
- [x] ~~Design packaging card dielines, folds, etc.~~
|
non_test
|
design instruction card for universe feather packaging description design instruction card for universe feather packaging tasks create document design packaging card for universe feather device design packaging card dielines folds etc
| 0
|
302,172
| 26,129,899,292
|
IssuesEvent
|
2022-12-29 02:16:51
|
apache/helix
|
https://api.github.com/repos/apache/helix
|
opened
|
[Failed CI Test] testGetAllMetadataStoreRealms(org.apache.helix.rest.server.TestMetadataStoreDirectoryAccessor)
|
FailedTestTracking
|
This issue is created for tracking unstable test: testGetAllMetadataStoreRealms(org.apache.helix.rest.server.TestMetadataStoreDirectoryAccessor)
|
1.0
|
[Failed CI Test] testGetAllMetadataStoreRealms(org.apache.helix.rest.server.TestMetadataStoreDirectoryAccessor) - This issue is created for tracking unstable test: testGetAllMetadataStoreRealms(org.apache.helix.rest.server.TestMetadataStoreDirectoryAccessor)
|
test
|
testgetallmetadatastorerealms org apache helix rest server testmetadatastoredirectoryaccessor this issue is created for tracking unstable test testgetallmetadatastorerealms org apache helix rest server testmetadatastoredirectoryaccessor
| 1
|
292,420
| 25,210,968,176
|
IssuesEvent
|
2022-11-14 03:50:17
|
pingcap/tidb
|
https://api.github.com/repos/pingcap/tidb
|
closed
|
DATA RACE in the (*Chunk).resetForReuse()
|
type/bug component/test severity/major affects-6.4
|
## Bug Report
Please answer these questions before submitting your issue. Thanks!
### 1. Minimal reproduce step (Required)
```
==================
WARNING: DATA RACE
Write at 0x00c02a98e450 by goroutine 702063:
github.com/pingcap/tidb/util/chunk.(*Chunk).resetForReuse()
util/chunk/chunk.go:137 +0x7b8
github.com/pingcap/tidb/util/chunk.(*allocator).Reset()
util/chunk/alloc.go:123 +0x12a
github.com/pingcap/tidb/testkit.(*TestKit).MustExec.func1()
testkit/testkit.go:120 +0xde
runtime.deferreturn()
GOROOT/src/runtime/panic.go:476 +0x32
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1561 +0x6aa
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute()
executor/projection.go:223 +0x651
github.com/pingcap/tidb/executor.(*ProjectionExec).Next()
executor/projection.go:181 +0xcc
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute()
executor/projection.go:223 +0x651
github.com/pingcap/tidb/executor.(*ProjectionExec).Next()
executor/projection.go:181 +0xcc
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute()
executor/projection.go:223 +0x651
github.com/pingcap/tidb/executor.(*ProjectionExec).Next()
executor/projection.go:181 +0xcc
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute()
executor/projection.go:223 +0x651
github.com/pingcap/tidb/executor.(*ProjectionExec).Next()
executor/projection.go:181 +0xcc
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows()
executor/aggregate.go:1382 +0x124
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1361 +0xad1
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk()
executor/aggregate.go:1429 +0x276
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1366 +0xb24
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows()
executor/aggregate.go:1382 +0x124
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1361 +0xad1
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk()
executor/aggregate.go:1429 +0x276
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1366 +0xb24
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows()
executor/aggregate.go:1382 +0x124
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1361 +0xad1
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows()
executor/aggregate.go:1382 +0x124
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1361 +0xad1
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows()
executor/aggregate.go:1382 +0x124
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeCurGroupRowsAndFetchChild()
executor/aggregate.go:1390 +0x64
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1343 +0x70c
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk()
executor/aggregate.go:1429 +0x276
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1366 +0xb24
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows()
executor/aggregate.go:1382 +0x124
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1361 +0xad1
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk()
executor/aggregate.go:1429 +0x276
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1366 +0xb24
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/tidb/executor.(*HashAggExec).Open()
executor/aggregate.go:311 +0x3ba
github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor()
executor/adapter.go:1111 +0xe9
github.com/pingcap/tidb/executor.(*ExecStmt).Exec()
executor/adapter.go:494 +0x994
github.com/pingcap/tidb/session.runStmt()
session/session.go:2333 +0x761
github.com/pingcap/tidb/session.(*session).ExecuteStmt()
session/session.go:2197 +0x1025
github.com/pingcap/tidb/testkit.(*TestKit).ExecWithContext()
testkit/testkit.go:296 +0x7c7
github.com/pingcap/tidb/testkit.(*TestKit).Exec()
testkit/testkit.go:270 +0x866
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1568 +0x829
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/tidb/executor.(*HashAggExec).Open()
executor/aggregate.go:311 +0x3ba
github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor()
executor/adapter.go:1111 +0xe9
github.com/pingcap/tidb/executor.(*ExecStmt).Exec()
executor/adapter.go:494 +0x994
github.com/pingcap/tidb/session.runStmt()
session/session.go:2333 +0x761
github.com/pingcap/tidb/session.(*session).ExecuteStmt()
session/session.go:2197 +0x1025
github.com/pingcap/tidb/testkit.(*TestKit).ExecWithContext()
testkit/testkit.go:296 +0x7c7
github.com/pingcap/tidb/testkit.(*TestKit).Exec()
testkit/testkit.go:270 +0x866
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1568 +0x829
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/tidb/executor.(*HashAggExec).Open()
executor/aggregate.go:311 +0x3ba
github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor()
executor/adapter.go:1111 +0xe9
github.com/pingcap/tidb/executor.(*ExecStmt).Exec()
executor/adapter.go:494 +0x994
github.com/pingcap/tidb/session.runStmt()
session/session.go:2333 +0x761
github.com/pingcap/tidb/session.(*session).ExecuteStmt()
session/session.go:2197 +0x1025
github.com/pingcap/tidb/testkit.(*TestKit).ExecWithContext()
testkit/testkit.go:296 +0x7c7
github.com/pingcap/tidb/testkit.(*TestKit).Exec()
testkit/testkit.go:270 +0x866
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1568 +0x829
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1004 +0x269
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/tidb/executor.(*HashAggExec).Open()
executor/aggregate.go:314 +0x3f1
github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor()
executor/adapter.go:1111 +0xe9
github.com/pingcap/tidb/executor.(*ExecStmt).Exec()
executor/adapter.go:494 +0x994
github.com/pingcap/tidb/session.runStmt()
session/session.go:2333 +0x761
github.com/pingcap/tidb/session.(*session).ExecuteStmt()
session/session.go:2197 +0x1025
Previous read at 0x00c02a98e450 by goroutine 716207:
github.com/pingcap/tidb/util/chunk.(*Chunk).NumRows()
util/chunk/chunk.go:352 +0x552
github.com/pingcap/tidb/executor.(*HashJoinExec).fetchBuildSideRows()
executor/join.go:308 +0x4d8
github.com/pingcap/tidb/executor.(*HashJoinExec).fetchAndBuildHashTable.func2()
executor/join.go:1170 +0xea
github.com/pingcap/tidb/util.WithRecovery()
util/misc.go:96 +0x6d
github.com/pingcap/tidb/executor.(*HashJoinExec).fetchAndBuildHashTable.func4()
executor/join.go:1167 +0x47
Goroutine 702063 (running) created at:
testing.(*T).Run()
GOROOT/src/testing/testing.go:1493 +0x75d
testing.runTests.func1()
GOROOT/src/testing/testing.go:1846 +0x99
testing.tRunner()
GOROOT/src/testing/testing.go:1446 +0x216
testing.runTests()
GOROOT/src/testing/testing.go:1844 +0x7ec
testing.(*M).Run()
GOROOT/src/testing/testing.go:1726 +0xa84
github.com/pingcap/tidb/testkit/testmain.(*testingM).Run()
testkit/testmain/wrapper.go:27 +0x42
go.uber.org/goleak.VerifyTestMain()
external/org_uber_go_goleak/testmain.go:53 +0x70
github.com/pingcap/tidb/executor_test.TestMain()
executor/main_test.go:70 +0xbc6
main.main()
bazel-out/k8-fastbuild/bin/executor/executor_test_/testmain.go:2288 +0x5e8
Goroutine 716207 (finished) created at:
github.com/pingcap/tidb/executor.(*HashJoinExec).fetchAndBuildHashTable()
executor/join.go:1167 +0x3d6
github.com/pingcap/tidb/executor.(*HashJoinExec).Next.func1()
executor/join.go:1126 +0xbc
github.com/pingcap/tidb/util.WithRecovery()
util/misc.go:96 +0x6d
github.com/pingcap/tidb/executor.(*HashJoinExec).Next.func2()
executor/join.go:1124 +0x47
==================
```
<!-- a step by step guide for reproducing the bug. -->
### 2. What did you expect to see? (Required)
### 3. What did you see instead (Required)
### 4. What is your TiDB version? (Required)
<!-- Paste the output of SELECT tidb_version() -->
|
1.0
|
DATA RACE in the (*Chunk).resetForReuse() - ## Bug Report
Please answer these questions before submitting your issue. Thanks!
### 1. Minimal reproduce step (Required)
```
==================
WARNING: DATA RACE
Write at 0x00c02a98e450 by goroutine 702063:
github.com/pingcap/tidb/util/chunk.(*Chunk).resetForReuse()
util/chunk/chunk.go:137 +0x7b8
github.com/pingcap/tidb/util/chunk.(*allocator).Reset()
util/chunk/alloc.go:123 +0x12a
github.com/pingcap/tidb/testkit.(*TestKit).MustExec.func1()
testkit/testkit.go:120 +0xde
runtime.deferreturn()
GOROOT/src/runtime/panic.go:476 +0x32
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1561 +0x6aa
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute()
executor/projection.go:223 +0x651
github.com/pingcap/tidb/executor.(*ProjectionExec).Next()
executor/projection.go:181 +0xcc
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute()
executor/projection.go:223 +0x651
github.com/pingcap/tidb/executor.(*ProjectionExec).Next()
executor/projection.go:181 +0xcc
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute()
executor/projection.go:223 +0x651
github.com/pingcap/tidb/executor.(*ProjectionExec).Next()
executor/projection.go:181 +0xcc
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*ProjectionExec).parallelExecute()
executor/projection.go:223 +0x651
github.com/pingcap/tidb/executor.(*ProjectionExec).Next()
executor/projection.go:181 +0xcc
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows()
executor/aggregate.go:1382 +0x124
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1361 +0xad1
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk()
executor/aggregate.go:1429 +0x276
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1366 +0xb24
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows()
executor/aggregate.go:1382 +0x124
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1361 +0xad1
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk()
executor/aggregate.go:1429 +0x276
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1366 +0xb24
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows()
executor/aggregate.go:1382 +0x124
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1361 +0xad1
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows()
executor/aggregate.go:1382 +0x124
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1361 +0xad1
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows()
executor/aggregate.go:1382 +0x124
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeCurGroupRowsAndFetchChild()
executor/aggregate.go:1390 +0x64
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1343 +0x70c
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk()
executor/aggregate.go:1429 +0x276
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1366 +0xb24
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeGroupRows()
executor/aggregate.go:1382 +0x124
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1361 +0xad1
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*StreamAggExec).appendResult2Chunk()
executor/aggregate.go:1429 +0x276
github.com/pingcap/tidb/executor.(*StreamAggExec).consumeOneGroup()
executor/aggregate.go:1366 +0xb24
github.com/pingcap/tidb/executor.(*StreamAggExec).Next()
executor/aggregate.go:1313 +0x215
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/tidb/executor.(*HashAggExec).Open()
executor/aggregate.go:311 +0x3ba
github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor()
executor/adapter.go:1111 +0xe9
github.com/pingcap/tidb/executor.(*ExecStmt).Exec()
executor/adapter.go:494 +0x994
github.com/pingcap/tidb/session.runStmt()
session/session.go:2333 +0x761
github.com/pingcap/tidb/session.(*session).ExecuteStmt()
session/session.go:2197 +0x1025
github.com/pingcap/tidb/testkit.(*TestKit).ExecWithContext()
testkit/testkit.go:296 +0x7c7
github.com/pingcap/tidb/testkit.(*TestKit).Exec()
testkit/testkit.go:270 +0x866
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1568 +0x829
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/tidb/executor.(*HashAggExec).Open()
executor/aggregate.go:311 +0x3ba
github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor()
executor/adapter.go:1111 +0xe9
github.com/pingcap/tidb/executor.(*ExecStmt).Exec()
executor/adapter.go:494 +0x994
github.com/pingcap/tidb/session.runStmt()
session/session.go:2333 +0x761
github.com/pingcap/tidb/session.(*session).ExecuteStmt()
session/session.go:2197 +0x1025
github.com/pingcap/tidb/testkit.(*TestKit).ExecWithContext()
testkit/testkit.go:296 +0x7c7
github.com/pingcap/tidb/testkit.(*TestKit).Exec()
testkit/testkit.go:270 +0x866
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1568 +0x829
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/tidb/executor.(*HashAggExec).Open()
executor/aggregate.go:311 +0x3ba
github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor()
executor/adapter.go:1111 +0xe9
github.com/pingcap/tidb/executor.(*ExecStmt).Exec()
executor/adapter.go:494 +0x994
github.com/pingcap/tidb/session.runStmt()
session/session.go:2333 +0x761
github.com/pingcap/tidb/session.(*session).ExecuteStmt()
session/session.go:2197 +0x1025
github.com/pingcap/tidb/testkit.(*TestKit).ExecWithContext()
testkit/testkit.go:296 +0x7c7
github.com/pingcap/tidb/testkit.(*TestKit).Exec()
testkit/testkit.go:270 +0x866
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1568 +0x829
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1004 +0x269
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/failpoint.Eval()
external/com_github_pingcap_failpoint/failpoints.go:271 +0x44
github.com/pingcap/tidb/executor.(*HashAggExec).getPartialResults()
executor/aggregate.go:1120 +0x1de
github.com/pingcap/tidb/executor.(*HashAggExec).execute()
executor/aggregate.go:1038 +0xee9
github.com/pingcap/tidb/executor.(*HashAggExec).unparallelExec()
executor/aggregate.go:963 +0x17a
github.com/pingcap/tidb/executor.(*HashAggExec).Next()
executor/aggregate.go:782 +0x106
github.com/pingcap/tidb/executor.Next()
executor/executor.go:325 +0x5c3
github.com/pingcap/tidb/executor.(*ExecStmt).next()
executor/adapter.go:1118 +0x89
github.com/pingcap/tidb/executor.(*recordSet).Next()
executor/adapter.go:153 +0x164
github.com/pingcap/tidb/session.(*execStmtResult).Next()
<autogenerated>:1 +0x76
github.com/pingcap/tidb/session.GetRows4Test()
session/tidb.go:336 +0xea
github.com/pingcap/tidb/executor_test.TestRandomPanicConsume()
executor/aggregate_test.go:1570 +0x904
github.com/pingcap/tidb/executor.(*HashAggExec).Open()
executor/aggregate.go:314 +0x3f1
github.com/pingcap/tidb/executor.(*ExecStmt).openExecutor()
executor/adapter.go:1111 +0xe9
github.com/pingcap/tidb/executor.(*ExecStmt).Exec()
executor/adapter.go:494 +0x994
github.com/pingcap/tidb/session.runStmt()
session/session.go:2333 +0x761
github.com/pingcap/tidb/session.(*session).ExecuteStmt()
session/session.go:2197 +0x1025
Previous read at 0x00c02a98e450 by goroutine 716207:
github.com/pingcap/tidb/util/chunk.(*Chunk).NumRows()
util/chunk/chunk.go:352 +0x552
github.com/pingcap/tidb/executor.(*HashJoinExec).fetchBuildSideRows()
executor/join.go:308 +0x4d8
github.com/pingcap/tidb/executor.(*HashJoinExec).fetchAndBuildHashTable.func2()
executor/join.go:1170 +0xea
github.com/pingcap/tidb/util.WithRecovery()
util/misc.go:96 +0x6d
github.com/pingcap/tidb/executor.(*HashJoinExec).fetchAndBuildHashTable.func4()
executor/join.go:1167 +0x47
Goroutine 702063 (running) created at:
testing.(*T).Run()
GOROOT/src/testing/testing.go:1493 +0x75d
testing.runTests.func1()
GOROOT/src/testing/testing.go:1846 +0x99
testing.tRunner()
GOROOT/src/testing/testing.go:1446 +0x216
testing.runTests()
GOROOT/src/testing/testing.go:1844 +0x7ec
testing.(*M).Run()
GOROOT/src/testing/testing.go:1726 +0xa84
github.com/pingcap/tidb/testkit/testmain.(*testingM).Run()
testkit/testmain/wrapper.go:27 +0x42
go.uber.org/goleak.VerifyTestMain()
external/org_uber_go_goleak/testmain.go:53 +0x70
github.com/pingcap/tidb/executor_test.TestMain()
executor/main_test.go:70 +0xbc6
main.main()
bazel-out/k8-fastbuild/bin/executor/executor_test_/testmain.go:2288 +0x5e8
Goroutine 716207 (finished) created at:
github.com/pingcap/tidb/executor.(*HashJoinExec).fetchAndBuildHashTable()
executor/join.go:1167 +0x3d6
github.com/pingcap/tidb/executor.(*HashJoinExec).Next.func1()
executor/join.go:1126 +0xbc
github.com/pingcap/tidb/util.WithRecovery()
util/misc.go:96 +0x6d
github.com/pingcap/tidb/executor.(*HashJoinExec).Next.func2()
executor/join.go:1124 +0x47
==================
```
<!-- a step by step guide for reproducing the bug. -->
### 2. What did you expect to see? (Required)
### 3. What did you see instead (Required)
### 4. What is your TiDB version? (Required)
<!-- Paste the output of SELECT tidb_version() -->
|
test
|
data race in the chunk resetforreuse bug report please answer these questions before submitting your issue thanks minimal reproduce step required warning data race write at by goroutine github com pingcap tidb util chunk chunk resetforreuse util chunk chunk go github com pingcap tidb util chunk allocator reset util chunk alloc go github com pingcap tidb testkit testkit mustexec testkit testkit go runtime deferreturn goroot src runtime panic go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor projectionexec parallelexecute executor projection go github com pingcap tidb executor projectionexec next executor projection go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor projectionexec parallelexecute executor projection go github com pingcap tidb executor projectionexec next executor projection go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor projectionexec parallelexecute executor projection go github com pingcap tidb executor projectionexec next executor projection go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor projectionexec parallelexecute executor projection go github com pingcap tidb executor projectionexec next executor projection go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec consumegrouprows executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec consumegrouprows executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec consumegrouprows executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec consumegrouprows executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec consumegrouprows executor aggregate go github com pingcap tidb executor streamaggexec consumecurgrouprowsandfetchchild executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec consumegrouprows executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor streamaggexec executor aggregate go github com pingcap tidb executor streamaggexec consumeonegroup executor aggregate go github com pingcap tidb executor streamaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap tidb executor hashaggexec open executor aggregate go github com pingcap tidb executor execstmt openexecutor executor adapter go github com pingcap tidb executor execstmt exec executor adapter go github com pingcap tidb session runstmt session session go github com pingcap tidb session session executestmt session session go github com pingcap tidb testkit testkit execwithcontext testkit testkit go github com pingcap tidb testkit testkit exec testkit testkit go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap tidb executor hashaggexec open executor aggregate go github com pingcap tidb executor execstmt openexecutor executor adapter go github com pingcap tidb executor execstmt exec executor adapter go github com pingcap tidb session runstmt session session go github com pingcap tidb session session executestmt session session go github com pingcap tidb testkit testkit execwithcontext testkit testkit go github com pingcap tidb testkit testkit exec testkit testkit go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap tidb executor hashaggexec open executor aggregate go github com pingcap tidb executor execstmt openexecutor executor adapter go github com pingcap tidb executor execstmt exec executor adapter go github com pingcap tidb session runstmt session session go github com pingcap tidb session session executestmt session session go github com pingcap tidb testkit testkit execwithcontext testkit testkit go github com pingcap tidb testkit testkit exec testkit testkit go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap failpoint eval external com github pingcap failpoint failpoints go github com pingcap tidb executor hashaggexec getpartialresults executor aggregate go github com pingcap tidb executor hashaggexec execute executor aggregate go github com pingcap tidb executor hashaggexec unparallelexec executor aggregate go github com pingcap tidb executor hashaggexec next executor aggregate go github com pingcap tidb executor next executor executor go github com pingcap tidb executor execstmt next executor adapter go github com pingcap tidb executor recordset next executor adapter go github com pingcap tidb session execstmtresult next github com pingcap tidb session session tidb go github com pingcap tidb executor test testrandompanicconsume executor aggregate test go github com pingcap tidb executor hashaggexec open executor aggregate go github com pingcap tidb executor execstmt openexecutor executor adapter go github com pingcap tidb executor execstmt exec executor adapter go github com pingcap tidb session runstmt session session go github com pingcap tidb session session executestmt session session go previous read at by goroutine github com pingcap tidb util chunk chunk numrows util chunk chunk go github com pingcap tidb executor hashjoinexec fetchbuildsiderows executor join go github com pingcap tidb executor hashjoinexec fetchandbuildhashtable executor join go github com pingcap tidb util withrecovery util misc go github com pingcap tidb executor hashjoinexec fetchandbuildhashtable executor join go goroutine running created at testing t run goroot src testing testing go testing runtests goroot src testing testing go testing trunner goroot src testing testing go testing runtests goroot src testing testing go testing m run goroot src testing testing go github com pingcap tidb testkit testmain testingm run testkit testmain wrapper go go uber org goleak verifytestmain external org uber go goleak testmain go github com pingcap tidb executor test testmain executor main test go main main bazel out fastbuild bin executor executor test testmain go goroutine finished created at github com pingcap tidb executor hashjoinexec fetchandbuildhashtable executor join go github com pingcap tidb executor hashjoinexec next executor join go github com pingcap tidb util withrecovery util misc go github com pingcap tidb executor hashjoinexec next executor join go what did you expect to see required what did you see instead required what is your tidb version required
| 1
|
90,531
| 11,415,000,646
|
IssuesEvent
|
2020-02-02 07:52:41
|
microsoft/ApplicationInspector
|
https://api.github.com/repos/microsoft/ApplicationInspector
|
closed
|
Report Show Skipped Code as Analyzed
|
working as designed
|
**Describe the bug**
The final report output.html show analyzed 0.99% but detects several kinds of code, I think the result is wrong and this could be the skipped %.
**Desktop (please complete the following information):**
- OS: Windows 10 - Linux
- Browser FireFox
<img width="271" alt="Captura de Pantalla 2020-01-27 a la(s) 08 55 57" src="https://user-images.githubusercontent.com/60227319/73173091-4eadfa80-40e3-11ea-98df-08319f96e7e0.png">
<img width="834" alt="Captura de Pantalla 2020-01-27 a la(s) 08 58 09" src="https://user-images.githubusercontent.com/60227319/73173092-4eadfa80-40e3-11ea-97f6-84b6f99bb62b.png">
|
1.0
|
Report Show Skipped Code as Analyzed - **Describe the bug**
The final report output.html show analyzed 0.99% but detects several kinds of code, I think the result is wrong and this could be the skipped %.
**Desktop (please complete the following information):**
- OS: Windows 10 - Linux
- Browser FireFox
<img width="271" alt="Captura de Pantalla 2020-01-27 a la(s) 08 55 57" src="https://user-images.githubusercontent.com/60227319/73173091-4eadfa80-40e3-11ea-98df-08319f96e7e0.png">
<img width="834" alt="Captura de Pantalla 2020-01-27 a la(s) 08 58 09" src="https://user-images.githubusercontent.com/60227319/73173092-4eadfa80-40e3-11ea-97f6-84b6f99bb62b.png">
|
non_test
|
report show skipped code as analyzed describe the bug the final report output html show analyzed but detects several kinds of code i think the result is wrong and this could be the skipped desktop please complete the following information os windows linux browser firefox img width alt captura de pantalla a la s src img width alt captura de pantalla a la s src
| 0
|
157,338
| 12,370,789,574
|
IssuesEvent
|
2020-05-18 17:24:22
|
hashgraph/hedera-wallet-ios
|
https://api.github.com/repos/hashgraph/hedera-wallet-ios
|
closed
|
Wrap hedera-cli CryptoGetAccountBalance
|
testing
|
Write a wrapper script for hedera-cli. The first command should be to get_account_balance and should return the balance in a script-friendly form.
|
1.0
|
Wrap hedera-cli CryptoGetAccountBalance - Write a wrapper script for hedera-cli. The first command should be to get_account_balance and should return the balance in a script-friendly form.
|
test
|
wrap hedera cli cryptogetaccountbalance write a wrapper script for hedera cli the first command should be to get account balance and should return the balance in a script friendly form
| 1
|
294,736
| 25,399,499,216
|
IssuesEvent
|
2022-11-22 11:00:02
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
sql/colexec: TestMergeJoinerMultiBatchRuns failed
|
C-test-failure O-robot branch-master
|
sql/colexec.TestMergeJoinerMultiBatchRuns [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7641761?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7641761?buildTab=artifacts#/) on master @ [dca415eddac0d659ae6d76b4e3dfdf4076adbd34](https://github.com/cockroachdb/cockroach/commits/dca415eddac0d659ae6d76b4e3dfdf4076adbd34):
Fatal error:
```
panic: test timed out after 14m55s
```
Stack:
```
goroutine 36988 [running]:
testing.(*M).startAlarm.func1()
GOROOT/src/testing/testing.go:2036 +0xbb
created by time.goFunc
GOROOT/src/time/sleep.go:176 +0x48
```
<details><summary>Log preceding fatal error</summary>
<p>
```
=== RUN TestMergeJoinerMultiBatchRuns
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/1727f600d839fa94e6186075a07a436e/logTestMergeJoinerMultiBatchRuns2968884249
test_log_scope.go:79: use -show-logs to present logs inline
=== RUN TestMergeJoinerMultiBatchRuns/groupSize=511/numInputBatches=2
=== RUN TestMergeJoinerMultiBatchRuns/groupSize=1023/numInputBatches=1
=== RUN TestMergeJoinerMultiBatchRuns/groupSize=2046/numInputBatches=1
=== RUN TestMergeJoinerMultiBatchRuns/groupSize=2046/numInputBatches=16
```
</p>
</details>
<p>Parameters: <code>TAGS=bazel,gss</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestMergeJoinerMultiBatchRuns.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
sql/colexec: TestMergeJoinerMultiBatchRuns failed - sql/colexec.TestMergeJoinerMultiBatchRuns [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7641761?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7641761?buildTab=artifacts#/) on master @ [dca415eddac0d659ae6d76b4e3dfdf4076adbd34](https://github.com/cockroachdb/cockroach/commits/dca415eddac0d659ae6d76b4e3dfdf4076adbd34):
Fatal error:
```
panic: test timed out after 14m55s
```
Stack:
```
goroutine 36988 [running]:
testing.(*M).startAlarm.func1()
GOROOT/src/testing/testing.go:2036 +0xbb
created by time.goFunc
GOROOT/src/time/sleep.go:176 +0x48
```
<details><summary>Log preceding fatal error</summary>
<p>
```
=== RUN TestMergeJoinerMultiBatchRuns
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/1727f600d839fa94e6186075a07a436e/logTestMergeJoinerMultiBatchRuns2968884249
test_log_scope.go:79: use -show-logs to present logs inline
=== RUN TestMergeJoinerMultiBatchRuns/groupSize=511/numInputBatches=2
=== RUN TestMergeJoinerMultiBatchRuns/groupSize=1023/numInputBatches=1
=== RUN TestMergeJoinerMultiBatchRuns/groupSize=2046/numInputBatches=1
=== RUN TestMergeJoinerMultiBatchRuns/groupSize=2046/numInputBatches=16
```
</p>
</details>
<p>Parameters: <code>TAGS=bazel,gss</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestMergeJoinerMultiBatchRuns.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
sql colexec testmergejoinermultibatchruns failed sql colexec testmergejoinermultibatchruns with on master fatal error panic test timed out after stack goroutine testing m startalarm goroot src testing testing go created by time gofunc goroot src time sleep go log preceding fatal error run testmergejoinermultibatchruns test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline run testmergejoinermultibatchruns groupsize numinputbatches run testmergejoinermultibatchruns groupsize numinputbatches run testmergejoinermultibatchruns groupsize numinputbatches run testmergejoinermultibatchruns groupsize numinputbatches parameters tags bazel gss help see also cc cockroachdb sql queries
| 1
|
340,146
| 30,495,814,143
|
IssuesEvent
|
2023-07-18 10:43:55
|
MohistMC/Mohist
|
https://api.github.com/repos/MohistMC/Mohist
|
reopened
|
[1.20.1]在服务器内合成物品时有小概率崩溃,且存档会受到部分回档
|
Wait Needs Testing
|
<!-- ISSUE_TEMPLATE_3 -> IMPORTANT: DO NOT DELETE THIS LINE.-->
<!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.-->
<!-- If you don't know where to upload your logs and crash reports, you can use these websites : -->
<!-- https://gist.github.com (recommended) -->
<!-- https://mclo.gs -->
<!-- https://haste.mohistmc.com -->
<!-- https://pastebin.com -->
<!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT -->
**Minecraft Version :** {1.20.1}
**Mohist Version :** {177}
**Operating System :** {windows 10 pro}
**Logs :** {> #
# A fatal error has been detected by the Java Runtime Environment:
#
# Internal Error (macroAssembler_x86.cpp:864), pid=7428, tid=10556
# fatal error: DEBUG MESSAGE: duplicated predicate failed which is impossible
#
# JRE version: Java(TM) SE Runtime Environment (17.0.7+8) (build 17.0.7+8-LTS-224)
# Java VM: Java HotSpot(TM) 64-Bit Server VM (17.0.7+8-LTS-224, mixed mode, sharing, tiered, compressed class ptrs, g1 gc, windows-amd64)
# No core dump will be written. Minidumps are not enabled by default on client versions of Windows
#
# An error report file with more information is saved as:
# D:\MC Server 1.20.1\hs_err_pid7428.log
#
# If you would like to submit a bug report, please visit:
# https://bugreport.java.com/bugreport/crash.jsp
#
服务器已关闭,将于20秒后重启}
**Mod list :** {jei-1.20.1-forge-15.1.0.19,create-1.20.1-0.5.1.d,curios-forge-5.2.0-beta.3+1.20.1,Jade-1.20-forge-11.1.4}
**Description of issue :** {在服务器内合成物品时有小概率崩溃,且存档会受到部分回档}
|
1.0
|
[1.20.1]在服务器内合成物品时有小概率崩溃,且存档会受到部分回档 - <!-- ISSUE_TEMPLATE_3 -> IMPORTANT: DO NOT DELETE THIS LINE.-->
<!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.-->
<!-- If you don't know where to upload your logs and crash reports, you can use these websites : -->
<!-- https://gist.github.com (recommended) -->
<!-- https://mclo.gs -->
<!-- https://haste.mohistmc.com -->
<!-- https://pastebin.com -->
<!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT -->
**Minecraft Version :** {1.20.1}
**Mohist Version :** {177}
**Operating System :** {windows 10 pro}
**Logs :** {> #
# A fatal error has been detected by the Java Runtime Environment:
#
# Internal Error (macroAssembler_x86.cpp:864), pid=7428, tid=10556
# fatal error: DEBUG MESSAGE: duplicated predicate failed which is impossible
#
# JRE version: Java(TM) SE Runtime Environment (17.0.7+8) (build 17.0.7+8-LTS-224)
# Java VM: Java HotSpot(TM) 64-Bit Server VM (17.0.7+8-LTS-224, mixed mode, sharing, tiered, compressed class ptrs, g1 gc, windows-amd64)
# No core dump will be written. Minidumps are not enabled by default on client versions of Windows
#
# An error report file with more information is saved as:
# D:\MC Server 1.20.1\hs_err_pid7428.log
#
# If you would like to submit a bug report, please visit:
# https://bugreport.java.com/bugreport/crash.jsp
#
服务器已关闭,将于20秒后重启}
**Mod list :** {jei-1.20.1-forge-15.1.0.19,create-1.20.1-0.5.1.d,curios-forge-5.2.0-beta.3+1.20.1,Jade-1.20-forge-11.1.4}
**Description of issue :** {在服务器内合成物品时有小概率崩溃,且存档会受到部分回档}
|
test
|
在服务器内合成物品时有小概率崩溃,且存档会受到部分回档 important do not delete this line minecraft version mohist version operating system windows pro logs a fatal error has been detected by the java runtime environment internal error macroassembler cpp pid tid fatal error debug message duplicated predicate failed which is impossible jre version java tm se runtime environment build lts java vm java hotspot tm bit server vm lts mixed mode sharing tiered compressed class ptrs gc windows no core dump will be written minidumps are not enabled by default on client versions of windows an error report file with more information is saved as d mc server hs err log if you would like to submit a bug report please visit 服务器已关闭, mod list jei forge ,create d,curios forge beta ,jade forge description of issue 在服务器内合成物品时有小概率崩溃,且存档会受到部分回档
| 1
|
349,005
| 31,766,072,644
|
IssuesEvent
|
2023-09-12 08:51:33
|
camunda/zeebe
|
https://api.github.com/repos/camunda/zeebe
|
closed
|
Add a test restore app class
|
area/test kind/task
|
**Description**
> **Note**
> Blocked by #13966
Based off of #13966, add a new class which allows launching a test instance of `RestoreApp`. This class should behave like any Spring application, with the added capability of specifying the `backupId` command line argument, and specifying the `RESTORE` profile.
See the hack day PR for an example: https://github.com/camunda/zeebe/blob/99892d55658536ed2715b0790ac2257cbf7293c6/qa/util/src/main/java/io/camunda/zeebe/qa/util/cluster/TestRestoreApp.java
|
1.0
|
Add a test restore app class - **Description**
> **Note**
> Blocked by #13966
Based off of #13966, add a new class which allows launching a test instance of `RestoreApp`. This class should behave like any Spring application, with the added capability of specifying the `backupId` command line argument, and specifying the `RESTORE` profile.
See the hack day PR for an example: https://github.com/camunda/zeebe/blob/99892d55658536ed2715b0790ac2257cbf7293c6/qa/util/src/main/java/io/camunda/zeebe/qa/util/cluster/TestRestoreApp.java
|
test
|
add a test restore app class description note blocked by based off of add a new class which allows launching a test instance of restoreapp this class should behave like any spring application with the added capability of specifying the backupid command line argument and specifying the restore profile see the hack day pr for an example
| 1
|
127,566
| 10,475,271,624
|
IssuesEvent
|
2019-09-23 15:59:38
|
kcigeospatial/Fred_Co_Land-Management
|
https://api.github.com/repos/kcigeospatial/Fred_Co_Land-Management
|
closed
|
Use Permit - Awaiting Fee Payment Notification - Standard Format
|
Ready for Test Env. Retest
|
Did not get an Awaiting Fee Payment notification for the Home Occupation permit.
|
2.0
|
Use Permit - Awaiting Fee Payment Notification - Standard Format - Did not get an Awaiting Fee Payment notification for the Home Occupation permit.
|
test
|
use permit awaiting fee payment notification standard format did not get an awaiting fee payment notification for the home occupation permit
| 1
|
347,371
| 31,160,491,061
|
IssuesEvent
|
2023-08-16 15:39:58
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
reopened
|
DISABLED test_mm_sparse_first_T_cuda_bfloat16 (__main__.TestSparseSemiStructuredCUDA)
|
module: sparse triaged module: flaky-tests skipped
|
Platforms: linux, slow
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_mm_sparse_first_T_cuda_bfloat16&suite=TestSparseSemiStructuredCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/15355252942).
Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_mm_sparse_first_T_cuda_bfloat16`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_sparse_semi_structured.py`
cc @alexsamardzic @nikitaved @pearu @cpuhrsch @amjames @bhosmer
|
1.0
|
DISABLED test_mm_sparse_first_T_cuda_bfloat16 (__main__.TestSparseSemiStructuredCUDA) - Platforms: linux, slow
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_mm_sparse_first_T_cuda_bfloat16&suite=TestSparseSemiStructuredCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/15355252942).
Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_mm_sparse_first_T_cuda_bfloat16`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_sparse_semi_structured.py`
cc @alexsamardzic @nikitaved @pearu @cpuhrsch @amjames @bhosmer
|
test
|
disabled test mm sparse first t cuda main testsparsesemistructuredcuda platforms linux slow this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test mm sparse first t cuda there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path test sparse semi structured py cc alexsamardzic nikitaved pearu cpuhrsch amjames bhosmer
| 1
|
61,782
| 14,640,710,321
|
IssuesEvent
|
2020-12-25 03:21:45
|
fu1771695yongxie/pm
|
https://api.github.com/repos/fu1771695yongxie/pm
|
opened
|
CVE-2018-3721 (Medium) detected in lodash-3.10.1.tgz
|
security vulnerability
|
## CVE-2018-3721 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary>
<p>The modern build of lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p>
<p>Path to dependency file: pm/package.json</p>
<p>Path to vulnerable library: pm/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- babel-core-5.8.38.tgz (Root Library)
- :x: **lodash-3.10.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/pm/commit/1c06cbe4c354bfe6922fec380958337d18de7e44">1c06cbe4c354bfe6922fec380958337d18de7e44</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash node module before 4.17.5 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via defaultsDeep, merge, and mergeWith functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3721>CVE-2018-3721</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3721">https://nvd.nist.gov/vuln/detail/CVE-2018-3721</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 4.17.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-3721 (Medium) detected in lodash-3.10.1.tgz - ## CVE-2018-3721 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary>
<p>The modern build of lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p>
<p>Path to dependency file: pm/package.json</p>
<p>Path to vulnerable library: pm/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- babel-core-5.8.38.tgz (Root Library)
- :x: **lodash-3.10.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/pm/commit/1c06cbe4c354bfe6922fec380958337d18de7e44">1c06cbe4c354bfe6922fec380958337d18de7e44</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash node module before 4.17.5 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via defaultsDeep, merge, and mergeWith functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3721>CVE-2018-3721</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3721">https://nvd.nist.gov/vuln/detail/CVE-2018-3721</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 4.17.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file pm package json path to vulnerable library pm node modules lodash package json dependency hierarchy babel core tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details lodash node module before suffers from a modification of assumed immutable data maid vulnerability via defaultsdeep merge and mergewith functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
1,847
| 2,576,389,011
|
IssuesEvent
|
2015-02-12 09:42:37
|
w3c/csswg-test
|
https://api.github.com/repos/w3c/csswg-test
|
closed
|
ime-mode
|
spec:ui testtwf
|
Might be particularly relevant here in Shenzhen, since most people presumably use an IME.
|
1.0
|
ime-mode - Might be particularly relevant here in Shenzhen, since most people presumably use an IME.
|
test
|
ime mode might be particularly relevant here in shenzhen since most people presumably use an ime
| 1
|
225,379
| 24,828,515,171
|
IssuesEvent
|
2022-10-25 23:39:26
|
snowdensb/jpo-ode
|
https://api.github.com/repos/snowdensb/jpo-ode
|
reopened
|
CVE-2022-40156 (High) detected in woodstox-core-6.2.4.jar
|
security vulnerability
|
## CVE-2022-40156 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>woodstox-core-6.2.4.jar</b></p></summary>
<p>Woodstox is a high-performance XML processor that implements Stax (JSR-173),
SAX2 and Stax2 APIs</p>
<p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p>
<p>Path to dependency file: /jpo-ode-svcs/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar</p>
<p>
Dependency Hierarchy:
- jackson-dataformat-xml-2.12.3.jar (Root Library)
- :x: **woodstox-core-6.2.4.jar** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Those using Xstream to seralize XML data may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. This effect may support a denial of service attack.
<p>Publish Date: 2022-09-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40156>CVE-2022-40156</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-09-16</p>
<p>Fix Resolution: woodstox-core-5.4.0,woodstox-core-6.4.0</p>
</p>
</details>
<p></p>
|
True
|
CVE-2022-40156 (High) detected in woodstox-core-6.2.4.jar - ## CVE-2022-40156 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>woodstox-core-6.2.4.jar</b></p></summary>
<p>Woodstox is a high-performance XML processor that implements Stax (JSR-173),
SAX2 and Stax2 APIs</p>
<p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p>
<p>Path to dependency file: /jpo-ode-svcs/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.4/woodstox-core-6.2.4.jar</p>
<p>
Dependency Hierarchy:
- jackson-dataformat-xml-2.12.3.jar (Root Library)
- :x: **woodstox-core-6.2.4.jar** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Those using Xstream to seralize XML data may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. This effect may support a denial of service attack.
<p>Publish Date: 2022-09-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40156>CVE-2022-40156</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-09-16</p>
<p>Fix Resolution: woodstox-core-5.4.0,woodstox-core-6.4.0</p>
</p>
</details>
<p></p>
|
non_test
|
cve high detected in woodstox core jar cve high severity vulnerability vulnerable library woodstox core jar woodstox is a high performance xml processor that implements stax jsr and apis library home page a href path to dependency file jpo ode svcs pom xml path to vulnerable library home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar dependency hierarchy jackson dataformat xml jar root library x woodstox core jar vulnerable library found in base branch dev vulnerability details those using xstream to seralize xml data may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow this effect may support a denial of service attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution woodstox core woodstox core
| 0
|
780,057
| 27,377,960,077
|
IssuesEvent
|
2023-02-28 07:56:14
|
ballerina-platform/ballerina-dev-website
|
https://api.github.com/repos/ballerina-platform/ballerina-dev-website
|
closed
|
Update `Style guide` page documentation to use our documenting practices
|
Priority/High Type/Task Points/0.5 Area/LearnPages Category/Content
|
## Description
$subject.
## Related website/documentation area
[1] https://ballerina.io/learn/style-guide/annotations-documentation-and-comments/#comments:~:text=%23%20%2B%20value%20%2D%20value%20input%20parameter
Our practice in documentation is,
- The description starts with a capital letter and ends with a period.
- Parameter doc lines start with a simple letter and no period at the end.
e.g.
```bal
# Adds two integers.
#
# + x - an integer
# + y - another integer
# + return - the sum of `x` and `y`
public function add(int x, int y) returns int {
return x + y;
}
```
We need to update the last example in [1] to adhere above practice.
|
1.0
|
Update `Style guide` page documentation to use our documenting practices - ## Description
$subject.
## Related website/documentation area
[1] https://ballerina.io/learn/style-guide/annotations-documentation-and-comments/#comments:~:text=%23%20%2B%20value%20%2D%20value%20input%20parameter
Our practice in documentation is,
- The description starts with a capital letter and ends with a period.
- Parameter doc lines start with a simple letter and no period at the end.
e.g.
```bal
# Adds two integers.
#
# + x - an integer
# + y - another integer
# + return - the sum of `x` and `y`
public function add(int x, int y) returns int {
return x + y;
}
```
We need to update the last example in [1] to adhere above practice.
|
non_test
|
update style guide page documentation to use our documenting practices description subject related website documentation area our practice in documentation is the description starts with a capital letter and ends with a period parameter doc lines start with a simple letter and no period at the end e g bal adds two integers x an integer y another integer return the sum of x and y public function add int x int y returns int return x y we need to update the last example in to adhere above practice
| 0
|
189,945
| 6,803,281,551
|
IssuesEvent
|
2017-11-02 23:55:50
|
sul-dlss/preservation_catalog
|
https://api.github.com/repos/sul-dlss/preservation_catalog
|
closed
|
(MV) Directories ONLY contain files
|
high priority ready
|
- content and metadata directories each contain only files.
With RSpec tests.
|
1.0
|
(MV) Directories ONLY contain files - - content and metadata directories each contain only files.
With RSpec tests.
|
non_test
|
mv directories only contain files content and metadata directories each contain only files with rspec tests
| 0
|
275,284
| 8,575,547,918
|
IssuesEvent
|
2018-11-12 17:34:56
|
aowen87/TicketTester
|
https://api.github.com/repos/aowen87/TicketTester
|
closed
|
Can we remove Sim V1 files?
|
Expected Use: 3 - Occasional Feature Impact: 3 - Medium Priority: Normal
|
I checked with Brad, and he thinks we can remove the Sim version reader/write and library files.
He suggest we keep the V2 naming, in case a V3 comes along.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1910
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Can we remove Sim V1 files?
Assigned to: Kathleen Biagas
Category:
Target version: 2.8
Author: Kathleen Biagas
Start: 07/11/2014
Due date:
% Done: 0
Estimated time:
Created: 07/11/2014 06:41 pm
Updated: 08/19/2014 08:00 pm
Likelihood:
Severity:
Found in version:
Impact: 3 - Medium
Expected Use: 3 - Occasional
OS: All
Support Group: Any
Description:
I checked with Brad, and he thinks we can remove the Sim version reader/write and library files.
He suggest we keep the V2 naming, in case a V3 comes along.
Comments:
I removed V1 from sim and removed the SimV1 reader and writer.
|
1.0
|
Can we remove Sim V1 files? - I checked with Brad, and he thinks we can remove the Sim version reader/write and library files.
He suggest we keep the V2 naming, in case a V3 comes along.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1910
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Can we remove Sim V1 files?
Assigned to: Kathleen Biagas
Category:
Target version: 2.8
Author: Kathleen Biagas
Start: 07/11/2014
Due date:
% Done: 0
Estimated time:
Created: 07/11/2014 06:41 pm
Updated: 08/19/2014 08:00 pm
Likelihood:
Severity:
Found in version:
Impact: 3 - Medium
Expected Use: 3 - Occasional
OS: All
Support Group: Any
Description:
I checked with Brad, and he thinks we can remove the Sim version reader/write and library files.
He suggest we keep the V2 naming, in case a V3 comes along.
Comments:
I removed V1 from sim and removed the SimV1 reader and writer.
|
non_test
|
can we remove sim files i checked with brad and he thinks we can remove the sim version reader write and library files he suggest we keep the naming in case a comes along redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker feature priority normal subject can we remove sim files assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood severity found in version impact medium expected use occasional os all support group any description i checked with brad and he thinks we can remove the sim version reader write and library files he suggest we keep the naming in case a comes along comments i removed from sim and removed the reader and writer
| 0
|
328,596
| 9,997,129,421
|
IssuesEvent
|
2019-07-12 02:54:40
|
momentum-mod/website
|
https://api.github.com/repos/momentum-mod/website
|
closed
|
Allow embedding YouTube videos for maps
|
Enhancement Priority: Medium Size: Medium
|
In accordance with #282 , we need to be able to support youtube videos in the map info table to be able to actually show them.
[youtubesurfvideos JSON data from BorkChops](https://github.com/momentum-mod/website/files/3364441/youtubesurfvideos.txt)
|
1.0
|
Allow embedding YouTube videos for maps - In accordance with #282 , we need to be able to support youtube videos in the map info table to be able to actually show them.
[youtubesurfvideos JSON data from BorkChops](https://github.com/momentum-mod/website/files/3364441/youtubesurfvideos.txt)
|
non_test
|
allow embedding youtube videos for maps in accordance with we need to be able to support youtube videos in the map info table to be able to actually show them
| 0
|
59,862
| 8,381,207,251
|
IssuesEvent
|
2018-10-07 22:30:17
|
damienbod/angular-auth-oidc-client
|
https://api.github.com/repos/damienbod/angular-auth-oidc-client
|
closed
|
Refreshing the page wipes all storage data
|
enhancement enhancement documentation
|
Hi,
i'm doing a POC using your library and i got this issue.
So basically i'm able to login, get token, send the token using http interceptors, all work fine. My api check the token, all good!
But if i refresh the page, all auth storage data is wiped!
On my debug, i found out that when calling authorizedCallback, it then calls resetAuthorizationData straight away wiping the data.
https://github.com/damienbod/angular-auth-oidc-client/blob/40790b8909f5218536c6f4f92d7f8329ac5b225d/src/services/oidc.security.service.ts#L194.
Is this the expected behaviour? i'm doing whats in the read.me (configure on the app.module, setting up in app.component...
any thoughts ?
Thanks in advance
|
1.0
|
Refreshing the page wipes all storage data - Hi,
i'm doing a POC using your library and i got this issue.
So basically i'm able to login, get token, send the token using http interceptors, all work fine. My api check the token, all good!
But if i refresh the page, all auth storage data is wiped!
On my debug, i found out that when calling authorizedCallback, it then calls resetAuthorizationData straight away wiping the data.
https://github.com/damienbod/angular-auth-oidc-client/blob/40790b8909f5218536c6f4f92d7f8329ac5b225d/src/services/oidc.security.service.ts#L194.
Is this the expected behaviour? i'm doing whats in the read.me (configure on the app.module, setting up in app.component...
any thoughts ?
Thanks in advance
|
non_test
|
refreshing the page wipes all storage data hi i m doing a poc using your library and i got this issue so basically i m able to login get token send the token using http interceptors all work fine my api check the token all good but if i refresh the page all auth storage data is wiped on my debug i found out that when calling authorizedcallback it then calls resetauthorizationdata straight away wiping the data is this the expected behaviour i m doing whats in the read me configure on the app module setting up in app component any thoughts thanks in advance
| 0
|
144,562
| 11,624,176,032
|
IssuesEvent
|
2020-02-27 10:17:41
|
SPW-DIG/metawal-core-geonetwork
|
https://api.github.com/repos/SPW-DIG/metawal-core-geonetwork
|
closed
|
Ressource en ligne / Protocole DB & FILE
|
Env prod - OK Env test - OK Env valid - OK
|
- [x] Editeur / Ajout de choix DB et FILE:RASTER/VECTOR dans la liste de choix
* https://github.com/geonetwork/core-geonetwork/pull/4383
* https://github.com/metadata101/iso19115-3.2018/pull/23
```xml
<mrd:onLine>
<cit:CI_OnlineResource>
<cit:linkage>
<gco:CharacterString>dbora:schema</gco:CharacterString>
</cit:linkage>
<cit:protocol>
<gco:CharacterString>DB:ORACLE</gco:CharacterString>
</cit:protocol>
<cit:name>
<gco:CharacterString>tablename</gco:CharacterString>
</cit:name>
<cit:function>
<cit:CI_OnLineFunctionCode codeList="http://standards.iso.org/iso/19139/resources/gmxCodelists.xml#CI_OnLineFunctionCode"
codeListValue="fileAccess"/>
</cit:function>
</cit:CI_OnlineResource>
</mrd:onLine>
<mrd:onLine>
<cit:CI_OnlineResource>
<cit:linkage gco:nilReason="withheld">
<gco:CharacterString>/geo/data/dgo3_dada.ecw</gco:CharacterString>
</cit:linkage>
<cit:protocol>
<gco:CharacterString>FILE:RASTER</gco:CharacterString>
</cit:protocol>
<cit:name>
<gco:CharacterString>Données au format ECW</gco:CharacterString>
</cit:name>
<cit:function>
<cit:CI_OnLineFunctionCode codeList="http://standards.iso.org/iso/19139/resources/gmxCodelists.xml#CI_OnLineFunctionCode"
codeListValue="fileAccess"/>
</cit:function>
</cit:CI_OnlineResource>
</mrd:onLine>
```
- [x] UFO / Pas de withheld
- [x] Publish / Do not check download by default. Intranet / Check all. GN ref https://github.com/geonetwork/core-geonetwork/pull/4405
- [ ] Tester que quand il y a download pour un groupe, les infos ne sont pas affichées
|
1.0
|
Ressource en ligne / Protocole DB & FILE - - [x] Editeur / Ajout de choix DB et FILE:RASTER/VECTOR dans la liste de choix
* https://github.com/geonetwork/core-geonetwork/pull/4383
* https://github.com/metadata101/iso19115-3.2018/pull/23
```xml
<mrd:onLine>
<cit:CI_OnlineResource>
<cit:linkage>
<gco:CharacterString>dbora:schema</gco:CharacterString>
</cit:linkage>
<cit:protocol>
<gco:CharacterString>DB:ORACLE</gco:CharacterString>
</cit:protocol>
<cit:name>
<gco:CharacterString>tablename</gco:CharacterString>
</cit:name>
<cit:function>
<cit:CI_OnLineFunctionCode codeList="http://standards.iso.org/iso/19139/resources/gmxCodelists.xml#CI_OnLineFunctionCode"
codeListValue="fileAccess"/>
</cit:function>
</cit:CI_OnlineResource>
</mrd:onLine>
<mrd:onLine>
<cit:CI_OnlineResource>
<cit:linkage gco:nilReason="withheld">
<gco:CharacterString>/geo/data/dgo3_dada.ecw</gco:CharacterString>
</cit:linkage>
<cit:protocol>
<gco:CharacterString>FILE:RASTER</gco:CharacterString>
</cit:protocol>
<cit:name>
<gco:CharacterString>Données au format ECW</gco:CharacterString>
</cit:name>
<cit:function>
<cit:CI_OnLineFunctionCode codeList="http://standards.iso.org/iso/19139/resources/gmxCodelists.xml#CI_OnLineFunctionCode"
codeListValue="fileAccess"/>
</cit:function>
</cit:CI_OnlineResource>
</mrd:onLine>
```
- [x] UFO / Pas de withheld
- [x] Publish / Do not check download by default. Intranet / Check all. GN ref https://github.com/geonetwork/core-geonetwork/pull/4405
- [ ] Tester que quand il y a download pour un groupe, les infos ne sont pas affichées
|
test
|
ressource en ligne protocole db file editeur ajout de choix db et file raster vector dans la liste de choix xml dbora schema db oracle tablename cit ci onlinefunctioncode codelist codelistvalue fileaccess geo data dada ecw file raster données au format ecw cit ci onlinefunctioncode codelist codelistvalue fileaccess ufo pas de withheld publish do not check download by default intranet check all gn ref tester que quand il y a download pour un groupe les infos ne sont pas affichées
| 1
|
219,908
| 17,119,793,843
|
IssuesEvent
|
2021-07-12 02:25:46
|
aimakerspace/PeekingDuck
|
https://api.github.com/repos/aimakerspace/PeekingDuck
|
closed
|
In the unit tests, windows appends double backslash instead of forward slash in os.path.join
|
testing
|
In windows, the os.path.join use double backslash, thus this affect the assertion checks in the unit test that assumed forward slash in the output. (As seen below)
Unit Tests affected:
- test_declarativeloader
- test_configloader
- test_runner
Suggestion:
To do a string replacement before assert statement to change double backslash into 1 forward slash

|
1.0
|
In the unit tests, windows appends double backslash instead of forward slash in os.path.join - In windows, the os.path.join use double backslash, thus this affect the assertion checks in the unit test that assumed forward slash in the output. (As seen below)
Unit Tests affected:
- test_declarativeloader
- test_configloader
- test_runner
Suggestion:
To do a string replacement before assert statement to change double backslash into 1 forward slash

|
test
|
in the unit tests windows appends double backslash instead of forward slash in os path join in windows the os path join use double backslash thus this affect the assertion checks in the unit test that assumed forward slash in the output as seen below unit tests affected test declarativeloader test configloader test runner suggestion to do a string replacement before assert statement to change double backslash into forward slash
| 1
|
380,050
| 26,399,494,053
|
IssuesEvent
|
2023-01-12 23:07:05
|
aptos-labs/aptos-core
|
https://api.github.com/repos/aptos-labs/aptos-core
|
opened
|
Add documentation for running fullnodes on AWS and Azure
|
documentation
|
# Add documentation for running fullnodes on AWS and Azure
<!-- A clear and concise description of the feature you are requesting -->
## Motivation
We have documentation for running fullnodes on source/docker and GCP but not for AWS and Azure
We should have multiple cloud provider options similar to our validator node docs
## Pitch
**Describe the solution you'd like**
Replicate this doc https://aptos.dev/nodes/full-node/run-a-fullnode-on-gcp/ for AWS and Azure
## Additional context
Refer to validator + VFN instructions:
https://aptos.dev/nodes/validator-node/operator/running-validator-node/run-validator-node-using-azure
https://aptos.dev/nodes/validator-node/operator/running-validator-node/run-validator-node-using-aws
Initial steps are the same but replace 'aptos-node' module with 'fullnode' module?
|
1.0
|
Add documentation for running fullnodes on AWS and Azure - # Add documentation for running fullnodes on AWS and Azure
<!-- A clear and concise description of the feature you are requesting -->
## Motivation
We have documentation for running fullnodes on source/docker and GCP but not for AWS and Azure
We should have multiple cloud provider options similar to our validator node docs
## Pitch
**Describe the solution you'd like**
Replicate this doc https://aptos.dev/nodes/full-node/run-a-fullnode-on-gcp/ for AWS and Azure
## Additional context
Refer to validator + VFN instructions:
https://aptos.dev/nodes/validator-node/operator/running-validator-node/run-validator-node-using-azure
https://aptos.dev/nodes/validator-node/operator/running-validator-node/run-validator-node-using-aws
Initial steps are the same but replace 'aptos-node' module with 'fullnode' module?
|
non_test
|
add documentation for running fullnodes on aws and azure add documentation for running fullnodes on aws and azure motivation we have documentation for running fullnodes on source docker and gcp but not for aws and azure we should have multiple cloud provider options similar to our validator node docs pitch describe the solution you d like replicate this doc for aws and azure additional context refer to validator vfn instructions initial steps are the same but replace aptos node module with fullnode module
| 0
|
344,511
| 30,749,907,916
|
IssuesEvent
|
2023-07-28 18:13:16
|
saltstack/salt
|
https://api.github.com/repos/saltstack/salt
|
opened
|
[Increase Test Coverage] Batch 9
|
Tests
|
Increase the code coverage percent on the following files to at least 80%.
File | Percent
salt/modules/win_pkg.py | 36
salt/auth/__init__.py | 62
salt/payload.py | 68
salt/states/service.py | 58
salt/modules/win_file.py | 61
Please be aware that currently the percentage might be inaccurate if the module uses __salt__ due to https://github.com/saltstack/salt/issues/64696
|
1.0
|
[Increase Test Coverage] Batch 9 - Increase the code coverage percent on the following files to at least 80%.
File | Percent
salt/modules/win_pkg.py | 36
salt/auth/__init__.py | 62
salt/payload.py | 68
salt/states/service.py | 58
salt/modules/win_file.py | 61
Please be aware that currently the percentage might be inaccurate if the module uses __salt__ due to https://github.com/saltstack/salt/issues/64696
|
test
|
batch increase the code coverage percent on the following files to at least file percent salt modules win pkg py salt auth init py salt payload py salt states service py salt modules win file py please be aware that currently the percentage might be inaccurate if the module uses salt due to
| 1
|
226,959
| 18,045,975,745
|
IssuesEvent
|
2021-09-18 22:42:44
|
logicmoo/logicmoo_workspace
|
https://api.github.com/repos/logicmoo/logicmoo_workspace
|
opened
|
logicmoo.pfc.test.sanity_base.MT_03 JUnit
|
Test_9999 logicmoo.pfc.test.sanity_base unit_test MT_03
|
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif mt_03.pl)
GH_MASTER_ISSUE_FINFO=
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AMT_03
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/1629eba4a2a1da0e1b731d156198a7168dafae44
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/68/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/1629eba4a2a1da0e1b731d156198a7168dafae44
https://github.com/logicmoo/logicmoo_workspace/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl
```
%
running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/mt_03.pl'),
%~ this_test_might_need( :-( use_module( library(logicmoo_plarkc))))
%~ this_test_might_need( :-( expects_dialect(pfc)))
%:- add_import_module(header_sane,baseKB,end).
:- set_defaultAssertMt(myMt).
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/mt_03.pl:17
%~ pfc_iri : include_module_file(myMt:library('pfclib/system_each_module.pfc'),myMt).
/*~
%~ pfc_iri:include_module_file(myMt:library('pfclib/system_each_module.pfc'),myMt)
~*/
:- expects_dialect(pfc).
:- mpred_trace_exec.
mtProlog(modA).
No source location!?
%~ message_hook_type(error)
%~ message_hook(
%~ error(
%~ permission_error(redefine,imported_procedure,baseKB:mtProlog/1),
%~ context(system:'$record_clause'/3,Context_Kw)),
%~ error,
%~ [ 'No permission to ~w ~w `~p\'' - [ redefine,
%~ imported_procedure,
%~ baseKB : mtProlog/1]])
/*~
No permission to redefine imported_procedure `baseKB:(mtProlog/1)'
ERROR: No permission to redefine imported_procedure `baseKB:(mtProlog/1)'
~*/
mtProlog(modB).
No source location!?
%~ message_hook_type(error)
%~ message_hook(
%~ error(
%~ permission_error(redefine,imported_procedure,baseKB:mtProlog/1),
%~ context(system:'$record_clause'/3,Context_Kw)),
%~ error,
%~ [ 'No permission to ~w ~w `~p\'' - [ redefine,
%~ imported_procedure,
%~ baseKB : mtProlog/1]])
/*~
No permission to redefine imported_procedure `baseKB:(mtProlog/1)'
ERROR: No permission to redefine imported_procedure `baseKB:(mtProlog/1)'
~*/
modA: (codeA:- printAll('$current_source_module'(_M)),codeB).
No source location!?
modB: (codeB).
%:- \+ modA:codeA.
No source location!?
%:- \+ modA:codeA.
genlMt(modA,modB).
% before test, to make sure codeA was not accdently defined in modB
% before test, to make sure codeA was not accdently defined in modB
:- sanity(\+ module_clause(modB:codeA,_)).
:- sanity(\+ module_clause(modA:codeB,_)).
:- sanity( module_clause(modA:codeA,_)).
:- sanity( module_clause(modB:codeB,_)).
% before test, genlMt makes the rule available and should not corrupt the modA module
% before test, genlMt makes the rule available and should not corrupt the modA module
:- warn_fail_TODO(clause_u(modA:codeB,_)).
% make sure genlMt didnt unassert
%~ :-( warn_fail_TODO( clause_u(modA:codeB,Kw))).
% make sure genlMt didnt unassert
:- sanity(clause_u(modB:codeB,_)).
% run the test
% run the test
modA: (:- codeA).
% to make codeB sure is available in modA
No source location!?
% to make codeB sure is available in modA
:- mpred_must( clause_u(modA:codeB,_)).
% to make sure codeA does not get accdently defined in modB
%~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56
%~ failed_mpred_test( clause_u(modA:codeB,Kw))
%~ FILE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56
%~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56
%~ FILE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/mt_03.pl:56
%~ DUMP_BREAK/0
%~ message_hook_type(error)
%~ message_hook( initialization_exception(abort),
%~ error,
%~ [ 'Prolog initialisation failed:', nl,'Unknown message: ~p'-[abort]])
%~ unused(save_junit_results)
```
totalTime=3
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AMT_03
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/1629eba4a2a1da0e1b731d156198a7168dafae44
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/68/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/1629eba4a2a1da0e1b731d156198a7168dafae44
https://github.com/logicmoo/logicmoo_workspace/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl
FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k mt_03.pl (returned 1)
|
3.0
|
logicmoo.pfc.test.sanity_base.MT_03 JUnit - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif mt_03.pl)
GH_MASTER_ISSUE_FINFO=
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AMT_03
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/1629eba4a2a1da0e1b731d156198a7168dafae44
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/68/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/1629eba4a2a1da0e1b731d156198a7168dafae44
https://github.com/logicmoo/logicmoo_workspace/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl
```
%
running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/mt_03.pl'),
%~ this_test_might_need( :-( use_module( library(logicmoo_plarkc))))
%~ this_test_might_need( :-( expects_dialect(pfc)))
%:- add_import_module(header_sane,baseKB,end).
:- set_defaultAssertMt(myMt).
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/mt_03.pl:17
%~ pfc_iri : include_module_file(myMt:library('pfclib/system_each_module.pfc'),myMt).
/*~
%~ pfc_iri:include_module_file(myMt:library('pfclib/system_each_module.pfc'),myMt)
~*/
:- expects_dialect(pfc).
:- mpred_trace_exec.
mtProlog(modA).
No source location!?
%~ message_hook_type(error)
%~ message_hook(
%~ error(
%~ permission_error(redefine,imported_procedure,baseKB:mtProlog/1),
%~ context(system:'$record_clause'/3,Context_Kw)),
%~ error,
%~ [ 'No permission to ~w ~w `~p\'' - [ redefine,
%~ imported_procedure,
%~ baseKB : mtProlog/1]])
/*~
No permission to redefine imported_procedure `baseKB:(mtProlog/1)'
ERROR: No permission to redefine imported_procedure `baseKB:(mtProlog/1)'
~*/
mtProlog(modB).
No source location!?
%~ message_hook_type(error)
%~ message_hook(
%~ error(
%~ permission_error(redefine,imported_procedure,baseKB:mtProlog/1),
%~ context(system:'$record_clause'/3,Context_Kw)),
%~ error,
%~ [ 'No permission to ~w ~w `~p\'' - [ redefine,
%~ imported_procedure,
%~ baseKB : mtProlog/1]])
/*~
No permission to redefine imported_procedure `baseKB:(mtProlog/1)'
ERROR: No permission to redefine imported_procedure `baseKB:(mtProlog/1)'
~*/
modA: (codeA:- printAll('$current_source_module'(_M)),codeB).
No source location!?
modB: (codeB).
%:- \+ modA:codeA.
No source location!?
%:- \+ modA:codeA.
genlMt(modA,modB).
% before test, to make sure codeA was not accdently defined in modB
% before test, to make sure codeA was not accdently defined in modB
:- sanity(\+ module_clause(modB:codeA,_)).
:- sanity(\+ module_clause(modA:codeB,_)).
:- sanity( module_clause(modA:codeA,_)).
:- sanity( module_clause(modB:codeB,_)).
% before test, genlMt makes the rule available and should not corrupt the modA module
% before test, genlMt makes the rule available and should not corrupt the modA module
:- warn_fail_TODO(clause_u(modA:codeB,_)).
% make sure genlMt didnt unassert
%~ :-( warn_fail_TODO( clause_u(modA:codeB,Kw))).
% make sure genlMt didnt unassert
:- sanity(clause_u(modB:codeB,_)).
% run the test
% run the test
modA: (:- codeA).
% to make codeB sure is available in modA
No source location!?
% to make codeB sure is available in modA
:- mpred_must( clause_u(modA:codeB,_)).
% to make sure codeA does not get accdently defined in modB
%~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56
%~ failed_mpred_test( clause_u(modA:codeB,Kw))
%~ FILE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56
%~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56
%~ FILE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/mt_03.pl#L56
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/mt_03.pl:56
%~ DUMP_BREAK/0
%~ message_hook_type(error)
%~ message_hook( initialization_exception(abort),
%~ error,
%~ [ 'Prolog initialisation failed:', nl,'Unknown message: ~p'-[abort]])
%~ unused(save_junit_results)
```
totalTime=3
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AMT_03
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/1629eba4a2a1da0e1b731d156198a7168dafae44
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/68/testReport/logicmoo.pfc.test.sanity_base/MT_03/logicmoo_pfc_test_sanity_base_MT_03_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/1629eba4a2a1da0e1b731d156198a7168dafae44
https://github.com/logicmoo/logicmoo_workspace/blob/1629eba4a2a1da0e1b731d156198a7168dafae44/packs_sys/pfc/t/sanity_base/mt_03.pl
FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k mt_03.pl (returned 1)
|
test
|
logicmoo pfc test sanity base mt junit cd var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base timeout foreground preserve status s sigkill k lmoo clif mt pl gh master issue finfo issue search gitlab latest this build github running var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base mt pl this test might need use module library logicmoo plarkc this test might need expects dialect pfc add import module header sane basekb end set defaultassertmt mymt var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base mt pl pfc iri include module file mymt library pfclib system each module pfc mymt pfc iri include module file mymt library pfclib system each module pfc mymt expects dialect pfc mpred trace exec mtprolog moda no source location message hook type error message hook error permission error redefine imported procedure basekb mtprolog context system record clause context kw error no permission to w w p redefine imported procedure basekb mtprolog no permission to redefine imported procedure basekb mtprolog error no permission to redefine imported procedure basekb mtprolog mtprolog modb no source location message hook type error message hook error permission error redefine imported procedure basekb mtprolog context system record clause context kw error no permission to w w p redefine imported procedure basekb mtprolog no permission to redefine imported procedure basekb mtprolog error no permission to redefine imported procedure basekb mtprolog moda codea printall current source module m codeb no source location modb codeb moda codea no source location moda codea genlmt moda modb before test to make sure codea was not accdently defined in modb before test to make sure codea was not accdently defined in modb sanity module clause modb codea sanity module clause moda codeb sanity module clause moda codea sanity module clause modb codeb before test genlmt makes the rule available and should not corrupt the moda module before test genlmt makes the rule available and should not corrupt the moda module warn fail todo clause u moda codeb make sure genlmt didnt unassert warn fail todo clause u moda codeb kw make sure genlmt didnt unassert sanity clause u modb codeb run the test run the test moda codea to make codeb sure is available in moda no source location to make codeb sure is available in moda mpred must clause u moda codeb to make sure codea does not get accdently defined in modb file failed mpred test clause u moda codeb kw file file file var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base mt pl dump break message hook type error message hook initialization exception abort error unused save junit results totaltime issue search gitlab latest this build github failed var lib jenkins workspace logicmoo workspace bin lmoo junit minor k mt pl returned
| 1
|
124,621
| 17,772,676,455
|
IssuesEvent
|
2021-08-30 15:18:44
|
kapseliboi/html2canvas
|
https://api.github.com/repos/kapseliboi/html2canvas
|
opened
|
CVE-2020-15168 (Medium) detected in node-fetch-1.7.3.tgz
|
security vulnerability
|
## CVE-2020-15168 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-fetch-1.7.3.tgz</b></p></summary>
<p>A light-weight module that brings window.fetch to node.js and io.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-1.7.3.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-1.7.3.tgz</a></p>
<p>Path to dependency file: html2canvas/www/package.json</p>
<p>Path to vulnerable library: html2canvas/www/node_modules/node-fetch/package.json</p>
<p>
Dependency Hierarchy:
- glamor-2.20.40.tgz (Root Library)
- fbjs-0.8.17.tgz
- isomorphic-fetch-2.2.1.tgz
- :x: **node-fetch-1.7.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/html2canvas/commit/0ae2bdc652fe2e15c2adc0e9e9d841a564f7053d">0ae2bdc652fe2e15c2adc0e9e9d841a564f7053d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
node-fetch before versions 2.6.1 and 3.0.0-beta.9 did not honor the size option after following a redirect, which means that when a content size was over the limit, a FetchError would never get thrown and the process would end without failure. For most people, this fix will have a little or no impact. However, if you are relying on node-fetch to gate files above a size, the impact could be significant, for example: If you don't double-check the size of the data after fetch() has completed, your JS thread could get tied up doing work on a large file (DoS) and/or cost you money in computing.
<p>Publish Date: 2020-09-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15168>CVE-2020-15168</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/node-fetch/node-fetch/security/advisories/GHSA-w7rc-rwvf-8q5r">https://github.com/node-fetch/node-fetch/security/advisories/GHSA-w7rc-rwvf-8q5r</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 2.6.1,3.0.0-beta.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-15168 (Medium) detected in node-fetch-1.7.3.tgz - ## CVE-2020-15168 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-fetch-1.7.3.tgz</b></p></summary>
<p>A light-weight module that brings window.fetch to node.js and io.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-1.7.3.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-1.7.3.tgz</a></p>
<p>Path to dependency file: html2canvas/www/package.json</p>
<p>Path to vulnerable library: html2canvas/www/node_modules/node-fetch/package.json</p>
<p>
Dependency Hierarchy:
- glamor-2.20.40.tgz (Root Library)
- fbjs-0.8.17.tgz
- isomorphic-fetch-2.2.1.tgz
- :x: **node-fetch-1.7.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/html2canvas/commit/0ae2bdc652fe2e15c2adc0e9e9d841a564f7053d">0ae2bdc652fe2e15c2adc0e9e9d841a564f7053d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
node-fetch before versions 2.6.1 and 3.0.0-beta.9 did not honor the size option after following a redirect, which means that when a content size was over the limit, a FetchError would never get thrown and the process would end without failure. For most people, this fix will have a little or no impact. However, if you are relying on node-fetch to gate files above a size, the impact could be significant, for example: If you don't double-check the size of the data after fetch() has completed, your JS thread could get tied up doing work on a large file (DoS) and/or cost you money in computing.
<p>Publish Date: 2020-09-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15168>CVE-2020-15168</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/node-fetch/node-fetch/security/advisories/GHSA-w7rc-rwvf-8q5r">https://github.com/node-fetch/node-fetch/security/advisories/GHSA-w7rc-rwvf-8q5r</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 2.6.1,3.0.0-beta.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in node fetch tgz cve medium severity vulnerability vulnerable library node fetch tgz a light weight module that brings window fetch to node js and io js library home page a href path to dependency file www package json path to vulnerable library www node modules node fetch package json dependency hierarchy glamor tgz root library fbjs tgz isomorphic fetch tgz x node fetch tgz vulnerable library found in head commit a href found in base branch master vulnerability details node fetch before versions and beta did not honor the size option after following a redirect which means that when a content size was over the limit a fetcherror would never get thrown and the process would end without failure for most people this fix will have a little or no impact however if you are relying on node fetch to gate files above a size the impact could be significant for example if you don t double check the size of the data after fetch has completed your js thread could get tied up doing work on a large file dos and or cost you money in computing publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution beta step up your open source security game with whitesource
| 0
|
144,608
| 19,292,292,920
|
IssuesEvent
|
2021-12-12 01:27:32
|
rvvergara/next-js-basic
|
https://api.github.com/repos/rvvergara/next-js-basic
|
opened
|
CVE-2021-43803 (High) detected in next-9.3.2.tgz
|
security vulnerability
|
## CVE-2021-43803 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>next-9.3.2.tgz</b></p></summary>
<p>The React Framework</p>
<p>Library home page: <a href="https://registry.npmjs.org/next/-/next-9.3.2.tgz">https://registry.npmjs.org/next/-/next-9.3.2.tgz</a></p>
<p>Path to dependency file: next-js-basic/package.json</p>
<p>Path to vulnerable library: next-js-basic/node_modules/next/package.json</p>
<p>
Dependency Hierarchy:
- :x: **next-9.3.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Next.js is a React framework. In versions of Next.js prior to 12.0.5 or 11.1.3, invalid or malformed URLs could lead to a server crash. In order to be affected by this issue, the deployment must use Next.js versions above 11.1.0 and below 12.0.5, Node.js above 15.0.0, and next start or a custom server. Deployments on Vercel are not affected, along with similar environments where invalid requests are filtered before reaching Next.js. Versions 12.0.5 and 11.1.3 contain patches for this issue.
<p>Publish Date: 2021-12-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43803>CVE-2021-43803</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/vercel/next.js/security/advisories/GHSA-25mp-g6fv-mqxx">https://github.com/vercel/next.js/security/advisories/GHSA-25mp-g6fv-mqxx</a></p>
<p>Release Date: 2021-12-10</p>
<p>Fix Resolution: next - 11.1.3,12.0.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-43803 (High) detected in next-9.3.2.tgz - ## CVE-2021-43803 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>next-9.3.2.tgz</b></p></summary>
<p>The React Framework</p>
<p>Library home page: <a href="https://registry.npmjs.org/next/-/next-9.3.2.tgz">https://registry.npmjs.org/next/-/next-9.3.2.tgz</a></p>
<p>Path to dependency file: next-js-basic/package.json</p>
<p>Path to vulnerable library: next-js-basic/node_modules/next/package.json</p>
<p>
Dependency Hierarchy:
- :x: **next-9.3.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Next.js is a React framework. In versions of Next.js prior to 12.0.5 or 11.1.3, invalid or malformed URLs could lead to a server crash. In order to be affected by this issue, the deployment must use Next.js versions above 11.1.0 and below 12.0.5, Node.js above 15.0.0, and next start or a custom server. Deployments on Vercel are not affected, along with similar environments where invalid requests are filtered before reaching Next.js. Versions 12.0.5 and 11.1.3 contain patches for this issue.
<p>Publish Date: 2021-12-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43803>CVE-2021-43803</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/vercel/next.js/security/advisories/GHSA-25mp-g6fv-mqxx">https://github.com/vercel/next.js/security/advisories/GHSA-25mp-g6fv-mqxx</a></p>
<p>Release Date: 2021-12-10</p>
<p>Fix Resolution: next - 11.1.3,12.0.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in next tgz cve high severity vulnerability vulnerable library next tgz the react framework library home page a href path to dependency file next js basic package json path to vulnerable library next js basic node modules next package json dependency hierarchy x next tgz vulnerable library found in base branch master vulnerability details next js is a react framework in versions of next js prior to or invalid or malformed urls could lead to a server crash in order to be affected by this issue the deployment must use next js versions above and below node js above and next start or a custom server deployments on vercel are not affected along with similar environments where invalid requests are filtered before reaching next js versions and contain patches for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution next step up your open source security game with whitesource
| 0
|
86,680
| 8,042,556,289
|
IssuesEvent
|
2018-07-31 08:32:48
|
SuperuserLabs/thankful
|
https://api.github.com/repos/SuperuserLabs/thankful
|
opened
|
Add ability to run as plain webapp
|
improves: testing/qa priority: low
|
Continues work started in #70
- [ ] Fix all the errors from missing `browser` etc
- [ ] Fix Metamask support when not in a WebExtension
- [ ] Polyfill `browser.storage.local` to use `window.localStorage` when in webapp-mode
|
1.0
|
Add ability to run as plain webapp - Continues work started in #70
- [ ] Fix all the errors from missing `browser` etc
- [ ] Fix Metamask support when not in a WebExtension
- [ ] Polyfill `browser.storage.local` to use `window.localStorage` when in webapp-mode
|
test
|
add ability to run as plain webapp continues work started in fix all the errors from missing browser etc fix metamask support when not in a webextension polyfill browser storage local to use window localstorage when in webapp mode
| 1
|
301,467
| 26,050,575,761
|
IssuesEvent
|
2022-12-22 18:13:31
|
vegaprotocol/vega
|
https://api.github.com/repos/vegaprotocol/vega
|
closed
|
Implement test coverage for 0038-OLIQ-008
|
feature tests
|
IN order to get test coverage for 0038-OLIQ-liquidity_provision_order_type.md we need to cover the following ACs
- [ ] 0038-OLIQ-008
|
1.0
|
Implement test coverage for 0038-OLIQ-008 - IN order to get test coverage for 0038-OLIQ-liquidity_provision_order_type.md we need to cover the following ACs
- [ ] 0038-OLIQ-008
|
test
|
implement test coverage for oliq in order to get test coverage for oliq liquidity provision order type md we need to cover the following acs oliq
| 1
|
183,261
| 14,219,402,452
|
IssuesEvent
|
2020-11-17 13:14:19
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
util/stop: TestStopperIsStopped failed
|
C-test-failure O-robot branch-master
|
[(util/stop).TestStopperIsStopped failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2448413&tab=buildLog) on [master@807755c5473da7f3ad384fe8bb04b804ab56f911](https://github.com/cockroachdb/cockroach/commits/807755c5473da7f3ad384fe8bb04b804ab56f911):
```
=== RUN TestStopperIsStopped
I201116 23:11:04.028342 83 util/stop/stopper.go:564 quiescing
stopper.go:98: leaked stopper, created at:
goroutine 82 [running]:
runtime/debug.Stack(0x3306240, 0xc000610000, 0x0)
/usr/local/go/src/runtime/debug/stack.go:24 +0xab
github.com/cockroachdb/cockroach/pkg/util/stop.register(0xc000158960)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:52 +0x45
github.com/cockroachdb/cockroach/pkg/util/stop.NewStopper(0x0, 0x0, 0x0, 0x5c7d51)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0x39c
github.com/cockroachdb/cockroach/pkg/util/stop_test.TestStopperIsStopped(0xc000092900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper_test.go:88 +0xa5
testing.tRunner(0xc000092900, 0x217b400)
/usr/local/go/src/testing/testing.go:1123 +0x203
created by testing.(*T).Run
/usr/local/go/src/testing/testing.go:1168 +0x5bc
--- FAIL: TestStopperIsStopped (0.10s)
```
<details><summary>More</summary><p>
Parameters:
- GOFLAGS=-json
```
make stressrace TESTS=TestStopperIsStopped PKG=./pkg/util/stop TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestStopperIsStopped.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
1.0
|
util/stop: TestStopperIsStopped failed - [(util/stop).TestStopperIsStopped failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2448413&tab=buildLog) on [master@807755c5473da7f3ad384fe8bb04b804ab56f911](https://github.com/cockroachdb/cockroach/commits/807755c5473da7f3ad384fe8bb04b804ab56f911):
```
=== RUN TestStopperIsStopped
I201116 23:11:04.028342 83 util/stop/stopper.go:564 quiescing
stopper.go:98: leaked stopper, created at:
goroutine 82 [running]:
runtime/debug.Stack(0x3306240, 0xc000610000, 0x0)
/usr/local/go/src/runtime/debug/stack.go:24 +0xab
github.com/cockroachdb/cockroach/pkg/util/stop.register(0xc000158960)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:52 +0x45
github.com/cockroachdb/cockroach/pkg/util/stop.NewStopper(0x0, 0x0, 0x0, 0x5c7d51)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:189 +0x39c
github.com/cockroachdb/cockroach/pkg/util/stop_test.TestStopperIsStopped(0xc000092900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper_test.go:88 +0xa5
testing.tRunner(0xc000092900, 0x217b400)
/usr/local/go/src/testing/testing.go:1123 +0x203
created by testing.(*T).Run
/usr/local/go/src/testing/testing.go:1168 +0x5bc
--- FAIL: TestStopperIsStopped (0.10s)
```
<details><summary>More</summary><p>
Parameters:
- GOFLAGS=-json
```
make stressrace TESTS=TestStopperIsStopped PKG=./pkg/util/stop TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestStopperIsStopped.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
test
|
util stop teststopperisstopped failed on run teststopperisstopped util stop stopper go quiescing stopper go leaked stopper created at goroutine runtime debug stack usr local go src runtime debug stack go github com cockroachdb cockroach pkg util stop register go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg util stop newstopper go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg util stop test teststopperisstopped go src github com cockroachdb cockroach pkg util stop stopper test go testing trunner usr local go src testing testing go created by testing t run usr local go src testing testing go fail teststopperisstopped more parameters goflags json make stressrace tests teststopperisstopped pkg pkg util stop testtimeout stressflags timeout powered by
| 1
|
31,252
| 7,333,057,416
|
IssuesEvent
|
2018-03-05 18:10:28
|
dotnet/roslyn-analyzers
|
https://api.github.com/repos/dotnet/roslyn-analyzers
|
closed
|
CA2100 does not fire for msdn/docs sample
|
Area-Microsoft.CodeQuality.Analyzers Bug Dependency-DataFlow
|
#### Analyzer package
Example: Microsoft.CodeQuality.Analyzers.Exp, version [2.6.1-beta1-62702-01](https://dotnet.myget.org/feed/roslyn-analyzers/package/nuget/Microsoft.CodeQuality.Analyzers.Exp/2.6.1-beta1-62702-01)
#### Analyzer
ReviewSqlQueriesForSecurityVulnerabilities
#### Repro steps
1. Copy the C#or VB sample from https://docs.microsoft.com/en-us/visualstudio/code-quality/ca2100-review-sql-queries-for-security-vulnerabilities
#### Expected behavior
CA2100 fires.
#### Actual behavior
No CA2100
This seems to be due to the fact that `SqlCommand.CommandText` overrides `DbCommand.CommandText`, and the latter implements `IDbCommand.CommandText`, while the analyzer only checks if the invoked property (i.e. `SqlCommand.CommandText`) directly implements `IDbCommand.CommandText`.
|
1.0
|
CA2100 does not fire for msdn/docs sample - #### Analyzer package
Example: Microsoft.CodeQuality.Analyzers.Exp, version [2.6.1-beta1-62702-01](https://dotnet.myget.org/feed/roslyn-analyzers/package/nuget/Microsoft.CodeQuality.Analyzers.Exp/2.6.1-beta1-62702-01)
#### Analyzer
ReviewSqlQueriesForSecurityVulnerabilities
#### Repro steps
1. Copy the C#or VB sample from https://docs.microsoft.com/en-us/visualstudio/code-quality/ca2100-review-sql-queries-for-security-vulnerabilities
#### Expected behavior
CA2100 fires.
#### Actual behavior
No CA2100
This seems to be due to the fact that `SqlCommand.CommandText` overrides `DbCommand.CommandText`, and the latter implements `IDbCommand.CommandText`, while the analyzer only checks if the invoked property (i.e. `SqlCommand.CommandText`) directly implements `IDbCommand.CommandText`.
|
non_test
|
does not fire for msdn docs sample analyzer package example microsoft codequality analyzers exp version analyzer reviewsqlqueriesforsecurityvulnerabilities repro steps copy the c or vb sample from expected behavior fires actual behavior no this seems to be due to the fact that sqlcommand commandtext overrides dbcommand commandtext and the latter implements idbcommand commandtext while the analyzer only checks if the invoked property i e sqlcommand commandtext directly implements idbcommand commandtext
| 0
|
482,901
| 13,915,543,286
|
IssuesEvent
|
2020-10-21 00:56:03
|
visit-dav/visit
|
https://api.github.com/repos/visit-dav/visit
|
closed
|
len_string is no longer part of exodus format
|
bug impact medium likelihood medium priority reviewed
|
https://github.com/visit-dav/visit/blob/629922a7ac2219d9f312aacbd2486e5f2a1c99f0/src/databases/Exodus/avtExodusFileFormat.C#L141
I think that "maximum_name_length" is the parameter needed here.
I asked Greg Sjaardema and it looks like len_str is only added when QA records are added to a file and was mistakenly added to files by ex_copy until some point between v 7.08 and v 7.18.
With the current logic, if len_string is not defined, VisIt does not see variables in a file.
I can provide example exodus files upon request
|
1.0
|
len_string is no longer part of exodus format - https://github.com/visit-dav/visit/blob/629922a7ac2219d9f312aacbd2486e5f2a1c99f0/src/databases/Exodus/avtExodusFileFormat.C#L141
I think that "maximum_name_length" is the parameter needed here.
I asked Greg Sjaardema and it looks like len_str is only added when QA records are added to a file and was mistakenly added to files by ex_copy until some point between v 7.08 and v 7.18.
With the current logic, if len_string is not defined, VisIt does not see variables in a file.
I can provide example exodus files upon request
|
non_test
|
len string is no longer part of exodus format i think that maximum name length is the parameter needed here i asked greg sjaardema and it looks like len str is only added when qa records are added to a file and was mistakenly added to files by ex copy until some point between v and v with the current logic if len string is not defined visit does not see variables in a file i can provide example exodus files upon request
| 0
|
71,161
| 7,236,035,860
|
IssuesEvent
|
2018-02-13 04:23:50
|
mono/mono
|
https://api.github.com/repos/mono/mono
|
closed
|
dim-diamondshape.exe and dim-sharedgenerics.exe tests fail on FullAOT
|
area-Runtime: AOT test enhancement
|
They were reenabled recently by https://github.com/mono/mono/commit/5c4510a8a7f50ebaa9ad1c19711df995dca15661 and now fail e.g. on https://jenkins.mono-project.com/job/test-mono-pull-request-amd64-fullaot/7814/testReport/
```
MonoTests.runtime.dim-diamondshape.exe
MESSAGE:
Calling IFoo.Foo on Foo - expecting exception. Exception caught: System.NotSupportedException: Interface method 'IFoo:Foo (int)' in class 'FooClass' has multiple candidate implementations.
at Program.Negative () [0x00015] in <ae5be20b4348427f93f5a12d54425b6c>:0
Calling I1.Func on I47Class - expecting exception Exception caught: System.NotSupportedException: Interface method 'I1:Func (int)' in class 'I47Class' has multiple candidate implementations.
at Program.Negative () [0x0005e] in <ae5be20b4348427f93f5a12d54425b6c>:0
Calling GI1<T>.Func on GI23Class<S> - expecting exception Exception caught: System.NotSupportedException: Interface method 'GI1`1:Func<S> (System.Type[]&)' in class 'GI23Class`1<object>' has multiple candidate implementations.
at Program.Negative () [0x000aa] in <ae5be20b4348427f93f5a12d54425b6c>:0
Calling I1.Func on I4Class - expecting I4.Func At I4.Func PASS
Calling I1.Func on I8Class - expecting I8.Func At I8.Func PASS
Calling GI1.Func on GI4Class<object> - expecting GI4.Func<S>
+++++++++++++++++++
STACK TRACE:
Unhandled Exception: System.ExecutionEngineException: Attempting to JIT compile method 'GI4`1<object>:GI1<T>.Func<string> (System.Type[]&)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information.
at Program.Positive () [0x00069] in <ae5be20b4348427f93f5a12d54425b6c>:0
at Program.Main () [0x00007] in <ae5be20b4348427f93f5a12d54425b6c>:0
[ERROR] FATAL UNHANDLED EXCEPTION: System.ExecutionEngineException: Attempting to JIT compile method 'GI4`1<object>:GI1<T>.Func<string> (System.Type[]&)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information.
at Program.Positive () [0x00069] in <ae5be20b4348427f93f5a12d54425b6c>:0
at Program.Main () [0x00007] in <ae5be20b4348427f93f5a12d54425b6c>:0 | 0.12 sec | 8
```
```
MonoTests.runtime.dim-sharedgenerics.exe
MESSAGE:
Calling IFoo<string>.Foo on FooBar<string, object> - expecting default method IFoo<string>.Foo At IFoo.Foo:Arg=ABC, TypeOf(T)=System.String PASS
Calling IBar<string[]>.Foo on FooBar<string, object> - expecting default method IBar<object>.Foo
+++++++++++++++++++
STACK TRACE:
Unhandled Exception: System.ExecutionEngineException: Attempting to JIT compile method 'IBar`1<object>:Bar (object)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information.
at Program.Main () [0x00046] in <e0066f8e0a474276ad30263edbd086bc>:0
[ERROR] FATAL UNHANDLED EXCEPTION: System.ExecutionEngineException: Attempting to JIT compile method 'IBar`1<object>:Bar (object)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information.
at Program.Main () [0x00046] in <e0066f8e0a474276ad30263edbd086bc>:0
```
|
1.0
|
dim-diamondshape.exe and dim-sharedgenerics.exe tests fail on FullAOT - They were reenabled recently by https://github.com/mono/mono/commit/5c4510a8a7f50ebaa9ad1c19711df995dca15661 and now fail e.g. on https://jenkins.mono-project.com/job/test-mono-pull-request-amd64-fullaot/7814/testReport/
```
MonoTests.runtime.dim-diamondshape.exe
MESSAGE:
Calling IFoo.Foo on Foo - expecting exception. Exception caught: System.NotSupportedException: Interface method 'IFoo:Foo (int)' in class 'FooClass' has multiple candidate implementations.
at Program.Negative () [0x00015] in <ae5be20b4348427f93f5a12d54425b6c>:0
Calling I1.Func on I47Class - expecting exception Exception caught: System.NotSupportedException: Interface method 'I1:Func (int)' in class 'I47Class' has multiple candidate implementations.
at Program.Negative () [0x0005e] in <ae5be20b4348427f93f5a12d54425b6c>:0
Calling GI1<T>.Func on GI23Class<S> - expecting exception Exception caught: System.NotSupportedException: Interface method 'GI1`1:Func<S> (System.Type[]&)' in class 'GI23Class`1<object>' has multiple candidate implementations.
at Program.Negative () [0x000aa] in <ae5be20b4348427f93f5a12d54425b6c>:0
Calling I1.Func on I4Class - expecting I4.Func At I4.Func PASS
Calling I1.Func on I8Class - expecting I8.Func At I8.Func PASS
Calling GI1.Func on GI4Class<object> - expecting GI4.Func<S>
+++++++++++++++++++
STACK TRACE:
Unhandled Exception: System.ExecutionEngineException: Attempting to JIT compile method 'GI4`1<object>:GI1<T>.Func<string> (System.Type[]&)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information.
at Program.Positive () [0x00069] in <ae5be20b4348427f93f5a12d54425b6c>:0
at Program.Main () [0x00007] in <ae5be20b4348427f93f5a12d54425b6c>:0
[ERROR] FATAL UNHANDLED EXCEPTION: System.ExecutionEngineException: Attempting to JIT compile method 'GI4`1<object>:GI1<T>.Func<string> (System.Type[]&)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information.
at Program.Positive () [0x00069] in <ae5be20b4348427f93f5a12d54425b6c>:0
at Program.Main () [0x00007] in <ae5be20b4348427f93f5a12d54425b6c>:0 | 0.12 sec | 8
```
```
MonoTests.runtime.dim-sharedgenerics.exe
MESSAGE:
Calling IFoo<string>.Foo on FooBar<string, object> - expecting default method IFoo<string>.Foo At IFoo.Foo:Arg=ABC, TypeOf(T)=System.String PASS
Calling IBar<string[]>.Foo on FooBar<string, object> - expecting default method IBar<object>.Foo
+++++++++++++++++++
STACK TRACE:
Unhandled Exception: System.ExecutionEngineException: Attempting to JIT compile method 'IBar`1<object>:Bar (object)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information.
at Program.Main () [0x00046] in <e0066f8e0a474276ad30263edbd086bc>:0
[ERROR] FATAL UNHANDLED EXCEPTION: System.ExecutionEngineException: Attempting to JIT compile method 'IBar`1<object>:Bar (object)' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information.
at Program.Main () [0x00046] in <e0066f8e0a474276ad30263edbd086bc>:0
```
|
test
|
dim diamondshape exe and dim sharedgenerics exe tests fail on fullaot they were reenabled recently by and now fail e g on monotests runtime dim diamondshape exe message calling ifoo foo on foo expecting exception exception caught system notsupportedexception interface method ifoo foo int in class fooclass has multiple candidate implementations at program negative in calling func on expecting exception exception caught system notsupportedexception interface method func int in class has multiple candidate implementations at program negative in calling func on expecting exception exception caught system notsupportedexception interface method func system type in class has multiple candidate implementations at program negative in calling func on expecting func at func pass calling func on expecting func at func pass calling func on expecting func stack trace unhandled exception system executionengineexception attempting to jit compile method func system type while running in aot only mode see for more information at program positive in at program main in fatal unhandled exception system executionengineexception attempting to jit compile method func system type while running in aot only mode see for more information at program positive in at program main in sec monotests runtime dim sharedgenerics exe message calling ifoo foo on foobar expecting default method ifoo foo at ifoo foo arg abc typeof t system string pass calling ibar foo on foobar expecting default method ibar foo stack trace unhandled exception system executionengineexception attempting to jit compile method ibar bar object while running in aot only mode see for more information at program main in fatal unhandled exception system executionengineexception attempting to jit compile method ibar bar object while running in aot only mode see for more information at program main in
| 1
|
43,406
| 5,537,627,344
|
IssuesEvent
|
2017-03-21 22:38:29
|
fossology/fossology
|
https://api.github.com/repos/fossology/fossology
|
closed
|
run Stress Testing weekly with latest code
|
bug Category: Testing Priority: Normal Status: New Tracker: Bug
|
---
Author Name: **larry shi**
Original Redmine Issue: 6987, http://www.fossology.org/issues/6987
Original Date: 2014/05/07
Original Assignee: Dong Ma
---
manually or automatically.
|
1.0
|
run Stress Testing weekly with latest code - ---
Author Name: **larry shi**
Original Redmine Issue: 6987, http://www.fossology.org/issues/6987
Original Date: 2014/05/07
Original Assignee: Dong Ma
---
manually or automatically.
|
test
|
run stress testing weekly with latest code author name larry shi original redmine issue original date original assignee dong ma manually or automatically
| 1
|
163,270
| 12,710,354,294
|
IssuesEvent
|
2020-06-23 13:46:21
|
pywbem/pywbemtools
|
https://api.github.com/repos/pywbem/pywbemtools
|
closed
|
Test: CIMInstanceName.to_wbem_uri() sorts by default in pywbem 1.0.0
|
area: test resolution: fixed type: bug
|
CIMInstanceName.to_wbem_uri() sorts by default since pywbem 1.0.0. This affects some of the pywbemcli tests. Since pywbemcli needs to run with both pywbem before and after 1.0.0, the tests need to be adjusted so that accomodate both.
|
1.0
|
Test: CIMInstanceName.to_wbem_uri() sorts by default in pywbem 1.0.0 - CIMInstanceName.to_wbem_uri() sorts by default since pywbem 1.0.0. This affects some of the pywbemcli tests. Since pywbemcli needs to run with both pywbem before and after 1.0.0, the tests need to be adjusted so that accomodate both.
|
test
|
test ciminstancename to wbem uri sorts by default in pywbem ciminstancename to wbem uri sorts by default since pywbem this affects some of the pywbemcli tests since pywbemcli needs to run with both pywbem before and after the tests need to be adjusted so that accomodate both
| 1
|
2,237
| 2,524,972,710
|
IssuesEvent
|
2015-01-20 21:20:29
|
graybeal/ont
|
https://api.github.com/repos/graybeal/ont
|
closed
|
Please set up periodic "harvest" for updated version of GCOOS ontology
|
1 star content enhancement imported Milestone-Beta1 ooici Priority-Medium
|
_From [steph_wa...@consolidated.net](https://code.google.com/u/112846428158176258467/) on September 14, 2009 19:27:24_
What capability do you want added or improved? Felimon at GCOOS requests a periodic "harvest" of the updated version of
GCOOS ontology for the MMI repository. Harvest from: http://gcoos.rsmas.miami.edu/dp/srv_gcoos_generateOWL.php Where do you want this capability to be accessible? to be automatic What sort of input/command mechanism do you want? What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies, text contents of vocabulary (voc2rdf), operating system, browser/version (Firefox, Safari, Chrome, IE, etc.), screenshot, etc.)
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=178_
|
1.0
|
Please set up periodic "harvest" for updated version of GCOOS ontology - _From [steph_wa...@consolidated.net](https://code.google.com/u/112846428158176258467/) on September 14, 2009 19:27:24_
What capability do you want added or improved? Felimon at GCOOS requests a periodic "harvest" of the updated version of
GCOOS ontology for the MMI repository. Harvest from: http://gcoos.rsmas.miami.edu/dp/srv_gcoos_generateOWL.php Where do you want this capability to be accessible? to be automatic What sort of input/command mechanism do you want? What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies, text contents of vocabulary (voc2rdf), operating system, browser/version (Firefox, Safari, Chrome, IE, etc.), screenshot, etc.)
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=178_
|
non_test
|
please set up periodic harvest for updated version of gcoos ontology from on september what capability do you want added or improved felimon at gcoos requests a periodic harvest of the updated version of gcoos ontology for the mmi repository harvest from where do you want this capability to be accessible to be automatic what sort of input command mechanism do you want what is the desired output content format location other details of your desired capability what version of the product are you using please provide any additional information below particular ontology ies text contents of vocabulary operating system browser version firefox safari chrome ie etc screenshot etc original issue
| 0
|
73,729
| 7,353,421,953
|
IssuesEvent
|
2018-03-09 00:37:10
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Add a common.log for tests
|
discuss test
|
- **Subsystem**: Test
It's come up a couple of times so far, so it's probably time for a proper discussion on the use of `console.log` in tests (i.e. should tests have any `console.log` output, and if so, when).
|
1.0
|
Add a common.log for tests - - **Subsystem**: Test
It's come up a couple of times so far, so it's probably time for a proper discussion on the use of `console.log` in tests (i.e. should tests have any `console.log` output, and if so, when).
|
test
|
add a common log for tests subsystem test it s come up a couple of times so far so it s probably time for a proper discussion on the use of console log in tests i e should tests have any console log output and if so when
| 1
|
153,174
| 5,886,738,916
|
IssuesEvent
|
2017-05-17 04:20:39
|
FDPA/fdpa
|
https://api.github.com/repos/FDPA/fdpa
|
closed
|
Local Resolution PDFs are returning 404
|
Priority ready for work
|
Craft thinks the PDF lives here http://107.170.57.118/assets/uploads/resolutions/Bethlehem-Township.pdf
But that link is 404. This is true of all PDFs.
|
1.0
|
Local Resolution PDFs are returning 404 - Craft thinks the PDF lives here http://107.170.57.118/assets/uploads/resolutions/Bethlehem-Township.pdf
But that link is 404. This is true of all PDFs.
|
non_test
|
local resolution pdfs are returning craft thinks the pdf lives here but that link is this is true of all pdfs
| 0
|
118,629
| 15,342,905,984
|
IssuesEvent
|
2021-02-27 18:05:03
|
plotn/coolreader
|
https://api.github.com/repos/plotn/coolreader
|
closed
|
Перенести настройки "Опции рендеринга" и "Уровень совместимости DOM"
|
design
|
Пользователь [написал](https://4pda.ru/forum/index.php?s=&showtopic=995536&view=findpost&p=104618639):
> "Опции рендеринга" и "Уровень совместимости DOM" на мой взгляд стоит убрать в "Редкие и экспериментальные". Понять что это такое у простого юзверя нет никаких шансов (даже невзирая на разъяснения), да и умолчальное их значение подойдёт подавляющему большинству. А то маячит, пугает, справки не даёт, дублируется в "Шрифтах" и "CSS"
Дальше ещё немного обсуждения:
> хорошая мысль. Но лучше тогда в "тонкие настройки шрифта", дополнительно переименовав их в "... рендеринга и шрифта"
|
1.0
|
Перенести настройки "Опции рендеринга" и "Уровень совместимости DOM" - Пользователь [написал](https://4pda.ru/forum/index.php?s=&showtopic=995536&view=findpost&p=104618639):
> "Опции рендеринга" и "Уровень совместимости DOM" на мой взгляд стоит убрать в "Редкие и экспериментальные". Понять что это такое у простого юзверя нет никаких шансов (даже невзирая на разъяснения), да и умолчальное их значение подойдёт подавляющему большинству. А то маячит, пугает, справки не даёт, дублируется в "Шрифтах" и "CSS"
Дальше ещё немного обсуждения:
> хорошая мысль. Но лучше тогда в "тонкие настройки шрифта", дополнительно переименовав их в "... рендеринга и шрифта"
|
non_test
|
перенести настройки опции рендеринга и уровень совместимости dom пользователь опции рендеринга и уровень совместимости dom на мой взгляд стоит убрать в редкие и экспериментальные понять что это такое у простого юзверя нет никаких шансов даже невзирая на разъяснения да и умолчальное их значение подойдёт подавляющему большинству а то маячит пугает справки не даёт дублируется в шрифтах и css дальше ещё немного обсуждения хорошая мысль но лучше тогда в тонкие настройки шрифта дополнительно переименовав их в рендеринга и шрифта
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.