Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 4 112 | repo_url stringlengths 33 141 | action stringclasses 3 values | title stringlengths 1 1.02k | labels stringlengths 4 1.54k | body stringlengths 1 262k | index stringclasses 17 values | text_combine stringlengths 95 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
531,709 | 15,503,167,957 | IssuesEvent | 2021-03-11 12:48:41 | FireDynamics/ARTSS | https://api.github.com/repos/FireDynamics/ARTSS | closed | obstacle domain restriction with multiple obstacles does not work | bug effort: high priority: high | - in a case, where obstacles would overlap, the calculation of inner cells would be wrong due to a false number of obstacle cells (du to duplicates in the obstacle list) | 1.0 | obstacle domain restriction with multiple obstacles does not work - - in a case, where obstacles would overlap, the calculation of inner cells would be wrong due to a false number of obstacle cells (du to duplicates in the obstacle list) | non_test | obstacle domain restriction with multiple obstacles does not work in a case where obstacles would overlap the calculation of inner cells would be wrong due to a false number of obstacle cells du to duplicates in the obstacle list | 0 |
129,378 | 10,572,871,167 | IssuesEvent | 2019-10-07 10:34:39 | robotology/gym-ignition | https://api.github.com/repos/robotology/gym-ignition | closed | Develop an example to compare Runtime classes | complexity::medium component::models component::test issue::status::in-progress issue::type::enhancement | At the moment the supported runtimes are the following:
- `GazeboRuntime`
- `PyBulletRuntime` (almost ready in #40)
It would be nice comparing simple environments like the existing CartPole, or an even simpler Pendulum. Without contacts, the two physics engines behaviors should mostly match. | 1.0 | Develop an example to compare Runtime classes - At the moment the supported runtimes are the following:
- `GazeboRuntime`
- `PyBulletRuntime` (almost ready in #40)
It would be nice comparing simple environments like the existing CartPole, or an even simpler Pendulum. Without contacts, the two physics engines behaviors should mostly match. | test | develop an example to compare runtime classes at the moment the supported runtimes are the following gazeboruntime pybulletruntime almost ready in it would be nice comparing simple environments like the existing cartpole or an even simpler pendulum without contacts the two physics engines behaviors should mostly match | 1 |
18,297 | 10,226,923,915 | IssuesEvent | 2019-08-16 19:12:56 | pcrane70/hadoop | https://api.github.com/repos/pcrane70/hadoop | opened | WS-2019-0103 (Medium) detected in handlebars-3.0.7.tgz | security vulnerability | ## WS-2019-0103 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-3.0.7.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-3.0.7.tgz">https://registry.npmjs.org/handlebars/-/handlebars-3.0.7.tgz</a></p>
<p>Path to dependency file: /hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/package.json</p>
<p>Path to vulnerable library: /tmp/git/hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- ember-cli-1.13.14.tgz (Root Library)
- broccoli-0.16.8.tgz
- :x: **handlebars-3.0.7.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/pcrane70/hadoop/commit/9996d65feb6ec3d97f72187616daad5418f51db5">9996d65feb6ec3d97f72187616daad5418f51db5</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Handlebars.js before 4.1.0 has Remote Code Execution (RCE)
<p>Publish Date: 2019-05-30
<p>URL: <a href=https://github.com/wycats/handlebars.js/issues/1267#issue-187151586>WS-2019-0103</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/wycats/handlebars.js/commit/edc6220d51139b32c28e51641fadad59a543ae57">https://github.com/wycats/handlebars.js/commit/edc6220d51139b32c28e51641fadad59a543ae57</a></p>
<p>Release Date: 2019-05-30</p>
<p>Fix Resolution: 4.0.13</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"3.0.7","isTransitiveDependency":true,"dependencyTree":"ember-cli:1.13.14;broccoli:0.16.8;handlebars:3.0.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0.13"}],"vulnerabilityIdentifier":"WS-2019-0103","vulnerabilityDetails":"Handlebars.js before 4.1.0 has Remote Code Execution (RCE)","vulnerabilityUrl":"https://github.com/wycats/handlebars.js/issues/1267#issue-187151586","cvss2Severity":"medium","cvss2Score":"5.5","extraData":{}}</REMEDIATE> --> | True | WS-2019-0103 (Medium) detected in handlebars-3.0.7.tgz - ## WS-2019-0103 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-3.0.7.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-3.0.7.tgz">https://registry.npmjs.org/handlebars/-/handlebars-3.0.7.tgz</a></p>
<p>Path to dependency file: /hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/package.json</p>
<p>Path to vulnerable library: /tmp/git/hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- ember-cli-1.13.14.tgz (Root Library)
- broccoli-0.16.8.tgz
- :x: **handlebars-3.0.7.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/pcrane70/hadoop/commit/9996d65feb6ec3d97f72187616daad5418f51db5">9996d65feb6ec3d97f72187616daad5418f51db5</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Handlebars.js before 4.1.0 has Remote Code Execution (RCE)
<p>Publish Date: 2019-05-30
<p>URL: <a href=https://github.com/wycats/handlebars.js/issues/1267#issue-187151586>WS-2019-0103</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/wycats/handlebars.js/commit/edc6220d51139b32c28e51641fadad59a543ae57">https://github.com/wycats/handlebars.js/commit/edc6220d51139b32c28e51641fadad59a543ae57</a></p>
<p>Release Date: 2019-05-30</p>
<p>Fix Resolution: 4.0.13</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"3.0.7","isTransitiveDependency":true,"dependencyTree":"ember-cli:1.13.14;broccoli:0.16.8;handlebars:3.0.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0.13"}],"vulnerabilityIdentifier":"WS-2019-0103","vulnerabilityDetails":"Handlebars.js before 4.1.0 has Remote Code Execution (RCE)","vulnerabilityUrl":"https://github.com/wycats/handlebars.js/issues/1267#issue-187151586","cvss2Severity":"medium","cvss2Score":"5.5","extraData":{}}</REMEDIATE> --> | non_test | ws medium detected in handlebars tgz ws medium severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file hadoop hadoop yarn project hadoop yarn hadoop yarn ui src main webapp package json path to vulnerable library tmp git hadoop hadoop yarn project hadoop yarn hadoop yarn ui src main webapp node modules handlebars package json dependency hierarchy ember cli tgz root library broccoli tgz x handlebars tgz vulnerable library found in head commit a href vulnerability details handlebars js before has remote code execution rce publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails handlebars js before has remote code execution rce vulnerabilityurl | 0 |
327,620 | 9,978,103,093 | IssuesEvent | 2019-07-09 18:59:37 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | opened | Filler values incorrect for structures | Area/Language Priority/Blocker Type/SpecDeviation | **Description:**
$title.
For example for records with fields with default values:
```ballerina
import ballerina/io;
type Bar record {
string s = "test 1";
};
public function main() {
Bar[] barr = [];
Bar b = { s: "test 2" };
barr[1] = b;
io:println(barr[0].s == "test 1"); // prints false, has empty string
}
```
Object arrays are just filled with null.
`getZeroValue()` methods need to be updated to use the value creator to create these filler values.
| 1.0 | Filler values incorrect for structures - **Description:**
$title.
For example for records with fields with default values:
```ballerina
import ballerina/io;
type Bar record {
string s = "test 1";
};
public function main() {
Bar[] barr = [];
Bar b = { s: "test 2" };
barr[1] = b;
io:println(barr[0].s == "test 1"); // prints false, has empty string
}
```
Object arrays are just filled with null.
`getZeroValue()` methods need to be updated to use the value creator to create these filler values.
| non_test | filler values incorrect for structures description title for example for records with fields with default values ballerina import ballerina io type bar record string s test public function main bar barr bar b s test barr b io println barr s test prints false has empty string object arrays are just filled with null getzerovalue methods need to be updated to use the value creator to create these filler values | 0 |
58,941 | 11,912,123,535 | IssuesEvent | 2020-03-31 09:44:49 | home-assistant/brands | https://api.github.com/repos/home-assistant/brands | closed | Brother Printer is missing brand images | domain-missing has-codeowner has-config-flow |
## The problem
The Brother Printer integration does not have brand images in
this repository.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/brother/icon.png`
- `src/brother/logo.png`
- `src/brother/icon@2x.png`
- `src/brother/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Updating the documentation repository
Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository.
If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: brother.png` line from this file:
<https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/brother.markdown>
**Note**: The documentation PR needs to be opened against the `current` branch.
**Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently.
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @bieniu! Mind taking a look at this issue as it is with an integration (brother) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/brother/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
| 1.0 | Brother Printer is missing brand images -
## The problem
The Brother Printer integration does not have brand images in
this repository.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/brother/icon.png`
- `src/brother/logo.png`
- `src/brother/icon@2x.png`
- `src/brother/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Updating the documentation repository
Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository.
If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: brother.png` line from this file:
<https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/brother.markdown>
**Note**: The documentation PR needs to be opened against the `current` branch.
**Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently.
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @bieniu! Mind taking a look at this issue as it is with an integration (brother) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/brother/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
| non_test | brother printer is missing brand images the problem the brother printer integration does not have brand images in this repository we recently started this brands repository to create a centralized storage of all brand related images these images are used on our website and the home assistant frontend the following images are missing and would ideally be added src brother icon png src brother logo png src brother icon png src brother logo png for image specifications and requirements please see updating the documentation repository our documentation repository already has a logo for this integration however it does not meet the image requirements of this new brands repository if adding images to this repository please open up a pr to the documentation repository as well removing the logo brother png line from this file note the documentation pr needs to be opened against the current branch please leave the actual logo file in the documentation repository it will be cleaned up differently additional information for more information about this repository read the file of this repository it contains information on how this repository works and image specification and requirements codeowner mention hi there bieniu mind taking a look at this issue as it is with an integration brother you are listed as a for thanks resolving this issue is not limited to codeowners if you want to help us out feel free to resolve this issue thanks already | 0 |
28,399 | 2,701,384,521 | IssuesEvent | 2015-04-05 07:39:52 | cs2103jan2015-t13-2c/main | https://api.github.com/repos/cs2103jan2015-t13-2c/main | closed | (GUI) As a user I can find my way around the program instinctively | Priority.high | so I do not spend time searching for commands | 1.0 | (GUI) As a user I can find my way around the program instinctively - so I do not spend time searching for commands | non_test | gui as a user i can find my way around the program instinctively so i do not spend time searching for commands | 0 |
156,969 | 12,341,953,121 | IssuesEvent | 2020-05-14 23:16:27 | Senetas/SureDrop | https://api.github.com/repos/Senetas/SureDrop | closed | This is a check list | Test Case | *This is the test plan for feature blah*
- [ ] Test item 1
- [ ] Test item 2
- [ ] Test item 3
- [ ] Test item 4 | 1.0 | This is a check list - *This is the test plan for feature blah*
- [ ] Test item 1
- [ ] Test item 2
- [ ] Test item 3
- [ ] Test item 4 | test | this is a check list this is the test plan for feature blah test item test item test item test item | 1 |
5,693 | 3,975,675,948 | IssuesEvent | 2016-05-05 07:13:46 | kolliSuman/issues | https://api.github.com/repos/kolliSuman/issues | closed | QA_Skeleton - Assembling, Identification & labeling_Simulation_p1 | Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open | Defect Description :
In the Skeleton-Assembling simulation page of "Skeleton - Assembling, Identification & labeling" experiment, the reset button is missing in the page instead reset button should be displayed on the screen inorder to clear the arranged parts from the positions
Actual Result :
In the Skeleton-Assembling simulation page of "Skeleton - Assembling, Identification & labeling" experiment, the reset button is missing in the page
Environment :
OS: Windows 7, Linux
Browsers: Firefox,Chrome
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/anthropology-iitg/blob/master/test-cases/integration_test-cases/Skeleton%20-%20Assembling%2C%20Identification%20%26%20labeling/Skeleton%20-%20Assembling%2C%20Identification%20%26%20labeling_12_Simulation_p1.org | True | QA_Skeleton - Assembling, Identification & labeling_Simulation_p1 - Defect Description :
In the Skeleton-Assembling simulation page of "Skeleton - Assembling, Identification & labeling" experiment, the reset button is missing in the page instead reset button should be displayed on the screen inorder to clear the arranged parts from the positions
Actual Result :
In the Skeleton-Assembling simulation page of "Skeleton - Assembling, Identification & labeling" experiment, the reset button is missing in the page
Environment :
OS: Windows 7, Linux
Browsers: Firefox,Chrome
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/anthropology-iitg/blob/master/test-cases/integration_test-cases/Skeleton%20-%20Assembling%2C%20Identification%20%26%20labeling/Skeleton%20-%20Assembling%2C%20Identification%20%26%20labeling_12_Simulation_p1.org | non_test | qa skeleton assembling identification labeling simulation defect description in the skeleton assembling simulation page of skeleton assembling identification labeling experiment the reset button is missing in the page instead reset button should be displayed on the screen inorder to clear the arranged parts from the positions actual result in the skeleton assembling simulation page of skeleton assembling identification labeling experiment the reset button is missing in the page environment os windows linux browsers firefox chrome bandwidth hardware configuration processor test step link | 0 |
167,363 | 13,023,311,649 | IssuesEvent | 2020-07-27 09:47:18 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | opened | Display 'off' for 'BlobPublicAccess' on blob container's properties dialog when 'Public Read Access' is off | :gear: blobs 🧪 testing | **Storage Explorer Version:** 1.15.0-dev
**Build**: 20200725.1
**Branch**: master
**Platform/OS:** Windows 10/ Linux Ubuntu 18.04/ macOS Catalina
**Architecture**: ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Expand one storage account -> Blob Containers.
2. Create one blob container then right click it -> Select 'Set Public Access Level...' -> Select 'No public access' -> Click 'Apply'.
3. Observe the value of ‘Public Read Access’ on Properties panel.
4. Open the blob container's properties dialog -> Observe the value of property 'BlobPublicAccess'.
**Expect Experience:**
Display 'off' instead of blank for 'BlobPublicAccess' on blob container's properties dialog.
**Actual Experience:**
Display blank for 'BlobPublicAccess' on blob container's properties dialog.

| 1.0 | Display 'off' for 'BlobPublicAccess' on blob container's properties dialog when 'Public Read Access' is off - **Storage Explorer Version:** 1.15.0-dev
**Build**: 20200725.1
**Branch**: master
**Platform/OS:** Windows 10/ Linux Ubuntu 18.04/ macOS Catalina
**Architecture**: ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Expand one storage account -> Blob Containers.
2. Create one blob container then right click it -> Select 'Set Public Access Level...' -> Select 'No public access' -> Click 'Apply'.
3. Observe the value of ‘Public Read Access’ on Properties panel.
4. Open the blob container's properties dialog -> Observe the value of property 'BlobPublicAccess'.
**Expect Experience:**
Display 'off' instead of blank for 'BlobPublicAccess' on blob container's properties dialog.
**Actual Experience:**
Display blank for 'BlobPublicAccess' on blob container's properties dialog.

| test | display off for blobpublicaccess on blob container s properties dialog when public read access is off storage explorer version dev build branch master platform os windows linux ubuntu macos catalina architecture regression from not a regression steps to reproduce expand one storage account blob containers create one blob container then right click it select set public access level select no public access click apply observe the value of ‘public read access’ on properties panel open the blob container s properties dialog observe the value of property blobpublicaccess expect experience display off instead of blank for blobpublicaccess on blob container s properties dialog actual experience display blank for blobpublicaccess on blob container s properties dialog | 1 |
243,841 | 20,592,767,565 | IssuesEvent | 2022-03-05 03:08:45 | RelativityMC/C2ME-fabric | https://api.github.com/repos/RelativityMC/C2ME-fabric | closed | Alpha 5.96 (1.18) crash | bug need testing | Every so often my server crashes for an unknown reason. I have gotten the same error message:
`Exception in server tick loop` 5 times already in the span of 2 days.
Here are the crash reports:
[crash-2021-12-09_19.02.45-server.txt](https://github.com/RelativityMC/C2ME-fabric/files/7694586/crash-2021-12-09_19.02.45-server.txt)
[crash-2021-12-09_22.41.21-server.txt](https://github.com/RelativityMC/C2ME-fabric/files/7694587/crash-2021-12-09_22.41.21-server.txt)
[crash-2021-12-09_19.02.45-server.txt](https://github.com/RelativityMC/C2ME-fabric/files/7694588/crash-2021-12-09_19.02.45-server.txt)
[crash-2021-12-09_22.41.21-server.txt](https://github.com/RelativityMC/C2ME-fabric/files/7694589/crash-2021-12-09_22.41.21-server.txt)
[crash-2021-12-10_07.20.19-server.txt](https://github.com/RelativityMC/C2ME-fabric/files/7694590/crash-2021-12-10_07.20.19-server.txt)
I also have latest.log:
[latest.log](https://github.com/RelativityMC/C2ME-fabric/files/7694595/latest.log)
Any idea what could be causing this? If it's not c2me's fault then please tell me where to post this in order to get it fixed.
Hardware:
i7-7700@4Ghz
32 GB DDR4 2400mhz ram
sata 4TB ssd with 500MB / s read and writes (where the server is on)
Windows 10
12GB of ram allocated to the server
Software:
Minecraft 1.18 fabric server

| 1.0 | Alpha 5.96 (1.18) crash - Every so often my server crashes for an unknown reason. I have gotten the same error message:
`Exception in server tick loop` 5 times already in the span of 2 days.
Here are the crash reports:
[crash-2021-12-09_19.02.45-server.txt](https://github.com/RelativityMC/C2ME-fabric/files/7694586/crash-2021-12-09_19.02.45-server.txt)
[crash-2021-12-09_22.41.21-server.txt](https://github.com/RelativityMC/C2ME-fabric/files/7694587/crash-2021-12-09_22.41.21-server.txt)
[crash-2021-12-09_19.02.45-server.txt](https://github.com/RelativityMC/C2ME-fabric/files/7694588/crash-2021-12-09_19.02.45-server.txt)
[crash-2021-12-09_22.41.21-server.txt](https://github.com/RelativityMC/C2ME-fabric/files/7694589/crash-2021-12-09_22.41.21-server.txt)
[crash-2021-12-10_07.20.19-server.txt](https://github.com/RelativityMC/C2ME-fabric/files/7694590/crash-2021-12-10_07.20.19-server.txt)
I also have latest.log:
[latest.log](https://github.com/RelativityMC/C2ME-fabric/files/7694595/latest.log)
Any idea what could be causing this? If it's not c2me's fault then please tell me where to post this in order to get it fixed.
Hardware:
i7-7700@4Ghz
32 GB DDR4 2400mhz ram
sata 4TB ssd with 500MB / s read and writes (where the server is on)
Windows 10
12GB of ram allocated to the server
Software:
Minecraft 1.18 fabric server

| test | alpha crash every so often my server crashes for an unknown reason i have gotten the same error message exception in server tick loop times already in the span of days here are the crash reports i also have latest log any idea what could be causing this if it s not s fault then please tell me where to post this in order to get it fixed hardware gb ram sata ssd with s read and writes where the server is on windows of ram allocated to the server software minecraft fabric server | 1 |
24,604 | 17,466,546,622 | IssuesEvent | 2021-08-06 17:44:14 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | closed | Documentation - EKS | operations infrastructure eks | ## Description
Make sure developers understand how we use Kubernetes in our environment.
- How to onboard a new project
- Application Manifest documentation (How to make a simple application manifest, etc)
- CI documentation (what CI needs to output to manifest)
- ArgoCD documentation (logging in, interacting with applications)
- How to use Kubernetes secrets in your application repo (external-secrets, parameter store)
## Background/context/resources
_Any additional context for the reader to know, if applicable_
## Technical notes
_Notes around work that is happening, if applicable_
---
## Tasks
- [ ] _Write documentation_
- [ ] Peer review / Edits
## Definition of Done
- [ ] _Developer documentation for EKS is written and available to developers_
---
### Reminders
- [ ] Please attach your team label and any other appropriate label(s)
- [ ] Please attach the needs grooming tag if needed
- [ ] Please connect to an epic
| 1.0 | Documentation - EKS - ## Description
Make sure developers understand how we use Kubernetes in our environment.
- How to onboard a new project
- Application Manifest documentation (How to make a simple application manifest, etc)
- CI documentation (what CI needs to output to manifest)
- ArgoCD documentation (logging in, interacting with applications)
- How to use Kubernetes secrets in your application repo (external-secrets, parameter store)
## Background/context/resources
_Any additional context for the reader to know, if applicable_
## Technical notes
_Notes around work that is happening, if applicable_
---
## Tasks
- [ ] _Write documentation_
- [ ] Peer review / Edits
## Definition of Done
- [ ] _Developer documentation for EKS is written and available to developers_
---
### Reminders
- [ ] Please attach your team label and any other appropriate label(s)
- [ ] Please attach the needs grooming tag if needed
- [ ] Please connect to an epic
| non_test | documentation eks description make sure developers understand how we use kubernetes in our environment how to onboard a new project application manifest documentation how to make a simple application manifest etc ci documentation what ci needs to output to manifest argocd documentation logging in interacting with applications how to use kubernetes secrets in your application repo external secrets parameter store background context resources any additional context for the reader to know if applicable technical notes notes around work that is happening if applicable tasks write documentation peer review edits definition of done developer documentation for eks is written and available to developers reminders please attach your team label and any other appropriate label s please attach the needs grooming tag if needed please connect to an epic | 0 |
260,797 | 8,214,905,502 | IssuesEvent | 2018-09-05 02:11:01 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.ibm.com - Unable to access the website - Secure Connection Failed error thrown | browser-firefox priority-important severity-critical type-ssl | <!-- @browser: Firefox 63.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://www.ibm.com/fr-fr/marketplace/requirements-management
**Browser / Version**: Firefox 63.0
**Operating System**: Windows 10
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: The secured connection is not working on the official website of IBM
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2018/9/3233d215-7854-4d2b-922c-d139f075fca6.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>buildID: 20180830123124</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.all: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>channel: aurora</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.ibm.com - Unable to access the website - Secure Connection Failed error thrown - <!-- @browser: Firefox 63.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://www.ibm.com/fr-fr/marketplace/requirements-management
**Browser / Version**: Firefox 63.0
**Operating System**: Windows 10
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: The secured connection is not working on the official website of IBM
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2018/9/3233d215-7854-4d2b-922c-d139f075fca6.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>buildID: 20180830123124</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.all: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>channel: aurora</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_test | unable to access the website secure connection failed error thrown url browser version firefox operating system windows tested another browser no problem type something else description the secured connection is not working on the official website of ibm steps to reproduce browser configuration mixed active content blocked false buildid tracking content blocked false gfx webrender blob images true gfx webrender all false mixed passive content blocked false gfx webrender enabled false image mem shared true channel aurora from with ❤️ | 0 |
234,107 | 19,095,681,890 | IssuesEvent | 2021-11-29 16:26:14 | GuabinaCore/WoWKoi | https://api.github.com/repos/GuabinaCore/WoWKoi | closed | [Tyce] - Spell Skyhorn Strafing Run (213467) for Quest Justice Rains from Above | Quest Object Zone:HIghmountain Test Fix | **Links:**
https://www.wowhead.com/quest=40594/justice-rains-from-above
**What is happening:**
1st spell on War Eagle is not working properly, it will cast 5-6 times and then it would say "You have no target"
Also, theres a lot mobs missing
VIDEO:
https://youtu.be/lFYFGhkGWxw
**What should happen:**
https://www.youtube.com/watch?v=4TXX-oCxY4Q | 1.0 | [Tyce] - Spell Skyhorn Strafing Run (213467) for Quest Justice Rains from Above - **Links:**
https://www.wowhead.com/quest=40594/justice-rains-from-above
**What is happening:**
1st spell on War Eagle is not working properly, it will cast 5-6 times and then it would say "You have no target"
Also, theres a lot mobs missing
VIDEO:
https://youtu.be/lFYFGhkGWxw
**What should happen:**
https://www.youtube.com/watch?v=4TXX-oCxY4Q | test | spell skyhorn strafing run for quest justice rains from above links what is happening spell on war eagle is not working properly it will cast times and then it would say you have no target also theres a lot mobs missing video what should happen | 1 |
139,900 | 11,298,713,979 | IssuesEvent | 2020-01-17 09:37:40 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | opened | Pop up a Cloud Explorer window when clicking the 'Refresh All' using the mouse middle key | 🧪 testing | **Storage Explorer Version**: 1.12.0
**Build**: [20200117.2](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3393793)
**Branch**: master
**Platform/OS**: Windows 10/ Linux Ubuntu 18.04/ MacOS High Sierra
**Architecture**: ia32/x64
**Regression From**: Not a regression
**Steps to reproduce:**
1. Open Storage Explorer.
2. Click 'Refresh All' using the mouse middle key.
3. Check the results.
**Expect Experience:**
No window pops up.
**Actual Experience:**
Pop up a Cloud Explorer window.

**More Info:**
This issue also reproduces when clicking 'Collapse All' using the mouse middle key. | 1.0 | Pop up a Cloud Explorer window when clicking the 'Refresh All' using the mouse middle key - **Storage Explorer Version**: 1.12.0
**Build**: [20200117.2](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3393793)
**Branch**: master
**Platform/OS**: Windows 10/ Linux Ubuntu 18.04/ MacOS High Sierra
**Architecture**: ia32/x64
**Regression From**: Not a regression
**Steps to reproduce:**
1. Open Storage Explorer.
2. Click 'Refresh All' using the mouse middle key.
3. Check the results.
**Expect Experience:**
No window pops up.
**Actual Experience:**
Pop up a Cloud Explorer window.

**More Info:**
This issue also reproduces when clicking 'Collapse All' using the mouse middle key. | test | pop up a cloud explorer window when clicking the refresh all using the mouse middle key storage explorer version build branch master platform os windows linux ubuntu macos high sierra architecture regression from not a regression steps to reproduce open storage explorer click refresh all using the mouse middle key check the results expect experience no window pops up actual experience pop up a cloud explorer window more info this issue also reproduces when clicking collapse all using the mouse middle key | 1 |
326,575 | 28,002,429,763 | IssuesEvent | 2023-03-27 13:34:31 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | opened | pull-kubernetes-dependencies is failing | kind/failing-test | ### Which jobs are failing?
pull-kubernetes-dependencies
### Which tests are failing?
verify.Vendor
### Since when has it been failing?
27 March 2023, 15:28 IST (9:58 UTC)
### Testgrid link
https://testgrid.k8s.io/presubmits-kubernetes-blocking#pull-kubernetes-dependencies
### Reason for failure (if possible)
```
Your vendored results are different:
diff -Naupr -x 'AUTHORS*' -x 'CONTRIBUTORS*' vendor/k8s.io/api/go.mod /home/prow/go/src/k8s.io/kubernetes/_tmp/kube-vendor.3Sq7Hh/kubernetes/vendor/k8s.io/api/go.mod
--- vendor/k8s.io/api/go.mod 2023-03-27 10:01:13.097931396 +0000
+++ /home/prow/go/src/k8s.io/kubernetes/_tmp/kube-vendor.3Sq7Hh/kubernetes/vendor/k8s.io/api/go.mod 2023-03-27 10:04:41.397038952 +0000
@@ -21,7 +21,7 @@ require (
github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd // indirect
github.com/modern-go/reflect2 v1.0.2 // indirect
github.com/pmezard/go-difflib v1.0.0 // indirect
- github.com/rogpeppe/go-internal v1.9.0 // indirect
+ github.com/rogpeppe/go-internal v1.10.0 // indirect
github.com/spf13/pflag v1.0.5 // indirect
golang.org/x/net v0.8.0 // indirect
golang.org/x/text v0.8.0 // indirect
```
### Anything else we need to know?
_No response_
### Relevant SIG(s)
/sig architecture testing | 1.0 | pull-kubernetes-dependencies is failing - ### Which jobs are failing?
pull-kubernetes-dependencies
### Which tests are failing?
verify.Vendor
### Since when has it been failing?
27 March 2023, 15:28 IST (9:58 UTC)
### Testgrid link
https://testgrid.k8s.io/presubmits-kubernetes-blocking#pull-kubernetes-dependencies
### Reason for failure (if possible)
```
Your vendored results are different:
diff -Naupr -x 'AUTHORS*' -x 'CONTRIBUTORS*' vendor/k8s.io/api/go.mod /home/prow/go/src/k8s.io/kubernetes/_tmp/kube-vendor.3Sq7Hh/kubernetes/vendor/k8s.io/api/go.mod
--- vendor/k8s.io/api/go.mod 2023-03-27 10:01:13.097931396 +0000
+++ /home/prow/go/src/k8s.io/kubernetes/_tmp/kube-vendor.3Sq7Hh/kubernetes/vendor/k8s.io/api/go.mod 2023-03-27 10:04:41.397038952 +0000
@@ -21,7 +21,7 @@ require (
github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd // indirect
github.com/modern-go/reflect2 v1.0.2 // indirect
github.com/pmezard/go-difflib v1.0.0 // indirect
- github.com/rogpeppe/go-internal v1.9.0 // indirect
+ github.com/rogpeppe/go-internal v1.10.0 // indirect
github.com/spf13/pflag v1.0.5 // indirect
golang.org/x/net v0.8.0 // indirect
golang.org/x/text v0.8.0 // indirect
```
### Anything else we need to know?
_No response_
### Relevant SIG(s)
/sig architecture testing | test | pull kubernetes dependencies is failing which jobs are failing pull kubernetes dependencies which tests are failing verify vendor since when has it been failing march ist utc testgrid link reason for failure if possible your vendored results are different diff naupr x authors x contributors vendor io api go mod home prow go src io kubernetes tmp kube vendor kubernetes vendor io api go mod vendor io api go mod home prow go src io kubernetes tmp kube vendor kubernetes vendor io api go mod require github com modern go concurrent indirect github com modern go indirect github com pmezard go difflib indirect github com rogpeppe go internal indirect github com rogpeppe go internal indirect github com pflag indirect golang org x net indirect golang org x text indirect anything else we need to know no response relevant sig s sig architecture testing | 1 |
160,836 | 20,120,307,424 | IssuesEvent | 2022-02-08 01:06:10 | AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches | https://api.github.com/repos/AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches | opened | CVE-2022-23589 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl | security vulnerability | ## CVE-2022-23589 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /FinalProject/requirements.txt</p>
<p>Path to vulnerable library: /teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Tensorflow is an Open Source Machine Learning Framework. Under certain scenarios, Grappler component of TensorFlow can trigger a null pointer dereference. There are 2 places where this can occur, for the same malicious alteration of a `SavedModel` file (fixing the first one would trigger the same dereference in the second place). First, during constant folding, the `GraphDef` might not have the required nodes for the binary operation. If a node is missing, the correposning `mul_*child` would be null, and the dereference in the subsequent line would be incorrect. We have a similar issue during `IsIdentityConsumingSwitch`. The fix will be included in TensorFlow 2.8.0. We will also cherrypick this commit on TensorFlow 2.7.1, TensorFlow 2.6.3, and TensorFlow 2.5.3, as these are also affected and still in supported range.
<p>Publish Date: 2022-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23589>CVE-2022-23589</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-9px9-73fg-3fqp">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-9px9-73fg-3fqp</a></p>
<p>Release Date: 2022-02-04</p>
<p>Fix Resolution: tensorflow - 2.5.3,2.6.3,2.7.1,2.8.0;tensorflow-cpu - 2.5.3,2.6.3,2.7.1,2.8.0;tensorflow-gpu - 2.5.3,2.6.3,2.7.1,2.8.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-23589 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2022-23589 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /FinalProject/requirements.txt</p>
<p>Path to vulnerable library: /teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Tensorflow is an Open Source Machine Learning Framework. Under certain scenarios, Grappler component of TensorFlow can trigger a null pointer dereference. There are 2 places where this can occur, for the same malicious alteration of a `SavedModel` file (fixing the first one would trigger the same dereference in the second place). First, during constant folding, the `GraphDef` might not have the required nodes for the binary operation. If a node is missing, the correposning `mul_*child` would be null, and the dereference in the subsequent line would be incorrect. We have a similar issue during `IsIdentityConsumingSwitch`. The fix will be included in TensorFlow 2.8.0. We will also cherrypick this commit on TensorFlow 2.7.1, TensorFlow 2.6.3, and TensorFlow 2.5.3, as these are also affected and still in supported range.
<p>Publish Date: 2022-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23589>CVE-2022-23589</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-9px9-73fg-3fqp">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-9px9-73fg-3fqp</a></p>
<p>Release Date: 2022-02-04</p>
<p>Fix Resolution: tensorflow - 2.5.3,2.6.3,2.7.1,2.8.0;tensorflow-cpu - 2.5.3,2.6.3,2.7.1,2.8.0;tensorflow-gpu - 2.5.3,2.6.3,2.7.1,2.8.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file finalproject requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an open source machine learning framework under certain scenarios grappler component of tensorflow can trigger a null pointer dereference there are places where this can occur for the same malicious alteration of a savedmodel file fixing the first one would trigger the same dereference in the second place first during constant folding the graphdef might not have the required nodes for the binary operation if a node is missing the correposning mul child would be null and the dereference in the subsequent line would be incorrect we have a similar issue during isidentityconsumingswitch the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource | 0 |
311,565 | 26,797,833,391 | IssuesEvent | 2023-02-01 13:10:59 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | kv/kvserver: TestReplicateQueueRebalanceMultiStore failed | C-test-failure O-robot branch-master T-kv no-test-failure-activity | kv/kvserver.TestReplicateQueueRebalanceMultiStore [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7609242?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7609242?buildTab=artifacts#/) on master @ [cfb5ae9a96e1770daa4aef1615a46e212b561a84](https://github.com/cockroachdb/cockroach/commits/cfb5ae9a96e1770daa4aef1615a46e212b561a84):
Fatal error:
```
panic: test timed out after 59m55s
```
Stack:
```
goroutine 225834439 [running]:
testing.(*M).startAlarm.func1()
GOROOT/src/testing/testing.go:2036 +0x8e
created by time.goFunc
GOROOT/src/time/sleep.go:176 +0x32
```
<details><summary>Log preceding fatal error</summary>
<p>
```
=== RUN TestReplicateQueueRebalanceMultiStore
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/33e1d369c27b9c01b2b6009c561815a3/logTestReplicateQueueRebalanceMultiStore360195895
test_log_scope.go:79: use -show-logs to present logs inline
=== RUN TestReplicateQueueRebalanceMultiStore/simple
```
</p>
</details>
<p>Parameters: <code>TAGS=bazel,gss,deadlock</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/kv
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestReplicateQueueRebalanceMultiStore.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-21654 | 2.0 | kv/kvserver: TestReplicateQueueRebalanceMultiStore failed - kv/kvserver.TestReplicateQueueRebalanceMultiStore [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7609242?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7609242?buildTab=artifacts#/) on master @ [cfb5ae9a96e1770daa4aef1615a46e212b561a84](https://github.com/cockroachdb/cockroach/commits/cfb5ae9a96e1770daa4aef1615a46e212b561a84):
Fatal error:
```
panic: test timed out after 59m55s
```
Stack:
```
goroutine 225834439 [running]:
testing.(*M).startAlarm.func1()
GOROOT/src/testing/testing.go:2036 +0x8e
created by time.goFunc
GOROOT/src/time/sleep.go:176 +0x32
```
<details><summary>Log preceding fatal error</summary>
<p>
```
=== RUN TestReplicateQueueRebalanceMultiStore
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/33e1d369c27b9c01b2b6009c561815a3/logTestReplicateQueueRebalanceMultiStore360195895
test_log_scope.go:79: use -show-logs to present logs inline
=== RUN TestReplicateQueueRebalanceMultiStore/simple
```
</p>
</details>
<p>Parameters: <code>TAGS=bazel,gss,deadlock</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/kv
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestReplicateQueueRebalanceMultiStore.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-21654 | test | kv kvserver testreplicatequeuerebalancemultistore failed kv kvserver testreplicatequeuerebalancemultistore with on master fatal error panic test timed out after stack goroutine testing m startalarm goroot src testing testing go created by time gofunc goroot src time sleep go log preceding fatal error run testreplicatequeuerebalancemultistore test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline run testreplicatequeuerebalancemultistore simple parameters tags bazel gss deadlock help see also cc cockroachdb kv jira issue crdb | 1 |
292,131 | 25,202,769,949 | IssuesEvent | 2022-11-13 10:05:29 | lowRISC/opentitan | https://api.github.com/repos/lowRISC/opentitan | closed | [rom-e2e] rom_e2e_shutdown_exception_asm | Type:Task SW:ROM Milestone:V2 Component:Rom/E2e/Test | **Testpoint name:** [rom_e2e_shutdown_exception_asm](https://cs.opensource.google/opentitan/opentitan/+/master:sw/device/silicon_creator/rom/data/rom_e2e_testplan.hjson?q=rom_e2e_shutdown_exception_asm)
**Contact person:** @alphan
**Description:** Verify that ROM asm exception handler resets the chip.
- Power on with the `CREATOR_SW_CFG_ROM_EXEC_EN` OTP item set to `0`.
- Execution should halt very early in `_rom_start_boot`.
- Connect the debugger and set a breakpoint at `_asm_exception_handler`.
- Note: We need to use a debugger for this test since `mtvec` points to C handlers
when `rom_main()` starts executing.
- Set `pc` to `0x10000000` (start of main SRAM) and execute one machine instruction,
i.e. `stepi`.
- Verify that execution stops at `_asm_exception_handler` since code execution from SRAM is
not enabled.
- Continue and verify that the asm exception handler resets the chip by confirming that
execution halts at `_rom_start_boot`.
| 1.0 | [rom-e2e] rom_e2e_shutdown_exception_asm - **Testpoint name:** [rom_e2e_shutdown_exception_asm](https://cs.opensource.google/opentitan/opentitan/+/master:sw/device/silicon_creator/rom/data/rom_e2e_testplan.hjson?q=rom_e2e_shutdown_exception_asm)
**Contact person:** @alphan
**Description:** Verify that ROM asm exception handler resets the chip.
- Power on with the `CREATOR_SW_CFG_ROM_EXEC_EN` OTP item set to `0`.
- Execution should halt very early in `_rom_start_boot`.
- Connect the debugger and set a breakpoint at `_asm_exception_handler`.
- Note: We need to use a debugger for this test since `mtvec` points to C handlers
when `rom_main()` starts executing.
- Set `pc` to `0x10000000` (start of main SRAM) and execute one machine instruction,
i.e. `stepi`.
- Verify that execution stops at `_asm_exception_handler` since code execution from SRAM is
not enabled.
- Continue and verify that the asm exception handler resets the chip by confirming that
execution halts at `_rom_start_boot`.
| test | rom shutdown exception asm testpoint name contact person alphan description verify that rom asm exception handler resets the chip power on with the creator sw cfg rom exec en otp item set to execution should halt very early in rom start boot connect the debugger and set a breakpoint at asm exception handler note we need to use a debugger for this test since mtvec points to c handlers when rom main starts executing set pc to start of main sram and execute one machine instruction i e stepi verify that execution stops at asm exception handler since code execution from sram is not enabled continue and verify that the asm exception handler resets the chip by confirming that execution halts at rom start boot | 1 |
162,312 | 12,643,204,260 | IssuesEvent | 2020-06-16 09:25:16 | WoWManiaUK/Redemption | https://api.github.com/repos/WoWManiaUK/Redemption | closed | [Item] Minor Darkmoon Prize - Does not contain items sometimes | Fix - Tester Confirmed | **Links:**
https://www.wow-mania.com/armory/?item=19298
**What is Happening:**
The item does not contain an item sometimes. https://i.imgur.com/DPz3TAV.jpg
**What Should happen:**
Item should always conain an item
| 1.0 | [Item] Minor Darkmoon Prize - Does not contain items sometimes - **Links:**
https://www.wow-mania.com/armory/?item=19298
**What is Happening:**
The item does not contain an item sometimes. https://i.imgur.com/DPz3TAV.jpg
**What Should happen:**
Item should always conain an item
| test | minor darkmoon prize does not contain items sometimes links what is happening the item does not contain an item sometimes what should happen item should always conain an item | 1 |
331,196 | 10,061,754,723 | IssuesEvent | 2019-07-22 22:19:45 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | sts.karnataka.gov.in - see bug description | browser-firefox-mobile engine-gecko priority-normal | <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
**URL**: https://sts.karnataka.gov.in/SATS/sts.htm#
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: Not able to logout .
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | sts.karnataka.gov.in - see bug description - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
**URL**: https://sts.karnataka.gov.in/SATS/sts.htm#
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: Not able to logout .
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_test | sts karnataka gov in see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description not able to logout steps to reproduce browser configuration none from with ❤️ | 0 |
144,346 | 11,613,237,414 | IssuesEvent | 2020-02-26 10:22:19 | eclipse/openj9 | https://api.github.com/repos/eclipse/openj9 | opened | JTReg Test Failure : java/lang/invoke/VarHandles/VarHandleTestMethodHandleAccessXXX.java (2) | test failure | Failure link
------------
Follow on from: https://github.com/eclipse/openj9/issues/3940
https://ci.adoptopenjdk.net/user/adam-thorpe/my-views/view/OpenJ9%20sanity%20openjdk/job/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/151/console
- test category: sanity.openjdk
- OS/architecture: jdk11+ all platforms
Optional info
-------------
- intermittent failure (yes|no): no
- regression or new test: regression
Failure output (captured from console output)
---------------------------------------------
I can see there's still some conversation and open PR's from the previous issue, I just wanted to raise this if this wasn't already a known issue.
```
00:38:27 Type=Segmentation error vmState=0x000501ff
00:38:27 J9Generic_Signal_Number=00000018 Signal_Number=0000000b Error_Value=00000000 Signal_Code=00000000
00:38:27 Handler1=000000000228F370 Handler2=0000000002487740
00:38:27 RDI=6E692E656C646E61 RSI=0000000000000003 RAX=0000000000000003 RBX=0000700000BBB160
00:38:27 RCX=0000000000000003 RDX=00000000770E3090 R8=00000000770D0BB0 R9=000000000000000A
00:38:27 R10=00000000037EE921 R11=00006FFFFD3CB63E R12=0000000000007FFF R13=0000000077097F20
00:38:27 R14=0000000000000003 R15=00000000037EE7DC
00:38:27 RIP=0000000002B2B58A GS=0000 FS=0000 RSP=0000700000BB9F88
00:38:27 RFlags=0000000000010202 CS=002B RBP=6E692E656C646E61 ERR=6921A00800000000
00:38:27 TRAPNO=000000000000000D CPU=A008000000000000 FAULTVADDR=00007FFD6921A008
00:38:27 XMM0 000000a5000000c0 (f: 192.000000, d: 3.501293e-312)
00:38:27 XMM1 00ff00ffffffffff (f: 4294967296.000000, d: 7.064164e-304)
00:38:27 XMM2 0000000000000000 (f: 0.000000, d: 0.000000e+00)
00:38:27 XMM3 00000000770503a0 (f: 1996817280.000000, d: 9.865588e-315)
00:38:27 XMM4 0000000077050ef0 (f: 1996820224.000000, d: 9.865603e-315)
00:38:27 XMM5 0000000077011df0 (f: 1996561920.000000, d: 9.864326e-315)
00:38:27 XMM6 000000007704f430 (f: 1996813312.000000, d: 9.865569e-315)
00:38:27 XMM7 00000000770503a0 (f: 1996817280.000000, d: 9.865588e-315)
00:38:27 XMM8 0000000077050ef0 (f: 1996820224.000000, d: 9.865603e-315)
00:38:27 XMM9 000000d0000000d0 (f: 208.000000, d: 4.413751e-312)
00:38:27 XMM10 0000002000000020 (f: 32.000000, d: 6.790387e-313)
00:38:27 XMM11 8000000080000000 (f: 2147483648.000000, d: -1.060998e-314)
00:38:27 XMM12 0000002000000020 (f: 32.000000, d: 6.790387e-313)
00:38:27 XMM13 0000000000000000 (f: 0.000000, d: 0.000000e+00)
00:38:27 XMM14 0000000000000000 (f: 0.000000, d: 0.000000e+00)
00:38:27 XMM15 0000000000000000 (f: 0.000000, d: 0.000000e+00)
00:38:27 Module=/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/lib/compressedrefs/libj9jit29.dylib
00:38:27 Module_base_address=00000000025F8000 Symbol=_ZN15TR_PrexArgument14knowledgeLevelEPS_
00:38:27 Symbol_address=0000000002B2B580
00:38:27
00:38:27 Method_being_compiled=java/lang/invoke/BruteArgumentMoverHandle.invokeExact_thunkArchetype_X(I)I
00:38:27 Target=2_90_20200225_487 (Mac OS X 10.14.5)
00:38:27 CPU=amd64 (4 logical CPUs) (0x200000000 RAM)
00:38:27 ----------- Stack Backtrace -----------
00:38:27 ---------------------------------------
00:38:27 JVMDUMP039I Processing dump event "gpf", detail "" at 2020/02/25 16:36:28 - please wait.
00:38:27 JVMDUMP032I JVM requested System dump using '/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/core.20200225.163628.92838.0001.dmp' in response to an event
00:38:27 JVMDUMP012E Error in System dump: The core file created by child process with pid = 92960 was not found. Expected to find core file with name "/cores/core.92960"
00:38:27 JVMDUMP032I JVM requested Java dump using '/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/javacore.20200225.163628.92838.0002.txt' in response to an event
00:38:27 JVMDUMP039I Processing dump event "gpf", detail "" at 2020/02/25 16:36:40 - please wait.
00:38:27 JVMDUMP032I JVM requested System dump using '/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/core.20200225.163640.92838.0003.dmp' in response to an event
00:38:27 JVMDUMP012E Error in System dump: The core file created by child process with pid = 92969 was not found. Expected to find core file with name "/cores/core.92969"
00:38:27 JVMDUMP032I JVM requested Java dump using '/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/javacore.20200225.163640.92838.0004.txt' in response to an event
00:38:27 JVMDUMP010I Java dump written to /Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/javacore.20200225.163628.92838.0002.txt
00:38:27 JVMDUMP032I JVM requested Snap dump using '/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/Snap.20200225.163628.92838.0005.trc' in response to an event
00:38:27 JVMDUMP010I Snap dump written to /Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/Snap.20200225.163628.92838.0005.trc
00:38:27 JVMDUMP007I JVM Requesting JIT dump using '/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/jitdump.20200225.163628.92838.0006.dmp'
00:38:27 JVMDUMP010I JIT dump written to /Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/jitdump.20200225.163628.92838.0006.dmp
00:38:27 JVMDUMP013I Processed dump event "gpf", detail "".
``` | 1.0 | JTReg Test Failure : java/lang/invoke/VarHandles/VarHandleTestMethodHandleAccessXXX.java (2) - Failure link
------------
Follow on from: https://github.com/eclipse/openj9/issues/3940
https://ci.adoptopenjdk.net/user/adam-thorpe/my-views/view/OpenJ9%20sanity%20openjdk/job/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/151/console
- test category: sanity.openjdk
- OS/architecture: jdk11+ all platforms
Optional info
-------------
- intermittent failure (yes|no): no
- regression or new test: regression
Failure output (captured from console output)
---------------------------------------------
I can see there's still some conversation and open PR's from the previous issue, I just wanted to raise this if this wasn't already a known issue.
```
00:38:27 Type=Segmentation error vmState=0x000501ff
00:38:27 J9Generic_Signal_Number=00000018 Signal_Number=0000000b Error_Value=00000000 Signal_Code=00000000
00:38:27 Handler1=000000000228F370 Handler2=0000000002487740
00:38:27 RDI=6E692E656C646E61 RSI=0000000000000003 RAX=0000000000000003 RBX=0000700000BBB160
00:38:27 RCX=0000000000000003 RDX=00000000770E3090 R8=00000000770D0BB0 R9=000000000000000A
00:38:27 R10=00000000037EE921 R11=00006FFFFD3CB63E R12=0000000000007FFF R13=0000000077097F20
00:38:27 R14=0000000000000003 R15=00000000037EE7DC
00:38:27 RIP=0000000002B2B58A GS=0000 FS=0000 RSP=0000700000BB9F88
00:38:27 RFlags=0000000000010202 CS=002B RBP=6E692E656C646E61 ERR=6921A00800000000
00:38:27 TRAPNO=000000000000000D CPU=A008000000000000 FAULTVADDR=00007FFD6921A008
00:38:27 XMM0 000000a5000000c0 (f: 192.000000, d: 3.501293e-312)
00:38:27 XMM1 00ff00ffffffffff (f: 4294967296.000000, d: 7.064164e-304)
00:38:27 XMM2 0000000000000000 (f: 0.000000, d: 0.000000e+00)
00:38:27 XMM3 00000000770503a0 (f: 1996817280.000000, d: 9.865588e-315)
00:38:27 XMM4 0000000077050ef0 (f: 1996820224.000000, d: 9.865603e-315)
00:38:27 XMM5 0000000077011df0 (f: 1996561920.000000, d: 9.864326e-315)
00:38:27 XMM6 000000007704f430 (f: 1996813312.000000, d: 9.865569e-315)
00:38:27 XMM7 00000000770503a0 (f: 1996817280.000000, d: 9.865588e-315)
00:38:27 XMM8 0000000077050ef0 (f: 1996820224.000000, d: 9.865603e-315)
00:38:27 XMM9 000000d0000000d0 (f: 208.000000, d: 4.413751e-312)
00:38:27 XMM10 0000002000000020 (f: 32.000000, d: 6.790387e-313)
00:38:27 XMM11 8000000080000000 (f: 2147483648.000000, d: -1.060998e-314)
00:38:27 XMM12 0000002000000020 (f: 32.000000, d: 6.790387e-313)
00:38:27 XMM13 0000000000000000 (f: 0.000000, d: 0.000000e+00)
00:38:27 XMM14 0000000000000000 (f: 0.000000, d: 0.000000e+00)
00:38:27 XMM15 0000000000000000 (f: 0.000000, d: 0.000000e+00)
00:38:27 Module=/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/lib/compressedrefs/libj9jit29.dylib
00:38:27 Module_base_address=00000000025F8000 Symbol=_ZN15TR_PrexArgument14knowledgeLevelEPS_
00:38:27 Symbol_address=0000000002B2B580
00:38:27
00:38:27 Method_being_compiled=java/lang/invoke/BruteArgumentMoverHandle.invokeExact_thunkArchetype_X(I)I
00:38:27 Target=2_90_20200225_487 (Mac OS X 10.14.5)
00:38:27 CPU=amd64 (4 logical CPUs) (0x200000000 RAM)
00:38:27 ----------- Stack Backtrace -----------
00:38:27 ---------------------------------------
00:38:27 JVMDUMP039I Processing dump event "gpf", detail "" at 2020/02/25 16:36:28 - please wait.
00:38:27 JVMDUMP032I JVM requested System dump using '/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/core.20200225.163628.92838.0001.dmp' in response to an event
00:38:27 JVMDUMP012E Error in System dump: The core file created by child process with pid = 92960 was not found. Expected to find core file with name "/cores/core.92960"
00:38:27 JVMDUMP032I JVM requested Java dump using '/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/javacore.20200225.163628.92838.0002.txt' in response to an event
00:38:27 JVMDUMP039I Processing dump event "gpf", detail "" at 2020/02/25 16:36:40 - please wait.
00:38:27 JVMDUMP032I JVM requested System dump using '/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/core.20200225.163640.92838.0003.dmp' in response to an event
00:38:27 JVMDUMP012E Error in System dump: The core file created by child process with pid = 92969 was not found. Expected to find core file with name "/cores/core.92969"
00:38:27 JVMDUMP032I JVM requested Java dump using '/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/javacore.20200225.163640.92838.0004.txt' in response to an event
00:38:27 JVMDUMP010I Java dump written to /Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/javacore.20200225.163628.92838.0002.txt
00:38:27 JVMDUMP032I JVM requested Snap dump using '/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/Snap.20200225.163628.92838.0005.trc' in response to an event
00:38:27 JVMDUMP010I Snap dump written to /Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/Snap.20200225.163628.92838.0005.trc
00:38:27 JVMDUMP007I JVM Requesting JIT dump using '/Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/jitdump.20200225.163628.92838.0006.dmp'
00:38:27 JVMDUMP010I JIT dump written to /Users/jenkins/workspace/Test_openjdk11_j9_sanity.openjdk_x86-64_mac/openjdk-tests/TKG/test_output_15826763584317/jdk_lang_0/work/scratch/1/jitdump.20200225.163628.92838.0006.dmp
00:38:27 JVMDUMP013I Processed dump event "gpf", detail "".
``` | test | jtreg test failure java lang invoke varhandles varhandletestmethodhandleaccessxxx java failure link follow on from test category sanity openjdk os architecture all platforms optional info intermittent failure yes no no regression or new test regression failure output captured from console output i can see there s still some conversation and open pr s from the previous issue i just wanted to raise this if this wasn t already a known issue type segmentation error vmstate signal number signal number error value signal code rdi rsi rax rbx rcx rdx rip gs fs rsp rflags cs rbp err trapno cpu faultvaddr f d f d f d f d f d f d f d f d f d f d f d f d f d f d f d f d module users jenkins workspace test sanity openjdk mac openjdkbinary image contents home lib compressedrefs dylib module base address symbol symbol address method being compiled java lang invoke bruteargumentmoverhandle invokeexact thunkarchetype x i i target mac os x cpu logical cpus ram stack backtrace processing dump event gpf detail at please wait jvm requested system dump using users jenkins workspace test sanity openjdk mac openjdk tests tkg test output jdk lang work scratch core dmp in response to an event error in system dump the core file created by child process with pid was not found expected to find core file with name cores core jvm requested java dump using users jenkins workspace test sanity openjdk mac openjdk tests tkg test output jdk lang work scratch javacore txt in response to an event processing dump event gpf detail at please wait jvm requested system dump using users jenkins workspace test sanity openjdk mac openjdk tests tkg test output jdk lang work scratch core dmp in response to an event error in system dump the core file created by child process with pid was not found expected to find core file with name cores core jvm requested java dump using users jenkins workspace test sanity openjdk mac openjdk tests tkg test output jdk lang work scratch javacore txt in response to an event java dump written to users jenkins workspace test sanity openjdk mac openjdk tests tkg test output jdk lang work scratch javacore txt jvm requested snap dump using users jenkins workspace test sanity openjdk mac openjdk tests tkg test output jdk lang work scratch snap trc in response to an event snap dump written to users jenkins workspace test sanity openjdk mac openjdk tests tkg test output jdk lang work scratch snap trc jvm requesting jit dump using users jenkins workspace test sanity openjdk mac openjdk tests tkg test output jdk lang work scratch jitdump dmp jit dump written to users jenkins workspace test sanity openjdk mac openjdk tests tkg test output jdk lang work scratch jitdump dmp processed dump event gpf detail | 1 |
172,534 | 13,309,418,831 | IssuesEvent | 2020-08-26 03:57:59 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | cdc: panic: semaphore release without acquire | A-cdc C-bug C-test-failure | Test flake?
```
=== RUN TestChangefeedNemeses/sinkless
panic: semaphore release without acquire [recovered]
panic: semaphore release without acquire
goroutine 218102 [running]:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).Recover(0xc00209c240, 0x5c2ff60, 0xc001494740)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:207 +0x11f
panic(0x44e0a40, 0x5b3b180)
/usr/local/go/src/runtime/panic.go:679 +0x1b2
github.com/marusama/semaphore.(*semaphore).Release(0xc003d35e60, 0x1, 0xc001d26000)
/go/src/github.com/cockroachdb/cockroach/vendor/github.com/marusama/semaphore/semaphore.go:170 +0x11d
github.com/cockroachdb/cockroach/pkg/util/limit.(*ConcurrentRequestLimiter).Finish(...)
/go/src/github.com/cockroachdb/cockroach/pkg/util/limit/limiter.go:54
github.com/cockroachdb/cockroach/pkg/kv/kvserver.iteratorWithCloser.Close(0x5c86b40, 0xc001d26000, 0xc009813150)
/go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/replica_rangefeed.go:120 +0x3b
github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed.(*registration).maybeRunCatchupScan.func1(0x5c87500, 0xc002a32dc0, 0xc001091ea0, 0x2453900d, 0xed6d61674, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:290 +0x31
github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed.(*registration).maybeRunCatchupScan(0xc001091ea0, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:411 +0xe22
```
https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_UnitTests/2217816? | 1.0 | cdc: panic: semaphore release without acquire - Test flake?
```
=== RUN TestChangefeedNemeses/sinkless
panic: semaphore release without acquire [recovered]
panic: semaphore release without acquire
goroutine 218102 [running]:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).Recover(0xc00209c240, 0x5c2ff60, 0xc001494740)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:207 +0x11f
panic(0x44e0a40, 0x5b3b180)
/usr/local/go/src/runtime/panic.go:679 +0x1b2
github.com/marusama/semaphore.(*semaphore).Release(0xc003d35e60, 0x1, 0xc001d26000)
/go/src/github.com/cockroachdb/cockroach/vendor/github.com/marusama/semaphore/semaphore.go:170 +0x11d
github.com/cockroachdb/cockroach/pkg/util/limit.(*ConcurrentRequestLimiter).Finish(...)
/go/src/github.com/cockroachdb/cockroach/pkg/util/limit/limiter.go:54
github.com/cockroachdb/cockroach/pkg/kv/kvserver.iteratorWithCloser.Close(0x5c86b40, 0xc001d26000, 0xc009813150)
/go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/replica_rangefeed.go:120 +0x3b
github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed.(*registration).maybeRunCatchupScan.func1(0x5c87500, 0xc002a32dc0, 0xc001091ea0, 0x2453900d, 0xed6d61674, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:290 +0x31
github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed.(*registration).maybeRunCatchupScan(0xc001091ea0, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/rangefeed/registry.go:411 +0xe22
```
https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_UnitTests/2217816? | test | cdc panic semaphore release without acquire test flake run testchangefeednemeses sinkless panic semaphore release without acquire panic semaphore release without acquire goroutine github com cockroachdb cockroach pkg util stop stopper recover go src github com cockroachdb cockroach pkg util stop stopper go panic usr local go src runtime panic go github com marusama semaphore semaphore release go src github com cockroachdb cockroach vendor github com marusama semaphore semaphore go github com cockroachdb cockroach pkg util limit concurrentrequestlimiter finish go src github com cockroachdb cockroach pkg util limit limiter go github com cockroachdb cockroach pkg kv kvserver iteratorwithcloser close go src github com cockroachdb cockroach pkg kv kvserver replica rangefeed go github com cockroachdb cockroach pkg kv kvserver rangefeed registration mayberuncatchupscan go src github com cockroachdb cockroach pkg kv kvserver rangefeed registry go github com cockroachdb cockroach pkg kv kvserver rangefeed registration mayberuncatchupscan go src github com cockroachdb cockroach pkg kv kvserver rangefeed registry go | 1 |
69,984 | 8,482,860,651 | IssuesEvent | 2018-10-25 19:48:14 | public-transport/friendly-public-transport-format | https://api.github.com/repos/public-transport/friendly-public-transport-format | closed | indicating fptf version in js modules | design help wanted | @derhuerst and I were just briefly discussing on how to indicate the "supported" version of FPTF in JavaScript public transport libraries.
We thought of three different options, but feel free to share any other ideas:
1. Require the `fptf` module (at a specific version) in the `devDependencies`.
2. Require the `validate-fptf` module (at a specific version) in the `devDependencies`.
3. Add a custom `fptf` key to the package.json file, which indicates the version (similar to the `browserify` key) | 1.0 | indicating fptf version in js modules - @derhuerst and I were just briefly discussing on how to indicate the "supported" version of FPTF in JavaScript public transport libraries.
We thought of three different options, but feel free to share any other ideas:
1. Require the `fptf` module (at a specific version) in the `devDependencies`.
2. Require the `validate-fptf` module (at a specific version) in the `devDependencies`.
3. Add a custom `fptf` key to the package.json file, which indicates the version (similar to the `browserify` key) | non_test | indicating fptf version in js modules derhuerst and i were just briefly discussing on how to indicate the supported version of fptf in javascript public transport libraries we thought of three different options but feel free to share any other ideas require the fptf module at a specific version in the devdependencies require the validate fptf module at a specific version in the devdependencies add a custom fptf key to the package json file which indicates the version similar to the browserify key | 0 |
38,689 | 6,689,035,747 | IssuesEvent | 2017-10-08 21:17:56 | matplotlib/matplotlib | https://api.github.com/repos/matplotlib/matplotlib | closed | re-write release guide | Documentation | current one is a tad out of date. This is a brain dump of what I have done for the last 3 releases. Here instead of a PR as I didn't plan to type out the list when I created the issue.
- [ ] make sure all of the testing and such is done (pretty easy now that we have travis)
- [ ] run github stats script to update the contributor and issue logs
- [ ] bump version to release target and commit
- [ ] create (and optionally sign) tag
- [ ] create doc branch off of the new tag
- [ ] version bump back to dev version and commit
- [ ] merge everything back to master
- [ ] checkout tag
- [ ] push tag and updated branches to github
- [ ] turn tag into a Release
- [ ] run `git clean -xfd` to make sure stray files don't make it into the dist
- [ ] run `python setup.py sdist` to make clean source tarball
- [ ] kick off building the mac wheels (https://travis-ci.org/MacPython/matplotlib-wheels)
- [ ] create folders on SF (matplotlib-X.Y.Z + windows, mac sub-folders)
- [ ] upload tarball to SF
- [ ] email windows builder
- [ ] email linux/conda/canopy maintainers
- [ ] run `python setup.py sdist upload --sign` to upload to pypi (this assumes .pypirc is set up)
- [ ] collect wheels + exe for windows and mac and upload to pypi (use twine) + SF
- [ ] upload all the binaries to pypi
- [ ] get a DOI from https://zenodo.org/
- [ ] manually update `citing.html` with DOI link in doc/_templates
- [ ] install the tagged version
- [ ] build the docs and copy to new folder (X.Y.Z) matplotlib.github.com repo
- [ ] copy new files to top level folder
- [ ] manually update `downloads.txt` + rerun `update_downloads.py`
- [ ] manually update `versions.html`
- [ ] commit all changes to `matplotlib.github.com` master and push upstream
- [ ] announce to all three mailing lists + g+ + twitter
| 1.0 | re-write release guide - current one is a tad out of date. This is a brain dump of what I have done for the last 3 releases. Here instead of a PR as I didn't plan to type out the list when I created the issue.
- [ ] make sure all of the testing and such is done (pretty easy now that we have travis)
- [ ] run github stats script to update the contributor and issue logs
- [ ] bump version to release target and commit
- [ ] create (and optionally sign) tag
- [ ] create doc branch off of the new tag
- [ ] version bump back to dev version and commit
- [ ] merge everything back to master
- [ ] checkout tag
- [ ] push tag and updated branches to github
- [ ] turn tag into a Release
- [ ] run `git clean -xfd` to make sure stray files don't make it into the dist
- [ ] run `python setup.py sdist` to make clean source tarball
- [ ] kick off building the mac wheels (https://travis-ci.org/MacPython/matplotlib-wheels)
- [ ] create folders on SF (matplotlib-X.Y.Z + windows, mac sub-folders)
- [ ] upload tarball to SF
- [ ] email windows builder
- [ ] email linux/conda/canopy maintainers
- [ ] run `python setup.py sdist upload --sign` to upload to pypi (this assumes .pypirc is set up)
- [ ] collect wheels + exe for windows and mac and upload to pypi (use twine) + SF
- [ ] upload all the binaries to pypi
- [ ] get a DOI from https://zenodo.org/
- [ ] manually update `citing.html` with DOI link in doc/_templates
- [ ] install the tagged version
- [ ] build the docs and copy to new folder (X.Y.Z) matplotlib.github.com repo
- [ ] copy new files to top level folder
- [ ] manually update `downloads.txt` + rerun `update_downloads.py`
- [ ] manually update `versions.html`
- [ ] commit all changes to `matplotlib.github.com` master and push upstream
- [ ] announce to all three mailing lists + g+ + twitter
| non_test | re write release guide current one is a tad out of date this is a brain dump of what i have done for the last releases here instead of a pr as i didn t plan to type out the list when i created the issue make sure all of the testing and such is done pretty easy now that we have travis run github stats script to update the contributor and issue logs bump version to release target and commit create and optionally sign tag create doc branch off of the new tag version bump back to dev version and commit merge everything back to master checkout tag push tag and updated branches to github turn tag into a release run git clean xfd to make sure stray files don t make it into the dist run python setup py sdist to make clean source tarball kick off building the mac wheels create folders on sf matplotlib x y z windows mac sub folders upload tarball to sf email windows builder email linux conda canopy maintainers run python setup py sdist upload sign to upload to pypi this assumes pypirc is set up collect wheels exe for windows and mac and upload to pypi use twine sf upload all the binaries to pypi get a doi from manually update citing html with doi link in doc templates install the tagged version build the docs and copy to new folder x y z matplotlib github com repo copy new files to top level folder manually update downloads txt rerun update downloads py manually update versions html commit all changes to matplotlib github com master and push upstream announce to all three mailing lists g twitter | 0 |
174,045 | 27,561,856,324 | IssuesEvent | 2023-03-07 22:50:43 | chapel-lang/chapel | https://api.github.com/repos/chapel-lang/chapel | closed | Memory.Initialization: Meta-Issue for Stabilization | type: Design area: Libraries / Modules type: Chapel 2.0 | This issue exists to track and coordinate topics for stabilizing the Memory.Initialization module.
See https://github.com/chapel-lang/chapel/issues/16172 for the original module proposal and design discussion.
Module Naming:
- [x] https://github.com/chapel-lang/chapel/issues/20425
Function Signature Issues:
- [x] https://github.com/chapel-lang/chapel/issues/20411
- [x] https://github.com/chapel-lang/chapel/issues/20322
- [x] https://github.com/chapel-lang/chapel/issues/20328
- [x] https://github.com/chapel-lang/chapel/issues/20329
- [x] https://github.com/chapel-lang/chapel/issues/20334
Moving Array Elements:
- [x] https://github.com/chapel-lang/chapel/issues/20338
- [x] https://github.com/chapel-lang/chapel/issues/20339
Semantic Issues:
- [x] https://github.com/chapel-lang/chapel/issues/20410 | 1.0 | Memory.Initialization: Meta-Issue for Stabilization - This issue exists to track and coordinate topics for stabilizing the Memory.Initialization module.
See https://github.com/chapel-lang/chapel/issues/16172 for the original module proposal and design discussion.
Module Naming:
- [x] https://github.com/chapel-lang/chapel/issues/20425
Function Signature Issues:
- [x] https://github.com/chapel-lang/chapel/issues/20411
- [x] https://github.com/chapel-lang/chapel/issues/20322
- [x] https://github.com/chapel-lang/chapel/issues/20328
- [x] https://github.com/chapel-lang/chapel/issues/20329
- [x] https://github.com/chapel-lang/chapel/issues/20334
Moving Array Elements:
- [x] https://github.com/chapel-lang/chapel/issues/20338
- [x] https://github.com/chapel-lang/chapel/issues/20339
Semantic Issues:
- [x] https://github.com/chapel-lang/chapel/issues/20410 | non_test | memory initialization meta issue for stabilization this issue exists to track and coordinate topics for stabilizing the memory initialization module see for the original module proposal and design discussion module naming function signature issues moving array elements semantic issues | 0 |
11,483 | 3,204,705,809 | IssuesEvent | 2015-10-03 11:23:15 | e-government-ua/i | https://api.github.com/repos/e-government-ua/i | closed | На централе (wf-central) реализовать набор сущностей с корнем Place, для организации всех регионов, районов, и населенных пунктов в единую унифицированную структуру, и сервисы работы с ними | active hi priority test | |Phase | Status | Responsible |
|------------ |------------|------------ |
|Dev. build |  | @dgroup |
|Dev. QA |  | @dgroup |
|BA (API) |  | @dgroup |
|QA |  | ? |
- [x] 1) Сущности:
- [x] 1.1) Сущность Place - "место"
Поля:
- nID - ИД-номер (автоитерируемый-уникальный)
- nID_PlaceType - ИД-номер типа
- sID_UA - ИД-строка (уникальный) общеукраинский код классификатора (КОАТИИ)
- sName - строка-название (например: "Днепропетровск", "Днепропетровская")
- sNameOriginal - строка-название оригинальное (например: "ЛУЧИСТІВСЬКА/С.ЛУЧИСТЕ")
- [x] 1.2) Сущность PlaceType - "тип места"
Поля:
- nID - ИД-номер (автоитерируемый-уникальный)
- nOrder - порядковый номер
- sName - строка-название (например: "Область", "Город", "Село")
- bArea - булевый, "Площадь" (true = область/район/регион и т.д.)
- bRoot - булевый, "Корень" (true = административная еденица страны)
- [x] 1.3) Сущность PlaceTree - "дерево мест"
Поля:
- nID_Place - ИД-номер места
- nID_Place_Root - ИД-номер места-корня
- nID_Place_Area - ИД-номер места-площади
- nID_Place_Parent - ИД-номер места-родителя
- [x] 2) Сервисы мест:
- [x] 2.1) getPlacesTree
Параметры:
- nID //опциональный, если задан то с указанного узла дерева
- sID_UA //опциональный, если задан то с указанного узла дерева
- nID_PlaceType //опциональный, если задан то только заданные типы узлов дерева
- bArea //опциональный, если задан - выбрать только те что с соответственным bArea
- bRoot //опциональный, если задан - выбрать только те что с соответственным bRoot
- nDeep //опциональный, если задан - выбрать только это число вложений по дереву
- [x] 2.2) getPlace
Параметры:
- nID //опциональный, если задан - sID_UA
- sID_UA //опциональный, если задан - nID
- bTree //опциональный, если задан - выводить рекурсивно все дерево до корня
- [x] 2.3) setPlace - инсертер и апдейтер одновременно
Параметры:
- nID //опциональный, если задан - sID_UA
- sID_UA //опциональный, если задан - nID
- остальные параметры сущности
- [x] 2.4) removePlace
Параметры:
- nID //опциональный, если задан - sID_UA
- sID_UA //опциональный, если задан - nID
- [x] 3) Сервисы типов мест:
- [x] 3.1) getPlaceTypes
Параметры:
- bArea //опциональный, если задан - выбрать только те что с соответственным bArea
- bRoot //опциональный, если задан - выбрать только те что с соответственным bRoot
- [x] 3.2) getPlaceType
Параметры:
- nID
- [x] 3.3) setPlaceType - инсертер и апдейтер одновременно
Параметры:
- nID
- остальные параметры сущности
- [x] 3.4) removePlaceType
Параметры:
- nID
- [ ] 4) Занести данные из класификатора в csv-шки
http://www.ukrstat.gov.ua/klasf/klasif/koatuu.rar
- [ ] 5) Добавить в доку
Уже в конце всего, после успешных тестов:
- [ ] 6) Добавить в сущность ServiceData поле nID_Place, и по проставлять соответствующие ИД-шники-синонимы текущих мест, и выдавать только ближайших +1 уровень (nDeep=2) | 1.0 | На централе (wf-central) реализовать набор сущностей с корнем Place, для организации всех регионов, районов, и населенных пунктов в единую унифицированную структуру, и сервисы работы с ними - |Phase | Status | Responsible |
|------------ |------------|------------ |
|Dev. build |  | @dgroup |
|Dev. QA |  | @dgroup |
|BA (API) |  | @dgroup |
|QA |  | ? |
- [x] 1) Сущности:
- [x] 1.1) Сущность Place - "место"
Поля:
- nID - ИД-номер (автоитерируемый-уникальный)
- nID_PlaceType - ИД-номер типа
- sID_UA - ИД-строка (уникальный) общеукраинский код классификатора (КОАТИИ)
- sName - строка-название (например: "Днепропетровск", "Днепропетровская")
- sNameOriginal - строка-название оригинальное (например: "ЛУЧИСТІВСЬКА/С.ЛУЧИСТЕ")
- [x] 1.2) Сущность PlaceType - "тип места"
Поля:
- nID - ИД-номер (автоитерируемый-уникальный)
- nOrder - порядковый номер
- sName - строка-название (например: "Область", "Город", "Село")
- bArea - булевый, "Площадь" (true = область/район/регион и т.д.)
- bRoot - булевый, "Корень" (true = административная еденица страны)
- [x] 1.3) Сущность PlaceTree - "дерево мест"
Поля:
- nID_Place - ИД-номер места
- nID_Place_Root - ИД-номер места-корня
- nID_Place_Area - ИД-номер места-площади
- nID_Place_Parent - ИД-номер места-родителя
- [x] 2) Сервисы мест:
- [x] 2.1) getPlacesTree
Параметры:
- nID //опциональный, если задан то с указанного узла дерева
- sID_UA //опциональный, если задан то с указанного узла дерева
- nID_PlaceType //опциональный, если задан то только заданные типы узлов дерева
- bArea //опциональный, если задан - выбрать только те что с соответственным bArea
- bRoot //опциональный, если задан - выбрать только те что с соответственным bRoot
- nDeep //опциональный, если задан - выбрать только это число вложений по дереву
- [x] 2.2) getPlace
Параметры:
- nID //опциональный, если задан - sID_UA
- sID_UA //опциональный, если задан - nID
- bTree //опциональный, если задан - выводить рекурсивно все дерево до корня
- [x] 2.3) setPlace - инсертер и апдейтер одновременно
Параметры:
- nID //опциональный, если задан - sID_UA
- sID_UA //опциональный, если задан - nID
- остальные параметры сущности
- [x] 2.4) removePlace
Параметры:
- nID //опциональный, если задан - sID_UA
- sID_UA //опциональный, если задан - nID
- [x] 3) Сервисы типов мест:
- [x] 3.1) getPlaceTypes
Параметры:
- bArea //опциональный, если задан - выбрать только те что с соответственным bArea
- bRoot //опциональный, если задан - выбрать только те что с соответственным bRoot
- [x] 3.2) getPlaceType
Параметры:
- nID
- [x] 3.3) setPlaceType - инсертер и апдейтер одновременно
Параметры:
- nID
- остальные параметры сущности
- [x] 3.4) removePlaceType
Параметры:
- nID
- [ ] 4) Занести данные из класификатора в csv-шки
http://www.ukrstat.gov.ua/klasf/klasif/koatuu.rar
- [ ] 5) Добавить в доку
Уже в конце всего, после успешных тестов:
- [ ] 6) Добавить в сущность ServiceData поле nID_Place, и по проставлять соответствующие ИД-шники-синонимы текущих мест, и выдавать только ближайших +1 уровень (nDeep=2) | test | на централе wf central реализовать набор сущностей с корнем place для организации всех регионов районов и населенных пунктов в единую унифицированную структуру и сервисы работы с ними phase status responsible dev build complete dgroup dev qa complete dgroup ba api complete dgroup qa complete сущности сущность place место поля nid ид номер автоитерируемый уникальный nid placetype ид номер типа sid ua ид строка уникальный общеукраинский код классификатора коатии sname строка название например днепропетровск днепропетровская snameoriginal строка название оригинальное например лучистівська с лучисте сущность placetype тип места поля nid ид номер автоитерируемый уникальный norder порядковый номер sname строка название например область город село barea булевый площадь true область район регион и т д broot булевый корень true административная еденица страны сущность placetree дерево мест поля nid place ид номер места nid place root ид номер места корня nid place area ид номер места площади nid place parent ид номер места родителя сервисы мест getplacestree параметры nid опциональный если задан то с указанного узла дерева sid ua опциональный если задан то с указанного узла дерева nid placetype опциональный если задан то только заданные типы узлов дерева barea опциональный если задан выбрать только те что с соответственным barea broot опциональный если задан выбрать только те что с соответственным broot ndeep опциональный если задан выбрать только это число вложений по дереву getplace параметры nid опциональный если задан sid ua sid ua опциональный если задан nid btree опциональный если задан выводить рекурсивно все дерево до корня setplace инсертер и апдейтер одновременно параметры nid опциональный если задан sid ua sid ua опциональный если задан nid остальные параметры сущности removeplace параметры nid опциональный если задан sid ua sid ua опциональный если задан nid сервисы типов мест getplacetypes параметры barea опциональный если задан выбрать только те что с соответственным barea broot опциональный если задан выбрать только те что с соответственным broot getplacetype параметры nid setplacetype инсертер и апдейтер одновременно параметры nid остальные параметры сущности removeplacetype параметры nid занести данные из класификатора в csv шки добавить в доку уже в конце всего после успешных тестов добавить в сущность servicedata поле nid place и по проставлять соответствующие ид шники синонимы текущих мест и выдавать только ближайших уровень ndeep | 1 |
241,446 | 7,812,013,889 | IssuesEvent | 2018-06-12 12:07:50 | abntex/abntex2 | https://api.github.com/repos/abntex/abntex2 | closed | Permitir enumeração nos cabeçalhos com Pandoc | Component-UI Milestone-ReleaseAny Priority-Low Type-Enhancement auto-migrated enhancement | ```
Which component is your suggestion addressed to?
[ ] class
[ ] packages
[ ] bibliography style
[ ] documentation
[ ] examples
[X] other: template
What is you suggestion?
Flag especificando se os cabeçalhos devem ser enumerados.
```
Original issue reported on code.google.com by `dudekt...@gmail.com` on 23 Jun 2014 at 10:53
| 1.0 | Permitir enumeração nos cabeçalhos com Pandoc - ```
Which component is your suggestion addressed to?
[ ] class
[ ] packages
[ ] bibliography style
[ ] documentation
[ ] examples
[X] other: template
What is you suggestion?
Flag especificando se os cabeçalhos devem ser enumerados.
```
Original issue reported on code.google.com by `dudekt...@gmail.com` on 23 Jun 2014 at 10:53
| non_test | permitir enumeração nos cabeçalhos com pandoc which component is your suggestion addressed to class packages bibliography style documentation examples other template what is you suggestion flag especificando se os cabeçalhos devem ser enumerados original issue reported on code google com by dudekt gmail com on jun at | 0 |
30,890 | 2,729,037,699 | IssuesEvent | 2015-04-16 04:25:49 | GoogleCloudPlatform/kubernetes | https://api.github.com/repos/GoogleCloudPlatform/kubernetes | opened | Why is api.Pod.Host not called api.Pod.Node? | priority/P2 team/UX | It's minor, but now that I see it, I can't unsee it.
@bgrant0607 | 1.0 | Why is api.Pod.Host not called api.Pod.Node? - It's minor, but now that I see it, I can't unsee it.
@bgrant0607 | non_test | why is api pod host not called api pod node it s minor but now that i see it i can t unsee it | 0 |
336,211 | 30,176,021,793 | IssuesEvent | 2023-07-04 04:45:54 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | closed | Fix jax_lax_operators.test_jax_lax_neg | JAX Frontend Sub Task Failing Test | | | |
|---|---|
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5450322083/jobs/9915454726"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5450322083/jobs/9915454726"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5450322083/jobs/9915454726"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5450322083/jobs/9915454726"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5450322083/jobs/9915454726"><img src=https://img.shields.io/badge/-success-success></a>
| 1.0 | Fix jax_lax_operators.test_jax_lax_neg - | | |
|---|---|
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5450322083/jobs/9915454726"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5450322083/jobs/9915454726"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5450322083/jobs/9915454726"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5450322083/jobs/9915454726"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5450322083/jobs/9915454726"><img src=https://img.shields.io/badge/-success-success></a>
| test | fix jax lax operators test jax lax neg torch a href src numpy a href src tensorflow a href src jax a href src paddle a href src | 1 |
16,541 | 6,221,898,768 | IssuesEvent | 2017-07-10 07:36:02 | GitTools/GitVersion | https://api.github.com/repos/GitTools/GitVersion | closed | Wrong variable names in documentation | Build Server: TFS / Team Services | The description in the [Visual Studio Marketplace](https://marketplace.visualstudio.com/items?itemName=gittools.gitversion) suggests using variables formatted as `$(GitVersion_NuGetVersion)` and `$(GitVersion_SemVer)`. However, when using VSTS, one should format the variables as `$(GitVersion.NuGetVersion)` and `$(GitVersion.SemVer)` (period instead of an underscore). I'm not sure where the text is hosted, so I can't make a PR to fix that. | 1.0 | Wrong variable names in documentation - The description in the [Visual Studio Marketplace](https://marketplace.visualstudio.com/items?itemName=gittools.gitversion) suggests using variables formatted as `$(GitVersion_NuGetVersion)` and `$(GitVersion_SemVer)`. However, when using VSTS, one should format the variables as `$(GitVersion.NuGetVersion)` and `$(GitVersion.SemVer)` (period instead of an underscore). I'm not sure where the text is hosted, so I can't make a PR to fix that. | non_test | wrong variable names in documentation the description in the suggests using variables formatted as gitversion nugetversion and gitversion semver however when using vsts one should format the variables as gitversion nugetversion and gitversion semver period instead of an underscore i m not sure where the text is hosted so i can t make a pr to fix that | 0 |
153,878 | 12,167,649,226 | IssuesEvent | 2020-04-27 11:17:54 | shaunakwyn/Meals4US | https://api.github.com/repos/shaunakwyn/Meals4US | closed | item options & Item Ingredients not showing | 7. Menu management Ready for test SP Sprint One | no data fetch for this two
**Also when you create any module atleast you have to check once that it's working then move that to done.**

| 1.0 | item options & Item Ingredients not showing - no data fetch for this two
**Also when you create any module atleast you have to check once that it's working then move that to done.**

| test | item options item ingredients not showing no data fetch for this two also when you create any module atleast you have to check once that it s working then move that to done | 1 |
179,521 | 13,885,357,463 | IssuesEvent | 2020-10-18 19:40:32 | ibm-openbmc/dev | https://api.github.com/repos/ibm-openbmc/dev | closed | HMC-BMC Dump Managememt-Stage2 | Epic Test on WSP-TAC | => Rework on the dump interface implementation once schema gets standardized.
=> Implement the Log Services for the manager
=> Implement redfish interface for the "User-initiated dump" (system Log Service)
=> Implement redfish interface for the "User-initiated dump" (Manager Log Service)
=> Implement the clear all /delete functionality for the BMC dumps. | 1.0 | HMC-BMC Dump Managememt-Stage2 - => Rework on the dump interface implementation once schema gets standardized.
=> Implement the Log Services for the manager
=> Implement redfish interface for the "User-initiated dump" (system Log Service)
=> Implement redfish interface for the "User-initiated dump" (Manager Log Service)
=> Implement the clear all /delete functionality for the BMC dumps. | test | hmc bmc dump managememt rework on the dump interface implementation once schema gets standardized implement the log services for the manager implement redfish interface for the user initiated dump system log service implement redfish interface for the user initiated dump manager log service implement the clear all delete functionality for the bmc dumps | 1 |
347,039 | 31,075,916,700 | IssuesEvent | 2023-08-12 13:29:16 | timemates/app | https://api.github.com/repos/timemates/app | closed | [Authorization] Write tests for `afterstart`, `new_account_info`, `configure_account` MVI | tests | > **Info**
> Before you go, check [examples](https://github.com/timemates/app/tree/master/feature/authorization/presentation/src/jvmTest/kotlin/io/timemates/app/authorization/ui) that already-made for `authorization` feature.
> Also, make sure that you [read](https://github.com/timemates/app/wiki/App-architecture) and understand our architecture and MVI approach.
>
> Other useful things to learn:
> - https://sharmaricha7724.medium.com/understanding-mockk-kotlin-86db80db07e6
> - https://www.vogella.com/tutorials/JUnit/article.html (we use kotlin.test package, but on JVM it uses jUnit under the hood)
> - https://mockk.io
Tasks to do:
- [x] #64
- [x] #65
- [x] #66 | 1.0 | [Authorization] Write tests for `afterstart`, `new_account_info`, `configure_account` MVI - > **Info**
> Before you go, check [examples](https://github.com/timemates/app/tree/master/feature/authorization/presentation/src/jvmTest/kotlin/io/timemates/app/authorization/ui) that already-made for `authorization` feature.
> Also, make sure that you [read](https://github.com/timemates/app/wiki/App-architecture) and understand our architecture and MVI approach.
>
> Other useful things to learn:
> - https://sharmaricha7724.medium.com/understanding-mockk-kotlin-86db80db07e6
> - https://www.vogella.com/tutorials/JUnit/article.html (we use kotlin.test package, but on JVM it uses jUnit under the hood)
> - https://mockk.io
Tasks to do:
- [x] #64
- [x] #65
- [x] #66 | test | write tests for afterstart new account info configure account mvi info before you go check that already made for authorization feature also make sure that you and understand our architecture and mvi approach other useful things to learn we use kotlin test package but on jvm it uses junit under the hood tasks to do | 1 |
65,806 | 12,687,128,280 | IssuesEvent | 2020-06-20 14:53:26 | Abbassihraf/P-curiosity-LAB | https://api.github.com/repos/Abbassihraf/P-curiosity-LAB | closed | 404 page | Code | - Version Desktop
- [x] 404 section
- Version Tablette
- [x] 404 section
- Version Mobile
- [x] 404 section
| 1.0 | 404 page - - Version Desktop
- [x] 404 section
- Version Tablette
- [x] 404 section
- Version Mobile
- [x] 404 section
| non_test | page version desktop section version tablette section version mobile section | 0 |
152,557 | 12,112,166,701 | IssuesEvent | 2020-04-21 13:25:18 | enonic/app-contentstudio | https://api.github.com/repos/enonic/app-contentstudio | opened | Add ui-test for checking of language in content wizard | Test | language should be automatically set and identical as selected in the project | 1.0 | Add ui-test for checking of language in content wizard - language should be automatically set and identical as selected in the project | test | add ui test for checking of language in content wizard language should be automatically set and identical as selected in the project | 1 |
234,994 | 19,286,862,130 | IssuesEvent | 2021-12-11 04:39:05 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Can't evaluate the debugger expression for record type has nested record field defined in a separate module | Type/Bug Priority/High Area/Debugger Team/DevTools Points/2.5 Debugger / Usability Test | **Description:**
Expression evaluation does not work for the Nested record type.
sample records
```ballerina
// This define in separate module
public type Pet2 record {
int id;
string name;
Type 'type;
};
// Type can be null or record.
public type Type record {
string 'type;
Owner owner;
}?;
public type Owner record {
string name;
string[]? address;
};
```
main.bal file
```ballerina
// Record with field without nested record
types:Pet1 pet1 = {
id: 1,
name: "Tommy"
};
// Record with field with nested record
types:Pet2 pet2 = {
id: 2,
name: "Rova",
'type: { // <- nested record
'type: "Dog",
owner: {
name: "lnash",
address: ()
}
}
};
```
When we are trying to access `pet2` variable by typing in debugger console, it returns this message
`instance of hansani.expressions$0046types.0.$value$Pet2(id=1558)` instead of the record details.
**screencast**
https://user-images.githubusercontent.com/25495666/143671530-daf5af5d-fad9-4bc4-a855-0c13b3d8ec92.mp4
**Steps to reproduce:**
1. Use the ballerina project source code here https://github.com/lnash94/reminders/tree/main/expressions
2. Try to debug and evaluate the expression for access `pet2` details.
**Affected Versions:**
Ballerina SwanLake beta4
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| 1.0 | Can't evaluate the debugger expression for record type has nested record field defined in a separate module - **Description:**
Expression evaluation does not work for the Nested record type.
sample records
```ballerina
// This define in separate module
public type Pet2 record {
int id;
string name;
Type 'type;
};
// Type can be null or record.
public type Type record {
string 'type;
Owner owner;
}?;
public type Owner record {
string name;
string[]? address;
};
```
main.bal file
```ballerina
// Record with field without nested record
types:Pet1 pet1 = {
id: 1,
name: "Tommy"
};
// Record with field with nested record
types:Pet2 pet2 = {
id: 2,
name: "Rova",
'type: { // <- nested record
'type: "Dog",
owner: {
name: "lnash",
address: ()
}
}
};
```
When we are trying to access `pet2` variable by typing in debugger console, it returns this message
`instance of hansani.expressions$0046types.0.$value$Pet2(id=1558)` instead of the record details.
**screencast**
https://user-images.githubusercontent.com/25495666/143671530-daf5af5d-fad9-4bc4-a855-0c13b3d8ec92.mp4
**Steps to reproduce:**
1. Use the ballerina project source code here https://github.com/lnash94/reminders/tree/main/expressions
2. Try to debug and evaluate the expression for access `pet2` details.
**Affected Versions:**
Ballerina SwanLake beta4
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| test | can t evaluate the debugger expression for record type has nested record field defined in a separate module description expression evaluation does not work for the nested record type sample records ballerina this define in separate module public type record int id string name type type type can be null or record public type type record string type owner owner public type owner record string name string address main bal file ballerina record with field without nested record types id name tommy record with field with nested record types id name rova type nested record type dog owner name lnash address when we are trying to access variable by typing in debugger console it returns this message instance of hansani expressions value id instead of the record details screencast steps to reproduce use the ballerina project source code here try to debug and evaluate the expression for access details affected versions ballerina swanlake os db other environment details and versions related issues optional suggested labels optional suggested assignees optional | 1 |
39,453 | 5,234,223,130 | IssuesEvent | 2017-01-30 15:09:11 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | github.com/cockroachdb/cockroach/pkg/kv: TestReverseScanWithSplitAndMerge failed under stress | Robot test-failure | SHA: https://github.com/cockroachdb/cockroach/commits/4c53128707d07d268833ff5ccb5acee9d8720544
Parameters:
```
COCKROACH_PROPOSER_EVALUATED_KV=false
TAGS=deadlock
GOFLAGS=
```
Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=133235&tab=buildLog
```
W170130 11:01:36.257030 3302503 server/status/runtime.go:116 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I170130 11:01:36.258365 3302503 server/config.go:456 1 storage engine initialized
I170130 11:01:36.259820 3302503 server/node.go:444 [n?] store [n0,s0] not bootstrapped
I170130 11:01:36.264400 3303097 storage/replica.go:4339 [n?,s1,r1/1:/M{in-ax},@c429a10c00] gossip not initialized
I170130 11:01:36.267423 3302503 server/node.go:373 [n?] **** cluster b8082567-1d2e-4585-9869-ab9dcf36adad has been created
I170130 11:01:36.267491 3302503 server/node.go:374 [n?] **** add additional nodes by specifying --join=127.0.0.1:37862
I170130 11:01:36.270997 3302503 storage/store.go:1255 [n1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I170130 11:01:36.273493 3302503 server/node.go:457 [n1] initialized store [n1,s1]: {Capacity:536870912 Available:536870912 RangeCount:1 LeaseCount:1}
I170130 11:01:36.273654 3302503 server/node.go:342 [n1] node ID 1 initialized
I170130 11:01:36.273809 3302503 gossip/gossip.go:293 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:37862" > attrs:<> locality:<>
I170130 11:01:36.274335 3302503 storage/stores.go:296 [n1] read 0 node addresses from persistent storage
I170130 11:01:36.274473 3302503 server/node.go:589 [n1] connecting to gossip network to verify cluster ID...
I170130 11:01:36.274630 3302503 server/node.go:613 [n1] node connected via gossip and verified as part of cluster "b8082567-1d2e-4585-9869-ab9dcf36adad"
I170130 11:01:36.282901 3302503 server/node.go:392 [n1] node=1: started with [[]=] engine(s) and attributes []
I170130 11:01:36.283019 3302503 sql/executor.go:332 [n1] creating distSQLPlanner with address {tcp 127.0.0.1:37862}
I170130 11:01:36.291882 3304019 sql/event_log.go:95 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:37862} Attrs: Locality:} ClusterID:b8082567-1d2e-4585-9869-ab9dcf36adad StartedAt:1485774096274724959}
I170130 11:01:36.295840 3302503 server/server.go:629 [n1] starting https server at 127.0.0.1:37008
I170130 11:01:36.295910 3302503 server/server.go:630 [n1] starting grpc/postgres server at 127.0.0.1:37862
I170130 11:01:36.295985 3302503 server/server.go:631 [n1] advertising CockroachDB node at 127.0.0.1:37862
I170130 11:01:36.318724 3302503 sql/event_log.go:95 [n1] Event: "alter_table", target: 12, info: {TableName:eventlog Statement:ALTER TABLE system.eventlog ALTER COLUMN uniqueID SET DEFAULT uuid_v4() User:node MutationID:0 CascadeDroppedViews:[]}
I170130 11:01:36.364863 3302503 server/server.go:686 [n1] done ensuring all necessary migrations have run
I170130 11:01:36.365557 3302503 server/server.go:688 [n1] serving sql connections
I170130 11:01:46.512165 3304670 vendor/google.golang.org/grpc/transport/http2_server.go:320 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:37862->127.0.0.1:34716: use of closed network connection
test_server_shim.go:133: had 1 ranges at startup, expected 6
``` | 1.0 | github.com/cockroachdb/cockroach/pkg/kv: TestReverseScanWithSplitAndMerge failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/4c53128707d07d268833ff5ccb5acee9d8720544
Parameters:
```
COCKROACH_PROPOSER_EVALUATED_KV=false
TAGS=deadlock
GOFLAGS=
```
Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=133235&tab=buildLog
```
W170130 11:01:36.257030 3302503 server/status/runtime.go:116 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I170130 11:01:36.258365 3302503 server/config.go:456 1 storage engine initialized
I170130 11:01:36.259820 3302503 server/node.go:444 [n?] store [n0,s0] not bootstrapped
I170130 11:01:36.264400 3303097 storage/replica.go:4339 [n?,s1,r1/1:/M{in-ax},@c429a10c00] gossip not initialized
I170130 11:01:36.267423 3302503 server/node.go:373 [n?] **** cluster b8082567-1d2e-4585-9869-ab9dcf36adad has been created
I170130 11:01:36.267491 3302503 server/node.go:374 [n?] **** add additional nodes by specifying --join=127.0.0.1:37862
I170130 11:01:36.270997 3302503 storage/store.go:1255 [n1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I170130 11:01:36.273493 3302503 server/node.go:457 [n1] initialized store [n1,s1]: {Capacity:536870912 Available:536870912 RangeCount:1 LeaseCount:1}
I170130 11:01:36.273654 3302503 server/node.go:342 [n1] node ID 1 initialized
I170130 11:01:36.273809 3302503 gossip/gossip.go:293 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:37862" > attrs:<> locality:<>
I170130 11:01:36.274335 3302503 storage/stores.go:296 [n1] read 0 node addresses from persistent storage
I170130 11:01:36.274473 3302503 server/node.go:589 [n1] connecting to gossip network to verify cluster ID...
I170130 11:01:36.274630 3302503 server/node.go:613 [n1] node connected via gossip and verified as part of cluster "b8082567-1d2e-4585-9869-ab9dcf36adad"
I170130 11:01:36.282901 3302503 server/node.go:392 [n1] node=1: started with [[]=] engine(s) and attributes []
I170130 11:01:36.283019 3302503 sql/executor.go:332 [n1] creating distSQLPlanner with address {tcp 127.0.0.1:37862}
I170130 11:01:36.291882 3304019 sql/event_log.go:95 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:37862} Attrs: Locality:} ClusterID:b8082567-1d2e-4585-9869-ab9dcf36adad StartedAt:1485774096274724959}
I170130 11:01:36.295840 3302503 server/server.go:629 [n1] starting https server at 127.0.0.1:37008
I170130 11:01:36.295910 3302503 server/server.go:630 [n1] starting grpc/postgres server at 127.0.0.1:37862
I170130 11:01:36.295985 3302503 server/server.go:631 [n1] advertising CockroachDB node at 127.0.0.1:37862
I170130 11:01:36.318724 3302503 sql/event_log.go:95 [n1] Event: "alter_table", target: 12, info: {TableName:eventlog Statement:ALTER TABLE system.eventlog ALTER COLUMN uniqueID SET DEFAULT uuid_v4() User:node MutationID:0 CascadeDroppedViews:[]}
I170130 11:01:36.364863 3302503 server/server.go:686 [n1] done ensuring all necessary migrations have run
I170130 11:01:36.365557 3302503 server/server.go:688 [n1] serving sql connections
I170130 11:01:46.512165 3304670 vendor/google.golang.org/grpc/transport/http2_server.go:320 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:37862->127.0.0.1:34716: use of closed network connection
test_server_shim.go:133: had 1 ranges at startup, expected 6
``` | test | github com cockroachdb cockroach pkg kv testreversescanwithsplitandmerge failed under stress sha parameters cockroach proposer evaluated kv false tags deadlock goflags stress build found a failed test server status runtime go could not parse build timestamp parsing time as cannot parse as server config go storage engine initialized server node go store not bootstrapped storage replica go gossip not initialized server node go cluster has been created server node go add additional nodes by specifying join storage store go failed initial metrics computation system config not yet available server node go initialized store capacity available rangecount leasecount server node go node id initialized gossip gossip go nodedescriptor set to node id address attrs locality storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes sql executor go creating distsqlplanner with address tcp sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality clusterid startedat server server go starting https server at server server go starting grpc postgres server at server server go advertising cockroachdb node at sql event log go event alter table target info tablename eventlog statement alter table system eventlog alter column uniqueid set default uuid user node mutationid cascadedroppedviews server server go done ensuring all necessary migrations have run server server go serving sql connections vendor google golang org grpc transport server go transport handlestreams failed to read frame read tcp use of closed network connection test server shim go had ranges at startup expected | 1 |
722,971 | 24,879,791,299 | IssuesEvent | 2022-10-27 23:09:00 | phetsims/joist | https://api.github.com/repos/phetsims/joist | closed | Long Title strings off center in home screen | type:bug priority:2-high status:blocks-publication | Found during https://github.com/phetsims/QA/issues/514. pH issue made in https://github.com/phetsims/ph-scale/issues/184.
This issue seems to be independent of platform. When strings are long, the title text on the home screen seems to be further to the left than center. Unsure if this ended up in any published sims.

Assigning @samreid due to recent Joist work. Please reassign as appropriate. | 1.0 | Long Title strings off center in home screen - Found during https://github.com/phetsims/QA/issues/514. pH issue made in https://github.com/phetsims/ph-scale/issues/184.
This issue seems to be independent of platform. When strings are long, the title text on the home screen seems to be further to the left than center. Unsure if this ended up in any published sims.

Assigning @samreid due to recent Joist work. Please reassign as appropriate. | non_test | long title strings off center in home screen found during ph issue made in this issue seems to be independent of platform when strings are long the title text on the home screen seems to be further to the left than center unsure if this ended up in any published sims assigning samreid due to recent joist work please reassign as appropriate | 0 |
77,986 | 9,651,694,741 | IssuesEvent | 2019-05-18 10:20:21 | BookStackApp/BookStack | https://api.github.com/repos/BookStackApp/BookStack | closed | Feature Request: More condensed view in general, improve navigation | :art: Design :paintbrush: View Customization | **Describe the feature you'd like**
During the past few weeks I gave BookStack, DocuWiki, XWiki, Wiki.js and several others a test run.
While I prefer Bookstack in terms of editor features and article layout I don't like the overall use of screen estate.
I would like to have a more condensed view e.g. for the list of pages and chapters within a book (there is much unused (= white) screen estate. And additionally I would like to have a classic navigation column on the left much like in classic wikis.
Navigation from page x in book a to page y in book b should take just one click. Right now, I have to click on "Books" in the navigation bar at the top, then choose the correct book, then klick on the page I want.
**Describe the benefits this feature would bring to BookStack users**
This would greatly improve the navigation experience and would increase the amount of information on one page.
| 1.0 | Feature Request: More condensed view in general, improve navigation - **Describe the feature you'd like**
During the past few weeks I gave BookStack, DocuWiki, XWiki, Wiki.js and several others a test run.
While I prefer Bookstack in terms of editor features and article layout I don't like the overall use of screen estate.
I would like to have a more condensed view e.g. for the list of pages and chapters within a book (there is much unused (= white) screen estate. And additionally I would like to have a classic navigation column on the left much like in classic wikis.
Navigation from page x in book a to page y in book b should take just one click. Right now, I have to click on "Books" in the navigation bar at the top, then choose the correct book, then klick on the page I want.
**Describe the benefits this feature would bring to BookStack users**
This would greatly improve the navigation experience and would increase the amount of information on one page.
| non_test | feature request more condensed view in general improve navigation describe the feature you d like during the past few weeks i gave bookstack docuwiki xwiki wiki js and several others a test run while i prefer bookstack in terms of editor features and article layout i don t like the overall use of screen estate i would like to have a more condensed view e g for the list of pages and chapters within a book there is much unused white screen estate and additionally i would like to have a classic navigation column on the left much like in classic wikis navigation from page x in book a to page y in book b should take just one click right now i have to click on books in the navigation bar at the top then choose the correct book then klick on the page i want describe the benefits this feature would bring to bookstack users this would greatly improve the navigation experience and would increase the amount of information on one page | 0 |
135,611 | 11,013,108,089 | IssuesEvent | 2019-12-04 19:44:46 | ReactionMechanismGenerator/RMG-Py | https://api.github.com/repos/ReactionMechanismGenerator/RMG-Py | closed | Scoop tests not being properly run | Topic: Parallel Type: Testing | It appears that the few unit tests we have for scoop are not being discovered by nosetests. The of the test classes present in `scoop_framework.utilTest`, only the `WorkerWrapperTest seems to be run.
Since we're looking to replace scoop, I'm not sure this is particularly high priority. | 1.0 | Scoop tests not being properly run - It appears that the few unit tests we have for scoop are not being discovered by nosetests. The of the test classes present in `scoop_framework.utilTest`, only the `WorkerWrapperTest seems to be run.
Since we're looking to replace scoop, I'm not sure this is particularly high priority. | test | scoop tests not being properly run it appears that the few unit tests we have for scoop are not being discovered by nosetests the of the test classes present in scoop framework utiltest only the workerwrappertest seems to be run since we re looking to replace scoop i m not sure this is particularly high priority | 1 |
27,919 | 22,602,913,639 | IssuesEvent | 2022-06-29 10:44:03 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Fix AbortTest.Abort_ConnectAndAbort_ThrowsWebSocketExceptionWithmessage on NodeJS | arch-wasm area-Infrastructure-mono in-pr | Running test on NodeJS causes uncaught exception
```
WebSocket uri 'wss://127.0.0.1:60051/WebSocket/EchoWebSocket.ashx?delay10sec'
OperationCanceledException 1
System.Runtime.InteropServices.JavaScript.JSException: Error: WebSocket was closed before the connection was established
at WebSocket.close (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\node_modules\ws\lib\websocket.js:284:14)
at Object.mono_wasm_web_socket_abort (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:3737:16)
at _mono_wasm_web_socket_abort (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:10771:71)
at do_icall (wasm://wasm/02c17daa:wasm-function[309]:0x2dd4d)
at do_icall_wrapper (wasm://wasm/02c17daa:wasm-function[268]:0x2c450)
at interp_exec_method (wasm://wasm/02c17daa:wasm-function[203]:0xc6ad)
at interp_runtime_invoke (wasm://wasm/02c17daa:wasm-function[202]:0xa8ce)
at mono_jit_runtime_invoke (wasm://wasm/02c17daa:wasm-function[14961]:0x288c2a)
at do_runtime_invoke (wasm://wasm/02c17daa:wasm-function[3336]:0xfe90a)
at mono_runtime_invoke_checked (wasm://wasm/02c17daa:wasm-function[3334]:0xfe7e6)
at mono_runtime_try_invoke_span (wasm://wasm/02c17daa:wasm-function[3479]:0x1073bf)
at mono_runtime_invoke_span_checked (wasm://wasm/02c17daa:wasm-function[3480]:0x107722)
at ves_icall_InternalInvoke (wasm://wasm/02c17daa:wasm-function[2016]:0xb53c7)
at ves_icall_InternalInvoke_raw (wasm://wasm/02c17daa:wasm-function[2252]:0xc515b)
at do_icall (wasm://wasm/02c17daa:wasm-function[309]:0x2de47)
at do_icall_wrapper (wasm://wasm/02c17daa:wasm-function[268]:0x2c450)
at interp_exec_method (wasm://wasm/02c17daa:wasm-function[203]:0xc6ad)
at interp_runtime_invoke (wasm://wasm/02c17daa:wasm-function[202]:0xa8ce)
at mono_jit_runtime_invoke (wasm://wasm/02c17daa:wasm-function[14961]:0x288c2a)
at do_runtime_invoke (wasm://wasm/02c17daa:wasm-function[3336]:0xfe90a)
at mono_runtime_try_invoke (wasm://wasm/02c17daa:wasm-function[3350]:0xff5d3)
at mono_runtime_invoke (wasm://wasm/02c17daa:wasm-function[3414]:0x1036d0)
at mono_wasm_invoke_method (wasm://wasm/02c17daa:wasm-function[137]:0x8ad8)
at Module._mono_wasm_invoke_method (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:11470:117)
at Object.wf.<computed> (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:126:20)
at _call_method_with_converted_args (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:2367:48)
at call_method (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:2329:16)
at C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:2397:20
at mono_call_assembly_entry_point (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:2404:67)
at Object.mono_run_main (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:3966:16)
at Object.init (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\test-main.js:212:43)
at Object.onDotnetReady (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\test-main.js:151:17)
at finalize_startup (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:3062:31)
at mono_wasm_on_runtime_initialized (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:2925:9)
at System.Runtime.InteropServices.JavaScript.Runtime.WebSocketAbort(JSObject webSocket) in C:\Development\dotnet\runtime\src\libraries\System.Private.Runtime.InteropServices.JavaScript\src\System\Runtime\InteropServices\JavaScript\Runtime.cs:line 297
at System.Net.WebSockets.BrowserWebSocket.Abort() in C:\Development\dotnet\runtime\src\libraries\System.Net.WebSockets.Client\src\System\Net\WebSockets\BrowserWebSockets\BrowserWebSocket.cs:line 146
at System.Net.WebSockets.WebSocketHandle.Abort() in C:\Development\dotnet\runtime\src\libraries\System.Net.WebSockets.Client\src\System\Net\WebSockets\WebSocketHandle.Browser.cs:line 27
at System.Net.WebSockets.ClientWebSocket.Abort() in C:\Development\dotnet\runtime\src\libraries\System.Net.WebSockets.Client\src\System\Net\WebSockets\ClientWebSocket.cs:line 148
at System.Net.WebSockets.Client.Tests.AbortTest.Abort_ConnectAndAbort_ThrowsWebSocketExceptionWithmessage(Uri server) in C:\Development\dotnet\runtime\src\libraries\System.Net.WebSockets.Client\tests\AbortTest.cs:line 32
```
`WebSocket.js` throws Error when WS is in the state `CONNECTING`. | 1.0 | Fix AbortTest.Abort_ConnectAndAbort_ThrowsWebSocketExceptionWithmessage on NodeJS - Running test on NodeJS causes uncaught exception
```
WebSocket uri 'wss://127.0.0.1:60051/WebSocket/EchoWebSocket.ashx?delay10sec'
OperationCanceledException 1
System.Runtime.InteropServices.JavaScript.JSException: Error: WebSocket was closed before the connection was established
at WebSocket.close (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\node_modules\ws\lib\websocket.js:284:14)
at Object.mono_wasm_web_socket_abort (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:3737:16)
at _mono_wasm_web_socket_abort (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:10771:71)
at do_icall (wasm://wasm/02c17daa:wasm-function[309]:0x2dd4d)
at do_icall_wrapper (wasm://wasm/02c17daa:wasm-function[268]:0x2c450)
at interp_exec_method (wasm://wasm/02c17daa:wasm-function[203]:0xc6ad)
at interp_runtime_invoke (wasm://wasm/02c17daa:wasm-function[202]:0xa8ce)
at mono_jit_runtime_invoke (wasm://wasm/02c17daa:wasm-function[14961]:0x288c2a)
at do_runtime_invoke (wasm://wasm/02c17daa:wasm-function[3336]:0xfe90a)
at mono_runtime_invoke_checked (wasm://wasm/02c17daa:wasm-function[3334]:0xfe7e6)
at mono_runtime_try_invoke_span (wasm://wasm/02c17daa:wasm-function[3479]:0x1073bf)
at mono_runtime_invoke_span_checked (wasm://wasm/02c17daa:wasm-function[3480]:0x107722)
at ves_icall_InternalInvoke (wasm://wasm/02c17daa:wasm-function[2016]:0xb53c7)
at ves_icall_InternalInvoke_raw (wasm://wasm/02c17daa:wasm-function[2252]:0xc515b)
at do_icall (wasm://wasm/02c17daa:wasm-function[309]:0x2de47)
at do_icall_wrapper (wasm://wasm/02c17daa:wasm-function[268]:0x2c450)
at interp_exec_method (wasm://wasm/02c17daa:wasm-function[203]:0xc6ad)
at interp_runtime_invoke (wasm://wasm/02c17daa:wasm-function[202]:0xa8ce)
at mono_jit_runtime_invoke (wasm://wasm/02c17daa:wasm-function[14961]:0x288c2a)
at do_runtime_invoke (wasm://wasm/02c17daa:wasm-function[3336]:0xfe90a)
at mono_runtime_try_invoke (wasm://wasm/02c17daa:wasm-function[3350]:0xff5d3)
at mono_runtime_invoke (wasm://wasm/02c17daa:wasm-function[3414]:0x1036d0)
at mono_wasm_invoke_method (wasm://wasm/02c17daa:wasm-function[137]:0x8ad8)
at Module._mono_wasm_invoke_method (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:11470:117)
at Object.wf.<computed> (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:126:20)
at _call_method_with_converted_args (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:2367:48)
at call_method (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:2329:16)
at C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:2397:20
at mono_call_assembly_entry_point (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:2404:67)
at Object.mono_run_main (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:3966:16)
at Object.init (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\test-main.js:212:43)
at Object.onDotnetReady (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\test-main.js:151:17)
at finalize_startup (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:3062:31)
at mono_wasm_on_runtime_initialized (C:\Development\dotnet\runtime\artifacts\bin\System.Net.WebSockets.Client.Tests\net7.0-Browser-Debug\browser-wasm\AppBundle\dotnet.js:2925:9)
at System.Runtime.InteropServices.JavaScript.Runtime.WebSocketAbort(JSObject webSocket) in C:\Development\dotnet\runtime\src\libraries\System.Private.Runtime.InteropServices.JavaScript\src\System\Runtime\InteropServices\JavaScript\Runtime.cs:line 297
at System.Net.WebSockets.BrowserWebSocket.Abort() in C:\Development\dotnet\runtime\src\libraries\System.Net.WebSockets.Client\src\System\Net\WebSockets\BrowserWebSockets\BrowserWebSocket.cs:line 146
at System.Net.WebSockets.WebSocketHandle.Abort() in C:\Development\dotnet\runtime\src\libraries\System.Net.WebSockets.Client\src\System\Net\WebSockets\WebSocketHandle.Browser.cs:line 27
at System.Net.WebSockets.ClientWebSocket.Abort() in C:\Development\dotnet\runtime\src\libraries\System.Net.WebSockets.Client\src\System\Net\WebSockets\ClientWebSocket.cs:line 148
at System.Net.WebSockets.Client.Tests.AbortTest.Abort_ConnectAndAbort_ThrowsWebSocketExceptionWithmessage(Uri server) in C:\Development\dotnet\runtime\src\libraries\System.Net.WebSockets.Client\tests\AbortTest.cs:line 32
```
`WebSocket.js` throws Error when WS is in the state `CONNECTING`. | non_test | fix aborttest abort connectandabort throwswebsocketexceptionwithmessage on nodejs running test on nodejs causes uncaught exception websocket uri wss websocket echowebsocket ashx operationcanceledexception system runtime interopservices javascript jsexception error websocket was closed before the connection was established at websocket close c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle node modules ws lib websocket js at object mono wasm web socket abort c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle dotnet js at mono wasm web socket abort c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle dotnet js at do icall wasm wasm wasm function at do icall wrapper wasm wasm wasm function at interp exec method wasm wasm wasm function at interp runtime invoke wasm wasm wasm function at mono jit runtime invoke wasm wasm wasm function at do runtime invoke wasm wasm wasm function at mono runtime invoke checked wasm wasm wasm function at mono runtime try invoke span wasm wasm wasm function at mono runtime invoke span checked wasm wasm wasm function at ves icall internalinvoke wasm wasm wasm function at ves icall internalinvoke raw wasm wasm wasm function at do icall wasm wasm wasm function at do icall wrapper wasm wasm wasm function at interp exec method wasm wasm wasm function at interp runtime invoke wasm wasm wasm function at mono jit runtime invoke wasm wasm wasm function at do runtime invoke wasm wasm wasm function at mono runtime try invoke wasm wasm wasm function at mono runtime invoke wasm wasm wasm function at mono wasm invoke method wasm wasm wasm function at module mono wasm invoke method c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle dotnet js at object wf c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle dotnet js at call method with converted args c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle dotnet js at call method c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle dotnet js at c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle dotnet js at mono call assembly entry point c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle dotnet js at object mono run main c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle dotnet js at object init c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle test main js at object ondotnetready c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle test main js at finalize startup c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle dotnet js at mono wasm on runtime initialized c development dotnet runtime artifacts bin system net websockets client tests browser debug browser wasm appbundle dotnet js at system runtime interopservices javascript runtime websocketabort jsobject websocket in c development dotnet runtime src libraries system private runtime interopservices javascript src system runtime interopservices javascript runtime cs line at system net websockets browserwebsocket abort in c development dotnet runtime src libraries system net websockets client src system net websockets browserwebsockets browserwebsocket cs line at system net websockets websockethandle abort in c development dotnet runtime src libraries system net websockets client src system net websockets websockethandle browser cs line at system net websockets clientwebsocket abort in c development dotnet runtime src libraries system net websockets client src system net websockets clientwebsocket cs line at system net websockets client tests aborttest abort connectandabort throwswebsocketexceptionwithmessage uri server in c development dotnet runtime src libraries system net websockets client tests aborttest cs line websocket js throws error when ws is in the state connecting | 0 |
169,159 | 13,127,013,286 | IssuesEvent | 2020-08-06 09:34:47 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | sql/colexec: TestAggregatorMultiFunc failed | C-test-failure O-robot branch-master | [(sql/colexec).TestAggregatorMultiFunc failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2082239&tab=buildLog) on [master@8e38c32b75b52c8d552a25c990cfa16f87e3f04d](https://github.com/cockroachdb/cockroach/commits/8e38c32b75b52c8d552a25c990cfa16f87e3f04d):
```
=== RUN TestAggregatorMultiFunc
E200711 05:47:42.374518 13 util/log/exit_override.go:59 logging error: device or resource busy
ERROR: exit status 2
3 runs completed, 1 failures, over 1m19s
context canceled
```
<details><summary>More</summary><p>
Parameters:
- TAGS=
- GOFLAGS=-parallel=4
```
make stressrace TESTS=TestAggregatorMultiFunc PKG=./pkg/sql/colexec TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestAggregatorMultiFunc.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| 1.0 | sql/colexec: TestAggregatorMultiFunc failed - [(sql/colexec).TestAggregatorMultiFunc failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2082239&tab=buildLog) on [master@8e38c32b75b52c8d552a25c990cfa16f87e3f04d](https://github.com/cockroachdb/cockroach/commits/8e38c32b75b52c8d552a25c990cfa16f87e3f04d):
```
=== RUN TestAggregatorMultiFunc
E200711 05:47:42.374518 13 util/log/exit_override.go:59 logging error: device or resource busy
ERROR: exit status 2
3 runs completed, 1 failures, over 1m19s
context canceled
```
<details><summary>More</summary><p>
Parameters:
- TAGS=
- GOFLAGS=-parallel=4
```
make stressrace TESTS=TestAggregatorMultiFunc PKG=./pkg/sql/colexec TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestAggregatorMultiFunc.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| test | sql colexec testaggregatormultifunc failed on run testaggregatormultifunc util log exit override go logging error device or resource busy error exit status runs completed failures over context canceled more parameters tags goflags parallel make stressrace tests testaggregatormultifunc pkg pkg sql colexec testtimeout stressflags timeout powered by | 1 |
130,404 | 10,608,278,270 | IssuesEvent | 2019-10-11 07:06:19 | rsx-labs/aide-frontend | https://api.github.com/repos/rsx-labs/aide-frontend | closed | Add Task 'Status' Data Change from 'Returned to Triage' to 'Returned to Submitter' | Bug For QA Testing | **Describe the bug**
Add Task 'Status' Data Change from 'Returned to Triage' to 'Returned to Submitter'
**Expected behavior**
Change 'Triage' to 'Submitter' to be more user friendly for other departments in case
**Version (please complete the following information):**
- Version 2.6 | 1.0 | Add Task 'Status' Data Change from 'Returned to Triage' to 'Returned to Submitter' - **Describe the bug**
Add Task 'Status' Data Change from 'Returned to Triage' to 'Returned to Submitter'
**Expected behavior**
Change 'Triage' to 'Submitter' to be more user friendly for other departments in case
**Version (please complete the following information):**
- Version 2.6 | test | add task status data change from returned to triage to returned to submitter describe the bug add task status data change from returned to triage to returned to submitter expected behavior change triage to submitter to be more user friendly for other departments in case version please complete the following information version | 1 |
198,758 | 14,996,287,200 | IssuesEvent | 2021-01-29 15:24:01 | sourcegraph/sourcegraph | https://api.github.com/repos/sourcegraph/sourcegraph | closed | Tracking issue: Dhall test components | dhall dhall-testing team/distribution tracking |
<!-- BEGIN WORK -->
<!-- BEGIN ASSIGNEE: daxmc99 -->
@daxmc99: __2.00d__
Completed: __2.00d__
- [x] (🏁 13 days ago) dhall: tests for postgres component ([~#16804~](https://github.com/sourcegraph/sourcegraph/issues/16804)) __0.50d__
- [x] (🏁 13 days ago) dhall: tests for grafana component ([~#16802~](https://github.com/sourcegraph/sourcegraph/issues/16802)) __0.50d__
- [x] (🏁 13 days ago) dhall: tests for repo-updater component ([~#16803~](https://github.com/sourcegraph/sourcegraph/issues/16803)) __0.50d__
- [x] (🏁 today) dhall: tests for gitserver component ([~#16797~](https://github.com/sourcegraph/sourcegraph/issues/16797)) __0.50d__
<!-- END ASSIGNEE -->
<!-- BEGIN ASSIGNEE: ggilmore -->
@ggilmore: __1.00d__
Completed: __1.00d__
- [x] (🏁 15 days ago) dhall: tests for github-proxy component ([~#16798~](https://github.com/sourcegraph/sourcegraph/issues/16798)) __0.50d__
- [x] (🏁 3 days ago) dhall: tests for symbols component ([~#16799~](https://github.com/sourcegraph/sourcegraph/issues/16799)) __0.50d__
<!-- END ASSIGNEE -->
<!-- BEGIN ASSIGNEE: uwedeportivo -->
@uwedeportivo: __1.00d__
Completed: __1.00d__
- [x] (🏁 3 days ago) dhall: tests for query-runner component ([~#16800~](https://github.com/sourcegraph/sourcegraph/issues/16800)) __0.50d__
- [x] (🏁 3 days ago) dhall: tests for searcher component ([~#16801~](https://github.com/sourcegraph/sourcegraph/issues/16801)) __0.50d__
<!-- END ASSIGNEE -->
<!-- END WORK -->
| 1.0 | Tracking issue: Dhall test components -
<!-- BEGIN WORK -->
<!-- BEGIN ASSIGNEE: daxmc99 -->
@daxmc99: __2.00d__
Completed: __2.00d__
- [x] (🏁 13 days ago) dhall: tests for postgres component ([~#16804~](https://github.com/sourcegraph/sourcegraph/issues/16804)) __0.50d__
- [x] (🏁 13 days ago) dhall: tests for grafana component ([~#16802~](https://github.com/sourcegraph/sourcegraph/issues/16802)) __0.50d__
- [x] (🏁 13 days ago) dhall: tests for repo-updater component ([~#16803~](https://github.com/sourcegraph/sourcegraph/issues/16803)) __0.50d__
- [x] (🏁 today) dhall: tests for gitserver component ([~#16797~](https://github.com/sourcegraph/sourcegraph/issues/16797)) __0.50d__
<!-- END ASSIGNEE -->
<!-- BEGIN ASSIGNEE: ggilmore -->
@ggilmore: __1.00d__
Completed: __1.00d__
- [x] (🏁 15 days ago) dhall: tests for github-proxy component ([~#16798~](https://github.com/sourcegraph/sourcegraph/issues/16798)) __0.50d__
- [x] (🏁 3 days ago) dhall: tests for symbols component ([~#16799~](https://github.com/sourcegraph/sourcegraph/issues/16799)) __0.50d__
<!-- END ASSIGNEE -->
<!-- BEGIN ASSIGNEE: uwedeportivo -->
@uwedeportivo: __1.00d__
Completed: __1.00d__
- [x] (🏁 3 days ago) dhall: tests for query-runner component ([~#16800~](https://github.com/sourcegraph/sourcegraph/issues/16800)) __0.50d__
- [x] (🏁 3 days ago) dhall: tests for searcher component ([~#16801~](https://github.com/sourcegraph/sourcegraph/issues/16801)) __0.50d__
<!-- END ASSIGNEE -->
<!-- END WORK -->
| test | tracking issue dhall test components completed 🏁 days ago dhall tests for postgres component 🏁 days ago dhall tests for grafana component 🏁 days ago dhall tests for repo updater component 🏁 today dhall tests for gitserver component ggilmore completed 🏁 days ago dhall tests for github proxy component 🏁 days ago dhall tests for symbols component uwedeportivo completed 🏁 days ago dhall tests for query runner component 🏁 days ago dhall tests for searcher component | 1 |
20,848 | 3,851,237,522 | IssuesEvent | 2016-04-06 00:38:48 | sass/libsass | https://api.github.com/repos/sass/libsass | closed | Invalidly escaped selectors get silently swallowed | Bug - Confirmed Bug - Selectors Bug - Should Error Dev - Test Written | In Libsass 3.3 when you have invalidly escaped selectors they can get silently swallowed and never output without throwing a compilation exception. Consider the following scss:
```scss
$key: 'bar';
.test11#{'\@#{$key}'} { content: '1.1'; }
.test12#{'\\@#{$key}'} { content: '1.2'; }
//---
$suffix1: '\@#{$key}';
.test21#{$suffix1} { content: '2.1'; }
$suffix2: '\\@#{$key}';
.test22#{$suffix2} { content: '2.2'; }
//---
.test31#{'\@baz'} { content: '3.1'; }
.test32#{'\\@baz'} { content: '3.2'; }
```
In libsass 3.3.1 (I'm using node-sass 3.4.1 as a wrapper) the above compiles to:
```css
.test11\@bar {
content: '1.1'; }
.test22\@bar {
content: '2.2'; }
.test32\@baz {
content: '3.2'; }
```
Note that the output for tests 1.2, 2.1 and 3.1 are all absent. Running this against Ruby Sass 3.4.14 (using sassmeister.com) gives compile errors that those tests have invalid CSS. Commenting out those test cases allows the scss to compile.
This is particularly an issue as libsass 3.2 (again, tested using sassmeister.com) treats cases 1.1, 2.1 and 3.2 as valid scss, but refuses to compile 1.2, 2.2 and 3.1 which means that when upgrading from libsass 3.2 to 3.3 there are no compile-time errors but parts of your CSS file go missing.
I would expect that the three cases 1.2, 2.1 and 3.1 would raise an exception when compiling in libsass 3.3 so that bad selectors issues are obviously highlighted rather than being glossed over.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/27897171-invalidly-escaped-selectors-get-silently-swallowed?utm_campaign=plugin&utm_content=tracker%2F283068&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F283068&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | Invalidly escaped selectors get silently swallowed - In Libsass 3.3 when you have invalidly escaped selectors they can get silently swallowed and never output without throwing a compilation exception. Consider the following scss:
```scss
$key: 'bar';
.test11#{'\@#{$key}'} { content: '1.1'; }
.test12#{'\\@#{$key}'} { content: '1.2'; }
//---
$suffix1: '\@#{$key}';
.test21#{$suffix1} { content: '2.1'; }
$suffix2: '\\@#{$key}';
.test22#{$suffix2} { content: '2.2'; }
//---
.test31#{'\@baz'} { content: '3.1'; }
.test32#{'\\@baz'} { content: '3.2'; }
```
In libsass 3.3.1 (I'm using node-sass 3.4.1 as a wrapper) the above compiles to:
```css
.test11\@bar {
content: '1.1'; }
.test22\@bar {
content: '2.2'; }
.test32\@baz {
content: '3.2'; }
```
Note that the output for tests 1.2, 2.1 and 3.1 are all absent. Running this against Ruby Sass 3.4.14 (using sassmeister.com) gives compile errors that those tests have invalid CSS. Commenting out those test cases allows the scss to compile.
This is particularly an issue as libsass 3.2 (again, tested using sassmeister.com) treats cases 1.1, 2.1 and 3.2 as valid scss, but refuses to compile 1.2, 2.2 and 3.1 which means that when upgrading from libsass 3.2 to 3.3 there are no compile-time errors but parts of your CSS file go missing.
I would expect that the three cases 1.2, 2.1 and 3.1 would raise an exception when compiling in libsass 3.3 so that bad selectors issues are obviously highlighted rather than being glossed over.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/27897171-invalidly-escaped-selectors-get-silently-swallowed?utm_campaign=plugin&utm_content=tracker%2F283068&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F283068&utm_medium=issues&utm_source=github).
</bountysource-plugin> | test | invalidly escaped selectors get silently swallowed in libsass when you have invalidly escaped selectors they can get silently swallowed and never output without throwing a compilation exception consider the following scss scss key bar key content key content key content key content baz content baz content in libsass i m using node sass as a wrapper the above compiles to css bar content bar content baz content note that the output for tests and are all absent running this against ruby sass using sassmeister com gives compile errors that those tests have invalid css commenting out those test cases allows the scss to compile this is particularly an issue as libsass again tested using sassmeister com treats cases and as valid scss but refuses to compile and which means that when upgrading from libsass to there are no compile time errors but parts of your css file go missing i would expect that the three cases and would raise an exception when compiling in libsass so that bad selectors issues are obviously highlighted rather than being glossed over want to back this issue we accept bounties via | 1 |
115,420 | 24,761,941,940 | IssuesEvent | 2022-10-22 02:54:55 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | RyuJIT: fgDebugCheckBBlist (or similar) should check loop annotations | enhancement Hackathon good first issue tenet-reliability area-CodeGen-coreclr JitUntriaged | Follow up to @briansull's https://github.com/dotnet/coreclr/pull/13957#issuecomment-329322217 : when we walk the block list to check invariants, we could/should make sure that:
- each block's `bbNatLoopNum` matches the index of the first loop in the table whose top/bottom blocks' numbers bracket it
- walking the lpParent links in the loop table from that innermost loop, all ancestors also have top/bottom blocks whose numbers bracket the current block's number
This would have given us an assertion failure during compilation of dotnet/runtime#8916 rather than silent bad codegen.
category:correctness
theme:flowgraph
skill-level:intermediate
cost:small | 1.0 | RyuJIT: fgDebugCheckBBlist (or similar) should check loop annotations - Follow up to @briansull's https://github.com/dotnet/coreclr/pull/13957#issuecomment-329322217 : when we walk the block list to check invariants, we could/should make sure that:
- each block's `bbNatLoopNum` matches the index of the first loop in the table whose top/bottom blocks' numbers bracket it
- walking the lpParent links in the loop table from that innermost loop, all ancestors also have top/bottom blocks whose numbers bracket the current block's number
This would have given us an assertion failure during compilation of dotnet/runtime#8916 rather than silent bad codegen.
category:correctness
theme:flowgraph
skill-level:intermediate
cost:small | non_test | ryujit fgdebugcheckbblist or similar should check loop annotations follow up to briansull s when we walk the block list to check invariants we could should make sure that each block s bbnatloopnum matches the index of the first loop in the table whose top bottom blocks numbers bracket it walking the lpparent links in the loop table from that innermost loop all ancestors also have top bottom blocks whose numbers bracket the current block s number this would have given us an assertion failure during compilation of dotnet runtime rather than silent bad codegen category correctness theme flowgraph skill level intermediate cost small | 0 |
534,561 | 15,625,073,261 | IssuesEvent | 2021-03-21 06:12:21 | hyperjumptech/monika | https://api.github.com/repos/hyperjumptech/monika | closed | Probe detection record | High Priority | Right now, for each of probe http call, the status and information is not recorded.
Thus is not possible to trace in historical manner of events and not able to see to se trends in graphs fashion.
We need to save probe detection information which saves minimum.
1. Probe Name / ID
2. Time stamp
3. HttpRequest status (success or fail)
4. Response code
5. Response time | 1.0 | Probe detection record - Right now, for each of probe http call, the status and information is not recorded.
Thus is not possible to trace in historical manner of events and not able to see to se trends in graphs fashion.
We need to save probe detection information which saves minimum.
1. Probe Name / ID
2. Time stamp
3. HttpRequest status (success or fail)
4. Response code
5. Response time | non_test | probe detection record right now for each of probe http call the status and information is not recorded thus is not possible to trace in historical manner of events and not able to see to se trends in graphs fashion we need to save probe detection information which saves minimum probe name id time stamp httprequest status success or fail response code response time | 0 |
123,671 | 10,278,964,880 | IssuesEvent | 2019-08-25 18:47:33 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | tests/ztest/mock: test_multi_value_test: Unused mocked return value | area: Tests bug priority: low | **To Reproduce**
Steps to reproduce the behavior:
1. mkdir build; cd build
2. cmake -DBOARD=altera_max10
3. make
4. make run
**Screenshots or console output**
```
***** Booting Zephyr OS zephyr-v1.14.0-18-g88a1fd7dd96b (delayed boot 3000ms) *****
Running test suite mock_framework_tests
===================================================================
starting test - test_return_value_tests
PASS - test_return_value_tests
===================================================================
starting test - test_multi_value_tests
Failed to find parameter a for expect_one_parameter
Test test_multi_value_tests failed: Unused mock return values
FAIL - test_multi_value_tests
===================================================================
Test suite mock_framework_tests failed.
===================================================================
```
**Environment (please complete the following information):**
- OS: fedora
- Toolchain: zephyr
- Commit SHA or Version used: 88a1fd7dd96b7c21f1bd0109c7c9df0249fc9447
**Additional context**
Add any other context about the problem here.
| 1.0 | tests/ztest/mock: test_multi_value_test: Unused mocked return value - **To Reproduce**
Steps to reproduce the behavior:
1. mkdir build; cd build
2. cmake -DBOARD=altera_max10
3. make
4. make run
**Screenshots or console output**
```
***** Booting Zephyr OS zephyr-v1.14.0-18-g88a1fd7dd96b (delayed boot 3000ms) *****
Running test suite mock_framework_tests
===================================================================
starting test - test_return_value_tests
PASS - test_return_value_tests
===================================================================
starting test - test_multi_value_tests
Failed to find parameter a for expect_one_parameter
Test test_multi_value_tests failed: Unused mock return values
FAIL - test_multi_value_tests
===================================================================
Test suite mock_framework_tests failed.
===================================================================
```
**Environment (please complete the following information):**
- OS: fedora
- Toolchain: zephyr
- Commit SHA or Version used: 88a1fd7dd96b7c21f1bd0109c7c9df0249fc9447
**Additional context**
Add any other context about the problem here.
| test | tests ztest mock test multi value test unused mocked return value to reproduce steps to reproduce the behavior mkdir build cd build cmake dboard altera make make run screenshots or console output booting zephyr os zephyr delayed boot running test suite mock framework tests starting test test return value tests pass test return value tests starting test test multi value tests failed to find parameter a for expect one parameter test test multi value tests failed unused mock return values fail test multi value tests test suite mock framework tests failed environment please complete the following information os fedora toolchain zephyr commit sha or version used additional context add any other context about the problem here | 1 |
296,901 | 25,583,388,349 | IssuesEvent | 2022-12-01 07:13:01 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: remove-invalid-database-privileges failed | C-test-failure O-robot O-roachtest release-blocker T-sql-experience branch-release-22.1 | roachtest.remove-invalid-database-privileges [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=7782673&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=7782673&tab=artifacts#/remove-invalid-database-privileges) on release-22.1 @ [f9730bda77ce8e3ecfa42302be58d81dcd04cd21](https://github.com/cockroachdb/cockroach/commits/f9730bda77ce8e3ecfa42302be58d81dcd04cd21):
```
The test failed on branch=release-22.1, cloud=gce:
test artifacts and logs in: /artifacts/remove-invalid-database-privileges/run_1
cluster.go:1849,versionupgrade.go:393,versionupgrade.go:354,versionupgrade.go:178,remove_invalid_database_privileges.go:117,remove_invalid_database_privileges.go:30,test_runner.go:883: one or more parallel execution failure
(1) attached stack trace
-- stack trace:
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2202
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2083
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:175
| github.com/cockroachdb/cockroach/pkg/roachprod.Start
| github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:660
| main.(*clusterImpl).StartE
| main/pkg/cmd/roachtest/cluster.go:1803
| main.(*clusterImpl).Start
| main/pkg/cmd/roachtest/cluster.go:1848
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.upgradeNodes
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/versionupgrade.go:393
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.binaryUpgradeStep.func1
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/versionupgrade.go:354
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.(*versionUpgradeTest).run
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/versionupgrade.go:178
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runRemoveInvalidDatabasePrivileges
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/remove_invalid_database_privileges.go:117
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerRemoveInvalidDatabasePrivileges.func1
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/remove_invalid_database_privileges.go:30
| main.(*testRunner).runTest.func2
| main/pkg/cmd/roachtest/test_runner.go:883
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1581
Wraps: (2) one or more parallel execution failure
Error types: (1) *withstack.withStack (2) *errutil.leafError
```
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*remove-invalid-database-privileges.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: remove-invalid-database-privileges failed - roachtest.remove-invalid-database-privileges [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=7782673&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=7782673&tab=artifacts#/remove-invalid-database-privileges) on release-22.1 @ [f9730bda77ce8e3ecfa42302be58d81dcd04cd21](https://github.com/cockroachdb/cockroach/commits/f9730bda77ce8e3ecfa42302be58d81dcd04cd21):
```
The test failed on branch=release-22.1, cloud=gce:
test artifacts and logs in: /artifacts/remove-invalid-database-privileges/run_1
cluster.go:1849,versionupgrade.go:393,versionupgrade.go:354,versionupgrade.go:178,remove_invalid_database_privileges.go:117,remove_invalid_database_privileges.go:30,test_runner.go:883: one or more parallel execution failure
(1) attached stack trace
-- stack trace:
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2202
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2083
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:175
| github.com/cockroachdb/cockroach/pkg/roachprod.Start
| github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:660
| main.(*clusterImpl).StartE
| main/pkg/cmd/roachtest/cluster.go:1803
| main.(*clusterImpl).Start
| main/pkg/cmd/roachtest/cluster.go:1848
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.upgradeNodes
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/versionupgrade.go:393
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.binaryUpgradeStep.func1
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/versionupgrade.go:354
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.(*versionUpgradeTest).run
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/versionupgrade.go:178
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runRemoveInvalidDatabasePrivileges
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/remove_invalid_database_privileges.go:117
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerRemoveInvalidDatabasePrivileges.func1
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/remove_invalid_database_privileges.go:30
| main.(*testRunner).runTest.func2
| main/pkg/cmd/roachtest/test_runner.go:883
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1581
Wraps: (2) one or more parallel execution failure
Error types: (1) *withstack.withStack (2) *errutil.leafError
```
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*remove-invalid-database-privileges.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| test | roachtest remove invalid database privileges failed roachtest remove invalid database privileges with on release the test failed on branch release cloud gce test artifacts and logs in artifacts remove invalid database privileges run cluster go versionupgrade go versionupgrade go versionupgrade go remove invalid database privileges go remove invalid database privileges go test runner go one or more parallel execution failure attached stack trace stack trace github com cockroachdb cockroach pkg roachprod install syncedcluster parallele github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster parallel github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster start github com cockroachdb cockroach pkg roachprod install cockroach go github com cockroachdb cockroach pkg roachprod start github com cockroachdb cockroach pkg roachprod roachprod go main clusterimpl starte main pkg cmd roachtest cluster go main clusterimpl start main pkg cmd roachtest cluster go github com cockroachdb cockroach pkg cmd roachtest tests upgradenodes github com cockroachdb cockroach pkg cmd roachtest tests versionupgrade go github com cockroachdb cockroach pkg cmd roachtest tests binaryupgradestep github com cockroachdb cockroach pkg cmd roachtest tests versionupgrade go github com cockroachdb cockroach pkg cmd roachtest tests versionupgradetest run github com cockroachdb cockroach pkg cmd roachtest tests versionupgrade go github com cockroachdb cockroach pkg cmd roachtest tests runremoveinvaliddatabaseprivileges github com cockroachdb cockroach pkg cmd roachtest tests remove invalid database privileges go github com cockroachdb cockroach pkg cmd roachtest tests registerremoveinvaliddatabaseprivileges github com cockroachdb cockroach pkg cmd roachtest tests remove invalid database privileges go main testrunner runtest main pkg cmd roachtest test runner go runtime goexit goroot src runtime asm s wraps one or more parallel execution failure error types withstack withstack errutil leaferror help see see cc cockroachdb sql experience | 1 |
139,957 | 11,300,063,609 | IssuesEvent | 2020-01-17 12:46:30 | ckeditor/ckeditor4 | https://api.github.com/repos/ckeditor/ckeditor4 | opened | Failing tests on major | status:confirmed type:failingtest | ## Type of report
Failing tests
## Actual results
### IE 8
<img width="967" alt="" src="https://user-images.githubusercontent.com/1078728/72612895-5d035600-392e-11ea-9a2f-46da557a6ae1.png">
### IE 11
<img width="570" alt="" src="https://user-images.githubusercontent.com/1078728/72612930-72788000-392e-11ea-887c-2d30185ae18b.png">
### Edge
<img width="1258" alt="" src="https://user-images.githubusercontent.com/1078728/72612954-83c18c80-392e-11ea-8e0e-e60e8cf2857d.png">
### Safari
<img width="598" alt="" src="https://user-images.githubusercontent.com/1078728/72613019-c1beb080-392e-11ea-97ef-00aa5918b984.png">
| 1.0 | Failing tests on major - ## Type of report
Failing tests
## Actual results
### IE 8
<img width="967" alt="" src="https://user-images.githubusercontent.com/1078728/72612895-5d035600-392e-11ea-9a2f-46da557a6ae1.png">
### IE 11
<img width="570" alt="" src="https://user-images.githubusercontent.com/1078728/72612930-72788000-392e-11ea-887c-2d30185ae18b.png">
### Edge
<img width="1258" alt="" src="https://user-images.githubusercontent.com/1078728/72612954-83c18c80-392e-11ea-8e0e-e60e8cf2857d.png">
### Safari
<img width="598" alt="" src="https://user-images.githubusercontent.com/1078728/72613019-c1beb080-392e-11ea-97ef-00aa5918b984.png">
| test | failing tests on major type of report failing tests actual results ie img width alt src ie img width alt src edge img width alt src safari img width alt src | 1 |
67,030 | 16,778,780,459 | IssuesEvent | 2021-06-15 03:25:59 | kobotoolbox/kpi | https://api.github.com/repos/kobotoolbox/kpi | closed | Broken skip logic on beta branch | bug formbuilder frontend | ## Description
<!-- Description of the 🐛 bug or a 🚀 feature; you can also add one of the labels: `bug`, `enhancement`, `ideas` or `question` -->
## Steps to Reproduce
1. Create new form
2. Add `select_one` question
3. Add `text` question
4. For the `text` question open "Settings" and "Skip Logic"
5. Use "+ ADD A CONDITION" button and select the `select_one` question from the list
6. Observe that after "=" selector there is no choice selector
## Expected behavior
<img width="1018" alt="Screenshot 2021-06-11 at 19 42 17" src="https://user-images.githubusercontent.com/2521888/121728164-cd42e680-cae4-11eb-87f2-c5956a474268.png">
## Actual behavior
(Disregard the different button styles and the "X" button)
<img width="754" alt="Screenshot 2021-06-11 at 19 41 37" src="https://user-images.githubusercontent.com/2521888/121728184-d338c780-cae4-11eb-909d-5757739afef2.png"> | 1.0 | Broken skip logic on beta branch - ## Description
<!-- Description of the 🐛 bug or a 🚀 feature; you can also add one of the labels: `bug`, `enhancement`, `ideas` or `question` -->
## Steps to Reproduce
1. Create new form
2. Add `select_one` question
3. Add `text` question
4. For the `text` question open "Settings" and "Skip Logic"
5. Use "+ ADD A CONDITION" button and select the `select_one` question from the list
6. Observe that after "=" selector there is no choice selector
## Expected behavior
<img width="1018" alt="Screenshot 2021-06-11 at 19 42 17" src="https://user-images.githubusercontent.com/2521888/121728164-cd42e680-cae4-11eb-87f2-c5956a474268.png">
## Actual behavior
(Disregard the different button styles and the "X" button)
<img width="754" alt="Screenshot 2021-06-11 at 19 41 37" src="https://user-images.githubusercontent.com/2521888/121728184-d338c780-cae4-11eb-909d-5757739afef2.png"> | non_test | broken skip logic on beta branch description steps to reproduce create new form add select one question add text question for the text question open settings and skip logic use add a condition button and select the select one question from the list observe that after selector there is no choice selector expected behavior img width alt screenshot at src actual behavior disregard the different button styles and the x button img width alt screenshot at src | 0 |
157,265 | 12,368,318,684 | IssuesEvent | 2020-05-18 13:39:53 | rethinkdb/rethinkdb | https://api.github.com/repos/rethinkdb/rethinkdb | closed | test-runner reports a server as crashed when it is still running | cp:testing | When running against an external server, e.g. `./test-runner polyglot -c 60436 -d 59436 -i py -j10`, I noticed errors like this:
```
Using rethinkdb binary /home/ssd2/grey/rethinkdb/build/debug/rethinkdb
python 2.6 interpreter: /usr/bin/python2.6, driver: /home/ssd2/grey/rethinkdb/drivers/python
== Starting: polyglot/arity.py2.6 (T+ 0.3 sec)
== Starting: polyglot/aggregation.py2.6 (T+ 0.3 sec)
== Starting: polyglot/changefeeds/edge.py2.6 (T+ 0.3 sec)
== Starting: polyglot/arraylimits.py2.6 (T+ 0.4 sec)
== Starting: polyglot/changefeeds/include_states.py2.6 (T+ 0.4 sec)
== Starting: polyglot/changefeeds/idxcopy.py2.6 (T+ 0.4 sec)
== Starting: polyglot/changefeeds/point.py2.6 (T+ 0.4 sec)
== Starting: polyglot/changefeeds/sindex.py2.6 (T+ 0.4 sec)
== Starting: polyglot/changefeeds/squash.py2.6 (T+ 0.4 sec)
== Starting: polyglot/changefeeds/table.py2.6 (T+ 0.4 sec)
>>> Crashed server polyglot/arity.py2.6 after 1.0 sec (T+ 1.0 sec)
Server crashed during testing
<<< end Crashed server: polyglot/arity.py2.6
Exception in thread Thread-2:
Traceback (most recent call last):
File "/usr/lib/python2.7/threading.py", line 551, in __bootstrap_inner
self.run()
File "./test-runner", line 1464, in run
self.clean_server()
File "./test-runner", line 1282, in clean_server
raise ValueError('clean_server called when there was no connection')
ValueError: clean_server called when there was no connection
```
The server was still running, but the client connection had been closed by the server (due to an `interrupted_exc_t`, still investigating the cause). I'm not sure if there's a better way for `test-runner` to check the status of the server, but it was a little confusing.
| 1.0 | test-runner reports a server as crashed when it is still running - When running against an external server, e.g. `./test-runner polyglot -c 60436 -d 59436 -i py -j10`, I noticed errors like this:
```
Using rethinkdb binary /home/ssd2/grey/rethinkdb/build/debug/rethinkdb
python 2.6 interpreter: /usr/bin/python2.6, driver: /home/ssd2/grey/rethinkdb/drivers/python
== Starting: polyglot/arity.py2.6 (T+ 0.3 sec)
== Starting: polyglot/aggregation.py2.6 (T+ 0.3 sec)
== Starting: polyglot/changefeeds/edge.py2.6 (T+ 0.3 sec)
== Starting: polyglot/arraylimits.py2.6 (T+ 0.4 sec)
== Starting: polyglot/changefeeds/include_states.py2.6 (T+ 0.4 sec)
== Starting: polyglot/changefeeds/idxcopy.py2.6 (T+ 0.4 sec)
== Starting: polyglot/changefeeds/point.py2.6 (T+ 0.4 sec)
== Starting: polyglot/changefeeds/sindex.py2.6 (T+ 0.4 sec)
== Starting: polyglot/changefeeds/squash.py2.6 (T+ 0.4 sec)
== Starting: polyglot/changefeeds/table.py2.6 (T+ 0.4 sec)
>>> Crashed server polyglot/arity.py2.6 after 1.0 sec (T+ 1.0 sec)
Server crashed during testing
<<< end Crashed server: polyglot/arity.py2.6
Exception in thread Thread-2:
Traceback (most recent call last):
File "/usr/lib/python2.7/threading.py", line 551, in __bootstrap_inner
self.run()
File "./test-runner", line 1464, in run
self.clean_server()
File "./test-runner", line 1282, in clean_server
raise ValueError('clean_server called when there was no connection')
ValueError: clean_server called when there was no connection
```
The server was still running, but the client connection had been closed by the server (due to an `interrupted_exc_t`, still investigating the cause). I'm not sure if there's a better way for `test-runner` to check the status of the server, but it was a little confusing.
| test | test runner reports a server as crashed when it is still running when running against an external server e g test runner polyglot c d i py i noticed errors like this using rethinkdb binary home grey rethinkdb build debug rethinkdb python interpreter usr bin driver home grey rethinkdb drivers python starting polyglot arity t sec starting polyglot aggregation t sec starting polyglot changefeeds edge t sec starting polyglot arraylimits t sec starting polyglot changefeeds include states t sec starting polyglot changefeeds idxcopy t sec starting polyglot changefeeds point t sec starting polyglot changefeeds sindex t sec starting polyglot changefeeds squash t sec starting polyglot changefeeds table t sec crashed server polyglot arity after sec t sec server crashed during testing end crashed server polyglot arity exception in thread thread traceback most recent call last file usr lib threading py line in bootstrap inner self run file test runner line in run self clean server file test runner line in clean server raise valueerror clean server called when there was no connection valueerror clean server called when there was no connection the server was still running but the client connection had been closed by the server due to an interrupted exc t still investigating the cause i m not sure if there s a better way for test runner to check the status of the server but it was a little confusing | 1 |
132,188 | 28,115,777,839 | IssuesEvent | 2023-03-31 10:36:27 | BlueBrain/nmodl | https://api.github.com/repos/BlueBrain/nmodl | opened | (One More) Difference with UNITS precision in code generation | bug codegen | Consider a UNITS block like:
```console
NEURON {
SUFFIX cagk
}
UNITS {
R = 8.313424 (joule/degC)
}
```
* NEURON generates
```cpp
static double R = 8.313424;
```
* MOD2C generates
```cpp
static double R = 8.313424;
```
* NMODL generates
```cpp
static const double R = 8.31342;
```
| 1.0 | (One More) Difference with UNITS precision in code generation - Consider a UNITS block like:
```console
NEURON {
SUFFIX cagk
}
UNITS {
R = 8.313424 (joule/degC)
}
```
* NEURON generates
```cpp
static double R = 8.313424;
```
* MOD2C generates
```cpp
static double R = 8.313424;
```
* NMODL generates
```cpp
static const double R = 8.31342;
```
| non_test | one more difference with units precision in code generation consider a units block like console neuron suffix cagk units r joule degc neuron generates cpp static double r generates cpp static double r nmodl generates cpp static const double r | 0 |
258,742 | 8,179,387,710 | IssuesEvent | 2018-08-28 16:16:46 | melsicon/melsicon.de | https://api.github.com/repos/melsicon/melsicon.de | opened | Add imprint (Impressum) | Priority: High Type: Enhancement | ### **Type of Issue**
- [ ] Bug
- [X] Enhancement
- [ ] Documentation
- [ ] Maintenance
- [ ] Question
#### Priority
- [ ] Low
- [ ] Medium
- [X] High
- [ ] Critical
## **Description**
Add an imprint/impressum to the website containing all legally required information.
| 1.0 | Add imprint (Impressum) - ### **Type of Issue**
- [ ] Bug
- [X] Enhancement
- [ ] Documentation
- [ ] Maintenance
- [ ] Question
#### Priority
- [ ] Low
- [ ] Medium
- [X] High
- [ ] Critical
## **Description**
Add an imprint/impressum to the website containing all legally required information.
| non_test | add imprint impressum type of issue bug enhancement documentation maintenance question priority low medium high critical description add an imprint impressum to the website containing all legally required information | 0 |
10,183 | 6,621,149,221 | IssuesEvent | 2017-09-21 18:03:39 | coreos/bugs | https://api.github.com/repos/coreos/bugs | closed | Double-escaping locksmithd strings | area/usability component/ct kind/bug team/tools | # Issue Report #
<!-- Fill in either the 'Bug' or 'Feature Request' section -->
### Container Linux Version ###
```
computer:container-linux-config-transpiler root# git status
On branch master
Your branch is up-to-date with 'origin/master'.
nothing to commit, working tree clean
```
### Expected Behavior ###
```
locksmith:
reboot_strategy: etcd-lock
etcd_endpoints: https://internal-etcd-balancer.us-east-2.elb.amazonaws.com:2379
etcd_cafile: /etc/ssl/certs/CA.pem
etcd_certfile: /etc/ssl/certs/etcd-cert.pem
etcd_keyfile: /etc/ssl/certs/etcd-key.pem
```
Should produce correct JSON for /etc/coreos/update.conf
### Actual Behavior ###
```
{
"filesystem" : "root",
"contents" : {
"verification" : {},
"source" : "data:,%0AREBOOT_STRATEGY%3D%22etcd-lock%22%0ALOCKSMITHD_ENDPOINT%3D%22%5C%22https%3A%2F%2Finternal-etcd-balancer.us-east-2.elb.amazonaws.com%3A2379%5C%22%22%0ALOCKSMITHD_ETCD_CAFILE%3D%22%5C%22%2Fetc%2Fssl%2Fcerts%2FCA.pem%5C%22%22%0ALOCKSMITHD_ETCD_CERTFILE%3D%22%5C%22%2Fetc%2Fssl%2Fcerts%2Fetcd-cert.pem%5C%22%22%0ALOCKSMITHD_ETCD_KEYFILE%3D%22%5C%22%2Fetc%2Fssl%2Fcerts%2Fetcd-key.pem%5C%22%22"
},
"user" : {},
"mode" : 420,
"group" : {},
"path" : "/etc/coreos/update.conf"
}
```
Which when decoded produces:
```
REBOOT_STRATEGY="etcd-lock"
LOCKSMITHD_ENDPOINT="\"https://internal-etcd-balancer.us-east-2.elb.amazonaws.com:2379\""
LOCKSMITHD_ETCD_CAFILE="\"/etc/ssl/certs/CA.pem\""
LOCKSMITHD_ETCD_CERTFILE="\"/etc/ssl/certs/etcd-cert.pem\""
LOCKSMITHD_ETCD_KEYFILE="\"/etc/ssl/certs/etcd-key.pem\""
```
### Reproduction Steps ###
See above
### Other Information ###
My guess would be that it has to do with the fact that it reads from a special "locksmithd:" section but implements by creating a file in the resulting "storage:" section. It might be getting escaped both at input and output.
| True | Double-escaping locksmithd strings - # Issue Report #
<!-- Fill in either the 'Bug' or 'Feature Request' section -->
### Container Linux Version ###
```
computer:container-linux-config-transpiler root# git status
On branch master
Your branch is up-to-date with 'origin/master'.
nothing to commit, working tree clean
```
### Expected Behavior ###
```
locksmith:
reboot_strategy: etcd-lock
etcd_endpoints: https://internal-etcd-balancer.us-east-2.elb.amazonaws.com:2379
etcd_cafile: /etc/ssl/certs/CA.pem
etcd_certfile: /etc/ssl/certs/etcd-cert.pem
etcd_keyfile: /etc/ssl/certs/etcd-key.pem
```
Should produce correct JSON for /etc/coreos/update.conf
### Actual Behavior ###
```
{
"filesystem" : "root",
"contents" : {
"verification" : {},
"source" : "data:,%0AREBOOT_STRATEGY%3D%22etcd-lock%22%0ALOCKSMITHD_ENDPOINT%3D%22%5C%22https%3A%2F%2Finternal-etcd-balancer.us-east-2.elb.amazonaws.com%3A2379%5C%22%22%0ALOCKSMITHD_ETCD_CAFILE%3D%22%5C%22%2Fetc%2Fssl%2Fcerts%2FCA.pem%5C%22%22%0ALOCKSMITHD_ETCD_CERTFILE%3D%22%5C%22%2Fetc%2Fssl%2Fcerts%2Fetcd-cert.pem%5C%22%22%0ALOCKSMITHD_ETCD_KEYFILE%3D%22%5C%22%2Fetc%2Fssl%2Fcerts%2Fetcd-key.pem%5C%22%22"
},
"user" : {},
"mode" : 420,
"group" : {},
"path" : "/etc/coreos/update.conf"
}
```
Which when decoded produces:
```
REBOOT_STRATEGY="etcd-lock"
LOCKSMITHD_ENDPOINT="\"https://internal-etcd-balancer.us-east-2.elb.amazonaws.com:2379\""
LOCKSMITHD_ETCD_CAFILE="\"/etc/ssl/certs/CA.pem\""
LOCKSMITHD_ETCD_CERTFILE="\"/etc/ssl/certs/etcd-cert.pem\""
LOCKSMITHD_ETCD_KEYFILE="\"/etc/ssl/certs/etcd-key.pem\""
```
### Reproduction Steps ###
See above
### Other Information ###
My guess would be that it has to do with the fact that it reads from a special "locksmithd:" section but implements by creating a file in the resulting "storage:" section. It might be getting escaped both at input and output.
| non_test | double escaping locksmithd strings issue report container linux version computer container linux config transpiler root git status on branch master your branch is up to date with origin master nothing to commit working tree clean expected behavior locksmith reboot strategy etcd lock etcd endpoints etcd cafile etc ssl certs ca pem etcd certfile etc ssl certs etcd cert pem etcd keyfile etc ssl certs etcd key pem should produce correct json for etc coreos update conf actual behavior filesystem root contents verification source data strategy lock endpoint etcd balancer us east elb amazonaws com etcd cafile pem etcd certfile cert pem etcd keyfile key pem user mode group path etc coreos update conf which when decoded produces reboot strategy etcd lock locksmithd endpoint locksmithd etcd cafile etc ssl certs ca pem locksmithd etcd certfile etc ssl certs etcd cert pem locksmithd etcd keyfile etc ssl certs etcd key pem reproduction steps see above other information my guess would be that it has to do with the fact that it reads from a special locksmithd section but implements by creating a file in the resulting storage section it might be getting escaped both at input and output | 0 |
100,725 | 8,752,752,230 | IssuesEvent | 2018-12-14 05:00:38 | humera987/FXLabs-Test-Automation | https://api.github.com/repos/humera987/FXLabs-Test-Automation | reopened | Testing 14 : ApiV1BotClustersGetPathParamPageMysqlSqlInjectionTimebound | Testing 14 | Project : Testing 14
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=OWMwZGUyY2MtOGFhNy00OGFkLWEzNzQtOGY0MGZiM2FhOTYw; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 14 Dec 2018 04:54:43 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/bot-clusters?page=
Request :
Response :
{
"timestamp" : "2018-12-14T04:54:43.499+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/bot-clusters"
}
Logs :
Assertion [@ResponseTime < 7000 OR @ResponseTime > 10000] resolved-to [465 < 7000 OR 465 > 10000] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot --- | 1.0 | Testing 14 : ApiV1BotClustersGetPathParamPageMysqlSqlInjectionTimebound - Project : Testing 14
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=OWMwZGUyY2MtOGFhNy00OGFkLWEzNzQtOGY0MGZiM2FhOTYw; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 14 Dec 2018 04:54:43 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/bot-clusters?page=
Request :
Response :
{
"timestamp" : "2018-12-14T04:54:43.499+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/bot-clusters"
}
Logs :
Assertion [@ResponseTime < 7000 OR @ResponseTime > 10000] resolved-to [465 < 7000 OR 465 > 10000] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot --- | test | testing project testing job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api bot clusters logs assertion resolved to result assertion resolved to result fx bot | 1 |
10,477 | 3,115,160,440 | IssuesEvent | 2015-09-03 13:12:19 | Leadera/ecoman_repo | https://api.github.com/repos/Leadera/ecoman_repo | closed | Product amount can only be entered as integer | bug Dataset management waiting for testing | should be changed, also eg. 0.5 should be possible | 1.0 | Product amount can only be entered as integer - should be changed, also eg. 0.5 should be possible | test | product amount can only be entered as integer should be changed also eg should be possible | 1 |
92,172 | 8,354,372,658 | IssuesEvent | 2018-10-02 13:09:06 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: jepsen/1/bank/majority-ring failed | C-test-failure O-robot | SHA: https://github.com/cockroachdb/cockroach/commits/0f5a289d49105cd317a2713b71546d84c25547f6
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stressrace instead of stress and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
make stress TESTS=jepsen/1/bank/majority-ring PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-stderr=false -maxtime 20m -timeout 10m'
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=937459&tab=buildLog
```
The test failed on master:
test.go:500,cluster.go:890,jepsen.go:86,jepsen.go:125,jepsen.go:285: /home/agent/work/.go/bin/roachprod run teamcity-937459-jepsen-1:1-6 -- sh -c "sudo apt-get -qqy upgrade -o Dpkg::Options::='--force-confold' > /dev/null 2>&1" returned:
stderr:
stdout:
.............................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................
1:
2:
exit status 100
3:
exit status 100
4:
exit status 100
5:
exit status 100
6:
exit status 100
Error: exit status 100
: exit status 1
test.go:500,cluster.go:890,jepsen.go:74,asm_amd64.s:573,panic.go:377,test.go:501,cluster.go:890,jepsen.go:86,jepsen.go:125,jepsen.go:285: test already failed
``` | 1.0 | roachtest: jepsen/1/bank/majority-ring failed - SHA: https://github.com/cockroachdb/cockroach/commits/0f5a289d49105cd317a2713b71546d84c25547f6
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stressrace instead of stress and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
make stress TESTS=jepsen/1/bank/majority-ring PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-stderr=false -maxtime 20m -timeout 10m'
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=937459&tab=buildLog
```
The test failed on master:
test.go:500,cluster.go:890,jepsen.go:86,jepsen.go:125,jepsen.go:285: /home/agent/work/.go/bin/roachprod run teamcity-937459-jepsen-1:1-6 -- sh -c "sudo apt-get -qqy upgrade -o Dpkg::Options::='--force-confold' > /dev/null 2>&1" returned:
stderr:
stdout:
.............................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................
1:
2:
exit status 100
3:
exit status 100
4:
exit status 100
5:
exit status 100
6:
exit status 100
Error: exit status 100
: exit status 1
test.go:500,cluster.go:890,jepsen.go:74,asm_amd64.s:573,panic.go:377,test.go:501,cluster.go:890,jepsen.go:86,jepsen.go:125,jepsen.go:285: test already failed
``` | test | roachtest jepsen bank majority ring failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stressrace instead of stress and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach make stress tests jepsen bank majority ring pkg roachtest testtimeout stressflags stderr false maxtime timeout failed test the test failed on master test go cluster go jepsen go jepsen go jepsen go home agent work go bin roachprod run teamcity jepsen sh c sudo apt get qqy upgrade o dpkg options force confold dev null returned stderr stdout exit status exit status exit status exit status exit status error exit status exit status test go cluster go jepsen go asm s panic go test go cluster go jepsen go jepsen go jepsen go test already failed | 1 |
198,734 | 14,995,022,759 | IssuesEvent | 2021-01-29 13:46:18 | DocCyblade/squirrelpak-mc12 | https://api.github.com/repos/DocCyblade/squirrelpak-mc12 | opened | Changes to TComplementModules.cfg | TEST enhancement | Below are changes made:
```
modules {
# Adds casting supplies made from porcelain. Default is true
B:CeramicsPlugin=false
``` | 1.0 | Changes to TComplementModules.cfg - Below are changes made:
```
modules {
# Adds casting supplies made from porcelain. Default is true
B:CeramicsPlugin=false
``` | test | changes to tcomplementmodules cfg below are changes made modules adds casting supplies made from porcelain default is true b ceramicsplugin false | 1 |
232,564 | 18,888,221,434 | IssuesEvent | 2021-11-15 10:17:29 | cyberbotics/webots | https://api.github.com/repos/cyberbotics/webots | closed | connector_detach test sometimes failing | test suite issue | **Describe the Bug**
The following test is sometimes giving False negative:
```
FAILURE with connector_detach: Connectors should not be detached when applying a force < 2.1 N.
``` | 1.0 | connector_detach test sometimes failing - **Describe the Bug**
The following test is sometimes giving False negative:
```
FAILURE with connector_detach: Connectors should not be detached when applying a force < 2.1 N.
``` | test | connector detach test sometimes failing describe the bug the following test is sometimes giving false negative failure with connector detach connectors should not be detached when applying a force n | 1 |
213,759 | 16,535,695,710 | IssuesEvent | 2021-05-27 11:35:19 | tqtezos/baseDAO | https://api.github.com/repos/tqtezos/baseDAO | closed | Move Proposal and Management tests into sub-modules | p0 tests | # Clarification and motivation
Both the `Test.Ligo.BaseDAO.Proposal` and `Test.Ligo.BaseDAO.Management` test modules contain _some_ of the definition of the scenarios to run as well as the `TestTree`.
This however results in long compilation times that can get to be quite inconvenient, especially in case of small changes.
Note: some sub-modules already exist, e.g. `Test.Ligo.BaseDAO.Proposal.Vote` for an entrypoint, others might need to be created (possible for a behavior).
Additionally, we should consider renaming `Test.Ligo.BaseDAO.Proposal.Proposal` to `Test.Ligo.BaseDAO.Proposal.Propose`.
# Acceptance criteria
- Scenarios for both proposals and management logic are defined in different sub-modules (existing or new).
- If there are no reason against it `Test.Ligo.BaseDAO.Proposal.Proposal` is renamed to `Test.Ligo.BaseDAO.Proposal.Propose`
| 1.0 | Move Proposal and Management tests into sub-modules - # Clarification and motivation
Both the `Test.Ligo.BaseDAO.Proposal` and `Test.Ligo.BaseDAO.Management` test modules contain _some_ of the definition of the scenarios to run as well as the `TestTree`.
This however results in long compilation times that can get to be quite inconvenient, especially in case of small changes.
Note: some sub-modules already exist, e.g. `Test.Ligo.BaseDAO.Proposal.Vote` for an entrypoint, others might need to be created (possible for a behavior).
Additionally, we should consider renaming `Test.Ligo.BaseDAO.Proposal.Proposal` to `Test.Ligo.BaseDAO.Proposal.Propose`.
# Acceptance criteria
- Scenarios for both proposals and management logic are defined in different sub-modules (existing or new).
- If there are no reason against it `Test.Ligo.BaseDAO.Proposal.Proposal` is renamed to `Test.Ligo.BaseDAO.Proposal.Propose`
| test | move proposal and management tests into sub modules clarification and motivation both the test ligo basedao proposal and test ligo basedao management test modules contain some of the definition of the scenarios to run as well as the testtree this however results in long compilation times that can get to be quite inconvenient especially in case of small changes note some sub modules already exist e g test ligo basedao proposal vote for an entrypoint others might need to be created possible for a behavior additionally we should consider renaming test ligo basedao proposal proposal to test ligo basedao proposal propose acceptance criteria scenarios for both proposals and management logic are defined in different sub modules existing or new if there are no reason against it test ligo basedao proposal proposal is renamed to test ligo basedao proposal propose | 1 |
273,695 | 23,778,686,773 | IssuesEvent | 2022-09-02 00:36:40 | pyinstaller/pyinstaller | https://api.github.com/repos/pyinstaller/pyinstaller | closed | Add a test-case for testing the correct order of imports | area:test-suite area:modulegraph | AFAIK if there is such test-case yet.
This may be a bit tricky since one *may* need to fetch log-messages (I think we have code for this in the test-suite, but maybe I'm mixing this up with [ my`gitflow` tool](https://github.com/htgoebel/gitflow)) or monkey-patch ModuleGraph to get the order. | 1.0 | Add a test-case for testing the correct order of imports - AFAIK if there is such test-case yet.
This may be a bit tricky since one *may* need to fetch log-messages (I think we have code for this in the test-suite, but maybe I'm mixing this up with [ my`gitflow` tool](https://github.com/htgoebel/gitflow)) or monkey-patch ModuleGraph to get the order. | test | add a test case for testing the correct order of imports afaik if there is such test case yet this may be a bit tricky since one may need to fetch log messages i think we have code for this in the test suite but maybe i m mixing this up with or monkey patch modulegraph to get the order | 1 |
231,797 | 18,795,633,420 | IssuesEvent | 2021-11-08 21:56:54 | devfile/api | https://api.github.com/repos/devfile/api | closed | Building ODC during CI test failed | area/integration-tests | **Which area/kind this issue is related to?**
/area integration-tests
## Issue Description
CI integration tests are failed on ODC build with the following error. This can be resolved by using Go 1.16 instead of the current Go 1.13.
```
STEP 7: RUN ./build.sh
build github.com/openshift/console/cmd/bridge: cannot load io/fs: open /go/src/github.com/openshift/console/vendor/io/fs: no such file or directory
``` | 1.0 | Building ODC during CI test failed - **Which area/kind this issue is related to?**
/area integration-tests
## Issue Description
CI integration tests are failed on ODC build with the following error. This can be resolved by using Go 1.16 instead of the current Go 1.13.
```
STEP 7: RUN ./build.sh
build github.com/openshift/console/cmd/bridge: cannot load io/fs: open /go/src/github.com/openshift/console/vendor/io/fs: no such file or directory
``` | test | building odc during ci test failed which area kind this issue is related to area integration tests issue description ci integration tests are failed on odc build with the following error this can be resolved by using go instead of the current go step run build sh build github com openshift console cmd bridge cannot load io fs open go src github com openshift console vendor io fs no such file or directory | 1 |
273,618 | 23,771,482,996 | IssuesEvent | 2022-09-01 16:42:28 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Grandchildren Scenes don't inherit scripts | bug topic:core needs testing | ### Godot version
3.4stable
### System information
Ubuntu 20.04
### Issue description
When creating long chains of inherited scenes in Godot, the information of the script attached is lost for grandchildren and beyond.
### Steps to reproduce
Create a scene A, then create an inherited scene B from A. After that create an inherited scene C from B. Now attach a script to A. B will have the script attached but C will not.
Edit: All scenes must be open in the editor.
### Minimal reproduction project
[InheritanceBug.zip](https://github.com/godotengine/godot/files/7827351/InheritanceBug.zip)
| 1.0 | Grandchildren Scenes don't inherit scripts - ### Godot version
3.4stable
### System information
Ubuntu 20.04
### Issue description
When creating long chains of inherited scenes in Godot, the information of the script attached is lost for grandchildren and beyond.
### Steps to reproduce
Create a scene A, then create an inherited scene B from A. After that create an inherited scene C from B. Now attach a script to A. B will have the script attached but C will not.
Edit: All scenes must be open in the editor.
### Minimal reproduction project
[InheritanceBug.zip](https://github.com/godotengine/godot/files/7827351/InheritanceBug.zip)
| test | grandchildren scenes don t inherit scripts godot version system information ubuntu issue description when creating long chains of inherited scenes in godot the information of the script attached is lost for grandchildren and beyond steps to reproduce create a scene a then create an inherited scene b from a after that create an inherited scene c from b now attach a script to a b will have the script attached but c will not edit all scenes must be open in the editor minimal reproduction project | 1 |
344,895 | 10,349,720,671 | IssuesEvent | 2019-09-04 23:37:59 | oslc-op/oslc-specs | https://api.github.com/repos/oslc-op/oslc-specs | closed | OSLC discovery supports discovering server capabilities, but not client needs | Core: Main Spec Priority: High Status: Deferred Xtra: Jira | Just something to think about. OSLC currently supports discovery of what a server provdes, but there is no way for a client to express what it needs. Therefore there is no means for some tool that provisions, deploys and integrates tools to be able to match client needs with server capabilities.
Possibly a client to use ServiceProvider resources to express the services it needs.
---
_Migrated from https://issues.oasis-open.org/browse/OSLCCORE-49 (opened by @jamsden; previously assigned to @jamsden)_
| 1.0 | OSLC discovery supports discovering server capabilities, but not client needs - Just something to think about. OSLC currently supports discovery of what a server provdes, but there is no way for a client to express what it needs. Therefore there is no means for some tool that provisions, deploys and integrates tools to be able to match client needs with server capabilities.
Possibly a client to use ServiceProvider resources to express the services it needs.
---
_Migrated from https://issues.oasis-open.org/browse/OSLCCORE-49 (opened by @jamsden; previously assigned to @jamsden)_
| non_test | oslc discovery supports discovering server capabilities but not client needs just something to think about oslc currently supports discovery of what a server provdes but there is no way for a client to express what it needs therefore there is no means for some tool that provisions deploys and integrates tools to be able to match client needs with server capabilities possibly a client to use serviceprovider resources to express the services it needs migrated from opened by jamsden previously assigned to jamsden | 0 |
186,796 | 14,409,128,217 | IssuesEvent | 2020-12-04 01:27:31 | knative/serving | https://api.github.com/repos/knative/serving | closed | Update documentation to install kubetest correctly (or move to kubetest2 when available) | area/test-and-release kind/bug kind/doc lifecycle/stale | /area test-and-release
/kind doc
Our documentation explicitly states you should install kubetest using go get: https://github.com/knative/serving/tree/master/test#presubmit-tests.
Kubetest installation via go get is not supported; the team has stated that the correct way to install is via git clone/go install. However in the near future kubetest will be replaced with kubetest2, which will be go-gettable: https://github.com/kubernetes/test-infra/issues/18656
Depending on how long it'll take for us to change to kubetest2 we should probably update our documentation. | 1.0 | Update documentation to install kubetest correctly (or move to kubetest2 when available) - /area test-and-release
/kind doc
Our documentation explicitly states you should install kubetest using go get: https://github.com/knative/serving/tree/master/test#presubmit-tests.
Kubetest installation via go get is not supported; the team has stated that the correct way to install is via git clone/go install. However in the near future kubetest will be replaced with kubetest2, which will be go-gettable: https://github.com/kubernetes/test-infra/issues/18656
Depending on how long it'll take for us to change to kubetest2 we should probably update our documentation. | test | update documentation to install kubetest correctly or move to when available area test and release kind doc our documentation explicitly states you should install kubetest using go get kubetest installation via go get is not supported the team has stated that the correct way to install is via git clone go install however in the near future kubetest will be replaced with which will be go gettable depending on how long it ll take for us to change to we should probably update our documentation | 1 |
783,052 | 27,516,911,186 | IssuesEvent | 2023-03-06 12:36:52 | informalsystems/quint | https://api.github.com/repos/informalsystems/quint | closed | Module imports and circular imports | bug Flanguage feedback product-priority | In the course of #489, I started to write a simple example of bank transfer. Somehow, my coding habits kicked in and I have produced the following code:
```scala
module bank {
type Addr = str
// we isolate the keeper interface in a separate namespace
module keeper {
// to my surprise, this import works
// import bank.*
pure def validateBalance(addr: Addr): bool = {
true
}
}
```
There are several points to discuss here:
- do we want `Addr` to be accessible in `keeper`? I think it is normal behavior in many languages.
- I can access `Addr` in `keeper` via `bank::Addr`, but is it what we want?
- Surprisingly, `import bank.*` is working inside `keeper`. It probably leads to circular references in `keeper`.
/cc @bugarela, @shonfeder
This is probably related to the issue #483 by @lasarojc. | 1.0 | Module imports and circular imports - In the course of #489, I started to write a simple example of bank transfer. Somehow, my coding habits kicked in and I have produced the following code:
```scala
module bank {
type Addr = str
// we isolate the keeper interface in a separate namespace
module keeper {
// to my surprise, this import works
// import bank.*
pure def validateBalance(addr: Addr): bool = {
true
}
}
```
There are several points to discuss here:
- do we want `Addr` to be accessible in `keeper`? I think it is normal behavior in many languages.
- I can access `Addr` in `keeper` via `bank::Addr`, but is it what we want?
- Surprisingly, `import bank.*` is working inside `keeper`. It probably leads to circular references in `keeper`.
/cc @bugarela, @shonfeder
This is probably related to the issue #483 by @lasarojc. | non_test | module imports and circular imports in the course of i started to write a simple example of bank transfer somehow my coding habits kicked in and i have produced the following code scala module bank type addr str we isolate the keeper interface in a separate namespace module keeper to my surprise this import works import bank pure def validatebalance addr addr bool true there are several points to discuss here do we want addr to be accessible in keeper i think it is normal behavior in many languages i can access addr in keeper via bank addr but is it what we want surprisingly import bank is working inside keeper it probably leads to circular references in keeper cc bugarela shonfeder this is probably related to the issue by lasarojc | 0 |
5,701 | 8,167,158,049 | IssuesEvent | 2018-08-25 18:27:24 | ValveSoftware/Proton | https://api.github.com/repos/ValveSoftware/Proton | reopened | World of Final Fantasy - DirectX 11 Error (552700) | Game compatibility - Unofficial | When starting World of Final Fantasy (aka WOFF), the game gives a garbled error message about DirectX 11.

Log: https://gist.github.com/fureloka/272fd6a6bef87610464a9cd93d21af6f
OS: Debian Sid
CPU: AMD FX-6100
GPU: Nvidia GTX 970
GPU Driver: Nvidia 390.77 | True | World of Final Fantasy - DirectX 11 Error (552700) - When starting World of Final Fantasy (aka WOFF), the game gives a garbled error message about DirectX 11.

Log: https://gist.github.com/fureloka/272fd6a6bef87610464a9cd93d21af6f
OS: Debian Sid
CPU: AMD FX-6100
GPU: Nvidia GTX 970
GPU Driver: Nvidia 390.77 | non_test | world of final fantasy directx error when starting world of final fantasy aka woff the game gives a garbled error message about directx log os debian sid cpu amd fx gpu nvidia gtx gpu driver nvidia | 0 |
151,297 | 12,030,727,371 | IssuesEvent | 2020-04-13 08:00:14 | 2ssue/login-project | https://api.github.com/repos/2ssue/login-project | opened | Headless Test에 대해 알아보기 | 🎯test 🗃 search | 현재 구현한 테스트는 프론트엔드에서 값을 잘 넘겨준다고 했을 때 서버의 응답을 테스트하는 형태이다.
브라우저가 응답을 잘 넘겨주는지를 테스트하기 위해서 헤드리스 크롬에 대해 알아보고, Mocha로 이런 테스트를 구현할 수 있는지도 알아본다. | 1.0 | Headless Test에 대해 알아보기 - 현재 구현한 테스트는 프론트엔드에서 값을 잘 넘겨준다고 했을 때 서버의 응답을 테스트하는 형태이다.
브라우저가 응답을 잘 넘겨주는지를 테스트하기 위해서 헤드리스 크롬에 대해 알아보고, Mocha로 이런 테스트를 구현할 수 있는지도 알아본다. | test | headless test에 대해 알아보기 현재 구현한 테스트는 프론트엔드에서 값을 잘 넘겨준다고 했을 때 서버의 응답을 테스트하는 형태이다 브라우저가 응답을 잘 넘겨주는지를 테스트하기 위해서 헤드리스 크롬에 대해 알아보고 mocha로 이런 테스트를 구현할 수 있는지도 알아본다 | 1 |
183,518 | 14,235,394,384 | IssuesEvent | 2020-11-18 14:46:47 | hashgraph/hedera-services | https://api.github.com/repos/hashgraph/hedera-services | closed | Investigate the cause for performance test failure after v0.9.0-alpha.6 | Test Development bug | 1. 15N-15C performance tests passed well on November 2nd without any errors
HCS : https://hedera-hashgraph.slack.com/archives/CKWHL8R9A/p1604301824106300.
Crypto : https://hedera-hashgraph.slack.com/archives/C018Y4E6ADT/p1604301876075100
Nov 17th : TPS is not stable at 10K https://hedera-hashgraph.slack.com/archives/C018Y4E6ADT/p1605619562096400
The same issue happened with 6N network
On Nov 2nd : https://hedera-hashgraph.slack.com/archives/CKWHL8R9A/p1604352713115500, with stable TPS of 8K
On Nov7th - TPS fluctuates and not stable at 8K https://hedera-hashgraph.slack.com/archives/C018Y4E6ADT/p1605601298096300
Investigate the issue with the performance. | 1.0 | Investigate the cause for performance test failure after v0.9.0-alpha.6 - 1. 15N-15C performance tests passed well on November 2nd without any errors
HCS : https://hedera-hashgraph.slack.com/archives/CKWHL8R9A/p1604301824106300.
Crypto : https://hedera-hashgraph.slack.com/archives/C018Y4E6ADT/p1604301876075100
Nov 17th : TPS is not stable at 10K https://hedera-hashgraph.slack.com/archives/C018Y4E6ADT/p1605619562096400
The same issue happened with 6N network
On Nov 2nd : https://hedera-hashgraph.slack.com/archives/CKWHL8R9A/p1604352713115500, with stable TPS of 8K
On Nov7th - TPS fluctuates and not stable at 8K https://hedera-hashgraph.slack.com/archives/C018Y4E6ADT/p1605601298096300
Investigate the issue with the performance. | test | investigate the cause for performance test failure after alpha performance tests passed well on november without any errors hcs crypto nov tps is not stable at the same issue happened with network on nov with stable tps of on tps fluctuates and not stable at investigate the issue with the performance | 1 |
96,286 | 8,601,494,123 | IssuesEvent | 2018-11-16 11:01:06 | GTNewHorizons/NewHorizons | https://api.github.com/repos/GTNewHorizons/NewHorizons | closed | Quest update (Minor): #1728 Heavy Duty Rocket Engine Tier 3 | FixedInDev need to be tested | Quest asks for 4x Tier 3 Booster. Should be 8x Tier 3 Booster for the engines. | 1.0 | Quest update (Minor): #1728 Heavy Duty Rocket Engine Tier 3 - Quest asks for 4x Tier 3 Booster. Should be 8x Tier 3 Booster for the engines. | test | quest update minor heavy duty rocket engine tier quest asks for tier booster should be tier booster for the engines | 1 |
93,063 | 8,392,952,836 | IssuesEvent | 2018-10-09 19:06:04 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | `gci-gke-alpha-features` test suite failing to bring up cluster | kind/bug kind/failing-test priority/critical-urgent sig/instrumentation sig/network | <!-- This form is for bug reports and feature requests ONLY!
If you're looking for help check [Stack Overflow](https://stackoverflow.com/questions/tagged/kubernetes) and the [troubleshooting guide](https://kubernetes.io/docs/tasks/debug-application-cluster/troubleshooting/).
If the matter is security related, please disclose it privately via https://kubernetes.io/security/.
-->
**Is this a BUG REPORT or FEATURE REQUEST?**:
> Uncomment only one, leave it on its own line:
>
/kind bug
> /kind feature
**What happened**:
https://k8s-testgrid.appspot.com/google-gke#gci-gke-alpha-features test suite has been failing since Sept 27. Failing to bring cluster up:
```
Test Failures
BeforeSuite 10m6s
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=BeforeSuite$'
_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:137
Oct 5 16:35:23.903: Error waiting for all pods to be running and ready: 2 / 19 pods in namespace "kube-system" are NOT in RUNNING and READY state in 10m0s
POD NODE PHASE GRACE CONDITIONS
kube-dns-7df7d6c969-7cr6s gke-e2e-32483-7506d-default-pool-eb898fb6-dpdq Running [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:22:23 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:22:23 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [kubedns sidecar]} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:22:23 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [kubedns sidecar]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:22:23 +0000 UTC Reason: Message:}]
metrics-server-v0.3.1-66499c78f6-xcqgr gke-e2e-32483-7506d-default-pool-eb898fb6-dpdq Running [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:22:23 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:32:37 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [metrics-server]} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:32:37 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [metrics-server]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:22:23 +0000 UTC Reason: Message:}]
_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:190
```
Example:
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gci-gke-alpha-features/32491
**What you expected to happen**:
**How to reproduce it (as minimally and precisely as possible)**:
**Anything else we need to know?**:
**Environment**:
- Kubernetes version (use `kubectl version`):
- Cloud provider or hardware configuration:
- OS (e.g. from /etc/os-release):
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
| 1.0 | `gci-gke-alpha-features` test suite failing to bring up cluster - <!-- This form is for bug reports and feature requests ONLY!
If you're looking for help check [Stack Overflow](https://stackoverflow.com/questions/tagged/kubernetes) and the [troubleshooting guide](https://kubernetes.io/docs/tasks/debug-application-cluster/troubleshooting/).
If the matter is security related, please disclose it privately via https://kubernetes.io/security/.
-->
**Is this a BUG REPORT or FEATURE REQUEST?**:
> Uncomment only one, leave it on its own line:
>
/kind bug
> /kind feature
**What happened**:
https://k8s-testgrid.appspot.com/google-gke#gci-gke-alpha-features test suite has been failing since Sept 27. Failing to bring cluster up:
```
Test Failures
BeforeSuite 10m6s
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=BeforeSuite$'
_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:137
Oct 5 16:35:23.903: Error waiting for all pods to be running and ready: 2 / 19 pods in namespace "kube-system" are NOT in RUNNING and READY state in 10m0s
POD NODE PHASE GRACE CONDITIONS
kube-dns-7df7d6c969-7cr6s gke-e2e-32483-7506d-default-pool-eb898fb6-dpdq Running [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:22:23 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:22:23 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [kubedns sidecar]} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:22:23 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [kubedns sidecar]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:22:23 +0000 UTC Reason: Message:}]
metrics-server-v0.3.1-66499c78f6-xcqgr gke-e2e-32483-7506d-default-pool-eb898fb6-dpdq Running [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:22:23 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:32:37 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [metrics-server]} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:32:37 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [metrics-server]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-10-05 16:22:23 +0000 UTC Reason: Message:}]
_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:190
```
Example:
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gci-gke-alpha-features/32491
**What you expected to happen**:
**How to reproduce it (as minimally and precisely as possible)**:
**Anything else we need to know?**:
**Environment**:
- Kubernetes version (use `kubectl version`):
- Cloud provider or hardware configuration:
- OS (e.g. from /etc/os-release):
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
| test | gci gke alpha features test suite failing to bring up cluster this form is for bug reports and feature requests only if you re looking for help check and the if the matter is security related please disclose it privately via is this a bug report or feature request uncomment only one leave it on its own line kind bug kind feature what happened test suite has been failing since sept failing to bring cluster up test failures beforesuite go run hack go v test test args ginkgo focus beforesuite output dockerized go src io kubernetes test go oct error waiting for all pods to be running and ready pods in namespace kube system are not in running and ready state in pod node phase grace conditions kube dns gke default pool dpdq running type containersready status false lastprobetime utc lasttransitiontime utc reason containersnotready message containers with unready status type podscheduled status true lastprobetime utc lasttransitiontime utc reason message metrics server xcqgr gke default pool dpdq running type containersready status false lastprobetime utc lasttransitiontime utc reason containersnotready message containers with unready status type podscheduled status true lastprobetime utc lasttransitiontime utc reason message output dockerized go src io kubernetes test go example what you expected to happen how to reproduce it as minimally and precisely as possible anything else we need to know environment kubernetes version use kubectl version cloud provider or hardware configuration os e g from etc os release kernel e g uname a install tools others | 1 |
45,745 | 5,730,517,514 | IssuesEvent | 2017-04-21 09:38:04 | xcat2/xcat-core | https://api.github.com/repos/xcat2/xcat-core | closed | [Test case]Refine installation test cases | component:test priority:high sprint1 | The installation test cases include the following 8 cases:
```
-rw-r--r-- 1 root root 1580 Mar 30 22:16 reg_linux_diskfull_installation_hierarchy
-rw-r--r-- 1 root root 4078 Mar 30 22:16 reg_linux_diskless_installation_flat
-rw-r--r-- 1 root root 3183 Mar 30 22:16 reg_linux_diskless_installation_hierarchy
-rw-r--r-- 1 root root 5374 Mar 30 22:16 reg_linux_statelite_installation_flat
-rw-r--r-- 1 root root 4348 Mar 30 22:16 reg_linux_statelite_installation_hierarchy_by_nfs
-rw-r--r-- 1 root root 4509 Mar 30 22:16 reg_linux_statelite_installation_hierarchy_by_ramdisk
-rw-r--r-- 1 root root 1449 Mar 30 22:16 setup_vm
-rw-r--r-- 1 root root 3823 Mar 30 22:16 SN_setup_case
```
1. syncfile cleanup for the first 6 installation process above
2. add debug information:
2.1 cat dhcp lease file
2.2 cat harddisk spaces
2.3 check whether fs is writable | 1.0 | [Test case]Refine installation test cases - The installation test cases include the following 8 cases:
```
-rw-r--r-- 1 root root 1580 Mar 30 22:16 reg_linux_diskfull_installation_hierarchy
-rw-r--r-- 1 root root 4078 Mar 30 22:16 reg_linux_diskless_installation_flat
-rw-r--r-- 1 root root 3183 Mar 30 22:16 reg_linux_diskless_installation_hierarchy
-rw-r--r-- 1 root root 5374 Mar 30 22:16 reg_linux_statelite_installation_flat
-rw-r--r-- 1 root root 4348 Mar 30 22:16 reg_linux_statelite_installation_hierarchy_by_nfs
-rw-r--r-- 1 root root 4509 Mar 30 22:16 reg_linux_statelite_installation_hierarchy_by_ramdisk
-rw-r--r-- 1 root root 1449 Mar 30 22:16 setup_vm
-rw-r--r-- 1 root root 3823 Mar 30 22:16 SN_setup_case
```
1. syncfile cleanup for the first 6 installation process above
2. add debug information:
2.1 cat dhcp lease file
2.2 cat harddisk spaces
2.3 check whether fs is writable | test | refine installation test cases the installation test cases include the following cases rw r r root root mar reg linux diskfull installation hierarchy rw r r root root mar reg linux diskless installation flat rw r r root root mar reg linux diskless installation hierarchy rw r r root root mar reg linux statelite installation flat rw r r root root mar reg linux statelite installation hierarchy by nfs rw r r root root mar reg linux statelite installation hierarchy by ramdisk rw r r root root mar setup vm rw r r root root mar sn setup case syncfile cleanup for the first installation process above add debug information cat dhcp lease file cat harddisk spaces check whether fs is writable | 1 |
211,916 | 23,856,781,823 | IssuesEvent | 2022-09-07 01:02:29 | kedacore/test-tools | https://api.github.com/repos/kedacore/test-tools | opened | CVE-2022-38751 (Medium) detected in snakeyaml-1.26.jar | security vulnerability | ## CVE-2022-38751 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.26.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /e2e/images/artemis/consumer/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.26/snakeyaml-1.26.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.26/snakeyaml-1.26.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-2.3.0.RELEASE.jar (Root Library)
- :x: **snakeyaml-1.26.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kedacore/test-tools/commit/2c144e12e5f278d59cbdc4f4eb3c652e0d62591e">2c144e12e5f278d59cbdc4f4eb3c652e0d62591e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow.
<p>Publish Date: 2022-09-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38751>CVE-2022-38751</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-38751 (Medium) detected in snakeyaml-1.26.jar - ## CVE-2022-38751 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.26.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /e2e/images/artemis/consumer/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.26/snakeyaml-1.26.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.26/snakeyaml-1.26.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-2.3.0.RELEASE.jar (Root Library)
- :x: **snakeyaml-1.26.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kedacore/test-tools/commit/2c144e12e5f278d59cbdc4f4eb3c652e0d62591e">2c144e12e5f278d59cbdc4f4eb3c652e0d62591e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow.
<p>Publish Date: 2022-09-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38751>CVE-2022-38751</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in snakeyaml jar cve medium severity vulnerability vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file images artemis consumer pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter release jar root library x snakeyaml jar vulnerable library found in head commit a href found in base branch main vulnerability details using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend | 0 |
10,610 | 4,074,410,431 | IssuesEvent | 2016-05-28 12:30:36 | SleepyTrousers/EnderIO | https://api.github.com/repos/SleepyTrousers/EnderIO | closed | Wrong value for burn rate in stirling generators | 1.7 Code Complete UI Bug | #### Issue Description:
Burn rate shows with octatic capacitor in stirling generator as: 1,3333334x which seems incorrect.
#### What happens:
Burn rate is 1,3333334x in stirling generator(when using octatic capacitor) without capacitor it shows 1.0x(which seems more logical then 1.3333334x)
#### What you expected to happen:
It would be logical if the burn rate followed the same logic, as without capacitor, for example 1,4x or 1,3x instead of 1.3333334x
#### Steps to reproduce:
1.Get EnderIO 424.
2.Place down stirliing generator with octatic capacitor.
3.Look that it shows 1,3333334x instead of round number.

...
____
#### Affected Versions (Do *not* use "latest"):
- EnderIO:beta 424
- EnderCore:beta 0.2.0.21
- Minecraft: 1.7.10
- Forge:1614
| 1.0 | Wrong value for burn rate in stirling generators - #### Issue Description:
Burn rate shows with octatic capacitor in stirling generator as: 1,3333334x which seems incorrect.
#### What happens:
Burn rate is 1,3333334x in stirling generator(when using octatic capacitor) without capacitor it shows 1.0x(which seems more logical then 1.3333334x)
#### What you expected to happen:
It would be logical if the burn rate followed the same logic, as without capacitor, for example 1,4x or 1,3x instead of 1.3333334x
#### Steps to reproduce:
1.Get EnderIO 424.
2.Place down stirliing generator with octatic capacitor.
3.Look that it shows 1,3333334x instead of round number.

...
____
#### Affected Versions (Do *not* use "latest"):
- EnderIO:beta 424
- EnderCore:beta 0.2.0.21
- Minecraft: 1.7.10
- Forge:1614
| non_test | wrong value for burn rate in stirling generators issue description burn rate shows with octatic capacitor in stirling generator as which seems incorrect what happens burn rate is in stirling generator when using octatic capacitor without capacitor it shows which seems more logical then what you expected to happen it would be logical if the burn rate followed the same logic as without capacitor for example or instead of steps to reproduce get enderio place down stirliing generator with octatic capacitor look that it shows instead of round number affected versions do not use latest enderio beta endercore beta minecraft forge | 0 |
31,652 | 4,713,514,255 | IssuesEvent | 2016-10-14 20:20:48 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | opened | kubernetes-e2e-gci-gke-reboot-release-1.4: broken test run | kind/flake priority/P2 team/test-infra | https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gci-gke-reboot-release-1.4/681/
Multiple broken tests:
Failed: [k8s.io] Reboot [Disruptive] [Feature:Reboot] each node by triggering kernel panic and ensure they function upon restart {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:104
Oct 14 12:11:55.486: Test failed; at least one node failed to reboot in the time given.
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:158
```
Issues about this test specifically: #34123
Failed: [k8s.io] Reboot [Disruptive] [Feature:Reboot] each node by ordering unclean reboot and ensure they function upon restart {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:98
Oct 14 12:22:03.947: Test failed; at least one node failed to reboot in the time given.
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:158
```
Issues about this test specifically: #33882
Failed: [k8s.io] Reboot [Disruptive] [Feature:Reboot] each node by dropping all outbound packets for a while and ensure they function afterwards {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:126
Oct 14 12:32:12.186: Test failed; at least one node failed to reboot in the time given.
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:158
```
Issues about this test specifically: #33703
Failed: [k8s.io] Reboot [Disruptive] [Feature:Reboot] each node by ordering clean reboot and ensure they function upon restart {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:92
Oct 14 12:42:20.737: Test failed; at least one node failed to reboot in the time given.
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:158
```
Issues about this test specifically: #33874
Failed: [k8s.io] Reboot [Disruptive] [Feature:Reboot] each node by switching off the network interface and ensure they function upon switch on {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:110
Oct 14 12:52:29.003: Test failed; at least one node failed to reboot in the time given.
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:158
```
Issues about this test specifically: #33407 #33623
Failed: [k8s.io] Reboot [Disruptive] [Feature:Reboot] each node by dropping all inbound packets for a while and ensure they function afterwards {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:118
Oct 14 13:02:37.484: Test failed; at least one node failed to reboot in the time given.
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:158
```
Issues about this test specifically: #33405
Previous issues for this suite: #33421 #34096 #34697
| 1.0 | kubernetes-e2e-gci-gke-reboot-release-1.4: broken test run - https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gci-gke-reboot-release-1.4/681/
Multiple broken tests:
Failed: [k8s.io] Reboot [Disruptive] [Feature:Reboot] each node by triggering kernel panic and ensure they function upon restart {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:104
Oct 14 12:11:55.486: Test failed; at least one node failed to reboot in the time given.
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:158
```
Issues about this test specifically: #34123
Failed: [k8s.io] Reboot [Disruptive] [Feature:Reboot] each node by ordering unclean reboot and ensure they function upon restart {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:98
Oct 14 12:22:03.947: Test failed; at least one node failed to reboot in the time given.
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:158
```
Issues about this test specifically: #33882
Failed: [k8s.io] Reboot [Disruptive] [Feature:Reboot] each node by dropping all outbound packets for a while and ensure they function afterwards {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:126
Oct 14 12:32:12.186: Test failed; at least one node failed to reboot in the time given.
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:158
```
Issues about this test specifically: #33703
Failed: [k8s.io] Reboot [Disruptive] [Feature:Reboot] each node by ordering clean reboot and ensure they function upon restart {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:92
Oct 14 12:42:20.737: Test failed; at least one node failed to reboot in the time given.
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:158
```
Issues about this test specifically: #33874
Failed: [k8s.io] Reboot [Disruptive] [Feature:Reboot] each node by switching off the network interface and ensure they function upon switch on {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:110
Oct 14 12:52:29.003: Test failed; at least one node failed to reboot in the time given.
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:158
```
Issues about this test specifically: #33407 #33623
Failed: [k8s.io] Reboot [Disruptive] [Feature:Reboot] each node by dropping all inbound packets for a while and ensure they function afterwards {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:118
Oct 14 13:02:37.484: Test failed; at least one node failed to reboot in the time given.
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/reboot.go:158
```
Issues about this test specifically: #33405
Previous issues for this suite: #33421 #34096 #34697
| test | kubernetes gci gke reboot release broken test run multiple broken tests failed reboot each node by triggering kernel panic and ensure they function upon restart kubernetes suite go src io kubernetes output dockerized go src io kubernetes test reboot go oct test failed at least one node failed to reboot in the time given go src io kubernetes output dockerized go src io kubernetes test reboot go issues about this test specifically failed reboot each node by ordering unclean reboot and ensure they function upon restart kubernetes suite go src io kubernetes output dockerized go src io kubernetes test reboot go oct test failed at least one node failed to reboot in the time given go src io kubernetes output dockerized go src io kubernetes test reboot go issues about this test specifically failed reboot each node by dropping all outbound packets for a while and ensure they function afterwards kubernetes suite go src io kubernetes output dockerized go src io kubernetes test reboot go oct test failed at least one node failed to reboot in the time given go src io kubernetes output dockerized go src io kubernetes test reboot go issues about this test specifically failed reboot each node by ordering clean reboot and ensure they function upon restart kubernetes suite go src io kubernetes output dockerized go src io kubernetes test reboot go oct test failed at least one node failed to reboot in the time given go src io kubernetes output dockerized go src io kubernetes test reboot go issues about this test specifically failed reboot each node by switching off the network interface and ensure they function upon switch on kubernetes suite go src io kubernetes output dockerized go src io kubernetes test reboot go oct test failed at least one node failed to reboot in the time given go src io kubernetes output dockerized go src io kubernetes test reboot go issues about this test specifically failed reboot each node by dropping all inbound packets for a while and ensure they function afterwards kubernetes suite go src io kubernetes output dockerized go src io kubernetes test reboot go oct test failed at least one node failed to reboot in the time given go src io kubernetes output dockerized go src io kubernetes test reboot go issues about this test specifically previous issues for this suite | 1 |
228,113 | 18,159,211,699 | IssuesEvent | 2021-09-27 07:39:11 | SAP/ui5-webcomponents | https://api.github.com/repos/SAP/ui5-webcomponents | closed | [Android] MultiComboBox's filter selected values button does nothing | Medium Prio TOPIC RL 1.0 Release Testing | ### **Bug Description**
When the user has something selected in the MultiComboBox, opens the picker and presses on the "Filter selected items" button (in the top right), nothing happens - all available items are still shown.
### **Expected Behavior**
The button should hide the unselected items and leave only the selected ones.
### **Steps to Reproduce**
1. Go to https://sap.github.io/ui5-webcomponents/master/playground/components/MultiComboBox/
2. Click on the MultiComboBox with items
3. Toggle the "Filter selected items" button - nothing happens
### **Context**
- UI5 Web Components version: master
- OS/Platform: Android 10
- Browser: Chrome
- Affected component: MultiComboBox
### **Priority**
- [ ] Low
- [x] Medium
- [ ] High
- [ ] Very High | 1.0 | [Android] MultiComboBox's filter selected values button does nothing - ### **Bug Description**
When the user has something selected in the MultiComboBox, opens the picker and presses on the "Filter selected items" button (in the top right), nothing happens - all available items are still shown.
### **Expected Behavior**
The button should hide the unselected items and leave only the selected ones.
### **Steps to Reproduce**
1. Go to https://sap.github.io/ui5-webcomponents/master/playground/components/MultiComboBox/
2. Click on the MultiComboBox with items
3. Toggle the "Filter selected items" button - nothing happens
### **Context**
- UI5 Web Components version: master
- OS/Platform: Android 10
- Browser: Chrome
- Affected component: MultiComboBox
### **Priority**
- [ ] Low
- [x] Medium
- [ ] High
- [ ] Very High | test | multicombobox s filter selected values button does nothing bug description when the user has something selected in the multicombobox opens the picker and presses on the filter selected items button in the top right nothing happens all available items are still shown expected behavior the button should hide the unselected items and leave only the selected ones steps to reproduce go to click on the multicombobox with items toggle the filter selected items button nothing happens context web components version master os platform android browser chrome affected component multicombobox priority low medium high very high | 1 |
10,509 | 6,774,536,554 | IssuesEvent | 2017-10-27 10:46:02 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Changing the language directly in the project manager | enhancement hacktoberfest topic:editor usability | Because we can, and since we spawn a new process for the editor anyway, it would be the right time to select another language.
In a more practical context, it would also make it very easy to find if we want to encourage people to make screenshots in English for bug reports, and not in their own locale.
| True | Changing the language directly in the project manager - Because we can, and since we spawn a new process for the editor anyway, it would be the right time to select another language.
In a more practical context, it would also make it very easy to find if we want to encourage people to make screenshots in English for bug reports, and not in their own locale.
| non_test | changing the language directly in the project manager because we can and since we spawn a new process for the editor anyway it would be the right time to select another language in a more practical context it would also make it very easy to find if we want to encourage people to make screenshots in english for bug reports and not in their own locale | 0 |
372,111 | 11,009,337,477 | IssuesEvent | 2019-12-04 12:27:07 | bounswe/bounswe2019group1 | https://api.github.com/repos/bounswe/bounswe2019group1 | closed | Create write article page. | Effort: Medium Platform: Android Priority: High | The page for writing articles must be implemented. The page will also contain a private/public button for the user so that the user can choose the privacy option of the article. | 1.0 | Create write article page. - The page for writing articles must be implemented. The page will also contain a private/public button for the user so that the user can choose the privacy option of the article. | non_test | create write article page the page for writing articles must be implemented the page will also contain a private public button for the user so that the user can choose the privacy option of the article | 0 |
272,472 | 23,677,231,725 | IssuesEvent | 2022-08-28 09:12:25 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: schemachange/invertedindex failed | C-test-failure O-robot O-roachtest release-blocker branch-release-22.1 | roachtest.schemachange/invertedindex [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=6267032&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=6267032&tab=artifacts#/schemachange/invertedindex) on release-22.1 @ [fd04c2f3d7675d5a4a5a6c7d0c53cc502b019529](https://github.com/cockroachdb/cockroach/commits/fd04c2f3d7675d5a4a5a6c7d0c53cc502b019529):
```
| 279.0s 0 642.1 676.9 1.3 3.5 5.5 8.1 read
| 279.0s 0 786.1 688.0 8.1 15.7 22.0 32.5 write
| 280.0s 0 745.6 677.2 1.2 3.5 5.5 7.9 read
| 280.0s 0 802.6 688.4 7.9 15.7 21.0 26.2 write
| _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
| 281.0s 0 810.1 677.7 1.2 3.3 4.2 11.0 read
| 281.0s 0 796.1 688.8 8.1 15.2 19.9 35.7 write
| 282.0s 0 745.9 677.9 1.2 4.2 9.4 13.1 read
| 282.0s 0 621.9 688.6 8.9 26.2 39.8 52.4 write
| 283.0s 0 603.2 677.6 1.2 6.3 11.0 14.7 read
| 283.0s 0 551.2 688.1 9.4 31.5 54.5 92.3 write
Wraps: (4) COMMAND_PROBLEM
Wraps: (5) Node 5. Command with error:
| ``````
| ./workload run json --read-percent=50 --duration 1h0m0s {pgurl:1-4} --sequential
| ``````
Wraps: (6) exit status 1
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.Cmd (5) *hintdetail.withDetail (6) *exec.ExitError
monitor.go:127,inverted_index.go:109,inverted_index.go:32,test_runner.go:883: monitor failure: monitor task failed: t.Fatal() was called
(1) attached stack trace
-- stack trace:
| main.(*monitorImpl).WaitE
| main/pkg/cmd/roachtest/monitor.go:115
| main.(*monitorImpl).Wait
| main/pkg/cmd/roachtest/monitor.go:123
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runSchemaChangeInvertedIndex
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/inverted_index.go:109
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerSchemaChangeInvertedIndex.func1
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/inverted_index.go:32
| main.(*testRunner).runTest.func2
| main/pkg/cmd/roachtest/test_runner.go:883
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
-- stack trace:
| main.(*monitorImpl).wait.func2
| main/pkg/cmd/roachtest/monitor.go:171
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
-- stack trace:
| main.init
| main/pkg/cmd/roachtest/monitor.go:80
| runtime.doInit
| GOROOT/src/runtime/proc.go:6498
| runtime.main
| GOROOT/src/runtime/proc.go:238
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1581
Wraps: (6) t.Fatal() was called
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.leafError
```
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-schema
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/invertedindex.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: schemachange/invertedindex failed - roachtest.schemachange/invertedindex [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=6267032&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=6267032&tab=artifacts#/schemachange/invertedindex) on release-22.1 @ [fd04c2f3d7675d5a4a5a6c7d0c53cc502b019529](https://github.com/cockroachdb/cockroach/commits/fd04c2f3d7675d5a4a5a6c7d0c53cc502b019529):
```
| 279.0s 0 642.1 676.9 1.3 3.5 5.5 8.1 read
| 279.0s 0 786.1 688.0 8.1 15.7 22.0 32.5 write
| 280.0s 0 745.6 677.2 1.2 3.5 5.5 7.9 read
| 280.0s 0 802.6 688.4 7.9 15.7 21.0 26.2 write
| _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
| 281.0s 0 810.1 677.7 1.2 3.3 4.2 11.0 read
| 281.0s 0 796.1 688.8 8.1 15.2 19.9 35.7 write
| 282.0s 0 745.9 677.9 1.2 4.2 9.4 13.1 read
| 282.0s 0 621.9 688.6 8.9 26.2 39.8 52.4 write
| 283.0s 0 603.2 677.6 1.2 6.3 11.0 14.7 read
| 283.0s 0 551.2 688.1 9.4 31.5 54.5 92.3 write
Wraps: (4) COMMAND_PROBLEM
Wraps: (5) Node 5. Command with error:
| ``````
| ./workload run json --read-percent=50 --duration 1h0m0s {pgurl:1-4} --sequential
| ``````
Wraps: (6) exit status 1
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.Cmd (5) *hintdetail.withDetail (6) *exec.ExitError
monitor.go:127,inverted_index.go:109,inverted_index.go:32,test_runner.go:883: monitor failure: monitor task failed: t.Fatal() was called
(1) attached stack trace
-- stack trace:
| main.(*monitorImpl).WaitE
| main/pkg/cmd/roachtest/monitor.go:115
| main.(*monitorImpl).Wait
| main/pkg/cmd/roachtest/monitor.go:123
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runSchemaChangeInvertedIndex
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/inverted_index.go:109
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerSchemaChangeInvertedIndex.func1
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/inverted_index.go:32
| main.(*testRunner).runTest.func2
| main/pkg/cmd/roachtest/test_runner.go:883
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
-- stack trace:
| main.(*monitorImpl).wait.func2
| main/pkg/cmd/roachtest/monitor.go:171
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
-- stack trace:
| main.init
| main/pkg/cmd/roachtest/monitor.go:80
| runtime.doInit
| GOROOT/src/runtime/proc.go:6498
| runtime.main
| GOROOT/src/runtime/proc.go:238
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1581
Wraps: (6) t.Fatal() was called
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.leafError
```
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-schema
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/invertedindex.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| test | roachtest schemachange invertedindex failed roachtest schemachange invertedindex with on release read write read write elapsed errors ops sec inst ops sec cum ms ms ms pmax ms read write read write read write wraps command problem wraps node command with error workload run json read percent duration pgurl sequential wraps exit status error types withstack withstack errutil withprefix cluster withcommanddetails errors cmd hintdetail withdetail exec exiterror monitor go inverted index go inverted index go test runner go monitor failure monitor task failed t fatal was called attached stack trace stack trace main monitorimpl waite main pkg cmd roachtest monitor go main monitorimpl wait main pkg cmd roachtest monitor go github com cockroachdb cockroach pkg cmd roachtest tests runschemachangeinvertedindex github com cockroachdb cockroach pkg cmd roachtest tests inverted index go github com cockroachdb cockroach pkg cmd roachtest tests registerschemachangeinvertedindex github com cockroachdb cockroach pkg cmd roachtest tests inverted index go main testrunner runtest main pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace stack trace main monitorimpl wait main pkg cmd roachtest monitor go wraps monitor task failed wraps attached stack trace stack trace main init main pkg cmd roachtest monitor go runtime doinit goroot src runtime proc go runtime main goroot src runtime proc go runtime goexit goroot src runtime asm s wraps t fatal was called error types withstack withstack errutil withprefix withstack withstack errutil withprefix withstack withstack errutil leaferror help see see cc cockroachdb sql schema | 1 |
77,458 | 7,575,047,240 | IssuesEvent | 2018-04-23 23:29:46 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | Flake TestRouterDuplications | component/networking kind/test-flake lifecycle/rotten priority/P1 | Seen in https://ci.openshift.redhat.com/jenkins/job/test_pull_request_origin_integration/989/consoleFull
```
--- FAIL: TestRouterDuplications (2.35s)
router_test.go:837: Unable to validate both routes in a duplicate service scenario. Resp 1: %!s(<nil>), Resp 2: Get http://0.0.0.0: read tcp 127.0.0.1:33922->127.0.0.1:80: read: connection reset by peer
```
| 1.0 | Flake TestRouterDuplications - Seen in https://ci.openshift.redhat.com/jenkins/job/test_pull_request_origin_integration/989/consoleFull
```
--- FAIL: TestRouterDuplications (2.35s)
router_test.go:837: Unable to validate both routes in a duplicate service scenario. Resp 1: %!s(<nil>), Resp 2: Get http://0.0.0.0: read tcp 127.0.0.1:33922->127.0.0.1:80: read: connection reset by peer
```
| test | flake testrouterduplications seen in fail testrouterduplications router test go unable to validate both routes in a duplicate service scenario resp s resp get read tcp read connection reset by peer | 1 |
16,207 | 3,510,306,310 | IssuesEvent | 2016-01-09 10:57:49 | radare/radare2 | https://api.github.com/repos/radare/radare2 | closed | ahi 2 mov eax, 'A' instead of mov eax, 'AAAA' | easy enhancement has-test test-attached | ahi 2 mov eax, 'A' instead of mov eax, 'AAAA'
https://github.com/radare/radare2-regressions/blob/6913c508304e3c25c1a04a81afe02be6f5634c98/t/cmd/ahi#L164 | 2.0 | ahi 2 mov eax, 'A' instead of mov eax, 'AAAA' - ahi 2 mov eax, 'A' instead of mov eax, 'AAAA'
https://github.com/radare/radare2-regressions/blob/6913c508304e3c25c1a04a81afe02be6f5634c98/t/cmd/ahi#L164 | test | ahi mov eax a instead of mov eax aaaa ahi mov eax a instead of mov eax aaaa | 1 |
330,386 | 28,373,159,794 | IssuesEvent | 2023-04-12 18:36:38 | EECE3093C/team-project-the-over-under | https://api.github.com/repos/EECE3093C/team-project-the-over-under | closed | Testing: Write unit test for webscrapers in Python | testing | Target:
Pass/fail on whether the webscrapers are running. This would help us be sure our code isn't getting flagged by bot detection.
Delierable:
test_Webscraper.py | 1.0 | Testing: Write unit test for webscrapers in Python - Target:
Pass/fail on whether the webscrapers are running. This would help us be sure our code isn't getting flagged by bot detection.
Delierable:
test_Webscraper.py | test | testing write unit test for webscrapers in python target pass fail on whether the webscrapers are running this would help us be sure our code isn t getting flagged by bot detection delierable test webscraper py | 1 |
95,356 | 8,556,239,136 | IssuesEvent | 2018-11-08 12:33:06 | mautic/mautic | https://api.github.com/repos/mautic/mautic | closed | frequency rule does not apply on campaign | Bug Ready To Test | **Please DO NOT report security vulnerabilities here. Send them to security@mautic.com instead.**
What type of report is this:
| Q | A
| ---| ---
| Bug report? | X
| Feature request? |
| Enhancement? |
## Description:
If you set up frequency rule in the general configuration, emails are still sent from campaigns.
## If a bug:
| Q | A
| --- | ---
| Mautic version | 2.14
| PHP version |
### Steps to reproduce:
1. Set up frequency rules (for example: 2 emails per week)
2. create a contact
3. Create 2 segment emails and send it to the contact
4. Create a campaign email then create a campaign and send this third email to the contact
5. See that contact received the email (it should not as we set up in the general configuration not more than 2 emails per week)
### Log errors:
_Please check for related errors in the latest log file in [mautic root]/app/log/ and/or the web server's logs and post them here. Be sure to remove sensitive information if applicable._
| 1.0 | frequency rule does not apply on campaign - **Please DO NOT report security vulnerabilities here. Send them to security@mautic.com instead.**
What type of report is this:
| Q | A
| ---| ---
| Bug report? | X
| Feature request? |
| Enhancement? |
## Description:
If you set up frequency rule in the general configuration, emails are still sent from campaigns.
## If a bug:
| Q | A
| --- | ---
| Mautic version | 2.14
| PHP version |
### Steps to reproduce:
1. Set up frequency rules (for example: 2 emails per week)
2. create a contact
3. Create 2 segment emails and send it to the contact
4. Create a campaign email then create a campaign and send this third email to the contact
5. See that contact received the email (it should not as we set up in the general configuration not more than 2 emails per week)
### Log errors:
_Please check for related errors in the latest log file in [mautic root]/app/log/ and/or the web server's logs and post them here. Be sure to remove sensitive information if applicable._
| test | frequency rule does not apply on campaign please do not report security vulnerabilities here send them to security mautic com instead what type of report is this q a bug report x feature request enhancement description if you set up frequency rule in the general configuration emails are still sent from campaigns if a bug q a mautic version php version steps to reproduce set up frequency rules for example emails per week create a contact create segment emails and send it to the contact create a campaign email then create a campaign and send this third email to the contact see that contact received the email it should not as we set up in the general configuration not more than emails per week log errors please check for related errors in the latest log file in app log and or the web server s logs and post them here be sure to remove sensitive information if applicable | 1 |
219,269 | 17,083,610,171 | IssuesEvent | 2021-07-08 08:59:03 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | kubeadm-kinder-rootless-latest task-08-upgrade failed | area/kubeadm kind/failing-test needs-triage sig/cluster-lifecycle | #### Which jobs are failing:
https://testgrid.k8s.io/sig-cluster-lifecycle-kubeadm#kubeadm-kinder-rootless-latest
#### Which test(s) are failing:
task-08-upgrade
#### Since when has it been failing:
https://github.com/kubernetes/kubernetes/compare/ec39cc2ea...17f6f2862?
#### Testgrid link:
#### Reason for failure:
> [preflight] Some fatal errors occurred:
> [ERROR ImagePull]: unsupported pull policy ""
> [preflight] If you know what you are doing, you can make a check non-fatal with `--ignore-preflight-errors=...`
> [upgrade/version] Found 1 potential version compatibility errors but skipping since the --force flag is set:
>
> - Specified version to upgrade to "v1.22.0-beta.0.547+ec39cc2eafffa5" is an unstable version and such upgrades weren't allowed via setting the --allow-*-
#### Anything else we need to know:
kubeadm 1.22 related issue: https://github.com/kubernetes/kubeadm/issues/2473
| 1.0 | kubeadm-kinder-rootless-latest task-08-upgrade failed - #### Which jobs are failing:
https://testgrid.k8s.io/sig-cluster-lifecycle-kubeadm#kubeadm-kinder-rootless-latest
#### Which test(s) are failing:
task-08-upgrade
#### Since when has it been failing:
https://github.com/kubernetes/kubernetes/compare/ec39cc2ea...17f6f2862?
#### Testgrid link:
#### Reason for failure:
> [preflight] Some fatal errors occurred:
> [ERROR ImagePull]: unsupported pull policy ""
> [preflight] If you know what you are doing, you can make a check non-fatal with `--ignore-preflight-errors=...`
> [upgrade/version] Found 1 potential version compatibility errors but skipping since the --force flag is set:
>
> - Specified version to upgrade to "v1.22.0-beta.0.547+ec39cc2eafffa5" is an unstable version and such upgrades weren't allowed via setting the --allow-*-
#### Anything else we need to know:
kubeadm 1.22 related issue: https://github.com/kubernetes/kubeadm/issues/2473
| test | kubeadm kinder rootless latest task upgrade failed which jobs are failing which test s are failing task upgrade since when has it been failing testgrid link reason for failure some fatal errors occurred unsupported pull policy if you know what you are doing you can make a check non fatal with ignore preflight errors found potential version compatibility errors but skipping since the force flag is set specified version to upgrade to beta is an unstable version and such upgrades weren t allowed via setting the allow anything else we need to know kubeadm related issue | 1 |
3,077 | 2,653,521,982 | IssuesEvent | 2015-03-17 00:12:26 | GoogleCloudPlatform/kubernetes | https://api.github.com/repos/GoogleCloudPlatform/kubernetes | closed | pkg/proxy/proxier_test.go crash on OSX | kind/bug team/testing | pkg/proxy/proxier_test.go crashes on macos
ChooseHostInterface() returns an error and so NewProxier is nil, so stuff crashes. | 1.0 | pkg/proxy/proxier_test.go crash on OSX - pkg/proxy/proxier_test.go crashes on macos
ChooseHostInterface() returns an error and so NewProxier is nil, so stuff crashes. | test | pkg proxy proxier test go crash on osx pkg proxy proxier test go crashes on macos choosehostinterface returns an error and so newproxier is nil so stuff crashes | 1 |
212,184 | 16,474,875,209 | IssuesEvent | 2021-05-24 02:58:59 | Berea-CS-Courses/capstone-project-EzraSkwarka | https://api.github.com/repos/Berea-CS-Courses/capstone-project-EzraSkwarka | opened | Minor: Change Node Spawns to use a Weight System | Minor Issue Polish documentation | Currently, the node spawning system uses raw percentile to determine what node to spawn. This makes the process of adding a new type of node into a region quite complicated, instead, I want to transition to using a weight-based system. For instance, I will designate an item as being common, uncommon, rare, or ultra-rare. I will then determine which Rarity I'm going to spawn, and then from within that Rarity choose any one of these suitable nodes.
As it stands right now, a particular case looks like this:
```
case region_list.mountian:
if((random_number >= 10)) {node_type = node_types.rock;}
else {node_type = node_types.tree;}
break;
```
but I want it to look something more akin to this:
```
case region_list.mountian:
Common = [Node1, Node2]
Uncommon = []
Rare = [Node 3]
UltraRare = [Node4, Node5]
break;
```
Once this is done, I will also need to update the "Adding New Objects to Systems" wiki page to reflect the changes
| 1.0 | Minor: Change Node Spawns to use a Weight System - Currently, the node spawning system uses raw percentile to determine what node to spawn. This makes the process of adding a new type of node into a region quite complicated, instead, I want to transition to using a weight-based system. For instance, I will designate an item as being common, uncommon, rare, or ultra-rare. I will then determine which Rarity I'm going to spawn, and then from within that Rarity choose any one of these suitable nodes.
As it stands right now, a particular case looks like this:
```
case region_list.mountian:
if((random_number >= 10)) {node_type = node_types.rock;}
else {node_type = node_types.tree;}
break;
```
but I want it to look something more akin to this:
```
case region_list.mountian:
Common = [Node1, Node2]
Uncommon = []
Rare = [Node 3]
UltraRare = [Node4, Node5]
break;
```
Once this is done, I will also need to update the "Adding New Objects to Systems" wiki page to reflect the changes
| non_test | minor change node spawns to use a weight system currently the node spawning system uses raw percentile to determine what node to spawn this makes the process of adding a new type of node into a region quite complicated instead i want to transition to using a weight based system for instance i will designate an item as being common uncommon rare or ultra rare i will then determine which rarity i m going to spawn and then from within that rarity choose any one of these suitable nodes as it stands right now a particular case looks like this case region list mountian if random number node type node types rock else node type node types tree break but i want it to look something more akin to this case region list mountian common uncommon rare ultrarare break once this is done i will also need to update the adding new objects to systems wiki page to reflect the changes | 0 |
305,479 | 26,390,743,811 | IssuesEvent | 2023-01-12 15:30:34 | CSOIreland/PxStat | https://api.github.com/repos/CSOIreland/PxStat | closed | [ENHANCEMENT] Saved queries can appear to jumble time periods compared to original due to lack of code/chronological order | enhancement released tested | **Is your feature request related to a problem? Please describe.**
Saved queries can appear to jumble time periods and list them by alphabetical label name rather than chronological on selection arrows this means that for example June of the same year can move between May and April which can be confusing if you have set for last three months. This appears in CPM11 for example. It is not as much an issue if the code and time period label are very tightly matched eg. years only and codes both the same 2022, 2021 etc. and no named months.
**Describe the solution you'd like**
Maintain the chronological order same as when queries are run and ordered by TLIST column in general if possible.
**Describe alternatives you've considered**
seems the TList code information is either not transferred or unused for saved queries
**Additional context**
In the original query order by month gives

or

in the saved query the chronological or code order is lost and labe alpha order seems to be used

or

| 1.0 | [ENHANCEMENT] Saved queries can appear to jumble time periods compared to original due to lack of code/chronological order - **Is your feature request related to a problem? Please describe.**
Saved queries can appear to jumble time periods and list them by alphabetical label name rather than chronological on selection arrows this means that for example June of the same year can move between May and April which can be confusing if you have set for last three months. This appears in CPM11 for example. It is not as much an issue if the code and time period label are very tightly matched eg. years only and codes both the same 2022, 2021 etc. and no named months.
**Describe the solution you'd like**
Maintain the chronological order same as when queries are run and ordered by TLIST column in general if possible.
**Describe alternatives you've considered**
seems the TList code information is either not transferred or unused for saved queries
**Additional context**
In the original query order by month gives

or

in the saved query the chronological or code order is lost and labe alpha order seems to be used

or

| test | saved queries can appear to jumble time periods compared to original due to lack of code chronological order is your feature request related to a problem please describe saved queries can appear to jumble time periods and list them by alphabetical label name rather than chronological on selection arrows this means that for example june of the same year can move between may and april which can be confusing if you have set for last three months this appears in for example it is not as much an issue if the code and time period label are very tightly matched eg years only and codes both the same etc and no named months describe the solution you d like maintain the chronological order same as when queries are run and ordered by tlist column in general if possible describe alternatives you ve considered seems the tlist code information is either not transferred or unused for saved queries additional context in the original query order by month gives or in the saved query the chronological or code order is lost and labe alpha order seems to be used or | 1 |
117,565 | 9,942,325,443 | IssuesEvent | 2019-07-03 13:40:31 | viszerale-therapie/vt.at-drupal | https://api.github.com/repos/viszerale-therapie/vt.at-drupal | closed | Fwd: Eigenartige Fehlermeldung Therapeutenliste | ready for test | *Sent by Andreas Schönefeldt. Created by [fire](https://fire.fundersclub.com/).*
---
> Lieber Andreas,
>
> beim Aktivieren von Therapeuten über den Link im Email kommt folgende Fehlermeldung.
>
> Auf der WebSite sieht aber auf den ersten Blick alles normal aus.
>
> Danke!
>
> Liebe Grüße, Ursula
>
>
> **Deprecated** : __autoload() is deprecated, use spl_autoload_register() instead in **/webseiten/therapeuten-liste/web/phpmailer/PHPMailerAutoload.php** on line **45**
>
>
>
> **Therapeut Daniela G aktiviert!**
| 1.0 | Fwd: Eigenartige Fehlermeldung Therapeutenliste - *Sent by Andreas Schönefeldt. Created by [fire](https://fire.fundersclub.com/).*
---
> Lieber Andreas,
>
> beim Aktivieren von Therapeuten über den Link im Email kommt folgende Fehlermeldung.
>
> Auf der WebSite sieht aber auf den ersten Blick alles normal aus.
>
> Danke!
>
> Liebe Grüße, Ursula
>
>
> **Deprecated** : __autoload() is deprecated, use spl_autoload_register() instead in **/webseiten/therapeuten-liste/web/phpmailer/PHPMailerAutoload.php** on line **45**
>
>
>
> **Therapeut Daniela G aktiviert!**
| test | fwd eigenartige fehlermeldung therapeutenliste sent by andreas schönefeldt created by lieber andreas beim aktivieren von therapeuten über den link im email kommt folgende fehlermeldung auf der website sieht aber auf den ersten blick alles normal aus danke liebe grüße ursula deprecated autoload is deprecated use spl autoload register instead in webseiten therapeuten liste web phpmailer phpmailerautoload php on line therapeut daniela g aktiviert | 1 |
97,832 | 11,033,966,155 | IssuesEvent | 2019-12-07 01:27:15 | streamlit/streamlit | https://api.github.com/repos/streamlit/streamlit | closed | installTracer option is not working | bug documentation regression | # Summary
The `installTracer` option in the Streamlit config does not do anything but slowing down the script to the point it is unusable. On top of this, there is no documentation on how to use this feature.
We should fix (or remove, if deemed obsolete) and add docuemntation.
# Steps to reproduce
Turn on `installTracer` in the Streamlit config. Run a script.
## Expected behavior:
I am expecting at least that the script runs, may with a little performance overhead.
## Actual behavior:
Script's state says "Running" forever.
## Is this a regression?
yes
# Debug info
- Streamlit version: (get it with `$ streamlit version`)
- Python version: (get it with `$ python --version`)
- Using Conda? PipEnv? PyEnv? Pex?
- OS version:
- Browser version:
# Additional information
https://discuss.streamlit.io/t/how-to-use-tracer/759/2
| 1.0 | installTracer option is not working - # Summary
The `installTracer` option in the Streamlit config does not do anything but slowing down the script to the point it is unusable. On top of this, there is no documentation on how to use this feature.
We should fix (or remove, if deemed obsolete) and add docuemntation.
# Steps to reproduce
Turn on `installTracer` in the Streamlit config. Run a script.
## Expected behavior:
I am expecting at least that the script runs, may with a little performance overhead.
## Actual behavior:
Script's state says "Running" forever.
## Is this a regression?
yes
# Debug info
- Streamlit version: (get it with `$ streamlit version`)
- Python version: (get it with `$ python --version`)
- Using Conda? PipEnv? PyEnv? Pex?
- OS version:
- Browser version:
# Additional information
https://discuss.streamlit.io/t/how-to-use-tracer/759/2
| non_test | installtracer option is not working summary the installtracer option in the streamlit config does not do anything but slowing down the script to the point it is unusable on top of this there is no documentation on how to use this feature we should fix or remove if deemed obsolete and add docuemntation steps to reproduce turn on installtracer in the streamlit config run a script expected behavior i am expecting at least that the script runs may with a little performance overhead actual behavior script s state says running forever is this a regression yes debug info streamlit version get it with streamlit version python version get it with python version using conda pipenv pyenv pex os version browser version additional information | 0 |
428,003 | 29,922,693,911 | IssuesEvent | 2023-06-22 00:49:39 | edgarrmondragon/citric | https://api.github.com/repos/edgarrmondragon/citric | opened | Publish documentation to Dash User-Contributed docsets | documentation enhancement | ### Feature Type
New RPC client method
### Description
* https://github.com/Kapeli/Dash-User-Contributions
* https://hynek.me/articles/productive-fruit-fly-programmer/ | 1.0 | Publish documentation to Dash User-Contributed docsets - ### Feature Type
New RPC client method
### Description
* https://github.com/Kapeli/Dash-User-Contributions
* https://hynek.me/articles/productive-fruit-fly-programmer/ | non_test | publish documentation to dash user contributed docsets feature type new rpc client method description | 0 |
94,710 | 8,514,093,675 | IssuesEvent | 2018-10-31 17:38:19 | Microsoft/vscode | https://api.github.com/repos/Microsoft/vscode | closed | Test multiline search | testplan-item | Test #13155
Complexity: 3
- [x] Mac @misolori
- [x] Windows @gregvanl
- [x] Linux @rebornix
Test that multiline search works and works the same as in the editor. A regex search executes in multiline mode if it has a `\n`. `\s` will match a newline only if there is a `\n` literal somewhere in the regex.
The search view should show the first line, and highlight the correct full range in the editor when clicked. | 1.0 | Test multiline search - Test #13155
Complexity: 3
- [x] Mac @misolori
- [x] Windows @gregvanl
- [x] Linux @rebornix
Test that multiline search works and works the same as in the editor. A regex search executes in multiline mode if it has a `\n`. `\s` will match a newline only if there is a `\n` literal somewhere in the regex.
The search view should show the first line, and highlight the correct full range in the editor when clicked. | test | test multiline search test complexity mac misolori windows gregvanl linux rebornix test that multiline search works and works the same as in the editor a regex search executes in multiline mode if it has a n s will match a newline only if there is a n literal somewhere in the regex the search view should show the first line and highlight the correct full range in the editor when clicked | 1 |
166,081 | 12,890,394,771 | IssuesEvent | 2020-07-13 15:56:05 | modin-project/modin | https://api.github.com/repos/modin-project/modin | opened | Test Failure on Dask Engine: Timeout | Dask :zap: Testing :chart_with_upwards_trend: dependencies :link: | The default to pandas test has been failing recently on Dask with Windows. I have searched and the errors thrown do not appear to be known issues at the moment.
Link to test failure: https://github.com/modin-project/modin/runs/862907959?check_suite_focus=true | 1.0 | Test Failure on Dask Engine: Timeout - The default to pandas test has been failing recently on Dask with Windows. I have searched and the errors thrown do not appear to be known issues at the moment.
Link to test failure: https://github.com/modin-project/modin/runs/862907959?check_suite_focus=true | test | test failure on dask engine timeout the default to pandas test has been failing recently on dask with windows i have searched and the errors thrown do not appear to be known issues at the moment link to test failure | 1 |
119,208 | 10,028,238,601 | IssuesEvent | 2019-07-17 11:09:21 | trezor/trezor-firmware | https://api.github.com/repos/trezor/trezor-firmware | opened | Shamir : device tests | Shamir tests | Add/modify device tests for shamir workflows
[ ] reset
[ ] recovery
[ ] dry run
[ ] persistence (have you tried turning it off and on again?)
[ ] warnings (identifier mismatch, share already added, invalid share)
| 1.0 | Shamir : device tests - Add/modify device tests for shamir workflows
[ ] reset
[ ] recovery
[ ] dry run
[ ] persistence (have you tried turning it off and on again?)
[ ] warnings (identifier mismatch, share already added, invalid share)
| test | shamir device tests add modify device tests for shamir workflows reset recovery dry run persistence have you tried turning it off and on again warnings identifier mismatch share already added invalid share | 1 |
232,462 | 18,876,340,598 | IssuesEvent | 2021-11-14 03:35:37 | CIS-3760-Team-25/unity-checkers | https://api.github.com/repos/CIS-3760-Team-25/unity-checkers | closed | S16. Automated Testing | story testing | As a developer I want to add the Unity Test Framework to the project and set up some basic tests so that the team can complete the DoD.
#### Acceptance Criteria
- [x] Create test assemblies for Play and/or Edit mode
- [x] Basic tests are written to verify GameObject initialization
#### Resources
https://docs.unity3d.com/Packages/com.unity.test-framework@1.1/manual/index.html
https://docs.unity3d.com/Packages/com.unity.test-framework@1.1/manual/workflow-create-test-assembly.html | 1.0 | S16. Automated Testing - As a developer I want to add the Unity Test Framework to the project and set up some basic tests so that the team can complete the DoD.
#### Acceptance Criteria
- [x] Create test assemblies for Play and/or Edit mode
- [x] Basic tests are written to verify GameObject initialization
#### Resources
https://docs.unity3d.com/Packages/com.unity.test-framework@1.1/manual/index.html
https://docs.unity3d.com/Packages/com.unity.test-framework@1.1/manual/workflow-create-test-assembly.html | test | automated testing as a developer i want to add the unity test framework to the project and set up some basic tests so that the team can complete the dod acceptance criteria create test assemblies for play and or edit mode basic tests are written to verify gameobject initialization resources | 1 |
114,400 | 17,209,452,915 | IssuesEvent | 2021-07-19 00:11:35 | turkdevops/javascript-sdk | https://api.github.com/repos/turkdevops/javascript-sdk | opened | WS-2018-0072 (High) detected in https-proxy-agent-1.0.0.tgz | security vulnerability | ## WS-2018-0072 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>https-proxy-agent-1.0.0.tgz</b></p></summary>
<p>An HTTP(s) proxy `http.Agent` implementation for HTTPS</p>
<p>Library home page: <a href="https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-1.0.0.tgz">https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-1.0.0.tgz</a></p>
<p>Path to dependency file: javascript-sdk/package.json</p>
<p>Path to vulnerable library: javascript-sdk/node_modules/https-proxy-agent/package.json</p>
<p>
Dependency Hierarchy:
- nightwatch-0.9.21.tgz (Root Library)
- proxy-agent-2.0.0.tgz
- :x: **https-proxy-agent-1.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/javascript-sdk/commit/2ed96566365ee89d8a9b1250ccd7c049281ed09c">2ed96566365ee89d8a9b1250ccd7c049281ed09c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of https-proxy-agent before 2.2.0 are vulnerable to a denial of service. This is due to unsanitized options (proxy.auth) being passed to Buffer().
<p>Publish Date: 2018-02-28
<p>URL: <a href=https://hackerone.com/reports/319532>WS-2018-0072</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nodesecurity.io/advisories/593">https://nodesecurity.io/advisories/593</a></p>
<p>Release Date: 2018-01-27</p>
<p>Fix Resolution: 2.2.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2018-0072 (High) detected in https-proxy-agent-1.0.0.tgz - ## WS-2018-0072 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>https-proxy-agent-1.0.0.tgz</b></p></summary>
<p>An HTTP(s) proxy `http.Agent` implementation for HTTPS</p>
<p>Library home page: <a href="https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-1.0.0.tgz">https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-1.0.0.tgz</a></p>
<p>Path to dependency file: javascript-sdk/package.json</p>
<p>Path to vulnerable library: javascript-sdk/node_modules/https-proxy-agent/package.json</p>
<p>
Dependency Hierarchy:
- nightwatch-0.9.21.tgz (Root Library)
- proxy-agent-2.0.0.tgz
- :x: **https-proxy-agent-1.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/javascript-sdk/commit/2ed96566365ee89d8a9b1250ccd7c049281ed09c">2ed96566365ee89d8a9b1250ccd7c049281ed09c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of https-proxy-agent before 2.2.0 are vulnerable to a denial of service. This is due to unsanitized options (proxy.auth) being passed to Buffer().
<p>Publish Date: 2018-02-28
<p>URL: <a href=https://hackerone.com/reports/319532>WS-2018-0072</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nodesecurity.io/advisories/593">https://nodesecurity.io/advisories/593</a></p>
<p>Release Date: 2018-01-27</p>
<p>Fix Resolution: 2.2.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | ws high detected in https proxy agent tgz ws high severity vulnerability vulnerable library https proxy agent tgz an http s proxy http agent implementation for https library home page a href path to dependency file javascript sdk package json path to vulnerable library javascript sdk node modules https proxy agent package json dependency hierarchy nightwatch tgz root library proxy agent tgz x https proxy agent tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions of https proxy agent before are vulnerable to a denial of service this is due to unsanitized options proxy auth being passed to buffer publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
53,728 | 6,342,835,556 | IssuesEvent | 2017-07-27 16:16:05 | katiekeel/little-shop | https://api.github.com/repos/katiekeel/little-shop | closed | Story 5: Removing an item | Controller in progress Styling Testing View | - Feature testing for cart show with item removal
- Cart controller - remove item
- Green flash notice for successful removal with link to item | 1.0 | Story 5: Removing an item - - Feature testing for cart show with item removal
- Cart controller - remove item
- Green flash notice for successful removal with link to item | test | story removing an item feature testing for cart show with item removal cart controller remove item green flash notice for successful removal with link to item | 1 |
165,153 | 26,110,666,317 | IssuesEvent | 2022-12-27 19:19:08 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | `SyntaxTree.GetChanges()` returns non-deterministic results | Area-Compilers Resolution-By Design | **Version Used**: 4.4.0
**Reproducible for**: C#, VB.NET
### Context
We were analyzing a problem reported by users, and ended up debugging and trying to understand the behaviour of `SyntaxTree.GetChanges()`. A 3rd-party-library uses the returned changes for code formatting, with sometimes completely unexpected results.
### Steps to Reproduce:
1. Run the provided sample program from [here](https://github.com/paulssn/RoslynGetChangesTest) (and see the explanation below).
2. => `SyntaxTree.GetChanges()` returns seemingly non-deterministic results
### Expected Behavior:
`SyntaxTree.GetChanges()` returns the same set of changes every time, if both trees are constructed from the same source text.
### Actual Behavior:
`SyntaxTree.GetChanges()` does not deliver consistent results on subsequent calls. Actually it seems to be not fully deterministic, as it also varies between Debug and Release configuration for example.
Consider the following sample program:
```csharp
using Microsoft.CodeAnalysis;
using Microsoft.CodeAnalysis.CSharp;
for (var i = 0; i < 10; i++)
{
CalculateChanges();
}
static void CalculateChanges()
{
var tree1 = GetTree("tree1.cs");
var tree2 = GetTree("tree2.cs");
var treeChanges = tree2.GetChanges(tree1);
Console.WriteLine($"Number of changes: {treeChanges.Count}");
}
static SyntaxTree GetTree(string file)
{
return CSharpSyntaxTree.ParseText(File.ReadAllText(file));
}
```
This prints the following for our specific example:
```
Number of changes: 950
Number of changes: 31
Number of changes: 27
Number of changes: 31
Number of changes: 31
Number of changes: 31
Number of changes: 31
Number of changes: 31
Number of changes: 720
Number of changes: 944
```
I can't share the SyntaxTree producing exactly these results, but it seems that one can reproduce similar behaviour, by either executing `GetChanges()` a lot of times (>1000), or by using large files with a lot of changes.
I tried it with a random big source file from the roslyn repository and a corresponding changed version (see attached project), and it produces the following results:
```
Number of changes: 3514
Number of changes: 9400
Number of changes: 9402
Number of changes: 25088
Number of changes: 9464
Number of changes: 9415
Number of changes: 9451
Number of changes: 22002
Number of changes: 3524
Number of changes: 9402
```
### Reason / what we know so far:
After some debugging we figured out, that the reason for this behaviour seems to be the following:
* Behind the scenes `GetChanges()` uses the `SyntaxDiffer` which is calculating the similarity of SyntaxNodes (`SyntaxDiffer.GetSimilarity()`).
* This is done by doing a reference comparison of the underlying `GreenNode`s.
* If we now inspect a comparison of specific nodes, it turns out that sometimes the nodes actually are equal and sometimes they are not.
* It seems like that's the case, because there's a global static cache for the nodes: If nodes are not re-used from the cache for some reason, they are treated as not equal, as another node is created and the comparison by reference fails.
* We didn't analyze the whole caching logic, and therefore I can't explain why a node is cached sometimes and sometimes it's not, but it seems there are several factors playing into that.
### Questions:
* Is this the intended behaviour or a bug?
* For me it feels wrong if caching (a performance optimization) would influence a feature like `GetChanges()`, ie has a functional impact.
* Is there something we missed, and the described behaviour can be explained by anything else?
* Let's assume this behaviour is intended, is there a way to get some kind of diff between SyntaxTrees or Documents in a consistent way? | 1.0 | `SyntaxTree.GetChanges()` returns non-deterministic results - **Version Used**: 4.4.0
**Reproducible for**: C#, VB.NET
### Context
We were analyzing a problem reported by users, and ended up debugging and trying to understand the behaviour of `SyntaxTree.GetChanges()`. A 3rd-party-library uses the returned changes for code formatting, with sometimes completely unexpected results.
### Steps to Reproduce:
1. Run the provided sample program from [here](https://github.com/paulssn/RoslynGetChangesTest) (and see the explanation below).
2. => `SyntaxTree.GetChanges()` returns seemingly non-deterministic results
### Expected Behavior:
`SyntaxTree.GetChanges()` returns the same set of changes every time, if both trees are constructed from the same source text.
### Actual Behavior:
`SyntaxTree.GetChanges()` does not deliver consistent results on subsequent calls. Actually it seems to be not fully deterministic, as it also varies between Debug and Release configuration for example.
Consider the following sample program:
```csharp
using Microsoft.CodeAnalysis;
using Microsoft.CodeAnalysis.CSharp;
for (var i = 0; i < 10; i++)
{
CalculateChanges();
}
static void CalculateChanges()
{
var tree1 = GetTree("tree1.cs");
var tree2 = GetTree("tree2.cs");
var treeChanges = tree2.GetChanges(tree1);
Console.WriteLine($"Number of changes: {treeChanges.Count}");
}
static SyntaxTree GetTree(string file)
{
return CSharpSyntaxTree.ParseText(File.ReadAllText(file));
}
```
This prints the following for our specific example:
```
Number of changes: 950
Number of changes: 31
Number of changes: 27
Number of changes: 31
Number of changes: 31
Number of changes: 31
Number of changes: 31
Number of changes: 31
Number of changes: 720
Number of changes: 944
```
I can't share the SyntaxTree producing exactly these results, but it seems that one can reproduce similar behaviour, by either executing `GetChanges()` a lot of times (>1000), or by using large files with a lot of changes.
I tried it with a random big source file from the roslyn repository and a corresponding changed version (see attached project), and it produces the following results:
```
Number of changes: 3514
Number of changes: 9400
Number of changes: 9402
Number of changes: 25088
Number of changes: 9464
Number of changes: 9415
Number of changes: 9451
Number of changes: 22002
Number of changes: 3524
Number of changes: 9402
```
### Reason / what we know so far:
After some debugging we figured out, that the reason for this behaviour seems to be the following:
* Behind the scenes `GetChanges()` uses the `SyntaxDiffer` which is calculating the similarity of SyntaxNodes (`SyntaxDiffer.GetSimilarity()`).
* This is done by doing a reference comparison of the underlying `GreenNode`s.
* If we now inspect a comparison of specific nodes, it turns out that sometimes the nodes actually are equal and sometimes they are not.
* It seems like that's the case, because there's a global static cache for the nodes: If nodes are not re-used from the cache for some reason, they are treated as not equal, as another node is created and the comparison by reference fails.
* We didn't analyze the whole caching logic, and therefore I can't explain why a node is cached sometimes and sometimes it's not, but it seems there are several factors playing into that.
### Questions:
* Is this the intended behaviour or a bug?
* For me it feels wrong if caching (a performance optimization) would influence a feature like `GetChanges()`, ie has a functional impact.
* Is there something we missed, and the described behaviour can be explained by anything else?
* Let's assume this behaviour is intended, is there a way to get some kind of diff between SyntaxTrees or Documents in a consistent way? | non_test | syntaxtree getchanges returns non deterministic results version used reproducible for c vb net context we were analyzing a problem reported by users and ended up debugging and trying to understand the behaviour of syntaxtree getchanges a party library uses the returned changes for code formatting with sometimes completely unexpected results steps to reproduce run the provided sample program from and see the explanation below syntaxtree getchanges returns seemingly non deterministic results expected behavior syntaxtree getchanges returns the same set of changes every time if both trees are constructed from the same source text actual behavior syntaxtree getchanges does not deliver consistent results on subsequent calls actually it seems to be not fully deterministic as it also varies between debug and release configuration for example consider the following sample program csharp using microsoft codeanalysis using microsoft codeanalysis csharp for var i i i calculatechanges static void calculatechanges var gettree cs var gettree cs var treechanges getchanges console writeline number of changes treechanges count static syntaxtree gettree string file return csharpsyntaxtree parsetext file readalltext file this prints the following for our specific example number of changes number of changes number of changes number of changes number of changes number of changes number of changes number of changes number of changes number of changes i can t share the syntaxtree producing exactly these results but it seems that one can reproduce similar behaviour by either executing getchanges a lot of times or by using large files with a lot of changes i tried it with a random big source file from the roslyn repository and a corresponding changed version see attached project and it produces the following results number of changes number of changes number of changes number of changes number of changes number of changes number of changes number of changes number of changes number of changes reason what we know so far after some debugging we figured out that the reason for this behaviour seems to be the following behind the scenes getchanges uses the syntaxdiffer which is calculating the similarity of syntaxnodes syntaxdiffer getsimilarity this is done by doing a reference comparison of the underlying greennode s if we now inspect a comparison of specific nodes it turns out that sometimes the nodes actually are equal and sometimes they are not it seems like that s the case because there s a global static cache for the nodes if nodes are not re used from the cache for some reason they are treated as not equal as another node is created and the comparison by reference fails we didn t analyze the whole caching logic and therefore i can t explain why a node is cached sometimes and sometimes it s not but it seems there are several factors playing into that questions is this the intended behaviour or a bug for me it feels wrong if caching a performance optimization would influence a feature like getchanges ie has a functional impact is there something we missed and the described behaviour can be explained by anything else let s assume this behaviour is intended is there a way to get some kind of diff between syntaxtrees or documents in a consistent way | 0 |
97,330 | 8,652,746,460 | IssuesEvent | 2018-11-27 09:01:25 | Microsoft/azure-pipelines-tasks | https://api.github.com/repos/Microsoft/azure-pipelines-tasks | closed | PublishTestResultsV2: Test result completed date must be greater than test result start date | Area: Test | [ScalaTest](http://www.scalatest.org) for some reason (bug or intentionally) sets time for ignored tests to **-0.001**.
```xml
<testcase classname="com.foo.bar" name="Foo bar should foo bar" time="-0.001">
<skipped/>
</testcase>
```
This prevents the report to be published, even if only one test was skipped and had negative time.
```
2018-09-12T08:32:52.8780751Z ##[warning]Failed to publish test results: Test result completed date must be greater than test result start date
```
Time for skipped tests doesn't have any meaning and could be ignored.
Sample ScalaTest code:
```scala
ignore should "foo bar" in {
???
}
```
Related issue: https://github.com/scalatest/scalatest/issues/1413 | 1.0 | PublishTestResultsV2: Test result completed date must be greater than test result start date - [ScalaTest](http://www.scalatest.org) for some reason (bug or intentionally) sets time for ignored tests to **-0.001**.
```xml
<testcase classname="com.foo.bar" name="Foo bar should foo bar" time="-0.001">
<skipped/>
</testcase>
```
This prevents the report to be published, even if only one test was skipped and had negative time.
```
2018-09-12T08:32:52.8780751Z ##[warning]Failed to publish test results: Test result completed date must be greater than test result start date
```
Time for skipped tests doesn't have any meaning and could be ignored.
Sample ScalaTest code:
```scala
ignore should "foo bar" in {
???
}
```
Related issue: https://github.com/scalatest/scalatest/issues/1413 | test | test result completed date must be greater than test result start date for some reason bug or intentionally sets time for ignored tests to xml this prevents the report to be published even if only one test was skipped and had negative time failed to publish test results test result completed date must be greater than test result start date time for skipped tests doesn t have any meaning and could be ignored sample scalatest code scala ignore should foo bar in related issue | 1 |
16,442 | 3,521,620,854 | IssuesEvent | 2016-01-13 03:17:08 | backdrop/backdrop-issues | https://api.github.com/repos/backdrop/backdrop-issues | closed | [UX] Collapse the input format options fieldset on node edit forms | pr - reviewed & tested by the community status - has pull request type - feature request | There are a whole bunch of modules on D.org dedicated to hiding this. Why, oh why, is it not collapsed by default?
Adding this to filter_process_format() in filter.module seems to do the trick (minus CSS adjustments):
```
'#title' => 'Formatting options',
'#collapsible' => TRUE,
'#collapsed' => TRUE,
```
Collapsed:

Expanded:

| 1.0 | [UX] Collapse the input format options fieldset on node edit forms - There are a whole bunch of modules on D.org dedicated to hiding this. Why, oh why, is it not collapsed by default?
Adding this to filter_process_format() in filter.module seems to do the trick (minus CSS adjustments):
```
'#title' => 'Formatting options',
'#collapsible' => TRUE,
'#collapsed' => TRUE,
```
Collapsed:

Expanded:

| test | collapse the input format options fieldset on node edit forms there are a whole bunch of modules on d org dedicated to hiding this why oh why is it not collapsed by default adding this to filter process format in filter module seems to do the trick minus css adjustments title formatting options collapsible true collapsed true collapsed expanded | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.