Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
321,295
| 27,520,368,428
|
IssuesEvent
|
2023-03-06 14:40:48
|
wazuh/wazuh-qa
|
https://api.github.com/repos/wazuh/wazuh-qa
|
closed
|
Add Cisco Umbrella integration tests
|
team/framework test/integration type/test-development role/qa-deprecaped feature/aws
|
As part of https://github.com/wazuh/wazuh-qa/issues/3333 we need to Implement the integration test cases related to `Cisco Umbrella` defined in https://github.com/wazuh/wazuh-qa/issues/3334.
# Test cases list
## Tier 0
- <details><summary>Umbrella integration works properly using the default configuration</summary>
### Pre-conditions
* There are already configured credentials for the `qa` profile.
* Debug mode is enabled in the internal_configuration file.
### Cases
<table>
<tr>
<td>Cases</td>
<td>Expected Result</td>
<td>Configuration</td>
</tr>
<tr>
<td>Use an existing bucket</td>
<td>The module is invoked with the expected parameters and no error occurs</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
</bucket>
</wodle>
```
</details>
</td>
</tr>
</table>
### Steps
1. Apply the test case configuration to the `ossec.conf`
1. Restart wazuh-manager service
1. Monitor the `ossec.log` waiting until the wazuh service is restarted or there is a timeout
1. Monitor the `ossec.log` waiting until the module is triggered or there is a timeout
1. Check the command was called with the correct parameters
1. Check no error messages are present in the `ossec.log`
</details>
- <details><summary>Umbrella integration works properly with remove_from_bucket</summary>
### Pre-conditions
* There are already configured credentials for the `qa` profile.
* Debug mode is enabled in the internal_configuration file.
### Cases
<table>
<tr>
<td>Cases</td>
<td>Expected Result</td>
<td>Configuration</td>
</tr>
<tr>
<td>Use an existing bucket, upload a file and consume it</td>
<td>The uploaded file was removed after the execution</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<remove_from_bucket>yes</remove_from_bucket>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
</bucket>
</wodle>
```
</details>
</td>
</tr>
</table>
### Steps
1. Apply the test case configuration to the `ossec.conf`
1. Restart wazuh-manager service
1. Monitor the `ossec.log` waiting until the wazuh service is restarted or there is a timeout
1. Monitor the `ossec.log` waiting until the module is triggered or there is a timeout
1. Check the command was called with the correct parameters
1. Make sure that the uploaded log was removed by the module after the execution
</details>
- <details><summary>Umbrella integration works properly with only_logs_after option</summary>
### Pre-conditions
* There are already configured credentials for the `qa` profile.
* Debug mode is enabled in the internal_configuration file.
* There are already logs with timestamps older than the date specified in the configuration.
* There are already logs with timestamps more recent timestamp than the date specified in the configuration.
### Cases
<table>
<tr>
<td>Cases</td>
<td>Expected Result</td>
<td>Configuration</td>
</tr>
<tr>
<td>Don't set a date in the configuration</td>
<td>Only the log uploaded during execution is processed</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
</bucket>
</wodle>
```
</details>
</td>
</tr>
<tr>
<td>Set a date for which there are logs available in the bucket</td>
<td>All logs with a timestamp greater than the only_logs_after value are processed</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
<only_logs_after>date-to-test</only_logs_after>
</bucket>
</wodle>
```
</details>
</td>
</tr>
</table>
### Steps
1. Delete the `s3_cloudtrail.db` file using a fixture
1. Apply the test case configuration to the `ossec.conf`
1. Upload a log file to the test bucket for the day of the test execution
1. Restart wazuh-manager service
1. Monitor the `ossec.log` waiting until the wazuh service is restarted or there is a timeout
1. Monitor the `ossec.log` waiting until the module is triggered or there is a timeout
1. Check the command was called with the correct parameters
1. Check the expected number of events were sent to analysisd. Only the logs whose timestamp is greater than the date specified in the configuration should be processed
1. Check the database was created and updated accordingly
</details>
- <details><summary>Umbrella integration works properly with path option</summary>
### Pre-conditions
* There are already configured credentials for the `qa` profile.
* Debug mode is enabled in the internal_configuration file.
* There are already logs generated after the date specified in the configuration inside of *test_prefix*.
### Cases
<table>
<tr>
<td>Cases</td>
<td>Expected Result</td>
<td>Configuration</td>
</tr>
<tr>
<td>Use an existing path with data</td>
<td>Only the contents of *test_prefix* are processed and sent to analysisd</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
<only_logs_after>date-to-test</only_logs_after>
<path>test_prefix</path>
</bucket>
</wodle>
```
</details>
</td>
</tr>
<tr>
<td>Use an existing path without data</td>
<td>No logs were processed. No logs were sent analysisd</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
<only_logs_after>date-to-test</only_logs_after>
<path>empty_prefix</path>
</bucket>
</wodle>
```
</details>
</td>
</tr>
<tr>
<td>Use an inexisting path</td>
<td>No logs were processed. No logs were sent analysisd. A message warning that the path does not exist should be displayed</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
<only_logs_after>date-to-test</only_logs_after>
<path>inexistent_prefix</path>
</bucket>
</wodle>
```
</details>
</td>
</tr>
</table>
### Steps
1. Delete the `s3_cloudtrail.db` file using a fixture
1. Apply the test case configuration to the `ossec.conf`
1. Restart wazuh-manager service
1. Monitor the `ossec.log` waiting until the wazuh service is restarted or there is a timeout
1. Monitor the `ossec.log` waiting until the module is triggered or there is a timeout
1. Check the command was called with the correct parameters
1. If a path that does not exist was specified, make sure that a message is displayed in the `ossec.log` warning the user
1. Check the expected number of events were forwarded to analysisd
1. Check the database was created and updated accordingly, using the correct path for each entry
</details>
- <details><summary>Umbrella integration works properly with discard_regex option</summary>
### Pre-conditions
* There are already configured credentials for the `qa` profile.
* Debug mode is enabled in the internal_configuration file.
* There are already logs generated after the date specified in the configuration.
### Cases
<table>
<tr>
<td>Cases</td>
<td>Expected Result</td>
<td>Configuration</td>
</tr>
<tr>
<td>Set a regex to one field</td>
<td>Fetch logs excluding the ones that match with the regex</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
<only_logs_after>date-to-test</only_logs_after>
<discard_regex field="field.to.watch">REGEX_TO_APPLY</discard_regex>
</bucket>
</wodle>
```
</details>
</td>
</tr>
</table>
### Steps
1. Delete the `s3_cloudtrail.db` file using a fixture
1. Apply the test case configuration to the `ossec.conf`
1. Restart wazuh-manager service
1. Monitor the `ossec.log` waiting until the wazuh service is restarted or there is a timeout
1. Monitor the `ossec.log` waiting until the module is triggered or there is a timeout
1. Check the command was called with the correct parameters
1. Check the expected number of events were forwarded to analysisd, only logs stored in the bucket and skips the ones that match with regex
1. Check the database was created and updated accordingly
</details>
## Tier 1
- <details><summary>Umbrella integration don't fetch logs that already were processed </summary>
### Pre-conditions
* There are already configured credentials for the `qa` profile.
* There are already logs generated after the date specified in the configurations.
### Cases
|Case|Expected Result|
|----|-----------|
|Call the AWS module multiple times with different `only_logs_after` values|No duplicates were processed|
### Steps
1. Delete the `s3_cloudtrail.db` file using a fixture
1. Call the module without `only_logs_after` and check that no logs were processed:
```bash
/var/ossec/wodles/aws/aws-s3 --bucket umbrella-tests \
--type cisco_umbrella \
--aws_profile qa \
--debug 2
```
1. Call the module with `only_logs_after` set in the past and check that the expected number of logs were processed:
```bash
/var/ossec/wodles/aws/aws-s3 --bucket umbrella-tests \
--type cisco_umbrella \
--only_logs_after <date-to-test> \
--aws_profile qa \
--debug 2
```
1. Call the module with the same parameters in `3` and check there were no duplicates
1. Call the module with `only_logs_after` set with an early date than setted in `3` and check that no logs were processed, there were no duplicates:
```bash
/var/ossec/wodles/aws/aws-s3 --bucket umbrella-tests \
--type cisco_umbrella \
--only_logs_after <early-date-than-step-6> \
--aws_profile qa \
--debug 2
```
1. Upload a log file to the test bucket for the day of the test execution
1. Call the module without `only_logs_after` and check that only the uploaded logs were processed and the last marker is specified in the DB.
```bash
/var/ossec/wodles/aws/aws-s3 --bucket umbrella-tests \
--type cisco_umbrella \
--aws_profile qa \
--debug 2
```
1. Delete the uploaded log file. This must be done even if the test fails
</details>
|
2.0
|
Add Cisco Umbrella integration tests - As part of https://github.com/wazuh/wazuh-qa/issues/3333 we need to Implement the integration test cases related to `Cisco Umbrella` defined in https://github.com/wazuh/wazuh-qa/issues/3334.
# Test cases list
## Tier 0
- <details><summary>Umbrella integration works properly using the default configuration</summary>
### Pre-conditions
* There are already configured credentials for the `qa` profile.
* Debug mode is enabled in the internal_configuration file.
### Cases
<table>
<tr>
<td>Cases</td>
<td>Expected Result</td>
<td>Configuration</td>
</tr>
<tr>
<td>Use an existing bucket</td>
<td>The module is invoked with the expected parameters and no error occurs</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
</bucket>
</wodle>
```
</details>
</td>
</tr>
</table>
### Steps
1. Apply the test case configuration to the `ossec.conf`
1. Restart wazuh-manager service
1. Monitor the `ossec.log` waiting until the wazuh service is restarted or there is a timeout
1. Monitor the `ossec.log` waiting until the module is triggered or there is a timeout
1. Check the command was called with the correct parameters
1. Check no error messages are present in the `ossec.log`
</details>
- <details><summary>Umbrella integration works properly with remove_from_bucket</summary>
### Pre-conditions
* There are already configured credentials for the `qa` profile.
* Debug mode is enabled in the internal_configuration file.
### Cases
<table>
<tr>
<td>Cases</td>
<td>Expected Result</td>
<td>Configuration</td>
</tr>
<tr>
<td>Use an existing bucket, upload a file and consume it</td>
<td>The uploaded file was removed after the execution</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<remove_from_bucket>yes</remove_from_bucket>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
</bucket>
</wodle>
```
</details>
</td>
</tr>
</table>
### Steps
1. Apply the test case configuration to the `ossec.conf`
1. Restart wazuh-manager service
1. Monitor the `ossec.log` waiting until the wazuh service is restarted or there is a timeout
1. Monitor the `ossec.log` waiting until the module is triggered or there is a timeout
1. Check the command was called with the correct parameters
1. Make sure that the uploaded log was removed by the module after the execution
</details>
- <details><summary>Umbrella integration works properly with only_logs_after option</summary>
### Pre-conditions
* There are already configured credentials for the `qa` profile.
* Debug mode is enabled in the internal_configuration file.
* There are already logs with timestamps older than the date specified in the configuration.
* There are already logs with timestamps more recent timestamp than the date specified in the configuration.
### Cases
<table>
<tr>
<td>Cases</td>
<td>Expected Result</td>
<td>Configuration</td>
</tr>
<tr>
<td>Don't set a date in the configuration</td>
<td>Only the log uploaded during execution is processed</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
</bucket>
</wodle>
```
</details>
</td>
</tr>
<tr>
<td>Set a date for which there are logs available in the bucket</td>
<td>All logs with a timestamp greater than the only_logs_after value are processed</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
<only_logs_after>date-to-test</only_logs_after>
</bucket>
</wodle>
```
</details>
</td>
</tr>
</table>
### Steps
1. Delete the `s3_cloudtrail.db` file using a fixture
1. Apply the test case configuration to the `ossec.conf`
1. Upload a log file to the test bucket for the day of the test execution
1. Restart wazuh-manager service
1. Monitor the `ossec.log` waiting until the wazuh service is restarted or there is a timeout
1. Monitor the `ossec.log` waiting until the module is triggered or there is a timeout
1. Check the command was called with the correct parameters
1. Check the expected number of events were sent to analysisd. Only the logs whose timestamp is greater than the date specified in the configuration should be processed
1. Check the database was created and updated accordingly
</details>
- <details><summary>Umbrella integration works properly with path option</summary>
### Pre-conditions
* There are already configured credentials for the `qa` profile.
* Debug mode is enabled in the internal_configuration file.
* There are already logs generated after the date specified in the configuration inside of *test_prefix*.
### Cases
<table>
<tr>
<td>Cases</td>
<td>Expected Result</td>
<td>Configuration</td>
</tr>
<tr>
<td>Use an existing path with data</td>
<td>Only the contents of *test_prefix* are processed and sent to analysisd</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
<only_logs_after>date-to-test</only_logs_after>
<path>test_prefix</path>
</bucket>
</wodle>
```
</details>
</td>
</tr>
<tr>
<td>Use an existing path without data</td>
<td>No logs were processed. No logs were sent analysisd</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
<only_logs_after>date-to-test</only_logs_after>
<path>empty_prefix</path>
</bucket>
</wodle>
```
</details>
</td>
</tr>
<tr>
<td>Use an inexisting path</td>
<td>No logs were processed. No logs were sent analysisd. A message warning that the path does not exist should be displayed</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
<only_logs_after>date-to-test</only_logs_after>
<path>inexistent_prefix</path>
</bucket>
</wodle>
```
</details>
</td>
</tr>
</table>
### Steps
1. Delete the `s3_cloudtrail.db` file using a fixture
1. Apply the test case configuration to the `ossec.conf`
1. Restart wazuh-manager service
1. Monitor the `ossec.log` waiting until the wazuh service is restarted or there is a timeout
1. Monitor the `ossec.log` waiting until the module is triggered or there is a timeout
1. Check the command was called with the correct parameters
1. If a path that does not exist was specified, make sure that a message is displayed in the `ossec.log` warning the user
1. Check the expected number of events were forwarded to analysisd
1. Check the database was created and updated accordingly, using the correct path for each entry
</details>
- <details><summary>Umbrella integration works properly with discard_regex option</summary>
### Pre-conditions
* There are already configured credentials for the `qa` profile.
* Debug mode is enabled in the internal_configuration file.
* There are already logs generated after the date specified in the configuration.
### Cases
<table>
<tr>
<td>Cases</td>
<td>Expected Result</td>
<td>Configuration</td>
</tr>
<tr>
<td>Set a regex to one field</td>
<td>Fetch logs excluding the ones that match with the regex</td>
<td>
<details><summary>ossec.conf</summary>
```XML
<wodle name="aws-s3">
<disabled>no</disabled>
<bucket type="cisco_umbrella">
<aws_profile>qa</aws_profile>
<name>umbrella-tests</name>
<only_logs_after>date-to-test</only_logs_after>
<discard_regex field="field.to.watch">REGEX_TO_APPLY</discard_regex>
</bucket>
</wodle>
```
</details>
</td>
</tr>
</table>
### Steps
1. Delete the `s3_cloudtrail.db` file using a fixture
1. Apply the test case configuration to the `ossec.conf`
1. Restart wazuh-manager service
1. Monitor the `ossec.log` waiting until the wazuh service is restarted or there is a timeout
1. Monitor the `ossec.log` waiting until the module is triggered or there is a timeout
1. Check the command was called with the correct parameters
1. Check the expected number of events were forwarded to analysisd, only logs stored in the bucket and skips the ones that match with regex
1. Check the database was created and updated accordingly
</details>
## Tier 1
- <details><summary>Umbrella integration don't fetch logs that already were processed </summary>
### Pre-conditions
* There are already configured credentials for the `qa` profile.
* There are already logs generated after the date specified in the configurations.
### Cases
|Case|Expected Result|
|----|-----------|
|Call the AWS module multiple times with different `only_logs_after` values|No duplicates were processed|
### Steps
1. Delete the `s3_cloudtrail.db` file using a fixture
1. Call the module without `only_logs_after` and check that no logs were processed:
```bash
/var/ossec/wodles/aws/aws-s3 --bucket umbrella-tests \
--type cisco_umbrella \
--aws_profile qa \
--debug 2
```
1. Call the module with `only_logs_after` set in the past and check that the expected number of logs were processed:
```bash
/var/ossec/wodles/aws/aws-s3 --bucket umbrella-tests \
--type cisco_umbrella \
--only_logs_after <date-to-test> \
--aws_profile qa \
--debug 2
```
1. Call the module with the same parameters in `3` and check there were no duplicates
1. Call the module with `only_logs_after` set with an early date than setted in `3` and check that no logs were processed, there were no duplicates:
```bash
/var/ossec/wodles/aws/aws-s3 --bucket umbrella-tests \
--type cisco_umbrella \
--only_logs_after <early-date-than-step-6> \
--aws_profile qa \
--debug 2
```
1. Upload a log file to the test bucket for the day of the test execution
1. Call the module without `only_logs_after` and check that only the uploaded logs were processed and the last marker is specified in the DB.
```bash
/var/ossec/wodles/aws/aws-s3 --bucket umbrella-tests \
--type cisco_umbrella \
--aws_profile qa \
--debug 2
```
1. Delete the uploaded log file. This must be done even if the test fails
</details>
|
test
|
add cisco umbrella integration tests as part of we need to implement the integration test cases related to cisco umbrella defined in test cases list tier umbrella integration works properly using the default configuration pre conditions there are already configured credentials for the qa profile debug mode is enabled in the internal configuration file cases cases expected result configuration use an existing bucket the module is invoked with the expected parameters and no error occurs ossec conf xml no qa umbrella tests steps apply the test case configuration to the ossec conf restart wazuh manager service monitor the ossec log waiting until the wazuh service is restarted or there is a timeout monitor the ossec log waiting until the module is triggered or there is a timeout check the command was called with the correct parameters check no error messages are present in the ossec log umbrella integration works properly with remove from bucket pre conditions there are already configured credentials for the qa profile debug mode is enabled in the internal configuration file cases cases expected result configuration use an existing bucket upload a file and consume it the uploaded file was removed after the execution ossec conf xml no yes qa umbrella tests steps apply the test case configuration to the ossec conf restart wazuh manager service monitor the ossec log waiting until the wazuh service is restarted or there is a timeout monitor the ossec log waiting until the module is triggered or there is a timeout check the command was called with the correct parameters make sure that the uploaded log was removed by the module after the execution umbrella integration works properly with only logs after option pre conditions there are already configured credentials for the qa profile debug mode is enabled in the internal configuration file there are already logs with timestamps older than the date specified in the configuration there are already logs with timestamps more recent timestamp than the date specified in the configuration cases cases expected result configuration don t set a date in the configuration only the log uploaded during execution is processed ossec conf xml no qa umbrella tests set a date for which there are logs available in the bucket all logs with a timestamp greater than the only logs after value are processed ossec conf xml no qa umbrella tests date to test steps delete the cloudtrail db file using a fixture apply the test case configuration to the ossec conf upload a log file to the test bucket for the day of the test execution restart wazuh manager service monitor the ossec log waiting until the wazuh service is restarted or there is a timeout monitor the ossec log waiting until the module is triggered or there is a timeout check the command was called with the correct parameters check the expected number of events were sent to analysisd only the logs whose timestamp is greater than the date specified in the configuration should be processed check the database was created and updated accordingly umbrella integration works properly with path option pre conditions there are already configured credentials for the qa profile debug mode is enabled in the internal configuration file there are already logs generated after the date specified in the configuration inside of test prefix cases cases expected result configuration use an existing path with data only the contents of test prefix are processed and sent to analysisd ossec conf xml no qa umbrella tests date to test test prefix use an existing path without data no logs were processed no logs were sent analysisd ossec conf xml no qa umbrella tests date to test empty prefix use an inexisting path no logs were processed no logs were sent analysisd a message warning that the path does not exist should be displayed ossec conf xml no qa umbrella tests date to test inexistent prefix steps delete the cloudtrail db file using a fixture apply the test case configuration to the ossec conf restart wazuh manager service monitor the ossec log waiting until the wazuh service is restarted or there is a timeout monitor the ossec log waiting until the module is triggered or there is a timeout check the command was called with the correct parameters if a path that does not exist was specified make sure that a message is displayed in the ossec log warning the user check the expected number of events were forwarded to analysisd check the database was created and updated accordingly using the correct path for each entry umbrella integration works properly with discard regex option pre conditions there are already configured credentials for the qa profile debug mode is enabled in the internal configuration file there are already logs generated after the date specified in the configuration cases cases expected result configuration set a regex to one field fetch logs excluding the ones that match with the regex ossec conf xml no qa umbrella tests date to test regex to apply steps delete the cloudtrail db file using a fixture apply the test case configuration to the ossec conf restart wazuh manager service monitor the ossec log waiting until the wazuh service is restarted or there is a timeout monitor the ossec log waiting until the module is triggered or there is a timeout check the command was called with the correct parameters check the expected number of events were forwarded to analysisd only logs stored in the bucket and skips the ones that match with regex check the database was created and updated accordingly tier umbrella integration don t fetch logs that already were processed pre conditions there are already configured credentials for the qa profile there are already logs generated after the date specified in the configurations cases case expected result call the aws module multiple times with different only logs after values no duplicates were processed steps delete the cloudtrail db file using a fixture call the module without only logs after and check that no logs were processed bash var ossec wodles aws aws bucket umbrella tests type cisco umbrella aws profile qa debug call the module with only logs after set in the past and check that the expected number of logs were processed bash var ossec wodles aws aws bucket umbrella tests type cisco umbrella only logs after aws profile qa debug call the module with the same parameters in and check there were no duplicates call the module with only logs after set with an early date than setted in and check that no logs were processed there were no duplicates bash var ossec wodles aws aws bucket umbrella tests type cisco umbrella only logs after aws profile qa debug upload a log file to the test bucket for the day of the test execution call the module without only logs after and check that only the uploaded logs were processed and the last marker is specified in the db bash var ossec wodles aws aws bucket umbrella tests type cisco umbrella aws profile qa debug delete the uploaded log file this must be done even if the test fails
| 1
|
205,654
| 15,652,380,394
|
IssuesEvent
|
2021-03-23 11:19:29
|
elastic/elasticsearch-net
|
https://api.github.com/repos/elastic/elasticsearch-net
|
closed
|
BulkApiTests sometimes result in process_cluster_event_timeout_exception from server
|
Flakey test
|
Per #5288 - Error below found on three tests:
- ReturnsExpectedResponse
- ReturnsExpectedIsValid
- ReturnsExpectedStatusCode
Expected response.IsValid to be true because Failed to set up pipeline named 'pipeline' required for bulk Invalid NEST response built from a unsuccessful (503) low level call on PUT: /_ingest/pipeline/pipeline?pretty=true&error_trace=true
Audit trail of this API call:
[1] BadResponse: Node: http://localhost:9200/ Took: 00:00:30.0051908
OriginalException: Elasticsearch.Net.ElasticsearchClientException: Request failed to execute. Call: Status code 503 from: PUT /_ingest/pipeline/pipeline?pretty=true&error_trace=true. ServerError: Type: process_cluster_event_timeout_exception Reason: "failed to process cluster event (put-pipeline-pipeline) within 30s"
Request:
{"processors":[{"set":{"field":"description","value":"Overridden"}}]}
Response:
{
"error" : {
"root_cause" : [
{
"type" : "process_cluster_event_timeout_exception",
"reason" : "failed to process cluster event (put-pipeline-pipeline) within 30s",
"stack_trace" : "ProcessClusterEventTimeoutException[failed to process cluster event (put-pipeline-pipeline) within 30s]\r\n\tat org.elasticsearch.cluster.service.MasterService$Batcher.lambda$onTimeout(MasterService.java:134)\r\n\tat java.base/java.util.ArrayList.forEach(ArrayList.java:1540)\r\n\tat org.elasticsearch.cluster.service.MasterService$Batcher.lambda$onTimeout$1(MasterService.java:133)\r\n\tat org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:688)\r\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\r\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\r\n\tat java.base/java.lang.Thread.run(Thread.java:835)\r\n"
}
],
"type" : "process_cluster_event_timeout_exception",
"reason" : "failed to process cluster event (put-pipeline-pipeline) within 30s",
"stack_trace" : "ProcessClusterEventTimeoutException[failed to process cluster event (put-pipeline-pipeline) within 30s]\r\n\tat org.elasticsearch.cluster.service.MasterService$Batcher.lambda$onTimeout(MasterService.java:134)\r\n\tat java.base/java.util.ArrayList.forEach(ArrayList.java:1540)\r\n\tat org.elasticsearch.cluster.service.MasterService$Batcher.lambda$onTimeout$1(MasterService.java:133)\r\n\tat org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:688)\r\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\r\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\r\n\tat java.base/java.lang.Thread.run(Thread.java:835)\r\n"
},
"status" : 503
}
TCP states:
Established: 51
TimeWait: 4
ThreadPool statistics:
Worker:
Busy: 5
Free: 32762
Min: 2
Max: 32767
IOCP:
Busy: 1
Free: 999
Min: 2
Max: 1000
, but found False.
|
1.0
|
BulkApiTests sometimes result in process_cluster_event_timeout_exception from server - Per #5288 - Error below found on three tests:
- ReturnsExpectedResponse
- ReturnsExpectedIsValid
- ReturnsExpectedStatusCode
Expected response.IsValid to be true because Failed to set up pipeline named 'pipeline' required for bulk Invalid NEST response built from a unsuccessful (503) low level call on PUT: /_ingest/pipeline/pipeline?pretty=true&error_trace=true
Audit trail of this API call:
[1] BadResponse: Node: http://localhost:9200/ Took: 00:00:30.0051908
OriginalException: Elasticsearch.Net.ElasticsearchClientException: Request failed to execute. Call: Status code 503 from: PUT /_ingest/pipeline/pipeline?pretty=true&error_trace=true. ServerError: Type: process_cluster_event_timeout_exception Reason: "failed to process cluster event (put-pipeline-pipeline) within 30s"
Request:
{"processors":[{"set":{"field":"description","value":"Overridden"}}]}
Response:
{
"error" : {
"root_cause" : [
{
"type" : "process_cluster_event_timeout_exception",
"reason" : "failed to process cluster event (put-pipeline-pipeline) within 30s",
"stack_trace" : "ProcessClusterEventTimeoutException[failed to process cluster event (put-pipeline-pipeline) within 30s]\r\n\tat org.elasticsearch.cluster.service.MasterService$Batcher.lambda$onTimeout(MasterService.java:134)\r\n\tat java.base/java.util.ArrayList.forEach(ArrayList.java:1540)\r\n\tat org.elasticsearch.cluster.service.MasterService$Batcher.lambda$onTimeout$1(MasterService.java:133)\r\n\tat org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:688)\r\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\r\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\r\n\tat java.base/java.lang.Thread.run(Thread.java:835)\r\n"
}
],
"type" : "process_cluster_event_timeout_exception",
"reason" : "failed to process cluster event (put-pipeline-pipeline) within 30s",
"stack_trace" : "ProcessClusterEventTimeoutException[failed to process cluster event (put-pipeline-pipeline) within 30s]\r\n\tat org.elasticsearch.cluster.service.MasterService$Batcher.lambda$onTimeout(MasterService.java:134)\r\n\tat java.base/java.util.ArrayList.forEach(ArrayList.java:1540)\r\n\tat org.elasticsearch.cluster.service.MasterService$Batcher.lambda$onTimeout$1(MasterService.java:133)\r\n\tat org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:688)\r\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)\r\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)\r\n\tat java.base/java.lang.Thread.run(Thread.java:835)\r\n"
},
"status" : 503
}
TCP states:
Established: 51
TimeWait: 4
ThreadPool statistics:
Worker:
Busy: 5
Free: 32762
Min: 2
Max: 32767
IOCP:
Busy: 1
Free: 999
Min: 2
Max: 1000
, but found False.
|
test
|
bulkapitests sometimes result in process cluster event timeout exception from server per error below found on three tests returnsexpectedresponse returnsexpectedisvalid returnsexpectedstatuscode expected response isvalid to be true because failed to set up pipeline named pipeline required for bulk invalid nest response built from a unsuccessful low level call on put ingest pipeline pipeline pretty true error trace true audit trail of this api call badresponse node took originalexception elasticsearch net elasticsearchclientexception request failed to execute call status code from put ingest pipeline pipeline pretty true error trace true servererror type process cluster event timeout exception reason failed to process cluster event put pipeline pipeline within request processors response error root cause type process cluster event timeout exception reason failed to process cluster event put pipeline pipeline within stack trace processclustereventtimeoutexception r n tat org elasticsearch cluster service masterservice batcher lambda ontimeout masterservice java r n tat java base java util arraylist foreach arraylist java r n tat org elasticsearch cluster service masterservice batcher lambda ontimeout masterservice java r n tat org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java r n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java r n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java r n tat java base java lang thread run thread java r n type process cluster event timeout exception reason failed to process cluster event put pipeline pipeline within stack trace processclustereventtimeoutexception r n tat org elasticsearch cluster service masterservice batcher lambda ontimeout masterservice java r n tat java base java util arraylist foreach arraylist java r n tat org elasticsearch cluster service masterservice batcher lambda ontimeout masterservice java r n tat org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java r n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java r n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java r n tat java base java lang thread run thread java r n status tcp states established timewait threadpool statistics worker busy free min max iocp busy free min max but found false
| 1
|
346,062
| 30,863,931,978
|
IssuesEvent
|
2023-08-03 06:34:28
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: network_logging failed
|
C-test-failure O-robot O-roachtest branch-master release-blocker T-observability-inf
|
roachtest.network_logging [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/11169499?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/11169499?buildTab=artifacts#/network_logging) on master @ [4fe2a80d81c6fc5a3da3c7c44c5fc38da67e0367](https://github.com/cockroachdb/cockroach/commits/4fe2a80d81c6fc5a3da3c7c44c5fc38da67e0367):
```
(monitor.go:137).Wait: monitor failure: monitor task failed: output in run_060402.371453719_n4_cockroach-workload-r: ./cockroach workload run kv --concurrency=32 --duration=1h 'postgres://root@10.142.1.19:26257?sslcert=certs%2Fclient.root.crt&sslkey=certs%2Fclient.root.key&sslmode=verify-full&sslrootcert=certs%2Fca.crt&statement_timeout=10s' 'postgres://root@10.142.1.71:26257?sslcert=certs%2Fclient.root.crt&sslkey=certs%2Fclient.root.key&sslmode=verify-full&sslrootcert=certs%2Fca.crt&statement_timeout=10s' 'postgres://root@10.142.1.111:26257?sslcert=certs%2Fclient.root.crt&sslkey=certs%2Fclient.root.key&sslmode=verify-full&sslrootcert=certs%2Fca.crt&statement_timeout=10s' returned: COMMAND_PROBLEM: exit status 1
test artifacts and logs in: /artifacts/network_logging/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
See: [Grafana](https://go.crdb.dev/p/roachfana/teamcity-11169499-1691041653-02-n4cpu4/1691042550200/1691044466863)
</p>
</details>
/cc @cockroachdb/obs-inf-prs
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*network_logging.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
2.0
|
roachtest: network_logging failed - roachtest.network_logging [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/11169499?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/11169499?buildTab=artifacts#/network_logging) on master @ [4fe2a80d81c6fc5a3da3c7c44c5fc38da67e0367](https://github.com/cockroachdb/cockroach/commits/4fe2a80d81c6fc5a3da3c7c44c5fc38da67e0367):
```
(monitor.go:137).Wait: monitor failure: monitor task failed: output in run_060402.371453719_n4_cockroach-workload-r: ./cockroach workload run kv --concurrency=32 --duration=1h 'postgres://root@10.142.1.19:26257?sslcert=certs%2Fclient.root.crt&sslkey=certs%2Fclient.root.key&sslmode=verify-full&sslrootcert=certs%2Fca.crt&statement_timeout=10s' 'postgres://root@10.142.1.71:26257?sslcert=certs%2Fclient.root.crt&sslkey=certs%2Fclient.root.key&sslmode=verify-full&sslrootcert=certs%2Fca.crt&statement_timeout=10s' 'postgres://root@10.142.1.111:26257?sslcert=certs%2Fclient.root.crt&sslkey=certs%2Fclient.root.key&sslmode=verify-full&sslrootcert=certs%2Fca.crt&statement_timeout=10s' returned: COMMAND_PROBLEM: exit status 1
test artifacts and logs in: /artifacts/network_logging/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
See: [Grafana](https://go.crdb.dev/p/roachfana/teamcity-11169499-1691041653-02-n4cpu4/1691042550200/1691044466863)
</p>
</details>
/cc @cockroachdb/obs-inf-prs
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*network_logging.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
roachtest network logging failed roachtest network logging with on master monitor go wait monitor failure monitor task failed output in run cockroach workload r cockroach workload run kv concurrency duration postgres root sslcert certs root crt sslkey certs root key sslmode verify full sslrootcert certs crt statement timeout postgres root sslcert certs root crt sslkey certs root key sslmode verify full sslrootcert certs crt statement timeout postgres root sslcert certs root crt sslkey certs root key sslmode verify full sslrootcert certs crt statement timeout returned command problem exit status test artifacts and logs in artifacts network logging run parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see see cc cockroachdb obs inf prs
| 1
|
332,718
| 29,491,356,806
|
IssuesEvent
|
2023-06-02 13:41:50
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Obras Públicas: Dados para acompanhamento - Santa Juliana
|
generalization test development template - GRP (27) tag - Obras Públicas subtag - Dados para acompanhamento
|
DoD: Realizar o teste de Generalização do validador da tag Obras Públicas: Dados para acompanhamento para o Município de Santa Juliana.
|
1.0
|
Teste de generalizacao para a tag Obras Públicas: Dados para acompanhamento - Santa Juliana - DoD: Realizar o teste de Generalização do validador da tag Obras Públicas: Dados para acompanhamento para o Município de Santa Juliana.
|
test
|
teste de generalizacao para a tag obras públicas dados para acompanhamento santa juliana dod realizar o teste de generalização do validador da tag obras públicas dados para acompanhamento para o município de santa juliana
| 1
|
142,241
| 11,459,976,077
|
IssuesEvent
|
2020-02-07 08:44:27
|
robotology/assistive-rehab
|
https://api.github.com/repos/robotology/assistive-rehab
|
closed
|
Verify temporal metric
|
✅ test
|
We want to be sure that the computed time is not affected by disturbances (such as questions during the test).
|
1.0
|
Verify temporal metric - We want to be sure that the computed time is not affected by disturbances (such as questions during the test).
|
test
|
verify temporal metric we want to be sure that the computed time is not affected by disturbances such as questions during the test
| 1
|
271,590
| 23,616,566,051
|
IssuesEvent
|
2022-08-24 16:21:59
|
gradle/gradle
|
https://api.github.com/repos/gradle/gradle
|
closed
|
Named argument notation not supported for dependencies in test suite plugin
|
a:feature @core in:dependency-declarations in:test-suites
|
### Expected Behavior
When declaring [dependencies of a test suite](https://docs.gradle.org/current/userguide/jvm_test_suite_plugin.html#configure_dependencies_of_a_test_suite), it should be possible to use map notation like in top level dependency declarations (e.g. `group = "org.junit.jupiter", name = "junit-jupiter", version = "5.8.2"`), rather than be forced to use string notation (e.g. `"org.junit.jupiter:junit-jupiter:5.8.2"`)
For example, the following should be valid:
```kotlin
testing {
suites {
val test by getting(JvmTestSuite::class) {
useJUnitJupiter()
dependencies {
implementation(group = "org.junit.jupiter", name = "junit-jupiter", version = "5.8.2")
runtimeOnly(group = "org.junit.jupiter", name = "junit-jupiter-engine", version = "5.8.2")
}
}
}
}
```
### Current Behavior
The test suite plugin requires dependencies to be declared using string notation. The example above has to be rewritten as:
```kotlin
testing {
suites {
val test by getting(JvmTestSuite::class) {
useJUnitJupiter()
dependencies {
implementation("org.junit.jupiter:junit-jupiter:5.8.2")
runtimeOnly("org.junit.jupiter:junit-jupiter-engine:5.8.2")
}
}
}
}
```
### Context
In the interests of consistency, and hence ease of migration and also intuitiveness, test suite dependencies should be declared in the same ways as regular module dependencies.
In my opinion, map notation is easier to understand, and less prone to fat-finger errors.
### Steps to Reproduce
See [example project](https://github.com/markslater/gradle-examples/tree/13356368fdc2e7f7356ed573e2a4e3830d62b051)
### Your Environment
Using Gradle 7.3.1
|
1.0
|
Named argument notation not supported for dependencies in test suite plugin - ### Expected Behavior
When declaring [dependencies of a test suite](https://docs.gradle.org/current/userguide/jvm_test_suite_plugin.html#configure_dependencies_of_a_test_suite), it should be possible to use map notation like in top level dependency declarations (e.g. `group = "org.junit.jupiter", name = "junit-jupiter", version = "5.8.2"`), rather than be forced to use string notation (e.g. `"org.junit.jupiter:junit-jupiter:5.8.2"`)
For example, the following should be valid:
```kotlin
testing {
suites {
val test by getting(JvmTestSuite::class) {
useJUnitJupiter()
dependencies {
implementation(group = "org.junit.jupiter", name = "junit-jupiter", version = "5.8.2")
runtimeOnly(group = "org.junit.jupiter", name = "junit-jupiter-engine", version = "5.8.2")
}
}
}
}
```
### Current Behavior
The test suite plugin requires dependencies to be declared using string notation. The example above has to be rewritten as:
```kotlin
testing {
suites {
val test by getting(JvmTestSuite::class) {
useJUnitJupiter()
dependencies {
implementation("org.junit.jupiter:junit-jupiter:5.8.2")
runtimeOnly("org.junit.jupiter:junit-jupiter-engine:5.8.2")
}
}
}
}
```
### Context
In the interests of consistency, and hence ease of migration and also intuitiveness, test suite dependencies should be declared in the same ways as regular module dependencies.
In my opinion, map notation is easier to understand, and less prone to fat-finger errors.
### Steps to Reproduce
See [example project](https://github.com/markslater/gradle-examples/tree/13356368fdc2e7f7356ed573e2a4e3830d62b051)
### Your Environment
Using Gradle 7.3.1
|
test
|
named argument notation not supported for dependencies in test suite plugin expected behavior when declaring it should be possible to use map notation like in top level dependency declarations e g group org junit jupiter name junit jupiter version rather than be forced to use string notation e g org junit jupiter junit jupiter for example the following should be valid kotlin testing suites val test by getting jvmtestsuite class usejunitjupiter dependencies implementation group org junit jupiter name junit jupiter version runtimeonly group org junit jupiter name junit jupiter engine version current behavior the test suite plugin requires dependencies to be declared using string notation the example above has to be rewritten as kotlin testing suites val test by getting jvmtestsuite class usejunitjupiter dependencies implementation org junit jupiter junit jupiter runtimeonly org junit jupiter junit jupiter engine context in the interests of consistency and hence ease of migration and also intuitiveness test suite dependencies should be declared in the same ways as regular module dependencies in my opinion map notation is easier to understand and less prone to fat finger errors steps to reproduce see your environment using gradle
| 1
|
227,307
| 18,054,565,542
|
IssuesEvent
|
2021-09-20 06:06:33
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Failing test: X-Pack API Integration Tests.x-pack/test/api_integration/apis/ml/anomaly_detectors/close_with_spaces·ts - apis Machine Learning anomaly detectors POST anomaly_detectors _close with spaces "before all" hook for "should close job from same space"
|
:ml failed-test
|
A test failed on a tracked branch
```
Error: [POST http://elastic:changeme@localhost:61191/api/kibana/settings] request failed (attempt=3/3): Request failed with status code 503 -- and ran out of retries
at KbnClientRequester.request (/dev/shm/workspace/parallel/19/kibana/node_modules/@kbn/test/target_node/kbn_client/kbn_client_requester.js:158:15)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (internal/process/task_queues.js:95:5)
at KbnClientUiSettings.update (/dev/shm/workspace/parallel/19/kibana/node_modules/@kbn/test/target_node/kbn_client/kbn_client_ui_settings.js:101:5)
at Object.setKibanaTimeZoneToUTC (test/functional/services/ml/test_resources.ts:35:7)
at Context.<anonymous> (test/api_integration/apis/ml/anomaly_detectors/close_with_spaces.ts:40:7)
at Object.apply (/dev/shm/workspace/parallel/19/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16)
```
First failure: [CI Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/16970/)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack API Integration Tests.x-pack/test/api_integration/apis/ml/anomaly_detectors/close_with_spaces·ts","test.name":"apis Machine Learning anomaly detectors POST anomaly_detectors _close with spaces \"before all\" hook for \"should close job from same space\"","test.failCount":1}} -->
|
1.0
|
Failing test: X-Pack API Integration Tests.x-pack/test/api_integration/apis/ml/anomaly_detectors/close_with_spaces·ts - apis Machine Learning anomaly detectors POST anomaly_detectors _close with spaces "before all" hook for "should close job from same space" - A test failed on a tracked branch
```
Error: [POST http://elastic:changeme@localhost:61191/api/kibana/settings] request failed (attempt=3/3): Request failed with status code 503 -- and ran out of retries
at KbnClientRequester.request (/dev/shm/workspace/parallel/19/kibana/node_modules/@kbn/test/target_node/kbn_client/kbn_client_requester.js:158:15)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (internal/process/task_queues.js:95:5)
at KbnClientUiSettings.update (/dev/shm/workspace/parallel/19/kibana/node_modules/@kbn/test/target_node/kbn_client/kbn_client_ui_settings.js:101:5)
at Object.setKibanaTimeZoneToUTC (test/functional/services/ml/test_resources.ts:35:7)
at Context.<anonymous> (test/api_integration/apis/ml/anomaly_detectors/close_with_spaces.ts:40:7)
at Object.apply (/dev/shm/workspace/parallel/19/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16)
```
First failure: [CI Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/16970/)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack API Integration Tests.x-pack/test/api_integration/apis/ml/anomaly_detectors/close_with_spaces·ts","test.name":"apis Machine Learning anomaly detectors POST anomaly_detectors _close with spaces \"before all\" hook for \"should close job from same space\"","test.failCount":1}} -->
|
test
|
failing test x pack api integration tests x pack test api integration apis ml anomaly detectors close with spaces·ts apis machine learning anomaly detectors post anomaly detectors close with spaces before all hook for should close job from same space a test failed on a tracked branch error request failed attempt request failed with status code and ran out of retries at kbnclientrequester request dev shm workspace parallel kibana node modules kbn test target node kbn client kbn client requester js at runmicrotasks at processticksandrejections internal process task queues js at kbnclientuisettings update dev shm workspace parallel kibana node modules kbn test target node kbn client kbn client ui settings js at object setkibanatimezonetoutc test functional services ml test resources ts at context test api integration apis ml anomaly detectors close with spaces ts at object apply dev shm workspace parallel kibana node modules kbn test target node functional test runner lib mocha wrap function js first failure
| 1
|
271,289
| 23,593,578,273
|
IssuesEvent
|
2022-08-23 17:12:05
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Obras públicas - Dados para acompanhamento - Cataguases
|
generalization test development template - Betha tag - Obras Públicas subtag - Dados para acompanhamento
|
DoD: Realizar o teste de Generalização do validador da tag Obras públicas - Dados para acompanhamento para o Município de Cataguases.
|
1.0
|
Teste de generalizacao para a tag Obras públicas - Dados para acompanhamento - Cataguases - DoD: Realizar o teste de Generalização do validador da tag Obras públicas - Dados para acompanhamento para o Município de Cataguases.
|
test
|
teste de generalizacao para a tag obras públicas dados para acompanhamento cataguases dod realizar o teste de generalização do validador da tag obras públicas dados para acompanhamento para o município de cataguases
| 1
|
335,522
| 30,038,108,227
|
IssuesEvent
|
2023-06-27 13:57:36
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Failing test: Jest Tests.x-pack/plugins/cases/public/components/case_view - CaseViewPage Tabs renders the activity tab when the query parameter tabId has an unknown value
|
failed-test skipped-test Team:ResponseOps Feature:Cases
|
A test failed on a tracked branch
```
TestingLibraryElementError: Unable to find an element by: [data-test-subj="case-view-tab-content-activity"]
Ignored nodes: comments, script, style
<body>
<div />
</body>
at Object.getElementError (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/config.js:40:19)
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:90:38
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:62:17
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:111:19
at getByTestId (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/case_view/case_view_page.test.tsx:512:23)
at batchedUpdates$1 (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/react-dom/cjs/react-dom.development.js:22380:12)
at act (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/react-dom/cjs/react-dom-test-utils.development.js:1042:14)
at Object.<anonymous> (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/case_view/case_view_page.test.tsx:511:16)
at Promise.then.completed (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:289:28)
at new Promise (<anonymous>)
at callAsyncCircusFn (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:222:10)
at _callCircusTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:248:40)
at runNextTicks (node:internal/process/task_queues:61:5)
at processTimers (node:internal/timers:499:9)
at _runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:184:3)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:86:9)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9)
at run (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:26:3)
at runAndTransformResultsToJestFormat (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:120:21)
at jestAdapter (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:79:19)
at runTestInternal (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:367:16)
at runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:444:34)
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/26147#0185f527-6859-4d01-ac37-e5483a535e48)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/cases/public/components/case_view","test.name":"CaseViewPage Tabs renders the activity tab when the query parameter tabId has an unknown value","test.failCount":4}} -->
|
2.0
|
Failing test: Jest Tests.x-pack/plugins/cases/public/components/case_view - CaseViewPage Tabs renders the activity tab when the query parameter tabId has an unknown value - A test failed on a tracked branch
```
TestingLibraryElementError: Unable to find an element by: [data-test-subj="case-view-tab-content-activity"]
Ignored nodes: comments, script, style
<body>
<div />
</body>
at Object.getElementError (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/config.js:40:19)
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:90:38
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:62:17
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:111:19
at getByTestId (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/case_view/case_view_page.test.tsx:512:23)
at batchedUpdates$1 (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/react-dom/cjs/react-dom.development.js:22380:12)
at act (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/react-dom/cjs/react-dom-test-utils.development.js:1042:14)
at Object.<anonymous> (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/case_view/case_view_page.test.tsx:511:16)
at Promise.then.completed (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:289:28)
at new Promise (<anonymous>)
at callAsyncCircusFn (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:222:10)
at _callCircusTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:248:40)
at runNextTicks (node:internal/process/task_queues:61:5)
at processTimers (node:internal/timers:499:9)
at _runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:184:3)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:86:9)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9)
at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9)
at run (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:26:3)
at runAndTransformResultsToJestFormat (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:120:21)
at jestAdapter (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:79:19)
at runTestInternal (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:367:16)
at runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-3f2434e6d3ddd216/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:444:34)
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/26147#0185f527-6859-4d01-ac37-e5483a535e48)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/cases/public/components/case_view","test.name":"CaseViewPage Tabs renders the activity tab when the query parameter tabId has an unknown value","test.failCount":4}} -->
|
test
|
failing test jest tests x pack plugins cases public components case view caseviewpage tabs renders the activity tab when the query parameter tabid has an unknown value a test failed on a tracked branch testinglibraryelementerror unable to find an element by ignored nodes comments script style at object getelementerror var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist config js at var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist query helpers js at var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist query helpers js at var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist query helpers js at getbytestid var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins cases public components case view case view page test tsx at batchedupdates var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules react dom cjs react dom development js at act var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules react dom cjs react dom test utils development js at object var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins cases public components case view case view page test tsx at promise then completed var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build utils js at new promise at callasynccircusfn var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build utils js at callcircustest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runnextticks node internal process task queues at processtimers node internal timers at runtest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at run var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runandtransformresultstojestformat var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build legacy code todo rewrite jestadapterinit js at jestadapter var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build legacy code todo rewrite jestadapter js at runtestinternal var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runner build runtest js at runtest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runner build runtest js first failure
| 1
|
411,766
| 27,830,449,666
|
IssuesEvent
|
2023-03-20 04:03:13
|
1C-Company/v8-code-style
|
https://api.github.com/repos/1C-Company/v8-code-style
|
opened
|
Добавить описание doc-comment-description-ends-on-dot.md
|
documentation
|
## Раздел документации или код проверки
<!-- Путь к разделу документации: -->
`doc-comment-description-ends-on-dot.md`
## Что необходимо улучшить
<!-- Кратко опишите, что нужно улучшить, исправить. -->
Добавить описание
|
1.0
|
Добавить описание doc-comment-description-ends-on-dot.md - ## Раздел документации или код проверки
<!-- Путь к разделу документации: -->
`doc-comment-description-ends-on-dot.md`
## Что необходимо улучшить
<!-- Кратко опишите, что нужно улучшить, исправить. -->
Добавить описание
|
non_test
|
добавить описание doc comment description ends on dot md раздел документации или код проверки doc comment description ends on dot md что необходимо улучшить добавить описание
| 0
|
12,860
| 15,108,285,034
|
IssuesEvent
|
2021-02-08 16:27:25
|
Conjurinc-workato-dev/evoke
|
https://api.github.com/repos/Conjurinc-workato-dev/evoke
|
opened
|
j2g bug
|
Bugtype/Compatibility ONYX-6615 severity/low team/A Team
|
##description
Steps to reproduce:
Current Results:ddddd
dsa
ads
a
sdas
Expected Results:
Error Messages:
Logs:
Other Symptoms:
Tenant ID / Pod Number:
##Found in version
10.8
##Workaround Complexity
NA
##Workaround Description
##Link to JIRA bug
https://ca-il-jira-test.il.cyber-ark.com/browse/ONYX-6615
|
True
|
j2g bug - ##description
Steps to reproduce:
Current Results:ddddd
dsa
ads
a
sdas
Expected Results:
Error Messages:
Logs:
Other Symptoms:
Tenant ID / Pod Number:
##Found in version
10.8
##Workaround Complexity
NA
##Workaround Description
##Link to JIRA bug
https://ca-il-jira-test.il.cyber-ark.com/browse/ONYX-6615
|
non_test
|
bug description steps to reproduce current results ddddd dsa ads a sdas expected results error messages logs other symptoms tenant id pod number found in version workaround complexity na workaround description link to jira bug
| 0
|
26,269
| 4,212,770,789
|
IssuesEvent
|
2016-06-29 17:10:31
|
ngageoint/hootenanny-ui
|
https://api.github.com/repos/ngageoint/hootenanny-ui
|
closed
|
Limit how far the user can expand the conflation area
|
Category: UI in progress Priority: Medium Status: Ready for Test
|
Currently, the dataset area can be expanded to cover the entire screen. This should be limited to at max, half the screen
Example:

|
1.0
|
Limit how far the user can expand the conflation area - Currently, the dataset area can be expanded to cover the entire screen. This should be limited to at max, half the screen
Example:

|
test
|
limit how far the user can expand the conflation area currently the dataset area can be expanded to cover the entire screen this should be limited to at max half the screen example
| 1
|
88,199
| 8,134,643,198
|
IssuesEvent
|
2018-08-19 18:12:16
|
YACS-RCOS/yacs-admin
|
https://api.github.com/repos/YACS-RCOS/yacs-admin
|
opened
|
Perform accessibility audit
|
accessibility testing
|
We need to perform an accessibility audit to make sure our application is fully accessible to people with disabilities.
Ideally, our project should meet the [WCAG 2.1](https://www.w3.org/WAI/WCAG21/quickref/) guidelines.
If we find any issues during the audit, we should create GitHub issues for each issue we find, and tag it with the "accessibility" label.
|
1.0
|
Perform accessibility audit - We need to perform an accessibility audit to make sure our application is fully accessible to people with disabilities.
Ideally, our project should meet the [WCAG 2.1](https://www.w3.org/WAI/WCAG21/quickref/) guidelines.
If we find any issues during the audit, we should create GitHub issues for each issue we find, and tag it with the "accessibility" label.
|
test
|
perform accessibility audit we need to perform an accessibility audit to make sure our application is fully accessible to people with disabilities ideally our project should meet the guidelines if we find any issues during the audit we should create github issues for each issue we find and tag it with the accessibility label
| 1
|
3,289
| 4,310,915,316
|
IssuesEvent
|
2016-07-21 20:49:37
|
mozilla/testpilot
|
https://api.github.com/repos/mozilla/testpilot
|
opened
|
Upgrade insecure Django dependency
|
security status: planned
|
This says our Django dependency is insecure:
https://requires.io/github/mozilla/testpilot/requirements/?branch=master
We should upgrade ASAP.
|
True
|
Upgrade insecure Django dependency - This says our Django dependency is insecure:
https://requires.io/github/mozilla/testpilot/requirements/?branch=master
We should upgrade ASAP.
|
non_test
|
upgrade insecure django dependency this says our django dependency is insecure we should upgrade asap
| 0
|
193,378
| 14,648,480,852
|
IssuesEvent
|
2020-12-27 03:16:45
|
github-vet/rangeloop-pointer-findings
|
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
|
closed
|
zhuowei/go-1-2-haiku: src/pkg/runtime/race/testdata/mop_test.go; 12 LoC
|
fresh small test
|
Found a possible issue in [zhuowei/go-1-2-haiku](https://www.github.com/zhuowei/go-1-2-haiku) at [src/pkg/runtime/race/testdata/mop_test.go](https://github.com/zhuowei/go-1-2-haiku/blob/c86129514c33070e68041753a5365fdbed1e0eed/src/pkg/runtime/race/testdata/mop_test.go#L327-L338)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable v used in defer or goroutine at line 332
[Click here to see the code in its original context.](https://github.com/zhuowei/go-1-2-haiku/blob/c86129514c33070e68041753a5365fdbed1e0eed/src/pkg/runtime/race/testdata/mop_test.go#L327-L338)
<details>
<summary>Click here to show the 12 line(s) of Go which triggered the analyzer.</summary>
```go
for i, v := range a {
go func(i int) {
// we don't want a write-vs-write race
// so there is no array b here
if i == 0 {
x = v
} else {
y = v
}
done <- true
}(i)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: c86129514c33070e68041753a5365fdbed1e0eed
|
1.0
|
zhuowei/go-1-2-haiku: src/pkg/runtime/race/testdata/mop_test.go; 12 LoC -
Found a possible issue in [zhuowei/go-1-2-haiku](https://www.github.com/zhuowei/go-1-2-haiku) at [src/pkg/runtime/race/testdata/mop_test.go](https://github.com/zhuowei/go-1-2-haiku/blob/c86129514c33070e68041753a5365fdbed1e0eed/src/pkg/runtime/race/testdata/mop_test.go#L327-L338)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable v used in defer or goroutine at line 332
[Click here to see the code in its original context.](https://github.com/zhuowei/go-1-2-haiku/blob/c86129514c33070e68041753a5365fdbed1e0eed/src/pkg/runtime/race/testdata/mop_test.go#L327-L338)
<details>
<summary>Click here to show the 12 line(s) of Go which triggered the analyzer.</summary>
```go
for i, v := range a {
go func(i int) {
// we don't want a write-vs-write race
// so there is no array b here
if i == 0 {
x = v
} else {
y = v
}
done <- true
}(i)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: c86129514c33070e68041753a5365fdbed1e0eed
|
test
|
zhuowei go haiku src pkg runtime race testdata mop test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable v used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for i v range a go func i int we don t want a write vs write race so there is no array b here if i x v else y v done true i leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 1
|
53,384
| 28,111,506,372
|
IssuesEvent
|
2023-03-31 07:34:42
|
bevyengine/bevy
|
https://api.github.com/repos/bevyengine/bevy
|
closed
|
Event microbenchmarks
|
C-Enhancement A-ECS C-Performance A-Diagnostics
|
## What problem does this solve or what need does it fill?
We currently don't have performance benchmarks for ECS events, and are driving blind when making any performance related decision for them.
## What solution would you like?
Basic benchmarks for writing and reading events, varying based on total number of events read/written and the size of the event type itself.
## What alternative(s) have you considered?
Continue driving blindly into event performance changes.
|
True
|
Event microbenchmarks - ## What problem does this solve or what need does it fill?
We currently don't have performance benchmarks for ECS events, and are driving blind when making any performance related decision for them.
## What solution would you like?
Basic benchmarks for writing and reading events, varying based on total number of events read/written and the size of the event type itself.
## What alternative(s) have you considered?
Continue driving blindly into event performance changes.
|
non_test
|
event microbenchmarks what problem does this solve or what need does it fill we currently don t have performance benchmarks for ecs events and are driving blind when making any performance related decision for them what solution would you like basic benchmarks for writing and reading events varying based on total number of events read written and the size of the event type itself what alternative s have you considered continue driving blindly into event performance changes
| 0
|
91,711
| 8,317,039,983
|
IssuesEvent
|
2018-09-25 10:48:55
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
teamcity: failed tests on release-2.1: testrace/TestChangefeedTimestamps, test/TestChangefeedTimestamps
|
C-test-failure O-robot
|
The following tests appear to have failed:
[#881965](https://teamcity.cockroachdb.com/viewLog.html?buildId=881965):
```
--- FAIL: testrace/TestChangefeedTimestamps (2.860s)
sql_runner.go:82: error scanning '&{<nil> 0xc4210a8a00}': pq: restart transaction: HandledRetryableTxnError: TransactionRetryError: retry txn (RETRY_SERIALIZABLE): "sql txn" id=b116853d key=/Table/53/1/1/0 rw=true pri=0.00280938 iso=SERIALIZABLE stat=PENDING epo=0 ts=1536081199.146700970,1 orig=1536081199.146327736,0 max=1536081199.146327736,0 wto=false rop=false seq=3
------- Stdout: -------
W180904 17:13:16.400301 2654 server/status/runtime.go:310 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180904 17:13:16.449082 2654 server/server.go:849 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180904 17:13:16.449613 2654 base/addr_validation.go:279 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180904 17:13:16.449738 2654 base/addr_validation.go:319 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180904 17:13:16.472540 2654 server/config.go:493 [n?] 1 storage engine initialized
I180904 17:13:16.472688 2654 server/config.go:496 [n?] RocksDB cache size: 128 MiB
I180904 17:13:16.472761 2654 server/config.go:496 [n?] store 0: in-memory, size 0 B
I180904 17:13:16.571035 2654 server/node.go:373 [n?] **** cluster c80e08e3-3704-495b-b2b6-ba94a069ce84 has been created
I180904 17:13:16.571164 2654 server/server.go:1411 [n?] **** add additional nodes by specifying --join=127.0.0.1:41567
I180904 17:13:16.572160 2654 gossip/gossip.go:407 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41567" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:12 > build_tag:"v2.1.0-beta.20180904-169-gee45f46" started_at:1536081196571841995
I180904 17:13:16.573738 2770 gossip/gossip.go:941 [n1] gossip connectivity
I180904 17:13:16.602470 2654 storage/store.go:1551 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180904 17:13:16.604347 2654 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180904 17:13:16.604774 2654 storage/stores.go:242 [n1] read 0 node addresses from persistent storage
I180904 17:13:16.605237 2654 server/node.go:699 [n1] connecting to gossip network to verify cluster ID...
I180904 17:13:16.608077 2654 server/node.go:724 [n1] node connected via gossip and verified as part of cluster "c80e08e3-3704-495b-b2b6-ba94a069ce84"
I180904 17:13:16.608603 2654 server/node.go:548 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180904 17:13:16.609872 2654 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180904 17:13:16.609961 2654 server/server.go:1802 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180904 17:13:16.611810 2654 server/server.go:1548 [n1] starting https server at 127.0.0.1:34509 (use: 127.0.0.1:34509)
I180904 17:13:16.611900 2654 server/server.go:1550 [n1] starting grpc/postgres server at 127.0.0.1:41567
I180904 17:13:16.611961 2654 server/server.go:1551 [n1] advertising CockroachDB node at 127.0.0.1:41567
W180904 17:13:16.612218 2654 jobs/registry.go:309 [n1] unable to get node liveness: node not in the liveness table
I180904 17:13:16.620837 2860 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180904 17:13:16.625737 2726 storage/replica_command.go:298 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180904 17:13:16.689606 2582 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180904 17:13:16.721624 2584 storage/replica_command.go:298 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180904 17:13:16.962147 2979 storage/replica_command.go:298 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180904 17:13:17.061541 2985 storage/replica_command.go:298 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180904 17:13:17.183997 2995 storage/replica_command.go:298 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
W180904 17:13:17.258856 2592 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=4ca8545b key=/Local/Range/System/tsd/RangeDescriptor rw=true pri=0.06644633 iso=SERIALIZABLE stat=PENDING epo=0 ts=1536081197.184085426,0 orig=1536081197.184085426,0 max=1536081197.184085426,0 wto=false rop=false seq=1
I180904 17:13:17.266243 2729 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:2.0-12 User:root}
I180904 17:13:17.298421 3026 storage/replica_command.go:298 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180904 17:13:17.404508 3014 storage/replica_command.go:298 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180904 17:13:17.452222 2970 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180904 17:13:17.535696 3019 storage/replica_command.go:298 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180904 17:13:17.609414 3049 storage/replica_command.go:298 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180904 17:13:17.700837 3032 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:46dd0cca-1384-4df9-b7de-aa2006edb19e User:root}
I180904 17:13:17.701886 3091 storage/replica_command.go:298 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180904 17:13:17.767415 3022 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180904 17:13:17.796843 2941 storage/replica_command.go:298 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180904 17:13:17.815467 2992 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established
I180904 17:13:17.851923 3054 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180904 17:13:17.934420 3039 storage/replica_command.go:298 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180904 17:13:17.971892 2654 server/server.go:1604 [n1] done ensuring all necessary migrations have run
I180904 17:13:17.972048 2654 server/server.go:1607 [n1] serving sql connections
I180904 17:13:18.002536 3141 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180904 17:13:18.006173 3161 storage/replica_command.go:298 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180904 17:13:18.014787 3143 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41567} Attrs: Locality: ServerVersion:2.0-12 BuildTag:v2.1.0-beta.20180904-169-gee45f46 StartedAt:1536081196571841995 LocalityAddress:[]} ClusterID:c80e08e3-3704-495b-b2b6-ba94a069ce84 StartedAt:1536081196571841995 LastUp:1536081196571841995}
I180904 17:13:18.125665 3148 storage/replica_command.go:298 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180904 17:13:18.186776 3171 storage/replica_command.go:298 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180904 17:13:18.228736 3132 storage/replica_command.go:298 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180904 17:13:18.298630 3219 storage/replica_command.go:298 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180904 17:13:18.407522 3118 storage/replica_command.go:298 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180904 17:13:18.472736 3228 storage/replica_command.go:298 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
I180904 17:13:18.483097 2770 gossip/gossip.go:941 [n1] gossip connectivity
n1 [sentinel];
I180904 17:13:18.530953 3252 storage/replica_command.go:298 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180904 17:13:18.581771 3268 storage/replica_command.go:298 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
I180904 17:13:18.773358 3200 sql/event_log.go:126 [n1,client=127.0.0.1:58192,user=root] Event: "set_cluster_setting", target: 0, info: {SettingName:changefeed.experimental_poll_interval Value:0s User:root}
I180904 17:13:18.814061 3241 storage/replica_command.go:298 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
I180904 17:13:18.816977 3200 sql/event_log.go:126 [n1,client=127.0.0.1:58192,user=root] Event: "create_database", target: 52, info: {DatabaseName:d Statement:CREATE DATABASE d User:root}
I180904 17:13:18.945427 3246 storage/replica_command.go:298 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
I180904 17:13:18.978208 3200 sql/event_log.go:126 [n1,client=127.0.0.1:58192,user=root] Event: "create_table", target: 53, info: {TableName:d.public.foo Statement:CREATE TABLE foo (a INT PRIMARY KEY) User:root}
I180904 17:13:19.170551 2654 util/stop/stopper.go:537 quiescing; tasks left:
1 node.Node: batch
1 [async] transport racer
1 [async] closedts-rangefeed-subscriber
I180904 17:13:19.170766 2951 kv/transport_race.go:91 transport race promotion: ran 20 iterations on up to 926 requests
I180904 17:13:19.172214 2654 util/stop/stopper.go:537 quiescing; tasks left:
1 node.Node: batch
1 [async] closedts-rangefeed-subscriber
--- FAIL: test/TestChangefeedTimestamps (3.250s)
------- Stdout: -------
W180904 17:08:47.564362 2365 server/status/runtime.go:310 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180904 17:08:47.577338 2365 server/server.go:849 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180904 17:08:47.577552 2365 base/addr_validation.go:279 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180904 17:08:47.577583 2365 base/addr_validation.go:319 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180904 17:08:47.581766 2365 server/config.go:493 [n?] 1 storage engine initialized
I180904 17:08:47.581832 2365 server/config.go:496 [n?] RocksDB cache size: 128 MiB
I180904 17:08:47.581851 2365 server/config.go:496 [n?] store 0: in-memory, size 0 B
I180904 17:08:47.585754 2365 server/node.go:373 [n?] **** cluster 6585ec15-e47f-48a4-8083-a5b0ebf6a22a has been created
I180904 17:08:47.585778 2365 server/server.go:1411 [n?] **** add additional nodes by specifying --join=127.0.0.1:42459
I180904 17:08:47.585924 2365 gossip/gossip.go:407 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:42459" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:12 > build_tag:"v2.1.0-beta.20180904-169-gee45f46" started_at:1536080927585860998
I180904 17:08:47.586750 2365 storage/store.go:1551 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180904 17:08:47.586841 2365 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180904 17:08:47.586883 2365 storage/stores.go:242 [n1] read 0 node addresses from persistent storage
I180904 17:08:47.586937 2365 server/node.go:699 [n1] connecting to gossip network to verify cluster ID...
I180904 17:08:47.587027 2810 gossip/gossip.go:941 [n1] gossip connectivity
I180904 17:08:47.588135 2365 server/node.go:724 [n1] node connected via gossip and verified as part of cluster "6585ec15-e47f-48a4-8083-a5b0ebf6a22a"
I180904 17:08:47.588159 2365 server/node.go:548 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180904 17:08:47.588418 2365 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180904 17:08:47.588436 2365 server/server.go:1802 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180904 17:08:47.588497 2365 server/server.go:1548 [n1] starting https server at 127.0.0.1:38037 (use: 127.0.0.1:38037)
I180904 17:08:47.588511 2365 server/server.go:1550 [n1] starting grpc/postgres server at 127.0.0.1:42459
I180904 17:08:47.588523 2365 server/server.go:1551 [n1] advertising CockroachDB node at 127.0.0.1:42459
I180904 17:08:47.592015 2983 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180904 17:08:47.611107 2471 storage/replica_command.go:298 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180904 17:08:47.632524 2996 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180904 17:08:47.636310 2473 storage/replica_command.go:298 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180904 17:08:47.647387 2710 storage/replica_command.go:298 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180904 17:08:47.655740 2715 storage/replica_command.go:298 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180904 17:08:47.670040 3011 storage/replica_command.go:298 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
I180904 17:08:47.679168 2797 storage/replica_command.go:298 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180904 17:08:47.679873 2887 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:2.0-12 User:root}
I180904 17:08:47.688228 2737 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180904 17:08:47.689762 3016 storage/replica_command.go:298 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180904 17:08:47.718364 3024 storage/replica_command.go:298 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180904 17:08:47.727098 3074 storage/replica_command.go:298 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180904 17:08:47.733167 3079 storage/replica_command.go:298 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180904 17:08:47.736721 3098 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:0af804fa-090b-4a56-9cc5-59e27c62e69a User:root}
I180904 17:08:47.738695 3108 storage/replica_command.go:298 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180904 17:08:47.744510 3125 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180904 17:08:47.745966 3119 storage/replica_command.go:298 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
W180904 17:08:47.749580 3143 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=ef3b26dd key=/Local/Range/Table/15/RangeDescriptor rw=true pri=0.03807233 iso=SERIALIZABLE stat=PENDING epo=0 ts=1536080927.745982141,0 orig=1536080927.745982141,0 max=1536080927.745982141,0 wto=false rop=false seq=1
I180904 17:08:47.751650 3032 storage/replica_command.go:298 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180904 17:08:47.758893 3155 storage/replica_command.go:298 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180904 17:08:47.761859 3135 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180904 17:08:47.763312 2365 server/server.go:1604 [n1] done ensuring all necessary migrations have run
I180904 17:08:47.763355 2365 server/server.go:1607 [n1] serving sql connections
I180904 17:08:47.766812 3189 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180904 17:08:47.766980 3191 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:42459} Attrs: Locality: ServerVersion:2.0-12 BuildTag:v2.1.0-beta.20180904-169-gee45f46 StartedAt:1536080927585860998 LocalityAddress:[]} ClusterID:6585ec15-e47f-48a4-8083-a5b0ebf6a22a StartedAt:1536080927585860998 LastUp:1536080927585860998}
I180904 17:08:47.771512 3194 storage/replica_command.go:298 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180904 17:08:47.779598 3178 storage/replica_command.go:298 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180904 17:08:47.784640 3218 storage/replica_command.go:298 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180904 17:08:47.799178 3045 storage/replica_command.go:298 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180904 17:08:47.818477 3162 storage/replica_command.go:298 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
I180904 17:08:47.825156 3239 storage/replica_command.go:298 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180904 17:08:47.834059 3246 storage/replica_command.go:298 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
I180904 17:08:47.863445 3272 sql/event_log.go:126 [n1,client=127.0.0.1:38428,user=root] Event: "set_cluster_setting", target: 0, info: {SettingName:changefeed.experimental_poll_interval Value:0s User:root}
I180904 17:08:47.873580 3298 storage/replica_command.go:298 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
I180904 17:08:47.879439 3272 sql/event_log.go:126 [n1,client=127.0.0.1:38428,user=root] Event: "create_database", target: 52, info: {DatabaseName:d Statement:CREATE DATABASE d User:root}
I180904 17:08:47.884615 3308 storage/replica_command.go:298 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
I180904 17:08:47.887470 3272 sql/event_log.go:126 [n1,client=127.0.0.1:38428,user=root] Event: "create_table", target: 53, info: {TableName:d.public.foo Statement:CREATE TABLE foo (a INT PRIMARY KEY) User:root}
I180904 17:08:47.947584 2365 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-rangefeed-subscribersql_runner.go:82: error scanning '&{<nil> 0xc4210a8a00}': pq: restart transaction: HandledRetryableTxnError: TransactionRetryError: retry txn (RETRY_SERIALIZABLE): "sql txn" id=b116853d key=/Table/53/1/1/0 rw=true pri=0.00280938 iso=SERIALIZABLE stat=PENDING epo=0 ts=1536081199.146700970,1 orig=1536081199.146327736,0 max=1536081199.146327736,0 wto=false rop=false seq=3
------- Stdout: -------
W180904 17:13:16.400301 2654 server/status/runtime.go:310 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180904 17:13:16.449082 2654 server/server.go:849 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180904 17:13:16.449613 2654 base/addr_validation.go:279 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180904 17:13:16.449738 2654 base/addr_validation.go:319 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180904 17:13:16.472540 2654 server/config.go:493 [n?] 1 storage engine initialized
I180904 17:13:16.472688 2654 server/config.go:496 [n?] RocksDB cache size: 128 MiB
I180904 17:13:16.472761 2654 server/config.go:496 [n?] store 0: in-memory, size 0 B
I180904 17:13:16.571035 2654 server/node.go:373 [n?] **** cluster c80e08e3-3704-495b-b2b6-ba94a069ce84 has been created
I180904 17:13:16.571164 2654 server/server.go:1411 [n?] **** add additional nodes by specifying --join=127.0.0.1:41567
I180904 17:13:16.572160 2654 gossip/gossip.go:407 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41567" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:12 > build_tag:"v2.1.0-beta.20180904-169-gee45f46" started_at:1536081196571841995
I180904 17:13:16.573738 2770 gossip/gossip.go:941 [n1] gossip connectivity
I180904 17:13:16.602470 2654 storage/store.go:1551 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180904 17:13:16.604347 2654 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180904 17:13:16.604774 2654 storage/stores.go:242 [n1] read 0 node addresses from persistent storage
I180904 17:13:16.605237 2654 server/node.go:699 [n1] connecting to gossip network to verify cluster ID...
I180904 17:13:16.608077 2654 server/node.go:724 [n1] node connected via gossip and verified as part of cluster "c80e08e3-3704-495b-b2b6-ba94a069ce84"
I180904 17:13:16.608603 2654 server/node.go:548 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180904 17:13:16.609872 2654 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180904 17:13:16.609961 2654 server/server.go:1802 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180904 17:13:16.611810 2654 server/server.go:1548 [n1] starting https server at 127.0.0.1:34509 (use: 127.0.0.1:34509)
I180904 17:13:16.611900 2654 server/server.go:1550 [n1] starting grpc/postgres server at 127.0.0.1:41567
I180904 17:13:16.611961 2654 server/server.go:1551 [n1] advertising CockroachDB node at 127.0.0.1:41567
W180904 17:13:16.612218 2654 jobs/registry.go:309 [n1] unable to get node liveness: node not in the liveness table
I180904 17:13:16.620837 2860 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180904 17:13:16.625737 2726 storage/replica_command.go:298 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180904 17:13:16.689606 2582 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180904 17:13:16.721624 2584 storage/replica_command.go:298 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180904 17:13:16.962147 2979 storage/replica_command.go:298 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180904 17:13:17.061541 2985 storage/replica_command.go:298 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180904 17:13:17.183997 2995 storage/replica_command.go:298 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
W180904 17:13:17.258856 2592 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=4ca8545b key=/Local/Range/System/tsd/RangeDescriptor rw=true pri=0.06644633 iso=SERIALIZABLE stat=PENDING epo=0 ts=1536081197.184085426,0 orig=1536081197.184085426,0 max=1536081197.184085426,0 wto=false rop=false seq=1
I180904 17:13:17.266243 2729 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:2.0-12 User:root}
I180904 17:13:17.298421 3026 storage/replica_command.go:298 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180904 17:13:17.404508 3014 storage/replica_command.go:298 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180904 17:13:17.452222 2970 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180904 17:13:17.535696 3019 storage/replica_command.go:298 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180904 17:13:17.609414 3049 storage/replica_command.go:298 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180904 17:13:17.700837 3032 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:46dd0cca-1384-4df9-b7de-aa2006edb19e User:root}
I180904 17:13:17.701886 3091 storage/replica_command.go:298 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180904 17:13:17.767415 3022 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180904 17:13:17.796843 2941 storage/replica_command.go:298 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180904 17:13:17.815467 2992 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established
I180904 17:13:17.851923 3054 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180904 17:13:17.934420 3039 storage/replica_command.go:298 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180904 17:13:17.971892 2654 server/server.go:1604 [n1] done ensuring all necessary migrations have run
I180904 17:13:17.972048 2654 server/server.go:1607 [n1] serving sql connections
I180904 17:13:18.002536 3141 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180904 17:13:18.006173 3161 storage/replica_command.go:298 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180904 17:13:18.014787 3143 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41567} Attrs: Locality: ServerVersion:2.0-12 BuildTag:v2.1.0-beta.20180904-169-gee45f46 StartedAt:1536081196571841995 LocalityAddress:[]} ClusterID:c80e08e3-3704-495b-b2b6-ba94a069ce84 StartedAt:1536081196571841995 LastUp:1536081196571841995}
I180904 17:13:18.125665 3148 storage/replica_command.go:298 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180904 17:13:18.186776 3171 storage/replica_command.go:298 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180904 17:13:18.228736 3132 storage/replica_command.go:298 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180904 17:13:18.298630 3219 storage/replica_command.go:298 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180904 17:13:18.407522 3118 storage/replica_command.go:298 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180904 17:13:18.472736 3228 storage/replica_command.go:298 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
I180904 17:13:18.483097 2770 gossip/gossip.go:941 [n1] gossip connectivity
n1 [sentinel];
I180904 17:13:18.530953 3252 storage/replica_command.go:298 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180904 17:13:18.581771 3268 storage/replica_command.go:298 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
I180904 17:13:18.773358 3200 sql/event_log.go:126 [n1,client=127.0.0.1:58192,user=root] Event: "set_cluster_setting", target: 0, info: {SettingName:changefeed.experimental_poll_interval Value:0s User:root}
I180904 17:13:18.814061 3241 storage/replica_command.go:298 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
I180904 17:13:18.816977 3200 sql/event_log.go:126 [n1,client=127.0.0.1:58192,user=root] Event: "create_database", target: 52, info: {DatabaseName:d Statement:CREATE DATABASE d User:root}
I180904 17:13:18.945427 3246 storage/replica_command.go:298 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
I180904 17:13:18.978208 3200 sql/event_log.go:126 [n1,client=127.0.0.1:58192,user=root] Event: "create_table", target: 53, info: {TableName:d.public.foo Statement:CREATE TABLE foo (a INT PRIMARY KEY) User:root}
I180904 17:13:19.170551 2654 util/stop/stopper.go:537 quiescing; tasks left:
1 node.Node: batch
1 [async] transport racer
1 [async] closedts-rangefeed-subscriber
I180904 17:13:19.170766 2951 kv/transport_race.go:91 transport race promotion: ran 20 iterations on up to 926 requests
I180904 17:13:19.172214 2654 util/stop/stopper.go:537 quiescing; tasks left:
1 node.Node: batch
1 [async] closedts-rangefeed-subscriber
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
teamcity: failed tests on release-2.1: testrace/TestChangefeedTimestamps, test/TestChangefeedTimestamps - The following tests appear to have failed:
[#881965](https://teamcity.cockroachdb.com/viewLog.html?buildId=881965):
```
--- FAIL: testrace/TestChangefeedTimestamps (2.860s)
sql_runner.go:82: error scanning '&{<nil> 0xc4210a8a00}': pq: restart transaction: HandledRetryableTxnError: TransactionRetryError: retry txn (RETRY_SERIALIZABLE): "sql txn" id=b116853d key=/Table/53/1/1/0 rw=true pri=0.00280938 iso=SERIALIZABLE stat=PENDING epo=0 ts=1536081199.146700970,1 orig=1536081199.146327736,0 max=1536081199.146327736,0 wto=false rop=false seq=3
------- Stdout: -------
W180904 17:13:16.400301 2654 server/status/runtime.go:310 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180904 17:13:16.449082 2654 server/server.go:849 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180904 17:13:16.449613 2654 base/addr_validation.go:279 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180904 17:13:16.449738 2654 base/addr_validation.go:319 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180904 17:13:16.472540 2654 server/config.go:493 [n?] 1 storage engine initialized
I180904 17:13:16.472688 2654 server/config.go:496 [n?] RocksDB cache size: 128 MiB
I180904 17:13:16.472761 2654 server/config.go:496 [n?] store 0: in-memory, size 0 B
I180904 17:13:16.571035 2654 server/node.go:373 [n?] **** cluster c80e08e3-3704-495b-b2b6-ba94a069ce84 has been created
I180904 17:13:16.571164 2654 server/server.go:1411 [n?] **** add additional nodes by specifying --join=127.0.0.1:41567
I180904 17:13:16.572160 2654 gossip/gossip.go:407 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41567" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:12 > build_tag:"v2.1.0-beta.20180904-169-gee45f46" started_at:1536081196571841995
I180904 17:13:16.573738 2770 gossip/gossip.go:941 [n1] gossip connectivity
I180904 17:13:16.602470 2654 storage/store.go:1551 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180904 17:13:16.604347 2654 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180904 17:13:16.604774 2654 storage/stores.go:242 [n1] read 0 node addresses from persistent storage
I180904 17:13:16.605237 2654 server/node.go:699 [n1] connecting to gossip network to verify cluster ID...
I180904 17:13:16.608077 2654 server/node.go:724 [n1] node connected via gossip and verified as part of cluster "c80e08e3-3704-495b-b2b6-ba94a069ce84"
I180904 17:13:16.608603 2654 server/node.go:548 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180904 17:13:16.609872 2654 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180904 17:13:16.609961 2654 server/server.go:1802 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180904 17:13:16.611810 2654 server/server.go:1548 [n1] starting https server at 127.0.0.1:34509 (use: 127.0.0.1:34509)
I180904 17:13:16.611900 2654 server/server.go:1550 [n1] starting grpc/postgres server at 127.0.0.1:41567
I180904 17:13:16.611961 2654 server/server.go:1551 [n1] advertising CockroachDB node at 127.0.0.1:41567
W180904 17:13:16.612218 2654 jobs/registry.go:309 [n1] unable to get node liveness: node not in the liveness table
I180904 17:13:16.620837 2860 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180904 17:13:16.625737 2726 storage/replica_command.go:298 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180904 17:13:16.689606 2582 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180904 17:13:16.721624 2584 storage/replica_command.go:298 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180904 17:13:16.962147 2979 storage/replica_command.go:298 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180904 17:13:17.061541 2985 storage/replica_command.go:298 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180904 17:13:17.183997 2995 storage/replica_command.go:298 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
W180904 17:13:17.258856 2592 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=4ca8545b key=/Local/Range/System/tsd/RangeDescriptor rw=true pri=0.06644633 iso=SERIALIZABLE stat=PENDING epo=0 ts=1536081197.184085426,0 orig=1536081197.184085426,0 max=1536081197.184085426,0 wto=false rop=false seq=1
I180904 17:13:17.266243 2729 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:2.0-12 User:root}
I180904 17:13:17.298421 3026 storage/replica_command.go:298 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180904 17:13:17.404508 3014 storage/replica_command.go:298 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180904 17:13:17.452222 2970 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180904 17:13:17.535696 3019 storage/replica_command.go:298 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180904 17:13:17.609414 3049 storage/replica_command.go:298 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180904 17:13:17.700837 3032 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:46dd0cca-1384-4df9-b7de-aa2006edb19e User:root}
I180904 17:13:17.701886 3091 storage/replica_command.go:298 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180904 17:13:17.767415 3022 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180904 17:13:17.796843 2941 storage/replica_command.go:298 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180904 17:13:17.815467 2992 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established
I180904 17:13:17.851923 3054 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180904 17:13:17.934420 3039 storage/replica_command.go:298 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180904 17:13:17.971892 2654 server/server.go:1604 [n1] done ensuring all necessary migrations have run
I180904 17:13:17.972048 2654 server/server.go:1607 [n1] serving sql connections
I180904 17:13:18.002536 3141 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180904 17:13:18.006173 3161 storage/replica_command.go:298 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180904 17:13:18.014787 3143 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41567} Attrs: Locality: ServerVersion:2.0-12 BuildTag:v2.1.0-beta.20180904-169-gee45f46 StartedAt:1536081196571841995 LocalityAddress:[]} ClusterID:c80e08e3-3704-495b-b2b6-ba94a069ce84 StartedAt:1536081196571841995 LastUp:1536081196571841995}
I180904 17:13:18.125665 3148 storage/replica_command.go:298 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180904 17:13:18.186776 3171 storage/replica_command.go:298 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180904 17:13:18.228736 3132 storage/replica_command.go:298 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180904 17:13:18.298630 3219 storage/replica_command.go:298 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180904 17:13:18.407522 3118 storage/replica_command.go:298 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180904 17:13:18.472736 3228 storage/replica_command.go:298 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
I180904 17:13:18.483097 2770 gossip/gossip.go:941 [n1] gossip connectivity
n1 [sentinel];
I180904 17:13:18.530953 3252 storage/replica_command.go:298 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180904 17:13:18.581771 3268 storage/replica_command.go:298 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
I180904 17:13:18.773358 3200 sql/event_log.go:126 [n1,client=127.0.0.1:58192,user=root] Event: "set_cluster_setting", target: 0, info: {SettingName:changefeed.experimental_poll_interval Value:0s User:root}
I180904 17:13:18.814061 3241 storage/replica_command.go:298 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
I180904 17:13:18.816977 3200 sql/event_log.go:126 [n1,client=127.0.0.1:58192,user=root] Event: "create_database", target: 52, info: {DatabaseName:d Statement:CREATE DATABASE d User:root}
I180904 17:13:18.945427 3246 storage/replica_command.go:298 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
I180904 17:13:18.978208 3200 sql/event_log.go:126 [n1,client=127.0.0.1:58192,user=root] Event: "create_table", target: 53, info: {TableName:d.public.foo Statement:CREATE TABLE foo (a INT PRIMARY KEY) User:root}
I180904 17:13:19.170551 2654 util/stop/stopper.go:537 quiescing; tasks left:
1 node.Node: batch
1 [async] transport racer
1 [async] closedts-rangefeed-subscriber
I180904 17:13:19.170766 2951 kv/transport_race.go:91 transport race promotion: ran 20 iterations on up to 926 requests
I180904 17:13:19.172214 2654 util/stop/stopper.go:537 quiescing; tasks left:
1 node.Node: batch
1 [async] closedts-rangefeed-subscriber
--- FAIL: test/TestChangefeedTimestamps (3.250s)
------- Stdout: -------
W180904 17:08:47.564362 2365 server/status/runtime.go:310 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180904 17:08:47.577338 2365 server/server.go:849 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180904 17:08:47.577552 2365 base/addr_validation.go:279 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180904 17:08:47.577583 2365 base/addr_validation.go:319 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180904 17:08:47.581766 2365 server/config.go:493 [n?] 1 storage engine initialized
I180904 17:08:47.581832 2365 server/config.go:496 [n?] RocksDB cache size: 128 MiB
I180904 17:08:47.581851 2365 server/config.go:496 [n?] store 0: in-memory, size 0 B
I180904 17:08:47.585754 2365 server/node.go:373 [n?] **** cluster 6585ec15-e47f-48a4-8083-a5b0ebf6a22a has been created
I180904 17:08:47.585778 2365 server/server.go:1411 [n?] **** add additional nodes by specifying --join=127.0.0.1:42459
I180904 17:08:47.585924 2365 gossip/gossip.go:407 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:42459" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:12 > build_tag:"v2.1.0-beta.20180904-169-gee45f46" started_at:1536080927585860998
I180904 17:08:47.586750 2365 storage/store.go:1551 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180904 17:08:47.586841 2365 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180904 17:08:47.586883 2365 storage/stores.go:242 [n1] read 0 node addresses from persistent storage
I180904 17:08:47.586937 2365 server/node.go:699 [n1] connecting to gossip network to verify cluster ID...
I180904 17:08:47.587027 2810 gossip/gossip.go:941 [n1] gossip connectivity
I180904 17:08:47.588135 2365 server/node.go:724 [n1] node connected via gossip and verified as part of cluster "6585ec15-e47f-48a4-8083-a5b0ebf6a22a"
I180904 17:08:47.588159 2365 server/node.go:548 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180904 17:08:47.588418 2365 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180904 17:08:47.588436 2365 server/server.go:1802 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180904 17:08:47.588497 2365 server/server.go:1548 [n1] starting https server at 127.0.0.1:38037 (use: 127.0.0.1:38037)
I180904 17:08:47.588511 2365 server/server.go:1550 [n1] starting grpc/postgres server at 127.0.0.1:42459
I180904 17:08:47.588523 2365 server/server.go:1551 [n1] advertising CockroachDB node at 127.0.0.1:42459
I180904 17:08:47.592015 2983 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180904 17:08:47.611107 2471 storage/replica_command.go:298 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180904 17:08:47.632524 2996 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180904 17:08:47.636310 2473 storage/replica_command.go:298 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180904 17:08:47.647387 2710 storage/replica_command.go:298 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180904 17:08:47.655740 2715 storage/replica_command.go:298 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180904 17:08:47.670040 3011 storage/replica_command.go:298 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
I180904 17:08:47.679168 2797 storage/replica_command.go:298 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180904 17:08:47.679873 2887 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:2.0-12 User:root}
I180904 17:08:47.688228 2737 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180904 17:08:47.689762 3016 storage/replica_command.go:298 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180904 17:08:47.718364 3024 storage/replica_command.go:298 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180904 17:08:47.727098 3074 storage/replica_command.go:298 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180904 17:08:47.733167 3079 storage/replica_command.go:298 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180904 17:08:47.736721 3098 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:0af804fa-090b-4a56-9cc5-59e27c62e69a User:root}
I180904 17:08:47.738695 3108 storage/replica_command.go:298 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180904 17:08:47.744510 3125 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180904 17:08:47.745966 3119 storage/replica_command.go:298 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
W180904 17:08:47.749580 3143 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=ef3b26dd key=/Local/Range/Table/15/RangeDescriptor rw=true pri=0.03807233 iso=SERIALIZABLE stat=PENDING epo=0 ts=1536080927.745982141,0 orig=1536080927.745982141,0 max=1536080927.745982141,0 wto=false rop=false seq=1
I180904 17:08:47.751650 3032 storage/replica_command.go:298 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180904 17:08:47.758893 3155 storage/replica_command.go:298 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180904 17:08:47.761859 3135 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180904 17:08:47.763312 2365 server/server.go:1604 [n1] done ensuring all necessary migrations have run
I180904 17:08:47.763355 2365 server/server.go:1607 [n1] serving sql connections
I180904 17:08:47.766812 3189 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180904 17:08:47.766980 3191 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:42459} Attrs: Locality: ServerVersion:2.0-12 BuildTag:v2.1.0-beta.20180904-169-gee45f46 StartedAt:1536080927585860998 LocalityAddress:[]} ClusterID:6585ec15-e47f-48a4-8083-a5b0ebf6a22a StartedAt:1536080927585860998 LastUp:1536080927585860998}
I180904 17:08:47.771512 3194 storage/replica_command.go:298 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180904 17:08:47.779598 3178 storage/replica_command.go:298 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180904 17:08:47.784640 3218 storage/replica_command.go:298 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180904 17:08:47.799178 3045 storage/replica_command.go:298 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180904 17:08:47.818477 3162 storage/replica_command.go:298 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
I180904 17:08:47.825156 3239 storage/replica_command.go:298 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180904 17:08:47.834059 3246 storage/replica_command.go:298 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
I180904 17:08:47.863445 3272 sql/event_log.go:126 [n1,client=127.0.0.1:38428,user=root] Event: "set_cluster_setting", target: 0, info: {SettingName:changefeed.experimental_poll_interval Value:0s User:root}
I180904 17:08:47.873580 3298 storage/replica_command.go:298 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
I180904 17:08:47.879439 3272 sql/event_log.go:126 [n1,client=127.0.0.1:38428,user=root] Event: "create_database", target: 52, info: {DatabaseName:d Statement:CREATE DATABASE d User:root}
I180904 17:08:47.884615 3308 storage/replica_command.go:298 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
I180904 17:08:47.887470 3272 sql/event_log.go:126 [n1,client=127.0.0.1:38428,user=root] Event: "create_table", target: 53, info: {TableName:d.public.foo Statement:CREATE TABLE foo (a INT PRIMARY KEY) User:root}
I180904 17:08:47.947584 2365 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-rangefeed-subscribersql_runner.go:82: error scanning '&{<nil> 0xc4210a8a00}': pq: restart transaction: HandledRetryableTxnError: TransactionRetryError: retry txn (RETRY_SERIALIZABLE): "sql txn" id=b116853d key=/Table/53/1/1/0 rw=true pri=0.00280938 iso=SERIALIZABLE stat=PENDING epo=0 ts=1536081199.146700970,1 orig=1536081199.146327736,0 max=1536081199.146327736,0 wto=false rop=false seq=3
------- Stdout: -------
W180904 17:13:16.400301 2654 server/status/runtime.go:310 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180904 17:13:16.449082 2654 server/server.go:849 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180904 17:13:16.449613 2654 base/addr_validation.go:279 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180904 17:13:16.449738 2654 base/addr_validation.go:319 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180904 17:13:16.472540 2654 server/config.go:493 [n?] 1 storage engine initialized
I180904 17:13:16.472688 2654 server/config.go:496 [n?] RocksDB cache size: 128 MiB
I180904 17:13:16.472761 2654 server/config.go:496 [n?] store 0: in-memory, size 0 B
I180904 17:13:16.571035 2654 server/node.go:373 [n?] **** cluster c80e08e3-3704-495b-b2b6-ba94a069ce84 has been created
I180904 17:13:16.571164 2654 server/server.go:1411 [n?] **** add additional nodes by specifying --join=127.0.0.1:41567
I180904 17:13:16.572160 2654 gossip/gossip.go:407 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41567" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:12 > build_tag:"v2.1.0-beta.20180904-169-gee45f46" started_at:1536081196571841995
I180904 17:13:16.573738 2770 gossip/gossip.go:941 [n1] gossip connectivity
I180904 17:13:16.602470 2654 storage/store.go:1551 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180904 17:13:16.604347 2654 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180904 17:13:16.604774 2654 storage/stores.go:242 [n1] read 0 node addresses from persistent storage
I180904 17:13:16.605237 2654 server/node.go:699 [n1] connecting to gossip network to verify cluster ID...
I180904 17:13:16.608077 2654 server/node.go:724 [n1] node connected via gossip and verified as part of cluster "c80e08e3-3704-495b-b2b6-ba94a069ce84"
I180904 17:13:16.608603 2654 server/node.go:548 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180904 17:13:16.609872 2654 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180904 17:13:16.609961 2654 server/server.go:1802 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180904 17:13:16.611810 2654 server/server.go:1548 [n1] starting https server at 127.0.0.1:34509 (use: 127.0.0.1:34509)
I180904 17:13:16.611900 2654 server/server.go:1550 [n1] starting grpc/postgres server at 127.0.0.1:41567
I180904 17:13:16.611961 2654 server/server.go:1551 [n1] advertising CockroachDB node at 127.0.0.1:41567
W180904 17:13:16.612218 2654 jobs/registry.go:309 [n1] unable to get node liveness: node not in the liveness table
I180904 17:13:16.620837 2860 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180904 17:13:16.625737 2726 storage/replica_command.go:298 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180904 17:13:16.689606 2582 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180904 17:13:16.721624 2584 storage/replica_command.go:298 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180904 17:13:16.962147 2979 storage/replica_command.go:298 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180904 17:13:17.061541 2985 storage/replica_command.go:298 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180904 17:13:17.183997 2995 storage/replica_command.go:298 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
W180904 17:13:17.258856 2592 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=4ca8545b key=/Local/Range/System/tsd/RangeDescriptor rw=true pri=0.06644633 iso=SERIALIZABLE stat=PENDING epo=0 ts=1536081197.184085426,0 orig=1536081197.184085426,0 max=1536081197.184085426,0 wto=false rop=false seq=1
I180904 17:13:17.266243 2729 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:2.0-12 User:root}
I180904 17:13:17.298421 3026 storage/replica_command.go:298 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180904 17:13:17.404508 3014 storage/replica_command.go:298 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180904 17:13:17.452222 2970 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180904 17:13:17.535696 3019 storage/replica_command.go:298 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180904 17:13:17.609414 3049 storage/replica_command.go:298 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180904 17:13:17.700837 3032 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:46dd0cca-1384-4df9-b7de-aa2006edb19e User:root}
I180904 17:13:17.701886 3091 storage/replica_command.go:298 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180904 17:13:17.767415 3022 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180904 17:13:17.796843 2941 storage/replica_command.go:298 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180904 17:13:17.815467 2992 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established
I180904 17:13:17.851923 3054 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180904 17:13:17.934420 3039 storage/replica_command.go:298 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180904 17:13:17.971892 2654 server/server.go:1604 [n1] done ensuring all necessary migrations have run
I180904 17:13:17.972048 2654 server/server.go:1607 [n1] serving sql connections
I180904 17:13:18.002536 3141 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180904 17:13:18.006173 3161 storage/replica_command.go:298 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180904 17:13:18.014787 3143 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41567} Attrs: Locality: ServerVersion:2.0-12 BuildTag:v2.1.0-beta.20180904-169-gee45f46 StartedAt:1536081196571841995 LocalityAddress:[]} ClusterID:c80e08e3-3704-495b-b2b6-ba94a069ce84 StartedAt:1536081196571841995 LastUp:1536081196571841995}
I180904 17:13:18.125665 3148 storage/replica_command.go:298 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180904 17:13:18.186776 3171 storage/replica_command.go:298 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180904 17:13:18.228736 3132 storage/replica_command.go:298 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180904 17:13:18.298630 3219 storage/replica_command.go:298 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180904 17:13:18.407522 3118 storage/replica_command.go:298 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180904 17:13:18.472736 3228 storage/replica_command.go:298 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
I180904 17:13:18.483097 2770 gossip/gossip.go:941 [n1] gossip connectivity
n1 [sentinel];
I180904 17:13:18.530953 3252 storage/replica_command.go:298 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180904 17:13:18.581771 3268 storage/replica_command.go:298 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
I180904 17:13:18.773358 3200 sql/event_log.go:126 [n1,client=127.0.0.1:58192,user=root] Event: "set_cluster_setting", target: 0, info: {SettingName:changefeed.experimental_poll_interval Value:0s User:root}
I180904 17:13:18.814061 3241 storage/replica_command.go:298 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
I180904 17:13:18.816977 3200 sql/event_log.go:126 [n1,client=127.0.0.1:58192,user=root] Event: "create_database", target: 52, info: {DatabaseName:d Statement:CREATE DATABASE d User:root}
I180904 17:13:18.945427 3246 storage/replica_command.go:298 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
I180904 17:13:18.978208 3200 sql/event_log.go:126 [n1,client=127.0.0.1:58192,user=root] Event: "create_table", target: 53, info: {TableName:d.public.foo Statement:CREATE TABLE foo (a INT PRIMARY KEY) User:root}
I180904 17:13:19.170551 2654 util/stop/stopper.go:537 quiescing; tasks left:
1 node.Node: batch
1 [async] transport racer
1 [async] closedts-rangefeed-subscriber
I180904 17:13:19.170766 2951 kv/transport_race.go:91 transport race promotion: ran 20 iterations on up to 926 requests
I180904 17:13:19.172214 2654 util/stop/stopper.go:537 quiescing; tasks left:
1 node.Node: batch
1 [async] closedts-rangefeed-subscriber
```
Please assign, take a look and update the issue accordingly.
|
test
|
teamcity failed tests on release testrace testchangefeedtimestamps test testchangefeedtimestamps the following tests appear to have failed fail testrace testchangefeedtimestamps sql runner go error scanning pq restart transaction handledretryabletxnerror transactionretryerror retry txn retry serializable sql txn id key table rw true pri iso serializable stat pending epo ts orig max wto false rop false seq stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag beta started at gossip gossip go gossip connectivity storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases queries writes bytesperreplica pmax writesperreplica pmax storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at jobs registry go unable to get node liveness node not in the liveness table server status recorder go available memory from cgroups eib exceeds system memory gib using system memory storage replica command go initiating a split of this range at key system sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user root storage replica command go initiating a split of this range at key system nodeliveness storage replica command go initiating a split of this range at key system nodelivenessmax storage replica command go initiating a split of this range at key system tsd storage replica command go initiating a split of this range at key system tse storage intent resolver go failed to push during intent resolution failed to push split id key local range system tsd rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq sql event log go event set cluster setting target info settingname version value user root storage replica command go initiating a split of this range at key table systemconfigspan start storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname trace debug enable value false user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname cluster secret value user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename defaultdb statement create database if not exists defaultdb user root storage replica command go initiating a split of this range at key table rpc nodedialer nodedialer go connection to established sql event log go event create database target info databasename postgres statement create database if not exists postgres user root storage replica command go initiating a split of this range at key table server server go done ensuring all necessary migrations have run server server go serving sql connections server server update go no need to upgrade cluster already at the newest version storage replica command go initiating a split of this range at key table sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag beta startedat localityaddress clusterid startedat lastup storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table gossip gossip go gossip connectivity storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname changefeed experimental poll interval value user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename d statement create database d user root storage replica command go initiating a split of this range at key table sql event log go event create table target info tablename d public foo statement create table foo a int primary key user root util stop stopper go quiescing tasks left node node batch transport racer closedts rangefeed subscriber kv transport race go transport race promotion ran iterations on up to requests util stop stopper go quiescing tasks left node node batch closedts rangefeed subscriber fail test testchangefeedtimestamps stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag beta started at storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases queries writes bytesperreplica pmax writesperreplica pmax storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id gossip gossip go gossip connectivity server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at server status recorder go available memory from cgroups eib exceeds system memory gib using system memory storage replica command go initiating a split of this range at key system sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user root storage replica command go initiating a split of this range at key system nodeliveness storage replica command go initiating a split of this range at key system nodelivenessmax storage replica command go initiating a split of this range at key system tsd storage replica command go initiating a split of this range at key system tse storage replica command go initiating a split of this range at key table systemconfigspan start sql event log go event set cluster setting target info settingname version value user root sql event log go event set cluster setting target info settingname trace debug enable value false user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname cluster secret value user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename defaultdb statement create database if not exists defaultdb user root storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename postgres statement create database if not exists postgres user root server server go done ensuring all necessary migrations have run server server go serving sql connections server server update go no need to upgrade cluster already at the newest version sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag beta startedat localityaddress clusterid startedat lastup storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname changefeed experimental poll interval value user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename d statement create database d user root storage replica command go initiating a split of this range at key table sql event log go event create table target info tablename d public foo statement create table foo a int primary key user root util stop stopper go quiescing tasks left closedts rangefeed subscribersql runner go error scanning pq restart transaction handledretryabletxnerror transactionretryerror retry txn retry serializable sql txn id key table rw true pri iso serializable stat pending epo ts orig max wto false rop false seq stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag beta started at gossip gossip go gossip connectivity storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases queries writes bytesperreplica pmax writesperreplica pmax storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at jobs registry go unable to get node liveness node not in the liveness table server status recorder go available memory from cgroups eib exceeds system memory gib using system memory storage replica command go initiating a split of this range at key system sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user root storage replica command go initiating a split of this range at key system nodeliveness storage replica command go initiating a split of this range at key system nodelivenessmax storage replica command go initiating a split of this range at key system tsd storage replica command go initiating a split of this range at key system tse storage intent resolver go failed to push during intent resolution failed to push split id key local range system tsd rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq sql event log go event set cluster setting target info settingname version value user root storage replica command go initiating a split of this range at key table systemconfigspan start storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname trace debug enable value false user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname cluster secret value user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename defaultdb statement create database if not exists defaultdb user root storage replica command go initiating a split of this range at key table rpc nodedialer nodedialer go connection to established sql event log go event create database target info databasename postgres statement create database if not exists postgres user root storage replica command go initiating a split of this range at key table server server go done ensuring all necessary migrations have run server server go serving sql connections server server update go no need to upgrade cluster already at the newest version storage replica command go initiating a split of this range at key table sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag beta startedat localityaddress clusterid startedat lastup storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table gossip gossip go gossip connectivity storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname changefeed experimental poll interval value user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename d statement create database d user root storage replica command go initiating a split of this range at key table sql event log go event create table target info tablename d public foo statement create table foo a int primary key user root util stop stopper go quiescing tasks left node node batch transport racer closedts rangefeed subscriber kv transport race go transport race promotion ran iterations on up to requests util stop stopper go quiescing tasks left node node batch closedts rangefeed subscriber please assign take a look and update the issue accordingly
| 1
|
281,836
| 24,424,154,471
|
IssuesEvent
|
2022-10-06 00:06:20
|
aws/amazon-vpc-cni-k8s
|
https://api.github.com/repos/aws/amazon-vpc-cni-k8s
|
closed
|
Scenario-based e2e testing
|
testing stale
|
Add parameterization for e2e testing to allow permutations/scenarios for:
* MTU
* routing behaviour
* num ENIs, IPs, instance types, etc
|
1.0
|
Scenario-based e2e testing - Add parameterization for e2e testing to allow permutations/scenarios for:
* MTU
* routing behaviour
* num ENIs, IPs, instance types, etc
|
test
|
scenario based testing add parameterization for testing to allow permutations scenarios for mtu routing behaviour num enis ips instance types etc
| 1
|
65,135
| 7,857,842,448
|
IssuesEvent
|
2018-06-21 12:12:15
|
toggl/mobileapp
|
https://api.github.com/repos/toggl/mobileapp
|
closed
|
Add haptic feedback to the iOS app
|
ios needs-design
|
This is a minor feature I want to add - many other iOS apps (facebook, twitter, ...) have this kind of feedback for the pull-to-refresh gesture where the user gets a light vibration when he can stop pulling and the app will refresh. I think it makes the app feel a lot better.
I added a PR with an implementation and it can be used in the future for reference: #1305
There are also probably many other places where this kind of feedback makes sense and would be a nice touch (e.g., pressing the start or stop buttons) and it might be worth it to go through the app and find some more.
|
1.0
|
Add haptic feedback to the iOS app - This is a minor feature I want to add - many other iOS apps (facebook, twitter, ...) have this kind of feedback for the pull-to-refresh gesture where the user gets a light vibration when he can stop pulling and the app will refresh. I think it makes the app feel a lot better.
I added a PR with an implementation and it can be used in the future for reference: #1305
There are also probably many other places where this kind of feedback makes sense and would be a nice touch (e.g., pressing the start or stop buttons) and it might be worth it to go through the app and find some more.
|
non_test
|
add haptic feedback to the ios app this is a minor feature i want to add many other ios apps facebook twitter have this kind of feedback for the pull to refresh gesture where the user gets a light vibration when he can stop pulling and the app will refresh i think it makes the app feel a lot better i added a pr with an implementation and it can be used in the future for reference there are also probably many other places where this kind of feedback makes sense and would be a nice touch e g pressing the start or stop buttons and it might be worth it to go through the app and find some more
| 0
|
141,128
| 11,395,153,755
|
IssuesEvent
|
2020-01-30 10:48:28
|
wazuh/wazuh-qa
|
https://api.github.com/repos/wazuh/wazuh-qa
|
opened
|
FIM System tests: Select required OS's and generate AMIs for them
|
fim-system-tests
|
It's required to generate AMIs from selected OS in order to freeze the machine in order to get a controlled environment that always generates the same alerts by default.
### Supported OS's
**Agent**:
- [ ] Redhat 7
- [ ] Redhat 8
- [ ] Centos 7
- [ ] Centos 8
- [ ] Ubuntu 16.04
- [ ] Ubuntu 18.04
- [ ] Ubuntu 14.04
- [ ] Amazon Linux 2
- [ ] Windows 2016
- [ ] Windows 2019
- [ ] Windows 2012
- [ ] Debian 9
- [ ] Debian 10
**Manager:**
- [ ] Centos 7
### Tasks
- [ ] Instantiate an AMI for every OS, remove automatic updates and install required dependencies. Then generate the AMI
- [ ] Test the changes using an AWS CF template and deploy using Ansible
Best regards,
Jose
Best
|
1.0
|
FIM System tests: Select required OS's and generate AMIs for them - It's required to generate AMIs from selected OS in order to freeze the machine in order to get a controlled environment that always generates the same alerts by default.
### Supported OS's
**Agent**:
- [ ] Redhat 7
- [ ] Redhat 8
- [ ] Centos 7
- [ ] Centos 8
- [ ] Ubuntu 16.04
- [ ] Ubuntu 18.04
- [ ] Ubuntu 14.04
- [ ] Amazon Linux 2
- [ ] Windows 2016
- [ ] Windows 2019
- [ ] Windows 2012
- [ ] Debian 9
- [ ] Debian 10
**Manager:**
- [ ] Centos 7
### Tasks
- [ ] Instantiate an AMI for every OS, remove automatic updates and install required dependencies. Then generate the AMI
- [ ] Test the changes using an AWS CF template and deploy using Ansible
Best regards,
Jose
Best
|
test
|
fim system tests select required os s and generate amis for them it s required to generate amis from selected os in order to freeze the machine in order to get a controlled environment that always generates the same alerts by default supported os s agent redhat redhat centos centos ubuntu ubuntu ubuntu amazon linux windows windows windows debian debian manager centos tasks instantiate an ami for every os remove automatic updates and install required dependencies then generate the ami test the changes using an aws cf template and deploy using ansible best regards jose best
| 1
|
42,212
| 22,352,358,262
|
IssuesEvent
|
2022-06-15 13:08:20
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
Performance drops after running tensor multiplication for 15 seconds on M1 MAX (Pytorch MPS).
|
module: performance triaged module: mps
|
### 🐛 Describe the bug
Performance drops to half of the original performance after running tensor multiplication for 15 seconds on M1 MAX (Pytorch MPS).
Python version: 3.9.7
OS: macOS 12.4
Pytorch version: 1.13.0.dev20220612
The code below reproduces the error:
```
from tqdm import tqdm
import torch
x = torch.rand((1024 * 8, 1024 * 8), dtype=torch.float32, device=torch.device('mps'))
y = torch.rand((1024 * 8, 1024 * 8), dtype=torch.float32, device=torch.device('mps'))
@torch.jit.script
def foo():
x = torch.rand((1024 * 8, 1024 * 8), dtype=torch.float32, device=torch.device('mps'))
y = torch.rand((1024 * 8, 1024 * 8), dtype=torch.float32, device=torch.device('mps'))
z = x * y
return z
if __name__ == '__main__':
z0 = None
for _ in tqdm(range(10000000000)):
zz = foo()
if z0 is None:
z0 = zz
else:
z0 += zz
torch.mps.empty_cache()
```
The above code works fine for RTX 3090 and performance never drops (by changing mps to cuda), but the performance will drop from 71 it/s to 25 it/s after 15 seconds running on M1 Max chip.
Please look into the issue, I suspect that's the reason why training speed of some transformer models dropped after running for a few minutes.
### Versions
Collecting environment information...
PyTorch version: 1.13.0.dev20220612
Is debug build: False
CUDA used to build PyTorch: None
ROCM used to build PyTorch: N/A
OS: macOS 12.4 (arm64)
GCC version: Could not collect
Clang version: 13.1.6 (clang-1316.0.21.2.5)
CMake version: version 3.22.3
Libc version: N/A
Python version: 3.9.7 | packaged by conda-forge | (default, Sep 29 2021, 19:24:02) [Clang 11.1.0 ] (64-bit runtime)
Python platform: macOS-12.4-arm64-arm-64bit
Is CUDA available: False
CUDA runtime version: No CUDA
GPU models and configuration: No CUDA
Nvidia driver version: No CUDA
cuDNN version: No CUDA
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
Versions of relevant libraries:
[pip3] numpy==1.22.3
[pip3] torch==1.13.0.dev20220612
[pip3] torchaudio==0.14.0.dev20220603
[pip3] torchtext==0.10.0
[pip3] torchvision==0.14.0a0+f9f721d
[conda] numpy 1.22.3 pypi_0 pypi
[conda] torch 1.13.0.dev20220612 pypi_0 pypi
[conda] torchaudio 0.14.0.dev20220603 pypi_0 pypi
[conda] torchtext 0.10.0 pypi_0 pypi
[conda] torchvision 0.14.0a0+f9f721d pypi_0 pypi
cc @VitalyFedyunin @ngimel @kulinseth @albanD
|
True
|
Performance drops after running tensor multiplication for 15 seconds on M1 MAX (Pytorch MPS). - ### 🐛 Describe the bug
Performance drops to half of the original performance after running tensor multiplication for 15 seconds on M1 MAX (Pytorch MPS).
Python version: 3.9.7
OS: macOS 12.4
Pytorch version: 1.13.0.dev20220612
The code below reproduces the error:
```
from tqdm import tqdm
import torch
x = torch.rand((1024 * 8, 1024 * 8), dtype=torch.float32, device=torch.device('mps'))
y = torch.rand((1024 * 8, 1024 * 8), dtype=torch.float32, device=torch.device('mps'))
@torch.jit.script
def foo():
x = torch.rand((1024 * 8, 1024 * 8), dtype=torch.float32, device=torch.device('mps'))
y = torch.rand((1024 * 8, 1024 * 8), dtype=torch.float32, device=torch.device('mps'))
z = x * y
return z
if __name__ == '__main__':
z0 = None
for _ in tqdm(range(10000000000)):
zz = foo()
if z0 is None:
z0 = zz
else:
z0 += zz
torch.mps.empty_cache()
```
The above code works fine for RTX 3090 and performance never drops (by changing mps to cuda), but the performance will drop from 71 it/s to 25 it/s after 15 seconds running on M1 Max chip.
Please look into the issue, I suspect that's the reason why training speed of some transformer models dropped after running for a few minutes.
### Versions
Collecting environment information...
PyTorch version: 1.13.0.dev20220612
Is debug build: False
CUDA used to build PyTorch: None
ROCM used to build PyTorch: N/A
OS: macOS 12.4 (arm64)
GCC version: Could not collect
Clang version: 13.1.6 (clang-1316.0.21.2.5)
CMake version: version 3.22.3
Libc version: N/A
Python version: 3.9.7 | packaged by conda-forge | (default, Sep 29 2021, 19:24:02) [Clang 11.1.0 ] (64-bit runtime)
Python platform: macOS-12.4-arm64-arm-64bit
Is CUDA available: False
CUDA runtime version: No CUDA
GPU models and configuration: No CUDA
Nvidia driver version: No CUDA
cuDNN version: No CUDA
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
Versions of relevant libraries:
[pip3] numpy==1.22.3
[pip3] torch==1.13.0.dev20220612
[pip3] torchaudio==0.14.0.dev20220603
[pip3] torchtext==0.10.0
[pip3] torchvision==0.14.0a0+f9f721d
[conda] numpy 1.22.3 pypi_0 pypi
[conda] torch 1.13.0.dev20220612 pypi_0 pypi
[conda] torchaudio 0.14.0.dev20220603 pypi_0 pypi
[conda] torchtext 0.10.0 pypi_0 pypi
[conda] torchvision 0.14.0a0+f9f721d pypi_0 pypi
cc @VitalyFedyunin @ngimel @kulinseth @albanD
|
non_test
|
performance drops after running tensor multiplication for seconds on max pytorch mps 🐛 describe the bug performance drops to half of the original performance after running tensor multiplication for seconds on max pytorch mps python version os macos pytorch version the code below reproduces the error from tqdm import tqdm import torch x torch rand dtype torch device torch device mps y torch rand dtype torch device torch device mps torch jit script def foo x torch rand dtype torch device torch device mps y torch rand dtype torch device torch device mps z x y return z if name main none for in tqdm range zz foo if is none zz else zz torch mps empty cache the above code works fine for rtx and performance never drops by changing mps to cuda but the performance will drop from it s to it s after seconds running on max chip please look into the issue i suspect that s the reason why training speed of some transformer models dropped after running for a few minutes versions collecting environment information pytorch version is debug build false cuda used to build pytorch none rocm used to build pytorch n a os macos gcc version could not collect clang version clang cmake version version libc version n a python version packaged by conda forge default sep bit runtime python platform macos arm is cuda available false cuda runtime version no cuda gpu models and configuration no cuda nvidia driver version no cuda cudnn version no cuda hip runtime version n a miopen runtime version n a is xnnpack available true versions of relevant libraries numpy torch torchaudio torchtext torchvision numpy pypi pypi torch pypi pypi torchaudio pypi pypi torchtext pypi pypi torchvision pypi pypi cc vitalyfedyunin ngimel kulinseth alband
| 0
|
38,552
| 15,727,676,612
|
IssuesEvent
|
2021-03-29 12:58:24
|
kyma-project/kyma
|
https://api.github.com/repos/kyma-project/kyma
|
opened
|
Networking component
|
Epic area/installation area/service-mesh
|
**Description**
Introduce a new component that can configure and manage the network setup for Kyma depending on the provider, topology etc...
**Reasons**
Correctly setup the network dynamically for all possible cases.
|
1.0
|
Networking component - **Description**
Introduce a new component that can configure and manage the network setup for Kyma depending on the provider, topology etc...
**Reasons**
Correctly setup the network dynamically for all possible cases.
|
non_test
|
networking component description introduce a new component that can configure and manage the network setup for kyma depending on the provider topology etc reasons correctly setup the network dynamically for all possible cases
| 0
|
183,214
| 14,933,997,520
|
IssuesEvent
|
2021-01-25 09:57:37
|
biopython/biopython
|
https://api.github.com/repos/biopython/biopython
|
opened
|
Update Bio.Application examples in Tutorial to use subprocess directly
|
Documentation help wanted
|
See discussion on #2877, closed via #3344 declaring Bio.Application obsolete, suggesting using subprocess directly.
We need to update the Tutorial to translate the command line wrapper examples to use subprocess directly. Working on Linux/macOS only would be a reasonable first pass but ideally tested as working on Windows too.
The tutorial is written in LaTeX, source here https://github.com/biopython/biopython/tree/master/Doc/Tutorial - on Linux/macOS using the ``make`` command in the ``Doc/`` directory will build this for you.
There is scope for multiple contributors to work on this, most simply by working on different sections of the Tutorial.
|
1.0
|
Update Bio.Application examples in Tutorial to use subprocess directly - See discussion on #2877, closed via #3344 declaring Bio.Application obsolete, suggesting using subprocess directly.
We need to update the Tutorial to translate the command line wrapper examples to use subprocess directly. Working on Linux/macOS only would be a reasonable first pass but ideally tested as working on Windows too.
The tutorial is written in LaTeX, source here https://github.com/biopython/biopython/tree/master/Doc/Tutorial - on Linux/macOS using the ``make`` command in the ``Doc/`` directory will build this for you.
There is scope for multiple contributors to work on this, most simply by working on different sections of the Tutorial.
|
non_test
|
update bio application examples in tutorial to use subprocess directly see discussion on closed via declaring bio application obsolete suggesting using subprocess directly we need to update the tutorial to translate the command line wrapper examples to use subprocess directly working on linux macos only would be a reasonable first pass but ideally tested as working on windows too the tutorial is written in latex source here on linux macos using the make command in the doc directory will build this for you there is scope for multiple contributors to work on this most simply by working on different sections of the tutorial
| 0
|
16,851
| 3,567,738,507
|
IssuesEvent
|
2016-01-26 00:27:24
|
d3athrow/vgstation13
|
https://api.github.com/repos/d3athrow/vgstation13
|
closed
|
(WEB REPORT BY: maddoscientisto REMOTE: 198.245.63.50:7777) Device analyzer can't unload items in the reverse engine
|
Needs Moar Testing
|
Revision: 0ab3a1198908b99f15d98e4b0a4b7f05f2b17978
>General description of the issue
I can't unload scanned items on the reverse engine neither from an analyzer nor from a pda, machine designs work fine but for example circuits don't.
Also the queue is empty and it's not doing any reversing.
>What you expected to happen
I expected the designs to unload in the machines and be available for printing
>What actually happened
nothing
>Steps to reproduce if possible
scan a circuit and try to load it in the reverse engine
|
1.0
|
(WEB REPORT BY: maddoscientisto REMOTE: 198.245.63.50:7777) Device analyzer can't unload items in the reverse engine - Revision: 0ab3a1198908b99f15d98e4b0a4b7f05f2b17978
>General description of the issue
I can't unload scanned items on the reverse engine neither from an analyzer nor from a pda, machine designs work fine but for example circuits don't.
Also the queue is empty and it's not doing any reversing.
>What you expected to happen
I expected the designs to unload in the machines and be available for printing
>What actually happened
nothing
>Steps to reproduce if possible
scan a circuit and try to load it in the reverse engine
|
test
|
web report by maddoscientisto remote device analyzer can t unload items in the reverse engine revision general description of the issue i can t unload scanned items on the reverse engine neither from an analyzer nor from a pda machine designs work fine but for example circuits don t also the queue is empty and it s not doing any reversing what you expected to happen i expected the designs to unload in the machines and be available for printing what actually happened nothing steps to reproduce if possible scan a circuit and try to load it in the reverse engine
| 1
|
29,507
| 7,103,489,181
|
IssuesEvent
|
2018-01-16 05:26:19
|
TornadoClientDev/Storm-Anticheat
|
https://api.github.com/repos/TornadoClientDev/Storm-Anticheat
|
reopened
|
1 Block Step Bypass - Envy 1.6
|
bug bypass code patching soon
|
1 block step works on storm anti-cheat using the client Envy 1.6.
Envy 1.6 download: https://www.youtube.com/watch?v=OmqNHo2-9ag
Showcase: https://gyazo.com/403effddabc84a030f3e5db226dd60ec
|
1.0
|
1 Block Step Bypass - Envy 1.6 - 1 block step works on storm anti-cheat using the client Envy 1.6.
Envy 1.6 download: https://www.youtube.com/watch?v=OmqNHo2-9ag
Showcase: https://gyazo.com/403effddabc84a030f3e5db226dd60ec
|
non_test
|
block step bypass envy block step works on storm anti cheat using the client envy envy download showcase
| 0
|
351,783
| 32,026,698,775
|
IssuesEvent
|
2023-09-22 09:20:12
|
project-codeflare/codeflare-operator
|
https://api.github.com/repos/project-codeflare/codeflare-operator
|
closed
|
e2e test support: Store test pod logs and events
|
testing triage/needs-triage
|
### Name of Feature or Improvement
Pod logs and events for test namespaces should be stored to help identifying test issues.
### Description of Problem the Feature Should Solve
Right now the test failures can be hard to investigate as the logs of pods in test namespaces aren't stored anywhere. Also events in the namespace should be stored, to identify issues like image pull problems.
### Describe the Solution You Would Like to See
e2e tests will store pod logs and namespace events in the output directory (where the Ray job logs are stored for now).
### Describe Alternatives You Have Considered
N/A
### Additional Context
|
1.0
|
e2e test support: Store test pod logs and events - ### Name of Feature or Improvement
Pod logs and events for test namespaces should be stored to help identifying test issues.
### Description of Problem the Feature Should Solve
Right now the test failures can be hard to investigate as the logs of pods in test namespaces aren't stored anywhere. Also events in the namespace should be stored, to identify issues like image pull problems.
### Describe the Solution You Would Like to See
e2e tests will store pod logs and namespace events in the output directory (where the Ray job logs are stored for now).
### Describe Alternatives You Have Considered
N/A
### Additional Context
|
test
|
test support store test pod logs and events name of feature or improvement pod logs and events for test namespaces should be stored to help identifying test issues description of problem the feature should solve right now the test failures can be hard to investigate as the logs of pods in test namespaces aren t stored anywhere also events in the namespace should be stored to identify issues like image pull problems describe the solution you would like to see tests will store pod logs and namespace events in the output directory where the ray job logs are stored for now describe alternatives you have considered n a additional context
| 1
|
53,386
| 6,719,260,298
|
IssuesEvent
|
2017-10-15 22:07:16
|
bnzk/djangocms-misc
|
https://api.github.com/repos/bnzk/djangocms-misc
|
closed
|
editmode_fallback: ordering and copy pasting not working...
|
bug design decision needed musthave
|
ordering and copy pasting is not working when done when fallback plugins are shown, as the plugin's placeholder id is the one from the original language the plugins where added. this is hard! monkey patch ahead :(
|
1.0
|
editmode_fallback: ordering and copy pasting not working... - ordering and copy pasting is not working when done when fallback plugins are shown, as the plugin's placeholder id is the one from the original language the plugins where added. this is hard! monkey patch ahead :(
|
non_test
|
editmode fallback ordering and copy pasting not working ordering and copy pasting is not working when done when fallback plugins are shown as the plugin s placeholder id is the one from the original language the plugins where added this is hard monkey patch ahead
| 0
|
521,534
| 15,110,654,509
|
IssuesEvent
|
2021-02-08 19:31:20
|
ansible/awx
|
https://api.github.com/repos/ansible/awx
|
closed
|
[ui_next] Remove User from Team's list
|
component:ui priority:high state:needs_devel type:feature
|
##### ISSUE TYPE
- Feature Idea
##### SUMMARY
The Team's list shows a list of users under 'Access'. We should allow users to be removed from a Team in the same way that they're removed from an Organization
|
1.0
|
[ui_next] Remove User from Team's list - ##### ISSUE TYPE
- Feature Idea
##### SUMMARY
The Team's list shows a list of users under 'Access'. We should allow users to be removed from a Team in the same way that they're removed from an Organization
|
non_test
|
remove user from team s list issue type feature idea summary the team s list shows a list of users under access we should allow users to be removed from a team in the same way that they re removed from an organization
| 0
|
31,905
| 6,016,962,719
|
IssuesEvent
|
2017-06-07 08:29:57
|
owncloud/core
|
https://api.github.com/repos/owncloud/core
|
closed
|
can't disable app with occ command
|
documentation
|
<!--
Thanks for reporting issues back to ownCloud! This is the issue tracker of ownCloud, if you have any support question please check out https://owncloud.org/support
This is the bug tracker for the Server component. Find other components at https://github.com/owncloud/core/blob/master/.github/CONTRIBUTING.md#guidelines
For reporting potential security issues please see https://owncloud.org/security/
To make it possible for us to help you please fill out below information carefully.
Before reporting any issues please make sure that you're using the latest available version for your major branch (e.g. 9.0.x), see https://owncloud.org/changelog/
-->
### Steps to reproduce
1. run occ comand: sudo -u www-data php occ app:disable files_external
2.
3.
### Expected behaviour
files_external disabled
### Actual behaviour
files_external can't be disabled.
### Server configuration
**Operating system**:
ubuntu 16.04
**Web server:**
apache2
**Database:**
mysql
**PHP version:**
7
**ownCloud version:** (see ownCloud admin page)
10.0.2
**Updated from an older ownCloud or fresh install:**
fresh
**Where did you install ownCloud from:**
owncloud.org
```
**The content of config/config.php:**
<?php
$CONFIG = array (
'filelocking.enabled' => true,
'memcache.locking' => '\OC\Memcache\Redis',
'redis' => array(
'host' => 'localhost',
'port' => 6379,
'timeout' => 0.0,
),
'memcache.local' => 'OC\Memcache\APCu',
'passwordsalt' => 'Bs6CIS8Xt2BMyU10dxWY74pOhffvgi',
'secret' => 'HTwmjbQNwbLz74WrPFcl72NU61vG/APBriU3lqEWjCUSGpBN',
'trusted_domains' =>
array (
0 => 'localhost',
),
'datadirectory' => '/var/www/owncloud/data',
'overwrite.cli.url' => 'http://localhost',
'dbtype' => 'mysql',
'version' => '10.0.2.1',
'dbname' => 'owncloud',
'dbhost' => 'localhost',
'dbtableprefix' => 'oc_',
'dbuser' => 'admin',
'dbpassword' => 'password',
'logtimezone' => 'UTC',
'installed' => true,
'instanceid' => 'ocdgiwkoc4qu',
);
```
```
**List of activated apps:**
root@oc10:/var/www/owncloud# sudo -u www-data php occ config:list system
{
"system": {
"filelocking.enabled": true,
"memcache.locking": "\\OC\\Memcache\\Redis",
"redis": {
"host": "localhost",
"port": 6379,
"timeout": 0
},
"memcache.local": "OC\\Memcache\\APCu",
"passwordsalt": "***REMOVED SENSITIVE VALUE***",
"secret": "***REMOVED SENSITIVE VALUE***",
"trusted_domains": [
"localhost"
],
"datadirectory": "\/var\/www\/owncloud\/data",
"overwrite.cli.url": "http:\/\/localhost",
"dbtype": "mysql",
"version": "10.0.2.1",
"dbname": "owncloud",
"dbhost": "localhost",
"dbtableprefix": "oc_",
"dbuser": "***REMOVED SENSITIVE VALUE***",
"dbpassword": "***REMOVED SENSITIVE VALUE***",
"logtimezone": "UTC",
"installed": true,
"instanceid": "ocdgiwkoc4qu"
}
}
```
```
root@oc10:/var/www/owncloud# sudo -u www-data php occ app:list
Enabled:
- comments: 0.3.0
- configreport: 0.1.1
- dav: 0.2.9
- federatedfilesharing: 0.3.0
- federation: 0.1.0
- files: 1.5.1
- files_external: 0.7.0
- files_sharing: 0.10.0
- files_trashbin: 0.9.0
- files_versions: 1.3.0
- files_videoplayer: 0.9.8
- firstrunwizard: 1.1
- market: 0.1.0
- notifications: 0.3.0
- provisioning_api: 0.5.0
- systemtags: 0.3.0
- templateeditor: 0.1
- updatenotification: 0.2.1
Disabled:
- encryption
- external
- files_antivirus
- theme-example
- user_external
```
**Are you using external storage, if yes which one:** local/smb/sftp/...
no
**Are you using encryption:** yes/no
no
**Are you using an external user-backend, if yes which one:** LDAP/ActiveDirectory/Webdav/...
no
### Client configuration
**Browser:**
firefox
**Operating system:**
ubuntu
### Logs
#### ownCloud log (data/owncloud.log)
```
root@oc10:/var/www/owncloud/data# more owncloud.log
{"reqId":"qDsDn6FcuGLiK3sYQfnp","level":3,"time":"2017-06-03T20:39:10+00:00","remoteAddr":"
","user":"--","app":"mysql.setup","method":"--","url":"--","message":
"Specific user creation failed: An exception occurred while executing
'SELECT user FROM mysql.user WHERE user=?'
with params [\"oc_admin\"]:\n\nSQLSTATE[42000]:
Syntax error or access violation: 1142 SELECT command denied to
user 'admin'@'localhost' for table 'user'"}
{"reqId":"c7f6KBjzstlbk9o1ybpz","level":3,"time":"2017-06-03T20:50:36+00:00",
"remoteAddr":"172.17.0.1","user":"--","app":"index","method":"GET","url":"\/owncloud\/index.php\/settings\/admin?sectionid=general","message":
"Exception: {\"Exception\":\"Error\",\"Message\":\"Class 'OCMemcacheAPCu' not found\",
\"Code\":0,\"Trace\":\"#0 \\\/var\\\/www\\\/owncloud\\\/lib\\\/private\\\/Server.php(381):
OC\\\\Memcache\\\\Factory->__construct('5d0de91ce88824c...', Object(OC\\\\Log), 'OCMemcacheAPCu', 'OCMemcacheAPCu', NULL)\\
n#1 \\\/var\\\/www\\\/owncloud\\\/lib\\\/composer\\\/pimple\\\/pimple
\\\/src\\\/Pimple\\\/Container.php(113): OC\\\\Server->OC\\\\{closure}(Object(OC\\\\Server))\\
n#2 \\\/var\\\/www\\\/owncloud\\\/lib\\\/private\\\/AppFramework\\\/Utility\\\/SimpleContainer.php(111
): Pimple\\\\Container->offsetGet('MemCacheFactory')\\
n#3 \\\/var\\\/www\\\/owncloud\\\/lib\\\/private\\\/ServerContainer.php(87):
OC\\\\AppFramework\\\\Utility\\\\SimpleContainer->query('MemCache
Factory')\\
n#4 \\\/var\\\/www\\\/owncloud\\\/lib\\\/private\\\/Server.php(1092): OC\\\\ServerContainer->query('MemCacheFactory')\\
n#5 \\\/var\\\/www\\\/owncloud\\\/lib\\\/base.php(821): OC\\\\Server->getMemCacheFactory()\\
n#6 \\\/var\\\/www\\\/owncloud\\\/lib\\\/base.php(560): OC::registerAutoloaderCache()\\
n#7 \\\/var\\\/www\\\/owncloud\\\/lib\\\/base.php(1003): OC::init()\\
n#8 \\\/var\\\/www\\\/owncloud\\\/index.php(47): require_once('\\\/var\\\/www\\\/ownclo...')\\
n#9 {main}\",\"File\":\"\\\/var\\\/www\\\/owncloud\\\/lib\\\/private\\\/Memcache\\\/Factory.php\",\"Line\":85}"}
```

|
1.0
|
can't disable app with occ command - <!--
Thanks for reporting issues back to ownCloud! This is the issue tracker of ownCloud, if you have any support question please check out https://owncloud.org/support
This is the bug tracker for the Server component. Find other components at https://github.com/owncloud/core/blob/master/.github/CONTRIBUTING.md#guidelines
For reporting potential security issues please see https://owncloud.org/security/
To make it possible for us to help you please fill out below information carefully.
Before reporting any issues please make sure that you're using the latest available version for your major branch (e.g. 9.0.x), see https://owncloud.org/changelog/
-->
### Steps to reproduce
1. run occ comand: sudo -u www-data php occ app:disable files_external
2.
3.
### Expected behaviour
files_external disabled
### Actual behaviour
files_external can't be disabled.
### Server configuration
**Operating system**:
ubuntu 16.04
**Web server:**
apache2
**Database:**
mysql
**PHP version:**
7
**ownCloud version:** (see ownCloud admin page)
10.0.2
**Updated from an older ownCloud or fresh install:**
fresh
**Where did you install ownCloud from:**
owncloud.org
```
**The content of config/config.php:**
<?php
$CONFIG = array (
'filelocking.enabled' => true,
'memcache.locking' => '\OC\Memcache\Redis',
'redis' => array(
'host' => 'localhost',
'port' => 6379,
'timeout' => 0.0,
),
'memcache.local' => 'OC\Memcache\APCu',
'passwordsalt' => 'Bs6CIS8Xt2BMyU10dxWY74pOhffvgi',
'secret' => 'HTwmjbQNwbLz74WrPFcl72NU61vG/APBriU3lqEWjCUSGpBN',
'trusted_domains' =>
array (
0 => 'localhost',
),
'datadirectory' => '/var/www/owncloud/data',
'overwrite.cli.url' => 'http://localhost',
'dbtype' => 'mysql',
'version' => '10.0.2.1',
'dbname' => 'owncloud',
'dbhost' => 'localhost',
'dbtableprefix' => 'oc_',
'dbuser' => 'admin',
'dbpassword' => 'password',
'logtimezone' => 'UTC',
'installed' => true,
'instanceid' => 'ocdgiwkoc4qu',
);
```
```
**List of activated apps:**
root@oc10:/var/www/owncloud# sudo -u www-data php occ config:list system
{
"system": {
"filelocking.enabled": true,
"memcache.locking": "\\OC\\Memcache\\Redis",
"redis": {
"host": "localhost",
"port": 6379,
"timeout": 0
},
"memcache.local": "OC\\Memcache\\APCu",
"passwordsalt": "***REMOVED SENSITIVE VALUE***",
"secret": "***REMOVED SENSITIVE VALUE***",
"trusted_domains": [
"localhost"
],
"datadirectory": "\/var\/www\/owncloud\/data",
"overwrite.cli.url": "http:\/\/localhost",
"dbtype": "mysql",
"version": "10.0.2.1",
"dbname": "owncloud",
"dbhost": "localhost",
"dbtableprefix": "oc_",
"dbuser": "***REMOVED SENSITIVE VALUE***",
"dbpassword": "***REMOVED SENSITIVE VALUE***",
"logtimezone": "UTC",
"installed": true,
"instanceid": "ocdgiwkoc4qu"
}
}
```
```
root@oc10:/var/www/owncloud# sudo -u www-data php occ app:list
Enabled:
- comments: 0.3.0
- configreport: 0.1.1
- dav: 0.2.9
- federatedfilesharing: 0.3.0
- federation: 0.1.0
- files: 1.5.1
- files_external: 0.7.0
- files_sharing: 0.10.0
- files_trashbin: 0.9.0
- files_versions: 1.3.0
- files_videoplayer: 0.9.8
- firstrunwizard: 1.1
- market: 0.1.0
- notifications: 0.3.0
- provisioning_api: 0.5.0
- systemtags: 0.3.0
- templateeditor: 0.1
- updatenotification: 0.2.1
Disabled:
- encryption
- external
- files_antivirus
- theme-example
- user_external
```
**Are you using external storage, if yes which one:** local/smb/sftp/...
no
**Are you using encryption:** yes/no
no
**Are you using an external user-backend, if yes which one:** LDAP/ActiveDirectory/Webdav/...
no
### Client configuration
**Browser:**
firefox
**Operating system:**
ubuntu
### Logs
#### ownCloud log (data/owncloud.log)
```
root@oc10:/var/www/owncloud/data# more owncloud.log
{"reqId":"qDsDn6FcuGLiK3sYQfnp","level":3,"time":"2017-06-03T20:39:10+00:00","remoteAddr":"
","user":"--","app":"mysql.setup","method":"--","url":"--","message":
"Specific user creation failed: An exception occurred while executing
'SELECT user FROM mysql.user WHERE user=?'
with params [\"oc_admin\"]:\n\nSQLSTATE[42000]:
Syntax error or access violation: 1142 SELECT command denied to
user 'admin'@'localhost' for table 'user'"}
{"reqId":"c7f6KBjzstlbk9o1ybpz","level":3,"time":"2017-06-03T20:50:36+00:00",
"remoteAddr":"172.17.0.1","user":"--","app":"index","method":"GET","url":"\/owncloud\/index.php\/settings\/admin?sectionid=general","message":
"Exception: {\"Exception\":\"Error\",\"Message\":\"Class 'OCMemcacheAPCu' not found\",
\"Code\":0,\"Trace\":\"#0 \\\/var\\\/www\\\/owncloud\\\/lib\\\/private\\\/Server.php(381):
OC\\\\Memcache\\\\Factory->__construct('5d0de91ce88824c...', Object(OC\\\\Log), 'OCMemcacheAPCu', 'OCMemcacheAPCu', NULL)\\
n#1 \\\/var\\\/www\\\/owncloud\\\/lib\\\/composer\\\/pimple\\\/pimple
\\\/src\\\/Pimple\\\/Container.php(113): OC\\\\Server->OC\\\\{closure}(Object(OC\\\\Server))\\
n#2 \\\/var\\\/www\\\/owncloud\\\/lib\\\/private\\\/AppFramework\\\/Utility\\\/SimpleContainer.php(111
): Pimple\\\\Container->offsetGet('MemCacheFactory')\\
n#3 \\\/var\\\/www\\\/owncloud\\\/lib\\\/private\\\/ServerContainer.php(87):
OC\\\\AppFramework\\\\Utility\\\\SimpleContainer->query('MemCache
Factory')\\
n#4 \\\/var\\\/www\\\/owncloud\\\/lib\\\/private\\\/Server.php(1092): OC\\\\ServerContainer->query('MemCacheFactory')\\
n#5 \\\/var\\\/www\\\/owncloud\\\/lib\\\/base.php(821): OC\\\\Server->getMemCacheFactory()\\
n#6 \\\/var\\\/www\\\/owncloud\\\/lib\\\/base.php(560): OC::registerAutoloaderCache()\\
n#7 \\\/var\\\/www\\\/owncloud\\\/lib\\\/base.php(1003): OC::init()\\
n#8 \\\/var\\\/www\\\/owncloud\\\/index.php(47): require_once('\\\/var\\\/www\\\/ownclo...')\\
n#9 {main}\",\"File\":\"\\\/var\\\/www\\\/owncloud\\\/lib\\\/private\\\/Memcache\\\/Factory.php\",\"Line\":85}"}
```

|
non_test
|
can t disable app with occ command thanks for reporting issues back to owncloud this is the issue tracker of owncloud if you have any support question please check out this is the bug tracker for the server component find other components at for reporting potential security issues please see to make it possible for us to help you please fill out below information carefully before reporting any issues please make sure that you re using the latest available version for your major branch e g x see steps to reproduce run occ comand sudo u www data php occ app disable files external expected behaviour files external disabled actual behaviour files external can t be disabled server configuration operating system ubuntu web server database mysql php version owncloud version see owncloud admin page updated from an older owncloud or fresh install fresh where did you install owncloud from owncloud org the content of config config php php config array filelocking enabled true memcache locking oc memcache redis redis array host localhost port timeout memcache local oc memcache apcu passwordsalt secret trusted domains array localhost datadirectory var www owncloud data overwrite cli url dbtype mysql version dbname owncloud dbhost localhost dbtableprefix oc dbuser admin dbpassword password logtimezone utc installed true instanceid list of activated apps root var www owncloud sudo u www data php occ config list system system filelocking enabled true memcache locking oc memcache redis redis host localhost port timeout memcache local oc memcache apcu passwordsalt removed sensitive value secret removed sensitive value trusted domains localhost datadirectory var www owncloud data overwrite cli url http localhost dbtype mysql version dbname owncloud dbhost localhost dbtableprefix oc dbuser removed sensitive value dbpassword removed sensitive value logtimezone utc installed true instanceid root var www owncloud sudo u www data php occ app list enabled comments configreport dav federatedfilesharing federation files files external files sharing files trashbin files versions files videoplayer firstrunwizard market notifications provisioning api systemtags templateeditor updatenotification disabled encryption external files antivirus theme example user external are you using external storage if yes which one local smb sftp no are you using encryption yes no no are you using an external user backend if yes which one ldap activedirectory webdav no client configuration browser firefox operating system ubuntu logs owncloud log data owncloud log root var www owncloud data more owncloud log reqid level time remoteaddr user app mysql setup method url message specific user creation failed an exception occurred while executing select user from mysql user where user with params n nsqlstate syntax error or access violation select command denied to user admin localhost for table user reqid level time remoteaddr user app index method get url owncloud index php settings admin sectionid general message exception exception error message class ocmemcacheapcu not found code trace var www owncloud lib private server php oc memcache factory construct object oc log ocmemcacheapcu ocmemcacheapcu null n var www owncloud lib composer pimple pimple src pimple container php oc server oc closure object oc server n var www owncloud lib private appframework utility simplecontainer php pimple container offsetget memcachefactory n var www owncloud lib private servercontainer php oc appframework utility simplecontainer query memcache factory n var www owncloud lib private server php oc servercontainer query memcachefactory n var www owncloud lib base php oc server getmemcachefactory n var www owncloud lib base php oc registerautoloadercache n var www owncloud lib base php oc init n var www owncloud index php require once var www ownclo n main file var www owncloud lib private memcache factory php line
| 0
|
27,374
| 4,308,332,939
|
IssuesEvent
|
2016-07-21 12:38:25
|
Imaginaerum/magento2-language-fr-fr
|
https://api.github.com/repos/Imaginaerum/magento2-language-fr-fr
|
closed
|
Error processing your request for french language
|
need tests
|
Magento gives me this error when I try to open the French Page:
```
a:4:{i:0;s:117:"Notice: Undefined offset: 1 in /var/www/html/magento/vendor/magento/framework/App/Language/Dictionary.php on line 196";i:1;s:9624:"#0 /var/www/html/magento/vendor/magento/framework/App/Language/Dictionary.php(196): Magento\Framework\App\ErrorHandler->handler(8, 'Undefined offse...', '/var/www/html/m...', 196, Array)
#1 /var/www/html/magento/vendor/magento/framework/App/Language/Dictionary.php(113): Magento\Framework\App\Language\Dictionary->readPackCsv('imaginaerum', 'fr_fr')
#2 /var/www/html/magento/vendor/magento/framework/Translate.php(323): Magento\Framework\App\Language\Dictionary->getDictionary('fr_FR')
#3 /var/www/html/magento/vendor/magento/framework/Translate.php(181): Magento\Framework\Translate->_loadPackTranslation()
#4 /var/www/html/magento/vendor/magento/framework/App/Area.php(240): Magento\Framework\Translate->loadData(NULL, false)
#5 /var/www/html/magento/vendor/magento/framework/App/Area.php(211): Magento\Framework\App\Area->_initTranslate()
#6 /var/www/html/magento/vendor/magento/framework/App/Area.php(138): Magento\Framework\App\Area->_loadPart('translate')
#7 /var/www/html/magento/vendor/magento/framework/View/DesignLoader.php(55): Magento\Framework\App\Area->load('translate')
#8 /var/www/html/magento/vendor/magento/framework/App/Action/Plugin/Design.php(48): Magento\Framework\View\DesignLoader->load()
#9 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(52): Magento\Framework\App\Action\Plugin\Design->beforeDispatch(Object(Magento\Cms\Controller\Index\Index\Interceptor), Object(Magento\Framework\App\Request\Http))
#10 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Cms\\Con...', 'dispatch', Object(Magento\Cms\Controller\Index\Index\Interceptor), Array, 'customer-app-ac...')
#11 /var/www/html/magento/vendor/magento/module-customer/Model/App/Action/ContextPlugin.php(61): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#12 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\Customer\Model\App\Action\ContextPlugin->aroundDispatch(Object(Magento\Cms\Controller\Index\Index\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#13 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Cms\\Con...', 'dispatch', Object(Magento\Cms\Controller\Index\Index\Interceptor), Array, 'contextPlugin')
#14 /var/www/html/magento/vendor/magento/module-store/App/Action/Plugin/Context.php(106): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#15 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\Store\App\Action\Plugin\Context->aroundDispatch(Object(Magento\Cms\Controller\Index\Index\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#16 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Cms\\Con...', 'dispatch', Object(Magento\Cms\Controller\Index\Index\Interceptor), Array, 'storeCheck')
#17 /var/www/html/magento/vendor/magento/module-store/App/Action/Plugin/StoreCheck.php(44): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#18 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\Store\App\Action\Plugin\StoreCheck->aroundDispatch(Object(Magento\Cms\Controller\Index\Index\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#19 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Cms\\Con...', 'dispatch', Object(Magento\Cms\Controller\Index\Index\Interceptor), Array, 'tax-app-action-...')
#20 /var/www/html/magento/vendor/magento/module-tax/Model/App/Action/ContextPlugin.php(91): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#21 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\Tax\Model\App\Action\ContextPlugin->aroundDispatch(Object(Magento\Cms\Controller\Index\Index\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#22 /var/www/html/magento/vendor/magento/framework/Interception/Interceptor.php(138): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Cms\\Con...', 'dispatch', Object(Magento\Cms\Controller\Index\Index\Interceptor), Array, 'weee-app-action...')
#23 /var/www/html/magento/vendor/magento/module-weee/Model/App/Action/ContextPlugin.php(112): Magento\Cms\Controller\Index\Index\Interceptor->Magento\Framework\Interception\{closure}(Object(Magento\Framework\App\Request\Http))
#24 /var/www/html/magento/vendor/magento/framework/Interception/Interceptor.php(142): Magento\Weee\Model\App\Action\ContextPlugin->aroundDispatch(Object(Magento\Cms\Controller\Index\Index\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#25 /var/www/html/magento/var/generation/Magento/Cms/Controller/Index/Index/Interceptor.php(39): Magento\Cms\Controller\Index\Index\Interceptor->___callPlugins('dispatch', Array, Array)
#26 /var/www/html/magento/vendor/magento/framework/App/FrontController.php(55): Magento\Cms\Controller\Index\Index\Interceptor->dispatch(Object(Magento\Framework\App\Request\Http))
#27 /var/www/html/magento/vendor/magento/framework/Interception/Interceptor.php(74): Magento\Framework\App\FrontController->dispatch(Object(Magento\Framework\App\Request\Http))
#28 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(70): Magento\Framework\App\FrontController\Interceptor->___callParent('dispatch', Array)
#29 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Framewo...', 'dispatch', Object(Magento\Framework\App\FrontController\Interceptor), Array, 'requestPreproce...')
#30 /var/www/html/magento/vendor/magento/module-store/App/FrontController/Plugin/RequestPreprocessor.php(94): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#31 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\Store\App\FrontController\Plugin\RequestPreprocessor->aroundDispatch(Object(Magento\Framework\App\FrontController\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#32 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Framewo...', 'dispatch', Object(Magento\Framework\App\FrontController\Interceptor), Array, 'install')
#33 /var/www/html/magento/vendor/magento/framework/Module/Plugin/DbStatusValidator.php(69): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#34 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\Framework\Module\Plugin\DbStatusValidator->aroundDispatch(Object(Magento\Framework\App\FrontController\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#35 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Framewo...', 'dispatch', Object(Magento\Framework\App\FrontController\Interceptor), Array, 'front-controlle...')
#36 /var/www/html/magento/vendor/magento/module-page-cache/Model/App/FrontController/BuiltinPlugin.php(68): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#37 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\PageCache\Model\App\FrontController\BuiltinPlugin->aroundDispatch(Object(Magento\Framework\App\FrontController\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#38 /var/www/html/magento/vendor/magento/framework/Interception/Interceptor.php(138): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Framewo...', 'dispatch', Object(Magento\Framework\App\FrontController\Interceptor), Array, 'front-controlle...')
#39 /var/www/html/magento/vendor/magento/module-page-cache/Model/App/FrontController/VarnishPlugin.php(55): Magento\Framework\App\FrontController\Interceptor->Magento\Framework\Interception\{closure}(Object(Magento\Framework\App\Request\Http))
#40 /var/www/html/magento/vendor/magento/framework/Interception/Interceptor.php(142): Magento\PageCache\Model\App\FrontController\VarnishPlugin->aroundDispatch(Object(Magento\Framework\App\FrontController\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#41 /var/www/html/magento/var/generation/Magento/Framework/App/FrontController/Interceptor.php(26): Magento\Framework\App\FrontController\Interceptor->___callPlugins('dispatch', Array, Array)
#42 /var/www/html/magento/vendor/magento/framework/App/Http.php(135): Magento\Framework\App\FrontController\Interceptor->dispatch(Object(Magento\Framework\App\Request\Http))
#43 /var/www/html/magento/vendor/magento/framework/App/Bootstrap.php(258): Magento\Framework\App\Http->launch()
#44 /var/www/html/magento/index.php(39): Magento\Framework\App\Bootstrap->run(Object(Magento\Framework\App\Http))
#45 {main}";s:3:"url";s:47:"/?___store=amira_fr&___from_store=amira_landing";s:11:"script_name";s:10:"/index.php";}
```
My imaginaerum installation is in /app/i18n/imaginaerum/fr_fr
|
1.0
|
Error processing your request for french language - Magento gives me this error when I try to open the French Page:
```
a:4:{i:0;s:117:"Notice: Undefined offset: 1 in /var/www/html/magento/vendor/magento/framework/App/Language/Dictionary.php on line 196";i:1;s:9624:"#0 /var/www/html/magento/vendor/magento/framework/App/Language/Dictionary.php(196): Magento\Framework\App\ErrorHandler->handler(8, 'Undefined offse...', '/var/www/html/m...', 196, Array)
#1 /var/www/html/magento/vendor/magento/framework/App/Language/Dictionary.php(113): Magento\Framework\App\Language\Dictionary->readPackCsv('imaginaerum', 'fr_fr')
#2 /var/www/html/magento/vendor/magento/framework/Translate.php(323): Magento\Framework\App\Language\Dictionary->getDictionary('fr_FR')
#3 /var/www/html/magento/vendor/magento/framework/Translate.php(181): Magento\Framework\Translate->_loadPackTranslation()
#4 /var/www/html/magento/vendor/magento/framework/App/Area.php(240): Magento\Framework\Translate->loadData(NULL, false)
#5 /var/www/html/magento/vendor/magento/framework/App/Area.php(211): Magento\Framework\App\Area->_initTranslate()
#6 /var/www/html/magento/vendor/magento/framework/App/Area.php(138): Magento\Framework\App\Area->_loadPart('translate')
#7 /var/www/html/magento/vendor/magento/framework/View/DesignLoader.php(55): Magento\Framework\App\Area->load('translate')
#8 /var/www/html/magento/vendor/magento/framework/App/Action/Plugin/Design.php(48): Magento\Framework\View\DesignLoader->load()
#9 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(52): Magento\Framework\App\Action\Plugin\Design->beforeDispatch(Object(Magento\Cms\Controller\Index\Index\Interceptor), Object(Magento\Framework\App\Request\Http))
#10 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Cms\\Con...', 'dispatch', Object(Magento\Cms\Controller\Index\Index\Interceptor), Array, 'customer-app-ac...')
#11 /var/www/html/magento/vendor/magento/module-customer/Model/App/Action/ContextPlugin.php(61): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#12 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\Customer\Model\App\Action\ContextPlugin->aroundDispatch(Object(Magento\Cms\Controller\Index\Index\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#13 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Cms\\Con...', 'dispatch', Object(Magento\Cms\Controller\Index\Index\Interceptor), Array, 'contextPlugin')
#14 /var/www/html/magento/vendor/magento/module-store/App/Action/Plugin/Context.php(106): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#15 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\Store\App\Action\Plugin\Context->aroundDispatch(Object(Magento\Cms\Controller\Index\Index\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#16 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Cms\\Con...', 'dispatch', Object(Magento\Cms\Controller\Index\Index\Interceptor), Array, 'storeCheck')
#17 /var/www/html/magento/vendor/magento/module-store/App/Action/Plugin/StoreCheck.php(44): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#18 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\Store\App\Action\Plugin\StoreCheck->aroundDispatch(Object(Magento\Cms\Controller\Index\Index\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#19 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Cms\\Con...', 'dispatch', Object(Magento\Cms\Controller\Index\Index\Interceptor), Array, 'tax-app-action-...')
#20 /var/www/html/magento/vendor/magento/module-tax/Model/App/Action/ContextPlugin.php(91): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#21 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\Tax\Model\App\Action\ContextPlugin->aroundDispatch(Object(Magento\Cms\Controller\Index\Index\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#22 /var/www/html/magento/vendor/magento/framework/Interception/Interceptor.php(138): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Cms\\Con...', 'dispatch', Object(Magento\Cms\Controller\Index\Index\Interceptor), Array, 'weee-app-action...')
#23 /var/www/html/magento/vendor/magento/module-weee/Model/App/Action/ContextPlugin.php(112): Magento\Cms\Controller\Index\Index\Interceptor->Magento\Framework\Interception\{closure}(Object(Magento\Framework\App\Request\Http))
#24 /var/www/html/magento/vendor/magento/framework/Interception/Interceptor.php(142): Magento\Weee\Model\App\Action\ContextPlugin->aroundDispatch(Object(Magento\Cms\Controller\Index\Index\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#25 /var/www/html/magento/var/generation/Magento/Cms/Controller/Index/Index/Interceptor.php(39): Magento\Cms\Controller\Index\Index\Interceptor->___callPlugins('dispatch', Array, Array)
#26 /var/www/html/magento/vendor/magento/framework/App/FrontController.php(55): Magento\Cms\Controller\Index\Index\Interceptor->dispatch(Object(Magento\Framework\App\Request\Http))
#27 /var/www/html/magento/vendor/magento/framework/Interception/Interceptor.php(74): Magento\Framework\App\FrontController->dispatch(Object(Magento\Framework\App\Request\Http))
#28 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(70): Magento\Framework\App\FrontController\Interceptor->___callParent('dispatch', Array)
#29 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Framewo...', 'dispatch', Object(Magento\Framework\App\FrontController\Interceptor), Array, 'requestPreproce...')
#30 /var/www/html/magento/vendor/magento/module-store/App/FrontController/Plugin/RequestPreprocessor.php(94): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#31 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\Store\App\FrontController\Plugin\RequestPreprocessor->aroundDispatch(Object(Magento\Framework\App\FrontController\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#32 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Framewo...', 'dispatch', Object(Magento\Framework\App\FrontController\Interceptor), Array, 'install')
#33 /var/www/html/magento/vendor/magento/framework/Module/Plugin/DbStatusValidator.php(69): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#34 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\Framework\Module\Plugin\DbStatusValidator->aroundDispatch(Object(Magento\Framework\App\FrontController\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#35 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(63): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Framewo...', 'dispatch', Object(Magento\Framework\App\FrontController\Interceptor), Array, 'front-controlle...')
#36 /var/www/html/magento/vendor/magento/module-page-cache/Model/App/FrontController/BuiltinPlugin.php(68): Magento\Framework\Interception\Chain\Chain->Magento\Framework\Interception\Chain\{closure}(Object(Magento\Framework\App\Request\Http))
#37 /var/www/html/magento/vendor/magento/framework/Interception/Chain/Chain.php(67): Magento\PageCache\Model\App\FrontController\BuiltinPlugin->aroundDispatch(Object(Magento\Framework\App\FrontController\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#38 /var/www/html/magento/vendor/magento/framework/Interception/Interceptor.php(138): Magento\Framework\Interception\Chain\Chain->invokeNext('Magento\\Framewo...', 'dispatch', Object(Magento\Framework\App\FrontController\Interceptor), Array, 'front-controlle...')
#39 /var/www/html/magento/vendor/magento/module-page-cache/Model/App/FrontController/VarnishPlugin.php(55): Magento\Framework\App\FrontController\Interceptor->Magento\Framework\Interception\{closure}(Object(Magento\Framework\App\Request\Http))
#40 /var/www/html/magento/vendor/magento/framework/Interception/Interceptor.php(142): Magento\PageCache\Model\App\FrontController\VarnishPlugin->aroundDispatch(Object(Magento\Framework\App\FrontController\Interceptor), Object(Closure), Object(Magento\Framework\App\Request\Http))
#41 /var/www/html/magento/var/generation/Magento/Framework/App/FrontController/Interceptor.php(26): Magento\Framework\App\FrontController\Interceptor->___callPlugins('dispatch', Array, Array)
#42 /var/www/html/magento/vendor/magento/framework/App/Http.php(135): Magento\Framework\App\FrontController\Interceptor->dispatch(Object(Magento\Framework\App\Request\Http))
#43 /var/www/html/magento/vendor/magento/framework/App/Bootstrap.php(258): Magento\Framework\App\Http->launch()
#44 /var/www/html/magento/index.php(39): Magento\Framework\App\Bootstrap->run(Object(Magento\Framework\App\Http))
#45 {main}";s:3:"url";s:47:"/?___store=amira_fr&___from_store=amira_landing";s:11:"script_name";s:10:"/index.php";}
```
My imaginaerum installation is in /app/i18n/imaginaerum/fr_fr
|
test
|
error processing your request for french language magento gives me this error when i try to open the french page a i s notice undefined offset in var www html magento vendor magento framework app language dictionary php on line i s var www html magento vendor magento framework app language dictionary php magento framework app errorhandler handler undefined offse var www html m array var www html magento vendor magento framework app language dictionary php magento framework app language dictionary readpackcsv imaginaerum fr fr var www html magento vendor magento framework translate php magento framework app language dictionary getdictionary fr fr var www html magento vendor magento framework translate php magento framework translate loadpacktranslation var www html magento vendor magento framework app area php magento framework translate loaddata null false var www html magento vendor magento framework app area php magento framework app area inittranslate var www html magento vendor magento framework app area php magento framework app area loadpart translate var www html magento vendor magento framework view designloader php magento framework app area load translate var www html magento vendor magento framework app action plugin design php magento framework view designloader load var www html magento vendor magento framework interception chain chain php magento framework app action plugin design beforedispatch object magento cms controller index index interceptor object magento framework app request http var www html magento vendor magento framework interception chain chain php magento framework interception chain chain invokenext magento cms con dispatch object magento cms controller index index interceptor array customer app ac var www html magento vendor magento module customer model app action contextplugin php magento framework interception chain chain magento framework interception chain closure object magento framework app request http var www html magento vendor magento framework interception chain chain php magento customer model app action contextplugin arounddispatch object magento cms controller index index interceptor object closure object magento framework app request http var www html magento vendor magento framework interception chain chain php magento framework interception chain chain invokenext magento cms con dispatch object magento cms controller index index interceptor array contextplugin var www html magento vendor magento module store app action plugin context php magento framework interception chain chain magento framework interception chain closure object magento framework app request http var www html magento vendor magento framework interception chain chain php magento store app action plugin context arounddispatch object magento cms controller index index interceptor object closure object magento framework app request http var www html magento vendor magento framework interception chain chain php magento framework interception chain chain invokenext magento cms con dispatch object magento cms controller index index interceptor array storecheck var www html magento vendor magento module store app action plugin storecheck php magento framework interception chain chain magento framework interception chain closure object magento framework app request http var www html magento vendor magento framework interception chain chain php magento store app action plugin storecheck arounddispatch object magento cms controller index index interceptor object closure object magento framework app request http var www html magento vendor magento framework interception chain chain php magento framework interception chain chain invokenext magento cms con dispatch object magento cms controller index index interceptor array tax app action var www html magento vendor magento module tax model app action contextplugin php magento framework interception chain chain magento framework interception chain closure object magento framework app request http var www html magento vendor magento framework interception chain chain php magento tax model app action contextplugin arounddispatch object magento cms controller index index interceptor object closure object magento framework app request http var www html magento vendor magento framework interception interceptor php magento framework interception chain chain invokenext magento cms con dispatch object magento cms controller index index interceptor array weee app action var www html magento vendor magento module weee model app action contextplugin php magento cms controller index index interceptor magento framework interception closure object magento framework app request http var www html magento vendor magento framework interception interceptor php magento weee model app action contextplugin arounddispatch object magento cms controller index index interceptor object closure object magento framework app request http var www html magento var generation magento cms controller index index interceptor php magento cms controller index index interceptor callplugins dispatch array array var www html magento vendor magento framework app frontcontroller php magento cms controller index index interceptor dispatch object magento framework app request http var www html magento vendor magento framework interception interceptor php magento framework app frontcontroller dispatch object magento framework app request http var www html magento vendor magento framework interception chain chain php magento framework app frontcontroller interceptor callparent dispatch array var www html magento vendor magento framework interception chain chain php magento framework interception chain chain invokenext magento framewo dispatch object magento framework app frontcontroller interceptor array requestpreproce var www html magento vendor magento module store app frontcontroller plugin requestpreprocessor php magento framework interception chain chain magento framework interception chain closure object magento framework app request http var www html magento vendor magento framework interception chain chain php magento store app frontcontroller plugin requestpreprocessor arounddispatch object magento framework app frontcontroller interceptor object closure object magento framework app request http var www html magento vendor magento framework interception chain chain php magento framework interception chain chain invokenext magento framewo dispatch object magento framework app frontcontroller interceptor array install var www html magento vendor magento framework module plugin dbstatusvalidator php magento framework interception chain chain magento framework interception chain closure object magento framework app request http var www html magento vendor magento framework interception chain chain php magento framework module plugin dbstatusvalidator arounddispatch object magento framework app frontcontroller interceptor object closure object magento framework app request http var www html magento vendor magento framework interception chain chain php magento framework interception chain chain invokenext magento framewo dispatch object magento framework app frontcontroller interceptor array front controlle var www html magento vendor magento module page cache model app frontcontroller builtinplugin php magento framework interception chain chain magento framework interception chain closure object magento framework app request http var www html magento vendor magento framework interception chain chain php magento pagecache model app frontcontroller builtinplugin arounddispatch object magento framework app frontcontroller interceptor object closure object magento framework app request http var www html magento vendor magento framework interception interceptor php magento framework interception chain chain invokenext magento framewo dispatch object magento framework app frontcontroller interceptor array front controlle var www html magento vendor magento module page cache model app frontcontroller varnishplugin php magento framework app frontcontroller interceptor magento framework interception closure object magento framework app request http var www html magento vendor magento framework interception interceptor php magento pagecache model app frontcontroller varnishplugin arounddispatch object magento framework app frontcontroller interceptor object closure object magento framework app request http var www html magento var generation magento framework app frontcontroller interceptor php magento framework app frontcontroller interceptor callplugins dispatch array array var www html magento vendor magento framework app http php magento framework app frontcontroller interceptor dispatch object magento framework app request http var www html magento vendor magento framework app bootstrap php magento framework app http launch var www html magento index php magento framework app bootstrap run object magento framework app http main s url s store amira fr from store amira landing s script name s index php my imaginaerum installation is in app imaginaerum fr fr
| 1
|
99,755
| 8,710,267,829
|
IssuesEvent
|
2018-12-06 16:01:06
|
italia/spid
|
https://api.github.com/repos/italia/spid
|
closed
|
Verifica metadata Comune di Borgo San Dalmazzo
|
aggiornamento md test metadata
|
Buongiorno,
Per conto del Comune di Borgo San Dalmazzo
abbiamo predisposto i metadata e pubblicati all'URL
https://borgosandalmazzo.multeonline.it/serviziSPID/metadata.xml
i metadata sono stati aggiornati con l'aggiunta di un secondo SP
[metadata_borgosandalmazzo2SP_new-signed.zip](https://github.com/italia/spid/files/2640432/metadata_borgosandalmazzo2SP_new-signed.zip)
cordiali saluti
Facondini Stefano
Maggioli spa
|
1.0
|
Verifica metadata Comune di Borgo San Dalmazzo - Buongiorno,
Per conto del Comune di Borgo San Dalmazzo
abbiamo predisposto i metadata e pubblicati all'URL
https://borgosandalmazzo.multeonline.it/serviziSPID/metadata.xml
i metadata sono stati aggiornati con l'aggiunta di un secondo SP
[metadata_borgosandalmazzo2SP_new-signed.zip](https://github.com/italia/spid/files/2640432/metadata_borgosandalmazzo2SP_new-signed.zip)
cordiali saluti
Facondini Stefano
Maggioli spa
|
test
|
verifica metadata comune di borgo san dalmazzo buongiorno per conto del comune di borgo san dalmazzo abbiamo predisposto i metadata e pubblicati all url i metadata sono stati aggiornati con l aggiunta di un secondo sp cordiali saluti facondini stefano maggioli spa
| 1
|
798,982
| 28,300,496,984
|
IssuesEvent
|
2023-04-10 05:22:18
|
googleapis/google-cloud-ruby
|
https://api.github.com/repos/googleapis/google-cloud-ruby
|
closed
|
[Nightly CI Failures] Failures detected for google-cloud-recommender-v1
|
type: bug priority: p1 nightly failure
|
At 2023-04-09 08:55:45 UTC, detected failures in google-cloud-recommender-v1 for: yard
report_key_b203859c45be229c6bf93346666a7a3d
|
1.0
|
[Nightly CI Failures] Failures detected for google-cloud-recommender-v1 - At 2023-04-09 08:55:45 UTC, detected failures in google-cloud-recommender-v1 for: yard
report_key_b203859c45be229c6bf93346666a7a3d
|
non_test
|
failures detected for google cloud recommender at utc detected failures in google cloud recommender for yard report key
| 0
|
7,831
| 2,859,235,834
|
IssuesEvent
|
2015-06-03 09:23:34
|
osakagamba/7GIQSKRNE5P3AVTZCEFXE3ON
|
https://api.github.com/repos/osakagamba/7GIQSKRNE5P3AVTZCEFXE3ON
|
closed
|
iZSlgSTuW3FRMA9yz7UXijsUSG/KRJ6/TVETrnWvWUkNJzkswUyiLfPFrh7VbrbEj5sfsxeTvGsa/Oa8GxNpaign7A6Coz5ae14rv+tWWi52l48VXNkrnOFkk5L8REpy9fwCGnYh2XkPvm5xC9IyvdajES8up8FlU+syhnN/c8U=
|
design
|
rBQKgP73P4u5KZfKwS2soWfb5zGicuy3cNM4BHOTJVjoAO1IsSYD3oY4iWeGJ84jizX6GxHehwIDrTpQos+3GWWWoZyjSr3PIj/7qkvlXXjB5CtDmfjlQhuYpmcaggw8NvWiRDK/VCUERMEkok9+FWQ/AyjaaAbcXtEXZlKgkG/DR6dREmQH8G+QfQXUI4DWi/cxzTH7FHVi1a1M/RPF+wAmw/CENeKHQMO9MrC3rlObDebNufa95slconUREJ/oetQttr5qw8C2Esfnxc06ea5rXuv2jLlZCpP3zr3oatTyfGeEhMrlklHgrZTi9YT7MtmR46jPMUchfzgFzakK07eXg7GdlhODfEIRGJ/v07jyfGeEhMrlklHgrZTi9YT7sFKjrrMZumu6Vewhs+BszVfWa513Cc60EIyq+CgYmbLXHkvPdwP6vCJnnwZCubw6EiY8hKF2e7ceByi9M1MzZlrfJsNA+9QNJ+sNIt4sVS0JGfp1E3qwXj9joQvOLHgeq6QzSYupudJ8EibtLmk3otxfLeG1FWpTAY5ykK6FQ8t82AQEORNRllbkdl6xdDFaG1pBnadiL+9Gjd0CZ5anTD3f2umiN1EYJTMlcR9LRlf3TzFrpiqI2EflbMhwxwIuLpob413lij8cg+STgQNE8JuU2eS1uHgYQOxc9/6Vg+WAw7tvgQOyfUUw1/vP6I9efvFHmNXpvzRe9iNZWSP/JbViigMj5KTO8vQFn09lvixVGyymPlLJvwhdURafkXXJ7sM/SLZ9cLGCZ58kQ4gMon9BPZPeDecI7bfntqZXdEIJ+n8BvMWfyOG3s5k0J39YVxv4F9ZXvTp76HtqdeWDrg==
|
1.0
|
iZSlgSTuW3FRMA9yz7UXijsUSG/KRJ6/TVETrnWvWUkNJzkswUyiLfPFrh7VbrbEj5sfsxeTvGsa/Oa8GxNpaign7A6Coz5ae14rv+tWWi52l48VXNkrnOFkk5L8REpy9fwCGnYh2XkPvm5xC9IyvdajES8up8FlU+syhnN/c8U= - rBQKgP73P4u5KZfKwS2soWfb5zGicuy3cNM4BHOTJVjoAO1IsSYD3oY4iWeGJ84jizX6GxHehwIDrTpQos+3GWWWoZyjSr3PIj/7qkvlXXjB5CtDmfjlQhuYpmcaggw8NvWiRDK/VCUERMEkok9+FWQ/AyjaaAbcXtEXZlKgkG/DR6dREmQH8G+QfQXUI4DWi/cxzTH7FHVi1a1M/RPF+wAmw/CENeKHQMO9MrC3rlObDebNufa95slconUREJ/oetQttr5qw8C2Esfnxc06ea5rXuv2jLlZCpP3zr3oatTyfGeEhMrlklHgrZTi9YT7MtmR46jPMUchfzgFzakK07eXg7GdlhODfEIRGJ/v07jyfGeEhMrlklHgrZTi9YT7sFKjrrMZumu6Vewhs+BszVfWa513Cc60EIyq+CgYmbLXHkvPdwP6vCJnnwZCubw6EiY8hKF2e7ceByi9M1MzZlrfJsNA+9QNJ+sNIt4sVS0JGfp1E3qwXj9joQvOLHgeq6QzSYupudJ8EibtLmk3otxfLeG1FWpTAY5ykK6FQ8t82AQEORNRllbkdl6xdDFaG1pBnadiL+9Gjd0CZ5anTD3f2umiN1EYJTMlcR9LRlf3TzFrpiqI2EflbMhwxwIuLpob413lij8cg+STgQNE8JuU2eS1uHgYQOxc9/6Vg+WAw7tvgQOyfUUw1/vP6I9efvFHmNXpvzRe9iNZWSP/JbViigMj5KTO8vQFn09lvixVGyymPlLJvwhdURafkXXJ7sM/SLZ9cLGCZ58kQ4gMon9BPZPeDecI7bfntqZXdEIJ+n8BvMWfyOG3s5k0J39YVxv4F9ZXvTp76HtqdeWDrg==
|
non_test
|
syhnn fwq ayjaaabcxtexzlkgkg rpf wamw
| 0
|
204,764
| 15,530,955,603
|
IssuesEvent
|
2021-03-13 21:16:37
|
isontheline/pro.webssh.net
|
https://api.github.com/repos/isontheline/pro.webssh.net
|
closed
|
SFTP : can’t refresh if short files list
|
bug testflight-release-needed
|
**Describe the feature**
If the files list doesn’t exceed the screen height size then we can’t use the pull to refresh.
|
1.0
|
SFTP : can’t refresh if short files list - **Describe the feature**
If the files list doesn’t exceed the screen height size then we can’t use the pull to refresh.
|
test
|
sftp can’t refresh if short files list describe the feature if the files list doesn’t exceed the screen height size then we can’t use the pull to refresh
| 1
|
52,630
| 22,326,842,527
|
IssuesEvent
|
2022-06-14 11:25:15
|
kyma-project/istio-operator
|
https://api.github.com/repos/kyma-project/istio-operator
|
closed
|
[POC] First implementation of kyma istio operator
|
area/service-mesh area/installation
|
<!-- Thank you for your contribution. Before you submit the issue:
1. Search open and closed issues for duplicates.
2. Read the contributing guidelines.
-->
**Description**
Implement kyma istio operator managaing kyma istio component. Kyma istio operator should be based on k8s operator watching kyma istio operator CR with limited options to configure (we don't want to expose full conf). Operator should be able to manage many istio installations on many clusters. Operator should be able install, update and reset to kyma default istio component.
Scope of this POC is figure out best installation method.
TODO:
- [x] investigate istio-operator -> timebox 2d (@veichtj)
- [x] investigate helm charts installation -> timebox 2d (@dariusztutaj)
- [ ] investigate how to migrate to new installation method avoiding down time
- [ ] investigate versioning and releasing (modularisation context)
- [ ] investigate performance of proposed solution
**Reasons**
Modularisation of kyma components
|
1.0
|
[POC] First implementation of kyma istio operator - <!-- Thank you for your contribution. Before you submit the issue:
1. Search open and closed issues for duplicates.
2. Read the contributing guidelines.
-->
**Description**
Implement kyma istio operator managaing kyma istio component. Kyma istio operator should be based on k8s operator watching kyma istio operator CR with limited options to configure (we don't want to expose full conf). Operator should be able to manage many istio installations on many clusters. Operator should be able install, update and reset to kyma default istio component.
Scope of this POC is figure out best installation method.
TODO:
- [x] investigate istio-operator -> timebox 2d (@veichtj)
- [x] investigate helm charts installation -> timebox 2d (@dariusztutaj)
- [ ] investigate how to migrate to new installation method avoiding down time
- [ ] investigate versioning and releasing (modularisation context)
- [ ] investigate performance of proposed solution
**Reasons**
Modularisation of kyma components
|
non_test
|
first implementation of kyma istio operator thank you for your contribution before you submit the issue search open and closed issues for duplicates read the contributing guidelines description implement kyma istio operator managaing kyma istio component kyma istio operator should be based on operator watching kyma istio operator cr with limited options to configure we don t want to expose full conf operator should be able to manage many istio installations on many clusters operator should be able install update and reset to kyma default istio component scope of this poc is figure out best installation method todo investigate istio operator timebox veichtj investigate helm charts installation timebox dariusztutaj investigate how to migrate to new installation method avoiding down time investigate versioning and releasing modularisation context investigate performance of proposed solution reasons modularisation of kyma components
| 0
|
202,816
| 15,302,366,661
|
IssuesEvent
|
2021-02-24 14:39:59
|
infinispan/infinispan-operator
|
https://api.github.com/repos/infinispan/infinispan-operator
|
closed
|
CI test didn't stop execute tests if first step is fails
|
test
|
There are two portions of tests are running on Travis CI now with
* RUN_SA_OPERATOR=TRUE make test PARALLEL_COUNT=2
* make multinamespace-test PARALLEL_COUNT=2
even if first test is failed, second continuing to execute instead of stop CI job execution with fail state
This is default TravisCI behavior and should be changed https://docs.travis-ci.com/user/job-lifecycle/#customizing-the-build-phase
|
1.0
|
CI test didn't stop execute tests if first step is fails - There are two portions of tests are running on Travis CI now with
* RUN_SA_OPERATOR=TRUE make test PARALLEL_COUNT=2
* make multinamespace-test PARALLEL_COUNT=2
even if first test is failed, second continuing to execute instead of stop CI job execution with fail state
This is default TravisCI behavior and should be changed https://docs.travis-ci.com/user/job-lifecycle/#customizing-the-build-phase
|
test
|
ci test didn t stop execute tests if first step is fails there are two portions of tests are running on travis ci now with run sa operator true make test parallel count make multinamespace test parallel count even if first test is failed second continuing to execute instead of stop ci job execution with fail state this is default travisci behavior and should be changed
| 1
|
180,224
| 13,926,744,788
|
IssuesEvent
|
2020-10-21 18:44:55
|
department-of-veterans-affairs/va.gov-cms
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
|
opened
|
Spike: E2E testing coverage
|
Automated testing
|
## Description
The CMS team is interested in having E2E test coverage to ensure consistent function of our platform.
The Facilities team has E2E testing on their roadmap for Q4. As we discuss the CMS strategy for E2E, we need to consider whether and how we may combine our E2E efforts w/ that of the Facilities team and others and flag any potential points of integration / risks that may arise.
## Acceptance Criteria
- [ ] define the meaning of E2E tests for the CMS team
- [ ] initial brainstorm tools/testing frameworks that we may want to start using and use cases
- [ ] flag any points of overlap / risks when integrating w/ other team's testing suites
- [ ] generate initial backlog if we're at a point in a conversation where we know what next steps we need to take in E2E initiative
participants:
@ElijahLynn @indytechcook @acrollet @swirtSJW @oksana-c [please add yourself here if interested]
|
1.0
|
Spike: E2E testing coverage - ## Description
The CMS team is interested in having E2E test coverage to ensure consistent function of our platform.
The Facilities team has E2E testing on their roadmap for Q4. As we discuss the CMS strategy for E2E, we need to consider whether and how we may combine our E2E efforts w/ that of the Facilities team and others and flag any potential points of integration / risks that may arise.
## Acceptance Criteria
- [ ] define the meaning of E2E tests for the CMS team
- [ ] initial brainstorm tools/testing frameworks that we may want to start using and use cases
- [ ] flag any points of overlap / risks when integrating w/ other team's testing suites
- [ ] generate initial backlog if we're at a point in a conversation where we know what next steps we need to take in E2E initiative
participants:
@ElijahLynn @indytechcook @acrollet @swirtSJW @oksana-c [please add yourself here if interested]
|
test
|
spike testing coverage description the cms team is interested in having test coverage to ensure consistent function of our platform the facilities team has testing on their roadmap for as we discuss the cms strategy for we need to consider whether and how we may combine our efforts w that of the facilities team and others and flag any potential points of integration risks that may arise acceptance criteria define the meaning of tests for the cms team initial brainstorm tools testing frameworks that we may want to start using and use cases flag any points of overlap risks when integrating w other team s testing suites generate initial backlog if we re at a point in a conversation where we know what next steps we need to take in initiative participants elijahlynn indytechcook acrollet swirtsjw oksana c
| 1
|
316,249
| 27,148,437,028
|
IssuesEvent
|
2023-02-16 22:09:16
|
Iridescent-CM/technovation-app
|
https://api.github.com/repos/Iridescent-CM/technovation-app
|
opened
|
Address location tests
|
testing
|
The following tests should be updated/fixed, or removed:
```
Saving a location in a geopolitically sensitive area prompts the user with a choice when Israel is the country
# Temporarily skipped with xit
# ./spec/system/location/sensitive_locations_spec.rb:17
Saving a location in a geopolitically sensitive area saves the changes in the database
# Temporarily skipped with xit
# ./spec/system/location/sensitive_locations_spec.rb:25
```
|
1.0
|
Address location tests - The following tests should be updated/fixed, or removed:
```
Saving a location in a geopolitically sensitive area prompts the user with a choice when Israel is the country
# Temporarily skipped with xit
# ./spec/system/location/sensitive_locations_spec.rb:17
Saving a location in a geopolitically sensitive area saves the changes in the database
# Temporarily skipped with xit
# ./spec/system/location/sensitive_locations_spec.rb:25
```
|
test
|
address location tests the following tests should be updated fixed or removed saving a location in a geopolitically sensitive area prompts the user with a choice when israel is the country temporarily skipped with xit spec system location sensitive locations spec rb saving a location in a geopolitically sensitive area saves the changes in the database temporarily skipped with xit spec system location sensitive locations spec rb
| 1
|
28,909
| 4,445,991,134
|
IssuesEvent
|
2016-08-20 11:31:21
|
herculeshssj/orcamento
|
https://api.github.com/repos/herculeshssj/orcamento
|
opened
|
Fechar fatura tornando a fatura vencida
|
Teste
|
Verificar se houve clique duplo por parte do usuário, ou realmente está ocorrendo o fechamento e posterior mudança para o status de vencida da fatura.
|
1.0
|
Fechar fatura tornando a fatura vencida - Verificar se houve clique duplo por parte do usuário, ou realmente está ocorrendo o fechamento e posterior mudança para o status de vencida da fatura.
|
test
|
fechar fatura tornando a fatura vencida verificar se houve clique duplo por parte do usuário ou realmente está ocorrendo o fechamento e posterior mudança para o status de vencida da fatura
| 1
|
177,077
| 13,683,012,424
|
IssuesEvent
|
2020-09-30 00:29:16
|
MiqueasAmorim/Pedido
|
https://api.github.com/repos/MiqueasAmorim/Pedido
|
closed
|
CT01 (ProdutoTest) - O valor unitário do produto não pode ser 0
|
test case
|
**Dados de entrada:**
- Nome : Borracha
- Valor unitário: 0
- Quantidade: 10
**Resultado esperado:**
- RuntimeException: "Valor inválido: 0.0"
|
1.0
|
CT01 (ProdutoTest) - O valor unitário do produto não pode ser 0 - **Dados de entrada:**
- Nome : Borracha
- Valor unitário: 0
- Quantidade: 10
**Resultado esperado:**
- RuntimeException: "Valor inválido: 0.0"
|
test
|
produtotest o valor unitário do produto não pode ser dados de entrada nome borracha valor unitário quantidade resultado esperado runtimeexception valor inválido
| 1
|
191,812
| 14,596,491,044
|
IssuesEvent
|
2020-12-20 16:04:07
|
github-vet/rangeloop-pointer-findings
|
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
|
closed
|
hello-mr-code/terraform-oci: oci/apigateway_api_test.go; 16 LoC
|
fresh small test
|
Found a possible issue in [hello-mr-code/terraform-oci](https://www.github.com/hello-mr-code/terraform-oci) at [oci/apigateway_api_test.go](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/apigateway_api_test.go#L277-L292)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> reference to apiId is reassigned at line 281
[Click here to see the code in its original context.](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/apigateway_api_test.go#L277-L292)
<details>
<summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary>
```go
for _, apiId := range apiIds {
if ok := SweeperDefaultResourceId[apiId]; !ok {
deleteApiRequest := oci_apigateway.DeleteApiRequest{}
deleteApiRequest.ApiId = &apiId
deleteApiRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "apigateway")
_, error := apiGatewayClient.DeleteApi(context.Background(), deleteApiRequest)
if error != nil {
fmt.Printf("Error deleting Api %s %s, It is possible that the resource is already deleted. Please verify manually \n", apiId, error)
continue
}
waitTillCondition(testAccProvider, &apiId, apiSweepWaitCondition, time.Duration(3*time.Minute),
apiSweepResponseFetchOperation, "apigateway", true)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 2f6aa93ef8643328af454512a5fe78ab006697f0
|
1.0
|
hello-mr-code/terraform-oci: oci/apigateway_api_test.go; 16 LoC -
Found a possible issue in [hello-mr-code/terraform-oci](https://www.github.com/hello-mr-code/terraform-oci) at [oci/apigateway_api_test.go](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/apigateway_api_test.go#L277-L292)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> reference to apiId is reassigned at line 281
[Click here to see the code in its original context.](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/apigateway_api_test.go#L277-L292)
<details>
<summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary>
```go
for _, apiId := range apiIds {
if ok := SweeperDefaultResourceId[apiId]; !ok {
deleteApiRequest := oci_apigateway.DeleteApiRequest{}
deleteApiRequest.ApiId = &apiId
deleteApiRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "apigateway")
_, error := apiGatewayClient.DeleteApi(context.Background(), deleteApiRequest)
if error != nil {
fmt.Printf("Error deleting Api %s %s, It is possible that the resource is already deleted. Please verify manually \n", apiId, error)
continue
}
waitTillCondition(testAccProvider, &apiId, apiSweepWaitCondition, time.Duration(3*time.Minute),
apiSweepResponseFetchOperation, "apigateway", true)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 2f6aa93ef8643328af454512a5fe78ab006697f0
|
test
|
hello mr code terraform oci oci apigateway api test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to apiid is reassigned at line click here to show the line s of go which triggered the analyzer go for apiid range apiids if ok sweeperdefaultresourceid ok deleteapirequest oci apigateway deleteapirequest deleteapirequest apiid apiid deleteapirequest requestmetadata retrypolicy getretrypolicy true apigateway error apigatewayclient deleteapi context background deleteapirequest if error nil fmt printf error deleting api s s it is possible that the resource is already deleted please verify manually n apiid error continue waittillcondition testaccprovider apiid apisweepwaitcondition time duration time minute apisweepresponsefetchoperation apigateway true leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 1
|
177,984
| 14,656,908,385
|
IssuesEvent
|
2020-12-28 14:27:51
|
beingmeta/kno
|
https://api.github.com/repos/beingmeta/kno
|
closed
|
PRINTOUT documentation
|
documentation
|
Describe basic model, use of return values and side-effects. Custom variants.
|
1.0
|
PRINTOUT documentation - Describe basic model, use of return values and side-effects. Custom variants.
|
non_test
|
printout documentation describe basic model use of return values and side effects custom variants
| 0
|
278,854
| 30,702,412,739
|
IssuesEvent
|
2023-07-27 01:28:01
|
pazhanivel07/linux_4.1.15
|
https://api.github.com/repos/pazhanivel07/linux_4.1.15
|
closed
|
CVE-2017-14156 (Medium) detected in linuxlinux-4.6 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2017-14156 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/linux_4.1.15/commit/2d5b61684c3e6642174b25f3a96ed7dd8bf327e3">2d5b61684c3e6642174b25f3a96ed7dd8bf327e3</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/video/fbdev/aty/atyfb_base.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The atyfb_ioctl function in drivers/video/fbdev/aty/atyfb_base.c in the Linux kernel through 4.12.10 does not initialize a certain data structure, which allows local users to obtain sensitive information from kernel stack memory by reading locations associated with padding bytes.
<p>Publish Date: 2017-09-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-14156>CVE-2017-14156</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-14156">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-14156</a></p>
<p>Release Date: 2017-09-05</p>
<p>Fix Resolution: v4.14-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-14156 (Medium) detected in linuxlinux-4.6 - autoclosed - ## CVE-2017-14156 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/linux_4.1.15/commit/2d5b61684c3e6642174b25f3a96ed7dd8bf327e3">2d5b61684c3e6642174b25f3a96ed7dd8bf327e3</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/video/fbdev/aty/atyfb_base.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The atyfb_ioctl function in drivers/video/fbdev/aty/atyfb_base.c in the Linux kernel through 4.12.10 does not initialize a certain data structure, which allows local users to obtain sensitive information from kernel stack memory by reading locations associated with padding bytes.
<p>Publish Date: 2017-09-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-14156>CVE-2017-14156</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-14156">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-14156</a></p>
<p>Release Date: 2017-09-05</p>
<p>Fix Resolution: v4.14-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers video fbdev aty atyfb base c vulnerability details the atyfb ioctl function in drivers video fbdev aty atyfb base c in the linux kernel through does not initialize a certain data structure which allows local users to obtain sensitive information from kernel stack memory by reading locations associated with padding bytes publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
13,514
| 5,392,625,845
|
IssuesEvent
|
2017-02-26 13:03:13
|
junit-team/junit5
|
https://api.github.com/repos/junit-team/junit5
|
closed
|
Upgrade to Gradle 3.4
|
build enhancement up-for-grabs
|
## Overview
The build fails after upgrading to Gradle 3.4 with following stacktrace:
```
...
:junit-platform-gradle-plugin:spotlessCheck
:junit-platform-gradle-plugin:compileTestJava NO-SOURCE
:junit-platform-gradle-plugin:compileTestGroovy
:junit-platform-gradle-plugin:processTestResources NO-SOURCE
:junit-platform-gradle-plugin:testClasses
:junit-platform-gradle-plugin:test
org.junit.platform.gradle.plugin.JUnitPlatformPluginSpec > adds dependencies to configuration FAILED
java.lang.UnsupportedOperationException
at org.gradle.api.internal.CompositeDomainObjectSet$DomainObjectCompositeCollection.containsAll(CompositeDomainObjectSet.java:228)
at org.gradle.api.internal.DefaultDomainObjectCollection.containsAll(DefaultDomainObjectCollection.java:241)
at org.gradle.api.internal.DelegatingDomainObjectSet.containsAll(DelegatingDomainObjectSet.java:98)
at org.gradle.api.internal.DelegatingDomainObjectSet.containsAll(DelegatingDomainObjectSet.java:98)
at org.junit.platform.gradle.plugin.JUnitPlatformPluginSpec.adds dependencies to configuration(JUnitPlatformPluginSpec.groovy:251)
10 tests completed, 1 failed
:junit-platform-gradle-plugin:test FAILED
```
## Deliverables
- [ ] Solve `UnsupportedOperationException`
- ~Make use of https://docs.gradle.org/current/userguide/java_library_plugin.html~ Moved to #691
|
1.0
|
Upgrade to Gradle 3.4 - ## Overview
The build fails after upgrading to Gradle 3.4 with following stacktrace:
```
...
:junit-platform-gradle-plugin:spotlessCheck
:junit-platform-gradle-plugin:compileTestJava NO-SOURCE
:junit-platform-gradle-plugin:compileTestGroovy
:junit-platform-gradle-plugin:processTestResources NO-SOURCE
:junit-platform-gradle-plugin:testClasses
:junit-platform-gradle-plugin:test
org.junit.platform.gradle.plugin.JUnitPlatformPluginSpec > adds dependencies to configuration FAILED
java.lang.UnsupportedOperationException
at org.gradle.api.internal.CompositeDomainObjectSet$DomainObjectCompositeCollection.containsAll(CompositeDomainObjectSet.java:228)
at org.gradle.api.internal.DefaultDomainObjectCollection.containsAll(DefaultDomainObjectCollection.java:241)
at org.gradle.api.internal.DelegatingDomainObjectSet.containsAll(DelegatingDomainObjectSet.java:98)
at org.gradle.api.internal.DelegatingDomainObjectSet.containsAll(DelegatingDomainObjectSet.java:98)
at org.junit.platform.gradle.plugin.JUnitPlatformPluginSpec.adds dependencies to configuration(JUnitPlatformPluginSpec.groovy:251)
10 tests completed, 1 failed
:junit-platform-gradle-plugin:test FAILED
```
## Deliverables
- [ ] Solve `UnsupportedOperationException`
- ~Make use of https://docs.gradle.org/current/userguide/java_library_plugin.html~ Moved to #691
|
non_test
|
upgrade to gradle overview the build fails after upgrading to gradle with following stacktrace junit platform gradle plugin spotlesscheck junit platform gradle plugin compiletestjava no source junit platform gradle plugin compiletestgroovy junit platform gradle plugin processtestresources no source junit platform gradle plugin testclasses junit platform gradle plugin test org junit platform gradle plugin junitplatformpluginspec adds dependencies to configuration failed java lang unsupportedoperationexception at org gradle api internal compositedomainobjectset domainobjectcompositecollection containsall compositedomainobjectset java at org gradle api internal defaultdomainobjectcollection containsall defaultdomainobjectcollection java at org gradle api internal delegatingdomainobjectset containsall delegatingdomainobjectset java at org gradle api internal delegatingdomainobjectset containsall delegatingdomainobjectset java at org junit platform gradle plugin junitplatformpluginspec adds dependencies to configuration junitplatformpluginspec groovy tests completed failed junit platform gradle plugin test failed deliverables solve unsupportedoperationexception make use of moved to
| 0
|
39,883
| 9,726,248,643
|
IssuesEvent
|
2019-05-30 10:56:28
|
primefaces/primereact
|
https://api.github.com/repos/primefaces/primereact
|
closed
|
DataTable expanded rows collapse when modifying one property of a record
|
defect
|
**I'm submitting a ...**
```
[X] bug report
[ ] feature request
[ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57
``
**Current behavior**
If you are using the row expansion feature, and saving the expanded row data in the state (as demonstrated here https://www.primefaces.org/primereact/#/datatable/rowexpand), all of the expanded rows will collapse if any of the records' fields are changed when passing them in from Redux.
It seems like the table loses track of the mapping between the expanded row data and the modified record data being passed in as a prop to 'values'.
**Expected behavior**
Expanded rows should not collapse when the records passed in to the table are not new, just one or more fields modified.
**Minimal reproduction of the problem with instructions**
1) Setup Prime React DataTable with expander column. Records should be provided from redux state slice.
2) expand row
3) have a button in the expanded row update a field on one of the records through a redux action/reducer.
4) You'll notice the row collapses, despite the expandedRowData still being present.
https://codesandbox.io/s/m752k9jq8y
**Please tell us about your environment:**
Windows/IntelliJ/Tomcat
* **React version:**
16.8.3
* **PrimeReact version:**
2.0.1
* **Browser:**
ALL
* **Language:**
ES6/7
|
1.0
|
DataTable expanded rows collapse when modifying one property of a record - **I'm submitting a ...**
```
[X] bug report
[ ] feature request
[ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57
``
**Current behavior**
If you are using the row expansion feature, and saving the expanded row data in the state (as demonstrated here https://www.primefaces.org/primereact/#/datatable/rowexpand), all of the expanded rows will collapse if any of the records' fields are changed when passing them in from Redux.
It seems like the table loses track of the mapping between the expanded row data and the modified record data being passed in as a prop to 'values'.
**Expected behavior**
Expanded rows should not collapse when the records passed in to the table are not new, just one or more fields modified.
**Minimal reproduction of the problem with instructions**
1) Setup Prime React DataTable with expander column. Records should be provided from redux state slice.
2) expand row
3) have a button in the expanded row update a field on one of the records through a redux action/reducer.
4) You'll notice the row collapses, despite the expandedRowData still being present.
https://codesandbox.io/s/m752k9jq8y
**Please tell us about your environment:**
Windows/IntelliJ/Tomcat
* **React version:**
16.8.3
* **PrimeReact version:**
2.0.1
* **Browser:**
ALL
* **Language:**
ES6/7
|
non_test
|
datatable expanded rows collapse when modifying one property of a record i m submitting a bug report feature request support request please do not submit support request here instead see current behavior if you are using the row expansion feature and saving the expanded row data in the state as demonstrated here all of the expanded rows will collapse if any of the records fields are changed when passing them in from redux it seems like the table loses track of the mapping between the expanded row data and the modified record data being passed in as a prop to values expected behavior expanded rows should not collapse when the records passed in to the table are not new just one or more fields modified minimal reproduction of the problem with instructions setup prime react datatable with expander column records should be provided from redux state slice expand row have a button in the expanded row update a field on one of the records through a redux action reducer you ll notice the row collapses despite the expandedrowdata still being present please tell us about your environment windows intellij tomcat react version primereact version browser all language
| 0
|
279,696
| 30,730,680,103
|
IssuesEvent
|
2023-07-28 01:06:06
|
Gal-Doron/Baragon-test-4
|
https://api.github.com/repos/Gal-Doron/Baragon-test-4
|
opened
|
aws-java-sdk-core-1.11.497.jar: 3 vulnerabilities (highest severity is: 7.5)
|
Mend: dependency security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>aws-java-sdk-core-1.11.497.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /BaragonServiceIntegrationTests/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (aws-java-sdk-core version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2020-28491](https://www.mend.io/vulnerability-database/CVE-2020-28491) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | jackson-dataformat-cbor-2.9.9.jar | Transitive | 1.11.498 | ✅ |
| [WS-2019-0379](https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | commons-codec-1.10.jar | Transitive | 1.11.657 | ✅ |
| [CVE-2020-13956](https://www.mend.io/vulnerability-database/CVE-2020-13956) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 5.3 | httpclient-4.5.5.jar | Transitive | 1.11.657 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2020-28491</summary>
### Vulnerable Library - <b>jackson-dataformat-cbor-2.9.9.jar</b></p>
<p>Support for reading and writing Concise Binary Object Representation
([CBOR](https://www.rfc-editor.org/info/rfc7049)
encoded data using Jackson abstractions (streaming API, data binding, tree model)</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson-dataformats-binary">http://github.com/FasterXML/jackson-dataformats-binary</a></p>
<p>Path to dependency file: /BaragonData/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- aws-java-sdk-core-1.11.497.jar (Root Library)
- :x: **jackson-dataformat-cbor-2.9.9.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
This affects the package com.fasterxml.jackson.dataformat:jackson-dataformat-cbor from 0 and before 2.11.4, from 2.12.0-rc1 and before 2.12.1. Unchecked allocation of byte buffer can cause a java.lang.OutOfMemoryError exception.
<p>Publish Date: 2021-02-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-28491>CVE-2020-28491</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28491">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28491</a></p>
<p>Release Date: 2021-02-18</p>
<p>Fix Resolution (com.fasterxml.jackson.dataformat:jackson-dataformat-cbor): 2.11.4</p>
<p>Direct dependency fix Resolution (com.amazonaws:aws-java-sdk-core): 1.11.498</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> WS-2019-0379</summary>
### Vulnerable Library - <b>commons-codec-1.10.jar</b></p>
<p>The Apache Commons Codec package contains simple encoder and decoders for
various formats such as Base64 and Hexadecimal. In addition to these
widely used encoders and decoders, the codec package also maintains a
collection of phonetic encoding utilities.</p>
<p>Path to dependency file: /BaragonWatcher/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar</p>
<p>
Dependency Hierarchy:
- aws-java-sdk-core-1.11.497.jar (Root Library)
- httpclient-4.5.5.jar
- :x: **commons-codec-1.10.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.
<p>Publish Date: 2019-05-20
<p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-05-20</p>
<p>Fix Resolution (commons-codec:commons-codec): 1.13</p>
<p>Direct dependency fix Resolution (com.amazonaws:aws-java-sdk-core): 1.11.657</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2020-13956</summary>
### Vulnerable Library - <b>httpclient-4.5.5.jar</b></p>
<p>Apache HttpComponents Client</p>
<p>Path to dependency file: /BaragonServiceIntegrationTests/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar</p>
<p>
Dependency Hierarchy:
- aws-java-sdk-core-1.11.497.jar (Root Library)
- :x: **httpclient-4.5.5.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-13956>CVE-2020-13956</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956</a></p>
<p>Release Date: 2020-12-02</p>
<p>Fix Resolution (org.apache.httpcomponents:httpclient): 4.5.13</p>
<p>Direct dependency fix Resolution (com.amazonaws:aws-java-sdk-core): 1.11.657</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
|
True
|
aws-java-sdk-core-1.11.497.jar: 3 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>aws-java-sdk-core-1.11.497.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /BaragonServiceIntegrationTests/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (aws-java-sdk-core version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2020-28491](https://www.mend.io/vulnerability-database/CVE-2020-28491) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | jackson-dataformat-cbor-2.9.9.jar | Transitive | 1.11.498 | ✅ |
| [WS-2019-0379](https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | commons-codec-1.10.jar | Transitive | 1.11.657 | ✅ |
| [CVE-2020-13956](https://www.mend.io/vulnerability-database/CVE-2020-13956) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 5.3 | httpclient-4.5.5.jar | Transitive | 1.11.657 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2020-28491</summary>
### Vulnerable Library - <b>jackson-dataformat-cbor-2.9.9.jar</b></p>
<p>Support for reading and writing Concise Binary Object Representation
([CBOR](https://www.rfc-editor.org/info/rfc7049)
encoded data using Jackson abstractions (streaming API, data binding, tree model)</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson-dataformats-binary">http://github.com/FasterXML/jackson-dataformats-binary</a></p>
<p>Path to dependency file: /BaragonData/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/dataformat/jackson-dataformat-cbor/2.9.9/jackson-dataformat-cbor-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- aws-java-sdk-core-1.11.497.jar (Root Library)
- :x: **jackson-dataformat-cbor-2.9.9.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
This affects the package com.fasterxml.jackson.dataformat:jackson-dataformat-cbor from 0 and before 2.11.4, from 2.12.0-rc1 and before 2.12.1. Unchecked allocation of byte buffer can cause a java.lang.OutOfMemoryError exception.
<p>Publish Date: 2021-02-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-28491>CVE-2020-28491</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28491">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28491</a></p>
<p>Release Date: 2021-02-18</p>
<p>Fix Resolution (com.fasterxml.jackson.dataformat:jackson-dataformat-cbor): 2.11.4</p>
<p>Direct dependency fix Resolution (com.amazonaws:aws-java-sdk-core): 1.11.498</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> WS-2019-0379</summary>
### Vulnerable Library - <b>commons-codec-1.10.jar</b></p>
<p>The Apache Commons Codec package contains simple encoder and decoders for
various formats such as Base64 and Hexadecimal. In addition to these
widely used encoders and decoders, the codec package also maintains a
collection of phonetic encoding utilities.</p>
<p>Path to dependency file: /BaragonWatcher/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar,/home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar</p>
<p>
Dependency Hierarchy:
- aws-java-sdk-core-1.11.497.jar (Root Library)
- httpclient-4.5.5.jar
- :x: **commons-codec-1.10.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.
<p>Publish Date: 2019-05-20
<p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-05-20</p>
<p>Fix Resolution (commons-codec:commons-codec): 1.13</p>
<p>Direct dependency fix Resolution (com.amazonaws:aws-java-sdk-core): 1.11.657</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2020-13956</summary>
### Vulnerable Library - <b>httpclient-4.5.5.jar</b></p>
<p>Apache HttpComponents Client</p>
<p>Path to dependency file: /BaragonServiceIntegrationTests/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar,/home/wss-scanner/.m2/repository/org/apache/httpcomponents/httpclient/4.5.5/httpclient-4.5.5.jar</p>
<p>
Dependency Hierarchy:
- aws-java-sdk-core-1.11.497.jar (Root Library)
- :x: **httpclient-4.5.5.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-13956>CVE-2020-13956</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956</a></p>
<p>Release Date: 2020-12-02</p>
<p>Fix Resolution (org.apache.httpcomponents:httpclient): 4.5.13</p>
<p>Direct dependency fix Resolution (com.amazonaws:aws-java-sdk-core): 1.11.657</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
|
non_test
|
aws java sdk core jar vulnerabilities highest severity is vulnerable library aws java sdk core jar path to dependency file baragonserviceintegrationtests pom xml path to vulnerable library home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar vulnerabilities cve severity cvss dependency type fixed in aws java sdk core version remediation available high jackson dataformat cbor jar transitive medium commons codec jar transitive medium httpclient jar transitive details cve vulnerable library jackson dataformat cbor jar support for reading and writing concise binary object representation encoded data using jackson abstractions streaming api data binding tree model library home page a href path to dependency file baragondata pom xml path to vulnerable library home wss scanner repository com fasterxml jackson dataformat jackson dataformat cbor jackson dataformat cbor jar home wss scanner repository com fasterxml jackson dataformat jackson dataformat cbor jackson dataformat cbor jar home wss scanner repository com fasterxml jackson dataformat jackson dataformat cbor jackson dataformat cbor jar home wss scanner repository com fasterxml jackson dataformat jackson dataformat cbor jackson dataformat cbor jar home wss scanner repository com fasterxml jackson dataformat jackson dataformat cbor jackson dataformat cbor jar home wss scanner repository com fasterxml jackson dataformat jackson dataformat cbor jackson dataformat cbor jar home wss scanner repository com fasterxml jackson dataformat jackson dataformat cbor jackson dataformat cbor jar dependency hierarchy aws java sdk core jar root library x jackson dataformat cbor jar vulnerable library found in base branch master vulnerability details this affects the package com fasterxml jackson dataformat jackson dataformat cbor from and before from and before unchecked allocation of byte buffer can cause a java lang outofmemoryerror exception publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson dataformat jackson dataformat cbor direct dependency fix resolution com amazonaws aws java sdk core rescue worker helmet automatic remediation is available for this issue ws vulnerable library commons codec jar the apache commons codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities path to dependency file baragonwatcher pom xml path to vulnerable library home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar home wss scanner repository commons codec commons codec commons codec jar dependency hierarchy aws java sdk core jar root library httpclient jar x commons codec jar vulnerable library found in base branch master vulnerability details apache commons codec before version “commons codec ” is vulnerable to information disclosure due to improper input validation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution commons codec commons codec direct dependency fix resolution com amazonaws aws java sdk core rescue worker helmet automatic remediation is available for this issue cve vulnerable library httpclient jar apache httpcomponents client path to dependency file baragonserviceintegrationtests pom xml path to vulnerable library home wss scanner repository org apache httpcomponents httpclient httpclient jar home wss scanner repository org apache httpcomponents httpclient httpclient jar home wss scanner repository org apache httpcomponents httpclient httpclient jar home wss scanner repository org apache httpcomponents httpclient httpclient jar home wss scanner repository org apache httpcomponents httpclient httpclient jar home wss scanner repository org apache httpcomponents httpclient httpclient jar home wss scanner repository org apache httpcomponents httpclient httpclient jar dependency hierarchy aws java sdk core jar root library x httpclient jar vulnerable library found in base branch master vulnerability details apache httpclient versions prior to version and can misinterpret malformed authority component in request uris passed to the library as java net uri object and pick the wrong target host for request execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache httpcomponents httpclient direct dependency fix resolution com amazonaws aws java sdk core rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
| 0
|
704,516
| 24,199,352,487
|
IssuesEvent
|
2022-09-24 10:31:27
|
ModuSynth/meta
|
https://api.github.com/repos/ModuSynth/meta
|
opened
|
Segregate the synthesizers per user
|
Priority P1 feature
|
## Context
Until now, every synthesizer created by any user will be regrouped under the same interface, and can be edited by whoever is on the interface. Until we decide to implement a collaborative way of editing synthesizers, only their creator should be able to modify them. To implement that, we need to implement a user's logic into our applicatiuon, and link resources to these users.
## Needs
As a _user of the application_
I want to be able to _subscribe, log in and log out_
So that I can _have my own synthesizer and be the only one able to modify them (outside of administrators_
|
1.0
|
Segregate the synthesizers per user - ## Context
Until now, every synthesizer created by any user will be regrouped under the same interface, and can be edited by whoever is on the interface. Until we decide to implement a collaborative way of editing synthesizers, only their creator should be able to modify them. To implement that, we need to implement a user's logic into our applicatiuon, and link resources to these users.
## Needs
As a _user of the application_
I want to be able to _subscribe, log in and log out_
So that I can _have my own synthesizer and be the only one able to modify them (outside of administrators_
|
non_test
|
segregate the synthesizers per user context until now every synthesizer created by any user will be regrouped under the same interface and can be edited by whoever is on the interface until we decide to implement a collaborative way of editing synthesizers only their creator should be able to modify them to implement that we need to implement a user s logic into our applicatiuon and link resources to these users needs as a user of the application i want to be able to subscribe log in and log out so that i can have my own synthesizer and be the only one able to modify them outside of administrators
| 0
|
278,415
| 24,152,376,272
|
IssuesEvent
|
2022-09-22 03:02:49
|
Brain-Bones/skeleton
|
https://api.github.com/repos/Brain-Bones/skeleton
|
closed
|
Add "Copy to Clipboard" feature to Code Blocks
|
enhancement ready to test
|
Reference:
https://mantine.dev/core/code/
This seems like an obvious use case for the newly requested action. Though the action will need to exist before this is possible.
This is a blocking prerequisite:
https://github.com/Brain-Bones/skeleton/issues/199
|
1.0
|
Add "Copy to Clipboard" feature to Code Blocks - Reference:
https://mantine.dev/core/code/
This seems like an obvious use case for the newly requested action. Though the action will need to exist before this is possible.
This is a blocking prerequisite:
https://github.com/Brain-Bones/skeleton/issues/199
|
test
|
add copy to clipboard feature to code blocks reference this seems like an obvious use case for the newly requested action though the action will need to exist before this is possible this is a blocking prerequisite
| 1
|
11,200
| 9,276,181,822
|
IssuesEvent
|
2019-03-20 01:46:00
|
Microsoft/vscode-cpptools
|
https://api.github.com/repos/Microsoft/vscode-cpptools
|
opened
|
Using the `Build and Debug Active File` command causes an "undefined task type" warning to get logged.
|
Language Service bug
|
With 0.22.0.
It doesn't seem to cause any other issue. Not sure why we never noticed this earlier.
|
1.0
|
Using the `Build and Debug Active File` command causes an "undefined task type" warning to get logged. - With 0.22.0.
It doesn't seem to cause any other issue. Not sure why we never noticed this earlier.
|
non_test
|
using the build and debug active file command causes an undefined task type warning to get logged with it doesn t seem to cause any other issue not sure why we never noticed this earlier
| 0
|
366,251
| 25,574,672,339
|
IssuesEvent
|
2022-11-30 20:54:49
|
lightninglabs/taro
|
https://api.github.com/repos/lightninglabs/taro
|
opened
|
BIP updates
|
documentation bips spec
|
Need to make multiple Taro BIP updates to reflect spec changes made during implementation.
Non-interactive transfers ( from #159 , #172 , #176 ) :
- [ ] Unconditional use of split commitments for non-interactive transfers
- [ ] Zero-value root locators for full-value sends
- [ ] NUMs key use for full-value sends
- [ ] Restricted support for collectibles in split commitments
- [ ] SplitCommitment clearing when computing input PkScripts for assets received non-interactively
Key encoding and storage ( from #187 )
- [ ] Handle all keys in addresses, assets, and records as compressed instead of x-only
Asset family to asset group rename ( from #200 )
- [ ] Rename asset family to asset group
|
1.0
|
BIP updates - Need to make multiple Taro BIP updates to reflect spec changes made during implementation.
Non-interactive transfers ( from #159 , #172 , #176 ) :
- [ ] Unconditional use of split commitments for non-interactive transfers
- [ ] Zero-value root locators for full-value sends
- [ ] NUMs key use for full-value sends
- [ ] Restricted support for collectibles in split commitments
- [ ] SplitCommitment clearing when computing input PkScripts for assets received non-interactively
Key encoding and storage ( from #187 )
- [ ] Handle all keys in addresses, assets, and records as compressed instead of x-only
Asset family to asset group rename ( from #200 )
- [ ] Rename asset family to asset group
|
non_test
|
bip updates need to make multiple taro bip updates to reflect spec changes made during implementation non interactive transfers from unconditional use of split commitments for non interactive transfers zero value root locators for full value sends nums key use for full value sends restricted support for collectibles in split commitments splitcommitment clearing when computing input pkscripts for assets received non interactively key encoding and storage from handle all keys in addresses assets and records as compressed instead of x only asset family to asset group rename from rename asset family to asset group
| 0
|
350,547
| 31,925,494,505
|
IssuesEvent
|
2023-09-19 01:12:36
|
AcalaNetwork/Acala
|
https://api.github.com/repos/AcalaNetwork/Acala
|
closed
|
broken e2e tests
|
T4-tests
|
https://github.com/AcalaNetwork/Acala/runs/7441526108?check_suite_focus=true
```
pthread lock: Invalid argument
error: test failed, to rerun pass '-p test-service --test standalone'
Caused by:
process didn't exit successfully: `/Acala/runner/work/Acala/Acala/target/release/deps/standalone-4010b5bae364c68f --include-ignored --skip test_full_node_catching_up --skip simple_balances_test` (signal: 6, SIGABRT: process abort signal)
make: *** [Makefile:150: test-e2e] Error 101
```
https://github.com/AcalaNetwork/Acala/runs/7399597230?check_suite_focus=true
```
terminate called without an active exception
test evm_fill_block_test ... ok
test result: ok. 5 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 83.84s
error: test failed, to rerun pass '-p test-service --test standalone'
Caused by:
process didn't exit successfully: `/Acala/runner/work/Acala/Acala/target/release/deps/standalone-8a55c432cd2ed2f8 --include-ignored --skip test_full_node_catching_up --skip simple_balances_test` (signal: 6, SIGABRT: process abort signal)
make: *** [Makefile:150: test-e2e] Error 101
```
looks like some concurrent issue with cargo test?
|
1.0
|
broken e2e tests - https://github.com/AcalaNetwork/Acala/runs/7441526108?check_suite_focus=true
```
pthread lock: Invalid argument
error: test failed, to rerun pass '-p test-service --test standalone'
Caused by:
process didn't exit successfully: `/Acala/runner/work/Acala/Acala/target/release/deps/standalone-4010b5bae364c68f --include-ignored --skip test_full_node_catching_up --skip simple_balances_test` (signal: 6, SIGABRT: process abort signal)
make: *** [Makefile:150: test-e2e] Error 101
```
https://github.com/AcalaNetwork/Acala/runs/7399597230?check_suite_focus=true
```
terminate called without an active exception
test evm_fill_block_test ... ok
test result: ok. 5 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 83.84s
error: test failed, to rerun pass '-p test-service --test standalone'
Caused by:
process didn't exit successfully: `/Acala/runner/work/Acala/Acala/target/release/deps/standalone-8a55c432cd2ed2f8 --include-ignored --skip test_full_node_catching_up --skip simple_balances_test` (signal: 6, SIGABRT: process abort signal)
make: *** [Makefile:150: test-e2e] Error 101
```
looks like some concurrent issue with cargo test?
|
test
|
broken tests pthread lock invalid argument error test failed to rerun pass p test service test standalone caused by process didn t exit successfully acala runner work acala acala target release deps standalone include ignored skip test full node catching up skip simple balances test signal sigabrt process abort signal make error terminate called without an active exception test evm fill block test ok test result ok passed failed ignored measured filtered out finished in error test failed to rerun pass p test service test standalone caused by process didn t exit successfully acala runner work acala acala target release deps standalone include ignored skip test full node catching up skip simple balances test signal sigabrt process abort signal make error looks like some concurrent issue with cargo test
| 1
|
658,529
| 21,895,982,931
|
IssuesEvent
|
2022-05-20 08:40:03
|
StartsMercury/simply-no-shading
|
https://api.github.com/repos/StartsMercury/simply-no-shading
|
closed
|
[5.0.0] Smart Chunk Reload with Iris Shader is Active
|
enhancement low priority
|
This mod is useless when any shader is enabled. Therefore reloading chunks on changing the config is not necessary.
|
1.0
|
[5.0.0] Smart Chunk Reload with Iris Shader is Active - This mod is useless when any shader is enabled. Therefore reloading chunks on changing the config is not necessary.
|
non_test
|
smart chunk reload with iris shader is active this mod is useless when any shader is enabled therefore reloading chunks on changing the config is not necessary
| 0
|
292,933
| 25,251,608,468
|
IssuesEvent
|
2022-11-15 15:04:18
|
ava-labs/spacesvm-rs
|
https://api.github.com/repos/ava-labs/spacesvm-rs
|
opened
|
add back integration tests
|
tests
|
These were removed because the code base was changed very quickly before release.
|
1.0
|
add back integration tests - These were removed because the code base was changed very quickly before release.
|
test
|
add back integration tests these were removed because the code base was changed very quickly before release
| 1
|
307,150
| 26,518,544,149
|
IssuesEvent
|
2023-01-18 23:18:57
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
DISABLED test_dynamo_min_operator_with_shape_dynamic_shapes (torch._dynamo.testing.make_test_cls_with_patches.<locals>.DummyTestClass)
|
module: flaky-tests skipped module: unknown
|
Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_dynamo_min_operator_with_shape_dynamic_shapes) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/10713308472).
Over the past 72 hours, it has flakily failed in 2 workflow(s).
**Debugging instructions (after clicking on the recent samples link):**
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Grep for `test_dynamo_min_operator_with_shape_dynamic_shapes`
Error retrieving /opt/conda/lib/python3.10/site-packages/torch/_dynamo/testing.py: Error: Statuscode 301
|
1.0
|
DISABLED test_dynamo_min_operator_with_shape_dynamic_shapes (torch._dynamo.testing.make_test_cls_with_patches.<locals>.DummyTestClass) - Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_dynamo_min_operator_with_shape_dynamic_shapes) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/10713308472).
Over the past 72 hours, it has flakily failed in 2 workflow(s).
**Debugging instructions (after clicking on the recent samples link):**
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Grep for `test_dynamo_min_operator_with_shape_dynamic_shapes`
Error retrieving /opt/conda/lib/python3.10/site-packages/torch/_dynamo/testing.py: Error: Statuscode 301
|
test
|
disabled test dynamo min operator with shape dynamic shapes torch dynamo testing make test cls with patches dummytestclass platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has flakily failed in workflow s debugging instructions after clicking on the recent samples link to find relevant log snippets click on the workflow logs linked above grep for test dynamo min operator with shape dynamic shapes error retrieving opt conda lib site packages torch dynamo testing py error statuscode
| 1
|
121,599
| 15,987,339,666
|
IssuesEvent
|
2021-04-19 00:11:48
|
almac-775/sp21-cse110-lab3
|
https://api.github.com/repos/almac-775/sp21-cse110-lab3
|
closed
|
Change Display
|
design documentation
|
<strong>What is the current issue? Please describe in detail</strong>
Have to mess with display values....
<strong>How would you like us to improve on this issue?</strong>
Not sure. Have to experiment with listed display values on the writeup...
<strong>If possible, please submit a picture or source for reference.</strong>
|
1.0
|
Change Display - <strong>What is the current issue? Please describe in detail</strong>
Have to mess with display values....
<strong>How would you like us to improve on this issue?</strong>
Not sure. Have to experiment with listed display values on the writeup...
<strong>If possible, please submit a picture or source for reference.</strong>
|
non_test
|
change display what is the current issue please describe in detail have to mess with display values how would you like us to improve on this issue not sure have to experiment with listed display values on the writeup if possible please submit a picture or source for reference
| 0
|
181,957
| 14,085,429,848
|
IssuesEvent
|
2020-11-05 00:58:33
|
GaloisInc/crucible
|
https://api.github.com/repos/GaloisInc/crucible
|
opened
|
crux-mir test suite failures
|
MIR testing
|
I can't figure out how to run the `crux-mir` test suite locally. When I run it, i get a cavalcade of similar errors.
```
Test suite test: RUNNING...
crux-mir
crux concrete
impl
simple: FAIL (0.25s)
Compiling and running oracle program (0.21s)
Oracle output: 42 (0.04s)
user error (Error 101 while running mir-json on test/conc_eval/impl/simple.rs)
self_mut: FAIL (0.28s)
Compiling and running oracle program (0.23s)
Oracle output: 42 (0.04s)
user error (Error 101 while running mir-json on test/conc_eval/impl/self_mut.rs)
self: FAIL (0.27s)
Compiling and running oracle program (0.22s)
Oracle output: 42 (0.05s)
user error (Error 101 while running mir-json on test/conc_eval/impl/self.rs)
```
This goes on for a long time, with a lot of `Error 101`. I don't know how to fix it.
|
1.0
|
crux-mir test suite failures - I can't figure out how to run the `crux-mir` test suite locally. When I run it, i get a cavalcade of similar errors.
```
Test suite test: RUNNING...
crux-mir
crux concrete
impl
simple: FAIL (0.25s)
Compiling and running oracle program (0.21s)
Oracle output: 42 (0.04s)
user error (Error 101 while running mir-json on test/conc_eval/impl/simple.rs)
self_mut: FAIL (0.28s)
Compiling and running oracle program (0.23s)
Oracle output: 42 (0.04s)
user error (Error 101 while running mir-json on test/conc_eval/impl/self_mut.rs)
self: FAIL (0.27s)
Compiling and running oracle program (0.22s)
Oracle output: 42 (0.05s)
user error (Error 101 while running mir-json on test/conc_eval/impl/self.rs)
```
This goes on for a long time, with a lot of `Error 101`. I don't know how to fix it.
|
test
|
crux mir test suite failures i can t figure out how to run the crux mir test suite locally when i run it i get a cavalcade of similar errors test suite test running crux mir crux concrete impl simple fail compiling and running oracle program oracle output user error error while running mir json on test conc eval impl simple rs self mut fail compiling and running oracle program oracle output user error error while running mir json on test conc eval impl self mut rs self fail compiling and running oracle program oracle output user error error while running mir json on test conc eval impl self rs this goes on for a long time with a lot of error i don t know how to fix it
| 1
|
20,745
| 6,101,957,506
|
IssuesEvent
|
2017-06-20 15:32:48
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
404 Category not found when article added in backend after upgrade to Joomla! 3.7.2
|
No Code Attached Yet
|
### Steps to reproduce the issue
1. Login to Joomla! 3.7.2 backend.
2. Add new article.
3. Save article or save and close article.
### Expected result
Remain on the article edit page with the link modified to: ...**edit&id=article_id** or open the article list page.
### Actual result
Front end 404 Category not found Error and the link is: **edit&id=0**
Article got saved and published if open a new window with Joomla! admin backend or website frontend.
### System information (as much as possible)
PHP Built On Linux ____________________________ #1 SMP Thu May 25 17:04:51 UTC 2017 x86_64
Versiune baza de date 5.5.5-10.1.24-MariaDB
Database Collation utf8mb4_unicode_ci
Colaționare conexiune Bază de date utf8mb4_unicode_ci
Versiune PHP 7.0.20
Server web Apache
Interfață PHP la server-ul web fpm-fcgi
Versiune Joomla! Joomla! 3.7.2 Stable [ Amani ] 22-May-2017 09:46 GMT
Versiune platformă Joomla! Joomla Platform 13.1.0 Stable [ Curiosity ] 24-Apr-2013 00:00 GMT
Agent utilizator Mozilla/5.0 (Windows NT 6.3; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.86 Safari/537.36
Setare Valoare
Safe Mode Oprit
Deshide basedir Nimic
Erori de afișare Oprit
Open Tags scurte Pornit
Urcări fișiere Pornit
Magic Quotes Oprit
Register Globals Oprit
Output Buffering Oprit
Cale de salvare sesiune /var/cpanel/php/sessions/ea-php70
Auto Start sesiune 0
XML activat Da
Zlib Activat Da
ZIP nativ activat Da
Funcții dezactivate exec,passthru,shell_exec,system
Șir Multibyte (mbstring) Activat Da
Iconv disponibil Nu
Mcrypt Activat Da
Maxim Variabile input 10000
### Additional comments
The problem is not always present but it's annoying
|
1.0
|
404 Category not found when article added in backend after upgrade to Joomla! 3.7.2 - ### Steps to reproduce the issue
1. Login to Joomla! 3.7.2 backend.
2. Add new article.
3. Save article or save and close article.
### Expected result
Remain on the article edit page with the link modified to: ...**edit&id=article_id** or open the article list page.
### Actual result
Front end 404 Category not found Error and the link is: **edit&id=0**
Article got saved and published if open a new window with Joomla! admin backend or website frontend.
### System information (as much as possible)
PHP Built On Linux ____________________________ #1 SMP Thu May 25 17:04:51 UTC 2017 x86_64
Versiune baza de date 5.5.5-10.1.24-MariaDB
Database Collation utf8mb4_unicode_ci
Colaționare conexiune Bază de date utf8mb4_unicode_ci
Versiune PHP 7.0.20
Server web Apache
Interfață PHP la server-ul web fpm-fcgi
Versiune Joomla! Joomla! 3.7.2 Stable [ Amani ] 22-May-2017 09:46 GMT
Versiune platformă Joomla! Joomla Platform 13.1.0 Stable [ Curiosity ] 24-Apr-2013 00:00 GMT
Agent utilizator Mozilla/5.0 (Windows NT 6.3; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.86 Safari/537.36
Setare Valoare
Safe Mode Oprit
Deshide basedir Nimic
Erori de afișare Oprit
Open Tags scurte Pornit
Urcări fișiere Pornit
Magic Quotes Oprit
Register Globals Oprit
Output Buffering Oprit
Cale de salvare sesiune /var/cpanel/php/sessions/ea-php70
Auto Start sesiune 0
XML activat Da
Zlib Activat Da
ZIP nativ activat Da
Funcții dezactivate exec,passthru,shell_exec,system
Șir Multibyte (mbstring) Activat Da
Iconv disponibil Nu
Mcrypt Activat Da
Maxim Variabile input 10000
### Additional comments
The problem is not always present but it's annoying
|
non_test
|
category not found when article added in backend after upgrade to joomla steps to reproduce the issue login to joomla backend add new article save article or save and close article expected result remain on the article edit page with the link modified to edit id article id or open the article list page actual result front end category not found error and the link is edit id article got saved and published if open a new window with joomla admin backend or website frontend system information as much as possible php built on linux smp thu may utc versiune baza de date mariadb database collation unicode ci colaționare conexiune bază de date unicode ci versiune php server web apache interfață php la server ul web fpm fcgi versiune joomla joomla stable may gmt versiune platformă joomla joomla platform stable apr gmt agent utilizator mozilla windows nt applewebkit khtml like gecko chrome safari setare valoare safe mode oprit deshide basedir nimic erori de afișare oprit open tags scurte pornit urcări fișiere pornit magic quotes oprit register globals oprit output buffering oprit cale de salvare sesiune var cpanel php sessions ea auto start sesiune xml activat da zlib activat da zip nativ activat da funcții dezactivate exec passthru shell exec system șir multibyte mbstring activat da iconv disponibil nu mcrypt activat da maxim variabile input additional comments the problem is not always present but it s annoying
| 0
|
15,147
| 3,927,180,808
|
IssuesEvent
|
2016-04-23 11:41:30
|
MarlinFirmware/Marlin
|
https://api.github.com/repos/MarlinFirmware/Marlin
|
closed
|
Difference between zprobe_zoffset and Z home offset?
|
Documentation Issue Inactive
|
Does anyone know the difference between these two values? It seems they both serve the same purpose. If that is true, can one of them be removed? If that is not true could someone explain to me what the difference is?
|
1.0
|
Difference between zprobe_zoffset and Z home offset? - Does anyone know the difference between these two values? It seems they both serve the same purpose. If that is true, can one of them be removed? If that is not true could someone explain to me what the difference is?
|
non_test
|
difference between zprobe zoffset and z home offset does anyone know the difference between these two values it seems they both serve the same purpose if that is true can one of them be removed if that is not true could someone explain to me what the difference is
| 0
|
41,632
| 5,379,112,423
|
IssuesEvent
|
2017-02-23 16:27:33
|
learn-co-curriculum/redux-reducer
|
https://api.github.com/repos/learn-co-curriculum/redux-reducer
|
closed
|
Reducer Tests
|
Test
|
manageFriends reducer tests for entire state to be returned.
managePresents reducer tests only for state.presents but the default still tests for the entire state. Update for consistency to avoid confusion?
|
1.0
|
Reducer Tests - manageFriends reducer tests for entire state to be returned.
managePresents reducer tests only for state.presents but the default still tests for the entire state. Update for consistency to avoid confusion?
|
test
|
reducer tests managefriends reducer tests for entire state to be returned managepresents reducer tests only for state presents but the default still tests for the entire state update for consistency to avoid confusion
| 1
|
180,173
| 30,456,132,622
|
IssuesEvent
|
2023-07-16 22:44:50
|
limlabs/saleor-storefront-starter
|
https://api.github.com/repos/limlabs/saleor-storefront-starter
|
closed
|
Theme
|
Epic Design
|
## Epic: [Theme]
### Description
[Provide a brief overview of the epic and its goals. Include any relevant background information or context.]
### Dependencies
- [Dependency 1]
- [Dependency 2]
- [Dependency 3]
### Links
- [Link 1]
- [Link 2]
- [Link 3]
### Tasks
- [Task 1]
- [Task 2]
- [Task 3]
### Acceptance Criteria
- [Acceptance criteria 1]
- [Acceptance criteria 2]
- [Acceptance criteria 3]
### Related Issues
- [Related issue 1]
- [Related issue 2]
- [Related issue 3]
### Additional Notes/Considerations
[Add any additional notes or considerations related to the epic.]
### Labels
[Specify relevant labels to categorize the epic, e.g., `epic`, `feature`, `priority`, etc.]
### Assignees
[Assign the relevant team members responsible for working on the epic.]
### Milestone
[Specify the milestone or release cycle associated with the epic.]
### Epic Owner
[Specify the person responsible for overseeing and coordinating the epic.]
|
1.0
|
Theme - ## Epic: [Theme]
### Description
[Provide a brief overview of the epic and its goals. Include any relevant background information or context.]
### Dependencies
- [Dependency 1]
- [Dependency 2]
- [Dependency 3]
### Links
- [Link 1]
- [Link 2]
- [Link 3]
### Tasks
- [Task 1]
- [Task 2]
- [Task 3]
### Acceptance Criteria
- [Acceptance criteria 1]
- [Acceptance criteria 2]
- [Acceptance criteria 3]
### Related Issues
- [Related issue 1]
- [Related issue 2]
- [Related issue 3]
### Additional Notes/Considerations
[Add any additional notes or considerations related to the epic.]
### Labels
[Specify relevant labels to categorize the epic, e.g., `epic`, `feature`, `priority`, etc.]
### Assignees
[Assign the relevant team members responsible for working on the epic.]
### Milestone
[Specify the milestone or release cycle associated with the epic.]
### Epic Owner
[Specify the person responsible for overseeing and coordinating the epic.]
|
non_test
|
theme epic description dependencies links tasks acceptance criteria related issues additional notes considerations labels assignees milestone epic owner
| 0
|
685,101
| 23,443,958,651
|
IssuesEvent
|
2022-08-15 17:37:43
|
cds-snc/notification-planning
|
https://api.github.com/repos/cds-snc/notification-planning
|
opened
|
Delete client data when their account is archived
|
Medium Priority | Priorité moyenne Privacy | Vie privée Policy l Politique Data l Données
|
_Delete client data when their account is archived _
## Description
As a client of Notify, particularly PTM, I need to be able to have my data (name, phone number, email address) deleted from GC Notify's records and database so that I can have my privacy respected if my account has been archived, or if I am no longer using Notify.
As a Notify admin, I need to be able to delete archived user data from our records and database if a account is archived so that I can respect PTM privacy policies and mitigate risk if there is a data breech.
WHY are we building?
As we expand and grow, we may have clients who fall under different privacy policies than the federal government. When an account it archived or deleted, we want to be able to delete that user data from our system.
WHAT are we building?
VALUE created by our solution
## Acceptance Criteria** (Definition of done)
_To be refined through discussion with the team_
Given some context, when (X) action occurs, then (Y) outcome is achieved
* A11y
* Bilingualism
* Privacy considerations
* Security controls in place
* Measuring success and metrics
## QA Steps
- [ ] Tested in a realistic production scenario
|
1.0
|
Delete client data when their account is archived - _Delete client data when their account is archived _
## Description
As a client of Notify, particularly PTM, I need to be able to have my data (name, phone number, email address) deleted from GC Notify's records and database so that I can have my privacy respected if my account has been archived, or if I am no longer using Notify.
As a Notify admin, I need to be able to delete archived user data from our records and database if a account is archived so that I can respect PTM privacy policies and mitigate risk if there is a data breech.
WHY are we building?
As we expand and grow, we may have clients who fall under different privacy policies than the federal government. When an account it archived or deleted, we want to be able to delete that user data from our system.
WHAT are we building?
VALUE created by our solution
## Acceptance Criteria** (Definition of done)
_To be refined through discussion with the team_
Given some context, when (X) action occurs, then (Y) outcome is achieved
* A11y
* Bilingualism
* Privacy considerations
* Security controls in place
* Measuring success and metrics
## QA Steps
- [ ] Tested in a realistic production scenario
|
non_test
|
delete client data when their account is archived delete client data when their account is archived description as a client of notify particularly ptm i need to be able to have my data name phone number email address deleted from gc notify s records and database so that i can have my privacy respected if my account has been archived or if i am no longer using notify as a notify admin i need to be able to delete archived user data from our records and database if a account is archived so that i can respect ptm privacy policies and mitigate risk if there is a data breech why are we building as we expand and grow we may have clients who fall under different privacy policies than the federal government when an account it archived or deleted we want to be able to delete that user data from our system what are we building value created by our solution acceptance criteria definition of done to be refined through discussion with the team given some context when x action occurs then y outcome is achieved bilingualism privacy considerations security controls in place measuring success and metrics qa steps tested in a realistic production scenario
| 0
|
70,292
| 15,079,827,269
|
IssuesEvent
|
2021-02-05 10:42:19
|
mkevenaar/chocolatey-packages
|
https://api.github.com/repos/mkevenaar/chocolatey-packages
|
closed
|
CVE-2015-3227 (Medium) detected in activesupport-3.2.22.5.gem - autoclosed
|
security vulnerability
|
## CVE-2015-3227 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>activesupport-3.2.22.5.gem</b></p></summary>
<p>A toolkit of support libraries and Ruby core extensions extracted from the Rails framework. Rich support for multibyte strings, internationalization, time zones, and testing.</p>
<p>Library home page: <a href="https://rubygems.org/gems/activesupport-3.2.22.5.gem">https://rubygems.org/gems/activesupport-3.2.22.5.gem</a></p>
<p>
Dependency Hierarchy:
- github-pages-211.gem (Root Library)
- jemoji-0.12.0.gem
- html-pipeline-2.14.0.gem
- :x: **activesupport-3.2.22.5.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mkevenaar/chocolatey-packages/commit/313d99806d6afee39ff0c246d553e8134ba4579e">313d99806d6afee39ff0c246d553e8134ba4579e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The (1) jdom.rb and (2) rexml.rb components in Active Support in Ruby on Rails before 4.1.11 and 4.2.x before 4.2.2, when JDOM or REXML is enabled, allow remote attackers to cause a denial of service (SystemStackError) via a large XML document depth.
<p>Publish Date: 2015-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-3227>CVE-2015-3227</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-3227">https://nvd.nist.gov/vuln/detail/CVE-2015-3227</a></p>
<p>Release Date: 2015-07-26</p>
<p>Fix Resolution: 4.1.11,4.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2015-3227 (Medium) detected in activesupport-3.2.22.5.gem - autoclosed - ## CVE-2015-3227 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>activesupport-3.2.22.5.gem</b></p></summary>
<p>A toolkit of support libraries and Ruby core extensions extracted from the Rails framework. Rich support for multibyte strings, internationalization, time zones, and testing.</p>
<p>Library home page: <a href="https://rubygems.org/gems/activesupport-3.2.22.5.gem">https://rubygems.org/gems/activesupport-3.2.22.5.gem</a></p>
<p>
Dependency Hierarchy:
- github-pages-211.gem (Root Library)
- jemoji-0.12.0.gem
- html-pipeline-2.14.0.gem
- :x: **activesupport-3.2.22.5.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mkevenaar/chocolatey-packages/commit/313d99806d6afee39ff0c246d553e8134ba4579e">313d99806d6afee39ff0c246d553e8134ba4579e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The (1) jdom.rb and (2) rexml.rb components in Active Support in Ruby on Rails before 4.1.11 and 4.2.x before 4.2.2, when JDOM or REXML is enabled, allow remote attackers to cause a denial of service (SystemStackError) via a large XML document depth.
<p>Publish Date: 2015-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-3227>CVE-2015-3227</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-3227">https://nvd.nist.gov/vuln/detail/CVE-2015-3227</a></p>
<p>Release Date: 2015-07-26</p>
<p>Fix Resolution: 4.1.11,4.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in activesupport gem autoclosed cve medium severity vulnerability vulnerable library activesupport gem a toolkit of support libraries and ruby core extensions extracted from the rails framework rich support for multibyte strings internationalization time zones and testing library home page a href dependency hierarchy github pages gem root library jemoji gem html pipeline gem x activesupport gem vulnerable library found in head commit a href found in base branch master vulnerability details the jdom rb and rexml rb components in active support in ruby on rails before and x before when jdom or rexml is enabled allow remote attackers to cause a denial of service systemstackerror via a large xml document depth publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
14,633
| 11,028,688,972
|
IssuesEvent
|
2019-12-06 12:18:11
|
killercup/cargo-edit
|
https://api.github.com/repos/killercup/cargo-edit
|
closed
|
Update CI to test on Linux, Mac, Windows
|
help wanted infrastructure
|
We should update our CI setup to test on all platforms. I recently saw that https://github.com/actions-rs/ has a collection of templates for Github actions that support testing and building Rust projects, so that might be an interesting alternative to travis.
My requirements:
- [ ] keep [bors](https://bors.tech/)
- [ ] test on Linux/macOS/Windows
Optional cool stuff:
- [ ] Build artifacts for releases
- [ ] Run cargo-audit as well
- [ ] Give inline error reports on PRs
|
1.0
|
Update CI to test on Linux, Mac, Windows - We should update our CI setup to test on all platforms. I recently saw that https://github.com/actions-rs/ has a collection of templates for Github actions that support testing and building Rust projects, so that might be an interesting alternative to travis.
My requirements:
- [ ] keep [bors](https://bors.tech/)
- [ ] test on Linux/macOS/Windows
Optional cool stuff:
- [ ] Build artifacts for releases
- [ ] Run cargo-audit as well
- [ ] Give inline error reports on PRs
|
non_test
|
update ci to test on linux mac windows we should update our ci setup to test on all platforms i recently saw that has a collection of templates for github actions that support testing and building rust projects so that might be an interesting alternative to travis my requirements keep test on linux macos windows optional cool stuff build artifacts for releases run cargo audit as well give inline error reports on prs
| 0
|
373,786
| 26,084,549,054
|
IssuesEvent
|
2022-12-25 23:15:11
|
bounswe/bounswe2022group1
|
https://api.github.com/repos/bounswe/bounswe2022group1
|
closed
|
Favorite learning spaces added to the home page
|
Type: Documentation Priority: Medium Status: Completed Android
|
Favorite learning spaces section added to the home page.
API connections will be implemented.
Page design looks like this:

*Reviewer:*
@omerozdemir1
*Review Deadline:*
27.12.2022
|
1.0
|
Favorite learning spaces added to the home page - Favorite learning spaces section added to the home page.
API connections will be implemented.
Page design looks like this:

*Reviewer:*
@omerozdemir1
*Review Deadline:*
27.12.2022
|
non_test
|
favorite learning spaces added to the home page favorite learning spaces section added to the home page api connections will be implemented page design looks like this reviewer review deadline
| 0
|
279,751
| 24,252,594,201
|
IssuesEvent
|
2022-09-27 15:12:20
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Test badge on TreeView
|
testplan-item
|
Refs: https://github.com/microsoft/vscode/issues/62783
- [x] anyOS @meganrogge
- [x] anyOS @bpasero
Complexity: 4
Roles: Developer
[Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23161780%0A%0A&assignees=alexr00)
---
We have newly finalized API for badges on TreeViews: https://github.com/microsoft/vscode/blob/9caff0557b9ed17a6ccb44b2f18c1f6915f2eb85/src/vscode-dts/vscode.d.ts#L10484-L10488
This badge shows as a circle with a number on the view's view container.
To test:
- Read the inline documentation for the API and verify that it's clear.
- Try out the API in a tree view. For example, in the tree-view-sample: https://github.com/microsoft/vscode-extension-samples/blob/31d3d4920b3fae52e03a185e434f51ee81965515/tree-view-sample/src/testView.ts#L6-L7
- Verify that you can set the badge number value.
- Verify that you can update the badge number value.
- Verify that you can cause the badge to hide by unsetting the badge number value.
|
1.0
|
Test badge on TreeView - Refs: https://github.com/microsoft/vscode/issues/62783
- [x] anyOS @meganrogge
- [x] anyOS @bpasero
Complexity: 4
Roles: Developer
[Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23161780%0A%0A&assignees=alexr00)
---
We have newly finalized API for badges on TreeViews: https://github.com/microsoft/vscode/blob/9caff0557b9ed17a6ccb44b2f18c1f6915f2eb85/src/vscode-dts/vscode.d.ts#L10484-L10488
This badge shows as a circle with a number on the view's view container.
To test:
- Read the inline documentation for the API and verify that it's clear.
- Try out the API in a tree view. For example, in the tree-view-sample: https://github.com/microsoft/vscode-extension-samples/blob/31d3d4920b3fae52e03a185e434f51ee81965515/tree-view-sample/src/testView.ts#L6-L7
- Verify that you can set the badge number value.
- Verify that you can update the badge number value.
- Verify that you can cause the badge to hide by unsetting the badge number value.
|
test
|
test badge on treeview refs anyos meganrogge anyos bpasero complexity roles developer we have newly finalized api for badges on treeviews this badge shows as a circle with a number on the view s view container to test read the inline documentation for the api and verify that it s clear try out the api in a tree view for example in the tree view sample verify that you can set the badge number value verify that you can update the badge number value verify that you can cause the badge to hide by unsetting the badge number value
| 1
|
250,591
| 21,316,166,221
|
IssuesEvent
|
2022-04-16 10:07:11
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: django failed
|
C-test-failure O-robot O-roachtest release-blocker branch-release-21.2
|
roachtest.django [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4907623&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4907623&tab=artifacts#/django) on release-21.2 @ [0f0029653c25772d09adae5be308ce0c45f84f0a](https://github.com/cockroachdb/cockroach/commits/0f0029653c25772d09adae5be308ce0c45f84f0a):
```
The test failed on branch=release-21.2, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/django/run_1
orm_helpers.go:245,orm_helpers.go:171,django.go:211,django.go:222,test_runner.go:777:
Tests run on Cockroach v21.2.7-232-g0f0029653c
Tests run against django cockroach-3.2.x
9042 Total Tests Run
9041 tests passed
1 test failed
613 tests skipped
3 tests ignored
0 tests passed unexpectedly
1 test failed unexpectedly
0 tests expected failed but skipped
0 tests expected failed but not run
---
--- SKIP: backends.mysql.test_creation.DatabaseCreationTests.test_create_test_db_unexpected_error due to MySQL tests (expected)
--- FAIL: queries.test_explain.ExplainTests.test_unknown_options (unexpected)
For a full summary look at the django artifacts
An updated blocklist (djangoBlocklist21_2) is available in the artifacts' django log
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #80042 roachtest: django failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-21.1 release-blocker]
- #79989 roachtest: django failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1 release-blocker]
- #79988 roachtest: django failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1.0 release-blocker]
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*django.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
2.0
|
roachtest: django failed - roachtest.django [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4907623&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4907623&tab=artifacts#/django) on release-21.2 @ [0f0029653c25772d09adae5be308ce0c45f84f0a](https://github.com/cockroachdb/cockroach/commits/0f0029653c25772d09adae5be308ce0c45f84f0a):
```
The test failed on branch=release-21.2, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/django/run_1
orm_helpers.go:245,orm_helpers.go:171,django.go:211,django.go:222,test_runner.go:777:
Tests run on Cockroach v21.2.7-232-g0f0029653c
Tests run against django cockroach-3.2.x
9042 Total Tests Run
9041 tests passed
1 test failed
613 tests skipped
3 tests ignored
0 tests passed unexpectedly
1 test failed unexpectedly
0 tests expected failed but skipped
0 tests expected failed but not run
---
--- SKIP: backends.mysql.test_creation.DatabaseCreationTests.test_create_test_db_unexpected_error due to MySQL tests (expected)
--- FAIL: queries.test_explain.ExplainTests.test_unknown_options (unexpected)
For a full summary look at the django artifacts
An updated blocklist (djangoBlocklist21_2) is available in the artifacts' django log
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #80042 roachtest: django failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-21.1 release-blocker]
- #79989 roachtest: django failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1 release-blocker]
- #79988 roachtest: django failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1.0 release-blocker]
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*django.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
roachtest django failed roachtest django with on release the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts django run orm helpers go orm helpers go django go django go test runner go tests run on cockroach tests run against django cockroach x total tests run tests passed test failed tests skipped tests ignored tests passed unexpectedly test failed unexpectedly tests expected failed but skipped tests expected failed but not run skip backends mysql test creation databasecreationtests test create test db unexpected error due to mysql tests expected fail queries test explain explaintests test unknown options unexpected for a full summary look at the django artifacts an updated blocklist is available in the artifacts django log reproduce see same failure on other branches roachtest django failed roachtest django failed roachtest django failed cc cockroachdb sql experience
| 1
|
1,906
| 4,046,950,467
|
IssuesEvent
|
2016-05-23 00:33:22
|
itkpi/events-parser
|
https://api.github.com/repos/itkpi/events-parser
|
closed
|
Black list
|
enhancement (service) enhancement (what user see)
|
Випилювати будь-які:
- по title
- по title || agenda
- Слово "Курсы" в title & назва компанії в title/agenda
|
1.0
|
Black list - Випилювати будь-які:
- по title
- по title || agenda
- Слово "Курсы" в title & назва компанії в title/agenda
|
non_test
|
black list випилювати будь які по title по title agenda слово курсы в title назва компанії в title agenda
| 0
|
102,212
| 8,821,982,921
|
IssuesEvent
|
2019-01-02 06:50:51
|
humera987/FXScript-Test-Functions
|
https://api.github.com/repos/humera987/FXScript-Test-Functions
|
closed
|
testing 2 : ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue
|
testing 2
|
Project : testing 2
Job : Default
Env : Default
Region : US_WEST
Result : fail
Status Code : 500
Headers : {}
Endpoint : http://13.56.210.25/api/v1/projects/{id}/auto-suggestions/null
Request :
Response :
Not enough variable values available to expand 'id'
Logs :
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : URL [http://13.56.210.25/api/v1/projects/{id}/auto-suggestions/null]
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Method [GET]
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Request []
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Request-Headers [{Content-Type=[application/json], Accept=[application/json], Authorization=[Basic SHVtZXJhLy9odW1lcmFAZnhsYWJzLmlvOmh1bWVyYTEyMyQ=]}]
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Response [Not enough variable values available to expand 'id']
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Response-Headers [{}]
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : StatusCode [500]
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Time [4]
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Size [51]
2019-01-02 06:44:45 INFO [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Assertion [@StatusCode != 401] resolved-to [500 != 401] result [Passed]
2019-01-02 06:44:45 ERROR [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Assertion [@StatusCode != 500] resolved-to [500 != 500] result [Failed]
2019-01-02 06:44:45 INFO [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Assertion [@StatusCode != 200] resolved-to [500 != 200] result [Passed]
2019-01-02 06:44:45 INFO [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Assertion [@StatusCode != 404] resolved-to [500 != 404] result [Passed]
--- FX Bot ---
|
1.0
|
testing 2 : ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue - Project : testing 2
Job : Default
Env : Default
Region : US_WEST
Result : fail
Status Code : 500
Headers : {}
Endpoint : http://13.56.210.25/api/v1/projects/{id}/auto-suggestions/null
Request :
Response :
Not enough variable values available to expand 'id'
Logs :
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : URL [http://13.56.210.25/api/v1/projects/{id}/auto-suggestions/null]
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Method [GET]
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Request []
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Request-Headers [{Content-Type=[application/json], Accept=[application/json], Authorization=[Basic SHVtZXJhLy9odW1lcmFAZnhsYWJzLmlvOmh1bWVyYTEyMyQ=]}]
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Response [Not enough variable values available to expand 'id']
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Response-Headers [{}]
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : StatusCode [500]
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Time [4]
2019-01-02 06:44:45 DEBUG [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Size [51]
2019-01-02 06:44:45 INFO [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Assertion [@StatusCode != 401] resolved-to [500 != 401] result [Passed]
2019-01-02 06:44:45 ERROR [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Assertion [@StatusCode != 500] resolved-to [500 != 500] result [Failed]
2019-01-02 06:44:45 INFO [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Assertion [@StatusCode != 200] resolved-to [500 != 200] result [Passed]
2019-01-02 06:44:45 INFO [ApiV1ProjectsIdAutoSuggestionsStatusGetPathParamStatusNullValue] : Assertion [@StatusCode != 404] resolved-to [500 != 404] result [Passed]
--- FX Bot ---
|
test
|
testing project testing job default env default region us west result fail status code headers endpoint request response not enough variable values available to expand id logs debug url debug method debug request debug request headers accept authorization debug response debug response headers debug statuscode debug time debug size info assertion resolved to result error assertion resolved to result info assertion resolved to result info assertion resolved to result fx bot
| 1
|
5,894
| 8,710,249,559
|
IssuesEvent
|
2018-12-06 15:58:35
|
Open-EO/openeo-api
|
https://api.github.com/repos/Open-EO/openeo-api
|
closed
|
Dedicated export processes in the process graph
|
job management processes vote
|
The definition of the process graph output format is located at the moment in the preview and job requests. However, IMHO it makes more sense to have export definitions in the process graph itself.
If you have complex process graphs you might want to export intermediate results and statistical analysis data. It is more flexible to have dedicated export processes, in which you can set the export format directly.
With export processes one can export many different formats in a single process graph.
For example, the following process will export all its inputs as GTiff files somewhere within a process graph and pipes the inputs upstream to the next process, so that further processing is possible with the same data:
```
"process_id": "raster_export",
"format": "GTiff",
"imagery": {
"process_id": "get_data",
"data_id": "nc_spm_08.landsat.raster.elevation",
"imagery": {
"process_id": "get_data",
"data_id": "nc_spm_08.landsat.raster.slope"
}
}
```
|
1.0
|
Dedicated export processes in the process graph - The definition of the process graph output format is located at the moment in the preview and job requests. However, IMHO it makes more sense to have export definitions in the process graph itself.
If you have complex process graphs you might want to export intermediate results and statistical analysis data. It is more flexible to have dedicated export processes, in which you can set the export format directly.
With export processes one can export many different formats in a single process graph.
For example, the following process will export all its inputs as GTiff files somewhere within a process graph and pipes the inputs upstream to the next process, so that further processing is possible with the same data:
```
"process_id": "raster_export",
"format": "GTiff",
"imagery": {
"process_id": "get_data",
"data_id": "nc_spm_08.landsat.raster.elevation",
"imagery": {
"process_id": "get_data",
"data_id": "nc_spm_08.landsat.raster.slope"
}
}
```
|
non_test
|
dedicated export processes in the process graph the definition of the process graph output format is located at the moment in the preview and job requests however imho it makes more sense to have export definitions in the process graph itself if you have complex process graphs you might want to export intermediate results and statistical analysis data it is more flexible to have dedicated export processes in which you can set the export format directly with export processes one can export many different formats in a single process graph for example the following process will export all its inputs as gtiff files somewhere within a process graph and pipes the inputs upstream to the next process so that further processing is possible with the same data process id raster export format gtiff imagery process id get data data id nc spm landsat raster elevation imagery process id get data data id nc spm landsat raster slope
| 0
|
106,031
| 13,240,630,754
|
IssuesEvent
|
2020-08-19 06:45:32
|
nikodemus/foolang
|
https://api.github.com/repos/nikodemus/foolang
|
opened
|
#min: and #max: confusing
|
design later quality-of-life sidetrack
|
Currently
```
x min: y
```
means `min(x,y)`, but reads as "x, or at minimum y", ie. max(x,y).
|
1.0
|
#min: and #max: confusing - Currently
```
x min: y
```
means `min(x,y)`, but reads as "x, or at minimum y", ie. max(x,y).
|
non_test
|
min and max confusing currently x min y means min x y but reads as x or at minimum y ie max x y
| 0
|
102,523
| 32,035,462,366
|
IssuesEvent
|
2023-09-22 15:02:40
|
statechannels/go-nitro
|
https://api.github.com/repos/statechannels/go-nitro
|
closed
|
The eth chain service should query for historic logs
|
:building_construction: Productionization
|
Right now the eth chain service calls [SubscribeFilterLogs](https://github.com/statechannels/go-nitro/blob/7cc2c812388313a57b996e9341cb6ae2ee535621/node/engine/chainservice/eth_chainservice.go#L292) to get contract events. However [SubscribeFilterLogs](https://github.com/ethereum/go-ethereum/issues/15063) only supports new events; it doesn't return any historic events that have happened.
That means if a nitro node misses a contract event (due to being offline) it will never catch up on that event, stalling out the directfund protocol.
Instead the chain service should fetch historic logs by calling `[FilterLogs](https://pkg.go.dev/github.com/ethereum/go-ethereum/ethclient#Client.FilterLogs)` on start up.
|
1.0
|
The eth chain service should query for historic logs - Right now the eth chain service calls [SubscribeFilterLogs](https://github.com/statechannels/go-nitro/blob/7cc2c812388313a57b996e9341cb6ae2ee535621/node/engine/chainservice/eth_chainservice.go#L292) to get contract events. However [SubscribeFilterLogs](https://github.com/ethereum/go-ethereum/issues/15063) only supports new events; it doesn't return any historic events that have happened.
That means if a nitro node misses a contract event (due to being offline) it will never catch up on that event, stalling out the directfund protocol.
Instead the chain service should fetch historic logs by calling `[FilterLogs](https://pkg.go.dev/github.com/ethereum/go-ethereum/ethclient#Client.FilterLogs)` on start up.
|
non_test
|
the eth chain service should query for historic logs right now the eth chain service calls to get contract events however only supports new events it doesn t return any historic events that have happened that means if a nitro node misses a contract event due to being offline it will never catch up on that event stalling out the directfund protocol instead the chain service should fetch historic logs by calling on start up
| 0
|
50,352
| 13,523,483,161
|
IssuesEvent
|
2020-09-15 10:01:47
|
AbdelhakAj/react-typescript
|
https://api.github.com/repos/AbdelhakAj/react-typescript
|
opened
|
CVE-2018-19797 (Medium) detected in opennmsopennms-source-22.0.1-1, node-sass-4.14.1.tgz
|
security vulnerability
|
## CVE-2018-19797 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opennmsopennms-source-22.0.1-1</b>, <b>node-sass-4.14.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: react-typescript/package.json</p>
<p>Path to vulnerable library: react-typescript/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/AbdelhakAj/react-typescript/commit/e4bd04280094e77c8687d4371ff697606b6d7237">e4bd04280094e77c8687d4371ff697606b6d7237</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a NULL Pointer Dereference in the function Sass::Selector_List::populate_extends in SharedPtr.hpp (used by ast.cpp and ast_selectors.cpp) may cause a Denial of Service (application crash) via a crafted sass input file.
<p>Publish Date: 2018-12-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19797>CVE-2018-19797</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19797</a></p>
<p>Release Date: 2018-12-03</p>
<p>Fix Resolution: libsass-3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-19797 (Medium) detected in opennmsopennms-source-22.0.1-1, node-sass-4.14.1.tgz - ## CVE-2018-19797 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opennmsopennms-source-22.0.1-1</b>, <b>node-sass-4.14.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: react-typescript/package.json</p>
<p>Path to vulnerable library: react-typescript/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/AbdelhakAj/react-typescript/commit/e4bd04280094e77c8687d4371ff697606b6d7237">e4bd04280094e77c8687d4371ff697606b6d7237</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a NULL Pointer Dereference in the function Sass::Selector_List::populate_extends in SharedPtr.hpp (used by ast.cpp and ast_selectors.cpp) may cause a Denial of Service (application crash) via a crafted sass input file.
<p>Publish Date: 2018-12-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19797>CVE-2018-19797</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19797">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19797</a></p>
<p>Release Date: 2018-12-03</p>
<p>Fix Resolution: libsass-3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in opennmsopennms source node sass tgz cve medium severity vulnerability vulnerable libraries opennmsopennms source node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file react typescript package json path to vulnerable library react typescript node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details in libsass a null pointer dereference in the function sass selector list populate extends in sharedptr hpp used by ast cpp and ast selectors cpp may cause a denial of service application crash via a crafted sass input file publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
| 0
|
301,971
| 26,113,660,842
|
IssuesEvent
|
2022-12-28 01:12:33
|
17ANT/INATON
|
https://api.github.com/repos/17ANT/INATON
|
closed
|
[PostModify] 테스트
|
✨ Feature ✅ Test
|
### 🌱 무엇을 하실 건지 설명해주세요!
- 오늘 작업 테스트 진행
### 🌱 구현방법 및 예상 동작
1. 게시글 수정
- 이미지 없는 게시글 수정 시 파일을 불러오지 못하는 이미지 출력X
- 게시글 수정 시 이미지 반영 여부
- 이미지 슬라이드 확인
2. 댓글 작성
- 댓글 입력시에만 게시 버튼 활성화
|
1.0
|
[PostModify] 테스트 - ### 🌱 무엇을 하실 건지 설명해주세요!
- 오늘 작업 테스트 진행
### 🌱 구현방법 및 예상 동작
1. 게시글 수정
- 이미지 없는 게시글 수정 시 파일을 불러오지 못하는 이미지 출력X
- 게시글 수정 시 이미지 반영 여부
- 이미지 슬라이드 확인
2. 댓글 작성
- 댓글 입력시에만 게시 버튼 활성화
|
test
|
테스트 🌱 무엇을 하실 건지 설명해주세요 오늘 작업 테스트 진행 🌱 구현방법 및 예상 동작 게시글 수정 이미지 없는 게시글 수정 시 파일을 불러오지 못하는 이미지 출력x 게시글 수정 시 이미지 반영 여부 이미지 슬라이드 확인 댓글 작성 댓글 입력시에만 게시 버튼 활성화
| 1
|
93,218
| 8,405,314,286
|
IssuesEvent
|
2018-10-11 14:58:30
|
chameleon-system/chameleon-system
|
https://api.github.com/repos/chameleon-system/chameleon-system
|
closed
|
Fix typos in XLIFF files
|
Status: Test Type: Bug
|
**Describe the bug**
There are some typos hiding in the XLIFF files.
**Affected version(s)**
6.2.0+
|
1.0
|
Fix typos in XLIFF files - **Describe the bug**
There are some typos hiding in the XLIFF files.
**Affected version(s)**
6.2.0+
|
test
|
fix typos in xliff files describe the bug there are some typos hiding in the xliff files affected version s
| 1
|
33,560
| 7,164,647,649
|
IssuesEvent
|
2018-01-29 11:59:27
|
scalameta/metals
|
https://api.github.com/repos/scalameta/metals
|
closed
|
Annoying error: scalameta.scalafmt.confPath=.scalafmt.conf is not a file
|
defect
|
While refactoring the scalafmt configuration handling I introduced a regression that causes the following error to appear whenever .scalafmt.conf is missing

|
1.0
|
Annoying error: scalameta.scalafmt.confPath=.scalafmt.conf is not a file - While refactoring the scalafmt configuration handling I introduced a regression that causes the following error to appear whenever .scalafmt.conf is missing

|
non_test
|
annoying error scalameta scalafmt confpath scalafmt conf is not a file while refactoring the scalafmt configuration handling i introduced a regression that causes the following error to appear whenever scalafmt conf is missing
| 0
|
346,052
| 30,862,815,016
|
IssuesEvent
|
2023-08-03 05:27:46
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
sql/tests: TestRandomSyntaxGeneration failed
|
C-test-failure O-robot branch-master T-sql-foundations
|
sql/tests.TestRandomSyntaxGeneration [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/11169507?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/11169507?buildTab=artifacts#/) on master @ [4fe2a80d81c6fc5a3da3c7c44c5fc38da67e0367](https://github.com/cockroachdb/cockroach/commits/4fe2a80d81c6fc5a3da3c7c44c5fc38da67e0367):
```
rsg_test.go:832: 1m15s of 5m0s: 159112 executions, 9071 successful
rsg_test.go:832: 1m20s of 5m0s: 170957 executions, 9699 successful
rsg_test.go:832: 1m25s of 5m0s: 182776 executions, 10297 successful
rsg_test.go:832: 1m30s of 5m0s: 194657 executions, 10882 successful
rsg_test.go:832: 1m35s of 5m0s: 206728 executions, 11480 successful
rsg_test.go:832: 1m40s of 5m0s: 218400 executions, 12040 successful
rsg_test.go:832: 1m45s of 5m0s: 229942 executions, 12604 successful
rsg_test.go:832: 1m50s of 5m0s: 241597 executions, 13158 successful
rsg_test.go:832: 1m55s of 5m0s: 253943 executions, 13761 successful
rsg_test.go:832: 2m0s of 5m0s: 266059 executions, 14317 successful
rsg_test.go:832: 2m5s of 5m0s: 276465 executions, 14805 successful
rsg_test.go:832: 2m10s of 5m0s: 288190 executions, 15370 successful
rsg_test.go:832: 2m15s of 5m0s: 300051 executions, 15934 successful
rsg_test.go:832: 2m20s of 5m0s: 311384 executions, 16457 successful
rsg_test.go:832: 2m25s of 5m0s: 322533 executions, 16966 successful
rsg_test.go:832: 2m30s of 5m0s: 331957 executions, 17442 successful
rsg_test.go:832: 2m35s of 5m0s: 343858 executions, 17963 successful
rsg_test.go:832: 2m40s of 5m0s: 356058 executions, 18530 successful
rsg_test.go:832: 2m45s of 5m0s: 368528 executions, 19092 successful
rsg_test.go:832: 2m50s of 5m0s: 379969 executions, 19621 successful
rsg_test.go:832: 2m55s of 5m0s: 390832 executions, 20162 successful
rsg_test.go:832: 3m0s of 5m0s: 401776 executions, 20646 successful
rsg_test.go:832: 3m5s of 5m0s: 412687 executions, 21138 successful
rsg_test.go:832: 3m10s of 5m0s: 424595 executions, 21717 successful
rsg_test.go:832: 3m15s of 5m0s: 436181 executions, 22288 successful
rsg_test.go:832: 3m20s of 5m0s: 447687 executions, 22797 successful
rsg_test.go:832: 3m25s of 5m0s: 459526 executions, 23357 successful
rsg_test.go:832: 3m30s of 5m0s: 470709 executions, 23854 successful
rsg_test.go:832: 3m35s of 5m0s: 481609 executions, 24407 successful
rsg_test.go:832: 3m40s of 5m0s: 486925 executions, 24647 successful
rsg_test.go:832: 3m45s of 5m0s: 491161 executions, 24853 successful
rsg_test.go:832: 3m50s of 5m0s: 495304 executions, 25048 successful
rsg_test.go:832: 3m55s of 5m0s: 499025 executions, 25228 successful
rsg_test.go:832: 4m0s of 5m0s: 503294 executions, 25446 successful
rsg_test.go:832: 4m5s of 5m0s: 508213 executions, 25671 successful
rsg_test.go:832: 4m10s of 5m0s: 513178 executions, 25906 successful
rsg_test.go:832: 4m15s of 5m0s: 518304 executions, 26148 successful
rsg_test.go:832: 4m20s of 5m0s: 523432 executions, 26391 successful
rsg_test.go:832: 4m25s of 5m0s: 528217 executions, 26594 successful
rsg_test.go:832: 4m30s of 5m0s: 533237 executions, 26800 successful
rsg_test.go:832: 4m35s of 5m0s: 538015 executions, 27020 successful
rsg_test.go:832: 4m40s of 5m0s: 543098 executions, 27234 successful
rsg_test.go:832: 4m45s of 5m0s: 548108 executions, 27466 successful
rsg_test.go:832: 4m50s of 5m0s: 553372 executions, 27686 successful
rsg_test.go:832: 4m55s of 5m0s: 558231 executions, 27907 successful
rsg_test.go:868: 562914 executions, 28098 successful
rsg_test.go:876: Parse followed by Format is not idempotent: "ALTER INDEX IF EXISTS FAMILY . FAMILY . ident VISIBILITY 7.379652426127388e-12" -> "ALTER INDEX IF EXISTS \"family\".\"family\".ident VISIBILITY 0.00" != "ALTER INDEX IF EXISTS \"family\".\"family\".ident NOT VISIBLE"
rsg_test.go:302: -- test log scope end --
test logs left over in: /artifacts/tmp/_tmp/d437d2c847dfedbc4972f231c3331c8e/logTestRandomSyntaxGeneration135218831
--- FAIL: TestRandomSyntaxGeneration (302.11s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-foundations
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRandomSyntaxGeneration.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
sql/tests: TestRandomSyntaxGeneration failed - sql/tests.TestRandomSyntaxGeneration [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/11169507?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/11169507?buildTab=artifacts#/) on master @ [4fe2a80d81c6fc5a3da3c7c44c5fc38da67e0367](https://github.com/cockroachdb/cockroach/commits/4fe2a80d81c6fc5a3da3c7c44c5fc38da67e0367):
```
rsg_test.go:832: 1m15s of 5m0s: 159112 executions, 9071 successful
rsg_test.go:832: 1m20s of 5m0s: 170957 executions, 9699 successful
rsg_test.go:832: 1m25s of 5m0s: 182776 executions, 10297 successful
rsg_test.go:832: 1m30s of 5m0s: 194657 executions, 10882 successful
rsg_test.go:832: 1m35s of 5m0s: 206728 executions, 11480 successful
rsg_test.go:832: 1m40s of 5m0s: 218400 executions, 12040 successful
rsg_test.go:832: 1m45s of 5m0s: 229942 executions, 12604 successful
rsg_test.go:832: 1m50s of 5m0s: 241597 executions, 13158 successful
rsg_test.go:832: 1m55s of 5m0s: 253943 executions, 13761 successful
rsg_test.go:832: 2m0s of 5m0s: 266059 executions, 14317 successful
rsg_test.go:832: 2m5s of 5m0s: 276465 executions, 14805 successful
rsg_test.go:832: 2m10s of 5m0s: 288190 executions, 15370 successful
rsg_test.go:832: 2m15s of 5m0s: 300051 executions, 15934 successful
rsg_test.go:832: 2m20s of 5m0s: 311384 executions, 16457 successful
rsg_test.go:832: 2m25s of 5m0s: 322533 executions, 16966 successful
rsg_test.go:832: 2m30s of 5m0s: 331957 executions, 17442 successful
rsg_test.go:832: 2m35s of 5m0s: 343858 executions, 17963 successful
rsg_test.go:832: 2m40s of 5m0s: 356058 executions, 18530 successful
rsg_test.go:832: 2m45s of 5m0s: 368528 executions, 19092 successful
rsg_test.go:832: 2m50s of 5m0s: 379969 executions, 19621 successful
rsg_test.go:832: 2m55s of 5m0s: 390832 executions, 20162 successful
rsg_test.go:832: 3m0s of 5m0s: 401776 executions, 20646 successful
rsg_test.go:832: 3m5s of 5m0s: 412687 executions, 21138 successful
rsg_test.go:832: 3m10s of 5m0s: 424595 executions, 21717 successful
rsg_test.go:832: 3m15s of 5m0s: 436181 executions, 22288 successful
rsg_test.go:832: 3m20s of 5m0s: 447687 executions, 22797 successful
rsg_test.go:832: 3m25s of 5m0s: 459526 executions, 23357 successful
rsg_test.go:832: 3m30s of 5m0s: 470709 executions, 23854 successful
rsg_test.go:832: 3m35s of 5m0s: 481609 executions, 24407 successful
rsg_test.go:832: 3m40s of 5m0s: 486925 executions, 24647 successful
rsg_test.go:832: 3m45s of 5m0s: 491161 executions, 24853 successful
rsg_test.go:832: 3m50s of 5m0s: 495304 executions, 25048 successful
rsg_test.go:832: 3m55s of 5m0s: 499025 executions, 25228 successful
rsg_test.go:832: 4m0s of 5m0s: 503294 executions, 25446 successful
rsg_test.go:832: 4m5s of 5m0s: 508213 executions, 25671 successful
rsg_test.go:832: 4m10s of 5m0s: 513178 executions, 25906 successful
rsg_test.go:832: 4m15s of 5m0s: 518304 executions, 26148 successful
rsg_test.go:832: 4m20s of 5m0s: 523432 executions, 26391 successful
rsg_test.go:832: 4m25s of 5m0s: 528217 executions, 26594 successful
rsg_test.go:832: 4m30s of 5m0s: 533237 executions, 26800 successful
rsg_test.go:832: 4m35s of 5m0s: 538015 executions, 27020 successful
rsg_test.go:832: 4m40s of 5m0s: 543098 executions, 27234 successful
rsg_test.go:832: 4m45s of 5m0s: 548108 executions, 27466 successful
rsg_test.go:832: 4m50s of 5m0s: 553372 executions, 27686 successful
rsg_test.go:832: 4m55s of 5m0s: 558231 executions, 27907 successful
rsg_test.go:868: 562914 executions, 28098 successful
rsg_test.go:876: Parse followed by Format is not idempotent: "ALTER INDEX IF EXISTS FAMILY . FAMILY . ident VISIBILITY 7.379652426127388e-12" -> "ALTER INDEX IF EXISTS \"family\".\"family\".ident VISIBILITY 0.00" != "ALTER INDEX IF EXISTS \"family\".\"family\".ident NOT VISIBLE"
rsg_test.go:302: -- test log scope end --
test logs left over in: /artifacts/tmp/_tmp/d437d2c847dfedbc4972f231c3331c8e/logTestRandomSyntaxGeneration135218831
--- FAIL: TestRandomSyntaxGeneration (302.11s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-foundations
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRandomSyntaxGeneration.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
sql tests testrandomsyntaxgeneration failed sql tests testrandomsyntaxgeneration with on master rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go executions successful rsg test go parse followed by format is not idempotent alter index if exists family family ident visibility alter index if exists family family ident visibility alter index if exists family family ident not visible rsg test go test log scope end test logs left over in artifacts tmp tmp fail testrandomsyntaxgeneration help see also cc cockroachdb sql foundations
| 1
|
17,327
| 3,604,777,956
|
IssuesEvent
|
2016-02-04 00:27:36
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
Add gce-slow suite to the merge queue blockers
|
area/test priority/P1 team/test-infra
|
There are few important tests in this suite so I believe it would be useful to add it to blockers. We need to deflake it first though:
- #17518
- #17781
cc @ixdy @smarterclayton @brendandburns @wojtek-t @wonderfly
|
2.0
|
Add gce-slow suite to the merge queue blockers - There are few important tests in this suite so I believe it would be useful to add it to blockers. We need to deflake it first though:
- #17518
- #17781
cc @ixdy @smarterclayton @brendandburns @wojtek-t @wonderfly
|
test
|
add gce slow suite to the merge queue blockers there are few important tests in this suite so i believe it would be useful to add it to blockers we need to deflake it first though cc ixdy smarterclayton brendandburns wojtek t wonderfly
| 1
|
301,198
| 22,740,991,212
|
IssuesEvent
|
2022-07-07 03:45:19
|
lucasGelpi/ISPCGrup5Aula11
|
https://api.github.com/repos/lucasGelpi/ISPCGrup5Aula11
|
closed
|
Extraer conceptos
|
documentation
|
Extraer del material de Scrum los siguientes conceptos: Ceremonias, Artefactos y Roles.
|
1.0
|
Extraer conceptos - Extraer del material de Scrum los siguientes conceptos: Ceremonias, Artefactos y Roles.
|
non_test
|
extraer conceptos extraer del material de scrum los siguientes conceptos ceremonias artefactos y roles
| 0
|
39,292
| 5,226,251,643
|
IssuesEvent
|
2017-01-27 20:46:42
|
isenseDev/rSENSE
|
https://api.github.com/repos/isenseDev/rSENSE
|
closed
|
Allow pie charts with data sets without number fields
|
Feature In Testing UI Vis
|
To simplify the use of pie charts, we'd like to modify pie charts so that when there are no number fields, the user can still see the chart, though restricted to only the row count view.
|
1.0
|
Allow pie charts with data sets without number fields - To simplify the use of pie charts, we'd like to modify pie charts so that when there are no number fields, the user can still see the chart, though restricted to only the row count view.
|
test
|
allow pie charts with data sets without number fields to simplify the use of pie charts we d like to modify pie charts so that when there are no number fields the user can still see the chart though restricted to only the row count view
| 1
|
11,557
| 3,007,165,050
|
IssuesEvent
|
2015-07-27 14:49:12
|
owncloud/gallery-old
|
https://api.github.com/repos/owncloud/gallery-old
|
closed
|
Change pictures view button is not visible using IE11 as browser
|
browser:ie bug design enhancement
|
### Steps to reproduce
1. Using IE 11 as browser, go to Pictures menu
2. Share the pictures by link
3. Using the previous link change the view of the pictures
### Expected behaviour
The change view button should be more visible and not located just in the right
### Actual behaviour
The button is only half-visible due to the slide bar
### Server configuration
**Operating system**: ubuntu 14.04
**Web server:** apache 2.4.7
**Database:** mysql
**PHP version:** 5.5.4
**ownCloud version:** OC 8.1.0.8
**Updated from an older ownCloud or fresh install:** fresh install


|
1.0
|
Change pictures view button is not visible using IE11 as browser - ### Steps to reproduce
1. Using IE 11 as browser, go to Pictures menu
2. Share the pictures by link
3. Using the previous link change the view of the pictures
### Expected behaviour
The change view button should be more visible and not located just in the right
### Actual behaviour
The button is only half-visible due to the slide bar
### Server configuration
**Operating system**: ubuntu 14.04
**Web server:** apache 2.4.7
**Database:** mysql
**PHP version:** 5.5.4
**ownCloud version:** OC 8.1.0.8
**Updated from an older ownCloud or fresh install:** fresh install


|
non_test
|
change pictures view button is not visible using as browser steps to reproduce using ie as browser go to pictures menu share the pictures by link using the previous link change the view of the pictures expected behaviour the change view button should be more visible and not located just in the right actual behaviour the button is only half visible due to the slide bar server configuration operating system ubuntu web server apache database mysql php version owncloud version oc updated from an older owncloud or fresh install fresh install
| 0
|
460,320
| 13,208,162,602
|
IssuesEvent
|
2020-08-15 02:48:10
|
larissatrochta/THP-Website-QA
|
https://api.github.com/repos/larissatrochta/THP-Website-QA
|
opened
|
Search Icon Margin Inconsistency and No Hovering Effect
|
enhancement low severity medium priority
|
On the top nav menu, the spacing between "My Account" and the Search icon isn't even - it seems a lot smaller than the rest of the list items. Also, there is no hover effect (changing color) when you hover it with your mouse, like the rest of the menu items.
Steps to reproduce
1) Visit https://thehoneypot.larissatrochta.ca/ on a desktop view
2) Hover mouse above the Search icon (magnifying glass)
Expected Results
Consistent styling is applied to the Search icon as with the rest of the menu items.
Actual Results
Search icon's margin and styling is inconsistent and should be amended to fit with the rest of the menu items.

|
1.0
|
Search Icon Margin Inconsistency and No Hovering Effect - On the top nav menu, the spacing between "My Account" and the Search icon isn't even - it seems a lot smaller than the rest of the list items. Also, there is no hover effect (changing color) when you hover it with your mouse, like the rest of the menu items.
Steps to reproduce
1) Visit https://thehoneypot.larissatrochta.ca/ on a desktop view
2) Hover mouse above the Search icon (magnifying glass)
Expected Results
Consistent styling is applied to the Search icon as with the rest of the menu items.
Actual Results
Search icon's margin and styling is inconsistent and should be amended to fit with the rest of the menu items.

|
non_test
|
search icon margin inconsistency and no hovering effect on the top nav menu the spacing between my account and the search icon isn t even it seems a lot smaller than the rest of the list items also there is no hover effect changing color when you hover it with your mouse like the rest of the menu items steps to reproduce visit on a desktop view hover mouse above the search icon magnifying glass expected results consistent styling is applied to the search icon as with the rest of the menu items actual results search icon s margin and styling is inconsistent and should be amended to fit with the rest of the menu items
| 0
|
10,646
| 13,446,217,972
|
IssuesEvent
|
2020-09-08 12:38:58
|
MHRA/products
|
https://api.github.com/repos/MHRA/products
|
closed
|
DOCUMENTATION - PARs Process Map
|
EPIC - PARs process
|
### User want
As a internal user I would like to have access to a PARs process map, so that I can understand / amend the process in the future.
### Acceptance Criteria
**Customer acceptance criteria**
The process map follows the BPMN methodology
Agency “House” Best Practice / Style guidelines have been used
The process map plots the end to end PARs upload / amend / delete process
The process map can be imported into Visual Paradigm (XMI or XML formats - XMI standards preferred)
**Technical acceptance criteria**
**Data acceptance criteria**
**Testing acceptance criteria**
**Data - Potential impact**
**Size**
**Value**
**Effort**
### Exit Criteria met
- [ ] Backlog
- [ ] Discovery
- [ ] DUXD
- [ ] Development
- [ ] Quality Assurance
- [ ] Release and Validate
|
1.0
|
DOCUMENTATION - PARs Process Map - ### User want
As a internal user I would like to have access to a PARs process map, so that I can understand / amend the process in the future.
### Acceptance Criteria
**Customer acceptance criteria**
The process map follows the BPMN methodology
Agency “House” Best Practice / Style guidelines have been used
The process map plots the end to end PARs upload / amend / delete process
The process map can be imported into Visual Paradigm (XMI or XML formats - XMI standards preferred)
**Technical acceptance criteria**
**Data acceptance criteria**
**Testing acceptance criteria**
**Data - Potential impact**
**Size**
**Value**
**Effort**
### Exit Criteria met
- [ ] Backlog
- [ ] Discovery
- [ ] DUXD
- [ ] Development
- [ ] Quality Assurance
- [ ] Release and Validate
|
non_test
|
documentation pars process map user want as a internal user i would like to have access to a pars process map so that i can understand amend the process in the future acceptance criteria customer acceptance criteria the process map follows the bpmn methodology agency “house” best practice style guidelines have been used the process map plots the end to end pars upload amend delete process the process map can be imported into visual paradigm xmi or xml formats xmi standards preferred technical acceptance criteria data acceptance criteria testing acceptance criteria data potential impact size value effort exit criteria met backlog discovery duxd development quality assurance release and validate
| 0
|
40,613
| 5,310,890,060
|
IssuesEvent
|
2017-02-12 23:46:40
|
nuxsmin/sysPass
|
https://api.github.com/repos/nuxsmin/sysPass
|
closed
|
Import fails due to Duplicated customer error
|
NeedTests
|
Hello.
If we have several accounts with the same customer it fails to import the entire list of accounts. porobably category has the same error.
This worked fine in version 1.x. I understand that if category exists it should grab the id and use that one, but not attempting to create it again from scratch.
Can you have a look please?
Albert.
|
1.0
|
Import fails due to Duplicated customer error - Hello.
If we have several accounts with the same customer it fails to import the entire list of accounts. porobably category has the same error.
This worked fine in version 1.x. I understand that if category exists it should grab the id and use that one, but not attempting to create it again from scratch.
Can you have a look please?
Albert.
|
test
|
import fails due to duplicated customer error hello if we have several accounts with the same customer it fails to import the entire list of accounts porobably category has the same error this worked fine in version x i understand that if category exists it should grab the id and use that one but not attempting to create it again from scratch can you have a look please albert
| 1
|
115,632
| 14,858,146,916
|
IssuesEvent
|
2021-01-18 16:22:59
|
carbon-design-system/carbon-for-ibm-dotcom
|
https://api.github.com/repos/carbon-design-system/carbon-for-ibm-dotcom
|
closed
|
Carousel component design specs - Target: Jan 15
|
Airtable Done Design specs design sprint demo
|
<!-- Avoid any type of solutions in this user story -->
<!-- replace _{{...}}_ with your own words or remove -->
#### User Story
<!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} -->
> As a `[user role below]`:
dds designer
> I need to:
provide carousel design specs
> so that:
developers and adopters can reference the specs
#### Additional information
After the decision to develop [Carousel](https://ibmdotcom-web-components-canary.mybluemix.net/?path=/story/components-carousel--default) as a separate component from [Card section – Carousel](https://ibmdotcom-web-components-canary.mybluemix.net/?path=/story/components-card-section-carousel--default), we need design specs for Carousel on its own.
Meet with James and possibly a developer to define what is possible within the Carousel component. Things to consider:
- single image
- multiple images
- within light box
#### Acceptance criteria
- [x] Design specs added to [Carousel component Box folder](https://ibm.ent.box.com/folder/121172699187?s=hmni7qhero9mf5rn593zij0tcay8h9dw)
- [ ] Any updates to Carousel component defined in issue(s) for development
- [x] Sign off from James and Wonil
<!-- Consider the following when writing Acceptance criteria for this story. -->
<!-- *** Each product backlog item or user story should have at least one Acceptance criteria. -->
<!-- *** Acceptance criteria defines a deliverable that can be completed in a single sprint -->
<!-- *** Each Acceptance criterion is independently testable. -->
<!-- *** Include functional as well as non-functional criteria – when relevant. -->
<!-- *** Team members write Acceptance criteria and the Product Owner verifies it. -->
|
2.0
|
Carousel component design specs - Target: Jan 15 - <!-- Avoid any type of solutions in this user story -->
<!-- replace _{{...}}_ with your own words or remove -->
#### User Story
<!-- {{Provide a detailed description of the user's need here, but avoid any type of solutions}} -->
> As a `[user role below]`:
dds designer
> I need to:
provide carousel design specs
> so that:
developers and adopters can reference the specs
#### Additional information
After the decision to develop [Carousel](https://ibmdotcom-web-components-canary.mybluemix.net/?path=/story/components-carousel--default) as a separate component from [Card section – Carousel](https://ibmdotcom-web-components-canary.mybluemix.net/?path=/story/components-card-section-carousel--default), we need design specs for Carousel on its own.
Meet with James and possibly a developer to define what is possible within the Carousel component. Things to consider:
- single image
- multiple images
- within light box
#### Acceptance criteria
- [x] Design specs added to [Carousel component Box folder](https://ibm.ent.box.com/folder/121172699187?s=hmni7qhero9mf5rn593zij0tcay8h9dw)
- [ ] Any updates to Carousel component defined in issue(s) for development
- [x] Sign off from James and Wonil
<!-- Consider the following when writing Acceptance criteria for this story. -->
<!-- *** Each product backlog item or user story should have at least one Acceptance criteria. -->
<!-- *** Acceptance criteria defines a deliverable that can be completed in a single sprint -->
<!-- *** Each Acceptance criterion is independently testable. -->
<!-- *** Include functional as well as non-functional criteria – when relevant. -->
<!-- *** Team members write Acceptance criteria and the Product Owner verifies it. -->
|
non_test
|
carousel component design specs target jan user story as a dds designer i need to provide carousel design specs so that developers and adopters can reference the specs additional information after the decision to develop as a separate component from we need design specs for carousel on its own meet with james and possibly a developer to define what is possible within the carousel component things to consider single image multiple images within light box acceptance criteria design specs added to any updates to carousel component defined in issue s for development sign off from james and wonil
| 0
|
714,902
| 24,580,067,350
|
IssuesEvent
|
2022-10-13 15:00:04
|
longhorn/longhorn
|
https://api.github.com/repos/longhorn/longhorn
|
closed
|
[TASK] Refresh distro version support for 1.4
|
priority/2 require/doc kind/task
|
## What's the task? Please describe
Longhorn supports FS volume which actually relies on some kernel modules to achieve that like ext4 and xfs. In different distros w/ different Kernel versions, probably there will some compatibility problems like unsupported mount options to cause the volume mount to fail. The doc needs to update to ensure people understand the criteria for the host kernel.
## Describe the items of the task (DoD, definition of done) you'd like
- [x] Update the supported distro in the master (1.4), 1.3, and 1.2 documentation (https://longhorn.io/docs/1.3.0/best-practices/#software) @yangchiu has updated versions for 1.4, https://longhorn.io/docs/1.4.0/best-practices/#software and for older releases, just keep it as is.
- [ ] ~Update the environment check script to provide early notice to users~
## Additional context
https://github.com/longhorn/longhorn/issues/2640
https://github.com/ceph/ceph-csi/issues/966#issuecomment-620655796
cc @longhorn/qa @PhanLe1010 @derekbit
|
1.0
|
[TASK] Refresh distro version support for 1.4 - ## What's the task? Please describe
Longhorn supports FS volume which actually relies on some kernel modules to achieve that like ext4 and xfs. In different distros w/ different Kernel versions, probably there will some compatibility problems like unsupported mount options to cause the volume mount to fail. The doc needs to update to ensure people understand the criteria for the host kernel.
## Describe the items of the task (DoD, definition of done) you'd like
- [x] Update the supported distro in the master (1.4), 1.3, and 1.2 documentation (https://longhorn.io/docs/1.3.0/best-practices/#software) @yangchiu has updated versions for 1.4, https://longhorn.io/docs/1.4.0/best-practices/#software and for older releases, just keep it as is.
- [ ] ~Update the environment check script to provide early notice to users~
## Additional context
https://github.com/longhorn/longhorn/issues/2640
https://github.com/ceph/ceph-csi/issues/966#issuecomment-620655796
cc @longhorn/qa @PhanLe1010 @derekbit
|
non_test
|
refresh distro version support for what s the task please describe longhorn supports fs volume which actually relies on some kernel modules to achieve that like and xfs in different distros w different kernel versions probably there will some compatibility problems like unsupported mount options to cause the volume mount to fail the doc needs to update to ensure people understand the criteria for the host kernel describe the items of the task dod definition of done you d like update the supported distro in the master and documentation yangchiu has updated versions for and for older releases just keep it as is update the environment check script to provide early notice to users additional context cc longhorn qa derekbit
| 0
|
647,829
| 21,156,775,584
|
IssuesEvent
|
2022-04-07 04:47:15
|
AY2122S2-CS2103T-T12-4/tp
|
https://api.github.com/repos/AY2122S2-CS2103T-T12-4/tp
|
closed
|
[PE-D] Duplicate tasks
|
priority.Medium
|
Are duplicate tasks allowed in the application? I am able to enter duplicate tasks (same name & same time)
Perhaps the team can clarify in the User Guide

<!--session: 1648792874626-9f3767a9-6a10-44f5-bc61-bc6b62ea7cef-->
<!--Version: Web v3.4.2-->
-------------
Labels: `severity.Medium` `type.DocumentationBug`
original: waffledood/ped#9
|
1.0
|
[PE-D] Duplicate tasks - Are duplicate tasks allowed in the application? I am able to enter duplicate tasks (same name & same time)
Perhaps the team can clarify in the User Guide

<!--session: 1648792874626-9f3767a9-6a10-44f5-bc61-bc6b62ea7cef-->
<!--Version: Web v3.4.2-->
-------------
Labels: `severity.Medium` `type.DocumentationBug`
original: waffledood/ped#9
|
non_test
|
duplicate tasks are duplicate tasks allowed in the application i am able to enter duplicate tasks same name same time perhaps the team can clarify in the user guide labels severity medium type documentationbug original waffledood ped
| 0
|
112,270
| 4,514,969,949
|
IssuesEvent
|
2016-09-05 03:41:31
|
OperationCode/operationcode
|
https://api.github.com/repos/OperationCode/operationcode
|
closed
|
Add split gem
|
Priority: Medium Status: Available Type: Enhancement
|
https://github.com/splitrb/split
This gem will allow us to do self hosted A/B testing.
Please check to make sure the management console is protected by at least basic auth
|
1.0
|
Add split gem - https://github.com/splitrb/split
This gem will allow us to do self hosted A/B testing.
Please check to make sure the management console is protected by at least basic auth
|
non_test
|
add split gem this gem will allow us to do self hosted a b testing please check to make sure the management console is protected by at least basic auth
| 0
|
14,871
| 11,207,889,680
|
IssuesEvent
|
2020-01-06 05:49:21
|
APSIMInitiative/ApsimX
|
https://api.github.com/repos/APSIMInitiative/ApsimX
|
closed
|
Scope caching issues
|
bug interface/infrastructure
|
Adding or removing models from a simulation doesn't cause the scope cache to be updated. This is easy enough to reproduce:
1. Open up any file with a simulation which doesn't contain a crop (e.g. factorial example)
2. Add fertilise at sowing script
3. Click on fertilise at sowing script - crop drop-down is empty (as it should be)
4. Add a crop to the paddock
5. Click on fertilise at sowing script - crop drop-down is still empty (it shouldn't be!)
The only workaround currently is to close the file and reopen it. I'm not sure what the best solution would be. If the user has clicked on every component in the file, then updating the cache for each component could be time-consuming, even if we only update those models in scope of the newly-added model.
A simpler option might be to erase the cache whenever we add/remove a model. This doesn't sound ideal but it would be faster than the first option so I think I will go with this option unless anyone has a better idea.
|
1.0
|
Scope caching issues - Adding or removing models from a simulation doesn't cause the scope cache to be updated. This is easy enough to reproduce:
1. Open up any file with a simulation which doesn't contain a crop (e.g. factorial example)
2. Add fertilise at sowing script
3. Click on fertilise at sowing script - crop drop-down is empty (as it should be)
4. Add a crop to the paddock
5. Click on fertilise at sowing script - crop drop-down is still empty (it shouldn't be!)
The only workaround currently is to close the file and reopen it. I'm not sure what the best solution would be. If the user has clicked on every component in the file, then updating the cache for each component could be time-consuming, even if we only update those models in scope of the newly-added model.
A simpler option might be to erase the cache whenever we add/remove a model. This doesn't sound ideal but it would be faster than the first option so I think I will go with this option unless anyone has a better idea.
|
non_test
|
scope caching issues adding or removing models from a simulation doesn t cause the scope cache to be updated this is easy enough to reproduce open up any file with a simulation which doesn t contain a crop e g factorial example add fertilise at sowing script click on fertilise at sowing script crop drop down is empty as it should be add a crop to the paddock click on fertilise at sowing script crop drop down is still empty it shouldn t be the only workaround currently is to close the file and reopen it i m not sure what the best solution would be if the user has clicked on every component in the file then updating the cache for each component could be time consuming even if we only update those models in scope of the newly added model a simpler option might be to erase the cache whenever we add remove a model this doesn t sound ideal but it would be faster than the first option so i think i will go with this option unless anyone has a better idea
| 0
|
24,638
| 24,078,334,211
|
IssuesEvent
|
2022-09-19 02:10:24
|
FreeTubeApp/FreeTube
|
https://api.github.com/repos/FreeTubeApp/FreeTube
|
closed
|
[Bug]: [Wayland] PiP video is laggy when the main Freetube window is minimised
|
bug U: Waiting for Response from Author B: visual B: usability
|
### Guidelines
- [X] I have encountered this bug in the [latest release of FreeTube](https://github.com/FreeTubeApp/FreeTube/releases).
- [X] I have searched the [issue tracker](https://github.com/FreeTubeApp/FreeTube/issues) for a bug report that matches the one I want to file, without success.
### Describe the bug
The video inside the picture in picture window lags when the main window is minimized. This does not happen when the main window is not minimized.
**How to reproduce?**
1. Launch Freetube under Wayland using `freetube --enable-features=UseOzonePlatform --ozone-platform=wayland`
2. Play a video.
3. Select Picture in picture mode.
4. Minimize the main window
5. The video in picture in picture lags as soon as we minimize the main Freetube window.
Here's the bug in action:
https://user-images.githubusercontent.com/32611672/152211552-a183b157-299f-45f5-b29a-b82a08a5c5a4.mp4
### Expected Behavior
The video in picture in picture should not lag even if the main window is minimized.
### Issue Labels
usability issue, visual bug
### FreeTube Version
v0.15.1 Beta
### Operating System Version
Arch Linux
### Installation Method (If applicable)
other
### Last Known Working FreeTube Version (If Any)
_No response_
### Primary API used
Local API
### Additional Information
_No response_
### Nightly Build
- [ ] I have encountered this bug in the latest [nightly build](https://docs.freetubeapp.io/development/nightly-builds).
|
True
|
[Bug]: [Wayland] PiP video is laggy when the main Freetube window is minimised - ### Guidelines
- [X] I have encountered this bug in the [latest release of FreeTube](https://github.com/FreeTubeApp/FreeTube/releases).
- [X] I have searched the [issue tracker](https://github.com/FreeTubeApp/FreeTube/issues) for a bug report that matches the one I want to file, without success.
### Describe the bug
The video inside the picture in picture window lags when the main window is minimized. This does not happen when the main window is not minimized.
**How to reproduce?**
1. Launch Freetube under Wayland using `freetube --enable-features=UseOzonePlatform --ozone-platform=wayland`
2. Play a video.
3. Select Picture in picture mode.
4. Minimize the main window
5. The video in picture in picture lags as soon as we minimize the main Freetube window.
Here's the bug in action:
https://user-images.githubusercontent.com/32611672/152211552-a183b157-299f-45f5-b29a-b82a08a5c5a4.mp4
### Expected Behavior
The video in picture in picture should not lag even if the main window is minimized.
### Issue Labels
usability issue, visual bug
### FreeTube Version
v0.15.1 Beta
### Operating System Version
Arch Linux
### Installation Method (If applicable)
other
### Last Known Working FreeTube Version (If Any)
_No response_
### Primary API used
Local API
### Additional Information
_No response_
### Nightly Build
- [ ] I have encountered this bug in the latest [nightly build](https://docs.freetubeapp.io/development/nightly-builds).
|
non_test
|
pip video is laggy when the main freetube window is minimised guidelines i have encountered this bug in the i have searched the for a bug report that matches the one i want to file without success describe the bug the video inside the picture in picture window lags when the main window is minimized this does not happen when the main window is not minimized how to reproduce launch freetube under wayland using freetube enable features useozoneplatform ozone platform wayland play a video select picture in picture mode minimize the main window the video in picture in picture lags as soon as we minimize the main freetube window here s the bug in action expected behavior the video in picture in picture should not lag even if the main window is minimized issue labels usability issue visual bug freetube version beta operating system version arch linux installation method if applicable other last known working freetube version if any no response primary api used local api additional information no response nightly build i have encountered this bug in the latest
| 0
|
288,109
| 24,882,768,478
|
IssuesEvent
|
2022-10-28 03:47:09
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Orçamento - Execução - Pedra do Indaiá
|
generalization test development template - Memory (66) tag - Orçamento subtag - Execução
|
DoD: Realizar o teste de Generalização do validador da tag Orçamento - Execução para o Município de Pedra do Indaiá.
|
1.0
|
Teste de generalizacao para a tag Orçamento - Execução - Pedra do Indaiá - DoD: Realizar o teste de Generalização do validador da tag Orçamento - Execução para o Município de Pedra do Indaiá.
|
test
|
teste de generalizacao para a tag orçamento execução pedra do indaiá dod realizar o teste de generalização do validador da tag orçamento execução para o município de pedra do indaiá
| 1
|
61,279
| 6,732,034,231
|
IssuesEvent
|
2017-10-18 09:52:35
|
tgstation/tgstation
|
https://api.github.com/repos/tgstation/tgstation
|
closed
|
Particle Accelerator\Engine Starter orientation
|
Bug Needs Reproducing/Testing
|
Though it's been a little while but there is a painfully annoying bug that makes it so that if you orientate the Particle Accelerator to shoot left or right the console detects that it's missing parts making setting up a makeshift tesla or singulo very annoying to set up if you make the mistake of orientating it to shoot at it from the left or right.
When I retest this on tg or on a single player server I'll get some pictures to show those viewing this post what I mean.
|
1.0
|
Particle Accelerator\Engine Starter orientation - Though it's been a little while but there is a painfully annoying bug that makes it so that if you orientate the Particle Accelerator to shoot left or right the console detects that it's missing parts making setting up a makeshift tesla or singulo very annoying to set up if you make the mistake of orientating it to shoot at it from the left or right.
When I retest this on tg or on a single player server I'll get some pictures to show those viewing this post what I mean.
|
test
|
particle accelerator engine starter orientation though it s been a little while but there is a painfully annoying bug that makes it so that if you orientate the particle accelerator to shoot left or right the console detects that it s missing parts making setting up a makeshift tesla or singulo very annoying to set up if you make the mistake of orientating it to shoot at it from the left or right when i retest this on tg or on a single player server i ll get some pictures to show those viewing this post what i mean
| 1
|
261,402
| 22,743,612,996
|
IssuesEvent
|
2022-07-07 07:09:21
|
WoWManiaUK/Redemption
|
https://api.github.com/repos/WoWManiaUK/Redemption
|
closed
|
[Zone] Wintergrasp spirit guides - Graveyard spirit guides arent switching appropriately
|
Fixed on PTR - Tester Confirmed
|
While we were fixing other WG things we noticed
**What is Happening:**
The graveyard spirit guides were never switching appropriately based on graveyard team control.
**What Should happen:**
Graveyard spirit guides should switch to approprate team control
Franca is already on this fix @GMZinius
|
1.0
|
[Zone] Wintergrasp spirit guides - Graveyard spirit guides arent switching appropriately - While we were fixing other WG things we noticed
**What is Happening:**
The graveyard spirit guides were never switching appropriately based on graveyard team control.
**What Should happen:**
Graveyard spirit guides should switch to approprate team control
Franca is already on this fix @GMZinius
|
test
|
wintergrasp spirit guides graveyard spirit guides arent switching appropriately while we were fixing other wg things we noticed what is happening the graveyard spirit guides were never switching appropriately based on graveyard team control what should happen graveyard spirit guides should switch to approprate team control franca is already on this fix gmzinius
| 1
|
339,280
| 30,387,880,100
|
IssuesEvent
|
2023-07-13 03:26:15
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
closed
|
Fix paddle_tensor.test_paddle_instance_acosh
|
Sub Task Failing Test Paddle Frontend
|
| | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5538552236/jobs/10108602079"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5538552236/jobs/10108602079"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5538552236/jobs/10108602079"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5538552236/jobs/10108602079"><img src=https://img.shields.io/badge/-success-success></a>
|
1.0
|
Fix paddle_tensor.test_paddle_instance_acosh - | | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5538552236/jobs/10108602079"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5538552236/jobs/10108602079"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5538552236/jobs/10108602079"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5538552236/jobs/10108602079"><img src=https://img.shields.io/badge/-success-success></a>
|
test
|
fix paddle tensor test paddle instance acosh numpy a href src jax a href src tensorflow a href src torch a href src
| 1
|
196,132
| 22,420,027,234
|
IssuesEvent
|
2022-06-20 01:17:49
|
DavidSpek/kubeflow
|
https://api.github.com/repos/DavidSpek/kubeflow
|
opened
|
CVE-2022-29241 (High) detected in jupyter_server-1.9.0-py3-none-any.whl
|
security vulnerability
|
## CVE-2022-29241 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jupyter_server-1.9.0-py3-none-any.whl</b></p></summary>
<p>The backend—i.e. core services, APIs, and REST endpoints—to Jupyter web applications.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/29/b7/7377d007118f7798b21362a6c0a0bf20c93cdc19345105276a862e1263d6/jupyter_server-1.9.0-py3-none-any.whl">https://files.pythonhosted.org/packages/29/b7/7377d007118f7798b21362a6c0a0bf20c93cdc19345105276a862e1263d6/jupyter_server-1.9.0-py3-none-any.whl</a></p>
<p>Path to dependency file: /components/crud-web-apps/common/backend</p>
<p>Path to vulnerable library: /components/crud-web-apps/common/backend</p>
<p>
Dependency Hierarchy:
- :x: **jupyter_server-1.9.0-py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/DavidSpek/kubeflow/commit/00cbc9d11a3306fed1e979d79dff6ae36749d4bd">00cbc9d11a3306fed1e979d79dff6ae36749d4bd</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Jupyter Server provides the backend (i.e. the core services, APIs, and REST endpoints) for Jupyter web applications like Jupyter Notebook. Prior to version 1.17.1, if notebook server is started with a value of `root_dir` that contains the starting user's home directory, then the underlying REST API can be used to leak the access token assigned at start time by guessing/brute forcing the PID of the jupyter server. While this requires an authenticated user session, this URL can be used from a cross-site scripting payload or from a hooked or otherwise compromised browser to leak this access token to a malicious third party. This token can be used along with the REST API to interact with Jupyter services/notebooks such as modifying or overwriting critical files, such as .bashrc or .ssh/authorized_keys, allowing a malicious user to read potentially sensitive data and possibly gain control of the impacted system. This issue is patched in version 1.17.1.
<p>Publish Date: 2022-06-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29241>CVE-2022-29241</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jupyter-server/jupyter_server/security/advisories/GHSA-q874-g24w-4q9g">https://github.com/jupyter-server/jupyter_server/security/advisories/GHSA-q874-g24w-4q9g</a></p>
<p>Release Date: 2022-06-14</p>
<p>Fix Resolution: jupyter-server - 1.17.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-29241 (High) detected in jupyter_server-1.9.0-py3-none-any.whl - ## CVE-2022-29241 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jupyter_server-1.9.0-py3-none-any.whl</b></p></summary>
<p>The backend—i.e. core services, APIs, and REST endpoints—to Jupyter web applications.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/29/b7/7377d007118f7798b21362a6c0a0bf20c93cdc19345105276a862e1263d6/jupyter_server-1.9.0-py3-none-any.whl">https://files.pythonhosted.org/packages/29/b7/7377d007118f7798b21362a6c0a0bf20c93cdc19345105276a862e1263d6/jupyter_server-1.9.0-py3-none-any.whl</a></p>
<p>Path to dependency file: /components/crud-web-apps/common/backend</p>
<p>Path to vulnerable library: /components/crud-web-apps/common/backend</p>
<p>
Dependency Hierarchy:
- :x: **jupyter_server-1.9.0-py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/DavidSpek/kubeflow/commit/00cbc9d11a3306fed1e979d79dff6ae36749d4bd">00cbc9d11a3306fed1e979d79dff6ae36749d4bd</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Jupyter Server provides the backend (i.e. the core services, APIs, and REST endpoints) for Jupyter web applications like Jupyter Notebook. Prior to version 1.17.1, if notebook server is started with a value of `root_dir` that contains the starting user's home directory, then the underlying REST API can be used to leak the access token assigned at start time by guessing/brute forcing the PID of the jupyter server. While this requires an authenticated user session, this URL can be used from a cross-site scripting payload or from a hooked or otherwise compromised browser to leak this access token to a malicious third party. This token can be used along with the REST API to interact with Jupyter services/notebooks such as modifying or overwriting critical files, such as .bashrc or .ssh/authorized_keys, allowing a malicious user to read potentially sensitive data and possibly gain control of the impacted system. This issue is patched in version 1.17.1.
<p>Publish Date: 2022-06-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29241>CVE-2022-29241</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jupyter-server/jupyter_server/security/advisories/GHSA-q874-g24w-4q9g">https://github.com/jupyter-server/jupyter_server/security/advisories/GHSA-q874-g24w-4q9g</a></p>
<p>Release Date: 2022-06-14</p>
<p>Fix Resolution: jupyter-server - 1.17.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in jupyter server none any whl cve high severity vulnerability vulnerable library jupyter server none any whl the backend—i e core services apis and rest endpoints—to jupyter web applications library home page a href path to dependency file components crud web apps common backend path to vulnerable library components crud web apps common backend dependency hierarchy x jupyter server none any whl vulnerable library found in head commit a href found in base branch master vulnerability details jupyter server provides the backend i e the core services apis and rest endpoints for jupyter web applications like jupyter notebook prior to version if notebook server is started with a value of root dir that contains the starting user s home directory then the underlying rest api can be used to leak the access token assigned at start time by guessing brute forcing the pid of the jupyter server while this requires an authenticated user session this url can be used from a cross site scripting payload or from a hooked or otherwise compromised browser to leak this access token to a malicious third party this token can be used along with the rest api to interact with jupyter services notebooks such as modifying or overwriting critical files such as bashrc or ssh authorized keys allowing a malicious user to read potentially sensitive data and possibly gain control of the impacted system this issue is patched in version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jupyter server step up your open source security game with mend
| 0
|
258,670
| 22,338,212,732
|
IssuesEvent
|
2022-06-14 20:48:13
|
zkSNACKs/WalletWasabi
|
https://api.github.com/repos/zkSNACKs/WalletWasabi
|
closed
|
Gained money by participating in CoinJoins?
|
question/research priority ww2 testing
|
### General Description
Did the Wallet just gained more money by doing lots of CoinJoin rounds??
The wallets (testnet) are from the non-stop coinjoining machine, so the wallets are a bit crazy. They got 1000+ UTXOs , 600+ used addresses and so.
### Screenshots
<img width="1440" alt="Screenshot 2022-06-13 at 13 30 11" src="https://user-images.githubusercontent.com/45069029/173344924-49300dd5-2835-46ba-a97a-dd4f419b952a.png">
<img width="1440" alt="Screenshot 2022-06-13 at 13 36 58" src="https://user-images.githubusercontent.com/45069029/173345424-cf7cd08b-bcf2-4a48-810b-921dbe66d0ea.png">
<img width="1440" alt="Screenshot 2022-06-13 at 13 45 33" src="https://user-images.githubusercontent.com/45069029/173346778-86de949f-00b6-4755-a4cd-dd793d892a5d.png">
### Operating System
macOS with the coinjoining "bots"
|
1.0
|
Gained money by participating in CoinJoins? - ### General Description
Did the Wallet just gained more money by doing lots of CoinJoin rounds??
The wallets (testnet) are from the non-stop coinjoining machine, so the wallets are a bit crazy. They got 1000+ UTXOs , 600+ used addresses and so.
### Screenshots
<img width="1440" alt="Screenshot 2022-06-13 at 13 30 11" src="https://user-images.githubusercontent.com/45069029/173344924-49300dd5-2835-46ba-a97a-dd4f419b952a.png">
<img width="1440" alt="Screenshot 2022-06-13 at 13 36 58" src="https://user-images.githubusercontent.com/45069029/173345424-cf7cd08b-bcf2-4a48-810b-921dbe66d0ea.png">
<img width="1440" alt="Screenshot 2022-06-13 at 13 45 33" src="https://user-images.githubusercontent.com/45069029/173346778-86de949f-00b6-4755-a4cd-dd793d892a5d.png">
### Operating System
macOS with the coinjoining "bots"
|
test
|
gained money by participating in coinjoins general description did the wallet just gained more money by doing lots of coinjoin rounds the wallets testnet are from the non stop coinjoining machine so the wallets are a bit crazy they got utxos used addresses and so screenshots img width alt screenshot at src img width alt screenshot at src img width alt screenshot at src operating system macos with the coinjoining bots
| 1
|
560,682
| 16,601,567,203
|
IssuesEvent
|
2021-06-01 20:15:00
|
google/ground-platform
|
https://api.github.com/repos/google/ground-platform
|
opened
|
[Terra Bio] Observations won't save for a layer with single optional question
|
priority: p0 type: bug
|
**Describe the bug**
Added observations aren't saving for a layers with imported JSON features:
https://user-images.githubusercontent.com/11237600/120384573-5479a880-c2f4-11eb-846c-89d88e3e712a.mov
|
1.0
|
[Terra Bio] Observations won't save for a layer with single optional question - **Describe the bug**
Added observations aren't saving for a layers with imported JSON features:
https://user-images.githubusercontent.com/11237600/120384573-5479a880-c2f4-11eb-846c-89d88e3e712a.mov
|
non_test
|
observations won t save for a layer with single optional question describe the bug added observations aren t saving for a layers with imported json features
| 0
|
279,062
| 24,196,045,956
|
IssuesEvent
|
2022-09-23 23:52:34
|
wazuh/wazuh-qa
|
https://api.github.com/repos/wazuh/wazuh-qa
|
opened
|
QA testing - Inaccurate macOS display name
|
team/qa type/qa-testing status/not-tracked role/qa-drop-table
|
| Target version | Related issue | Related PR |
|--------------------|--------------------|-----------------|
|4.4|https://github.com/wazuh/wazuh#14729 |wazuh/wazuh#14822|
## Description
This issue aims to request Manual testing on PR wazuh/wazuh#14822.
**Short problem description**: some macOS OS versions (from Sierra <oldest Wazuh support> to Catalina) show themselves (in the default OS information tool) as `MacOS X`, while their proper name is `macOS`. This inaccurate OS name is affecting both OS inventory information and Manager's agent list information. [Analysis](https://github.com/wazuh/wazuh/issues/14729#issuecomment-1239632323)
### Changes
- **Short problem proposed solution**: change default tool `sw_vers` to `system_profiler` to internally retrieve OS name
- Change inventory retrieving module (syscollector module) key component (sysinfo/data_provider library) responsible for retrieving OS information. **THIS CODE IS ONLY COMPILED AND USED ON MACOS SYSTEMS**.
- Change **agent side** mechanism responsible for updating agent information on the manager (wazuh-agentd' keepalive)
- **Known/considered scope**
- Agent inventory (OS)
- Agent information stored on the Manager agent list
- Vuln detector: **out research** indicate that VD is not affected by proposed changes
## Steps to reproduce
- Install Wazuh agent on any macOS system from Sierra to the last supported one (there's no official information from Wazuh side. @tsarquis88 test on Monterrey )
## Results before changes
- Wazuh Manager show, in every possible data interface (like API, Dashboard, Manager's CLI tools) `MacOS X` as OS name from Sierra to Catalina, and `macOS` from BigSur to the last supported one
## Results after changes
- Wazuh Manager show, in every possible data interface (like API, Dashboard, Manager's CLI tools) `macOS` as OS name from Sierra to the last supported one
|
1.0
|
QA testing - Inaccurate macOS display name - | Target version | Related issue | Related PR |
|--------------------|--------------------|-----------------|
|4.4|https://github.com/wazuh/wazuh#14729 |wazuh/wazuh#14822|
## Description
This issue aims to request Manual testing on PR wazuh/wazuh#14822.
**Short problem description**: some macOS OS versions (from Sierra <oldest Wazuh support> to Catalina) show themselves (in the default OS information tool) as `MacOS X`, while their proper name is `macOS`. This inaccurate OS name is affecting both OS inventory information and Manager's agent list information. [Analysis](https://github.com/wazuh/wazuh/issues/14729#issuecomment-1239632323)
### Changes
- **Short problem proposed solution**: change default tool `sw_vers` to `system_profiler` to internally retrieve OS name
- Change inventory retrieving module (syscollector module) key component (sysinfo/data_provider library) responsible for retrieving OS information. **THIS CODE IS ONLY COMPILED AND USED ON MACOS SYSTEMS**.
- Change **agent side** mechanism responsible for updating agent information on the manager (wazuh-agentd' keepalive)
- **Known/considered scope**
- Agent inventory (OS)
- Agent information stored on the Manager agent list
- Vuln detector: **out research** indicate that VD is not affected by proposed changes
## Steps to reproduce
- Install Wazuh agent on any macOS system from Sierra to the last supported one (there's no official information from Wazuh side. @tsarquis88 test on Monterrey )
## Results before changes
- Wazuh Manager show, in every possible data interface (like API, Dashboard, Manager's CLI tools) `MacOS X` as OS name from Sierra to Catalina, and `macOS` from BigSur to the last supported one
## Results after changes
- Wazuh Manager show, in every possible data interface (like API, Dashboard, Manager's CLI tools) `macOS` as OS name from Sierra to the last supported one
|
test
|
qa testing inaccurate macos display name target version related issue related pr wazuh wazuh description this issue aims to request manual testing on pr wazuh wazuh short problem description some macos os versions from sierra to catalina show themselves in the default os information tool as macos x while their proper name is macos this inaccurate os name is affecting both os inventory information and manager s agent list information changes short problem proposed solution change default tool sw vers to system profiler to internally retrieve os name change inventory retrieving module syscollector module key component sysinfo data provider library responsible for retrieving os information this code is only compiled and used on macos systems change agent side mechanism responsible for updating agent information on the manager wazuh agentd keepalive known considered scope agent inventory os agent information stored on the manager agent list vuln detector out research indicate that vd is not affected by proposed changes steps to reproduce install wazuh agent on any macos system from sierra to the last supported one there s no official information from wazuh side test on monterrey results before changes wazuh manager show in every possible data interface like api dashboard manager s cli tools macos x as os name from sierra to catalina and macos from bigsur to the last supported one results after changes wazuh manager show in every possible data interface like api dashboard manager s cli tools macos as os name from sierra to the last supported one
| 1
|
429,158
| 12,421,612,518
|
IssuesEvent
|
2020-05-23 17:42:51
|
harmony-one/experiment-deploy
|
https://api.github.com/repos/harmony-one/experiment-deploy
|
closed
|
[grafana] monitor sockstat from node_exporter
|
enhancement medium priority
|
create additional chart to monitor sockstat on each node. let's monitor inuse and mem of TCP to start with.
maybe we have to create a new dashboard for network, the current dashboard is too crowded already.
|
1.0
|
[grafana] monitor sockstat from node_exporter - create additional chart to monitor sockstat on each node. let's monitor inuse and mem of TCP to start with.
maybe we have to create a new dashboard for network, the current dashboard is too crowded already.
|
non_test
|
monitor sockstat from node exporter create additional chart to monitor sockstat on each node let s monitor inuse and mem of tcp to start with maybe we have to create a new dashboard for network the current dashboard is too crowded already
| 0
|
294,629
| 25,388,816,862
|
IssuesEvent
|
2022-11-22 01:11:07
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
pkg/ccl/logictestccl/tests/fakedist-disk/fakedist-disk_test: TestCCLLogic_new_schema_changer failed
|
C-test-failure O-robot T-sql-queries branch-release-22.2
|
pkg/ccl/logictestccl/tests/fakedist-disk/fakedist-disk_test.TestCCLLogic_new_schema_changer [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7605513?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7605513?buildTab=artifacts#/) on release-22.2 @ [bf7c1cf83275be4645fc77581be2cb9d5e8502c4](https://github.com/cockroachdb/cockroach/commits/bf7c1cf83275be4645fc77581be2cb9d5e8502c4):
```
=== RUN TestCCLLogic_new_schema_changer
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/effab48683274210e068775452544fff/logTestCCLLogic_new_schema_changer356649684
test_log_scope.go:79: use -show-logs to present logs inline
logic.go:1290: apply knob opt logictest.knobOptDisableCorpusGeneration
*
* INFO: Running test with the default test tenant. If you are only seeing a test case failure when this message appears, there may be a problem with your test case running within tenants.
*
[06:20:13] setting distsql_workmem='40191B';
[06:20:13] rng seed: 2726134000941929125
=== CONT TestCCLLogic_new_schema_changer
logic.go:3927: -- test log scope end --
generated_test.go:81: Leaked goroutine: goroutine 8388 [select]:
github.com/cockroachdb/pebble/vfs.(*diskHealthCheckingFS).startTickerLocked.func1()
github.com/cockroachdb/pebble/vfs/external/com_github_cockroachdb_pebble/vfs/disk_health.go:302 +0xe5
created by github.com/cockroachdb/pebble/vfs.(*diskHealthCheckingFS).startTickerLocked
github.com/cockroachdb/pebble/vfs/external/com_github_cockroachdb_pebble/vfs/disk_health.go:297 +0x7a
--- FAIL: TestCCLLogic_new_schema_changer (7.82s)
```
<p>Parameters: <code>TAGS=bazel,gss</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestCCLLogic_new_schema_changer.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-21645
|
1.0
|
pkg/ccl/logictestccl/tests/fakedist-disk/fakedist-disk_test: TestCCLLogic_new_schema_changer failed - pkg/ccl/logictestccl/tests/fakedist-disk/fakedist-disk_test.TestCCLLogic_new_schema_changer [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7605513?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7605513?buildTab=artifacts#/) on release-22.2 @ [bf7c1cf83275be4645fc77581be2cb9d5e8502c4](https://github.com/cockroachdb/cockroach/commits/bf7c1cf83275be4645fc77581be2cb9d5e8502c4):
```
=== RUN TestCCLLogic_new_schema_changer
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/effab48683274210e068775452544fff/logTestCCLLogic_new_schema_changer356649684
test_log_scope.go:79: use -show-logs to present logs inline
logic.go:1290: apply knob opt logictest.knobOptDisableCorpusGeneration
*
* INFO: Running test with the default test tenant. If you are only seeing a test case failure when this message appears, there may be a problem with your test case running within tenants.
*
[06:20:13] setting distsql_workmem='40191B';
[06:20:13] rng seed: 2726134000941929125
=== CONT TestCCLLogic_new_schema_changer
logic.go:3927: -- test log scope end --
generated_test.go:81: Leaked goroutine: goroutine 8388 [select]:
github.com/cockroachdb/pebble/vfs.(*diskHealthCheckingFS).startTickerLocked.func1()
github.com/cockroachdb/pebble/vfs/external/com_github_cockroachdb_pebble/vfs/disk_health.go:302 +0xe5
created by github.com/cockroachdb/pebble/vfs.(*diskHealthCheckingFS).startTickerLocked
github.com/cockroachdb/pebble/vfs/external/com_github_cockroachdb_pebble/vfs/disk_health.go:297 +0x7a
--- FAIL: TestCCLLogic_new_schema_changer (7.82s)
```
<p>Parameters: <code>TAGS=bazel,gss</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestCCLLogic_new_schema_changer.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-21645
|
test
|
pkg ccl logictestccl tests fakedist disk fakedist disk test testccllogic new schema changer failed pkg ccl logictestccl tests fakedist disk fakedist disk test testccllogic new schema changer with on release run testccllogic new schema changer test log scope go test logs captured to artifacts tmp tmp logtestccllogic new schema test log scope go use show logs to present logs inline logic go apply knob opt logictest knoboptdisablecorpusgeneration info running test with the default test tenant if you are only seeing a test case failure when this message appears there may be a problem with your test case running within tenants setting distsql workmem rng seed cont testccllogic new schema changer logic go test log scope end generated test go leaked goroutine goroutine github com cockroachdb pebble vfs diskhealthcheckingfs starttickerlocked github com cockroachdb pebble vfs external com github cockroachdb pebble vfs disk health go created by github com cockroachdb pebble vfs diskhealthcheckingfs starttickerlocked github com cockroachdb pebble vfs external com github cockroachdb pebble vfs disk health go fail testccllogic new schema changer parameters tags bazel gss help see also cc cockroachdb sql queries jira issue crdb
| 1
|
158,454
| 12,415,886,893
|
IssuesEvent
|
2020-05-22 17:02:21
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
Can't provision RKE clusters due to Rkeaddons.management.cattle.io "nginxingress-v1.15.12" not found error
|
[zube]: To Test area/rke kind/bug-qa status/blocker
|
**What kind of request is this:**
Bug
**Steps to reproduce (least amount of steps as possible):**
- Install Rancher `master-3004-head` and `v2.4-3003-head`
- Deploy a RKE DO cluster
**Result:**
Provisioning will fail with
```
[ERROR] getAddonTemplates: k8sVersion v1.17.6-rancher1-1 addon nginxingress-v1.15.12 [rkeaddons.management.cattle.io "nginxingress-v1.15.12" not found]
```
**Other details that may be helpful:**
**Environment information**
- Rancher version: `master-3004-head` and `v2.4-3003-head` (05/21/2020)
- Installation option (single install/HA): single
|
1.0
|
Can't provision RKE clusters due to Rkeaddons.management.cattle.io "nginxingress-v1.15.12" not found error - **What kind of request is this:**
Bug
**Steps to reproduce (least amount of steps as possible):**
- Install Rancher `master-3004-head` and `v2.4-3003-head`
- Deploy a RKE DO cluster
**Result:**
Provisioning will fail with
```
[ERROR] getAddonTemplates: k8sVersion v1.17.6-rancher1-1 addon nginxingress-v1.15.12 [rkeaddons.management.cattle.io "nginxingress-v1.15.12" not found]
```
**Other details that may be helpful:**
**Environment information**
- Rancher version: `master-3004-head` and `v2.4-3003-head` (05/21/2020)
- Installation option (single install/HA): single
|
test
|
can t provision rke clusters due to rkeaddons management cattle io nginxingress not found error what kind of request is this bug steps to reproduce least amount of steps as possible install rancher master head and head deploy a rke do cluster result provisioning will fail with getaddontemplates addon nginxingress other details that may be helpful environment information rancher version master head and head installation option single install ha single
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.