Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
209,567
| 16,040,865,686
|
IssuesEvent
|
2021-04-22 07:42:32
|
appium/appium
|
https://api.github.com/repos/appium/appium
|
closed
|
Appium v1.21.0-rc.1 and desktop 1.20.2-4 for IOS React Native app is very slow and taking twice the time of android runs
|
ThirdParty XCUITest
|
## The problem
Appium v1.21.0-rc.1 and desktop 1.20.2-4 for IOS React Native app is very slow and taking twice the time of android executions.
In appium desktop 1.8.3 versions the execution is equally fast for IOS and android.
## Environment
MAC OS : BigSure
Appium server : v1.21.0-rc.1
OR
Appium desktop : 1.20.2-4
## Details
Appium v1.21.0-rc.1 and desktop 1.20.2-4 for IOS React Native app is very slow and taking twice the time of android executions.
In appium desktop 1.8.3 versions the execution is equally fast for IOS and android.
Tried with the following capabilities to improve speed , but they did not have a impact :
capabilities.setCapability("waitForQuiescence", false);
capabilities.setCapability("useJSONSource", true);
capabilities.setCapability("simpleIsVisibleCheck",true);
|
1.0
|
Appium v1.21.0-rc.1 and desktop 1.20.2-4 for IOS React Native app is very slow and taking twice the time of android runs - ## The problem
Appium v1.21.0-rc.1 and desktop 1.20.2-4 for IOS React Native app is very slow and taking twice the time of android executions.
In appium desktop 1.8.3 versions the execution is equally fast for IOS and android.
## Environment
MAC OS : BigSure
Appium server : v1.21.0-rc.1
OR
Appium desktop : 1.20.2-4
## Details
Appium v1.21.0-rc.1 and desktop 1.20.2-4 for IOS React Native app is very slow and taking twice the time of android executions.
In appium desktop 1.8.3 versions the execution is equally fast for IOS and android.
Tried with the following capabilities to improve speed , but they did not have a impact :
capabilities.setCapability("waitForQuiescence", false);
capabilities.setCapability("useJSONSource", true);
capabilities.setCapability("simpleIsVisibleCheck",true);
|
test
|
appium rc and desktop for ios react native app is very slow and taking twice the time of android runs the problem appium rc and desktop for ios react native app is very slow and taking twice the time of android executions in appium desktop versions the execution is equally fast for ios and android environment mac os bigsure appium server rc or appium desktop details appium rc and desktop for ios react native app is very slow and taking twice the time of android executions in appium desktop versions the execution is equally fast for ios and android tried with the following capabilities to improve speed but they did not have a impact capabilities setcapability waitforquiescence false capabilities setcapability usejsonsource true capabilities setcapability simpleisvisiblecheck true
| 1
|
54,009
| 6,360,027,639
|
IssuesEvent
|
2017-07-31 09:01:27
|
medic/medic-webapp
|
https://api.github.com/repos/medic/medic-webapp
|
closed
|
Some reports not being replicated
|
4 - Acceptance Testing Bug
|
As a CHW supervisor I should see all reports that were submitted by people I supervise. I have been seeing odd replication issues where a supervisor does not see certain reports. One such case where a supervisor user account does not see reports from their CHW is if the `patient_id` did not match a patient. This scenario seems reproducible.
For instance, [this report](https://standard.app.medicmobile.org/medic/_design/medic/_rewrite/#/reports/0ba96f94b93899b06a4136ab5c409de6) can be seen by admin, but not Marni's supervisor `mch` who is assigned to the MCH Health Center:


|
1.0
|
Some reports not being replicated - As a CHW supervisor I should see all reports that were submitted by people I supervise. I have been seeing odd replication issues where a supervisor does not see certain reports. One such case where a supervisor user account does not see reports from their CHW is if the `patient_id` did not match a patient. This scenario seems reproducible.
For instance, [this report](https://standard.app.medicmobile.org/medic/_design/medic/_rewrite/#/reports/0ba96f94b93899b06a4136ab5c409de6) can be seen by admin, but not Marni's supervisor `mch` who is assigned to the MCH Health Center:


|
test
|
some reports not being replicated as a chw supervisor i should see all reports that were submitted by people i supervise i have been seeing odd replication issues where a supervisor does not see certain reports one such case where a supervisor user account does not see reports from their chw is if the patient id did not match a patient this scenario seems reproducible for instance can be seen by admin but not marni s supervisor mch who is assigned to the mch health center
| 1
|
253,546
| 21,688,440,058
|
IssuesEvent
|
2022-05-09 13:27:10
|
dusk-network/dusk-blockchain
|
https://api.github.com/repos/dusk-network/dusk-blockchain
|
closed
|
Optimize api.db on cluster
|
mark:testnet
|
**Describe the bug**
It turns out that api.db is growing fast in size on any cluster node. We should find a way to clean it up on regular base to avoid wasting disk space.
```
1.2G Apr 5 09:53 /opt/dusk/dusk_data/chain/api.db
```
|
1.0
|
Optimize api.db on cluster - **Describe the bug**
It turns out that api.db is growing fast in size on any cluster node. We should find a way to clean it up on regular base to avoid wasting disk space.
```
1.2G Apr 5 09:53 /opt/dusk/dusk_data/chain/api.db
```
|
test
|
optimize api db on cluster describe the bug it turns out that api db is growing fast in size on any cluster node we should find a way to clean it up on regular base to avoid wasting disk space apr opt dusk dusk data chain api db
| 1
|
894
| 2,656,821,418
|
IssuesEvent
|
2015-03-18 01:21:03
|
mesosphere/marathon
|
https://api.github.com/repos/mesosphere/marathon
|
opened
|
AS user I WANT to access my logs easily
|
usability
|
Couple of distinct use cases:
1. I WANT to see the current log of my running app.
2. My app is flapping and I want to see the log of my last failed run.
|
True
|
AS user I WANT to access my logs easily - Couple of distinct use cases:
1. I WANT to see the current log of my running app.
2. My app is flapping and I want to see the log of my last failed run.
|
non_test
|
as user i want to access my logs easily couple of distinct use cases i want to see the current log of my running app my app is flapping and i want to see the log of my last failed run
| 0
|
326,762
| 28,017,380,669
|
IssuesEvent
|
2023-03-28 00:34:29
|
Azure/azure-sdk-for-js
|
https://api.github.com/repos/Azure/azure-sdk-for-js
|
closed
|
Azure Communication Email Readme Issue
|
Client Docs test-manual-pass Communication - Email
|
1.
Section [link](https://github.com/Azure/azure-sdk-for-js/tree/main/sdk/communication/communication-email#send-email-with-attachments)

**Reason:**
Function 'send' does not exist on class 'EmailClient'.
**Suggestion:**
Update the code as following:
``` javascript
const poller = await emailClient.beginSend(message);
const response = await poller.pollUntilDone();
```
@joheredi, @mayurid, @yogeshmo for notification.
|
1.0
|
Azure Communication Email Readme Issue - 1.
Section [link](https://github.com/Azure/azure-sdk-for-js/tree/main/sdk/communication/communication-email#send-email-with-attachments)

**Reason:**
Function 'send' does not exist on class 'EmailClient'.
**Suggestion:**
Update the code as following:
``` javascript
const poller = await emailClient.beginSend(message);
const response = await poller.pollUntilDone();
```
@joheredi, @mayurid, @yogeshmo for notification.
|
test
|
azure communication email readme issue section reason function send does not exist on class emailclient suggestion update the code as following javascript const poller await emailclient beginsend message const response await poller polluntildone joheredi mayurid yogeshmo for notification
| 1
|
277,678
| 24,094,790,093
|
IssuesEvent
|
2022-09-19 17:43:54
|
microsoft/AzureStorageExplorer
|
https://api.github.com/repos/microsoft/AzureStorageExplorer
|
closed
|
It is better to add the service type in the 'pin/unpin to Quick Access' activity log
|
:bulb: feature request :heavy_check_mark: won't fix 🧪 testing :gear: quick access
|
**Storage Explorer Version**: 1.26.0-dev
**Build Number**: 20220915.4
**Branch**: main
**Platform/OS**: Windows 10/Linux Ubuntu 22.04/MacOS Monterey 12.5.1 (Apple M1 Pro)
**Architecture** ia32\x64
**How Found**: Ad-hoc testing
**Regression From**: Not a regression
## Steps to Reproduce ##
1. Expand one storage account -> Tables.
2. Create a table named 'test1' -> Pin the table to Quick Access.
3. Observe the activity log.
4. Check whether the service type 'table' displays in the activity log.
## Expected Experience ##
It is better to add the service type 'table' display in the activity log.
Like: **Pinned table 'test' to Quick Access.**
## Actual Experience ##
No service type 'table' displays in the activity log.

|
1.0
|
It is better to add the service type in the 'pin/unpin to Quick Access' activity log - **Storage Explorer Version**: 1.26.0-dev
**Build Number**: 20220915.4
**Branch**: main
**Platform/OS**: Windows 10/Linux Ubuntu 22.04/MacOS Monterey 12.5.1 (Apple M1 Pro)
**Architecture** ia32\x64
**How Found**: Ad-hoc testing
**Regression From**: Not a regression
## Steps to Reproduce ##
1. Expand one storage account -> Tables.
2. Create a table named 'test1' -> Pin the table to Quick Access.
3. Observe the activity log.
4. Check whether the service type 'table' displays in the activity log.
## Expected Experience ##
It is better to add the service type 'table' display in the activity log.
Like: **Pinned table 'test' to Quick Access.**
## Actual Experience ##
No service type 'table' displays in the activity log.

|
test
|
it is better to add the service type in the pin unpin to quick access activity log storage explorer version dev build number branch main platform os windows linux ubuntu macos monterey apple pro architecture how found ad hoc testing regression from not a regression steps to reproduce expand one storage account tables create a table named pin the table to quick access observe the activity log check whether the service type table displays in the activity log expected experience it is better to add the service type table display in the activity log like pinned table test to quick access actual experience no service type table displays in the activity log
| 1
|
242,769
| 20,262,557,821
|
IssuesEvent
|
2022-02-15 09:02:36
|
thesofproject/sof
|
https://api.github.com/repos/thesofproject/sof
|
closed
|
[BUG][BYT][BDW] There are host position update warnings always
|
bug BYT P1 BDW Intel Linux Daily tests
|
There are many host position update warnings from sof-logger according to recent daily test on Intel's BDW and BYT platforms:
```
[ 291465955.814017] ( 62.499996) c0 ll-schedule ./schedule/ll_schedule.c:355 INFO num_tasks 1 total_num_tasks 1
[ 292697318.629670] ( 1231362.875000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 294113987.635877] ( 1416669.000000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 295530656.277500] ( 1416668.625000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 296963985.022628] ( 1433328.750000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 298380654.133001] ( 1416669.125000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 299797322.774625] ( 1416668.625000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 301230651.623919] ( 1433328.875000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 302647320.525959] ( 1416668.875000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 304063989.271749] ( 1416668.750000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 305497318.121044] ( 1433328.875000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 306913987.127251] ( 1416669.000000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 308330655.664707] ( 1416668.500000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
```
**To Reproduce**
Just run any playback or capture streams on BDW/BYT, and check sof-logger logs.
**Reproduction Rate**
It happens in about every 500ms~2s when the stream is active
**Expected behavior**
We didn't have this warning previous.
**Impact**
Not sure if it will impact any audio quality, but this warning should be taken care anyway.
**Environment**
1) Branch name and commit hash of the 2 repositories: sof (firmware/topology) and linux (kernel driver).
* Kernel: topic/sof-dev Commit: a9c617ef
* SOF: main Commit: cd48b895ebda
2) Name of the topology file
* Topology: sof-bdw-rt286.tplg on BDW and sof-byt-nocodec.tplg on BYT.
3) Name of the platform(s) on which the bug is observed.
* Platform: BDW/BYT.
|
1.0
|
[BUG][BYT][BDW] There are host position update warnings always - There are many host position update warnings from sof-logger according to recent daily test on Intel's BDW and BYT platforms:
```
[ 291465955.814017] ( 62.499996) c0 ll-schedule ./schedule/ll_schedule.c:355 INFO num_tasks 1 total_num_tasks 1
[ 292697318.629670] ( 1231362.875000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 294113987.635877] ( 1416669.000000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 295530656.277500] ( 1416668.625000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 296963985.022628] ( 1433328.750000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 298380654.133001] ( 1416669.125000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 299797322.774625] ( 1416668.625000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 301230651.623919] ( 1433328.875000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 302647320.525959] ( 1416668.875000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 304063989.271749] ( 1416668.750000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 305497318.121044] ( 1433328.875000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 306913987.127251] ( 1416669.000000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
[ 308330655.664707] ( 1416668.500000) c0 pipe 1.10 ....../pipeline-stream.c:473 WARN pipeline_get_timestamp(): DAI position update failed
```
**To Reproduce**
Just run any playback or capture streams on BDW/BYT, and check sof-logger logs.
**Reproduction Rate**
It happens in about every 500ms~2s when the stream is active
**Expected behavior**
We didn't have this warning previous.
**Impact**
Not sure if it will impact any audio quality, but this warning should be taken care anyway.
**Environment**
1) Branch name and commit hash of the 2 repositories: sof (firmware/topology) and linux (kernel driver).
* Kernel: topic/sof-dev Commit: a9c617ef
* SOF: main Commit: cd48b895ebda
2) Name of the topology file
* Topology: sof-bdw-rt286.tplg on BDW and sof-byt-nocodec.tplg on BYT.
3) Name of the platform(s) on which the bug is observed.
* Platform: BDW/BYT.
|
test
|
there are host position update warnings always there are many host position update warnings from sof logger according to recent daily test on intel s bdw and byt platforms ll schedule schedule ll schedule c info num tasks total num tasks pipe pipeline stream c warn pipeline get timestamp dai position update failed pipe pipeline stream c warn pipeline get timestamp dai position update failed pipe pipeline stream c warn pipeline get timestamp dai position update failed pipe pipeline stream c warn pipeline get timestamp dai position update failed pipe pipeline stream c warn pipeline get timestamp dai position update failed pipe pipeline stream c warn pipeline get timestamp dai position update failed pipe pipeline stream c warn pipeline get timestamp dai position update failed pipe pipeline stream c warn pipeline get timestamp dai position update failed pipe pipeline stream c warn pipeline get timestamp dai position update failed pipe pipeline stream c warn pipeline get timestamp dai position update failed pipe pipeline stream c warn pipeline get timestamp dai position update failed pipe pipeline stream c warn pipeline get timestamp dai position update failed to reproduce just run any playback or capture streams on bdw byt and check sof logger logs reproduction rate it happens in about every when the stream is active expected behavior we didn t have this warning previous impact not sure if it will impact any audio quality but this warning should be taken care anyway environment branch name and commit hash of the repositories sof firmware topology and linux kernel driver kernel topic sof dev commit sof main commit name of the topology file topology sof bdw tplg on bdw and sof byt nocodec tplg on byt name of the platform s on which the bug is observed platform bdw byt
| 1
|
288,424
| 24,905,112,176
|
IssuesEvent
|
2022-10-29 06:10:35
|
chadvandy/cbfm_wh3
|
https://api.github.com/repos/chadvandy/cbfm_wh3
|
closed
|
Grom and Drycha enemy morale debuff when fighting Elves skill doesn't work (Mandras)
|
needs-testing resolved
|
Originally by Mandras
https://cdn.discordapp.com/attachments/581350140942483456/1019249381234454598/Grom_Mandras.pack
|
1.0
|
Grom and Drycha enemy morale debuff when fighting Elves skill doesn't work (Mandras) - Originally by Mandras
https://cdn.discordapp.com/attachments/581350140942483456/1019249381234454598/Grom_Mandras.pack
|
test
|
grom and drycha enemy morale debuff when fighting elves skill doesn t work mandras originally by mandras
| 1
|
65,549
| 6,967,835,389
|
IssuesEvent
|
2017-12-10 14:09:37
|
GoogleCloudPlatform/google-cloud-cpp
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-cpp
|
closed
|
Create a build that runs clang-tidy
|
static-analysis testing
|
At the end of this task we will have a build in the matrix that runs clang-tidy. Ideally the results will be pushed as comments to PRs, but even if they are in the build log that would be good enough.
|
1.0
|
Create a build that runs clang-tidy - At the end of this task we will have a build in the matrix that runs clang-tidy. Ideally the results will be pushed as comments to PRs, but even if they are in the build log that would be good enough.
|
test
|
create a build that runs clang tidy at the end of this task we will have a build in the matrix that runs clang tidy ideally the results will be pushed as comments to prs but even if they are in the build log that would be good enough
| 1
|
824,717
| 31,167,892,155
|
IssuesEvent
|
2023-08-16 21:22:06
|
michalspano/saol.se-cli
|
https://api.github.com/repos/michalspano/saol.se-cli
|
opened
|
Add support for nouns
|
feature priority:high
|
## Issue Description
<!-- Describe the idea (i.e. issue) in a comprehensive manner (natural language, requirement's specification, user story, etc.) -->
The `CLI` tool shall be able to query a **string** which holds a **noun**, store the noun in a specific data structure (i.e. order), and display the **grammatical phenomena** in a human-readable format to environment (i.e. a formatted sequence of characters printed to the standard output).
### Checklist
<!-- Provide acceptance criteria (i.e. checklist) for the proposed issue -->
- [ ] Data (such as the grammatical rules, etc.) are successfully extracted (_scraped_) from the web-based `saol` utility given a **string** that is lexically a **noun** (via a command-line argument).
- [ ] Such extracted data are stored in an efficient data structure/order (with the use of `Go`'s `struct` construct) are are made accessible in the required _"places"_ (such as function calls, parameters, etc.).
- [ ] The extracted data about a particular noun is displayed in the desired environment (e.g. standard output) in a human-readable format (based on the formatting given by the web-based `saol` utility).
- [ ] A noun which is not recognized by the web-based `saol` utility is handled in a way, such that the user is informed that the chosen noun is invalid.
<!-- Link issues that are related to the proposed issue (if not required, leave the following line commented) -->
<!--[Related to: #<issue_number_1>, ..., #<issue_number_n>]-->
### Demo
Let `-word=hus`, we run the _scraping_ against [svenska.se/saol/?sok=hus](https://svenska.se/saol/?sok=hus), which yields the following:
<img width="616" alt="Screenshot 2023-08-16 at 23 19 04" src="https://github.com/michalspano/saol.se-cli/assets/71947840/f801fca7-7d0c-4bc6-b89e-5c101f100613">
In the CLI-based format, we'd expect the following formatting (having the same query):
```txt
hus substantiv ~et; pl. ~
<INSERT_SEMANTICS>
Singular
ett hus obestämd form
ett hus obestämd form genitiv
huset bestämd form
husets bestämd form genitiv
Plural
hus obestämd form
hus obestämd form genitiv
husen bestämd form
husens bestämd form genitiv
Övrig(a) form(er)
huse i vissa uttryck
```
|
1.0
|
Add support for nouns - ## Issue Description
<!-- Describe the idea (i.e. issue) in a comprehensive manner (natural language, requirement's specification, user story, etc.) -->
The `CLI` tool shall be able to query a **string** which holds a **noun**, store the noun in a specific data structure (i.e. order), and display the **grammatical phenomena** in a human-readable format to environment (i.e. a formatted sequence of characters printed to the standard output).
### Checklist
<!-- Provide acceptance criteria (i.e. checklist) for the proposed issue -->
- [ ] Data (such as the grammatical rules, etc.) are successfully extracted (_scraped_) from the web-based `saol` utility given a **string** that is lexically a **noun** (via a command-line argument).
- [ ] Such extracted data are stored in an efficient data structure/order (with the use of `Go`'s `struct` construct) are are made accessible in the required _"places"_ (such as function calls, parameters, etc.).
- [ ] The extracted data about a particular noun is displayed in the desired environment (e.g. standard output) in a human-readable format (based on the formatting given by the web-based `saol` utility).
- [ ] A noun which is not recognized by the web-based `saol` utility is handled in a way, such that the user is informed that the chosen noun is invalid.
<!-- Link issues that are related to the proposed issue (if not required, leave the following line commented) -->
<!--[Related to: #<issue_number_1>, ..., #<issue_number_n>]-->
### Demo
Let `-word=hus`, we run the _scraping_ against [svenska.se/saol/?sok=hus](https://svenska.se/saol/?sok=hus), which yields the following:
<img width="616" alt="Screenshot 2023-08-16 at 23 19 04" src="https://github.com/michalspano/saol.se-cli/assets/71947840/f801fca7-7d0c-4bc6-b89e-5c101f100613">
In the CLI-based format, we'd expect the following formatting (having the same query):
```txt
hus substantiv ~et; pl. ~
<INSERT_SEMANTICS>
Singular
ett hus obestämd form
ett hus obestämd form genitiv
huset bestämd form
husets bestämd form genitiv
Plural
hus obestämd form
hus obestämd form genitiv
husen bestämd form
husens bestämd form genitiv
Övrig(a) form(er)
huse i vissa uttryck
```
|
non_test
|
add support for nouns issue description the cli tool shall be able to query a string which holds a noun store the noun in a specific data structure i e order and display the grammatical phenomena in a human readable format to environment i e a formatted sequence of characters printed to the standard output checklist data such as the grammatical rules etc are successfully extracted scraped from the web based saol utility given a string that is lexically a noun via a command line argument such extracted data are stored in an efficient data structure order with the use of go s struct construct are are made accessible in the required places such as function calls parameters etc the extracted data about a particular noun is displayed in the desired environment e g standard output in a human readable format based on the formatting given by the web based saol utility a noun which is not recognized by the web based saol utility is handled in a way such that the user is informed that the chosen noun is invalid demo let word hus we run the scraping against which yields the following img width alt screenshot at src in the cli based format we d expect the following formatting having the same query txt hus substantiv et pl singular ett hus obestämd form ett hus obestämd form genitiv huset bestämd form husets bestämd form genitiv plural hus obestämd form hus obestämd form genitiv husen bestämd form husens bestämd form genitiv övrig a form er huse i vissa uttryck
| 0
|
3,417
| 2,671,724,713
|
IssuesEvent
|
2015-03-24 09:25:32
|
radare/radare2
|
https://api.github.com/repos/radare/radare2
|
closed
|
aa0 broken by a commit
|
bug regression test-attached
|
Radare2-Regressions-format-elf shows a regression coming from e1edc18a688a47c02b0a346b4cab276778e46430 @alvarofe
http://ci.rada.re/view/File%20Formats/job/radare2-regressions-formats-elf/2142/console
[XX] 5 helloworld-gcc-elf: flags spaces after analysis
|
1.0
|
aa0 broken by a commit - Radare2-Regressions-format-elf shows a regression coming from e1edc18a688a47c02b0a346b4cab276778e46430 @alvarofe
http://ci.rada.re/view/File%20Formats/job/radare2-regressions-formats-elf/2142/console
[XX] 5 helloworld-gcc-elf: flags spaces after analysis
|
test
|
broken by a commit regressions format elf shows a regression coming from alvarofe helloworld gcc elf flags spaces after analysis
| 1
|
20,658
| 6,077,146,614
|
IssuesEvent
|
2017-06-16 02:27:09
|
dotnet/coreclr
|
https://api.github.com/repos/dotnet/coreclr
|
closed
|
[RyuJIT/arm32] Assertion failed 'isDoubleReg(reg1)'
|
arch-arm32 area-CodeGen bug
|
At emitarm.cpp Line: 2083
Example:
```
Assert failure(PID 4252 [0x0000109c], Thread: 11220 [0x2bd4]): Assertion failed 'isDoubleReg(reg1)' in 'FractalPerf.Julia:Render():double:this' (IL size 160)
File: c:\gh\coreclr\src\jit\emitarm.cpp Line: 2083
Image: c:\brucefo\tests\Windows_NT.arm.Checked\Tests\Core_Root\CoreRun.exe
```
Tests with this assert:
```
JIT\Performance\CodeQuality\FractalPerf\FractalPerf\FractalPerf.cmd
```
|
1.0
|
[RyuJIT/arm32] Assertion failed 'isDoubleReg(reg1)' - At emitarm.cpp Line: 2083
Example:
```
Assert failure(PID 4252 [0x0000109c], Thread: 11220 [0x2bd4]): Assertion failed 'isDoubleReg(reg1)' in 'FractalPerf.Julia:Render():double:this' (IL size 160)
File: c:\gh\coreclr\src\jit\emitarm.cpp Line: 2083
Image: c:\brucefo\tests\Windows_NT.arm.Checked\Tests\Core_Root\CoreRun.exe
```
Tests with this assert:
```
JIT\Performance\CodeQuality\FractalPerf\FractalPerf\FractalPerf.cmd
```
|
non_test
|
assertion failed isdoublereg at emitarm cpp line example assert failure pid thread assertion failed isdoublereg in fractalperf julia render double this il size file c gh coreclr src jit emitarm cpp line image c brucefo tests windows nt arm checked tests core root corerun exe tests with this assert jit performance codequality fractalperf fractalperf fractalperf cmd
| 0
|
149,352
| 11,890,781,444
|
IssuesEvent
|
2020-03-28 19:57:29
|
eBay/skin
|
https://api.github.com/repos/eBay/skin
|
closed
|
Storybook & Percy: Phase 2
|
aspect: percy aspect: storybook aspect: tests fellowship: good first project resolution: done
|
We've ported over a good deal of the old test pages to Storybook stories, and we have Percy running visual regression tests against those stories (either manually or via CI). Phase 1 complete!
Phase 2 is to have another sweep through the stories, have a bit of a cleanup, and ensure there is only one instance of a module per story. We still have quite a few stories that are showing multiple states of a module in a single story (for example, expanded and collapsed, or button & fake button).
|
1.0
|
Storybook & Percy: Phase 2 - We've ported over a good deal of the old test pages to Storybook stories, and we have Percy running visual regression tests against those stories (either manually or via CI). Phase 1 complete!
Phase 2 is to have another sweep through the stories, have a bit of a cleanup, and ensure there is only one instance of a module per story. We still have quite a few stories that are showing multiple states of a module in a single story (for example, expanded and collapsed, or button & fake button).
|
test
|
storybook percy phase we ve ported over a good deal of the old test pages to storybook stories and we have percy running visual regression tests against those stories either manually or via ci phase complete phase is to have another sweep through the stories have a bit of a cleanup and ensure there is only one instance of a module per story we still have quite a few stories that are showing multiple states of a module in a single story for example expanded and collapsed or button fake button
| 1
|
244,361
| 20,626,378,254
|
IssuesEvent
|
2022-03-07 23:08:08
|
coder/code-server
|
https://api.github.com/repos/coder/code-server
|
closed
|
[Testing]: write tests for logLevel in setDefaults
|
testing
|
We're missing coverage for L490-492 where we set logLevel to Error.
```typescript
case LogLevel.Error:
logger.level = Level.Error
args.verbose = false
break
```
https://github.com/coder/code-server/blob/main/src/node/cli.ts#L490-L492
write test for `setDefaults` and pass in error for log level. assert level and verbose
|
1.0
|
[Testing]: write tests for logLevel in setDefaults - We're missing coverage for L490-492 where we set logLevel to Error.
```typescript
case LogLevel.Error:
logger.level = Level.Error
args.verbose = false
break
```
https://github.com/coder/code-server/blob/main/src/node/cli.ts#L490-L492
write test for `setDefaults` and pass in error for log level. assert level and verbose
|
test
|
write tests for loglevel in setdefaults we re missing coverage for where we set loglevel to error typescript case loglevel error logger level level error args verbose false break write test for setdefaults and pass in error for log level assert level and verbose
| 1
|
40,930
| 12,800,077,605
|
IssuesEvent
|
2020-07-02 16:25:44
|
flyingcircusio/nixpkgs
|
https://api.github.com/repos/flyingcircusio/nixpkgs
|
opened
|
Vulnerability roundup 4: i2p-0.9.39: 1 advisory [7.8]
|
1.severity: security
|
[search](https://search.nix.gsc.io/?q=i2p&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=i2p+in%3Apath&type=Code)
* [ ] [CVE-2020-13431](https://nvd.nist.gov/vuln/detail/CVE-2020-13431) CVSSv3=7.8 (nixos-19.03)
Scanned versions: nixos-19.03: f156ee5cbf2. May contain false positives.
|
True
|
Vulnerability roundup 4: i2p-0.9.39: 1 advisory [7.8] - [search](https://search.nix.gsc.io/?q=i2p&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=i2p+in%3Apath&type=Code)
* [ ] [CVE-2020-13431](https://nvd.nist.gov/vuln/detail/CVE-2020-13431) CVSSv3=7.8 (nixos-19.03)
Scanned versions: nixos-19.03: f156ee5cbf2. May contain false positives.
|
non_test
|
vulnerability roundup advisory nixos scanned versions nixos may contain false positives
| 0
|
335,687
| 30,082,038,080
|
IssuesEvent
|
2023-06-29 04:56:53
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: failover/liveness/blackhole-recv/lease=expiration failed
|
C-test-failure O-robot O-roachtest branch-master release-blocker T-kv
|
roachtest.failover/liveness/blackhole-recv/lease=expiration [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestWeeklyBazel/10720111?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestWeeklyBazel/10720111?buildTab=artifacts#/failover/liveness/blackhole-recv/lease=expiration) on master @ [7fd4c21157221eae9e7d5892d89d2b5a671aba3e](https://github.com/cockroachdb/cockroach/commits/7fd4c21157221eae9e7d5892d89d2b5a671aba3e):
```
(assertions.go:333).Fail:
Error Trace: github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:1665
github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:911
github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:155
main/pkg/cmd/roachtest/test_runner.go:1060
GOROOT/src/runtime/asm_amd64.s:1594
Error: Received unexpected error:
pq: error getting span statistics - number of spans in request payload (1052) exceeds 'server.span_stats.span_batch_limit' cluster setting limit (500)
Test: failover/liveness/blackhole-recv/lease=expiration
(require.go:1360).NoError: FailNow called
(cluster.go:2279).Run: cluster.RunE: context canceled
test artifacts and logs in: /artifacts/failover/liveness/blackhole-recv/lease=expiration/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=2</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*failover/liveness/blackhole-recv/lease=expiration.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29184
|
2.0
|
roachtest: failover/liveness/blackhole-recv/lease=expiration failed - roachtest.failover/liveness/blackhole-recv/lease=expiration [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestWeeklyBazel/10720111?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestWeeklyBazel/10720111?buildTab=artifacts#/failover/liveness/blackhole-recv/lease=expiration) on master @ [7fd4c21157221eae9e7d5892d89d2b5a671aba3e](https://github.com/cockroachdb/cockroach/commits/7fd4c21157221eae9e7d5892d89d2b5a671aba3e):
```
(assertions.go:333).Fail:
Error Trace: github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:1665
github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:911
github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:155
main/pkg/cmd/roachtest/test_runner.go:1060
GOROOT/src/runtime/asm_amd64.s:1594
Error: Received unexpected error:
pq: error getting span statistics - number of spans in request payload (1052) exceeds 'server.span_stats.span_batch_limit' cluster setting limit (500)
Test: failover/liveness/blackhole-recv/lease=expiration
(require.go:1360).NoError: FailNow called
(cluster.go:2279).Run: cluster.RunE: context canceled
test artifacts and logs in: /artifacts/failover/liveness/blackhole-recv/lease=expiration/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=2</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*failover/liveness/blackhole-recv/lease=expiration.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29184
|
test
|
roachtest failover liveness blackhole recv lease expiration failed roachtest failover liveness blackhole recv lease expiration with on master assertions go fail error trace github com cockroachdb cockroach pkg cmd roachtest tests failover go github com cockroachdb cockroach pkg cmd roachtest tests failover go github com cockroachdb cockroach pkg cmd roachtest tests failover go main pkg cmd roachtest test runner go goroot src runtime asm s error received unexpected error pq error getting span statistics number of spans in request payload exceeds server span stats span batch limit cluster setting limit test failover liveness blackhole recv lease expiration require go noerror failnow called cluster go run cluster rune context canceled test artifacts and logs in artifacts failover liveness blackhole recv lease expiration run parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see cc cockroachdb kv triage jira issue crdb
| 1
|
226,379
| 18,014,998,658
|
IssuesEvent
|
2021-09-16 13:03:15
|
TTitcombe/Greta
|
https://api.github.com/repos/TTitcombe/Greta
|
opened
|
Reduce "responses" test repetition
|
testing
|
We use "responses" library to mock-out API calls. Currently this requires a `@responses.active` wrapper on every test function and a responses data setup.
Find a way to simplify this for test files so responses code does not need to be written so much
|
1.0
|
Reduce "responses" test repetition - We use "responses" library to mock-out API calls. Currently this requires a `@responses.active` wrapper on every test function and a responses data setup.
Find a way to simplify this for test files so responses code does not need to be written so much
|
test
|
reduce responses test repetition we use responses library to mock out api calls currently this requires a responses active wrapper on every test function and a responses data setup find a way to simplify this for test files so responses code does not need to be written so much
| 1
|
7,565
| 3,103,030,667
|
IssuesEvent
|
2015-08-31 06:43:49
|
saltstack/salt
|
https://api.github.com/repos/saltstack/salt
|
reopened
|
salt-cloud (EC2): Salt-Minion-2014.7.1-AMD64-Setup.exe doesn't install salt-minion service
|
Bug Documentation High Severity P2 Platform Regression Salt-Cloud Windows
|
Hi,
while bootstrapping a Windows Server 2008R2 AWS instance via salt-cloud, the Salt Minion installer (2014.7.1) doesn't create the salt-minion service. OTOH, salt-cloud doesn't notice it's not there:
```
[INFO ] Running command under pid 17978: 'winexe -U Administrator%secret_passwd //10.11.12.13 "sc stop salt-minion"'
[INFO ] Running command under pid 17980: 'winexe -U Administrator%secret_passwd //10.11.12.13 "sc start salt-minion"'
[INFO ] Salt installed on salt-cloud-test
[INFO ] Created Cloud VM 'salt-cloud-test'
```
When running the installer manually, the service IS created, though.
Bye...
Dirk
|
1.0
|
salt-cloud (EC2): Salt-Minion-2014.7.1-AMD64-Setup.exe doesn't install salt-minion service - Hi,
while bootstrapping a Windows Server 2008R2 AWS instance via salt-cloud, the Salt Minion installer (2014.7.1) doesn't create the salt-minion service. OTOH, salt-cloud doesn't notice it's not there:
```
[INFO ] Running command under pid 17978: 'winexe -U Administrator%secret_passwd //10.11.12.13 "sc stop salt-minion"'
[INFO ] Running command under pid 17980: 'winexe -U Administrator%secret_passwd //10.11.12.13 "sc start salt-minion"'
[INFO ] Salt installed on salt-cloud-test
[INFO ] Created Cloud VM 'salt-cloud-test'
```
When running the installer manually, the service IS created, though.
Bye...
Dirk
|
non_test
|
salt cloud salt minion setup exe doesn t install salt minion service hi while bootstrapping a windows server aws instance via salt cloud the salt minion installer doesn t create the salt minion service otoh salt cloud doesn t notice it s not there running command under pid winexe u administrator secret passwd sc stop salt minion running command under pid winexe u administrator secret passwd sc start salt minion salt installed on salt cloud test created cloud vm salt cloud test when running the installer manually the service is created though bye dirk
| 0
|
248,592
| 21,043,131,398
|
IssuesEvent
|
2022-03-31 13:57:13
|
zkSNACKs/WalletWasabi
|
https://api.github.com/repos/zkSNACKs/WalletWasabi
|
closed
|
Fee slider Continue button issue
|
debug UI ww2 testing
|
1. Open the fee slider while it is somewhere in the middle.
2. Click Continue without changing anything.
Notice that the tx fee decreases with a couple of sats whenever you repeat that.
|
1.0
|
Fee slider Continue button issue - 1. Open the fee slider while it is somewhere in the middle.
2. Click Continue without changing anything.
Notice that the tx fee decreases with a couple of sats whenever you repeat that.
|
test
|
fee slider continue button issue open the fee slider while it is somewhere in the middle click continue without changing anything notice that the tx fee decreases with a couple of sats whenever you repeat that
| 1
|
753,291
| 26,343,594,949
|
IssuesEvent
|
2023-01-10 19:52:18
|
KingSupernova31/RulesGuru
|
https://api.github.com/repos/KingSupernova31/RulesGuru
|
closed
|
Quotes in card names break things
|
bug medium priority good first issue
|
Quotes don't seem to be properly escaped somewhere. The card `Pang Tong, "Young Phoenix"` shows up as just `Pang Tong, ` in the list of card names.
|
1.0
|
Quotes in card names break things - Quotes don't seem to be properly escaped somewhere. The card `Pang Tong, "Young Phoenix"` shows up as just `Pang Tong, ` in the list of card names.
|
non_test
|
quotes in card names break things quotes don t seem to be properly escaped somewhere the card pang tong young phoenix shows up as just pang tong in the list of card names
| 0
|
280,725
| 24,326,941,240
|
IssuesEvent
|
2022-09-30 15:33:57
|
void-linux/void-packages
|
https://api.github.com/repos/void-linux/void-packages
|
closed
|
ZFS DKMS build fails on freshly installed system
|
bug needs-testing
|
### Is this a new report?
Yes
### System Info
Void 5.19.10_1 x86_64 GenuineIntel/VM uptodate rFF
### Package(s) Affected
zfs-2.1.5_3
### Does a report exist for this bug with the project's home (upstream) and/or another distro?
_No response_
### Expected behaviour
I installed a fresh Void system, updated with xbps-install -Su, rebooted into kernel 5.19 (also tested with linux-lts, same result), and attempted to install ZFS.
### Actual behaviour
On installing (and reinstalling) ZFS, this is the end of the xbps output:
```zfs-2.1.5_3: configuring ...
Added DKMS module 'zfs-2.1.5'.
Skipping kernel-5.13.19_1. kernel-headers package not installed...
Building DKMS module 'zfs-2.1.5' for kernel-5.19.10_1... Killed
```
Manually building with the DKMS module (sudo dkms build zfs/2.1.5) ended with a segmentation fault in gcc and the attached [make.log](https://github.com/void-linux/void-packages/files/9682842/make.log) file.
### Steps to reproduce
(Please note, this is my first time using void. I reinstalled and tried again to make sure it wasn't a fluke)
1. Install a fresh void system (tried with both local and netinstall)
2. Sync/update system and reboot
3. xbps-install zfs
4. DKMS fails, even on kernels <5.18
|
1.0
|
ZFS DKMS build fails on freshly installed system - ### Is this a new report?
Yes
### System Info
Void 5.19.10_1 x86_64 GenuineIntel/VM uptodate rFF
### Package(s) Affected
zfs-2.1.5_3
### Does a report exist for this bug with the project's home (upstream) and/or another distro?
_No response_
### Expected behaviour
I installed a fresh Void system, updated with xbps-install -Su, rebooted into kernel 5.19 (also tested with linux-lts, same result), and attempted to install ZFS.
### Actual behaviour
On installing (and reinstalling) ZFS, this is the end of the xbps output:
```zfs-2.1.5_3: configuring ...
Added DKMS module 'zfs-2.1.5'.
Skipping kernel-5.13.19_1. kernel-headers package not installed...
Building DKMS module 'zfs-2.1.5' for kernel-5.19.10_1... Killed
```
Manually building with the DKMS module (sudo dkms build zfs/2.1.5) ended with a segmentation fault in gcc and the attached [make.log](https://github.com/void-linux/void-packages/files/9682842/make.log) file.
### Steps to reproduce
(Please note, this is my first time using void. I reinstalled and tried again to make sure it wasn't a fluke)
1. Install a fresh void system (tried with both local and netinstall)
2. Sync/update system and reboot
3. xbps-install zfs
4. DKMS fails, even on kernels <5.18
|
test
|
zfs dkms build fails on freshly installed system is this a new report yes system info void genuineintel vm uptodate rff package s affected zfs does a report exist for this bug with the project s home upstream and or another distro no response expected behaviour i installed a fresh void system updated with xbps install su rebooted into kernel also tested with linux lts same result and attempted to install zfs actual behaviour on installing and reinstalling zfs this is the end of the xbps output zfs configuring added dkms module zfs skipping kernel kernel headers package not installed building dkms module zfs for kernel killed manually building with the dkms module sudo dkms build zfs ended with a segmentation fault in gcc and the attached file steps to reproduce please note this is my first time using void i reinstalled and tried again to make sure it wasn t a fluke install a fresh void system tried with both local and netinstall sync update system and reboot xbps install zfs dkms fails even on kernels
| 1
|
32,989
| 6,992,159,906
|
IssuesEvent
|
2017-12-15 04:48:04
|
Quantum64/ExGregilo
|
https://api.github.com/repos/Quantum64/ExGregilo
|
closed
|
Issue with Auto Sieve recipe
|
defect
|
I can not craft the auto sieve using a LV Electric Motor I made myself, but if I swap the motor with a LV Electric Motor that I cheated in using NEI the recipe is complete.
Edit: I have done further testing and this also happens with the Conveyor Module, Robot Arm and Piston as well.


|
1.0
|
Issue with Auto Sieve recipe - I can not craft the auto sieve using a LV Electric Motor I made myself, but if I swap the motor with a LV Electric Motor that I cheated in using NEI the recipe is complete.
Edit: I have done further testing and this also happens with the Conveyor Module, Robot Arm and Piston as well.


|
non_test
|
issue with auto sieve recipe i can not craft the auto sieve using a lv electric motor i made myself but if i swap the motor with a lv electric motor that i cheated in using nei the recipe is complete edit i have done further testing and this also happens with the conveyor module robot arm and piston as well
| 0
|
350,513
| 31,898,394,984
|
IssuesEvent
|
2023-09-18 05:25:26
|
NVIDIA/spark-rapids
|
https://api.github.com/repos/NVIDIA/spark-rapids
|
closed
|
Statistics tests for Parquet files written by GPU
|
test task
|
When the GPU writes a Parquet file, it's important to get the file, row group, and page statistics correct. We should have better tests, ideally including tests at larger scale than a few thousand rows typical for unit/integration tests, for verifying the statistics written to the Parquet file are correct.
|
1.0
|
Statistics tests for Parquet files written by GPU - When the GPU writes a Parquet file, it's important to get the file, row group, and page statistics correct. We should have better tests, ideally including tests at larger scale than a few thousand rows typical for unit/integration tests, for verifying the statistics written to the Parquet file are correct.
|
test
|
statistics tests for parquet files written by gpu when the gpu writes a parquet file it s important to get the file row group and page statistics correct we should have better tests ideally including tests at larger scale than a few thousand rows typical for unit integration tests for verifying the statistics written to the parquet file are correct
| 1
|
156,830
| 19,907,448,563
|
IssuesEvent
|
2022-01-25 14:11:46
|
Recidiviz/supervision-success-component
|
https://api.github.com/repos/Recidiviz/supervision-success-component
|
closed
|
Security Alert - Package: node-forge; Severity: LOW; Vuln ID: GHSA-gf8q-jrpm-jvxq
|
Subject: Security Subject: Vulnerability Severity: LOW
|
---
due: 2022-04-24
---
A new vulnerability has been reported by Dependabot. The criticality of this vulnerability is LOW.
LOW vulnerabilities have an SLA of 90 days according to our policy.
Affected package: node-forge
Ecosystem: NPM
Affected version range: < 1.0.0
Summary: URL parsing in node-forge could lead to undesired behavior.
Description: ### Impact
The regex used for the `forge.util.parseUrl` API would not properly parse certain inputs resulting in a parsed data structure that could lead to undesired behavior.
### Patches
`forge.util.parseUrl` and other very old related URL APIs were removed in 1.0.0 in favor of letting applications use the more modern WHATWG URL Standard API.
### Workarounds
Ensure code does not directly or indirectly call `forge.util.parseUrl` with untrusted input.
### References
- https://www.huntr.dev/bounties/41852c50-3c6d-4703-8c55-4db27164a4ae/
### For more information
If you have any questions or comments about this advisory:
* Open an issue in [forge](https://github.com/digitalbazaar/forge)
* Email us at support@digitalbazaar.com
identifiers: [{'type': 'GHSA', 'value': 'GHSA-gf8q-jrpm-jvxq'}]
Fixed Version: 1.0.0
Created Date = January 18, 2022
***Additional Context***
https://github.com/Recidiviz/supervision-success-component/security/dependabot?q=is%3Aopen+sort%3Anewest
|
True
|
Security Alert - Package: node-forge; Severity: LOW; Vuln ID: GHSA-gf8q-jrpm-jvxq -
---
due: 2022-04-24
---
A new vulnerability has been reported by Dependabot. The criticality of this vulnerability is LOW.
LOW vulnerabilities have an SLA of 90 days according to our policy.
Affected package: node-forge
Ecosystem: NPM
Affected version range: < 1.0.0
Summary: URL parsing in node-forge could lead to undesired behavior.
Description: ### Impact
The regex used for the `forge.util.parseUrl` API would not properly parse certain inputs resulting in a parsed data structure that could lead to undesired behavior.
### Patches
`forge.util.parseUrl` and other very old related URL APIs were removed in 1.0.0 in favor of letting applications use the more modern WHATWG URL Standard API.
### Workarounds
Ensure code does not directly or indirectly call `forge.util.parseUrl` with untrusted input.
### References
- https://www.huntr.dev/bounties/41852c50-3c6d-4703-8c55-4db27164a4ae/
### For more information
If you have any questions or comments about this advisory:
* Open an issue in [forge](https://github.com/digitalbazaar/forge)
* Email us at support@digitalbazaar.com
identifiers: [{'type': 'GHSA', 'value': 'GHSA-gf8q-jrpm-jvxq'}]
Fixed Version: 1.0.0
Created Date = January 18, 2022
***Additional Context***
https://github.com/Recidiviz/supervision-success-component/security/dependabot?q=is%3Aopen+sort%3Anewest
|
non_test
|
security alert package node forge severity low vuln id ghsa jrpm jvxq due a new vulnerability has been reported by dependabot the criticality of this vulnerability is low low vulnerabilities have an sla of days according to our policy affected package node forge ecosystem npm affected version range summary url parsing in node forge could lead to undesired behavior description impact the regex used for the forge util parseurl api would not properly parse certain inputs resulting in a parsed data structure that could lead to undesired behavior patches forge util parseurl and other very old related url apis were removed in in favor of letting applications use the more modern whatwg url standard api workarounds ensure code does not directly or indirectly call forge util parseurl with untrusted input references for more information if you have any questions or comments about this advisory open an issue in email us at support digitalbazaar com identifiers fixed version created date january additional context
| 0
|
322,001
| 23,884,206,862
|
IssuesEvent
|
2022-09-08 06:01:59
|
grindylow/ahoy
|
https://api.github.com/repos/grindylow/ahoy
|
closed
|
Feature Request : Limit-Anzeige auf Website in gewählter Einheit "Prozent" oder "Watt" anzeigen
|
documentation enhancement wontfix ESP
|
Begrenze ich meinen HM-600 auf 320 W sehe ich in der seriellen Debugausgabe auch diesen Wert, nach kurzer Zeit regelt sich der Inverter auf diesen Wert ein. Das finde ich als prima Feature.
Allerdings scheint ahoy dies intern immer in Prozent umzurechnen, denn auf der Website steht dann z.B. 53%. Dies finde ich etwas verwirrend:

Ich schlage vor, auf der Website den eingestellten Wert anzuzeigen, und zwar in Prozent oder Watt - je nachdem wie dieser eingestellt wurde. Noch hilfreicher finde ich die Zusatzangabe "persistent" oder "non persistent", so dass man mit einem Blick sieht, was konkret eingestellt wurde.
|
1.0
|
Feature Request : Limit-Anzeige auf Website in gewählter Einheit "Prozent" oder "Watt" anzeigen - Begrenze ich meinen HM-600 auf 320 W sehe ich in der seriellen Debugausgabe auch diesen Wert, nach kurzer Zeit regelt sich der Inverter auf diesen Wert ein. Das finde ich als prima Feature.
Allerdings scheint ahoy dies intern immer in Prozent umzurechnen, denn auf der Website steht dann z.B. 53%. Dies finde ich etwas verwirrend:

Ich schlage vor, auf der Website den eingestellten Wert anzuzeigen, und zwar in Prozent oder Watt - je nachdem wie dieser eingestellt wurde. Noch hilfreicher finde ich die Zusatzangabe "persistent" oder "non persistent", so dass man mit einem Blick sieht, was konkret eingestellt wurde.
|
non_test
|
feature request limit anzeige auf website in gewählter einheit prozent oder watt anzeigen begrenze ich meinen hm auf w sehe ich in der seriellen debugausgabe auch diesen wert nach kurzer zeit regelt sich der inverter auf diesen wert ein das finde ich als prima feature allerdings scheint ahoy dies intern immer in prozent umzurechnen denn auf der website steht dann z b dies finde ich etwas verwirrend ich schlage vor auf der website den eingestellten wert anzuzeigen und zwar in prozent oder watt je nachdem wie dieser eingestellt wurde noch hilfreicher finde ich die zusatzangabe persistent oder non persistent so dass man mit einem blick sieht was konkret eingestellt wurde
| 0
|
245,204
| 20,752,623,906
|
IssuesEvent
|
2022-03-15 09:12:33
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
sql/rowexec: TestJoinReaderUsesBatchLimit failed
|
C-test-failure O-robot branch-master
|
sql/rowexec.TestJoinReaderUsesBatchLimit [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4576247&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4576247&tab=artifacts#/) on master @ [72b0b023832066469dd63017160653236187df6a](https://github.com/cockroachdb/cockroach/commits/72b0b023832066469dd63017160653236187df6a):
```
=== RUN TestJoinReaderUsesBatchLimit
test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/fffb2e619f32b4cb8b294b200fe6717b/logTestJoinReaderUsesBatchLimit1516313893
test_log_scope.go:80: use -show-logs to present logs inline
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
Parameters in this failure:
- TAGS=bazel,gss,deadlock
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestJoinReaderUsesBatchLimit.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
sql/rowexec: TestJoinReaderUsesBatchLimit failed - sql/rowexec.TestJoinReaderUsesBatchLimit [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4576247&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4576247&tab=artifacts#/) on master @ [72b0b023832066469dd63017160653236187df6a](https://github.com/cockroachdb/cockroach/commits/72b0b023832066469dd63017160653236187df6a):
```
=== RUN TestJoinReaderUsesBatchLimit
test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/fffb2e619f32b4cb8b294b200fe6717b/logTestJoinReaderUsesBatchLimit1516313893
test_log_scope.go:80: use -show-logs to present logs inline
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
Parameters in this failure:
- TAGS=bazel,gss,deadlock
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestJoinReaderUsesBatchLimit.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
sql rowexec testjoinreaderusesbatchlimit failed sql rowexec testjoinreaderusesbatchlimit with on master run testjoinreaderusesbatchlimit test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline help see also parameters in this failure tags bazel gss deadlock cc cockroachdb sql queries
| 1
|
799,710
| 28,312,426,280
|
IssuesEvent
|
2023-04-10 16:34:57
|
dynamicslab/hydrogym
|
https://api.github.com/repos/dynamicslab/hydrogym
|
closed
|
PyTorch Update
|
priority Paper-Release
|
PyTorch to be update to 2.0, and ideally utilize the [core features](https://pytorch.org/blog/pytorch-2.0-release/) of the 2.0 release cycle.
|
1.0
|
PyTorch Update - PyTorch to be update to 2.0, and ideally utilize the [core features](https://pytorch.org/blog/pytorch-2.0-release/) of the 2.0 release cycle.
|
non_test
|
pytorch update pytorch to be update to and ideally utilize the of the release cycle
| 0
|
174,421
| 14,481,810,690
|
IssuesEvent
|
2020-12-10 13:09:46
|
Joystream/joystream
|
https://api.github.com/repos/Joystream/joystream
|
opened
|
Initial constitution
|
documentation
|
What should it contain?
- founding members list.
- social encoding standards for the chain?
- tbd.
|
1.0
|
Initial constitution - What should it contain?
- founding members list.
- social encoding standards for the chain?
- tbd.
|
non_test
|
initial constitution what should it contain founding members list social encoding standards for the chain tbd
| 0
|
222,854
| 17,497,067,457
|
IssuesEvent
|
2021-08-10 02:54:02
|
OvercastCommunity/CommunityMaps
|
https://api.github.com/repos/OvercastCommunity/CommunityMaps
|
closed
|
[CTW/40v40] chle_
|
map submission contest
|
### Checklist
Check what applies to you. *Add an X between the brackets or click the checkboxes when you have submitted the issue.*
- [x] I have [pruned](https://pgm.dev/docs/guides/packaging/pruning-chunks) the map
- [x] I have agreed with assigning the CC BY-SA license to this map, as mentioned in the README
- [x] I have created an XML file
- [x] I have created an map image
- [x] I have uploaded the map zip file to a file sharing service
- [x] The map has been tested locally to make sure it works
<img src="https://i.imgur.com/LC0aHbQ.png" alt="map" style="max-width:100%;">
# Map Name
chle_ (yeah, thats the name)
## Gamemode & Map Description
This map does not have a void players have to bridge over. Instead I went fully experimental and made a huge wall players have to dig through in order to get to the other side. While in the wall, players are given night vision and haste 2. The rest of the map is just straight forward. Oh, and no, the large roof supported by the pillars is just there for decoration.
The woolroom however, I believe it has the largest interior of any woolroom in OCN history, so be sure to check out every corner of it!
I have absolutely no idea how this is gonna play lmao
## Team Sizes
Red vs. Blue 40v40
## Screenshots
[here u go](https://imgur.com/a/AYxPmV4)
# XML
[XML](https://gist.github.com/chleongithub/cd69c8b7545a7c7c054c59779b96c037)
# Map Image
<img src="https://i.imgur.com/9pZHhZL.png" alt="map" style="max-width:100%;">
# Download Link
[Download](https://www.dropbox.com/s/i89krl5fziz4bzu/chle_.zip?dl=0)
|
1.0
|
[CTW/40v40] chle_ - ### Checklist
Check what applies to you. *Add an X between the brackets or click the checkboxes when you have submitted the issue.*
- [x] I have [pruned](https://pgm.dev/docs/guides/packaging/pruning-chunks) the map
- [x] I have agreed with assigning the CC BY-SA license to this map, as mentioned in the README
- [x] I have created an XML file
- [x] I have created an map image
- [x] I have uploaded the map zip file to a file sharing service
- [x] The map has been tested locally to make sure it works
<img src="https://i.imgur.com/LC0aHbQ.png" alt="map" style="max-width:100%;">
# Map Name
chle_ (yeah, thats the name)
## Gamemode & Map Description
This map does not have a void players have to bridge over. Instead I went fully experimental and made a huge wall players have to dig through in order to get to the other side. While in the wall, players are given night vision and haste 2. The rest of the map is just straight forward. Oh, and no, the large roof supported by the pillars is just there for decoration.
The woolroom however, I believe it has the largest interior of any woolroom in OCN history, so be sure to check out every corner of it!
I have absolutely no idea how this is gonna play lmao
## Team Sizes
Red vs. Blue 40v40
## Screenshots
[here u go](https://imgur.com/a/AYxPmV4)
# XML
[XML](https://gist.github.com/chleongithub/cd69c8b7545a7c7c054c59779b96c037)
# Map Image
<img src="https://i.imgur.com/9pZHhZL.png" alt="map" style="max-width:100%;">
# Download Link
[Download](https://www.dropbox.com/s/i89krl5fziz4bzu/chle_.zip?dl=0)
|
test
|
chle checklist check what applies to you add an x between the brackets or click the checkboxes when you have submitted the issue i have the map i have agreed with assigning the cc by sa license to this map as mentioned in the readme i have created an xml file i have created an map image i have uploaded the map zip file to a file sharing service the map has been tested locally to make sure it works map name chle yeah thats the name gamemode map description this map does not have a void players have to bridge over instead i went fully experimental and made a huge wall players have to dig through in order to get to the other side while in the wall players are given night vision and haste the rest of the map is just straight forward oh and no the large roof supported by the pillars is just there for decoration the woolroom however i believe it has the largest interior of any woolroom in ocn history so be sure to check out every corner of it i have absolutely no idea how this is gonna play lmao team sizes red vs blue screenshots xml map image download link
| 1
|
18,461
| 2,615,171,586
|
IssuesEvent
|
2015-03-01 06:53:34
|
chrsmith/html5rocks
|
https://api.github.com/repos/chrsmith/html5rocks
|
opened
|
Updating your Resource Section
|
auto-migrated Priority-P2 Type-Bug
|
```
Hi, first of all I would like to show my thanks to any researchers, developers
or designers that have developed HTML5Rocks, it has and will continue to be a
superb resource for all things HTML5.
A few months a go I developed a similar site (http://www.html5tuts.co.uk) for a
university project and i have recently been updating it in terms of content and
the design of the website. The reason I decided to send this message is because
I noticed in your resource section that you have a number of good sites but it
hasnt been updated for a while. As HTML5 is all about spreading the word and
pushing the standard out to as many people as possible I thought I would ask if
you could feature my site on the resource page?
Secondly, I wanted to ask about the possiblity of writing content for the
website, I'd need to go to here:
http://code.google.com/p/html5rocks/wiki/ContributorsGuide to test my article
but I wondered if there were any particular topics that the HTML5Rocks team
would like on the site at the moment?
Thanks and Best Regards,
Alexander Jones
```
Original issue reported on code.google.com by `aljones....@gmail.com` on 7 Dec 2011 at 2:29
|
1.0
|
Updating your Resource Section - ```
Hi, first of all I would like to show my thanks to any researchers, developers
or designers that have developed HTML5Rocks, it has and will continue to be a
superb resource for all things HTML5.
A few months a go I developed a similar site (http://www.html5tuts.co.uk) for a
university project and i have recently been updating it in terms of content and
the design of the website. The reason I decided to send this message is because
I noticed in your resource section that you have a number of good sites but it
hasnt been updated for a while. As HTML5 is all about spreading the word and
pushing the standard out to as many people as possible I thought I would ask if
you could feature my site on the resource page?
Secondly, I wanted to ask about the possiblity of writing content for the
website, I'd need to go to here:
http://code.google.com/p/html5rocks/wiki/ContributorsGuide to test my article
but I wondered if there were any particular topics that the HTML5Rocks team
would like on the site at the moment?
Thanks and Best Regards,
Alexander Jones
```
Original issue reported on code.google.com by `aljones....@gmail.com` on 7 Dec 2011 at 2:29
|
non_test
|
updating your resource section hi first of all i would like to show my thanks to any researchers developers or designers that have developed it has and will continue to be a superb resource for all things a few months a go i developed a similar site for a university project and i have recently been updating it in terms of content and the design of the website the reason i decided to send this message is because i noticed in your resource section that you have a number of good sites but it hasnt been updated for a while as is all about spreading the word and pushing the standard out to as many people as possible i thought i would ask if you could feature my site on the resource page secondly i wanted to ask about the possiblity of writing content for the website i d need to go to here to test my article but i wondered if there were any particular topics that the team would like on the site at the moment thanks and best regards alexander jones original issue reported on code google com by aljones gmail com on dec at
| 0
|
13,232
| 22,346,754,366
|
IssuesEvent
|
2022-06-15 08:30:08
|
4l3x-suvnet/todo_calendar
|
https://api.github.com/repos/4l3x-suvnet/todo_calendar
|
closed
|
Save todo's in local storage
|
additional requirement
|
On page refresh, the todo's should be saved in local storage using localStorage with getItem and setItem(?)
|
1.0
|
Save todo's in local storage - On page refresh, the todo's should be saved in local storage using localStorage with getItem and setItem(?)
|
non_test
|
save todo s in local storage on page refresh the todo s should be saved in local storage using localstorage with getitem and setitem
| 0
|
625,758
| 19,765,019,976
|
IssuesEvent
|
2022-01-17 00:07:56
|
internetarchive/openlibrary
|
https://api.github.com/repos/internetarchive/openlibrary
|
opened
|
Better Share Preview for Lists & Reading Log (e.g. Twitter Social Card)
|
Type: Bug Theme: Lists Priority: 2 Lead: @mekarpeles Theme: Distribution
|
<!-- What problem are we solving? What does the experience look like today? What are the symptoms? -->
This is a create "social card" view of a book list (from fivebooks). Would be great if Open Library lists generated this type of preview :slightly_smiling_face:
see: https://twitter.com/jabuppartyon/status/1482812981591347206

### Evidence / Screenshot (if possible)
### Relevant url?
<!-- `https://openlibrary.org/...` -->
### Steps to Reproduce
<!-- What steps caused you to find the bug? -->
1. Go to ...
2. Do ...
<!-- What actually happened after these steps? What did you expect to happen? -->
* Actual:
* Expected:
### Details
- **Logged in (Y/N)?**
- **Browser type/version?**
- **Operating system?**
- **Environment (prod/dev/local)?** prod
<!-- If not sure, put prod -->
### Proposal & Constraints
<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->
### Related files
<!-- Files related to this issue; this is super useful for new contributors who might want to help! If you're not sure, leave this blank; a maintainer will add them. -->
### Stakeholders
<!-- @ tag stakeholders of this bug -->
|
1.0
|
Better Share Preview for Lists & Reading Log (e.g. Twitter Social Card) - <!-- What problem are we solving? What does the experience look like today? What are the symptoms? -->
This is a create "social card" view of a book list (from fivebooks). Would be great if Open Library lists generated this type of preview :slightly_smiling_face:
see: https://twitter.com/jabuppartyon/status/1482812981591347206

### Evidence / Screenshot (if possible)
### Relevant url?
<!-- `https://openlibrary.org/...` -->
### Steps to Reproduce
<!-- What steps caused you to find the bug? -->
1. Go to ...
2. Do ...
<!-- What actually happened after these steps? What did you expect to happen? -->
* Actual:
* Expected:
### Details
- **Logged in (Y/N)?**
- **Browser type/version?**
- **Operating system?**
- **Environment (prod/dev/local)?** prod
<!-- If not sure, put prod -->
### Proposal & Constraints
<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->
### Related files
<!-- Files related to this issue; this is super useful for new contributors who might want to help! If you're not sure, leave this blank; a maintainer will add them. -->
### Stakeholders
<!-- @ tag stakeholders of this bug -->
|
non_test
|
better share preview for lists reading log e g twitter social card this is a create social card view of a book list from fivebooks would be great if open library lists generated this type of preview slightly smiling face see evidence screenshot if possible relevant url steps to reproduce go to do actual expected details logged in y n browser type version operating system environment prod dev local prod proposal constraints related files stakeholders
| 0
|
164,962
| 12,823,992,653
|
IssuesEvent
|
2020-07-06 12:45:31
|
ICIJ/datashare
|
https://api.github.com/repos/ICIJ/datashare
|
closed
|
Replace OKhttp by the ESClient
|
back enhancement need testing
|
As Bruno found out during our performance sprint, we can replace OkHttp by ESclient for more stability and less complexity.
|
1.0
|
Replace OKhttp by the ESClient - As Bruno found out during our performance sprint, we can replace OkHttp by ESclient for more stability and less complexity.
|
test
|
replace okhttp by the esclient as bruno found out during our performance sprint we can replace okhttp by esclient for more stability and less complexity
| 1
|
82,144
| 7,818,366,307
|
IssuesEvent
|
2018-06-13 12:05:50
|
magento-engcom/msi
|
https://api.github.com/repos/magento-engcom/msi
|
opened
|
TechDebt for #1371 eliminate fixture source_items_on_mixed_sources
|
Award: test coverage
|
TechDebt for #1371 eliminate fixture introduced in PR #1371 source_items_on_mixed_sources.php
and re-use existing fixtures:
https://github.com/magento-engcom/msi/blob/2.3-develop/app/code/Magento/InventoryCatalog/Test/_files/source_items_on_default_source.php
https://github.com/magento-engcom/msi/blob/2.3-develop/app/code/Magento/InventoryApi/Test/_files/source_items.php
|
1.0
|
TechDebt for #1371 eliminate fixture source_items_on_mixed_sources - TechDebt for #1371 eliminate fixture introduced in PR #1371 source_items_on_mixed_sources.php
and re-use existing fixtures:
https://github.com/magento-engcom/msi/blob/2.3-develop/app/code/Magento/InventoryCatalog/Test/_files/source_items_on_default_source.php
https://github.com/magento-engcom/msi/blob/2.3-develop/app/code/Magento/InventoryApi/Test/_files/source_items.php
|
test
|
techdebt for eliminate fixture source items on mixed sources techdebt for eliminate fixture introduced in pr source items on mixed sources php and re use existing fixtures
| 1
|
281,996
| 24,443,017,010
|
IssuesEvent
|
2022-10-06 15:46:06
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
[Azure AD] Always check access token from Microsoft
|
kind/enhancement area/authentication [zube]: To Test team/area1
|
It would be good to add this guardrail to prevent users from losing access to Rancher via Azure AD or getting their Azure AD setup in an invalid state. This can happen if they don't follow the documentation in setting the necessary permissions for their App registrations in Azure for Rancher v2.6.7+.
On initial Azure AD setup or subsequent login or search attempts, always check the access token to Microsoft Graph if it does not have the necessary permissions.
|
1.0
|
[Azure AD] Always check access token from Microsoft - It would be good to add this guardrail to prevent users from losing access to Rancher via Azure AD or getting their Azure AD setup in an invalid state. This can happen if they don't follow the documentation in setting the necessary permissions for their App registrations in Azure for Rancher v2.6.7+.
On initial Azure AD setup or subsequent login or search attempts, always check the access token to Microsoft Graph if it does not have the necessary permissions.
|
test
|
always check access token from microsoft it would be good to add this guardrail to prevent users from losing access to rancher via azure ad or getting their azure ad setup in an invalid state this can happen if they don t follow the documentation in setting the necessary permissions for their app registrations in azure for rancher on initial azure ad setup or subsequent login or search attempts always check the access token to microsoft graph if it does not have the necessary permissions
| 1
|
211,770
| 7,204,539,897
|
IssuesEvent
|
2018-02-06 13:04:32
|
Chingu-cohorts/voyage-wiki
|
https://api.github.com/repos/Chingu-cohorts/voyage-wiki
|
closed
|
Link to Carnegie wisdom article missing from FAQ
|
priority:must have scope:story state:done type:bug
|
In `Section 99. Frequently Asked Questions`, in the section titled "[Are Soft Skills Important?](https://github.com/Chingu-cohorts/voyage-wiki/wiki/Section-99.-Frequently-Asked-Questions#are-soft-skills-important)"
The last line, which reads `Here are 30 soft skill principles that can level-up your skills `, should (presumably) have a link to the following article:
https://medium.com/chingu/30-ways-to-level-up-your-skills-stack-carnegies-wisdom-remix-4c532aabd0c0
(As per the "Chingu Voyage-4 Prep Pack" email)
|
1.0
|
Link to Carnegie wisdom article missing from FAQ - In `Section 99. Frequently Asked Questions`, in the section titled "[Are Soft Skills Important?](https://github.com/Chingu-cohorts/voyage-wiki/wiki/Section-99.-Frequently-Asked-Questions#are-soft-skills-important)"
The last line, which reads `Here are 30 soft skill principles that can level-up your skills `, should (presumably) have a link to the following article:
https://medium.com/chingu/30-ways-to-level-up-your-skills-stack-carnegies-wisdom-remix-4c532aabd0c0
(As per the "Chingu Voyage-4 Prep Pack" email)
|
non_test
|
link to carnegie wisdom article missing from faq in section frequently asked questions in the section titled the last line which reads here are soft skill principles that can level up your skills should presumably have a link to the following article as per the chingu voyage prep pack email
| 0
|
34,414
| 4,918,458,018
|
IssuesEvent
|
2016-11-24 08:59:42
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
github.com/cockroachdb/cockroach/vendor/github.com/coreos/etcd/wal: TestCut failed under stress
|
Robot test-failure
|
SHA: https://github.com/cockroachdb/cockroach/commits/b54490b2cf70c155ec2b7af5133276ffe24dc02c
Parameters:
```
COCKROACH_PROPOSER_EVALUATED_KV=true
TAGS=stress
GOFLAGS=
```
Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=58266&tab=buildLog
```
wal_test.go:163: no space left on device
```
|
1.0
|
github.com/cockroachdb/cockroach/vendor/github.com/coreos/etcd/wal: TestCut failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/b54490b2cf70c155ec2b7af5133276ffe24dc02c
Parameters:
```
COCKROACH_PROPOSER_EVALUATED_KV=true
TAGS=stress
GOFLAGS=
```
Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=58266&tab=buildLog
```
wal_test.go:163: no space left on device
```
|
test
|
github com cockroachdb cockroach vendor github com coreos etcd wal testcut failed under stress sha parameters cockroach proposer evaluated kv true tags stress goflags stress build found a failed test wal test go no space left on device
| 1
|
412,779
| 27,872,520,404
|
IssuesEvent
|
2023-03-21 14:17:09
|
hyperledger/aries-framework-go
|
https://api.github.com/repos/hyperledger/aries-framework-go
|
closed
|
SD-JWT Documentation
|
documentation
|
Document SD-JWT packages:
- high level intro
- issuer
- holder
- verifier
- common
- examples
|
1.0
|
SD-JWT Documentation - Document SD-JWT packages:
- high level intro
- issuer
- holder
- verifier
- common
- examples
|
non_test
|
sd jwt documentation document sd jwt packages high level intro issuer holder verifier common examples
| 0
|
182,973
| 14,926,065,434
|
IssuesEvent
|
2021-01-24 09:45:31
|
goat-community/goat
|
https://api.github.com/repos/goat-community/goat
|
opened
|
Add documentation scenario:bridge
|
documentation
|
We should somewhere explain what is the difference between a new bridge and a new road.
-> which properties are assumes
-> that the differentiation is only important for the routing -> a new “road” in GOAT can be anything – a footpath, a pedestrian zone, a living street, a residential road, etc.
-> surface is important for the cycling speed
-> wheelchair usability is important for the wheelchair routing
I am just not sure yet, where to put this information. Maybe adding a new chapter to Docs -> Technical Documentation -> "Scenarios"
|
1.0
|
Add documentation scenario:bridge - We should somewhere explain what is the difference between a new bridge and a new road.
-> which properties are assumes
-> that the differentiation is only important for the routing -> a new “road” in GOAT can be anything – a footpath, a pedestrian zone, a living street, a residential road, etc.
-> surface is important for the cycling speed
-> wheelchair usability is important for the wheelchair routing
I am just not sure yet, where to put this information. Maybe adding a new chapter to Docs -> Technical Documentation -> "Scenarios"
|
non_test
|
add documentation scenario bridge we should somewhere explain what is the difference between a new bridge and a new road which properties are assumes that the differentiation is only important for the routing a new “road” in goat can be anything – a footpath a pedestrian zone a living street a residential road etc surface is important for the cycling speed wheelchair usability is important for the wheelchair routing i am just not sure yet where to put this information maybe adding a new chapter to docs technical documentation scenarios
| 0
|
284,128
| 24,579,990,395
|
IssuesEvent
|
2022-10-13 14:57:14
|
PennLINC/xcp_d
|
https://api.github.com/repos/PennLINC/xcp_d
|
closed
|
[RF] Move PyTests to relevant directory
|
refactor testing
|
So that it is easier to link the relevant test and the relevant function
|
1.0
|
[RF] Move PyTests to relevant directory - So that it is easier to link the relevant test and the relevant function
|
test
|
move pytests to relevant directory so that it is easier to link the relevant test and the relevant function
| 1
|
124,511
| 17,772,595,721
|
IssuesEvent
|
2021-08-30 15:13:52
|
kapseliboi/energy-futures-vis-avenir-energetique
|
https://api.github.com/repos/kapseliboi/energy-futures-vis-avenir-energetique
|
opened
|
CVE-2021-23362 (Medium) detected in hosted-git-info-2.7.1.tgz
|
security vulnerability
|
## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.7.1.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.7.1.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.7.1.tgz</a></p>
<p>Path to dependency file: energy-futures-vis-avenir-energetique/package.json</p>
<p>Path to vulnerable library: energy-futures-vis-avenir-energetique/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- conventional-changelog-cli-2.0.28.tgz (Root Library)
- meow-5.0.0.tgz
- normalize-package-data-2.4.0.tgz
- :x: **hosted-git-info-2.7.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/energy-futures-vis-avenir-energetique/commit/907b3c15edb7159764857453edc4f32b2432cdd4">907b3c15edb7159764857453edc4f32b2432cdd4</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: hosted-git-info - 2.8.9,3.0.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23362 (Medium) detected in hosted-git-info-2.7.1.tgz - ## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.7.1.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.7.1.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.7.1.tgz</a></p>
<p>Path to dependency file: energy-futures-vis-avenir-energetique/package.json</p>
<p>Path to vulnerable library: energy-futures-vis-avenir-energetique/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- conventional-changelog-cli-2.0.28.tgz (Root Library)
- meow-5.0.0.tgz
- normalize-package-data-2.4.0.tgz
- :x: **hosted-git-info-2.7.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/energy-futures-vis-avenir-energetique/commit/907b3c15edb7159764857453edc4f32b2432cdd4">907b3c15edb7159764857453edc4f32b2432cdd4</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: hosted-git-info - 2.8.9,3.0.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in hosted git info tgz cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file energy futures vis avenir energetique package json path to vulnerable library energy futures vis avenir energetique node modules hosted git info package json dependency hierarchy conventional changelog cli tgz root library meow tgz normalize package data tgz x hosted git info tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info step up your open source security game with whitesource
| 0
|
46,241
| 19,012,777,804
|
IssuesEvent
|
2021-11-23 11:08:33
|
kyma-incubator/reconciler
|
https://api.github.com/repos/kyma-incubator/reconciler
|
opened
|
Istio reconciler support istio 1.12
|
area/service-mesh area/security
|
<!-- Thank you for your contribution. Before you submit the issue:
1. Search open and closed issues for duplicates.
2. Read the contributing guidelines.
-->
**Description**
Implement necessary changes to install / upgrade istio in 1.12 version. For this story we do not need multiple version support.
- [ ] istioctl 1.12 is used for installation / upgrade
- [ ] istio running in 1.12 version
- [ ] sidecars running in 1.12 version
- [ ] reconciler bumped in kyma cli
**Reasons**
Keeping Istio up to date
**Attachments**
part of https://github.com/kyma-project/kyma/issues/12668
|
1.0
|
Istio reconciler support istio 1.12 - <!-- Thank you for your contribution. Before you submit the issue:
1. Search open and closed issues for duplicates.
2. Read the contributing guidelines.
-->
**Description**
Implement necessary changes to install / upgrade istio in 1.12 version. For this story we do not need multiple version support.
- [ ] istioctl 1.12 is used for installation / upgrade
- [ ] istio running in 1.12 version
- [ ] sidecars running in 1.12 version
- [ ] reconciler bumped in kyma cli
**Reasons**
Keeping Istio up to date
**Attachments**
part of https://github.com/kyma-project/kyma/issues/12668
|
non_test
|
istio reconciler support istio thank you for your contribution before you submit the issue search open and closed issues for duplicates read the contributing guidelines description implement necessary changes to install upgrade istio in version for this story we do not need multiple version support istioctl is used for installation upgrade istio running in version sidecars running in version reconciler bumped in kyma cli reasons keeping istio up to date attachments part of
| 0
|
286,754
| 24,782,234,357
|
IssuesEvent
|
2022-10-24 06:40:00
|
tgstation/tgstation
|
https://api.github.com/repos/tgstation/tgstation
|
closed
|
Paper is missing on-mob sprites
|
Sprites Bug Tested/Reproduced
|
Reporting client version: 514.1589
<!-- Write **BELOW** The Headers and **ABOVE** The comments else it may not be viewable -->
## Round ID:
[192873](https://scrubby.melonmesa.com/round/192873)
<!--- **INCLUDE THE ROUND ID**
If you discovered this issue from playing tgstation hosted servers:
[Round ID]: # (It can be found in the Status panel or retrieved from https://sb.atlantaned.space/rounds ! The round id let's us look up valuable information and logs for the round the bug happened.)-->
<!-- If you are reporting an issue found in another branch or codebase, you MUST link the branch or codebase repo in your issue report or it will be closed. For branches, If you have not pushed your code up, you must either reproduce it on master or push your code up before making an issue report. For other codebases, if you do not have a public code repo you will be refused help unless you can completely reproduce the issue on our code. -->
## Testmerges:
- [JSON Savefiles | Player Saves use JSON](https://www.github.com/tgstation/tgstation/pull/70492)
<!-- If you're certain the issue is to be caused by a test merge [OOC tab -> Show Server Revision], report it in the pull request's comment section rather than on the tracker(If you're unsure you can refer to the issue number by prefixing said number with #. The issue number can be found beside the title after submitting it to the tracker).If no testmerges are active, feel free to remove this section. -->
## Reproduction:
1. Pick up a piece of paper.
2. Put the piece of paper on your head.
3. Missing Textures.
Putting a piece of paper on your head normally gives you a paper hat, but the sprite seems to be missing.
<!-- Explain your issue in detail, including the steps to reproduce it. Issues without proper reproduction steps or explanation are open to being ignored/closed by maintainers.-->
<!-- **For Admins:** Oddities induced by var-edits and other admin tools are not necessarily bugs. Verify that your issues occur under regular circumstances before reporting them. -->
<!-- If you are reporting a runtime error you must include the runtime in your report or your report will be closed. -->
|
1.0
|
Paper is missing on-mob sprites - Reporting client version: 514.1589
<!-- Write **BELOW** The Headers and **ABOVE** The comments else it may not be viewable -->
## Round ID:
[192873](https://scrubby.melonmesa.com/round/192873)
<!--- **INCLUDE THE ROUND ID**
If you discovered this issue from playing tgstation hosted servers:
[Round ID]: # (It can be found in the Status panel or retrieved from https://sb.atlantaned.space/rounds ! The round id let's us look up valuable information and logs for the round the bug happened.)-->
<!-- If you are reporting an issue found in another branch or codebase, you MUST link the branch or codebase repo in your issue report or it will be closed. For branches, If you have not pushed your code up, you must either reproduce it on master or push your code up before making an issue report. For other codebases, if you do not have a public code repo you will be refused help unless you can completely reproduce the issue on our code. -->
## Testmerges:
- [JSON Savefiles | Player Saves use JSON](https://www.github.com/tgstation/tgstation/pull/70492)
<!-- If you're certain the issue is to be caused by a test merge [OOC tab -> Show Server Revision], report it in the pull request's comment section rather than on the tracker(If you're unsure you can refer to the issue number by prefixing said number with #. The issue number can be found beside the title after submitting it to the tracker).If no testmerges are active, feel free to remove this section. -->
## Reproduction:
1. Pick up a piece of paper.
2. Put the piece of paper on your head.
3. Missing Textures.
Putting a piece of paper on your head normally gives you a paper hat, but the sprite seems to be missing.
<!-- Explain your issue in detail, including the steps to reproduce it. Issues without proper reproduction steps or explanation are open to being ignored/closed by maintainers.-->
<!-- **For Admins:** Oddities induced by var-edits and other admin tools are not necessarily bugs. Verify that your issues occur under regular circumstances before reporting them. -->
<!-- If you are reporting a runtime error you must include the runtime in your report or your report will be closed. -->
|
test
|
paper is missing on mob sprites reporting client version round id include the round id if you discovered this issue from playing tgstation hosted servers it can be found in the status panel or retrieved from the round id let s us look up valuable information and logs for the round the bug happened testmerges reproduction pick up a piece of paper put the piece of paper on your head missing textures putting a piece of paper on your head normally gives you a paper hat but the sprite seems to be missing
| 1
|
49,788
| 6,040,112,469
|
IssuesEvent
|
2017-06-10 10:59:17
|
Princeton-CDH/winthrop-django
|
https://api.github.com/repos/Princeton-CDH/winthrop-django
|
closed
|
annotation + subject
|
awaiting testing
|
As an annotation data editor, I want the option of associating annotations with any of the subjects in the database that are used for labeling books.
|
1.0
|
annotation + subject - As an annotation data editor, I want the option of associating annotations with any of the subjects in the database that are used for labeling books.
|
test
|
annotation subject as an annotation data editor i want the option of associating annotations with any of the subjects in the database that are used for labeling books
| 1
|
192,166
| 14,601,685,258
|
IssuesEvent
|
2020-12-21 09:04:40
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
opened
|
First click on AD notification will not open AD in a view mode
|
OS/Desktop QA/Test-Plan-Specified QA/Yes bug feature/ads regression
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
First click on AD notification will not open AD in a view mode
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile 1.20.x staging environment
2. Enable rewards
3. View an AD
4. Click on AD notification to view an AD
5. AD isn't displayed in VIEW mode
6. Repeat step 4, AD isn't displayed in VIEW mode
7. Repeat step 4, AD is displayed in the view mode
## Actual result:
<!--Please add screenshots if needed-->
The first click on AD notification will not open AD in a view mode
## Expected result:
Click on AD notification should open the AD in a view mode
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easy
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.20.40 Chromium: 88.0.4315.5 (Official Build) nightly (64-bit)
-- | --
Revision | d2abe4588adaeb0d66aa0a0106675be1dff12c64-refs/branch-heads/4315@{#11}
OS | Windows 10 OS Version 2004 (Build 19041.685)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? No
- Can you reproduce this issue with the beta channel? No
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? na
- Does the issue resolve itself when disabling Brave Rewards? na
- Is the issue reproducible on the latest version of Chrome? na
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc: @brave/legacy_qa @rebron @tmancey
|
1.0
|
First click on AD notification will not open AD in a view mode - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
First click on AD notification will not open AD in a view mode
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile 1.20.x staging environment
2. Enable rewards
3. View an AD
4. Click on AD notification to view an AD
5. AD isn't displayed in VIEW mode
6. Repeat step 4, AD isn't displayed in VIEW mode
7. Repeat step 4, AD is displayed in the view mode
## Actual result:
<!--Please add screenshots if needed-->
The first click on AD notification will not open AD in a view mode
## Expected result:
Click on AD notification should open the AD in a view mode
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easy
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.20.40 Chromium: 88.0.4315.5 (Official Build) nightly (64-bit)
-- | --
Revision | d2abe4588adaeb0d66aa0a0106675be1dff12c64-refs/branch-heads/4315@{#11}
OS | Windows 10 OS Version 2004 (Build 19041.685)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? No
- Can you reproduce this issue with the beta channel? No
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? na
- Does the issue resolve itself when disabling Brave Rewards? na
- Is the issue reproducible on the latest version of Chrome? na
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc: @brave/legacy_qa @rebron @tmancey
|
test
|
first click on ad notification will not open ad in a view mode have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description first click on ad notification will not open ad in a view mode steps to reproduce clean profile x staging environment enable rewards view an ad click on ad notification to view an ad ad isn t displayed in view mode repeat step ad isn t displayed in view mode repeat step ad is displayed in the view mode actual result the first click on ad notification will not open ad in a view mode expected result click on ad notification should open the ad in a view mode reproduces how often easy brave version brave version info brave chromium official build nightly bit revision refs branch heads os windows os version build version channel information can you reproduce this issue with the current release no can you reproduce this issue with the beta channel no can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information cc brave legacy qa rebron tmancey
| 1
|
200,183
| 15,792,875,797
|
IssuesEvent
|
2021-04-02 07:56:51
|
SeitaBV/flexmeasures
|
https://api.github.com/repos/SeitaBV/flexmeasures
|
closed
|
Restructure documentation
|
documentation
|
We want to guide people with different intentions to different parts of the docs ― for people interesting in the purpose and benefits of FlexMeasures or for developers looking for installation and setup advice.
Both can use high quality content like tutorials, but for now we are looking for the right structure.
|
1.0
|
Restructure documentation - We want to guide people with different intentions to different parts of the docs ― for people interesting in the purpose and benefits of FlexMeasures or for developers looking for installation and setup advice.
Both can use high quality content like tutorials, but for now we are looking for the right structure.
|
non_test
|
restructure documentation we want to guide people with different intentions to different parts of the docs ― for people interesting in the purpose and benefits of flexmeasures or for developers looking for installation and setup advice both can use high quality content like tutorials but for now we are looking for the right structure
| 0
|
9,120
| 24,055,646,504
|
IssuesEvent
|
2022-09-16 16:34:55
|
MicrosoftDocs/architecture-center
|
https://api.github.com/repos/MicrosoftDocs/architecture-center
|
closed
|
Possibly wrong category for scenario
|
doc-enhancement assigned-to-author triaged architecture-center/svc example-scenario/subsvc Pri2
|
Should this scenario really be categorized/organized as "DevOps"? Perhaps Web or Mobile would be more fitting?
/C
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 36cc3d81-d33a-bc9b-ab91-d7c4c8e26a95
* Version Independent ID: d3ce3924-f004-a61a-867c-b81fae8cf5f9
* Content: [Sharing location in real time using low-cost serverless Azure services - Azure Example Scenarios](https://docs.microsoft.com/en-us/azure/architecture/example-scenario/signalr/)
* Content Source: [docs/example-scenario/signalr/index.yml](https://github.com/microsoftdocs/architecture-center/blob/master/docs/example-scenario/signalr/index.yml)
* Service: **architecture-center**
* Sub-service: **example-scenario**
* GitHub Login: @talhanaveed
* Microsoft Alias: **pnp**
|
1.0
|
Possibly wrong category for scenario - Should this scenario really be categorized/organized as "DevOps"? Perhaps Web or Mobile would be more fitting?
/C
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 36cc3d81-d33a-bc9b-ab91-d7c4c8e26a95
* Version Independent ID: d3ce3924-f004-a61a-867c-b81fae8cf5f9
* Content: [Sharing location in real time using low-cost serverless Azure services - Azure Example Scenarios](https://docs.microsoft.com/en-us/azure/architecture/example-scenario/signalr/)
* Content Source: [docs/example-scenario/signalr/index.yml](https://github.com/microsoftdocs/architecture-center/blob/master/docs/example-scenario/signalr/index.yml)
* Service: **architecture-center**
* Sub-service: **example-scenario**
* GitHub Login: @talhanaveed
* Microsoft Alias: **pnp**
|
non_test
|
possibly wrong category for scenario should this scenario really be categorized organized as devops perhaps web or mobile would be more fitting c document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service example scenario github login talhanaveed microsoft alias pnp
| 0
|
328,617
| 28,128,072,212
|
IssuesEvent
|
2023-03-31 19:39:26
|
StrataSource/Engine
|
https://api.github.com/repos/StrataSource/Engine
|
closed
|
Bug: Timescale doesn't affect sounds without sv_cheats 1
|
Type: Bug Status: Needs Testing
|
### Describe the bug
The title have described the issue pretty clearly - Timescale changes done by +slowtime or logic_timescale **will not** change the playback speed and pitch of the sounds if sv_cheats is set to 0
https://user-images.githubusercontent.com/57415598/192872606-f9e8c5ff-886b-4274-9402-b2f14026a04d.mp4
### To Reproduce
1. Open a map
2. Make sure sv_cheats is set to 0
3. Change your timescale via +slowtime or logic_timescale
### Issue Map
This issue is not tied to specific maps and happens everywhere
### Expected Behavior
Timescale affect sounds regardless of sv_cheats, like in original Portal 2
### Operating System
Windows 10 Home
|
1.0
|
Bug: Timescale doesn't affect sounds without sv_cheats 1 - ### Describe the bug
The title have described the issue pretty clearly - Timescale changes done by +slowtime or logic_timescale **will not** change the playback speed and pitch of the sounds if sv_cheats is set to 0
https://user-images.githubusercontent.com/57415598/192872606-f9e8c5ff-886b-4274-9402-b2f14026a04d.mp4
### To Reproduce
1. Open a map
2. Make sure sv_cheats is set to 0
3. Change your timescale via +slowtime or logic_timescale
### Issue Map
This issue is not tied to specific maps and happens everywhere
### Expected Behavior
Timescale affect sounds regardless of sv_cheats, like in original Portal 2
### Operating System
Windows 10 Home
|
test
|
bug timescale doesn t affect sounds without sv cheats describe the bug the title have described the issue pretty clearly timescale changes done by slowtime or logic timescale will not change the playback speed and pitch of the sounds if sv cheats is set to to reproduce open a map make sure sv cheats is set to change your timescale via slowtime or logic timescale issue map this issue is not tied to specific maps and happens everywhere expected behavior timescale affect sounds regardless of sv cheats like in original portal operating system windows home
| 1
|
39,711
| 5,243,171,364
|
IssuesEvent
|
2017-01-31 19:58:50
|
mailchimp/mc-magento
|
https://api.github.com/repos/mailchimp/mc-magento
|
closed
|
Wrong format for mailchimp_sync_delta field
|
done no interface testing
|
Hi there!
**Preconditions**
1. Magento 1.9.2.4
2. mc-magento 1.1.2
3. Real data is included and Products are exported to MailChimp
**Steps to reproduce**
1. Go to Catalog Product Edit Page in Admin Interface
2. Edit any data
3. Press "Save button"
4. Look how _mailchimp_sync_delta_ field is processing in **Mage_Core_Controller_Varien_Action->_filterDates()**
**Expected result**
1. After **$filterInput->filter** _mailchimp_sync_delta_ becomes correct date value
**Actual result**
1. It becomes uncorrect, for example:
* 2016-11-29 18:25:35 => 016 -11 -29 18 :25 :35
It brokes correct behaviour:
1. **Zend_Locale_Format::checkDateFormat** can't parse this date
2. After this **$filterInternal** get _mailchimp_sync_delta_ as float and tries to get precision by subtraction, it causes _A non well formed numeric value encountered_ error
Hope it will be fixed soon.
Regards, Nikolay
|
1.0
|
Wrong format for mailchimp_sync_delta field - Hi there!
**Preconditions**
1. Magento 1.9.2.4
2. mc-magento 1.1.2
3. Real data is included and Products are exported to MailChimp
**Steps to reproduce**
1. Go to Catalog Product Edit Page in Admin Interface
2. Edit any data
3. Press "Save button"
4. Look how _mailchimp_sync_delta_ field is processing in **Mage_Core_Controller_Varien_Action->_filterDates()**
**Expected result**
1. After **$filterInput->filter** _mailchimp_sync_delta_ becomes correct date value
**Actual result**
1. It becomes uncorrect, for example:
* 2016-11-29 18:25:35 => 016 -11 -29 18 :25 :35
It brokes correct behaviour:
1. **Zend_Locale_Format::checkDateFormat** can't parse this date
2. After this **$filterInternal** get _mailchimp_sync_delta_ as float and tries to get precision by subtraction, it causes _A non well formed numeric value encountered_ error
Hope it will be fixed soon.
Regards, Nikolay
|
test
|
wrong format for mailchimp sync delta field hi there preconditions magento mc magento real data is included and products are exported to mailchimp steps to reproduce go to catalog product edit page in admin interface edit any data press save button look how mailchimp sync delta field is processing in mage core controller varien action filterdates expected result after filterinput filter mailchimp sync delta becomes correct date value actual result it becomes uncorrect for example it brokes correct behaviour zend locale format checkdateformat can t parse this date after this filterinternal get mailchimp sync delta as float and tries to get precision by subtraction it causes a non well formed numeric value encountered error hope it will be fixed soon regards nikolay
| 1
|
278,894
| 24,183,668,488
|
IssuesEvent
|
2022-09-23 11:19:39
|
pravega/pravega
|
https://api.github.com/repos/pravega/pravega
|
opened
|
AssertExtensions assertEventuallyEquals error message is misleading.
|
area/testing
|
**Describe the bug**
assertEventuallyEquals printed logs are not correct and gives a false impression of Null pointer as below. where as observed value is not null its some different value other then expected.
java.util.concurrent.TimeoutException: Expected value: true observed: null
at io.pravega.test.common.AssertExtensions.assertEventuallyEquals(AssertExtensions.java:70)
Actual Exception message is to be thrown from awaitEvaluateExpr method like
throw new TimeoutException("Expected value: " + expected + " observed: " + result);
**To Reproduce**
Steps to reproduce the behavior:
call assertEventuallyEquals with condition that expected and observed are never same.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Additional information**
Add any other information about the problem here, including the root cause if known.
|
1.0
|
AssertExtensions assertEventuallyEquals error message is misleading. - **Describe the bug**
assertEventuallyEquals printed logs are not correct and gives a false impression of Null pointer as below. where as observed value is not null its some different value other then expected.
java.util.concurrent.TimeoutException: Expected value: true observed: null
at io.pravega.test.common.AssertExtensions.assertEventuallyEquals(AssertExtensions.java:70)
Actual Exception message is to be thrown from awaitEvaluateExpr method like
throw new TimeoutException("Expected value: " + expected + " observed: " + result);
**To Reproduce**
Steps to reproduce the behavior:
call assertEventuallyEquals with condition that expected and observed are never same.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Additional information**
Add any other information about the problem here, including the root cause if known.
|
test
|
assertextensions asserteventuallyequals error message is misleading describe the bug asserteventuallyequals printed logs are not correct and gives a false impression of null pointer as below where as observed value is not null its some different value other then expected java util concurrent timeoutexception expected value true observed null at io pravega test common assertextensions asserteventuallyequals assertextensions java actual exception message is to be thrown from awaitevaluateexpr method like throw new timeoutexception expected value expected observed result to reproduce steps to reproduce the behavior call asserteventuallyequals with condition that expected and observed are never same screenshots if applicable add screenshots to help explain your problem additional information add any other information about the problem here including the root cause if known
| 1
|
338,570
| 30,306,459,999
|
IssuesEvent
|
2023-07-10 09:48:21
|
n4af/TR4W
|
https://api.github.com/repos/n4af/TR4W
|
closed
|
PLAY MESSAGE_ACTIVE not working for K3
|
bug In testing Radio Control DVK
|
With the right codes in the programmed key, program is not playing the message on a K3. This message appeared in the debug log:
debug TR4WDebugLog - [K3-MemoryKeyer] Playing memory 0
I set the string to CQ SSB MEMORY F1=<03>PLAYMESSAGE_ACTIVE=1<04>
Perhaps it needs to be =01. If so, that should be fixed.
|
1.0
|
PLAY MESSAGE_ACTIVE not working for K3 - With the right codes in the programmed key, program is not playing the message on a K3. This message appeared in the debug log:
debug TR4WDebugLog - [K3-MemoryKeyer] Playing memory 0
I set the string to CQ SSB MEMORY F1=<03>PLAYMESSAGE_ACTIVE=1<04>
Perhaps it needs to be =01. If so, that should be fixed.
|
test
|
play message active not working for with the right codes in the programmed key program is not playing the message on a this message appeared in the debug log debug playing memory i set the string to cq ssb memory playmessage active perhaps it needs to be if so that should be fixed
| 1
|
159,931
| 13,776,156,478
|
IssuesEvent
|
2020-10-08 09:03:56
|
RedHatInsights/insights-operator-cli
|
https://api.github.com/repos/RedHatInsights/insights-operator-cli
|
closed
|
Added link to generated documentation for `restapi/utils.go`
|
documentation
|
Added link to generated documentation for `restapi/utils.go`
|
1.0
|
Added link to generated documentation for `restapi/utils.go` - Added link to generated documentation for `restapi/utils.go`
|
non_test
|
added link to generated documentation for restapi utils go added link to generated documentation for restapi utils go
| 0
|
93,366
| 8,411,012,965
|
IssuesEvent
|
2018-10-12 12:38:58
|
edenlabllc/ehealth.api
|
https://api.github.com/repos/edenlabllc/ehealth.api
|
closed
|
Dictionaries
|
kind/task project/medical_events status/test
|
- [x] Use dictionaries from ehealth.api fro autovalidation
- [x] dynamic codeable concepto validatione
Add dictionaries validations on all posts, patches in ME
|
1.0
|
Dictionaries - - [x] Use dictionaries from ehealth.api fro autovalidation
- [x] dynamic codeable concepto validatione
Add dictionaries validations on all posts, patches in ME
|
test
|
dictionaries use dictionaries from ehealth api fro autovalidation dynamic codeable concepto validatione add dictionaries validations on all posts patches in me
| 1
|
3,006
| 2,651,942,382
|
IssuesEvent
|
2015-03-16 14:49:37
|
metafizzy/isotope
|
https://api.github.com/repos/metafizzy/isotope
|
closed
|
Isotope Container height not updating on "append" [Help]
|
test case required
|
Hello, I am using isotope "append" method for my WordPress portfolio site to load more content by ajax and within isotope container. So far everything is working nicely except one thing that is -
1. The isotope container height is not updating automatically after each append occurs. But if I refresh the page once, and click on "load more" button, then the isotope container height is updating.
So the issue is, first time when I load my page, and then if I press "load more" button, the content is appending but the isotope container height is not updating automatically.
Please help me to fix this issue. I have less experience in jQuery and ajax and learning now.
Using isotope 2.0
Here is my code below -
<pre>
$(window).load(function(){
var $container = $('.portfolioContainer');
$container.isotope({
filter: '*',
});
/* dynamically fetching the portfolio items */
var page = 1;
var loadMorePosts = jQuery('.view-more-btn').text();
function loadMore() {
page++;
$.ajax({
method: "get",
url: './page/' + page + '/',
beforeSend: function() {
jQuery('.view-more-btn').html("<i class='fa fa-spinner fa-spin'></i>");
},
complete: function() {
},
success: function (data) {
var $data = jQuery(data).find('.portfolio');
if($data.length > 0) {
jQuery('.view-more-btn').html(loadMorePosts);
// isotope items apend
$container.isotope()
.append( $data )
.isotope( 'appended', $data )
.isotope( 'layout' );
}
else {
jQuery('.view-more-btn').html('No More Items');
}
},
error: function() {
jQuery('.view-more-btn').html('Nothing Left');
$(".view-more-btn").removeAttr('href');
}
});
}
jQuery('.view-more-btn').click(function(e) {
e.preventDefault();
loadMore();
});
/* end dynamically fetching the portfolio items */
$('.portfolioFilter a').click(function(){
$('.portfolioFilter .current').removeClass('current');
$(this).addClass('current');
var selector = $(this).attr('data-filter');
$container.isotope({
filter: selector,
});
return false;
});
});
</pre>
|
1.0
|
Isotope Container height not updating on "append" [Help] - Hello, I am using isotope "append" method for my WordPress portfolio site to load more content by ajax and within isotope container. So far everything is working nicely except one thing that is -
1. The isotope container height is not updating automatically after each append occurs. But if I refresh the page once, and click on "load more" button, then the isotope container height is updating.
So the issue is, first time when I load my page, and then if I press "load more" button, the content is appending but the isotope container height is not updating automatically.
Please help me to fix this issue. I have less experience in jQuery and ajax and learning now.
Using isotope 2.0
Here is my code below -
<pre>
$(window).load(function(){
var $container = $('.portfolioContainer');
$container.isotope({
filter: '*',
});
/* dynamically fetching the portfolio items */
var page = 1;
var loadMorePosts = jQuery('.view-more-btn').text();
function loadMore() {
page++;
$.ajax({
method: "get",
url: './page/' + page + '/',
beforeSend: function() {
jQuery('.view-more-btn').html("<i class='fa fa-spinner fa-spin'></i>");
},
complete: function() {
},
success: function (data) {
var $data = jQuery(data).find('.portfolio');
if($data.length > 0) {
jQuery('.view-more-btn').html(loadMorePosts);
// isotope items apend
$container.isotope()
.append( $data )
.isotope( 'appended', $data )
.isotope( 'layout' );
}
else {
jQuery('.view-more-btn').html('No More Items');
}
},
error: function() {
jQuery('.view-more-btn').html('Nothing Left');
$(".view-more-btn").removeAttr('href');
}
});
}
jQuery('.view-more-btn').click(function(e) {
e.preventDefault();
loadMore();
});
/* end dynamically fetching the portfolio items */
$('.portfolioFilter a').click(function(){
$('.portfolioFilter .current').removeClass('current');
$(this).addClass('current');
var selector = $(this).attr('data-filter');
$container.isotope({
filter: selector,
});
return false;
});
});
</pre>
|
test
|
isotope container height not updating on append hello i am using isotope append method for my wordpress portfolio site to load more content by ajax and within isotope container so far everything is working nicely except one thing that is the isotope container height is not updating automatically after each append occurs but if i refresh the page once and click on load more button then the isotope container height is updating so the issue is first time when i load my page and then if i press load more button the content is appending but the isotope container height is not updating automatically please help me to fix this issue i have less experience in jquery and ajax and learning now using isotope here is my code below window load function var container portfoliocontainer container isotope filter dynamically fetching the portfolio items var page var loadmoreposts jquery view more btn text function loadmore page ajax method get url page page beforesend function jquery view more btn html complete function success function data var data jquery data find portfolio if data length jquery view more btn html loadmoreposts isotope items apend container isotope append data isotope appended data isotope layout else jquery view more btn html no more items error function jquery view more btn html nothing left view more btn removeattr href jquery view more btn click function e e preventdefault loadmore end dynamically fetching the portfolio items portfoliofilter a click function portfoliofilter current removeclass current this addclass current var selector this attr data filter container isotope filter selector return false
| 1
|
739,584
| 25,603,559,401
|
IssuesEvent
|
2022-12-01 22:39:24
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
[c++] channel.WaitForConnected assertion failed while trying for mTLS connection
|
kind/bug lang/core priority/P2 untriaged
|
**Using gRPC with C++ Language
Linux version - Linux 4.19.81-OpenNetworkLinux #1 SMP Mon Jun 20 01:56:19 UTC 2022 x86_64 GNU/Linux
Python 2.7.16
gcc (Debian 8.3.0-6) 8.3.0**
**gRPC secure communication:
Created a custom channel and populated SslCredentials with ca cert, client cert and client key. Below assertion is observed once channel.WaitForConnected is called.**
I0913 17:07:25.303815824 17165 init.cc:165] grpc_init(void)
I0913 17:07:25.303940980 17165 init.cc:212] grpc_shutdown(void)
I0913 17:07:25.304056078 17165 init.cc:212] grpc_shutdown(void)
I0913 17:07:25.304331812 17165 init.cc:165] grpc_init(void)
I0913 17:07:25.304439088 17165 secure_channel_create.cc:201] grpc_secure_channel_create(creds=0x7fa6ec010d80, target=[10.220.6.50:50058](http://10.220.6.50:50058/), args=0x7fa705760590, reserved=(nil))
I0913 17:07:25.304579888 17165 init.cc:165] grpc_init(void)
I0913 17:07:25.305033297 17165 ev_posix.cc:352] (polling-api) pollset_set_create(0x7fa6ec00fc70)
I0913 17:07:25.305103697 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 1 -> 2
I0913 17:07:25.305229970 17165 client_channel.cc:1402] chand=0x7fa6ec00fa08: creating client_channel for channel stack 0x7fa6ec00f958
I0913 17:07:25.305320485 17165 ev_posix.cc:304] (polling-api) pollset_init(0x7fa6ec011c70)
I0913 17:07:25.305436421 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:796 ref 1 -> 2 pollset
I0913 17:07:25.305555151 17165 timer_generic.cc:364] TIMER 0x7fa6ec00e4d0: SET 5003 now 3 call 0x7fa6ec00e508[0x7fa70b4823cd]
I0913 17:07:25.305665221 17165 timer_generic.cc:401] .. add to shard 0 with queue_deadline_cap=0 => is_first_timer=false
I0913 17:07:25.305775850 17165 ev_posix.cc:364] (polling-api) pollset_set_add_pollset(0x7fa6ec00fc70, 0x7fa6ec011c70)
I0913 17:07:25.305887875 17165 ev_epollex_linux.cc:1450] PSS:0x7fa6ec00fc70: add pollset 0x7fa6ec011c70
I0913 17:07:25.306001577 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1241 ref 2 -> 3 pollset_as_multipollable
I0913 17:07:25.306742453 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1244 unref 3 -> 2 pollset
I0913 17:07:25.306846936 17165 ev_epollex_linux.cc:569] (fd-trace) Pollable_create: created epfd: 16 (type: 0)
I0913 17:07:25.306956168 17165 ev_epollex_linux.cc:1249] PS:0x7fa6ec011c70 active pollable transition from empty to multi
I0913 17:07:25.307059253 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec011dc8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1280 ref 1 -> 2 pollset_set
I0913 17:07:25.307174631 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1281 unref 2 -> 1 pollset_as_multipollable
I0913 17:07:25.307288891 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec011dc8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1478 unref 2 -> 1 pollset_set
I0913 17:07:25.307496739 17165 init.cc:165] grpc_init(void)
I0913 17:07:25.334856920 17165 init.cc:212] grpc_shutdown(void)
I0913 17:07:25.335164222 17165 channel_connectivity.cc:43] grpc_channel_check_connectivity_state(channel=0x7fa6ec00f8c0, try_to_connect=1)
I0913 17:07:25.335212831 17165 connectivity_state.cc:174] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: get current state: IDLE
I0913 17:07:25.335308933 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 1 -> 2 TryToConnect
I0913 17:07:25.335442749 17165 work_serializer.cc:55] WorkSerializer::Run() 0x7fa6ec00bf80 Scheduling callback [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:1926]
I0913 17:07:25.335517898 17165 work_serializer.cc:65] Executing immediately
I0913 17:07:25.335660933 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 2 -> 3 ClientChannelControlHelper
I0913 17:07:25.335822965 17165 ev_posix.cc:352] (polling-api) pollset_set_create(0x7fa6ec0105f0)
I0913 17:07:25.335905657 17165 ref_counted.h:92] lb_policy_refcount:0x7fa6ec001688 ref 1 -> 2
I0913 17:07:25.336194241 17165 ev_posix.cc:352] (polling-api) pollset_set_create(0x7fa6ec010670)
I0913 17:07:25.336355155 17165 ev_posix.cc:377] (polling-api) pollset_set_add_pollset_set(0x7fa6ec010670, 0x7fa6ec0105f0)
I0913 17:07:25.336457682 17165 ev_epollex_linux.cc:1487] PSS: merge (0x7fa6ec010670, 0x7fa6ec0105f0)
I0913 17:07:25.307856561 17166 executor.cc:224] EXECUTOR (default-executor) [0]: step (sub_depth=0)
I0913 17:07:25.307920536 17167 executor.cc:224] EXECUTOR (resolver-executor) [0]: step (sub_depth=0)
I0913 17:07:25.336576133 17165 ev_epollex_linux.cc:1521] PSS: parent 0x7fa6ec010670 to 0x7fa6ec0105f0
I0913 17:07:25.337203867 17165 resolving_lb_policy.cc:163] resolving_lb=0x7fa6ec001680: starting name resolution
I0913 17:07:25.337258063 17165 ref_counted.h:92] lb_policy_refcount:0x7fa6ec001688 ref 2 -> 3
I0913 17:07:25.337429314 17165 client_channel.cc:1306] chand=0x7fa6ec00fa08: update: state=CONNECTING picker=0x7fa6ec0127e0
I0913 17:07:25.337497200 17165 connectivity_state.cc:152] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: IDLE -> CONNECTING (helper)
I0913 17:07:25.337667613 17165 ref_counted.h:103] resolver_refcount:0x7fa6ec0129d8 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc:434 ref 1 -> 2 dns-resolving
I0913 17:07:25.338277467 17165 grpc_ares_wrapper.cc:128] (c-ares resolver) request:0x7fa6ec012b80 c-ares address sorting: input[0]=[10.220.6.50:50058](http://10.220.6.50:50058/)
I0913 17:07:25.338486711 17165 grpc_ares_wrapper.cc:128] (c-ares resolver) request:0x7fa6ec012b80 c-ares address sorting: output[0]=[10.220.6.50:50058](http://10.220.6.50:50058/)
I0913 17:07:25.338643994 17165 ev_posix.cc:377] (polling-api) pollset_set_add_pollset_set(0x7fa6ec0105f0, 0x7fa6ec00fc70)
I0913 17:07:25.338706851 17165 ev_epollex_linux.cc:1487] PSS: merge (0x7fa6ec0105f0, 0x7fa6ec00fc70)
I0913 17:07:25.338813848 17165 ev_epollex_linux.cc:1521] PSS: parent 0x7fa6ec0105f0 to 0x7fa6ec00fc70
I0913 17:07:25.339155232 17165 client_channel.cc:1581] chand=0x7fa6ec00fa08: creat[4173222.493507] serial8250: too much work for irq4
ed resolving_lb_policy=0x7fa6ec001680
I0913 17:07:25.339274241 17165 ref_counted.h:199] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:105 unref 3 -> 2 TryToConnect
I0913 17:07:25.339380121 17165 work_serializer.cc:102] WorkSerializer::DrainQueue() 0x7fa6ec00bf80
I0913 17:07:25.339499410 17165 work_serializer.cc:117] Queue Drained
I0913 17:07:25.339617022 17165 work_serializer.cc:55] WorkSerializer::Run() 0x7fa6ec00bf80 Scheduling callback [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc:327]
I0913 17:07:25.339708934 17165 work_serializer.cc:65] Executing immediately
I0913 17:07:25.339876553 17165 resolving_lb_policy.cc:293] resolving_lb=0x7fa6ec001680: got resolver result
I0913 17:07:25.340316832 17165 ref_counted.h:92] lb_policy_refcount:0x7fa6ec001688 ref 3 -> 4
I0913 17:07:25.340460426 17165 ev_posix.cc:352] (polling-api) pollset_set_create(0x7fa6ec012e20)
I0913 17:07:25.340571334 17165 resolving_lb_policy.cc:258] resolving_lb=0x7fa6ec001680: created new LB policy 0x7fa6ec012b80
I0913 17:07:25.340654584 17165 ev_posix.cc:377] (polling-api) pollset_set_add_pollset_set(0x7fa6ec012e20, 0x7fa6ec0105f0)
I0913 17:07:25.380510767 17165 ev_epollex_linux.cc:1487] PSS: merge (0x7fa6ec012e20, 0x7fa6ec0105f0)
I0913 17:07:25.380744037 17165 ev_epollex_linux.cc:1521] PSS: parent 0x7fa6ec012e20 to 0x7fa6ec00fc70
I0913 17:07:25.380469980 17168 timer_generic.cc:716] TIMER CHECK BEGIN: now=78 next=9223372036854775807 tls_min=0 glob_min=0
I0913 17:07:25.385597447 17168 timer_generic.cc:611] .. shard[0]->min_deadline = 1
I0913 17:07:25.386931416 17168 timer_generic.cc:534] .. shard[0]: heap_empty=true
I0913 17:07:25.391432826 17168 timer_generic.cc:509] .. shard[0]->queue_deadline_cap --> 1078
I0913 17:07:25.391458807 17168 timer_generic.cc:574] .. shard[0] popped 0
I0913 17:07:25.392370096 17168 timer_generic.cc:629] .. result --> 1, shard[0]->min_deadline 1 --> 1079, now=78
I0913 17:07:25.393245905 17168 timer_generic.cc:534] .. shard[1]: heap_empty=true
I0913 17:07:25.394042375 17168 timer_generic.cc:509] .. shard[1]->queue_deadline_cap --> 1078
I0913 17:07:25.394793309 17168 timer_generic.cc:574] .. shard[1] popped 0
I0913 17:07:25.395162909 17168 timer_generic.cc:629] .. result --> 1, shard[1]->min_deadline 1 --> 1079, now=78
I0913 17:07:25.384528317 17165 resolving_lb_policy.cc:240] resolving_lb=0x7fa6ec001680: Updating child policy 0x7fa6ec012b80
I0913 17:07:25.396767303 17165 child_policy_handler.cc:211] [child_policy_handler 0x7fa6ec012b80] creating new child policy pick_first
I0913 17:07:25.397911862 17165 ref_counted.h:103] lb_policy_refcount:0x7fa6ec012b88 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/lb_policy/child_policy_handler.cc:258 ref 1 -> 2 Helper
I0913 17:07:25.397978351 17165 ev_posix.cc:352] (polling-api) pollset_set_create(0x7fa6ec0130c0)
I0913 17:07:25.398036179 17165 pick_first.cc:146] Pick First 0x7fa6ec010af0 created.
I0913 17:07:25.399215100 17165 child_policy_handler.cc:272] [child_policy_handler 0x7fa6ec012b80] created new LB policy "pick_first" (0x7fa6ec010af0)
I0913 17:07:25.399387189 17165 ev_posix.cc:377] (polling-api) pollset_set_add_pollset_set(0x7fa6ec0130c0, 0x7fa6ec012e20)
I0913 17:07:25.399397525 17165 ev_epollex_linux.cc:1487] PSS: merge (0x7fa6ec0130c0, 0x7fa6ec012e20)
I0913 17:07:25.399502008 17165 ev_epollex_linux.cc:1521] PSS: parent 0x7fa6ec0130c0 to 0x7fa6ec00fc70
I0913 17:07:25.399616268 17165 child_policy_handler.cc:230] [child_policy_handler 0x7fa6ec012b80] updating child policy 0x7fa6ec010af0
I0913 17:07:25.399732763 17165 pick_first.cc:265] Pick First 0x7fa6ec010af0 received update with 1 addresses
I0913 17:07:25.399861830 17165 subchannel_list.h:361] [pick_first 0x7fa6ec010af0] Creating subchannel list 0x7fa6ec00ed40 for 1 subchannels
I0913 17:07:25.400047329 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 2 -> 3
I0913 17:07:25.400099011 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 3 -> 4
I0913 17:07:25.407755012 17168 timer_generic.cc:534] .. shard[2]: heap_empty=true
I0913 17:07:25.407776244 17168 timer_generic.cc:509] .. shard[2]->queue_deadline_cap --> 1078
I0913 17:07:25.407922352 17168 timer_generic.cc:574] .. shard[2] popped 0
I0913 17:07:25.408062314 17168 timer_generic.cc:629] .. result --> 1, shard[2]->min_deadline 1 --> 1079, now=78
I0913 17:07:25.408211215 17168 timer_generic.cc:534] .. shard[3]: heap_empty=true
I0913 17:07:25.408946504 17168 timer_generic.cc:509] .. shard[3]->queue_deadline_cap --> 1078
I0913 17:07:25.409368625 17168 timer_generic.cc:574] .. shard[3] popped 0
I0913 17:07:25.409511939 17168 timer_generic.cc:629] .. result --> 1, shard[3]->min_deadline 1 --> 1079, now=78
I0913 17:07:25.409652739 17168 timer_generic.cc:739] TIMER CHECK END: r=1; next=1079
I0913 17:07:25.409791304 17168 timer_manager.cc:188] sleep for a 1001 milliseconds
I0913 17:07:25.410554530 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 4 -> 5
I0913 17:07:25.410646720 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 5 -> 6
I0913 17:07:25.410764054 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:92 ref 1 -> 2 connector_arg_copy
I0913 17:07:25.410875800 17165 ref_counted.h:199] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/ext/transport/chttp2/client/secure/secure_channel_create.cc:144 unref 2 -> 1 lb_channel_create
I0913 17:07:25.410993133 17165 ref_counted.h:183] subchannel_pool:0x7fa6ec00b9c8 unref 6 -> 5
I0913 17:07:25.411102086 17165 ref_counted.h:183] subchannel_pool:0x7fa6ec00b9c8 unref 5 -> 4
I0913 17:07:25.411285908 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:92 ref 1 -> 2 connector_arg_copy
I0913 17:07:25.411333959 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 4 -> 5
I0913 17:07:25.411511914 17165 ev_posix.cc:352] (polling-api) pollset_set_create(0x7fa6ec03d640)
I0913 17:07:25.411604943 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 5 -> 6
I0913 17:07:25.411669755 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:92 ref 2 -> 3 connector_arg_copy
I0913 17:07:25.413236714 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:92 ref 3 -> 4 connector_arg_copy
I0913 17:07:25.413338403 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 6 -> 7
I0913 17:07:25.413450429 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 7 -> 8
I0913 17:07:25.413561895 17165 ref_counted.h:183] subchannel_pool:0x7fa6ec00b9c8 unref 8 -> 7
I0913 17:07:25.413666937 17165 ref_counted.h:199] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:86 unref 4 -> 3 connector_arg_destroy
I0913 17:07:25.413773375 17165 ref_counted.h:183] subchannel_pool:0x7fa6ec00b9c8 unref 7 -> 6
I0913 17:07:25.414899775 17165 client_channel.cc:868] chand=0x7fa6ec00fa08: creating subchannel wrapper 0x7fa6ec03da30 for subchannel 0x7fa6ec033280
I0913 17:07:25.414985540 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 2 -> 3 SubchannelWrapper
I0913 17:07:25.415218810 17165 subchannel_list.h:404] [pick_first 0x7fa6ec010af0] subchannel list 0x7fa6ec00ed40 index 0: Created subchannel 0x7fa6ec03da30 for address uri ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/)
I0913 17:07:25.415344245 17165 ref_counted.h:103] lb_policy_refcount:0x7fa6ec010af8 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/lb_policy/pick_first/pick_first.cc:93 ref 1 -> 2 subchannel_list
I0913 17:07:25.415441464 17165 subchannel_list.h:308] [pick_first 0x7fa6ec010af0] subchannel list 0x7fa6ec00ed40 index 0 of 1 (subchannel 0x7fa6ec03da30): starting watch (from IDLE)
I0913 17:07:25.415542035 17165 ref_counted.h:103] pick_first:0x7fa6ec00ed48 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/lb_policy/subchannel_list.h:317 ref 1 -> 2 Watcher
I0913 17:07:25.415705464 17165 ref_counted.h:103] client_channel_routing:0x7fa6ec03da38 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:921 ref 1 -> 2 WatcherWrapper
I0913 17:07:25.443438598 17165 ev_posix.cc:377] (polling-api) pollset_set_add_pollset_set(0x7fa6ec03d640, 0x7fa6ec0130c0)
I0913 17:07:25.443463740 17165 ev_epollex_linux.cc:1487] PSS: merge (0x7fa6ec03d640, 0x7fa6ec0130c0)
I0913 17:07:25.443569620 17165 ev_epollex_linux.cc:1521] PSS: parent 0x7fa6ec03d640 to 0x7fa6ec00fc70
I0913 17:07:25.443969391 17165 socket_utils_common_posix.cc:320] Enabling TCP_USER_TIMEOUT with a timeout of 10000 ms
I0913 17:07:25.445888071 17165 ev_posix.cc:254] (polling-api) fd_create(18, tcp-client:ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/), 1)
I0913 17:07:25.446187550 17165 ev_posix.cc:255] (fd-trace) fd_create(18, tcp-client:ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/), 1)
I0913 17:07:25.446316058 17165 ev_posix.cc:394] (polling-api) pollset_set_add_fd(0x7fa6ec03d640, 18)
I0913 17:07:25.446402382 17165 ev_epollex_linux.cc:1353] PSS:0x7fa6ec03d640: add fd 0x7fa6ec03e720 (18)
I0913 17:07:25.446514687 17165 ev_epollex_linux.cc:616] add fd 0x7fa6ec03e720 (18) to pollable 0x7fa6ec011dc0
I0913 17:07:25.446658560 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 6 -> 7
I0913 17:07:25.446745163 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:92 ref 3 -> 4 connector_arg_copy
I0913 17:07:25.446854395 17165 tcp_client_posix.c[4173222.577355] serial8250: too much work for irq4
c:329] CLIENT_CONNECT: ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/): asynchronously connecting fd 0x7fa6ec03e720
I0913 17:07:25.446979271 17165 timer_generic.cc:364] TIMER 0x7fa6ec012c20: SET 20036 now 36 call 0x7fa6ec012c58[0x7fa70b61f33b]
I0913 17:07:25.447319258 17165 timer_generic.cc:401] .. add to shard 0 with queue_deadline_cap=1078 => is_first_timer=false
I0913 17:07:25.447601138 17165 ref_counted.h:199] resolver_refcount:0x7fa6ec0129d8 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc:396 unref 2 -> 1 dns-resolving
I0913 17:07:25.447646674 17165 work_serializer.cc:102] WorkSerializer::DrainQueue() 0x7fa6ec00bf80
I0913 17:07:25.447753112 17165 work_serializer.cc:117] Queue Drained
I0913 17:07:25.447868211 17165 client_channel.cc:1033] chand=0x7fa6ec00fa08: connectivity change for subchannel wrapper 0x7fa6ec03da30 subchannel 0x7fa6ec033280; hopping into work_serializer
I0913 17:07:25.447979119 17165 work_serializer.cc:55] WorkSerializer::Run() 0x7fa6ec00bf80 Scheduling callback [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:1044]
I0913 17:07:25.448078573 17165 work_serializer.cc:65] Executing immediately
I0913 17:07:25.448185570 17165 client_channel.cc:1066] chand=0x7fa6ec00fa08: processing connectivity change in work serializer for subchannel wrapper 0x7fa6ec03da30 subchannel 0x7fa6ec033280 watcher=0x7fa6ec03ba20
I0913 17:07:25.448334751 17165 subchannel_list.h:241] [pick_first 0x7fa6ec010af0] subchannel list 0x7fa6ec00ed40 index 0 of 1 (subchannel 0x7fa6ec03da30): connectivity changed: state=CONNECTING, shutting_down=0, pending_watcher=0x7fa6ec03ba20
I0913 17:07:25.448437557 17165 ref_counted.h:103] lb_policy_refcount:0x7fa6ec010af8 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/lb_policy/pick_first/pick_first.cc:410 ref 2 -> 3 QueuePicker
I0913 17:07:25.448549582 17165 client_channel.cc:1306] chand=0x7fa6ec00fa08: update: state=CONNECTING picker=0x7fa6ec013ac0
I0913 17:07:25.448671944 17165 ref_counted.h:199] lb_policy_refcount:0x7fa6ec001688 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/lb_policy.h:375 unref 4 -> 3 QueuePicker
I0913 17:07:25.448745697 17165 work_serializer.cc:102] WorkSerializer::DrainQueue() 0x7fa6ec00bf80
I0913 17:07:25.449227881 17165 work_serializer.cc:117] Queue Drained
I0913 17:07:25.449335436 17165 init.cc:165] grpc_init(void)
I0913 17:07:25.449470928 17165 completion_queue.cc:526] grpc_completion_queue_create_internal(completion_type=0, polling_type=0)
I0913 17:07:25.449589100 17165 ev_posix.cc:304] (polling-api) pollset_init(0x7fa6ec03f298)
I0913 17:07:25.449735767 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:796 ref 1 -> 2 pollset
I0913 17:07:25.480508672 17165 channel_connectivity.cc:232] grpc_channel_watch_connectivity_state(channel=0x7fa6ec00f8c0, last_observed_state=0, deadline=gpr_timespec { tv_sec: 1663089445, tv_nsec: 335137682, clock_type: 1 }, cq=0x7fa6ec03f190, tag=0x7fa6ec013a50)
I0913 17:07:25.480579072 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 3 -> 4 watch_channel_connectivity
I0913 17:07:25.480707859 17165 ev_posix.cc:364] (polling-api) pollset_set_add_pollset(0x7fa6ec00fc70, 0x7fa6ec03f298)
I0913 17:07:25.480900901 17165 ev_epollex_linux.cc:1450] PSS:0x7fa6ec00fc70: add pollset 0x7fa6ec03f298
I0913 17:07:25.481010971 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1241 ref 2 -> 3 pollset_as_multipollable
I0913 17:07:25.481119085 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1244 unref 3 -> 2 pollset
I0913 17:07:25.481230272 17165 ev_epollex_linux.cc:569] (fd-trace) Pollable_create: created epfd: 19 (type: 0)
I0913 17:07:25.481345650 17165 ev_epollex_linux.cc:1249] PS:0x7fa6ec03f298 active pollable transition from empty to multi
I0913 17:07:25.481428063 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec03f638 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1280 ref 1 -> 2 pollset_set
I0913 17:07:25.481534501 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1281 unref 2 -> 1 pollset_as_multipollable
I0913 17:07:25.481642056 17165 ev_epollex_linux.cc:616] add fd 0x7fa6ec03e720 (18) to pollable 0x7fa6ec03f630
I0913 17:07:25.481755758 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec03f638 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1478 unref 2 -> 1 pollset_set
I0913 17:07:25.481859402 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 4 -> 5 ExternalConnectivityWatcher
I0913 17:07:25.482000482 17165 work_serializer.cc:55] WorkSerializer::Run() 0x7fa6ec00bf80 Scheduling callback [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:1155]
I0913 17:07:25.482080380 17165 work_serializer.cc:65] Executing immediately
I0913 17:07:25.482211402 17165 timer_generic.cc:364] TIMER 0x7fa6ec03f418: SET 600034 now 180 call 0x7fa6ec03f388[0x7fa70b482e3b]
I0913 17:07:25.482299402 17165 timer_generic.cc:401] .. add to shard 0 with queue_deadline_cap=1078 => is_first_timer=false
I0913 17:07:25.482430983 17165 connectivity_state.cc:118] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: add watcher 0x7fa6ec03f130
I0913 17:07:25.482516190 17165 connectivity_state.cc:124] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: notifying watcher 0x7fa6ec03f130: IDLE -> CONNECTING
I0913 17:07:25.482647771 17165 work_serializer.cc:55] WorkSerializer::Run() 0x7fa6ec00bf80 Scheduling callback [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:1175]
I0913 17:07:25.482732139 17165 work_serializer.cc:76] Scheduling on queue : item 0x7fa6ec03ec80
I0913 17:07:25.482900317 17165 work_serializer.cc:102] WorkSerializer::DrainQueue() 0x7fa6ec00bf80
I0913 17:07:25.482950044 17165 work_serializer.cc:134] Running item 0x7fa6ec03ec80 : callback scheduled at [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:1175]
I0913 17:07:25.483054526 17165 connectivity_state.cc:141] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: remove watcher 0x7fa6ec03f130
I0913 17:07:25.483184711 17165 ev_posix.cc:371] (polling-api) pollset_set_del_pollset(0x7fa6ec00fc70, 0x7fa6ec03f298)
I0913 17:07:25.483269917 17165 ev_epollex_linux.cc:1398] PSS:0x7fa6ec00fc70: del pollset 0x7fa6ec03f298
I0913 17:07:25.483379707 17165 ev_epollex_linux.cc:664] PS:0x7fa6ec03f298 (pollable:0x7fa6ec03f630) maybe_finish_shutdown sc=(nil) (target:!NULL) rw=(nil) (target:NULL) cpsc=0 (target:0)
I0913 17:07:25.511424054 17165 ref_counted.h:199] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:105 unref 5 -> 4 ExternalConnectivityWatcher
I0913 17:07:25.511541387 17165 work_serializer.cc:102] WorkSerializer::DrainQueue() 0x7fa6ec00bf80
I0913 17:07:25.511615978 17165 work_serializer.cc:117] Queue Drained
I0913 17:07:25.511634695 17165 timer_generic.cc:467] TIMER 0x7fa6ec03f418: CANCEL pending=true
I0913 17:07:25.511748397 17165 completion_queue.cc:707] cq_end_op_for_next(cq=0x7fa6ec03f190, tag=0x7fa6ec013a50, error="No Error", done=0x7fa70b482a2e, done_arg=0x7fa6ec03f310, storage=0x7fa6ec03f460)
I0913 17:07:25.511850644 17165 ev_posix.cc:331] (polling-api) pollset_kick(0x7fa6ec03f298, (nil))
I0913 17:07:25.511959597 1[4173222.646290] serial8250: too much work for irq4
7165 ev_epollex_linux.cc:731] PS:0x7fa6ec03f298 kick (nil) tls_pollset=(nil) tls_worker=(nil) pollset.root_worker=(nil)
I0913 17:07:25.512064638 17165 ev_epollex_linux.cc:741] PS:0x7fa6ec03f298 kicked_any_without_poller
I0913 17:07:25.512201248 17165 completion_queue.cc:982] grpc_completion_queue_next(cq=0x7fa6ec03f190, deadline=gpr_timespec { tv_sec: 9223372036854775807, tv_nsec: 0, clock_type: 1 }, reserved=(nil))
I0913 17:07:25.512371381 17165 ref_counted.h:199] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:105 unref 4 -> 3 watch_channel_connectivity
I0913 17:07:25.512474746 17165 completion_queue.cc:1083] RETURN_EVENT[0x7fa6ec03f190]: OP_COMPLETE: tag:0x7fa6ec013a50 OK
I0913 17:07:25.512537882 17165 completion_queue.cc:1425] grpc_completion_queue_destroy(cq=0x7fa6ec03f190)
I0913 17:07:25.512637895 17165 completion_queue.cc:1419] grpc_completion_queue_shutdown(cq=0x7fa6ec03f190)
I0913 17:07:25.512750200 17165 ev_posix.cc:309] (polling-api) pollset_shutdown(0x7fa6ec03f298)
I0913 17:07:25.512975648 17165 ev_epollex_linux.cc:664] PS:0x7fa6ec03f298 (pollable:0x7fa6ec03f630) maybe_finish_shutdown sc=0x7fa6ec03f1d0 (target:!NULL) rw=(nil) (target:NULL) cpsc=0 (target:0)
I0913 17:07:25.513091305 17165 ev_posix.cc:314] (polling-api) pollset_destroy(0x7fa6ec03f298)
I0913 17:07:25.513200816 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec03f638 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:927 unref 1 -> 0 pollset
I0913 17:07:25.513311165 17165 ev_epollex_linux.cc:149] (fd-trace) pollable_unref: Closing epfd: 19
I0913 17:07:25.513445540 17165 init.cc:212] grpc_shutdown(void)
I0913 17:07:25.513544994 17165 channel_connectivity.cc:43] grpc_channel_check_connectivity_state(channel=0x7fa6ec00f8c0, try_to_connect=1)
I0913 17:07:25.513658695 17165 connectivity_state.cc:174] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: get current state: CONNECTING
I0913 17:07:25.513767368 17165 init.cc:165] grpc_init(void)
I0913 17:07:25.513882746 17165 completion_queue.cc:526] grpc_completion_queue_create_internal(completion_type=0, polling_type=0)
I0913 17:07:25.513992257 17165 ev_posix.cc:304] (polling-api) pollset_init(0x7fa6ec03f738)
I0913 17:07:25.514111267 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:796 ref 1 -> 2 pollset
I0913 17:07:25.514229997 17165 channel_connectivity.cc:232] grpc_channel_watch_connectivity_state(channel=0x7fa6ec00f8c0, last_observed_state=1, deadline=gpr_timespec { tv_sec: 1663089445, tv_nsec: 335137682, clock_type: 1 }, cq=0x7fa6ec03f630, tag=0x7fa6ec013b20)
I0913 17:07:25.514340067 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 3 -> 4 watch_channel_connectivity
I0913 17:07:25.514452930 17165 ev_posix.cc:364] (polling-api) pollset_set_add_pollset(0x7fa6ec00fc70, 0x7fa6ec03f738)
I0913 17:07:25.514564397 17165 ev_epollex_linux.cc:1450] PSS:0x7fa6ec00fc70: add pollset 0x7fa6ec03f738
I0913 17:07:25.514675584 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1241 ref 2 -> 3 pollset_as_multipollable
I0913 17:07:25.542637797 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1244 unref 3 -> 2 pollset
I0913 17:07:25.542729429 17165 ev_epollex_linux.cc:569] (fd-trace) Pollable_create: created epfd: 19 (type: 0)
I0913 17:07:25.542869950 17165 ev_epollex_linux.cc:1249] PS:0x7fa6ec03f738 active pollable transition from empty to multi
I0913 17:07:25.542934483 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec03fdd8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1280 ref 1 -> 2 pollset_set
I0913 17:07:25.543148477 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1281 unref 2 -> 1 pollset_as_multipollable
I0913 17:07:25.543274750 17165 ev_epollex_linux.cc:616] add fd 0x7fa6ec03e720 (18) to pollable 0x7fa6ec03fdd0
I0913 17:07:25.543391245 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec03fdd8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1478 unref 2 -> 1 pollset_set
I0913 17:07:25.543491817 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 4 -> 5 ExternalConnectivityWatcher
I0913 17:07:25.543632617 17165 work_serializer.cc:55] WorkSerializer::Run() 0x7fa6ec00bf80 Scheduling callback [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:1155]
I0913 17:07:25.543722572 17165 work_serializer.cc:65] Executing immediately
I0913 17:07:25.543974280 17165 timer_generic.cc:364] TIMER 0x7fa6ec03f418: SET 600034 now 242 call 0x7fa6ec03f388[0x7fa70b482e3b]
I0913 17:07:25.544049988 17165 timer_generic.cc:401] .. add to shard 0 with queue_deadline_cap=1078 => is_first_timer=false
I0913 17:07:25.544172909 17165 connectivity_state.cc:118] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: add watcher 0x7fa6ec03f5d0
I0913 17:07:25.544299461 17165 work_serializer.cc:102] WorkSerializer::DrainQueue() 0x7fa6ec00bf80
I0913 17:07:25.544386902 17165 work_serializer.cc:117] Queue Drained
I0913 17:07:25.544504236 17165 completion_queue.cc:982] grpc_completion_queue_next(cq=0x7fa6ec03f630, deadline=gpr_timespec { tv_sec: 9223372036854775807, tv_nsec: 0, clock_type: 1 }, reserved=(nil))
I0913 17:07:25.544651741 17165 ev_posix.cc:322] (polling-api) pollset_work(0x7fa6ec03f738, 9223372036854775807) begin
I0913 17:07:25.544765442 17165 ev_epollex_linux.cc:1120] PS:0x7fa6ec03f738 work hdl=(nil) worker=0x7fa7057602f0 now=243 deadline=9223372036854775807 kwp=0 pollable=0x7fa6ec03fdd0
I0913 17:07:25.544948147 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec03fdd8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1018 ref 1 -> 2 pollset_worker
I0913 17:07:25.545075537 17165 ev_epollex_linux.cc:938] POLLABLE:0x7fa6ec03fdd0[type=pollset epfd=19 wakeup=20] poll for -1ms
I0913 17:07:25.571483922 17165 ev_epollex_linux.cc:957] POLLABLE:0x7fa6ec03fdd0 got 1 events
I0913 17:07:25.571542309 17165 ev_epollex_linux.cc:905] PS:0x7fa6ec03f738 got fd 0x7fa6ec03e720: cancel=0 read=0 write=1
I0913 17:07:25.571681153 17165 tcp_client_posix.cc:142] CLIENT_CONNECT: ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/): on_writable: error="No Error"
I0913 17:07:25.571790944 17165 timer_generic.cc:467] TIMER 0x7fa6ec012c20: CANCEL pending=true
I0913 17:07:25.571914423 17165 ev_posix.cc:400] (polling-api) pollset_set_del_fd(0x7fa6ec03d640, 18)
I0913 17:07:25.572013319 17165 ev_epollex_linux.cc:1377] PSS:0x7fa6ec03d640: del fd 0x7fa6ec03e720
I0913 17:07:25.572390182 17165 combiner.cc:61] C:0x7fa6ec03f4f0 create
I0913 17:07:25.572544950 17165 ref_counted.h:103] tcp:0x7fa6ec0403d8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/tcp_posix.cc:1823 ref 1 -> 2 error-tracking
I0913 17:07:25.572643566 17165 tcp_client_posix.cc:105] CLIENT_CONNECT: ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/): on_alarm: error="Cancelled"
I0913 17:07:25.572758665 17165 ref_counted.h:183] subchannel_pool:0x7fa6ec00b9c8 unref 7 -> 6
I0913 17:07:25.572988862 17165 ref_counted.h:199] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:86 unref 4 -> 3 connector_arg_destroy
I0913 17:07:25.573173801 17165 handshaker.cc:99] handshake_manager 0x7fa6ec038720: adding handshaker http_connect [0x7fa6ec040f40] at index 0
I0913 17:07:25.573545916 17165 ssl_transport_security.cc:220] HANDSHAKE START - before SSL initialization - PINIT
I0913 17:07:25.573650398 17165 ssl_transport_security.cc:220] LOOP - before SSL initialization - PINIT
I0913 17:07:25.573789243 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS write client hello - TWCH
I0913 17:07:25.573852379 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/transport/security_handshaker.cc:113 ref 3 -> 4 handshake
I0913 17:07:25.573949878 17165 handshaker.cc:99] handshake_manager 0x7fa6ec038720: adding handshaker security [0x7fa6ec05c6e0] at index 1
I0913 17:07:25.574059668 17165 ev_posix.cc:394] (polling-api) pollset_set_add_fd(0x7fa6ec03d640, 18)
I0913 17:07:25.574176722 17165 ev_epollex_linux.cc:1353] PSS:0x7fa6ec03d640: add fd 0x7fa6ec03e720 (18)
I0913 17:07:25.574290144 17165 ev_epollex_linux.cc:616] add fd 0x7fa6ec03e720 (18) to pollable 0x7fa6ec011dc0
I0913 17:07:25.574407198 17165 ev_epollex_linux.cc:616] add fd 0x7fa6ec03e720 (18) to pollable 0x7fa6ec03fdd0
I0913 17:07:25.574524252 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 6 -> 7
I0913 17:07:25.574623148 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:92 ref 4 -> 5 connector_arg_copy
I0913 17:07:25.574756125 17165 timer_generic.cc:364] TIMER 0x7fa6ec0387d8: SET 20036 now 273 call 0x7fa6ec038810[0x7fa70b5f12d2]
I0913 17:07:25.574843287 17165 timer_generic.cc:401] .. add to shard 0 with queue_deadline_cap=1078 => is_first_timer=false
I0913 17:07:25.575114551 17165 handshaker.cc:129] handshake_manager 0x7fa6ec038720: error="No Error" shutdown=0 index=0, args={endpoint=0x7fa6ec0403b0, args=0x7fa6ec012830 {size=12: grpc.primary_user_agent=grpc-c++/1.30.0, grpc.keepalive_time_ms=10000, grpc.keepalive_timeout_ms=10000, grpc.ssl_target_name_override=Infinera.com, grpc.client_channel_factory=0x7fa6ec00e9f0, grpc.channel_credentials=0x7fa6ec010d80, grpc.server_uri=dns:///[10.220.6.50:50058](http://10.220.6.50:50058/), grpc.default_authority=Infinera.com, grpc.subchannel_pool=0x7fa6ec00b9c0, grpc.http2_scheme=https, grpc.security_connector=0x7fa6ec013030, grpc.subchannel_address=ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/)}, read_buffer=0x7fa6ec012bf0 (length=0), exit_early=0}
I0913 17:07:25.575178246 17165 handshaker.cc:176] handshake_manager 0x7fa6ec038720: calling handshaker http_connect [0x7fa6ec040f40] at index 0
I0913 17:07:25.575312900 17165 handshaker.cc:129] handshake_manager 0x7fa6ec038720: error="No Error" shutdown=0 index=1, args={endpoint=0x7fa6ec0403b0, args=0x7fa6ec012830 {size=12: grpc.primary_user_agent=grpc-c++/1.30.0, grpc.keepalive_time_ms=10000, grpc.keepalive_timeout_ms=10000, grpc.ssl_target_name_override=Infinera.com, grpc.client_channel_factory=0x7fa6ec00e9f0, grpc.channel_credentials=0x7fa6ec010d80, grpc.server_uri=dns:///[10.220.6.50:50058](http://10.220.6.50:50058/), grpc.default_authority=Infinera.com, grpc.subchannel_pool=0x7fa6ec00b9c0, grpc.http2_scheme=https, grpc.security_connector=0x7fa6ec013030, grpc.subchannel_address=ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/)}, read_buffer=0x7fa6ec012bf0 (length=0), exit_early=0}
I0913 17:07:25.603279024 17165 handshaker.cc:176] handshake_manager 0x7fa6ec038720: calling handshaker security [0x7fa6ec05c6e0] at index 1
I0913 17:07:25.603416192 17165 tcp_posix.cc:1566] WRITE 0x7fa6ec0403b0 (peer=ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/))
I0913 17:07:25.603619012 17165 tcp_posix.cc:1616] write: "No Error"
I0913 17:07:25.603753386 17165 ref_counted.h:103] tcp:0x7fa6ec0403d8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/tcp_posix.cc:908 ref 2 -> 3 read
I0913 17:07:25.604010961 17165 tcp_posix.cc:541] TCP:0x7fa6ec0403b0 notify_on_read
I0913 17:07:25.604401513 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec03fdd8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1085 unref 2 -> 1 pollset_worker
I0913 17:07:25.604615228 17165 ev_epollex_linux.cc:664] PS:0x7fa6ec03f738 (pollable:0x7fa6ec03fdd0) maybe_finish_shutdown sc=(nil) (target:!NULL) rw=(nil) (target:NULL) cpsc=1 (target:0)
I0913 17:07:25.604730047 17165 ev_posix.cc:325] (polling-api) pollset_work(0x7fa6ec03f738, 9223372036854775807) end
I0913 17:07:25.604973653 17165 ev_posix.cc:322] (polling-api) pollset_work(0x7fa6ec03f738, 9223372036854775807) begin
I0913 17:07:25.605089869 17165 ev_epollex_linux.cc:1120] PS:0x7fa6ec03f738 work hdl=(nil) worker=0x7fa7057602f0 now=273 deadline=9223372036854775807 kwp=0 pollable=0x7fa6ec03fdd0
I0913 17:07:25.605204967 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec03fdd8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1018 ref 1 -> 2 pollset_worker
I0913 17:07:25.605318669 17165 ev_epollex_linux.cc:938] POLLABLE:0x7fa6ec03fdd0[type=pollset epfd=19 wakeup=20] poll for -1ms
I0913 17:07:25.624281973 17165 ev_epollex_linux.cc:957] POLLABLE:0x7fa6ec03fdd0 got 1 events
I0913 17:07:25.625191306 17165 ev_epollex_linux.cc:905] PS:0x7fa6ec03f738 got fd 0x7fa6ec03e720: cancel=0 read=1 write=1
I0913 17:07:25.626231103 17165 tcp_posix.cc:887] TCP:0x7fa6ec0403b0 got_read: "No Error"
I0913 17:07:25.627087357 17165 tcp_posix.cc:869] TCP:0x7fa6ec0403b0 alloc_slices
I0913 17:07:25.627883827 17165 resource_quota.cc:890] RQ anonymous_pool_7fa6ec03f7b0 ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/): alloc 8192; free_pool -> -8192
I0913 17:07:25.629396031 17165 combiner.cc:135] C:0x7fa6ec03f4f0 grpc_combiner_execute c=0x7fa6ec040cf8 last=1
I0913 17:07:25.630779167 17165 combiner.cc:199] C:0x7fa6ec03f4f0 grpc_combiner_continue_exec_ctx contended=0 exec_ctx_ready_to_finish=0 time_to_execute_final_list=0
I0913 17:07:25.633678419 17165 combiner.cc:222] C:0x7fa6ec03f4f0 maybe_finish_one n=0x7fa6ec040cf8
I0913 17:07:25.635280857 17165 combiner.cc:308] C:0x7fa6ec03f4f0 grpc_combiner_execute_finally c=0x7fa6ec03f820; ac=0x7fa6ec03f4f0
I0913 17:07:25.636448324 17165 combiner.cc:264] C:0x7fa6ec03f4f0 finish old_state=5
I0913 17:07:25.637369670 17165 combiner.cc:199] C:0x7fa6ec03f4f0 grpc_combiner_continue_exec_ctx contended=0 exec_ctx_ready_to_finish=0 time_to_execute_final_list=1
I0913 17:07:25.638760349 17165 combiner.cc:246] C:0x7fa6ec03f4f0 execute_final[0] c=0x7fa6ec03f820
I0913 17:07:25.639790648 17165 resource_quota.cc:321] RQ: check allocation for user 0x7fa6ec040cf0 shutdown=0 free_pool=-8192
I0913 17:07:25.640919004 17165 resource_quota.cc:347] RQ anonymous_pool_7fa6ec03f7b0 ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/): grant alloc 8192 bytes; rq_free_pool -> 9223372036854767615
I0913 17:07:25.642398521 17165 combiner.cc:264] C:0x7fa6ec03f4f0 finish old_state=3
I0913 17:07:25.643722712 17165 tcp_posix.cc:850] TCP:0x7fa6ec0403b0 read_allocation_done: "No Error"
I0913 17:07:25.644955550 17165 tcp_posix.cc:680] TCP:0x7fa6ec0403b0 call_cb 0x7fa6ec05c8d8 0x7fa70b65e450:0x7fa6ec05c6e0
I0913 17:07:25.646910547 17165 tcp_posix.cc:683] READ 0x7fa6ec0403b0 (peer=ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/)) error="No Error"
I0913 17:07:25.648767487 17165 ref_counted.h:199] tcp:0x7fa6ec0403d8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/tcp_posix.cc:844 unref 3 -> 2 read
I0913 17:07:25.650205100 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS write client hello - TWCH
I0913 17:07:25.652048630 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read server hello - TRSH
I0913 17:07:25.653957812 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read server certific - TRSC
I0913 17:07:25.655300999 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read server key exch - TRSKE
I0913 17:07:25.656334650 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read server done - TRSD
I0913 17:07:25.659340898 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS write client key exc - TWCKE
I0913 17:07:25.660446066 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS write change cipher - TWCCS
I0913 17:07:25.661612974 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS write finished - TWFIN
I0913 17:07:25.662694117 17165 tcp_posix.cc:1566] WRITE 0x7fa6ec0403b0 (peer=ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/))
I0913 17:07:25.664612238 17165 tcp_posix.cc:1616] write: "No Error"
I0913 17:07:25.665905140 17165 ref_counted.h:103] tcp:0x7fa6ec0403d8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/tcp_posix.cc:908 ref 2 -> 3 read
I0913 17:07:25.667535235 17165 tcp_posix.cc:887] TCP:0x7fa6ec0403b0 got_read: "No Error"
I0913 17:07:25.668567210 17165 tcp_posix.cc:879] TCP:0x7fa6ec0403b0 do_read
I0913 17:07:25.669521521 17165 tcp_posix.cc:680] TCP:0x7fa6ec0403b0 call_cb 0x7fa6ec05c8d8 0x7fa70b65e450:0x7fa6ec05c6e0
I0913 17:07:25.670683960 17165 tcp_posix.cc:683] READ 0x7fa6ec0403b0 (peer=ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/)) error="No Error"
I0913 17:07:25.671891934 17165 ref_counted.h:199] tcp:0x7fa6ec0403d8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/tcp_posix.cc:844 unref 3 -> 2 read
I0913 17:07:25.673510017 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS write finished - TWFIN
I0913 17:07:25.674696760 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read server session - TRST
I0913 17:07:25.675933230 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read change cipher s - TRCCS
I0913 17:07:25.677194564 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read finished - TRFIN
I0913 17:07:25.678290792 17165 ssl_transport_security.cc:220] HANDSHAKE DONE - SSL negotiation finished succe - SSLOK
E0913 17:07:25.680856761 17165 ssl_transport_security.cc:470] assertion failed: (int)peer->property_count == current_insert_index
signal 6[SIGABRT] caught by process [/opt/infinera/thanos/local/bin/tunnelclient]
#####DumpStacktrace ########
#0 /opt/infinera/thanos/local/lib/libBaseSystem.so unsigned long backward::details::unwind<backward::StackTraceImpl<backward::system_tag::linux_tag>::callback>(backward::StackTraceImpl<backward::system_tag::linux_tag>::callback, unsigned long) [0x7fa709dd4c72]
#1 /opt/infinera/thanos/local/lib/libBaseSystem.so backward::StackTraceImpl<backward::system_tag::linux_tag>::load_here(unsigned long) [0x7fa709dd4008]
#2 /opt/infinera/thanos/local/lib/libBaseSystem.so util::Backtrace::DumpStacktrace() [0x7fa709dd2c5e]
#3 /opt/infinera/thanos/local/lib/libBaseSystem.so OsProcess::MyFailureHandler(int) [0x7fa709db54e7]
#4 /lib/x86_64-linux-gnu/libc.so.6 [0x7fa70988683f]
#5 /lib/x86_64-linux-gnu/libc.so.6 gsignal [0x7fa7098867bb]
#6 /lib/x86_64-linux-gnu/libc.so.6 abort [0x7fa709871534]
#7 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b6a1dfd]
#8 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b6a3acf]
#9 /opt/infinera/thanos/local/lib/libgrpc.so.10 tsi_handshaker_result_extract_peer(tsi_handshaker_result const*, tsi_peer*) [0x7fa70b6a6db1]
#10 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b65df19]
#11 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b65e2b2]
#12 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b65e44b]
#13 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b65e5fe]
#14 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b6117ac]
#15 /opt/infinera/thanos/local/lib/libgrpc.so.10 grpc_core::ExecCtx::Flush() [0x7fa70b611cdc]
#16 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b6098b7]
#17 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b6110bd]
#18 /opt/infinera/thanos/local/lib/libgrpc.so.10 grpc_pollset_work(grpc_pollset*, grpc_pollset_worker**, long) [0x7fa70b61859f]
#19 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b674c8f]
#20 /opt/infinera/thanos/local/lib/libgrpc.so.10 grpc_completion_queue_next [0x7fa70b675154]
#21 /opt/infinera/thanos/local/lib/libgrpc++.so.1 grpc_impl::CompletionQueue::AsyncNextInternal(void**, bool*, gpr_timespec) [0x7fa70b09a675]
#22 /opt/infinera/thanos/local/lib/libgrpc++.so.1 grpc_impl::CompletionQueue::Next(void**, bool*) [0x7fa70b088ff6]
#23 /opt/infinera/thanos/local/lib/libgrpc++.so.1 grpc_impl::Channel::WaitForStateChangeImpl(grpc_connectivity_state, gpr_timespec) [0x7fa70b0883b4]
#24 /opt/infinera/thanos/local/bin/tunnelclient bool grpc::ChannelInterface::WaitForStateChange<gpr_timespec>(grpc_connectivity_state, gpr_timespec) [0x55d52d47a6e3]
#25 /opt/infinera/thanos/local/bin/tunnelclient bool grpc::ChannelInterface::WaitForConnected<gpr_timespec>(gpr_timespec) [0x55d52d4797fe]
#26 /opt/infinera/thanos/local/bin/tunnelclient ConnectionManager::createChannel(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::shared_ptr<grpc_impl::Channel>&) [0x55d52d472e6f]
#27 /opt/infinera/thanos/local/bin/tunnelclient ConnectionManager::connectToServer() [0x55d52d473b81]
#28 /opt/infinera/thanos/local/bin/tunnelclient boost::_mfi::mf0<void, ConnectionManager>::operator()(ConnectionManager*) const [0x55d52d47d4bc]
#29 /opt/infinera/thanos/local/bin/tunnelclient void boost::_bi::list1<boost::_bi::value<ConnectionManager*> >::operator()<boost::_mfi::mf0<void, ConnectionManager>, boost::_bi::list0>(boost::_bi::type<void>, boost::_mfi::mf0<void, ConnectionManager>&, boost::_bi::list0&, int) [0x55d52d47d41e]
#30 /opt/infinera/thanos/local/bin/tunnelclient boost::_bi::bind_t<void, boost::_mfi::mf0<void, ConnectionManager>, boost::_bi::list1<boost::_bi::value<ConnectionManager*> > >::operator()() [0x55d52d47d2cc]
#31 /opt/infinera/thanos/local/bin/tunnelclient boost::detail::thread_data<boost::_bi::bind_t<void, boost::_mfi::mf0<void, ConnectionManager>, boost::_bi::list1<boost::_bi::value<ConnectionManager*> > > >::run() [0x55d52d47d0f3]
#####End DumpStacktrace ########
**Ideally it must work fine or through some error. But here it is getting crashed.**
**Below is the code snippets.**
std::shared_ptr<grpc::ChannelCredentials> ConnectionManager::createChannelCredentials()
{
grpc::SslCredentialsOptions sslCredOptions;
std::string rootCertPath("/var/security/x509/trustedCerts/RSA_2048_SHA_512_root.crt");
std::string clientKeyPath("/var/security/x509/keys/RSA_2048_SHA_512_client.key");
std::string clientCertPath("/var/security/x509/certs/RSA_2048_SHA_512_client.crt");
//Here caCert, clientCert, ClientKey is of type string
getFileData(rootCertPath, caCert);//getFileData() will read and populate cert/key from the provided loc
sslCredOptions.pem_root_certs = caCert;
getFileData(clientKeyPath, clientKey);
sslCredOptions.pem_private_key = clientKey;
getFileData(clientCertPath, clientCert);
sslCredOptions.pem_cert_chain = clientCert;
return grpc::SslCredentials(sslCredOptions);
}
//Here server
std::shared_ptr<grpc::Channel> ConnectionManager::createChannel(const std::string &serverAddress, std::shared_ptr<grpc::Channel>& channel_)
{
std::shared_ptr<grpc::ChannelCredentials> channelCreds = ::grpc::InsecureChannelCredentials();
channelCreds = createChannelCredentials();
ChannelArguments channelArg;
std::string san("<CertSANName>");
channelArg.SetInt(GRPC_ARG_KEEPALIVE_TIME_MS, keepAliveTime_);
channelArg.SetInt(GRPC_ARG_KEEPALIVE_TIMEOUT_MS, keepAliveTime_);
channelArg.SetSslTargetNameOverride(san);
channel_ = grpc::CreateCustomChannel(serverAddress_, channelCreds, channelArg);
std::chrono::time_point<std::chrono::system_clock> deadline = std::chrono::system_clock::now() + std::chrono::seconds(10);
//if(!channel_->WaitForConnected(gpr_time_add(gpr_now(GPR_CLOCK_REALTIME), gpr_time_from_seconds(600, GPR_TIMESPAN))) || channel_->GetState(false) != grpc_connectivity_state::GRPC_CHANNEL_READY)
if(!channel_->WaitForConnected(deadline) || channel_->GetState(false) != grpc_connectivity_state::GRPC_CHANNEL_READY)//It is crashing here
{
//notify and start some backoff logic
}
else
{
//Secure connection established
....
....
....
}
return channel_;
}
**Tested the same with below openssl utility and I can see in the pcap that handshake is happening successfully.**
"sudo openssl s_client -connect <ip>:<port>-cert /var/security/x509/certs/RSA_2048_SHA_512_client.crt -key /var/security/x509/keys/RSA_2048_SHA_512_client.key -CAfile /var/security/x509/trusted_certs/RSA_2048_SHA_512_root.crt -state -debug"
|
1.0
|
[c++] channel.WaitForConnected assertion failed while trying for mTLS connection - **Using gRPC with C++ Language
Linux version - Linux 4.19.81-OpenNetworkLinux #1 SMP Mon Jun 20 01:56:19 UTC 2022 x86_64 GNU/Linux
Python 2.7.16
gcc (Debian 8.3.0-6) 8.3.0**
**gRPC secure communication:
Created a custom channel and populated SslCredentials with ca cert, client cert and client key. Below assertion is observed once channel.WaitForConnected is called.**
I0913 17:07:25.303815824 17165 init.cc:165] grpc_init(void)
I0913 17:07:25.303940980 17165 init.cc:212] grpc_shutdown(void)
I0913 17:07:25.304056078 17165 init.cc:212] grpc_shutdown(void)
I0913 17:07:25.304331812 17165 init.cc:165] grpc_init(void)
I0913 17:07:25.304439088 17165 secure_channel_create.cc:201] grpc_secure_channel_create(creds=0x7fa6ec010d80, target=[10.220.6.50:50058](http://10.220.6.50:50058/), args=0x7fa705760590, reserved=(nil))
I0913 17:07:25.304579888 17165 init.cc:165] grpc_init(void)
I0913 17:07:25.305033297 17165 ev_posix.cc:352] (polling-api) pollset_set_create(0x7fa6ec00fc70)
I0913 17:07:25.305103697 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 1 -> 2
I0913 17:07:25.305229970 17165 client_channel.cc:1402] chand=0x7fa6ec00fa08: creating client_channel for channel stack 0x7fa6ec00f958
I0913 17:07:25.305320485 17165 ev_posix.cc:304] (polling-api) pollset_init(0x7fa6ec011c70)
I0913 17:07:25.305436421 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:796 ref 1 -> 2 pollset
I0913 17:07:25.305555151 17165 timer_generic.cc:364] TIMER 0x7fa6ec00e4d0: SET 5003 now 3 call 0x7fa6ec00e508[0x7fa70b4823cd]
I0913 17:07:25.305665221 17165 timer_generic.cc:401] .. add to shard 0 with queue_deadline_cap=0 => is_first_timer=false
I0913 17:07:25.305775850 17165 ev_posix.cc:364] (polling-api) pollset_set_add_pollset(0x7fa6ec00fc70, 0x7fa6ec011c70)
I0913 17:07:25.305887875 17165 ev_epollex_linux.cc:1450] PSS:0x7fa6ec00fc70: add pollset 0x7fa6ec011c70
I0913 17:07:25.306001577 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1241 ref 2 -> 3 pollset_as_multipollable
I0913 17:07:25.306742453 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1244 unref 3 -> 2 pollset
I0913 17:07:25.306846936 17165 ev_epollex_linux.cc:569] (fd-trace) Pollable_create: created epfd: 16 (type: 0)
I0913 17:07:25.306956168 17165 ev_epollex_linux.cc:1249] PS:0x7fa6ec011c70 active pollable transition from empty to multi
I0913 17:07:25.307059253 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec011dc8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1280 ref 1 -> 2 pollset_set
I0913 17:07:25.307174631 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1281 unref 2 -> 1 pollset_as_multipollable
I0913 17:07:25.307288891 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec011dc8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1478 unref 2 -> 1 pollset_set
I0913 17:07:25.307496739 17165 init.cc:165] grpc_init(void)
I0913 17:07:25.334856920 17165 init.cc:212] grpc_shutdown(void)
I0913 17:07:25.335164222 17165 channel_connectivity.cc:43] grpc_channel_check_connectivity_state(channel=0x7fa6ec00f8c0, try_to_connect=1)
I0913 17:07:25.335212831 17165 connectivity_state.cc:174] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: get current state: IDLE
I0913 17:07:25.335308933 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 1 -> 2 TryToConnect
I0913 17:07:25.335442749 17165 work_serializer.cc:55] WorkSerializer::Run() 0x7fa6ec00bf80 Scheduling callback [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:1926]
I0913 17:07:25.335517898 17165 work_serializer.cc:65] Executing immediately
I0913 17:07:25.335660933 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 2 -> 3 ClientChannelControlHelper
I0913 17:07:25.335822965 17165 ev_posix.cc:352] (polling-api) pollset_set_create(0x7fa6ec0105f0)
I0913 17:07:25.335905657 17165 ref_counted.h:92] lb_policy_refcount:0x7fa6ec001688 ref 1 -> 2
I0913 17:07:25.336194241 17165 ev_posix.cc:352] (polling-api) pollset_set_create(0x7fa6ec010670)
I0913 17:07:25.336355155 17165 ev_posix.cc:377] (polling-api) pollset_set_add_pollset_set(0x7fa6ec010670, 0x7fa6ec0105f0)
I0913 17:07:25.336457682 17165 ev_epollex_linux.cc:1487] PSS: merge (0x7fa6ec010670, 0x7fa6ec0105f0)
I0913 17:07:25.307856561 17166 executor.cc:224] EXECUTOR (default-executor) [0]: step (sub_depth=0)
I0913 17:07:25.307920536 17167 executor.cc:224] EXECUTOR (resolver-executor) [0]: step (sub_depth=0)
I0913 17:07:25.336576133 17165 ev_epollex_linux.cc:1521] PSS: parent 0x7fa6ec010670 to 0x7fa6ec0105f0
I0913 17:07:25.337203867 17165 resolving_lb_policy.cc:163] resolving_lb=0x7fa6ec001680: starting name resolution
I0913 17:07:25.337258063 17165 ref_counted.h:92] lb_policy_refcount:0x7fa6ec001688 ref 2 -> 3
I0913 17:07:25.337429314 17165 client_channel.cc:1306] chand=0x7fa6ec00fa08: update: state=CONNECTING picker=0x7fa6ec0127e0
I0913 17:07:25.337497200 17165 connectivity_state.cc:152] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: IDLE -> CONNECTING (helper)
I0913 17:07:25.337667613 17165 ref_counted.h:103] resolver_refcount:0x7fa6ec0129d8 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc:434 ref 1 -> 2 dns-resolving
I0913 17:07:25.338277467 17165 grpc_ares_wrapper.cc:128] (c-ares resolver) request:0x7fa6ec012b80 c-ares address sorting: input[0]=[10.220.6.50:50058](http://10.220.6.50:50058/)
I0913 17:07:25.338486711 17165 grpc_ares_wrapper.cc:128] (c-ares resolver) request:0x7fa6ec012b80 c-ares address sorting: output[0]=[10.220.6.50:50058](http://10.220.6.50:50058/)
I0913 17:07:25.338643994 17165 ev_posix.cc:377] (polling-api) pollset_set_add_pollset_set(0x7fa6ec0105f0, 0x7fa6ec00fc70)
I0913 17:07:25.338706851 17165 ev_epollex_linux.cc:1487] PSS: merge (0x7fa6ec0105f0, 0x7fa6ec00fc70)
I0913 17:07:25.338813848 17165 ev_epollex_linux.cc:1521] PSS: parent 0x7fa6ec0105f0 to 0x7fa6ec00fc70
I0913 17:07:25.339155232 17165 client_channel.cc:1581] chand=0x7fa6ec00fa08: creat[4173222.493507] serial8250: too much work for irq4
ed resolving_lb_policy=0x7fa6ec001680
I0913 17:07:25.339274241 17165 ref_counted.h:199] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:105 unref 3 -> 2 TryToConnect
I0913 17:07:25.339380121 17165 work_serializer.cc:102] WorkSerializer::DrainQueue() 0x7fa6ec00bf80
I0913 17:07:25.339499410 17165 work_serializer.cc:117] Queue Drained
I0913 17:07:25.339617022 17165 work_serializer.cc:55] WorkSerializer::Run() 0x7fa6ec00bf80 Scheduling callback [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc:327]
I0913 17:07:25.339708934 17165 work_serializer.cc:65] Executing immediately
I0913 17:07:25.339876553 17165 resolving_lb_policy.cc:293] resolving_lb=0x7fa6ec001680: got resolver result
I0913 17:07:25.340316832 17165 ref_counted.h:92] lb_policy_refcount:0x7fa6ec001688 ref 3 -> 4
I0913 17:07:25.340460426 17165 ev_posix.cc:352] (polling-api) pollset_set_create(0x7fa6ec012e20)
I0913 17:07:25.340571334 17165 resolving_lb_policy.cc:258] resolving_lb=0x7fa6ec001680: created new LB policy 0x7fa6ec012b80
I0913 17:07:25.340654584 17165 ev_posix.cc:377] (polling-api) pollset_set_add_pollset_set(0x7fa6ec012e20, 0x7fa6ec0105f0)
I0913 17:07:25.380510767 17165 ev_epollex_linux.cc:1487] PSS: merge (0x7fa6ec012e20, 0x7fa6ec0105f0)
I0913 17:07:25.380744037 17165 ev_epollex_linux.cc:1521] PSS: parent 0x7fa6ec012e20 to 0x7fa6ec00fc70
I0913 17:07:25.380469980 17168 timer_generic.cc:716] TIMER CHECK BEGIN: now=78 next=9223372036854775807 tls_min=0 glob_min=0
I0913 17:07:25.385597447 17168 timer_generic.cc:611] .. shard[0]->min_deadline = 1
I0913 17:07:25.386931416 17168 timer_generic.cc:534] .. shard[0]: heap_empty=true
I0913 17:07:25.391432826 17168 timer_generic.cc:509] .. shard[0]->queue_deadline_cap --> 1078
I0913 17:07:25.391458807 17168 timer_generic.cc:574] .. shard[0] popped 0
I0913 17:07:25.392370096 17168 timer_generic.cc:629] .. result --> 1, shard[0]->min_deadline 1 --> 1079, now=78
I0913 17:07:25.393245905 17168 timer_generic.cc:534] .. shard[1]: heap_empty=true
I0913 17:07:25.394042375 17168 timer_generic.cc:509] .. shard[1]->queue_deadline_cap --> 1078
I0913 17:07:25.394793309 17168 timer_generic.cc:574] .. shard[1] popped 0
I0913 17:07:25.395162909 17168 timer_generic.cc:629] .. result --> 1, shard[1]->min_deadline 1 --> 1079, now=78
I0913 17:07:25.384528317 17165 resolving_lb_policy.cc:240] resolving_lb=0x7fa6ec001680: Updating child policy 0x7fa6ec012b80
I0913 17:07:25.396767303 17165 child_policy_handler.cc:211] [child_policy_handler 0x7fa6ec012b80] creating new child policy pick_first
I0913 17:07:25.397911862 17165 ref_counted.h:103] lb_policy_refcount:0x7fa6ec012b88 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/lb_policy/child_policy_handler.cc:258 ref 1 -> 2 Helper
I0913 17:07:25.397978351 17165 ev_posix.cc:352] (polling-api) pollset_set_create(0x7fa6ec0130c0)
I0913 17:07:25.398036179 17165 pick_first.cc:146] Pick First 0x7fa6ec010af0 created.
I0913 17:07:25.399215100 17165 child_policy_handler.cc:272] [child_policy_handler 0x7fa6ec012b80] created new LB policy "pick_first" (0x7fa6ec010af0)
I0913 17:07:25.399387189 17165 ev_posix.cc:377] (polling-api) pollset_set_add_pollset_set(0x7fa6ec0130c0, 0x7fa6ec012e20)
I0913 17:07:25.399397525 17165 ev_epollex_linux.cc:1487] PSS: merge (0x7fa6ec0130c0, 0x7fa6ec012e20)
I0913 17:07:25.399502008 17165 ev_epollex_linux.cc:1521] PSS: parent 0x7fa6ec0130c0 to 0x7fa6ec00fc70
I0913 17:07:25.399616268 17165 child_policy_handler.cc:230] [child_policy_handler 0x7fa6ec012b80] updating child policy 0x7fa6ec010af0
I0913 17:07:25.399732763 17165 pick_first.cc:265] Pick First 0x7fa6ec010af0 received update with 1 addresses
I0913 17:07:25.399861830 17165 subchannel_list.h:361] [pick_first 0x7fa6ec010af0] Creating subchannel list 0x7fa6ec00ed40 for 1 subchannels
I0913 17:07:25.400047329 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 2 -> 3
I0913 17:07:25.400099011 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 3 -> 4
I0913 17:07:25.407755012 17168 timer_generic.cc:534] .. shard[2]: heap_empty=true
I0913 17:07:25.407776244 17168 timer_generic.cc:509] .. shard[2]->queue_deadline_cap --> 1078
I0913 17:07:25.407922352 17168 timer_generic.cc:574] .. shard[2] popped 0
I0913 17:07:25.408062314 17168 timer_generic.cc:629] .. result --> 1, shard[2]->min_deadline 1 --> 1079, now=78
I0913 17:07:25.408211215 17168 timer_generic.cc:534] .. shard[3]: heap_empty=true
I0913 17:07:25.408946504 17168 timer_generic.cc:509] .. shard[3]->queue_deadline_cap --> 1078
I0913 17:07:25.409368625 17168 timer_generic.cc:574] .. shard[3] popped 0
I0913 17:07:25.409511939 17168 timer_generic.cc:629] .. result --> 1, shard[3]->min_deadline 1 --> 1079, now=78
I0913 17:07:25.409652739 17168 timer_generic.cc:739] TIMER CHECK END: r=1; next=1079
I0913 17:07:25.409791304 17168 timer_manager.cc:188] sleep for a 1001 milliseconds
I0913 17:07:25.410554530 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 4 -> 5
I0913 17:07:25.410646720 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 5 -> 6
I0913 17:07:25.410764054 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:92 ref 1 -> 2 connector_arg_copy
I0913 17:07:25.410875800 17165 ref_counted.h:199] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/ext/transport/chttp2/client/secure/secure_channel_create.cc:144 unref 2 -> 1 lb_channel_create
I0913 17:07:25.410993133 17165 ref_counted.h:183] subchannel_pool:0x7fa6ec00b9c8 unref 6 -> 5
I0913 17:07:25.411102086 17165 ref_counted.h:183] subchannel_pool:0x7fa6ec00b9c8 unref 5 -> 4
I0913 17:07:25.411285908 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:92 ref 1 -> 2 connector_arg_copy
I0913 17:07:25.411333959 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 4 -> 5
I0913 17:07:25.411511914 17165 ev_posix.cc:352] (polling-api) pollset_set_create(0x7fa6ec03d640)
I0913 17:07:25.411604943 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 5 -> 6
I0913 17:07:25.411669755 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:92 ref 2 -> 3 connector_arg_copy
I0913 17:07:25.413236714 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:92 ref 3 -> 4 connector_arg_copy
I0913 17:07:25.413338403 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 6 -> 7
I0913 17:07:25.413450429 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 7 -> 8
I0913 17:07:25.413561895 17165 ref_counted.h:183] subchannel_pool:0x7fa6ec00b9c8 unref 8 -> 7
I0913 17:07:25.413666937 17165 ref_counted.h:199] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:86 unref 4 -> 3 connector_arg_destroy
I0913 17:07:25.413773375 17165 ref_counted.h:183] subchannel_pool:0x7fa6ec00b9c8 unref 7 -> 6
I0913 17:07:25.414899775 17165 client_channel.cc:868] chand=0x7fa6ec00fa08: creating subchannel wrapper 0x7fa6ec03da30 for subchannel 0x7fa6ec033280
I0913 17:07:25.414985540 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 2 -> 3 SubchannelWrapper
I0913 17:07:25.415218810 17165 subchannel_list.h:404] [pick_first 0x7fa6ec010af0] subchannel list 0x7fa6ec00ed40 index 0: Created subchannel 0x7fa6ec03da30 for address uri ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/)
I0913 17:07:25.415344245 17165 ref_counted.h:103] lb_policy_refcount:0x7fa6ec010af8 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/lb_policy/pick_first/pick_first.cc:93 ref 1 -> 2 subchannel_list
I0913 17:07:25.415441464 17165 subchannel_list.h:308] [pick_first 0x7fa6ec010af0] subchannel list 0x7fa6ec00ed40 index 0 of 1 (subchannel 0x7fa6ec03da30): starting watch (from IDLE)
I0913 17:07:25.415542035 17165 ref_counted.h:103] pick_first:0x7fa6ec00ed48 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/lb_policy/subchannel_list.h:317 ref 1 -> 2 Watcher
I0913 17:07:25.415705464 17165 ref_counted.h:103] client_channel_routing:0x7fa6ec03da38 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:921 ref 1 -> 2 WatcherWrapper
I0913 17:07:25.443438598 17165 ev_posix.cc:377] (polling-api) pollset_set_add_pollset_set(0x7fa6ec03d640, 0x7fa6ec0130c0)
I0913 17:07:25.443463740 17165 ev_epollex_linux.cc:1487] PSS: merge (0x7fa6ec03d640, 0x7fa6ec0130c0)
I0913 17:07:25.443569620 17165 ev_epollex_linux.cc:1521] PSS: parent 0x7fa6ec03d640 to 0x7fa6ec00fc70
I0913 17:07:25.443969391 17165 socket_utils_common_posix.cc:320] Enabling TCP_USER_TIMEOUT with a timeout of 10000 ms
I0913 17:07:25.445888071 17165 ev_posix.cc:254] (polling-api) fd_create(18, tcp-client:ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/), 1)
I0913 17:07:25.446187550 17165 ev_posix.cc:255] (fd-trace) fd_create(18, tcp-client:ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/), 1)
I0913 17:07:25.446316058 17165 ev_posix.cc:394] (polling-api) pollset_set_add_fd(0x7fa6ec03d640, 18)
I0913 17:07:25.446402382 17165 ev_epollex_linux.cc:1353] PSS:0x7fa6ec03d640: add fd 0x7fa6ec03e720 (18)
I0913 17:07:25.446514687 17165 ev_epollex_linux.cc:616] add fd 0x7fa6ec03e720 (18) to pollable 0x7fa6ec011dc0
I0913 17:07:25.446658560 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 6 -> 7
I0913 17:07:25.446745163 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:92 ref 3 -> 4 connector_arg_copy
I0913 17:07:25.446854395 17165 tcp_client_posix.c[4173222.577355] serial8250: too much work for irq4
c:329] CLIENT_CONNECT: ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/): asynchronously connecting fd 0x7fa6ec03e720
I0913 17:07:25.446979271 17165 timer_generic.cc:364] TIMER 0x7fa6ec012c20: SET 20036 now 36 call 0x7fa6ec012c58[0x7fa70b61f33b]
I0913 17:07:25.447319258 17165 timer_generic.cc:401] .. add to shard 0 with queue_deadline_cap=1078 => is_first_timer=false
I0913 17:07:25.447601138 17165 ref_counted.h:199] resolver_refcount:0x7fa6ec0129d8 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc:396 unref 2 -> 1 dns-resolving
I0913 17:07:25.447646674 17165 work_serializer.cc:102] WorkSerializer::DrainQueue() 0x7fa6ec00bf80
I0913 17:07:25.447753112 17165 work_serializer.cc:117] Queue Drained
I0913 17:07:25.447868211 17165 client_channel.cc:1033] chand=0x7fa6ec00fa08: connectivity change for subchannel wrapper 0x7fa6ec03da30 subchannel 0x7fa6ec033280; hopping into work_serializer
I0913 17:07:25.447979119 17165 work_serializer.cc:55] WorkSerializer::Run() 0x7fa6ec00bf80 Scheduling callback [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:1044]
I0913 17:07:25.448078573 17165 work_serializer.cc:65] Executing immediately
I0913 17:07:25.448185570 17165 client_channel.cc:1066] chand=0x7fa6ec00fa08: processing connectivity change in work serializer for subchannel wrapper 0x7fa6ec03da30 subchannel 0x7fa6ec033280 watcher=0x7fa6ec03ba20
I0913 17:07:25.448334751 17165 subchannel_list.h:241] [pick_first 0x7fa6ec010af0] subchannel list 0x7fa6ec00ed40 index 0 of 1 (subchannel 0x7fa6ec03da30): connectivity changed: state=CONNECTING, shutting_down=0, pending_watcher=0x7fa6ec03ba20
I0913 17:07:25.448437557 17165 ref_counted.h:103] lb_policy_refcount:0x7fa6ec010af8 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/lb_policy/pick_first/pick_first.cc:410 ref 2 -> 3 QueuePicker
I0913 17:07:25.448549582 17165 client_channel.cc:1306] chand=0x7fa6ec00fa08: update: state=CONNECTING picker=0x7fa6ec013ac0
I0913 17:07:25.448671944 17165 ref_counted.h:199] lb_policy_refcount:0x7fa6ec001688 /home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/lb_policy.h:375 unref 4 -> 3 QueuePicker
I0913 17:07:25.448745697 17165 work_serializer.cc:102] WorkSerializer::DrainQueue() 0x7fa6ec00bf80
I0913 17:07:25.449227881 17165 work_serializer.cc:117] Queue Drained
I0913 17:07:25.449335436 17165 init.cc:165] grpc_init(void)
I0913 17:07:25.449470928 17165 completion_queue.cc:526] grpc_completion_queue_create_internal(completion_type=0, polling_type=0)
I0913 17:07:25.449589100 17165 ev_posix.cc:304] (polling-api) pollset_init(0x7fa6ec03f298)
I0913 17:07:25.449735767 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:796 ref 1 -> 2 pollset
I0913 17:07:25.480508672 17165 channel_connectivity.cc:232] grpc_channel_watch_connectivity_state(channel=0x7fa6ec00f8c0, last_observed_state=0, deadline=gpr_timespec { tv_sec: 1663089445, tv_nsec: 335137682, clock_type: 1 }, cq=0x7fa6ec03f190, tag=0x7fa6ec013a50)
I0913 17:07:25.480579072 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 3 -> 4 watch_channel_connectivity
I0913 17:07:25.480707859 17165 ev_posix.cc:364] (polling-api) pollset_set_add_pollset(0x7fa6ec00fc70, 0x7fa6ec03f298)
I0913 17:07:25.480900901 17165 ev_epollex_linux.cc:1450] PSS:0x7fa6ec00fc70: add pollset 0x7fa6ec03f298
I0913 17:07:25.481010971 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1241 ref 2 -> 3 pollset_as_multipollable
I0913 17:07:25.481119085 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1244 unref 3 -> 2 pollset
I0913 17:07:25.481230272 17165 ev_epollex_linux.cc:569] (fd-trace) Pollable_create: created epfd: 19 (type: 0)
I0913 17:07:25.481345650 17165 ev_epollex_linux.cc:1249] PS:0x7fa6ec03f298 active pollable transition from empty to multi
I0913 17:07:25.481428063 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec03f638 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1280 ref 1 -> 2 pollset_set
I0913 17:07:25.481534501 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1281 unref 2 -> 1 pollset_as_multipollable
I0913 17:07:25.481642056 17165 ev_epollex_linux.cc:616] add fd 0x7fa6ec03e720 (18) to pollable 0x7fa6ec03f630
I0913 17:07:25.481755758 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec03f638 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1478 unref 2 -> 1 pollset_set
I0913 17:07:25.481859402 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 4 -> 5 ExternalConnectivityWatcher
I0913 17:07:25.482000482 17165 work_serializer.cc:55] WorkSerializer::Run() 0x7fa6ec00bf80 Scheduling callback [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:1155]
I0913 17:07:25.482080380 17165 work_serializer.cc:65] Executing immediately
I0913 17:07:25.482211402 17165 timer_generic.cc:364] TIMER 0x7fa6ec03f418: SET 600034 now 180 call 0x7fa6ec03f388[0x7fa70b482e3b]
I0913 17:07:25.482299402 17165 timer_generic.cc:401] .. add to shard 0 with queue_deadline_cap=1078 => is_first_timer=false
I0913 17:07:25.482430983 17165 connectivity_state.cc:118] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: add watcher 0x7fa6ec03f130
I0913 17:07:25.482516190 17165 connectivity_state.cc:124] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: notifying watcher 0x7fa6ec03f130: IDLE -> CONNECTING
I0913 17:07:25.482647771 17165 work_serializer.cc:55] WorkSerializer::Run() 0x7fa6ec00bf80 Scheduling callback [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:1175]
I0913 17:07:25.482732139 17165 work_serializer.cc:76] Scheduling on queue : item 0x7fa6ec03ec80
I0913 17:07:25.482900317 17165 work_serializer.cc:102] WorkSerializer::DrainQueue() 0x7fa6ec00bf80
I0913 17:07:25.482950044 17165 work_serializer.cc:134] Running item 0x7fa6ec03ec80 : callback scheduled at [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:1175]
I0913 17:07:25.483054526 17165 connectivity_state.cc:141] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: remove watcher 0x7fa6ec03f130
I0913 17:07:25.483184711 17165 ev_posix.cc:371] (polling-api) pollset_set_del_pollset(0x7fa6ec00fc70, 0x7fa6ec03f298)
I0913 17:07:25.483269917 17165 ev_epollex_linux.cc:1398] PSS:0x7fa6ec00fc70: del pollset 0x7fa6ec03f298
I0913 17:07:25.483379707 17165 ev_epollex_linux.cc:664] PS:0x7fa6ec03f298 (pollable:0x7fa6ec03f630) maybe_finish_shutdown sc=(nil) (target:!NULL) rw=(nil) (target:NULL) cpsc=0 (target:0)
I0913 17:07:25.511424054 17165 ref_counted.h:199] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:105 unref 5 -> 4 ExternalConnectivityWatcher
I0913 17:07:25.511541387 17165 work_serializer.cc:102] WorkSerializer::DrainQueue() 0x7fa6ec00bf80
I0913 17:07:25.511615978 17165 work_serializer.cc:117] Queue Drained
I0913 17:07:25.511634695 17165 timer_generic.cc:467] TIMER 0x7fa6ec03f418: CANCEL pending=true
I0913 17:07:25.511748397 17165 completion_queue.cc:707] cq_end_op_for_next(cq=0x7fa6ec03f190, tag=0x7fa6ec013a50, error="No Error", done=0x7fa70b482a2e, done_arg=0x7fa6ec03f310, storage=0x7fa6ec03f460)
I0913 17:07:25.511850644 17165 ev_posix.cc:331] (polling-api) pollset_kick(0x7fa6ec03f298, (nil))
I0913 17:07:25.511959597 1[4173222.646290] serial8250: too much work for irq4
7165 ev_epollex_linux.cc:731] PS:0x7fa6ec03f298 kick (nil) tls_pollset=(nil) tls_worker=(nil) pollset.root_worker=(nil)
I0913 17:07:25.512064638 17165 ev_epollex_linux.cc:741] PS:0x7fa6ec03f298 kicked_any_without_poller
I0913 17:07:25.512201248 17165 completion_queue.cc:982] grpc_completion_queue_next(cq=0x7fa6ec03f190, deadline=gpr_timespec { tv_sec: 9223372036854775807, tv_nsec: 0, clock_type: 1 }, reserved=(nil))
I0913 17:07:25.512371381 17165 ref_counted.h:199] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:105 unref 4 -> 3 watch_channel_connectivity
I0913 17:07:25.512474746 17165 completion_queue.cc:1083] RETURN_EVENT[0x7fa6ec03f190]: OP_COMPLETE: tag:0x7fa6ec013a50 OK
I0913 17:07:25.512537882 17165 completion_queue.cc:1425] grpc_completion_queue_destroy(cq=0x7fa6ec03f190)
I0913 17:07:25.512637895 17165 completion_queue.cc:1419] grpc_completion_queue_shutdown(cq=0x7fa6ec03f190)
I0913 17:07:25.512750200 17165 ev_posix.cc:309] (polling-api) pollset_shutdown(0x7fa6ec03f298)
I0913 17:07:25.512975648 17165 ev_epollex_linux.cc:664] PS:0x7fa6ec03f298 (pollable:0x7fa6ec03f630) maybe_finish_shutdown sc=0x7fa6ec03f1d0 (target:!NULL) rw=(nil) (target:NULL) cpsc=0 (target:0)
I0913 17:07:25.513091305 17165 ev_posix.cc:314] (polling-api) pollset_destroy(0x7fa6ec03f298)
I0913 17:07:25.513200816 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec03f638 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:927 unref 1 -> 0 pollset
I0913 17:07:25.513311165 17165 ev_epollex_linux.cc:149] (fd-trace) pollable_unref: Closing epfd: 19
I0913 17:07:25.513445540 17165 init.cc:212] grpc_shutdown(void)
I0913 17:07:25.513544994 17165 channel_connectivity.cc:43] grpc_channel_check_connectivity_state(channel=0x7fa6ec00f8c0, try_to_connect=1)
I0913 17:07:25.513658695 17165 connectivity_state.cc:174] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: get current state: CONNECTING
I0913 17:07:25.513767368 17165 init.cc:165] grpc_init(void)
I0913 17:07:25.513882746 17165 completion_queue.cc:526] grpc_completion_queue_create_internal(completion_type=0, polling_type=0)
I0913 17:07:25.513992257 17165 ev_posix.cc:304] (polling-api) pollset_init(0x7fa6ec03f738)
I0913 17:07:25.514111267 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:796 ref 1 -> 2 pollset
I0913 17:07:25.514229997 17165 channel_connectivity.cc:232] grpc_channel_watch_connectivity_state(channel=0x7fa6ec00f8c0, last_observed_state=1, deadline=gpr_timespec { tv_sec: 1663089445, tv_nsec: 335137682, clock_type: 1 }, cq=0x7fa6ec03f630, tag=0x7fa6ec013b20)
I0913 17:07:25.514340067 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 3 -> 4 watch_channel_connectivity
I0913 17:07:25.514452930 17165 ev_posix.cc:364] (polling-api) pollset_set_add_pollset(0x7fa6ec00fc70, 0x7fa6ec03f738)
I0913 17:07:25.514564397 17165 ev_epollex_linux.cc:1450] PSS:0x7fa6ec00fc70: add pollset 0x7fa6ec03f738
I0913 17:07:25.514675584 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1241 ref 2 -> 3 pollset_as_multipollable
I0913 17:07:25.542637797 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1244 unref 3 -> 2 pollset
I0913 17:07:25.542729429 17165 ev_epollex_linux.cc:569] (fd-trace) Pollable_create: created epfd: 19 (type: 0)
I0913 17:07:25.542869950 17165 ev_epollex_linux.cc:1249] PS:0x7fa6ec03f738 active pollable transition from empty to multi
I0913 17:07:25.542934483 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec03fdd8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1280 ref 1 -> 2 pollset_set
I0913 17:07:25.543148477 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec00ac08 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1281 unref 2 -> 1 pollset_as_multipollable
I0913 17:07:25.543274750 17165 ev_epollex_linux.cc:616] add fd 0x7fa6ec03e720 (18) to pollable 0x7fa6ec03fdd0
I0913 17:07:25.543391245 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec03fdd8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1478 unref 2 -> 1 pollset_set
I0913 17:07:25.543491817 17165 ref_counted.h:133] stream_refcount:0x7fa6ec00f958 /home/rcheppudira/new_grpc/grpc/src/core/lib/transport/transport.h:88 ref 4 -> 5 ExternalConnectivityWatcher
I0913 17:07:25.543632617 17165 work_serializer.cc:55] WorkSerializer::Run() 0x7fa6ec00bf80 Scheduling callback [/home/rcheppudira/new_grpc/grpc/src/core/ext/filters/client_channel/client_channel.cc:1155]
I0913 17:07:25.543722572 17165 work_serializer.cc:65] Executing immediately
I0913 17:07:25.543974280 17165 timer_generic.cc:364] TIMER 0x7fa6ec03f418: SET 600034 now 242 call 0x7fa6ec03f388[0x7fa70b482e3b]
I0913 17:07:25.544049988 17165 timer_generic.cc:401] .. add to shard 0 with queue_deadline_cap=1078 => is_first_timer=false
I0913 17:07:25.544172909 17165 connectivity_state.cc:118] ConnectivityStateTracker client_channel[0x7fa6ec00fac8]: add watcher 0x7fa6ec03f5d0
I0913 17:07:25.544299461 17165 work_serializer.cc:102] WorkSerializer::DrainQueue() 0x7fa6ec00bf80
I0913 17:07:25.544386902 17165 work_serializer.cc:117] Queue Drained
I0913 17:07:25.544504236 17165 completion_queue.cc:982] grpc_completion_queue_next(cq=0x7fa6ec03f630, deadline=gpr_timespec { tv_sec: 9223372036854775807, tv_nsec: 0, clock_type: 1 }, reserved=(nil))
I0913 17:07:25.544651741 17165 ev_posix.cc:322] (polling-api) pollset_work(0x7fa6ec03f738, 9223372036854775807) begin
I0913 17:07:25.544765442 17165 ev_epollex_linux.cc:1120] PS:0x7fa6ec03f738 work hdl=(nil) worker=0x7fa7057602f0 now=243 deadline=9223372036854775807 kwp=0 pollable=0x7fa6ec03fdd0
I0913 17:07:25.544948147 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec03fdd8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1018 ref 1 -> 2 pollset_worker
I0913 17:07:25.545075537 17165 ev_epollex_linux.cc:938] POLLABLE:0x7fa6ec03fdd0[type=pollset epfd=19 wakeup=20] poll for -1ms
I0913 17:07:25.571483922 17165 ev_epollex_linux.cc:957] POLLABLE:0x7fa6ec03fdd0 got 1 events
I0913 17:07:25.571542309 17165 ev_epollex_linux.cc:905] PS:0x7fa6ec03f738 got fd 0x7fa6ec03e720: cancel=0 read=0 write=1
I0913 17:07:25.571681153 17165 tcp_client_posix.cc:142] CLIENT_CONNECT: ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/): on_writable: error="No Error"
I0913 17:07:25.571790944 17165 timer_generic.cc:467] TIMER 0x7fa6ec012c20: CANCEL pending=true
I0913 17:07:25.571914423 17165 ev_posix.cc:400] (polling-api) pollset_set_del_fd(0x7fa6ec03d640, 18)
I0913 17:07:25.572013319 17165 ev_epollex_linux.cc:1377] PSS:0x7fa6ec03d640: del fd 0x7fa6ec03e720
I0913 17:07:25.572390182 17165 combiner.cc:61] C:0x7fa6ec03f4f0 create
I0913 17:07:25.572544950 17165 ref_counted.h:103] tcp:0x7fa6ec0403d8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/tcp_posix.cc:1823 ref 1 -> 2 error-tracking
I0913 17:07:25.572643566 17165 tcp_client_posix.cc:105] CLIENT_CONNECT: ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/): on_alarm: error="Cancelled"
I0913 17:07:25.572758665 17165 ref_counted.h:183] subchannel_pool:0x7fa6ec00b9c8 unref 7 -> 6
I0913 17:07:25.572988862 17165 ref_counted.h:199] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:86 unref 4 -> 3 connector_arg_destroy
I0913 17:07:25.573173801 17165 handshaker.cc:99] handshake_manager 0x7fa6ec038720: adding handshaker http_connect [0x7fa6ec040f40] at index 0
I0913 17:07:25.573545916 17165 ssl_transport_security.cc:220] HANDSHAKE START - before SSL initialization - PINIT
I0913 17:07:25.573650398 17165 ssl_transport_security.cc:220] LOOP - before SSL initialization - PINIT
I0913 17:07:25.573789243 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS write client hello - TWCH
I0913 17:07:25.573852379 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/transport/security_handshaker.cc:113 ref 3 -> 4 handshake
I0913 17:07:25.573949878 17165 handshaker.cc:99] handshake_manager 0x7fa6ec038720: adding handshaker security [0x7fa6ec05c6e0] at index 1
I0913 17:07:25.574059668 17165 ev_posix.cc:394] (polling-api) pollset_set_add_fd(0x7fa6ec03d640, 18)
I0913 17:07:25.574176722 17165 ev_epollex_linux.cc:1353] PSS:0x7fa6ec03d640: add fd 0x7fa6ec03e720 (18)
I0913 17:07:25.574290144 17165 ev_epollex_linux.cc:616] add fd 0x7fa6ec03e720 (18) to pollable 0x7fa6ec011dc0
I0913 17:07:25.574407198 17165 ev_epollex_linux.cc:616] add fd 0x7fa6ec03e720 (18) to pollable 0x7fa6ec03fdd0
I0913 17:07:25.574524252 17165 ref_counted.h:92] subchannel_pool:0x7fa6ec00b9c8 ref 6 -> 7
I0913 17:07:25.574623148 17165 ref_counted.h:103] security_connector_refcount:0x7fa6ec013038 /home/rcheppudira/new_grpc/grpc/src/core/lib/security/security_connector/security_connector.cc:92 ref 4 -> 5 connector_arg_copy
I0913 17:07:25.574756125 17165 timer_generic.cc:364] TIMER 0x7fa6ec0387d8: SET 20036 now 273 call 0x7fa6ec038810[0x7fa70b5f12d2]
I0913 17:07:25.574843287 17165 timer_generic.cc:401] .. add to shard 0 with queue_deadline_cap=1078 => is_first_timer=false
I0913 17:07:25.575114551 17165 handshaker.cc:129] handshake_manager 0x7fa6ec038720: error="No Error" shutdown=0 index=0, args={endpoint=0x7fa6ec0403b0, args=0x7fa6ec012830 {size=12: grpc.primary_user_agent=grpc-c++/1.30.0, grpc.keepalive_time_ms=10000, grpc.keepalive_timeout_ms=10000, grpc.ssl_target_name_override=Infinera.com, grpc.client_channel_factory=0x7fa6ec00e9f0, grpc.channel_credentials=0x7fa6ec010d80, grpc.server_uri=dns:///[10.220.6.50:50058](http://10.220.6.50:50058/), grpc.default_authority=Infinera.com, grpc.subchannel_pool=0x7fa6ec00b9c0, grpc.http2_scheme=https, grpc.security_connector=0x7fa6ec013030, grpc.subchannel_address=ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/)}, read_buffer=0x7fa6ec012bf0 (length=0), exit_early=0}
I0913 17:07:25.575178246 17165 handshaker.cc:176] handshake_manager 0x7fa6ec038720: calling handshaker http_connect [0x7fa6ec040f40] at index 0
I0913 17:07:25.575312900 17165 handshaker.cc:129] handshake_manager 0x7fa6ec038720: error="No Error" shutdown=0 index=1, args={endpoint=0x7fa6ec0403b0, args=0x7fa6ec012830 {size=12: grpc.primary_user_agent=grpc-c++/1.30.0, grpc.keepalive_time_ms=10000, grpc.keepalive_timeout_ms=10000, grpc.ssl_target_name_override=Infinera.com, grpc.client_channel_factory=0x7fa6ec00e9f0, grpc.channel_credentials=0x7fa6ec010d80, grpc.server_uri=dns:///[10.220.6.50:50058](http://10.220.6.50:50058/), grpc.default_authority=Infinera.com, grpc.subchannel_pool=0x7fa6ec00b9c0, grpc.http2_scheme=https, grpc.security_connector=0x7fa6ec013030, grpc.subchannel_address=ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/)}, read_buffer=0x7fa6ec012bf0 (length=0), exit_early=0}
I0913 17:07:25.603279024 17165 handshaker.cc:176] handshake_manager 0x7fa6ec038720: calling handshaker security [0x7fa6ec05c6e0] at index 1
I0913 17:07:25.603416192 17165 tcp_posix.cc:1566] WRITE 0x7fa6ec0403b0 (peer=ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/))
I0913 17:07:25.603619012 17165 tcp_posix.cc:1616] write: "No Error"
I0913 17:07:25.603753386 17165 ref_counted.h:103] tcp:0x7fa6ec0403d8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/tcp_posix.cc:908 ref 2 -> 3 read
I0913 17:07:25.604010961 17165 tcp_posix.cc:541] TCP:0x7fa6ec0403b0 notify_on_read
I0913 17:07:25.604401513 17165 ref_counted.h:199] pollable_refcount:0x7fa6ec03fdd8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1085 unref 2 -> 1 pollset_worker
I0913 17:07:25.604615228 17165 ev_epollex_linux.cc:664] PS:0x7fa6ec03f738 (pollable:0x7fa6ec03fdd0) maybe_finish_shutdown sc=(nil) (target:!NULL) rw=(nil) (target:NULL) cpsc=1 (target:0)
I0913 17:07:25.604730047 17165 ev_posix.cc:325] (polling-api) pollset_work(0x7fa6ec03f738, 9223372036854775807) end
I0913 17:07:25.604973653 17165 ev_posix.cc:322] (polling-api) pollset_work(0x7fa6ec03f738, 9223372036854775807) begin
I0913 17:07:25.605089869 17165 ev_epollex_linux.cc:1120] PS:0x7fa6ec03f738 work hdl=(nil) worker=0x7fa7057602f0 now=273 deadline=9223372036854775807 kwp=0 pollable=0x7fa6ec03fdd0
I0913 17:07:25.605204967 17165 ref_counted.h:103] pollable_refcount:0x7fa6ec03fdd8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/ev_epollex_linux.cc:1018 ref 1 -> 2 pollset_worker
I0913 17:07:25.605318669 17165 ev_epollex_linux.cc:938] POLLABLE:0x7fa6ec03fdd0[type=pollset epfd=19 wakeup=20] poll for -1ms
I0913 17:07:25.624281973 17165 ev_epollex_linux.cc:957] POLLABLE:0x7fa6ec03fdd0 got 1 events
I0913 17:07:25.625191306 17165 ev_epollex_linux.cc:905] PS:0x7fa6ec03f738 got fd 0x7fa6ec03e720: cancel=0 read=1 write=1
I0913 17:07:25.626231103 17165 tcp_posix.cc:887] TCP:0x7fa6ec0403b0 got_read: "No Error"
I0913 17:07:25.627087357 17165 tcp_posix.cc:869] TCP:0x7fa6ec0403b0 alloc_slices
I0913 17:07:25.627883827 17165 resource_quota.cc:890] RQ anonymous_pool_7fa6ec03f7b0 ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/): alloc 8192; free_pool -> -8192
I0913 17:07:25.629396031 17165 combiner.cc:135] C:0x7fa6ec03f4f0 grpc_combiner_execute c=0x7fa6ec040cf8 last=1
I0913 17:07:25.630779167 17165 combiner.cc:199] C:0x7fa6ec03f4f0 grpc_combiner_continue_exec_ctx contended=0 exec_ctx_ready_to_finish=0 time_to_execute_final_list=0
I0913 17:07:25.633678419 17165 combiner.cc:222] C:0x7fa6ec03f4f0 maybe_finish_one n=0x7fa6ec040cf8
I0913 17:07:25.635280857 17165 combiner.cc:308] C:0x7fa6ec03f4f0 grpc_combiner_execute_finally c=0x7fa6ec03f820; ac=0x7fa6ec03f4f0
I0913 17:07:25.636448324 17165 combiner.cc:264] C:0x7fa6ec03f4f0 finish old_state=5
I0913 17:07:25.637369670 17165 combiner.cc:199] C:0x7fa6ec03f4f0 grpc_combiner_continue_exec_ctx contended=0 exec_ctx_ready_to_finish=0 time_to_execute_final_list=1
I0913 17:07:25.638760349 17165 combiner.cc:246] C:0x7fa6ec03f4f0 execute_final[0] c=0x7fa6ec03f820
I0913 17:07:25.639790648 17165 resource_quota.cc:321] RQ: check allocation for user 0x7fa6ec040cf0 shutdown=0 free_pool=-8192
I0913 17:07:25.640919004 17165 resource_quota.cc:347] RQ anonymous_pool_7fa6ec03f7b0 ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/): grant alloc 8192 bytes; rq_free_pool -> 9223372036854767615
I0913 17:07:25.642398521 17165 combiner.cc:264] C:0x7fa6ec03f4f0 finish old_state=3
I0913 17:07:25.643722712 17165 tcp_posix.cc:850] TCP:0x7fa6ec0403b0 read_allocation_done: "No Error"
I0913 17:07:25.644955550 17165 tcp_posix.cc:680] TCP:0x7fa6ec0403b0 call_cb 0x7fa6ec05c8d8 0x7fa70b65e450:0x7fa6ec05c6e0
I0913 17:07:25.646910547 17165 tcp_posix.cc:683] READ 0x7fa6ec0403b0 (peer=ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/)) error="No Error"
I0913 17:07:25.648767487 17165 ref_counted.h:199] tcp:0x7fa6ec0403d8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/tcp_posix.cc:844 unref 3 -> 2 read
I0913 17:07:25.650205100 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS write client hello - TWCH
I0913 17:07:25.652048630 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read server hello - TRSH
I0913 17:07:25.653957812 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read server certific - TRSC
I0913 17:07:25.655300999 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read server key exch - TRSKE
I0913 17:07:25.656334650 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read server done - TRSD
I0913 17:07:25.659340898 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS write client key exc - TWCKE
I0913 17:07:25.660446066 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS write change cipher - TWCCS
I0913 17:07:25.661612974 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS write finished - TWFIN
I0913 17:07:25.662694117 17165 tcp_posix.cc:1566] WRITE 0x7fa6ec0403b0 (peer=ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/))
I0913 17:07:25.664612238 17165 tcp_posix.cc:1616] write: "No Error"
I0913 17:07:25.665905140 17165 ref_counted.h:103] tcp:0x7fa6ec0403d8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/tcp_posix.cc:908 ref 2 -> 3 read
I0913 17:07:25.667535235 17165 tcp_posix.cc:887] TCP:0x7fa6ec0403b0 got_read: "No Error"
I0913 17:07:25.668567210 17165 tcp_posix.cc:879] TCP:0x7fa6ec0403b0 do_read
I0913 17:07:25.669521521 17165 tcp_posix.cc:680] TCP:0x7fa6ec0403b0 call_cb 0x7fa6ec05c8d8 0x7fa70b65e450:0x7fa6ec05c6e0
I0913 17:07:25.670683960 17165 tcp_posix.cc:683] READ 0x7fa6ec0403b0 (peer=ipv4:[10.220.6.50:50058](http://10.220.6.50:50058/)) error="No Error"
I0913 17:07:25.671891934 17165 ref_counted.h:199] tcp:0x7fa6ec0403d8 /home/rcheppudira/new_grpc/grpc/src/core/lib/iomgr/tcp_posix.cc:844 unref 3 -> 2 read
I0913 17:07:25.673510017 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS write finished - TWFIN
I0913 17:07:25.674696760 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read server session - TRST
I0913 17:07:25.675933230 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read change cipher s - TRCCS
I0913 17:07:25.677194564 17165 ssl_transport_security.cc:220] LOOP - SSLv3/TLS read finished - TRFIN
I0913 17:07:25.678290792 17165 ssl_transport_security.cc:220] HANDSHAKE DONE - SSL negotiation finished succe - SSLOK
E0913 17:07:25.680856761 17165 ssl_transport_security.cc:470] assertion failed: (int)peer->property_count == current_insert_index
signal 6[SIGABRT] caught by process [/opt/infinera/thanos/local/bin/tunnelclient]
#####DumpStacktrace ########
#0 /opt/infinera/thanos/local/lib/libBaseSystem.so unsigned long backward::details::unwind<backward::StackTraceImpl<backward::system_tag::linux_tag>::callback>(backward::StackTraceImpl<backward::system_tag::linux_tag>::callback, unsigned long) [0x7fa709dd4c72]
#1 /opt/infinera/thanos/local/lib/libBaseSystem.so backward::StackTraceImpl<backward::system_tag::linux_tag>::load_here(unsigned long) [0x7fa709dd4008]
#2 /opt/infinera/thanos/local/lib/libBaseSystem.so util::Backtrace::DumpStacktrace() [0x7fa709dd2c5e]
#3 /opt/infinera/thanos/local/lib/libBaseSystem.so OsProcess::MyFailureHandler(int) [0x7fa709db54e7]
#4 /lib/x86_64-linux-gnu/libc.so.6 [0x7fa70988683f]
#5 /lib/x86_64-linux-gnu/libc.so.6 gsignal [0x7fa7098867bb]
#6 /lib/x86_64-linux-gnu/libc.so.6 abort [0x7fa709871534]
#7 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b6a1dfd]
#8 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b6a3acf]
#9 /opt/infinera/thanos/local/lib/libgrpc.so.10 tsi_handshaker_result_extract_peer(tsi_handshaker_result const*, tsi_peer*) [0x7fa70b6a6db1]
#10 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b65df19]
#11 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b65e2b2]
#12 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b65e44b]
#13 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b65e5fe]
#14 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b6117ac]
#15 /opt/infinera/thanos/local/lib/libgrpc.so.10 grpc_core::ExecCtx::Flush() [0x7fa70b611cdc]
#16 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b6098b7]
#17 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b6110bd]
#18 /opt/infinera/thanos/local/lib/libgrpc.so.10 grpc_pollset_work(grpc_pollset*, grpc_pollset_worker**, long) [0x7fa70b61859f]
#19 /opt/infinera/thanos/local/lib/libgrpc.so.10 [0x7fa70b674c8f]
#20 /opt/infinera/thanos/local/lib/libgrpc.so.10 grpc_completion_queue_next [0x7fa70b675154]
#21 /opt/infinera/thanos/local/lib/libgrpc++.so.1 grpc_impl::CompletionQueue::AsyncNextInternal(void**, bool*, gpr_timespec) [0x7fa70b09a675]
#22 /opt/infinera/thanos/local/lib/libgrpc++.so.1 grpc_impl::CompletionQueue::Next(void**, bool*) [0x7fa70b088ff6]
#23 /opt/infinera/thanos/local/lib/libgrpc++.so.1 grpc_impl::Channel::WaitForStateChangeImpl(grpc_connectivity_state, gpr_timespec) [0x7fa70b0883b4]
#24 /opt/infinera/thanos/local/bin/tunnelclient bool grpc::ChannelInterface::WaitForStateChange<gpr_timespec>(grpc_connectivity_state, gpr_timespec) [0x55d52d47a6e3]
#25 /opt/infinera/thanos/local/bin/tunnelclient bool grpc::ChannelInterface::WaitForConnected<gpr_timespec>(gpr_timespec) [0x55d52d4797fe]
#26 /opt/infinera/thanos/local/bin/tunnelclient ConnectionManager::createChannel(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::shared_ptr<grpc_impl::Channel>&) [0x55d52d472e6f]
#27 /opt/infinera/thanos/local/bin/tunnelclient ConnectionManager::connectToServer() [0x55d52d473b81]
#28 /opt/infinera/thanos/local/bin/tunnelclient boost::_mfi::mf0<void, ConnectionManager>::operator()(ConnectionManager*) const [0x55d52d47d4bc]
#29 /opt/infinera/thanos/local/bin/tunnelclient void boost::_bi::list1<boost::_bi::value<ConnectionManager*> >::operator()<boost::_mfi::mf0<void, ConnectionManager>, boost::_bi::list0>(boost::_bi::type<void>, boost::_mfi::mf0<void, ConnectionManager>&, boost::_bi::list0&, int) [0x55d52d47d41e]
#30 /opt/infinera/thanos/local/bin/tunnelclient boost::_bi::bind_t<void, boost::_mfi::mf0<void, ConnectionManager>, boost::_bi::list1<boost::_bi::value<ConnectionManager*> > >::operator()() [0x55d52d47d2cc]
#31 /opt/infinera/thanos/local/bin/tunnelclient boost::detail::thread_data<boost::_bi::bind_t<void, boost::_mfi::mf0<void, ConnectionManager>, boost::_bi::list1<boost::_bi::value<ConnectionManager*> > > >::run() [0x55d52d47d0f3]
#####End DumpStacktrace ########
**Ideally it must work fine or through some error. But here it is getting crashed.**
**Below is the code snippets.**
std::shared_ptr<grpc::ChannelCredentials> ConnectionManager::createChannelCredentials()
{
grpc::SslCredentialsOptions sslCredOptions;
std::string rootCertPath("/var/security/x509/trustedCerts/RSA_2048_SHA_512_root.crt");
std::string clientKeyPath("/var/security/x509/keys/RSA_2048_SHA_512_client.key");
std::string clientCertPath("/var/security/x509/certs/RSA_2048_SHA_512_client.crt");
//Here caCert, clientCert, ClientKey is of type string
getFileData(rootCertPath, caCert);//getFileData() will read and populate cert/key from the provided loc
sslCredOptions.pem_root_certs = caCert;
getFileData(clientKeyPath, clientKey);
sslCredOptions.pem_private_key = clientKey;
getFileData(clientCertPath, clientCert);
sslCredOptions.pem_cert_chain = clientCert;
return grpc::SslCredentials(sslCredOptions);
}
//Here server
std::shared_ptr<grpc::Channel> ConnectionManager::createChannel(const std::string &serverAddress, std::shared_ptr<grpc::Channel>& channel_)
{
std::shared_ptr<grpc::ChannelCredentials> channelCreds = ::grpc::InsecureChannelCredentials();
channelCreds = createChannelCredentials();
ChannelArguments channelArg;
std::string san("<CertSANName>");
channelArg.SetInt(GRPC_ARG_KEEPALIVE_TIME_MS, keepAliveTime_);
channelArg.SetInt(GRPC_ARG_KEEPALIVE_TIMEOUT_MS, keepAliveTime_);
channelArg.SetSslTargetNameOverride(san);
channel_ = grpc::CreateCustomChannel(serverAddress_, channelCreds, channelArg);
std::chrono::time_point<std::chrono::system_clock> deadline = std::chrono::system_clock::now() + std::chrono::seconds(10);
//if(!channel_->WaitForConnected(gpr_time_add(gpr_now(GPR_CLOCK_REALTIME), gpr_time_from_seconds(600, GPR_TIMESPAN))) || channel_->GetState(false) != grpc_connectivity_state::GRPC_CHANNEL_READY)
if(!channel_->WaitForConnected(deadline) || channel_->GetState(false) != grpc_connectivity_state::GRPC_CHANNEL_READY)//It is crashing here
{
//notify and start some backoff logic
}
else
{
//Secure connection established
....
....
....
}
return channel_;
}
**Tested the same with below openssl utility and I can see in the pcap that handshake is happening successfully.**
"sudo openssl s_client -connect <ip>:<port>-cert /var/security/x509/certs/RSA_2048_SHA_512_client.crt -key /var/security/x509/keys/RSA_2048_SHA_512_client.key -CAfile /var/security/x509/trusted_certs/RSA_2048_SHA_512_root.crt -state -debug"
|
non_test
|
channel waitforconnected assertion failed while trying for mtls connection using grpc with c language linux version linux opennetworklinux smp mon jun utc gnu linux python gcc debian grpc secure communication created a custom channel and populated sslcredentials with ca cert client cert and client key below assertion is observed once channel waitforconnected is called init cc grpc init void init cc grpc shutdown void init cc grpc shutdown void init cc grpc init void secure channel create cc grpc secure channel create creds target args reserved nil init cc grpc init void ev posix cc polling api pollset set create ref counted h subchannel pool ref client channel cc chand creating client channel for channel stack ev posix cc polling api pollset init ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc ref pollset timer generic cc timer set now call timer generic cc add to shard with queue deadline cap is first timer false ev posix cc polling api pollset set add pollset ev epollex linux cc pss add pollset ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc ref pollset as multipollable ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc unref pollset ev epollex linux cc fd trace pollable create created epfd type ev epollex linux cc ps active pollable transition from empty to multi ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc ref pollset set ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc unref pollset as multipollable ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc unref pollset set init cc grpc init void init cc grpc shutdown void channel connectivity cc grpc channel check connectivity state channel try to connect connectivity state cc connectivitystatetracker client channel get current state idle ref counted h stream refcount home rcheppudira new grpc grpc src core lib transport transport h ref trytoconnect work serializer cc workserializer run scheduling callback work serializer cc executing immediately ref counted h stream refcount home rcheppudira new grpc grpc src core lib transport transport h ref clientchannelcontrolhelper ev posix cc polling api pollset set create ref counted h lb policy refcount ref ev posix cc polling api pollset set create ev posix cc polling api pollset set add pollset set ev epollex linux cc pss merge executor cc executor default executor step sub depth executor cc executor resolver executor step sub depth ev epollex linux cc pss parent to resolving lb policy cc resolving lb starting name resolution ref counted h lb policy refcount ref client channel cc chand update state connecting picker connectivity state cc connectivitystatetracker client channel idle connecting helper ref counted h resolver refcount home rcheppudira new grpc grpc src core ext filters client channel resolver dns c ares dns resolver ares cc ref dns resolving grpc ares wrapper cc c ares resolver request c ares address sorting input grpc ares wrapper cc c ares resolver request c ares address sorting output ev posix cc polling api pollset set add pollset set ev epollex linux cc pss merge ev epollex linux cc pss parent to client channel cc chand creat too much work for ed resolving lb policy ref counted h stream refcount home rcheppudira new grpc grpc src core lib transport transport h unref trytoconnect work serializer cc workserializer drainqueue work serializer cc queue drained work serializer cc workserializer run scheduling callback work serializer cc executing immediately resolving lb policy cc resolving lb got resolver result ref counted h lb policy refcount ref ev posix cc polling api pollset set create resolving lb policy cc resolving lb created new lb policy ev posix cc polling api pollset set add pollset set ev epollex linux cc pss merge ev epollex linux cc pss parent to timer generic cc timer check begin now next tls min glob min timer generic cc shard min deadline timer generic cc shard heap empty true timer generic cc shard queue deadline cap timer generic cc shard popped timer generic cc result shard min deadline now timer generic cc shard heap empty true timer generic cc shard queue deadline cap timer generic cc shard popped timer generic cc result shard min deadline now resolving lb policy cc resolving lb updating child policy child policy handler cc creating new child policy pick first ref counted h lb policy refcount home rcheppudira new grpc grpc src core ext filters client channel lb policy child policy handler cc ref helper ev posix cc polling api pollset set create pick first cc pick first created child policy handler cc created new lb policy pick first ev posix cc polling api pollset set add pollset set ev epollex linux cc pss merge ev epollex linux cc pss parent to child policy handler cc updating child policy pick first cc pick first received update with addresses subchannel list h creating subchannel list for subchannels ref counted h subchannel pool ref ref counted h subchannel pool ref timer generic cc shard heap empty true timer generic cc shard queue deadline cap timer generic cc shard popped timer generic cc result shard min deadline now timer generic cc shard heap empty true timer generic cc shard queue deadline cap timer generic cc shard popped timer generic cc result shard min deadline now timer generic cc timer check end r next timer manager cc sleep for a milliseconds ref counted h subchannel pool ref ref counted h subchannel pool ref ref counted h security connector refcount home rcheppudira new grpc grpc src core lib security security connector security connector cc ref connector arg copy ref counted h security connector refcount home rcheppudira new grpc grpc src core ext transport client secure secure channel create cc unref lb channel create ref counted h subchannel pool unref ref counted h subchannel pool unref ref counted h security connector refcount home rcheppudira new grpc grpc src core lib security security connector security connector cc ref connector arg copy ref counted h subchannel pool ref ev posix cc polling api pollset set create ref counted h subchannel pool ref ref counted h security connector refcount home rcheppudira new grpc grpc src core lib security security connector security connector cc ref connector arg copy ref counted h security connector refcount home rcheppudira new grpc grpc src core lib security security connector security connector cc ref connector arg copy ref counted h subchannel pool ref ref counted h subchannel pool ref ref counted h subchannel pool unref ref counted h security connector refcount home rcheppudira new grpc grpc src core lib security security connector security connector cc unref connector arg destroy ref counted h subchannel pool unref client channel cc chand creating subchannel wrapper for subchannel ref counted h stream refcount home rcheppudira new grpc grpc src core lib transport transport h ref subchannelwrapper subchannel list h subchannel list index created subchannel for address uri ref counted h lb policy refcount home rcheppudira new grpc grpc src core ext filters client channel lb policy pick first pick first cc ref subchannel list subchannel list h subchannel list index of subchannel starting watch from idle ref counted h pick first home rcheppudira new grpc grpc src core ext filters client channel lb policy subchannel list h ref watcher ref counted h client channel routing home rcheppudira new grpc grpc src core ext filters client channel client channel cc ref watcherwrapper ev posix cc polling api pollset set add pollset set ev epollex linux cc pss merge ev epollex linux cc pss parent to socket utils common posix cc enabling tcp user timeout with a timeout of ms ev posix cc polling api fd create tcp client ev posix cc fd trace fd create tcp client ev posix cc polling api pollset set add fd ev epollex linux cc pss add fd ev epollex linux cc add fd to pollable ref counted h subchannel pool ref ref counted h security connector refcount home rcheppudira new grpc grpc src core lib security security connector security connector cc ref connector arg copy tcp client posix c too much work for c client connect asynchronously connecting fd timer generic cc timer set now call timer generic cc add to shard with queue deadline cap is first timer false ref counted h resolver refcount home rcheppudira new grpc grpc src core ext filters client channel resolver dns c ares dns resolver ares cc unref dns resolving work serializer cc workserializer drainqueue work serializer cc queue drained client channel cc chand connectivity change for subchannel wrapper subchannel hopping into work serializer work serializer cc workserializer run scheduling callback work serializer cc executing immediately client channel cc chand processing connectivity change in work serializer for subchannel wrapper subchannel watcher subchannel list h subchannel list index of subchannel connectivity changed state connecting shutting down pending watcher ref counted h lb policy refcount home rcheppudira new grpc grpc src core ext filters client channel lb policy pick first pick first cc ref queuepicker client channel cc chand update state connecting picker ref counted h lb policy refcount home rcheppudira new grpc grpc src core ext filters client channel lb policy h unref queuepicker work serializer cc workserializer drainqueue work serializer cc queue drained init cc grpc init void completion queue cc grpc completion queue create internal completion type polling type ev posix cc polling api pollset init ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc ref pollset channel connectivity cc grpc channel watch connectivity state channel last observed state deadline gpr timespec tv sec tv nsec clock type cq tag ref counted h stream refcount home rcheppudira new grpc grpc src core lib transport transport h ref watch channel connectivity ev posix cc polling api pollset set add pollset ev epollex linux cc pss add pollset ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc ref pollset as multipollable ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc unref pollset ev epollex linux cc fd trace pollable create created epfd type ev epollex linux cc ps active pollable transition from empty to multi ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc ref pollset set ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc unref pollset as multipollable ev epollex linux cc add fd to pollable ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc unref pollset set ref counted h stream refcount home rcheppudira new grpc grpc src core lib transport transport h ref externalconnectivitywatcher work serializer cc workserializer run scheduling callback work serializer cc executing immediately timer generic cc timer set now call timer generic cc add to shard with queue deadline cap is first timer false connectivity state cc connectivitystatetracker client channel add watcher connectivity state cc connectivitystatetracker client channel notifying watcher idle connecting work serializer cc workserializer run scheduling callback work serializer cc scheduling on queue item work serializer cc workserializer drainqueue work serializer cc running item callback scheduled at connectivity state cc connectivitystatetracker client channel remove watcher ev posix cc polling api pollset set del pollset ev epollex linux cc pss del pollset ev epollex linux cc ps pollable maybe finish shutdown sc nil target null rw nil target null cpsc target ref counted h stream refcount home rcheppudira new grpc grpc src core lib transport transport h unref externalconnectivitywatcher work serializer cc workserializer drainqueue work serializer cc queue drained timer generic cc timer cancel pending true completion queue cc cq end op for next cq tag error no error done done arg storage ev posix cc polling api pollset kick nil too much work for ev epollex linux cc ps kick nil tls pollset nil tls worker nil pollset root worker nil ev epollex linux cc ps kicked any without poller completion queue cc grpc completion queue next cq deadline gpr timespec tv sec tv nsec clock type reserved nil ref counted h stream refcount home rcheppudira new grpc grpc src core lib transport transport h unref watch channel connectivity completion queue cc return event op complete tag ok completion queue cc grpc completion queue destroy cq completion queue cc grpc completion queue shutdown cq ev posix cc polling api pollset shutdown ev epollex linux cc ps pollable maybe finish shutdown sc target null rw nil target null cpsc target ev posix cc polling api pollset destroy ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc unref pollset ev epollex linux cc fd trace pollable unref closing epfd init cc grpc shutdown void channel connectivity cc grpc channel check connectivity state channel try to connect connectivity state cc connectivitystatetracker client channel get current state connecting init cc grpc init void completion queue cc grpc completion queue create internal completion type polling type ev posix cc polling api pollset init ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc ref pollset channel connectivity cc grpc channel watch connectivity state channel last observed state deadline gpr timespec tv sec tv nsec clock type cq tag ref counted h stream refcount home rcheppudira new grpc grpc src core lib transport transport h ref watch channel connectivity ev posix cc polling api pollset set add pollset ev epollex linux cc pss add pollset ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc ref pollset as multipollable ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc unref pollset ev epollex linux cc fd trace pollable create created epfd type ev epollex linux cc ps active pollable transition from empty to multi ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc ref pollset set ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc unref pollset as multipollable ev epollex linux cc add fd to pollable ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc unref pollset set ref counted h stream refcount home rcheppudira new grpc grpc src core lib transport transport h ref externalconnectivitywatcher work serializer cc workserializer run scheduling callback work serializer cc executing immediately timer generic cc timer set now call timer generic cc add to shard with queue deadline cap is first timer false connectivity state cc connectivitystatetracker client channel add watcher work serializer cc workserializer drainqueue work serializer cc queue drained completion queue cc grpc completion queue next cq deadline gpr timespec tv sec tv nsec clock type reserved nil ev posix cc polling api pollset work begin ev epollex linux cc ps work hdl nil worker now deadline kwp pollable ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc ref pollset worker ev epollex linux cc pollable poll for ev epollex linux cc pollable got events ev epollex linux cc ps got fd cancel read write tcp client posix cc client connect on writable error no error timer generic cc timer cancel pending true ev posix cc polling api pollset set del fd ev epollex linux cc pss del fd combiner cc c create ref counted h tcp home rcheppudira new grpc grpc src core lib iomgr tcp posix cc ref error tracking tcp client posix cc client connect on alarm error cancelled ref counted h subchannel pool unref ref counted h security connector refcount home rcheppudira new grpc grpc src core lib security security connector security connector cc unref connector arg destroy handshaker cc handshake manager adding handshaker http connect at index ssl transport security cc handshake start before ssl initialization pinit ssl transport security cc loop before ssl initialization pinit ssl transport security cc loop tls write client hello twch ref counted h security connector refcount home rcheppudira new grpc grpc src core lib security transport security handshaker cc ref handshake handshaker cc handshake manager adding handshaker security at index ev posix cc polling api pollset set add fd ev epollex linux cc pss add fd ev epollex linux cc add fd to pollable ev epollex linux cc add fd to pollable ref counted h subchannel pool ref ref counted h security connector refcount home rcheppudira new grpc grpc src core lib security security connector security connector cc ref connector arg copy timer generic cc timer set now call timer generic cc add to shard with queue deadline cap is first timer false handshaker cc handshake manager error no error shutdown index args endpoint args size grpc primary user agent grpc c grpc keepalive time ms grpc keepalive timeout ms grpc ssl target name override infinera com grpc client channel factory grpc channel credentials grpc server uri dns grpc default authority infinera com grpc subchannel pool grpc scheme https grpc security connector grpc subchannel address read buffer length exit early handshaker cc handshake manager calling handshaker http connect at index handshaker cc handshake manager error no error shutdown index args endpoint args size grpc primary user agent grpc c grpc keepalive time ms grpc keepalive timeout ms grpc ssl target name override infinera com grpc client channel factory grpc channel credentials grpc server uri dns grpc default authority infinera com grpc subchannel pool grpc scheme https grpc security connector grpc subchannel address read buffer length exit early handshaker cc handshake manager calling handshaker security at index tcp posix cc write peer tcp posix cc write no error ref counted h tcp home rcheppudira new grpc grpc src core lib iomgr tcp posix cc ref read tcp posix cc tcp notify on read ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc unref pollset worker ev epollex linux cc ps pollable maybe finish shutdown sc nil target null rw nil target null cpsc target ev posix cc polling api pollset work end ev posix cc polling api pollset work begin ev epollex linux cc ps work hdl nil worker now deadline kwp pollable ref counted h pollable refcount home rcheppudira new grpc grpc src core lib iomgr ev epollex linux cc ref pollset worker ev epollex linux cc pollable poll for ev epollex linux cc pollable got events ev epollex linux cc ps got fd cancel read write tcp posix cc tcp got read no error tcp posix cc tcp alloc slices resource quota cc rq anonymous pool alloc free pool combiner cc c grpc combiner execute c last combiner cc c grpc combiner continue exec ctx contended exec ctx ready to finish time to execute final list combiner cc c maybe finish one n combiner cc c grpc combiner execute finally c ac combiner cc c finish old state combiner cc c grpc combiner continue exec ctx contended exec ctx ready to finish time to execute final list combiner cc c execute final c resource quota cc rq check allocation for user shutdown free pool resource quota cc rq anonymous pool grant alloc bytes rq free pool combiner cc c finish old state tcp posix cc tcp read allocation done no error tcp posix cc tcp call cb tcp posix cc read peer error no error ref counted h tcp home rcheppudira new grpc grpc src core lib iomgr tcp posix cc unref read ssl transport security cc loop tls write client hello twch ssl transport security cc loop tls read server hello trsh ssl transport security cc loop tls read server certific trsc ssl transport security cc loop tls read server key exch trske ssl transport security cc loop tls read server done trsd ssl transport security cc loop tls write client key exc twcke ssl transport security cc loop tls write change cipher twccs ssl transport security cc loop tls write finished twfin tcp posix cc write peer tcp posix cc write no error ref counted h tcp home rcheppudira new grpc grpc src core lib iomgr tcp posix cc ref read tcp posix cc tcp got read no error tcp posix cc tcp do read tcp posix cc tcp call cb tcp posix cc read peer error no error ref counted h tcp home rcheppudira new grpc grpc src core lib iomgr tcp posix cc unref read ssl transport security cc loop tls write finished twfin ssl transport security cc loop tls read server session trst ssl transport security cc loop tls read change cipher s trccs ssl transport security cc loop tls read finished trfin ssl transport security cc handshake done ssl negotiation finished succe sslok ssl transport security cc assertion failed int peer property count current insert index signal caught by process dumpstacktrace opt infinera thanos local lib libbasesystem so unsigned long backward details unwind callback backward stacktraceimpl callback unsigned long opt infinera thanos local lib libbasesystem so backward stacktraceimpl load here unsigned long opt infinera thanos local lib libbasesystem so util backtrace dumpstacktrace opt infinera thanos local lib libbasesystem so osprocess myfailurehandler int lib linux gnu libc so lib linux gnu libc so gsignal lib linux gnu libc so abort opt infinera thanos local lib libgrpc so opt infinera thanos local lib libgrpc so opt infinera thanos local lib libgrpc so tsi handshaker result extract peer tsi handshaker result const tsi peer opt infinera thanos local lib libgrpc so opt infinera thanos local lib libgrpc so opt infinera thanos local lib libgrpc so opt infinera thanos local lib libgrpc so opt infinera thanos local lib libgrpc so opt infinera thanos local lib libgrpc so grpc core execctx flush opt infinera thanos local lib libgrpc so opt infinera thanos local lib libgrpc so opt infinera thanos local lib libgrpc so grpc pollset work grpc pollset grpc pollset worker long opt infinera thanos local lib libgrpc so opt infinera thanos local lib libgrpc so grpc completion queue next opt infinera thanos local lib libgrpc so grpc impl completionqueue asyncnextinternal void bool gpr timespec opt infinera thanos local lib libgrpc so grpc impl completionqueue next void bool opt infinera thanos local lib libgrpc so grpc impl channel waitforstatechangeimpl grpc connectivity state gpr timespec opt infinera thanos local bin tunnelclient bool grpc channelinterface waitforstatechange grpc connectivity state gpr timespec opt infinera thanos local bin tunnelclient bool grpc channelinterface waitforconnected gpr timespec opt infinera thanos local bin tunnelclient connectionmanager createchannel std basic string std allocator const std shared ptr opt infinera thanos local bin tunnelclient connectionmanager connecttoserver opt infinera thanos local bin tunnelclient boost mfi operator connectionmanager const opt infinera thanos local bin tunnelclient void boost bi operator boost bi boost bi type boost mfi boost bi int opt infinera thanos local bin tunnelclient boost bi bind t boost bi operator opt infinera thanos local bin tunnelclient boost detail thread data boost bi run end dumpstacktrace ideally it must work fine or through some error but here it is getting crashed below is the code snippets std shared ptr connectionmanager createchannelcredentials grpc sslcredentialsoptions sslcredoptions std string rootcertpath var security trustedcerts rsa sha root crt std string clientkeypath var security keys rsa sha client key std string clientcertpath var security certs rsa sha client crt here cacert clientcert clientkey is of type string getfiledata rootcertpath cacert getfiledata will read and populate cert key from the provided loc sslcredoptions pem root certs cacert getfiledata clientkeypath clientkey sslcredoptions pem private key clientkey getfiledata clientcertpath clientcert sslcredoptions pem cert chain clientcert return grpc sslcredentials sslcredoptions here server std shared ptr connectionmanager createchannel const std string serveraddress std shared ptr channel std shared ptr channelcreds grpc insecurechannelcredentials channelcreds createchannelcredentials channelarguments channelarg std string san channelarg setint grpc arg keepalive time ms keepalivetime channelarg setint grpc arg keepalive timeout ms keepalivetime channelarg setssltargetnameoverride san channel grpc createcustomchannel serveraddress channelcreds channelarg std chrono time point deadline std chrono system clock now std chrono seconds if channel waitforconnected gpr time add gpr now gpr clock realtime gpr time from seconds gpr timespan channel getstate false grpc connectivity state grpc channel ready if channel waitforconnected deadline channel getstate false grpc connectivity state grpc channel ready it is crashing here notify and start some backoff logic else secure connection established return channel tested the same with below openssl utility and i can see in the pcap that handshake is happening successfully sudo openssl s client connect cert var security certs rsa sha client crt key var security keys rsa sha client key cafile var security trusted certs rsa sha root crt state debug
| 0
|
442,500
| 12,746,403,020
|
IssuesEvent
|
2020-06-26 15:52:46
|
hochschule-darmstadt/openartbrowser
|
https://api.github.com/repos/hochschule-darmstadt/openartbrowser
|
closed
|
Refactor split_languages.py
|
high priority
|
**Describe the bug**
The split_languages.py is very memory consuming since for each language one file is generated which is later written to a file. Because of that we get a memory error from python.
**Expected behavior**
The script should be more memory efficient.
|
1.0
|
Refactor split_languages.py - **Describe the bug**
The split_languages.py is very memory consuming since for each language one file is generated which is later written to a file. Because of that we get a memory error from python.
**Expected behavior**
The script should be more memory efficient.
|
non_test
|
refactor split languages py describe the bug the split languages py is very memory consuming since for each language one file is generated which is later written to a file because of that we get a memory error from python expected behavior the script should be more memory efficient
| 0
|
72,671
| 13,905,014,647
|
IssuesEvent
|
2020-10-20 09:19:52
|
Regalis11/Barotrauma
|
https://api.github.com/repos/Regalis11/Barotrauma
|
closed
|
Level generation: the target outpos is sometimes positioned too close to the level walls
|
Bug Code Need more info
|
Sometimes the target outpost is positioned so that it's difficult or even impossible to dock to the outpost, because the level walls are too close in the left. The case pictured below is not the worst, because here it's possible to dock, if you are really careful.


|
1.0
|
Level generation: the target outpos is sometimes positioned too close to the level walls - Sometimes the target outpost is positioned so that it's difficult or even impossible to dock to the outpost, because the level walls are too close in the left. The case pictured below is not the worst, because here it's possible to dock, if you are really careful.


|
non_test
|
level generation the target outpos is sometimes positioned too close to the level walls sometimes the target outpost is positioned so that it s difficult or even impossible to dock to the outpost because the level walls are too close in the left the case pictured below is not the worst because here it s possible to dock if you are really careful
| 0
|
237,435
| 19,634,116,217
|
IssuesEvent
|
2022-01-08 01:31:35
|
qixils/java-crowd-control
|
https://api.github.com/repos/qixils/java-crowd-control
|
closed
|
Minimize usage of builders/constructors inside assertThrows calls
|
tests
|
Spotted a method which was throwing the expected exception but for the wrong reason. This perhaps could be mitigated by introducing custom exceptions but for now the easiest solution is to verify that all such method calls are only using pre-built objects unless necessary.
|
1.0
|
Minimize usage of builders/constructors inside assertThrows calls - Spotted a method which was throwing the expected exception but for the wrong reason. This perhaps could be mitigated by introducing custom exceptions but for now the easiest solution is to verify that all such method calls are only using pre-built objects unless necessary.
|
test
|
minimize usage of builders constructors inside assertthrows calls spotted a method which was throwing the expected exception but for the wrong reason this perhaps could be mitigated by introducing custom exceptions but for now the easiest solution is to verify that all such method calls are only using pre built objects unless necessary
| 1
|
133,145
| 5,197,843,425
|
IssuesEvent
|
2017-01-23 16:36:00
|
openshift/origin
|
https://api.github.com/repos/openshift/origin
|
closed
|
oc cluster up don't work with --public-hostname
|
component/composition kind/enhancement priority/P2
|
#I'm creating a cluster and trying to bind it to a internal private ip that I have as additional loopback in my box.
```
$ oc cluster up --public-hostname '11.2.2.2' --host-data-dir '/Users/jmorales/.oc/profiles/test3/data' --host-config-dir '/Users/jmorales/.oc/profiles/test3/config' --routing-suffix 'apps.lcup' --use-existing-config
-- Checking OpenShift client ... OK
-- Checking Docker client ... OK
-- Checking Docker version ... OK
-- Checking for existing OpenShift container ... OK
-- Checking for openshift/origin:v1.4.0-rc1 image ... OK
-- Checking Docker daemon configuration ... OK
-- Checking for available ports ... OK
-- Checking type of volume mount ...
Using Docker shared volumes for OpenShift volumes
-- Creating host directories ... OK
-- Finding server IP ...
Using public hostname IP 11.2.2.2 as the host IP
Using 127.0.0.1 as the server IP
-- Starting OpenShift container ...
Creating initial OpenShift configuration
Starting OpenShift using container 'origin'
Waiting for API server to start listening
**FAIL
Error: cannot access master readiness URL https://127.0.0.1:8443/healthz/ready
Details:
Last 10 lines of "origin" container log:
I1223 13:02:46.179272 5153 ensure.go:209] Created default security context constraint restricted
I1223 13:02:46.181262 5153 ensure.go:209] Created default security context constraint anyuid
I1223 13:02:46.182939 5153 ensure.go:209] Created default security context constraint hostnetwork
I1223 13:02:46.963136 5153 trace.go:61] Trace "Create /oapi/v1/namespaces/default/rolebindings" (started 2016-12-23 13:02:46.194574735 +0000 UTC):
[15.288µs] [15.288µs] About to convert to expected version
[36.073µs] [20.785µs] Conversion done
[765.1238ms] [765.087727ms] About to store object in database
[768.481978ms] [3.358178ms] Object stored in database
[768.494868ms] [12.89µs] Self-link added
[768.530741ms] [35.873µs] END**
[INFO] Cluster created succesfully
```
##### Version
```
$ oc version
oc v1.4.0-rc1+b4e0954
kubernetes v1.4.0+776c994
features: Basic-Auth
```
##### Steps To Reproduce
1. Create an internal loopback interface
2. Start the cluster binding to that internal loopback interface
##### Current Result
Cluster start succesful
##### Expected Result
Error
|
1.0
|
oc cluster up don't work with --public-hostname - #I'm creating a cluster and trying to bind it to a internal private ip that I have as additional loopback in my box.
```
$ oc cluster up --public-hostname '11.2.2.2' --host-data-dir '/Users/jmorales/.oc/profiles/test3/data' --host-config-dir '/Users/jmorales/.oc/profiles/test3/config' --routing-suffix 'apps.lcup' --use-existing-config
-- Checking OpenShift client ... OK
-- Checking Docker client ... OK
-- Checking Docker version ... OK
-- Checking for existing OpenShift container ... OK
-- Checking for openshift/origin:v1.4.0-rc1 image ... OK
-- Checking Docker daemon configuration ... OK
-- Checking for available ports ... OK
-- Checking type of volume mount ...
Using Docker shared volumes for OpenShift volumes
-- Creating host directories ... OK
-- Finding server IP ...
Using public hostname IP 11.2.2.2 as the host IP
Using 127.0.0.1 as the server IP
-- Starting OpenShift container ...
Creating initial OpenShift configuration
Starting OpenShift using container 'origin'
Waiting for API server to start listening
**FAIL
Error: cannot access master readiness URL https://127.0.0.1:8443/healthz/ready
Details:
Last 10 lines of "origin" container log:
I1223 13:02:46.179272 5153 ensure.go:209] Created default security context constraint restricted
I1223 13:02:46.181262 5153 ensure.go:209] Created default security context constraint anyuid
I1223 13:02:46.182939 5153 ensure.go:209] Created default security context constraint hostnetwork
I1223 13:02:46.963136 5153 trace.go:61] Trace "Create /oapi/v1/namespaces/default/rolebindings" (started 2016-12-23 13:02:46.194574735 +0000 UTC):
[15.288µs] [15.288µs] About to convert to expected version
[36.073µs] [20.785µs] Conversion done
[765.1238ms] [765.087727ms] About to store object in database
[768.481978ms] [3.358178ms] Object stored in database
[768.494868ms] [12.89µs] Self-link added
[768.530741ms] [35.873µs] END**
[INFO] Cluster created succesfully
```
##### Version
```
$ oc version
oc v1.4.0-rc1+b4e0954
kubernetes v1.4.0+776c994
features: Basic-Auth
```
##### Steps To Reproduce
1. Create an internal loopback interface
2. Start the cluster binding to that internal loopback interface
##### Current Result
Cluster start succesful
##### Expected Result
Error
|
non_test
|
oc cluster up don t work with public hostname i m creating a cluster and trying to bind it to a internal private ip that i have as additional loopback in my box oc cluster up public hostname host data dir users jmorales oc profiles data host config dir users jmorales oc profiles config routing suffix apps lcup use existing config checking openshift client ok checking docker client ok checking docker version ok checking for existing openshift container ok checking for openshift origin image ok checking docker daemon configuration ok checking for available ports ok checking type of volume mount using docker shared volumes for openshift volumes creating host directories ok finding server ip using public hostname ip as the host ip using as the server ip starting openshift container creating initial openshift configuration starting openshift using container origin waiting for api server to start listening fail error cannot access master readiness url details last lines of origin container log ensure go created default security context constraint restricted ensure go created default security context constraint anyuid ensure go created default security context constraint hostnetwork trace go trace create oapi namespaces default rolebindings started utc about to convert to expected version conversion done about to store object in database object stored in database self link added end cluster created succesfully version oc version oc kubernetes features basic auth steps to reproduce create an internal loopback interface start the cluster binding to that internal loopback interface current result cluster start succesful expected result error
| 0
|
607,678
| 18,788,523,435
|
IssuesEvent
|
2021-11-08 14:36:22
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.goal.com - see bug description
|
priority-normal browser-fenix engine-gecko
|
<!-- @browser: Firefox Mobile 96.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:96.0) Gecko/96.0 Firefox/96.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/92480 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.goal.com/en/lists/ballon-dor-2021-power-rankings/1l03egk461k3a1em4lb6gh8mhu#cs3e22030e352c0b4a
**Browser / Version**: Firefox Mobile 96.0
**Operating System**: Android 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: back button does not work and pressing it reloads the page and it sends me on random points of the site
**Steps to Reproduce**:
make the back button work as intended and dont make me go to the every stage of the site I have been in
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/11/9e82ba9e-ba10-4e38-bb55-c064707b99d8.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20211104094642</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/11/6db835d1-6bfb-4cd2-a93b-f6ff4908f639)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.goal.com - see bug description - <!-- @browser: Firefox Mobile 96.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:96.0) Gecko/96.0 Firefox/96.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/92480 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.goal.com/en/lists/ballon-dor-2021-power-rankings/1l03egk461k3a1em4lb6gh8mhu#cs3e22030e352c0b4a
**Browser / Version**: Firefox Mobile 96.0
**Operating System**: Android 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: back button does not work and pressing it reloads the page and it sends me on random points of the site
**Steps to Reproduce**:
make the back button work as intended and dont make me go to the every stage of the site I have been in
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/11/9e82ba9e-ba10-4e38-bb55-c064707b99d8.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20211104094642</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/11/6db835d1-6bfb-4cd2-a93b-f6ff4908f639)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description back button does not work and pressing it reloads the page and it sends me on random points of the site steps to reproduce make the back button work as intended and dont make me go to the every stage of the site i have been in view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
211,304
| 23,805,548,326
|
IssuesEvent
|
2022-09-04 01:09:42
|
joaonetoo/libeasy
|
https://api.github.com/repos/joaonetoo/libeasy
|
opened
|
WS-2022-0284 (Medium) detected in moment-timezone-0.5.14.tgz
|
security vulnerability
|
## WS-2022-0284 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moment-timezone-0.5.14.tgz</b></p></summary>
<p>Parse and display moments in any timezone.</p>
<p>Library home page: <a href="https://registry.npmjs.org/moment-timezone/-/moment-timezone-0.5.14.tgz">https://registry.npmjs.org/moment-timezone/-/moment-timezone-0.5.14.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/moment-timezone/package.json</p>
<p>
Dependency Hierarchy:
- sequelize-4.37.6.tgz (Root Library)
- :x: **moment-timezone-0.5.14.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/joaonetoo/libeasy/commit/a5ba48c09c952b045a2b0a710642c33a2b86164c">a5ba48c09c952b045a2b0a710642c33a2b86164c</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cleartext Transmission of Sensitive Information in moment-timezone
<p>Publish Date: 2022-08-30
<p>URL: <a href=https://github.com/moment/moment-timezone/commit/7915ac567ab19700e44ad6b5d8ef0b85e48a9e75>WS-2022-0284</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-v78c-4p63-2j6c">https://github.com/advisories/GHSA-v78c-4p63-2j6c</a></p>
<p>Release Date: 2022-08-30</p>
<p>Fix Resolution (moment-timezone): 0.5.35</p>
<p>Direct dependency fix Resolution (sequelize): 6.16.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2022-0284 (Medium) detected in moment-timezone-0.5.14.tgz - ## WS-2022-0284 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moment-timezone-0.5.14.tgz</b></p></summary>
<p>Parse and display moments in any timezone.</p>
<p>Library home page: <a href="https://registry.npmjs.org/moment-timezone/-/moment-timezone-0.5.14.tgz">https://registry.npmjs.org/moment-timezone/-/moment-timezone-0.5.14.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/moment-timezone/package.json</p>
<p>
Dependency Hierarchy:
- sequelize-4.37.6.tgz (Root Library)
- :x: **moment-timezone-0.5.14.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/joaonetoo/libeasy/commit/a5ba48c09c952b045a2b0a710642c33a2b86164c">a5ba48c09c952b045a2b0a710642c33a2b86164c</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cleartext Transmission of Sensitive Information in moment-timezone
<p>Publish Date: 2022-08-30
<p>URL: <a href=https://github.com/moment/moment-timezone/commit/7915ac567ab19700e44ad6b5d8ef0b85e48a9e75>WS-2022-0284</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-v78c-4p63-2j6c">https://github.com/advisories/GHSA-v78c-4p63-2j6c</a></p>
<p>Release Date: 2022-08-30</p>
<p>Fix Resolution (moment-timezone): 0.5.35</p>
<p>Direct dependency fix Resolution (sequelize): 6.16.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
ws medium detected in moment timezone tgz ws medium severity vulnerability vulnerable library moment timezone tgz parse and display moments in any timezone library home page a href path to dependency file package json path to vulnerable library node modules moment timezone package json dependency hierarchy sequelize tgz root library x moment timezone tgz vulnerable library found in head commit a href vulnerability details cleartext transmission of sensitive information in moment timezone publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution moment timezone direct dependency fix resolution sequelize step up your open source security game with mend
| 0
|
76,431
| 15,496,025,669
|
IssuesEvent
|
2021-03-11 01:55:47
|
jatcode/hookstictactoe
|
https://api.github.com/repos/jatcode/hookstictactoe
|
opened
|
CVE-2020-7789 (Medium) detected in node-notifier-5.4.0.tgz
|
security vulnerability
|
## CVE-2020-7789 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-notifier-5.4.0.tgz</b></p></summary>
<p>A Node.js module for sending notifications on native Mac, Windows (post and pre 8) and Linux (or Growl as fallback)</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-notifier/-/node-notifier-5.4.0.tgz">https://registry.npmjs.org/node-notifier/-/node-notifier-5.4.0.tgz</a></p>
<p>Path to dependency file: /hookstictactoe/package.json</p>
<p>Path to vulnerable library: hookstictactoe/node_modules/node-notifier/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-2.1.8.tgz (Root Library)
- jest-23.6.0.tgz
- jest-cli-23.6.0.tgz
- :x: **node-notifier-5.4.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package node-notifier before 9.0.0. It allows an attacker to run arbitrary commands on Linux machines due to the options params not being sanitised when being passed an array.
<p>Publish Date: 2020-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7789>CVE-2020-7789</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7789">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7789</a></p>
<p>Release Date: 2020-12-11</p>
<p>Fix Resolution: 9.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7789 (Medium) detected in node-notifier-5.4.0.tgz - ## CVE-2020-7789 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-notifier-5.4.0.tgz</b></p></summary>
<p>A Node.js module for sending notifications on native Mac, Windows (post and pre 8) and Linux (or Growl as fallback)</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-notifier/-/node-notifier-5.4.0.tgz">https://registry.npmjs.org/node-notifier/-/node-notifier-5.4.0.tgz</a></p>
<p>Path to dependency file: /hookstictactoe/package.json</p>
<p>Path to vulnerable library: hookstictactoe/node_modules/node-notifier/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-2.1.8.tgz (Root Library)
- jest-23.6.0.tgz
- jest-cli-23.6.0.tgz
- :x: **node-notifier-5.4.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package node-notifier before 9.0.0. It allows an attacker to run arbitrary commands on Linux machines due to the options params not being sanitised when being passed an array.
<p>Publish Date: 2020-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7789>CVE-2020-7789</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7789">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7789</a></p>
<p>Release Date: 2020-12-11</p>
<p>Fix Resolution: 9.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in node notifier tgz cve medium severity vulnerability vulnerable library node notifier tgz a node js module for sending notifications on native mac windows post and pre and linux or growl as fallback library home page a href path to dependency file hookstictactoe package json path to vulnerable library hookstictactoe node modules node notifier package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz x node notifier tgz vulnerable library vulnerability details this affects the package node notifier before it allows an attacker to run arbitrary commands on linux machines due to the options params not being sanitised when being passed an array publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
43,802
| 5,561,693,834
|
IssuesEvent
|
2017-03-24 23:40:50
|
golang/go
|
https://api.github.com/repos/golang/go
|
closed
|
net/http: ptrace protections result in hung sendfile test
|
HelpWanted Testing
|
On this revision:
```
commit 886b02d705ffb1be8b4974ac4c355d480a24e3ec
Author: Andrew Gerrand <adg@golang.org>
Date: Thu Jan 15 15:21:14 2015 +1100
[release-branch.go1.4] go1.4.1
Change-Id: If275a5caa07cfd16b7052ad50709e1d0f1258223
Reviewed-on: https://go-review.googlesource.com/2856
Reviewed-by: Rob Pike <r@golang.org>
```
Run `all.bash`, one of the testcases failed:
```
panic: test timed out after 2m0s
goroutine 264 [running]:
testing.func·008()
/home/david/go-git/src/testing/testing.go:681 +0x12f
created by time.goFunc
/home/david/go-git/src/time/sleep.go:129 +0x4b
goroutine 1 [chan receive]:
testing.RunTests(0x8fcff8, 0xa371e0, 0xd2, 0xd2, 0xc20804c001)
/home/david/go-git/src/testing/testing.go:556 +0xad6
testing.(*M).Run(0xc20804c0f0, 0x0)
/home/david/go-git/src/testing/testing.go:485 +0x6c
net/http_test.TestMain(0xc20804c0f0)
/home/david/go-git/src/net/http/main_test.go:19 +0x28
main.main()
net/http/_test/_testmain.go:510 +0x1d1
goroutine 262 [IO wait]:
net.(*pollDesc).Wait(0xc20821d790, 0x72, 0x0, 0x0)
/home/david/go-git/src/net/fd_poll_runtime.go:84 +0x47
net.(*pollDesc).WaitRead(0xc20821d790, 0x0, 0x0)
/home/david/go-git/src/net/fd_poll_runtime.go:89 +0x43
net.(*netFD).Read(0xc20821d730, 0xc2080c2000, 0x1000, 0x1000, 0x0, 0x7ff8d880c070, 0xc20819e4b8)
/home/david/go-git/src/net/fd_unix.go:242 +0x40f
net.(*conn).Read(0xc208032248, 0xc2080c2000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
/home/david/go-git/src/net/net.go:121 +0xdc
net/http.noteEOFReader.Read(0x7ff8d880f2e0, 0xc208032248, 0xc20817a7e8, 0xc2080c2000, 0x1000, 0x1000, 0x7ff8d87fa010, 0x0, 0x0)
/home/david/go-git/src/net/http/transport.go:1270 +0x6e
net/http.(*noteEOFReader).Read(0xc208181f80, 0xc2080c2000, 0x1000, 0x1000, 0xc207ffcdda, 0x0, 0x0)
<autogenerated>:125 +0xd4
bufio.(*Reader).fill(0xc2080a1980)
/home/david/go-git/src/bufio/bufio.go:97 +0x1ce
bufio.(*Reader).Peek(0xc2080a1980, 0x1, 0x0, 0x0, 0x0, 0x0, 0x0)
/home/david/go-git/src/bufio/bufio.go:132 +0xf0
net/http.(*persistConn).readLoop(0xc20817a790)
/home/david/go-git/src/net/http/transport.go:842 +0xa4
created by net/http.(*Transport).dialConn
/home/david/go-git/src/net/http/transport.go:660 +0xc9f
goroutine 259 [select]:
net/http.(*persistConn).roundTrip(0xc20817a790, 0xc20819e420, 0x0, 0x0, 0x0)
/home/david/go-git/src/net/http/transport.go:1082 +0x7ad
net/http.(*Transport).RoundTrip(0xc208076000, 0xc2081fd1e0, 0xc208181e60, 0x0, 0x0)
/home/david/go-git/src/net/http/transport.go:235 +0x558
net/http.send(0xc2081fd1e0, 0x7ff8d880c468, 0xc208076000, 0x17, 0x0, 0x0)
/home/david/go-git/src/net/http/client.go:219 +0x4fc
net/http.(*Client).send(0xa395a0, 0xc2081fd1e0, 0x17, 0x0, 0x0)
/home/david/go-git/src/net/http/client.go:142 +0x15b
net/http.(*Client).doFollowingRedirects(0xa395a0, 0xc2081fd1e0, 0x8fd318, 0x0, 0x0, 0x0)
/home/david/go-git/src/net/http/client.go:367 +0xb25
net/http.(*Client).Get(0xa395a0, 0xc208181e00, 0x17, 0xc2081fc0d0, 0x0, 0x0)
/home/david/go-git/src/net/http/client.go:299 +0xba
net/http.Get(0xc208181e00, 0x17, 0xc20805be58, 0x0, 0x0)
/home/david/go-git/src/net/http/client.go:276 +0x50
net/http_test.TestLinuxSendfile(0xc2080774d0)
/home/david/go-git/src/net/http/fs_test.go:822 +0xb14
testing.tRunner(0xc2080774d0, 0xa378a0)
/home/david/go-git/src/testing/testing.go:447 +0xbf
created by testing.RunTests
/home/david/go-git/src/testing/testing.go:555 +0xa8b
goroutine 263 [select]:
net/http.(*persistConn).writeLoop(0xc20817a790)
/home/david/go-git/src/net/http/transport.go:945 +0x41d
created by net/http.(*Transport).dialConn
/home/david/go-git/src/net/http/transport.go:661 +0xcbc
FAIL net/http 120.018s
```
|
1.0
|
net/http: ptrace protections result in hung sendfile test - On this revision:
```
commit 886b02d705ffb1be8b4974ac4c355d480a24e3ec
Author: Andrew Gerrand <adg@golang.org>
Date: Thu Jan 15 15:21:14 2015 +1100
[release-branch.go1.4] go1.4.1
Change-Id: If275a5caa07cfd16b7052ad50709e1d0f1258223
Reviewed-on: https://go-review.googlesource.com/2856
Reviewed-by: Rob Pike <r@golang.org>
```
Run `all.bash`, one of the testcases failed:
```
panic: test timed out after 2m0s
goroutine 264 [running]:
testing.func·008()
/home/david/go-git/src/testing/testing.go:681 +0x12f
created by time.goFunc
/home/david/go-git/src/time/sleep.go:129 +0x4b
goroutine 1 [chan receive]:
testing.RunTests(0x8fcff8, 0xa371e0, 0xd2, 0xd2, 0xc20804c001)
/home/david/go-git/src/testing/testing.go:556 +0xad6
testing.(*M).Run(0xc20804c0f0, 0x0)
/home/david/go-git/src/testing/testing.go:485 +0x6c
net/http_test.TestMain(0xc20804c0f0)
/home/david/go-git/src/net/http/main_test.go:19 +0x28
main.main()
net/http/_test/_testmain.go:510 +0x1d1
goroutine 262 [IO wait]:
net.(*pollDesc).Wait(0xc20821d790, 0x72, 0x0, 0x0)
/home/david/go-git/src/net/fd_poll_runtime.go:84 +0x47
net.(*pollDesc).WaitRead(0xc20821d790, 0x0, 0x0)
/home/david/go-git/src/net/fd_poll_runtime.go:89 +0x43
net.(*netFD).Read(0xc20821d730, 0xc2080c2000, 0x1000, 0x1000, 0x0, 0x7ff8d880c070, 0xc20819e4b8)
/home/david/go-git/src/net/fd_unix.go:242 +0x40f
net.(*conn).Read(0xc208032248, 0xc2080c2000, 0x1000, 0x1000, 0x0, 0x0, 0x0)
/home/david/go-git/src/net/net.go:121 +0xdc
net/http.noteEOFReader.Read(0x7ff8d880f2e0, 0xc208032248, 0xc20817a7e8, 0xc2080c2000, 0x1000, 0x1000, 0x7ff8d87fa010, 0x0, 0x0)
/home/david/go-git/src/net/http/transport.go:1270 +0x6e
net/http.(*noteEOFReader).Read(0xc208181f80, 0xc2080c2000, 0x1000, 0x1000, 0xc207ffcdda, 0x0, 0x0)
<autogenerated>:125 +0xd4
bufio.(*Reader).fill(0xc2080a1980)
/home/david/go-git/src/bufio/bufio.go:97 +0x1ce
bufio.(*Reader).Peek(0xc2080a1980, 0x1, 0x0, 0x0, 0x0, 0x0, 0x0)
/home/david/go-git/src/bufio/bufio.go:132 +0xf0
net/http.(*persistConn).readLoop(0xc20817a790)
/home/david/go-git/src/net/http/transport.go:842 +0xa4
created by net/http.(*Transport).dialConn
/home/david/go-git/src/net/http/transport.go:660 +0xc9f
goroutine 259 [select]:
net/http.(*persistConn).roundTrip(0xc20817a790, 0xc20819e420, 0x0, 0x0, 0x0)
/home/david/go-git/src/net/http/transport.go:1082 +0x7ad
net/http.(*Transport).RoundTrip(0xc208076000, 0xc2081fd1e0, 0xc208181e60, 0x0, 0x0)
/home/david/go-git/src/net/http/transport.go:235 +0x558
net/http.send(0xc2081fd1e0, 0x7ff8d880c468, 0xc208076000, 0x17, 0x0, 0x0)
/home/david/go-git/src/net/http/client.go:219 +0x4fc
net/http.(*Client).send(0xa395a0, 0xc2081fd1e0, 0x17, 0x0, 0x0)
/home/david/go-git/src/net/http/client.go:142 +0x15b
net/http.(*Client).doFollowingRedirects(0xa395a0, 0xc2081fd1e0, 0x8fd318, 0x0, 0x0, 0x0)
/home/david/go-git/src/net/http/client.go:367 +0xb25
net/http.(*Client).Get(0xa395a0, 0xc208181e00, 0x17, 0xc2081fc0d0, 0x0, 0x0)
/home/david/go-git/src/net/http/client.go:299 +0xba
net/http.Get(0xc208181e00, 0x17, 0xc20805be58, 0x0, 0x0)
/home/david/go-git/src/net/http/client.go:276 +0x50
net/http_test.TestLinuxSendfile(0xc2080774d0)
/home/david/go-git/src/net/http/fs_test.go:822 +0xb14
testing.tRunner(0xc2080774d0, 0xa378a0)
/home/david/go-git/src/testing/testing.go:447 +0xbf
created by testing.RunTests
/home/david/go-git/src/testing/testing.go:555 +0xa8b
goroutine 263 [select]:
net/http.(*persistConn).writeLoop(0xc20817a790)
/home/david/go-git/src/net/http/transport.go:945 +0x41d
created by net/http.(*Transport).dialConn
/home/david/go-git/src/net/http/transport.go:661 +0xcbc
FAIL net/http 120.018s
```
|
test
|
net http ptrace protections result in hung sendfile test on this revision commit author andrew gerrand date thu jan change id reviewed on reviewed by rob pike run all bash one of the testcases failed panic test timed out after goroutine testing func· home david go git src testing testing go created by time gofunc home david go git src time sleep go goroutine testing runtests home david go git src testing testing go testing m run home david go git src testing testing go net http test testmain home david go git src net http main test go main main net http test testmain go goroutine net polldesc wait home david go git src net fd poll runtime go net polldesc waitread home david go git src net fd poll runtime go net netfd read home david go git src net fd unix go net conn read home david go git src net net go net http noteeofreader read home david go git src net http transport go net http noteeofreader read bufio reader fill home david go git src bufio bufio go bufio reader peek home david go git src bufio bufio go net http persistconn readloop home david go git src net http transport go created by net http transport dialconn home david go git src net http transport go goroutine net http persistconn roundtrip home david go git src net http transport go net http transport roundtrip home david go git src net http transport go net http send home david go git src net http client go net http client send home david go git src net http client go net http client dofollowingredirects home david go git src net http client go net http client get home david go git src net http client go net http get home david go git src net http client go net http test testlinuxsendfile home david go git src net http fs test go testing trunner home david go git src testing testing go created by testing runtests home david go git src testing testing go goroutine net http persistconn writeloop home david go git src net http transport go created by net http transport dialconn home david go git src net http transport go fail net http
| 1
|
594,145
| 18,024,870,120
|
IssuesEvent
|
2021-09-17 02:13:49
|
LcorpOfficial/LCC-Minecraft-Server
|
https://api.github.com/repos/LcorpOfficial/LCC-Minecraft-Server
|
closed
|
investigate customcrafting+autocrafter integration error after implementation
|
bug enhancement help wanted wontfix medium priority
|
**Describe the bug**
A clear and concise description of what the bug is.
```
[05:14:46 WARN]: [AutomatedCrafting] Task #9 for AutomatedCrafting v2.4 generated an exception
java.lang.NoSuchMethodError: 'me.wolfyscript.customcrafting.handlers.RecipeHandler me.wolfyscript.customcrafting.CustomCrafting.getRecipeHandler()'
at nl.dgoossens.autocraft.compat.CustomCraftingCompat.load(CustomCraftingCompat.java:25) ~[?:?]
at nl.dgoossens.autocraft.RecipeLoader.reload(RecipeLoader.java:97) ~[?:?]
at nl.dgoossens.autocraft.RecipeLoader.lambda$new$0(RecipeLoader.java:27) ~[?:?]
at org.bukkit.craftbukkit.v1_16_R3.scheduler.CraftTask.run(CraftTask.java:99) ~[patched_1.16.4.jar:git-Tuinity-"52f4d08"]
at org.bukkit.craftbukkit.v1_16_R3.scheduler.CraftScheduler.mainThreadHeartbeat(CraftScheduler.java:468) ~[patched_1.16.4.jar:git-Tuinity-"52f4d08"]
at net.minecraft.server.v1_16_R3.MinecraftServer.w(MinecraftServer.java:949) ~[patched_1.16.4.jar:git-Tuinity-"52f4d08"]
at net.minecraft.server.v1_16_R3.MinecraftServer.lambda$a$0(MinecraftServer.java:178) ~[patched_1.16.4.jar:git-Tuinity-"52f4d08"]
at java.lang.Thread.run(Thread.java:834) [?:?]
```
**To Reproduce**
Steps to reproduce the behavior:
1. Load the server
2. Panic
**Expected behavior**
A clear and concise description of what you expected to happen.
automatedcrafting should hook into customcrafting correctly to recieve recipes
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Minecraft Info (please complete the following information):**
Alphaship Config
**Additional context notes**
Add any other context about the problem here.
contact both devs
|
1.0
|
investigate customcrafting+autocrafter integration error after implementation - **Describe the bug**
A clear and concise description of what the bug is.
```
[05:14:46 WARN]: [AutomatedCrafting] Task #9 for AutomatedCrafting v2.4 generated an exception
java.lang.NoSuchMethodError: 'me.wolfyscript.customcrafting.handlers.RecipeHandler me.wolfyscript.customcrafting.CustomCrafting.getRecipeHandler()'
at nl.dgoossens.autocraft.compat.CustomCraftingCompat.load(CustomCraftingCompat.java:25) ~[?:?]
at nl.dgoossens.autocraft.RecipeLoader.reload(RecipeLoader.java:97) ~[?:?]
at nl.dgoossens.autocraft.RecipeLoader.lambda$new$0(RecipeLoader.java:27) ~[?:?]
at org.bukkit.craftbukkit.v1_16_R3.scheduler.CraftTask.run(CraftTask.java:99) ~[patched_1.16.4.jar:git-Tuinity-"52f4d08"]
at org.bukkit.craftbukkit.v1_16_R3.scheduler.CraftScheduler.mainThreadHeartbeat(CraftScheduler.java:468) ~[patched_1.16.4.jar:git-Tuinity-"52f4d08"]
at net.minecraft.server.v1_16_R3.MinecraftServer.w(MinecraftServer.java:949) ~[patched_1.16.4.jar:git-Tuinity-"52f4d08"]
at net.minecraft.server.v1_16_R3.MinecraftServer.lambda$a$0(MinecraftServer.java:178) ~[patched_1.16.4.jar:git-Tuinity-"52f4d08"]
at java.lang.Thread.run(Thread.java:834) [?:?]
```
**To Reproduce**
Steps to reproduce the behavior:
1. Load the server
2. Panic
**Expected behavior**
A clear and concise description of what you expected to happen.
automatedcrafting should hook into customcrafting correctly to recieve recipes
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Minecraft Info (please complete the following information):**
Alphaship Config
**Additional context notes**
Add any other context about the problem here.
contact both devs
|
non_test
|
investigate customcrafting autocrafter integration error after implementation describe the bug a clear and concise description of what the bug is task for automatedcrafting generated an exception java lang nosuchmethoderror me wolfyscript customcrafting handlers recipehandler me wolfyscript customcrafting customcrafting getrecipehandler at nl dgoossens autocraft compat customcraftingcompat load customcraftingcompat java at nl dgoossens autocraft recipeloader reload recipeloader java at nl dgoossens autocraft recipeloader lambda new recipeloader java at org bukkit craftbukkit scheduler crafttask run crafttask java at org bukkit craftbukkit scheduler craftscheduler mainthreadheartbeat craftscheduler java at net minecraft server minecraftserver w minecraftserver java at net minecraft server minecraftserver lambda a minecraftserver java at java lang thread run thread java to reproduce steps to reproduce the behavior load the server panic expected behavior a clear and concise description of what you expected to happen automatedcrafting should hook into customcrafting correctly to recieve recipes screenshots if applicable add screenshots to help explain your problem minecraft info please complete the following information alphaship config additional context notes add any other context about the problem here contact both devs
| 0
|
64,750
| 18,870,080,555
|
IssuesEvent
|
2021-11-13 02:37:45
|
svalinn/DAGMC
|
https://api.github.com/repos/svalinn/DAGMC
|
closed
|
Cannot find source file: .../src/pyne/pyne.cpp
|
Type: Defect
|
rev. 472ac87 fails like this when being configured with ```-DBUILD_STATIC_LIBS=OFF -DBUILD_TESTS=OFF```
Git submodule is present but it doesn't contain ```pyne.cpp```.
OS: FreeBSD 13
|
1.0
|
Cannot find source file: .../src/pyne/pyne.cpp - rev. 472ac87 fails like this when being configured with ```-DBUILD_STATIC_LIBS=OFF -DBUILD_TESTS=OFF```
Git submodule is present but it doesn't contain ```pyne.cpp```.
OS: FreeBSD 13
|
non_test
|
cannot find source file src pyne pyne cpp rev fails like this when being configured with dbuild static libs off dbuild tests off git submodule is present but it doesn t contain pyne cpp os freebsd
| 0
|
101,437
| 8,787,919,531
|
IssuesEvent
|
2018-12-20 20:18:22
|
Princeton-CDH/ppa-django
|
https://api.github.com/repos/Princeton-CDH/ppa-django
|
closed
|
As a user, I want my search results from the Title field to prioritize unstemmed matches and boost title over subtitle.
|
awaiting testing
|
## Notes for testing
- title field search should search on both title and subtitle
- title matches should be prioritized over subtitle
- unstemmed matches should be prioritized over stemmed
- should still work with multiple terms, quotes, etc.
|
1.0
|
As a user, I want my search results from the Title field to prioritize unstemmed matches and boost title over subtitle. - ## Notes for testing
- title field search should search on both title and subtitle
- title matches should be prioritized over subtitle
- unstemmed matches should be prioritized over stemmed
- should still work with multiple terms, quotes, etc.
|
test
|
as a user i want my search results from the title field to prioritize unstemmed matches and boost title over subtitle notes for testing title field search should search on both title and subtitle title matches should be prioritized over subtitle unstemmed matches should be prioritized over stemmed should still work with multiple terms quotes etc
| 1
|
644,732
| 20,986,104,710
|
IssuesEvent
|
2022-03-29 03:27:01
|
EspressoSystems/cape
|
https://api.github.com/repos/EspressoSystems/cape
|
opened
|
Wallet :name parameter
|
bug priority: high wallet security wallet API unplanned
|
The `:name` parameter is used directly as the wallet file name, so this should be changed to something assured to be a valid file name, e.g. no path separators and length supported across platforms Linux, MacOS, and Windows.
|
1.0
|
Wallet :name parameter - The `:name` parameter is used directly as the wallet file name, so this should be changed to something assured to be a valid file name, e.g. no path separators and length supported across platforms Linux, MacOS, and Windows.
|
non_test
|
wallet name parameter the name parameter is used directly as the wallet file name so this should be changed to something assured to be a valid file name e g no path separators and length supported across platforms linux macos and windows
| 0
|
113,918
| 9,668,680,688
|
IssuesEvent
|
2019-05-21 15:36:48
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: kv95/enc=false/nodes=1 failed
|
C-test-failure O-roachtest O-robot
|
SHA: https://github.com/cockroachdb/cockroach/commits/9671342fead0509bec0913bae4ae1f244660788e
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=kv95/enc=false/nodes=1 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1298500&tab=buildLog
```
The test failed on branch=release-19.1, cloud=gce:
cluster.go:1474,kv.go:69,cluster.go:1812,errgroup.go:57: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1298500-kv95-enc-false-nodes-1:2 -- ./workload run kv --init --splits=1000 --histograms=logs/stats.json --concurrency=64 --duration=10m --read-percent=95 {pgurl:1-1} returned:
stderr:
stdout:
I190521 11:04:27.021703 1 workload/workload.go:562 starting 1000 splits
Error: ALTER TABLE kv SPLIT AT VALUES (9214157878976512): pq: splits would be immediately discarded by merge queue; disable the merge queue first by running 'SET CLUSTER SETTING kv.range_merge.queue_enabled = false'
Error: ssh verbose log retained in /root/.roachprod/debug/ssh_35.237.155.229_2019-05-21T11:04:25Z: exit status 1
: exit status 1
cluster.go:1833,kv.go:72,kv.go:137,test.go:1251: Goexit() was called
```
|
2.0
|
roachtest: kv95/enc=false/nodes=1 failed - SHA: https://github.com/cockroachdb/cockroach/commits/9671342fead0509bec0913bae4ae1f244660788e
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=kv95/enc=false/nodes=1 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1298500&tab=buildLog
```
The test failed on branch=release-19.1, cloud=gce:
cluster.go:1474,kv.go:69,cluster.go:1812,errgroup.go:57: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1298500-kv95-enc-false-nodes-1:2 -- ./workload run kv --init --splits=1000 --histograms=logs/stats.json --concurrency=64 --duration=10m --read-percent=95 {pgurl:1-1} returned:
stderr:
stdout:
I190521 11:04:27.021703 1 workload/workload.go:562 starting 1000 splits
Error: ALTER TABLE kv SPLIT AT VALUES (9214157878976512): pq: splits would be immediately discarded by merge queue; disable the merge queue first by running 'SET CLUSTER SETTING kv.range_merge.queue_enabled = false'
Error: ssh verbose log retained in /root/.roachprod/debug/ssh_35.237.155.229_2019-05-21T11:04:25Z: exit status 1
: exit status 1
cluster.go:1833,kv.go:72,kv.go:137,test.go:1251: Goexit() was called
```
|
test
|
roachtest enc false nodes failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests enc false nodes pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on branch release cloud gce cluster go kv go cluster go errgroup go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity enc false nodes workload run kv init splits histograms logs stats json concurrency duration read percent pgurl returned stderr stdout workload workload go starting splits error alter table kv split at values pq splits would be immediately discarded by merge queue disable the merge queue first by running set cluster setting kv range merge queue enabled false error ssh verbose log retained in root roachprod debug ssh exit status exit status cluster go kv go kv go test go goexit was called
| 1
|
322,072
| 27,579,121,860
|
IssuesEvent
|
2023-03-08 15:03:51
|
onc-healthit/onc-certification-g10-test-kit
|
https://api.github.com/repos/onc-healthit/onc-certification-g10-test-kit
|
closed
|
Suppress warnings in Limited Access Launch for STU 2
|
g10-test-kit v3.5.0
|
The token response body test for the Limited Access App with SMART App Launch STU 2 displays a warning for all of the scopes which aren't granted. These warnings should be suppressed like they are for STU 1.
See: https://chat.fhir.org/#narrow/stream/179309-inferno/topic/Token.20Exchange.20response.20body.20warning
|
1.0
|
Suppress warnings in Limited Access Launch for STU 2 - The token response body test for the Limited Access App with SMART App Launch STU 2 displays a warning for all of the scopes which aren't granted. These warnings should be suppressed like they are for STU 1.
See: https://chat.fhir.org/#narrow/stream/179309-inferno/topic/Token.20Exchange.20response.20body.20warning
|
test
|
suppress warnings in limited access launch for stu the token response body test for the limited access app with smart app launch stu displays a warning for all of the scopes which aren t granted these warnings should be suppressed like they are for stu see
| 1
|
213,841
| 16,540,742,953
|
IssuesEvent
|
2021-05-27 16:28:25
|
rapidsai/cuml
|
https://api.github.com/repos/rapidsai/cuml
|
closed
|
[BUG] Speed up test_umap (single-GPU)
|
Cython / Python Perf bug tests
|
It looks like the single GPU umap pytests are taking 3.5 minutes and could be sped up.
Specifically, the `test_umap_fit_transform` and `test_umap_reproducibility` tests should be investigated.
|
1.0
|
[BUG] Speed up test_umap (single-GPU) - It looks like the single GPU umap pytests are taking 3.5 minutes and could be sped up.
Specifically, the `test_umap_fit_transform` and `test_umap_reproducibility` tests should be investigated.
|
test
|
speed up test umap single gpu it looks like the single gpu umap pytests are taking minutes and could be sped up specifically the test umap fit transform and test umap reproducibility tests should be investigated
| 1
|
162,673
| 12,685,538,602
|
IssuesEvent
|
2020-06-20 05:09:19
|
eclipse/openj9
|
https://api.github.com/repos/eclipse/openj9
|
closed
|
AArch64: Failure in cmdLineTester_jvmtitests_hcr_3
|
arch:aarch64 test failure
|
Failure link
------------
https://ci.eclipse.org/openj9/job/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/1/
rc007 and ta001 in cmdLineTester_jvmtitests_hcr_3
```
[2020-03-29T02:09:50.064Z] Testing: rc007
[2020-03-29T02:09:50.064Z] Test start time: 2020/03/29 02:09:48 Coordinated Universal Time
[2020-03-29T02:09:50.064Z] Running command: "/home/jenkins/workspace/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/openjdkbinary/j2sdk-image/bin/java" -Xnocompressedrefs -Xgcpolicy:gencon -Xjit:count=0 -Xnocompressedrefs -Xdump -agentlib:jvmtitest=test:rc007 -cp "/home/jenkins/workspace/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/openjdk-tests/TKG/../../jvmtest/functional/cmdLineTests/jvmtitests/jvmtitest.jar:/home/jenkins/workspace/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/openjdk-tests/TKG/../TKG/lib/asm-all.jar" com.ibm.jvmti.tests.util.TestRunner
[2020-03-29T02:09:50.064Z] Time spent starting: 4 milliseconds
[2020-03-29T02:09:55.835Z] Time spent executing: 6652 milliseconds
[2020-03-29T02:09:55.835Z] Test result: FAILED
[2020-03-29T02:09:55.835Z] Output from test:
[2020-03-29T02:09:55.835Z] [OUT] *** Testing [1/3]: testMultipleClassRedefinesToNewVersions
[2020-03-29T02:09:55.835Z] [OUT] *** Test took 134 milliseconds
[2020-03-29T02:09:55.835Z] [OUT] OK
[2020-03-29T02:09:55.835Z] [OUT]
[2020-03-29T02:09:55.835Z] [OUT] *** Testing [2/3]: testMultipleClassRedefinesToThemselves
[2020-03-29T02:09:55.835Z] [OUT] *** Test took 16 milliseconds
[2020-03-29T02:09:55.835Z] [OUT] FAILED
[2020-03-29T02:09:55.835Z] [OUT]
[2020-03-29T02:09:55.835Z] [OUT] *** Testing [3/3]: testMultipleRedefinesFromSameHierarchy
[2020-03-29T02:09:55.835Z] [OUT] *** Test took 29 milliseconds
[2020-03-29T02:09:55.835Z] [OUT] OK
[2020-03-29T02:09:55.835Z] [OUT]
[2020-03-29T02:09:55.835Z] >> Success condition was not found: [Return code: 0]
```
```
[2020-03-29T02:13:52.483Z] Testing: ta001
[2020-03-29T02:13:52.483Z] Test start time: 2020/03/29 02:13:52 Coordinated Universal Time
[2020-03-29T02:13:52.483Z] Running command: "/home/jenkins/workspace/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/openjdkbinary/j2sdk-image/bin/java" -Xnocompressedrefs -Xgcpolicy:gencon -Xjit:count=0 -Xnocompressedrefs -Xdump -agentlib:jvmtitest=test:ta001 -cp "/home/jenkins/workspace/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/openjdk-tests/TKG/../../jvmtest/functional/cmdLineTests/jvmtitests/jvmtitest.jar:/home/jenkins/workspace/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/openjdk-tests/TKG/../TKG/lib/asm-all.jar" com.ibm.jvmti.tests.util.TestRunner
[2020-03-29T02:13:52.483Z] Time spent starting: 4 milliseconds
[2020-03-29T02:14:01.327Z] Time spent executing: 9074 milliseconds
[2020-03-29T02:14:01.327Z] Test result: FAILED
[2020-03-29T02:14:01.327Z] Output from test:
[2020-03-29T02:14:01.327Z] [OUT] *** Testing [1/13]: testBCIUsingASM_InjectNPELogic
[2020-03-29T02:14:01.327Z] [OUT] Pre Replace: 1
[2020-03-29T02:14:01.327Z] [OUT] Expected NullPointerException received
[2020-03-29T02:14:01.327Z] [OUT] *** Test took 323 milliseconds
[2020-03-29T02:14:01.327Z] [OUT] OK
(... lines omitted ...)
[2020-03-29T02:14:01.327Z] [OUT]
[2020-03-29T02:14:01.327Z] [OUT] *** Testing [6/13]: testBCIUsingASM_inject_CatchWithSelfGOTO
[2020-03-29T02:14:01.327Z] [ERR]
[2020-03-29T02:14:01.327Z] [ERR]
[2020-03-29T02:14:01.327Z] [ERR] *** Invalid JIT return address 0000FFFF9C303D70 in 0000FFFF9C201D08
[2020-03-29T02:14:01.327Z] [ERR]
[2020-03-29T02:14:01.327Z] [ERR] 02:13:58.965 0xffff9c201a00 j9vm.249 * ** ASSERTION FAILED ** at swalk.c:1565: ((0 ))
```
|
1.0
|
AArch64: Failure in cmdLineTester_jvmtitests_hcr_3 - Failure link
------------
https://ci.eclipse.org/openj9/job/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/1/
rc007 and ta001 in cmdLineTester_jvmtitests_hcr_3
```
[2020-03-29T02:09:50.064Z] Testing: rc007
[2020-03-29T02:09:50.064Z] Test start time: 2020/03/29 02:09:48 Coordinated Universal Time
[2020-03-29T02:09:50.064Z] Running command: "/home/jenkins/workspace/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/openjdkbinary/j2sdk-image/bin/java" -Xnocompressedrefs -Xgcpolicy:gencon -Xjit:count=0 -Xnocompressedrefs -Xdump -agentlib:jvmtitest=test:rc007 -cp "/home/jenkins/workspace/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/openjdk-tests/TKG/../../jvmtest/functional/cmdLineTests/jvmtitests/jvmtitest.jar:/home/jenkins/workspace/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/openjdk-tests/TKG/../TKG/lib/asm-all.jar" com.ibm.jvmti.tests.util.TestRunner
[2020-03-29T02:09:50.064Z] Time spent starting: 4 milliseconds
[2020-03-29T02:09:55.835Z] Time spent executing: 6652 milliseconds
[2020-03-29T02:09:55.835Z] Test result: FAILED
[2020-03-29T02:09:55.835Z] Output from test:
[2020-03-29T02:09:55.835Z] [OUT] *** Testing [1/3]: testMultipleClassRedefinesToNewVersions
[2020-03-29T02:09:55.835Z] [OUT] *** Test took 134 milliseconds
[2020-03-29T02:09:55.835Z] [OUT] OK
[2020-03-29T02:09:55.835Z] [OUT]
[2020-03-29T02:09:55.835Z] [OUT] *** Testing [2/3]: testMultipleClassRedefinesToThemselves
[2020-03-29T02:09:55.835Z] [OUT] *** Test took 16 milliseconds
[2020-03-29T02:09:55.835Z] [OUT] FAILED
[2020-03-29T02:09:55.835Z] [OUT]
[2020-03-29T02:09:55.835Z] [OUT] *** Testing [3/3]: testMultipleRedefinesFromSameHierarchy
[2020-03-29T02:09:55.835Z] [OUT] *** Test took 29 milliseconds
[2020-03-29T02:09:55.835Z] [OUT] OK
[2020-03-29T02:09:55.835Z] [OUT]
[2020-03-29T02:09:55.835Z] >> Success condition was not found: [Return code: 0]
```
```
[2020-03-29T02:13:52.483Z] Testing: ta001
[2020-03-29T02:13:52.483Z] Test start time: 2020/03/29 02:13:52 Coordinated Universal Time
[2020-03-29T02:13:52.483Z] Running command: "/home/jenkins/workspace/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/openjdkbinary/j2sdk-image/bin/java" -Xnocompressedrefs -Xgcpolicy:gencon -Xjit:count=0 -Xnocompressedrefs -Xdump -agentlib:jvmtitest=test:ta001 -cp "/home/jenkins/workspace/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/openjdk-tests/TKG/../../jvmtest/functional/cmdLineTests/jvmtitests/jvmtitest.jar:/home/jenkins/workspace/Test_openjdk11_j9_sanity.functional_aarch64_linux_xl_Release/openjdk-tests/TKG/../TKG/lib/asm-all.jar" com.ibm.jvmti.tests.util.TestRunner
[2020-03-29T02:13:52.483Z] Time spent starting: 4 milliseconds
[2020-03-29T02:14:01.327Z] Time spent executing: 9074 milliseconds
[2020-03-29T02:14:01.327Z] Test result: FAILED
[2020-03-29T02:14:01.327Z] Output from test:
[2020-03-29T02:14:01.327Z] [OUT] *** Testing [1/13]: testBCIUsingASM_InjectNPELogic
[2020-03-29T02:14:01.327Z] [OUT] Pre Replace: 1
[2020-03-29T02:14:01.327Z] [OUT] Expected NullPointerException received
[2020-03-29T02:14:01.327Z] [OUT] *** Test took 323 milliseconds
[2020-03-29T02:14:01.327Z] [OUT] OK
(... lines omitted ...)
[2020-03-29T02:14:01.327Z] [OUT]
[2020-03-29T02:14:01.327Z] [OUT] *** Testing [6/13]: testBCIUsingASM_inject_CatchWithSelfGOTO
[2020-03-29T02:14:01.327Z] [ERR]
[2020-03-29T02:14:01.327Z] [ERR]
[2020-03-29T02:14:01.327Z] [ERR] *** Invalid JIT return address 0000FFFF9C303D70 in 0000FFFF9C201D08
[2020-03-29T02:14:01.327Z] [ERR]
[2020-03-29T02:14:01.327Z] [ERR] 02:13:58.965 0xffff9c201a00 j9vm.249 * ** ASSERTION FAILED ** at swalk.c:1565: ((0 ))
```
|
test
|
failure in cmdlinetester jvmtitests hcr failure link and in cmdlinetester jvmtitests hcr testing test start time coordinated universal time running command home jenkins workspace test sanity functional linux xl release openjdkbinary image bin java xnocompressedrefs xgcpolicy gencon xjit count xnocompressedrefs xdump agentlib jvmtitest test cp home jenkins workspace test sanity functional linux xl release openjdk tests tkg jvmtest functional cmdlinetests jvmtitests jvmtitest jar home jenkins workspace test sanity functional linux xl release openjdk tests tkg tkg lib asm all jar com ibm jvmti tests util testrunner time spent starting milliseconds time spent executing milliseconds test result failed output from test testing testmultipleclassredefinestonewversions test took milliseconds ok testing testmultipleclassredefinestothemselves test took milliseconds failed testing testmultipleredefinesfromsamehierarchy test took milliseconds ok success condition was not found testing test start time coordinated universal time running command home jenkins workspace test sanity functional linux xl release openjdkbinary image bin java xnocompressedrefs xgcpolicy gencon xjit count xnocompressedrefs xdump agentlib jvmtitest test cp home jenkins workspace test sanity functional linux xl release openjdk tests tkg jvmtest functional cmdlinetests jvmtitests jvmtitest jar home jenkins workspace test sanity functional linux xl release openjdk tests tkg tkg lib asm all jar com ibm jvmti tests util testrunner time spent starting milliseconds time spent executing milliseconds test result failed output from test testing testbciusingasm injectnpelogic pre replace expected nullpointerexception received test took milliseconds ok lines omitted testing testbciusingasm inject catchwithselfgoto invalid jit return address in assertion failed at swalk c
| 1
|
176,050
| 13,625,046,805
|
IssuesEvent
|
2020-09-24 08:56:55
|
cerberustesting/cerberus-source
|
https://api.github.com/repos/cerberustesting/cerberus-source
|
opened
|
Action Class and Table refactoring with columns renaming and adding
|
Nat : refactor Perim : GUITest Prio : 2 normal
|
### Renaming and type modifications (in parentheses):
- [ ] testCase > testcase
- [ ] sequence> actionId
- [ ] step > stepId
- [ ] object > value1
- [ ] property > value2
- [ ] forceExeStatus > isFatal **(BOOLEAN)**
- [ ] conditionOper > conditionOperator
- [ ] last_modified > dateModif
### New columns:
- [ ] usrCreated (VARCHAR)
- [ ] usrModif (VARCHAR)
- [ ] dateCreated (TIMESTAMP)
### Action object:
```json
{
"sort": 1,
"stepId": 2,
"actionId": 4,
"description": "click search icon",
"action": "click",
"value1": "id=com.ocito.laredoute:id/iv_search_glass",
"value2": "",
"value3": "",
"conditionOper": "always",
"conditionVal1": "",
"conditionVal2": "",
"conditionVal3": "",
"isFatal": true,
"screenshotFilename": "",
"controls": [],
"test": "ENGINE : Action",
"testcase": "0017B",
"usrCreated": "",
"usrModif": "",
"dateCreated": "",
"dateModif": ""
}
```
|
1.0
|
Action Class and Table refactoring with columns renaming and adding - ### Renaming and type modifications (in parentheses):
- [ ] testCase > testcase
- [ ] sequence> actionId
- [ ] step > stepId
- [ ] object > value1
- [ ] property > value2
- [ ] forceExeStatus > isFatal **(BOOLEAN)**
- [ ] conditionOper > conditionOperator
- [ ] last_modified > dateModif
### New columns:
- [ ] usrCreated (VARCHAR)
- [ ] usrModif (VARCHAR)
- [ ] dateCreated (TIMESTAMP)
### Action object:
```json
{
"sort": 1,
"stepId": 2,
"actionId": 4,
"description": "click search icon",
"action": "click",
"value1": "id=com.ocito.laredoute:id/iv_search_glass",
"value2": "",
"value3": "",
"conditionOper": "always",
"conditionVal1": "",
"conditionVal2": "",
"conditionVal3": "",
"isFatal": true,
"screenshotFilename": "",
"controls": [],
"test": "ENGINE : Action",
"testcase": "0017B",
"usrCreated": "",
"usrModif": "",
"dateCreated": "",
"dateModif": ""
}
```
|
test
|
action class and table refactoring with columns renaming and adding renaming and type modifications in parentheses testcase testcase sequence actionid step stepid object property forceexestatus isfatal boolean conditionoper conditionoperator last modified datemodif new columns usrcreated varchar usrmodif varchar datecreated timestamp action object json sort stepid actionid description click search icon action click id com ocito laredoute id iv search glass conditionoper always isfatal true screenshotfilename controls test engine action testcase usrcreated usrmodif datecreated datemodif
| 1
|
753,033
| 26,338,456,946
|
IssuesEvent
|
2023-01-10 15:56:24
|
hashicorp/terraform-cdk
|
https://api.github.com/repos/hashicorp/terraform-cdk
|
closed
|
Improve error message when trying to deploy to non-existing TFC workspace
|
bug terraform cloud cdktf-cli priority/important-soon ux/cli size/small
|
<!--- Please keep this note for the community --->
### Community Note
- Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
- Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
- If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### cdktf & Language Versions
* 0.12.2
* any language
<!--- Please make note of the version you are using for the `hashicorp/terraform-cdk` and language bindings you are using.
If you are not running the latest version, please upgrade because your issue may have already been fixed. --->
### Affected Resource(s)
* `CloudBackend`
* `RemoteBackend`
<!--- Please list the affected resources. --->
### Debug Output
<img width="793" alt="image" src="https://user-images.githubusercontent.com/1112056/190597779-051555f5-d213-40b0-a3c4-1808aea6bdea.png">
<!---
Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist.
--->
### Expected Behavior
Should print a good error message or even ask whether the cdktf-cli should create that workspace and try again.
<!--- What should have happened? --->
### Actual Behavior
Error does not help at all
<!--- What actually happened? --->
### Steps to Reproduce
Try to deploy to a non-existing workspace in Terraform Cloud using the `CloudBackend` or `RemoteBackend` construct.
<!--- Please list the steps required to reproduce the issue. --->
### Important Factoids
<!--- Are there anything atypical about your accounts that we should know? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example:
--->
- #0000
|
1.0
|
Improve error message when trying to deploy to non-existing TFC workspace - <!--- Please keep this note for the community --->
### Community Note
- Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
- Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
- If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### cdktf & Language Versions
* 0.12.2
* any language
<!--- Please make note of the version you are using for the `hashicorp/terraform-cdk` and language bindings you are using.
If you are not running the latest version, please upgrade because your issue may have already been fixed. --->
### Affected Resource(s)
* `CloudBackend`
* `RemoteBackend`
<!--- Please list the affected resources. --->
### Debug Output
<img width="793" alt="image" src="https://user-images.githubusercontent.com/1112056/190597779-051555f5-d213-40b0-a3c4-1808aea6bdea.png">
<!---
Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist.
--->
### Expected Behavior
Should print a good error message or even ask whether the cdktf-cli should create that workspace and try again.
<!--- What should have happened? --->
### Actual Behavior
Error does not help at all
<!--- What actually happened? --->
### Steps to Reproduce
Try to deploy to a non-existing workspace in Terraform Cloud using the `CloudBackend` or `RemoteBackend` construct.
<!--- Please list the steps required to reproduce the issue. --->
### Important Factoids
<!--- Are there anything atypical about your accounts that we should know? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example:
--->
- #0000
|
non_test
|
improve error message when trying to deploy to non existing tfc workspace community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment cdktf language versions any language please make note of the version you are using for the hashicorp terraform cdk and language bindings you are using if you are not running the latest version please upgrade because your issue may have already been fixed affected resource s cloudbackend remotebackend debug output img width alt image src please provide a link to a github gist containing the complete debug output please do not paste the debug output in the issue just paste a link to the gist expected behavior should print a good error message or even ask whether the cdktf cli should create that workspace and try again actual behavior error does not help at all steps to reproduce try to deploy to a non existing workspace in terraform cloud using the cloudbackend or remotebackend construct important factoids references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor documentation for example
| 0
|
95,278
| 8,555,386,630
|
IssuesEvent
|
2018-11-08 09:53:51
|
humera987/FXLabs-Test-Automation
|
https://api.github.com/repos/humera987/FXLabs-Test-Automation
|
opened
|
testing8 : ApiV1ProjectsProjectidAutocodeconfigGetAnonymousInvalid
|
testing8 testing8
|
Project : testing8
Job : UAT
Env : UAT
Region : US_WEST_3
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Thu, 08 Nov 2018 09:53:49 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/projects/278864660/autocodeconfig
Request :
Response :
{
"timestamp" : "2018-11-08T09:53:50.504+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/projects/278864660/autocodeconfig"
}
Logs :
Assertion [@StatusCode == 401 OR @StatusCode == 403] resolved-to [404 == 401 OR 404 == 403] result [Failed]
--- FX Bot ---
|
2.0
|
testing8 : ApiV1ProjectsProjectidAutocodeconfigGetAnonymousInvalid - Project : testing8
Job : UAT
Env : UAT
Region : US_WEST_3
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Thu, 08 Nov 2018 09:53:49 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/projects/278864660/autocodeconfig
Request :
Response :
{
"timestamp" : "2018-11-08T09:53:50.504+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/projects/278864660/autocodeconfig"
}
Logs :
Assertion [@StatusCode == 401 OR @StatusCode == 403] resolved-to [404 == 401 OR 404 == 403] result [Failed]
--- FX Bot ---
|
test
|
project job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api projects autocodeconfig logs assertion resolved to result fx bot
| 1
|
121,562
| 10,173,699,884
|
IssuesEvent
|
2019-08-08 13:39:45
|
CityOfBoston/boston.gov-d8
|
https://api.github.com/repos/CityOfBoston/boston.gov-d8
|
opened
|
Global: extra element "And Apply" at the headers of all pages
|
max-testing
|
On the top header of all pages on the D8 site, there is an extra element title "And Apply" at the top of the page.
D7:

D8:

Tested on Chrome 76 on OS X Mojave
|
1.0
|
Global: extra element "And Apply" at the headers of all pages - On the top header of all pages on the D8 site, there is an extra element title "And Apply" at the top of the page.
D7:

D8:

Tested on Chrome 76 on OS X Mojave
|
test
|
global extra element and apply at the headers of all pages on the top header of all pages on the site there is an extra element title and apply at the top of the page tested on chrome on os x mojave
| 1
|
288,674
| 21,716,750,498
|
IssuesEvent
|
2022-05-10 18:43:26
|
mp-se/gravitymon
|
https://api.github.com/repos/mp-se/gravitymon
|
closed
|
Create instructions for contribution
|
documentation
|
Add a section on contribution to this project. Probably need to have a section describing the software setup and how to add an new push target.
|
1.0
|
Create instructions for contribution - Add a section on contribution to this project. Probably need to have a section describing the software setup and how to add an new push target.
|
non_test
|
create instructions for contribution add a section on contribution to this project probably need to have a section describing the software setup and how to add an new push target
| 0
|
154,846
| 19,758,416,509
|
IssuesEvent
|
2022-01-16 01:36:26
|
turkdevops/play-with-docker
|
https://api.github.com/repos/turkdevops/play-with-docker
|
opened
|
CVE-2021-27918 (High) detected in github.com/miekg/dns-v1.0.0
|
security vulnerability
|
## CVE-2021-27918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/miekg/dns-v1.0.0</b></p></summary>
<p>DNS library in Go</p>
<p>
Dependency Hierarchy:
- :x: **github.com/miekg/dns-v1.0.0** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/play-with-docker/commit/27377d4ea18db54381a8dc972091f3c342337ec9">27377d4ea18db54381a8dc972091f3c342337ec9</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
encoding/xml in Go before 1.15.9 and 1.16.x before 1.16.1 has an infinite loop if a custom TokenReader (for xml.NewTokenDecoder) returns EOF in the middle of an element. This can occur in the Decode, DecodeElement, or Skip method.
<p>Publish Date: 2021-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27918>CVE-2021-27918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://groups.google.com/g/golang-announce/c/MfiLYjG-RAw">https://groups.google.com/g/golang-announce/c/MfiLYjG-RAw</a></p>
<p>Release Date: 2021-03-11</p>
<p>Fix Resolution: 1.15.9, 1.16.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-27918 (High) detected in github.com/miekg/dns-v1.0.0 - ## CVE-2021-27918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/miekg/dns-v1.0.0</b></p></summary>
<p>DNS library in Go</p>
<p>
Dependency Hierarchy:
- :x: **github.com/miekg/dns-v1.0.0** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/play-with-docker/commit/27377d4ea18db54381a8dc972091f3c342337ec9">27377d4ea18db54381a8dc972091f3c342337ec9</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
encoding/xml in Go before 1.15.9 and 1.16.x before 1.16.1 has an infinite loop if a custom TokenReader (for xml.NewTokenDecoder) returns EOF in the middle of an element. This can occur in the Decode, DecodeElement, or Skip method.
<p>Publish Date: 2021-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27918>CVE-2021-27918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://groups.google.com/g/golang-announce/c/MfiLYjG-RAw">https://groups.google.com/g/golang-announce/c/MfiLYjG-RAw</a></p>
<p>Release Date: 2021-03-11</p>
<p>Fix Resolution: 1.15.9, 1.16.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in github com miekg dns cve high severity vulnerability vulnerable library github com miekg dns dns library in go dependency hierarchy x github com miekg dns vulnerable library found in head commit a href found in base branch master vulnerability details encoding xml in go before and x before has an infinite loop if a custom tokenreader for xml newtokendecoder returns eof in the middle of an element this can occur in the decode decodeelement or skip method publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
132,044
| 12,497,341,245
|
IssuesEvent
|
2020-06-01 16:19:47
|
vitaliytasksync/QA-Destination
|
https://api.github.com/repos/vitaliytasksync/QA-Destination
|
opened
|
Task8 trello Emphasis1_anna (жирный)
|
bug documentation
|
**Meet my family. There are five of us – my parents, my elder brother,
my baby sister and me. First, meet my mum and dad, Jane and Michael. My
mum enjoys reading and my dad enjoys playing chess with my brother Ken**
|
1.0
|
Task8 trello Emphasis1_anna (жирный) - **Meet my family. There are five of us – my parents, my elder brother,
my baby sister and me. First, meet my mum and dad, Jane and Michael. My
mum enjoys reading and my dad enjoys playing chess with my brother Ken**
|
non_test
|
trello anna жирный meet my family there are five of us – my parents my elder brother my baby sister and me first meet my mum and dad jane and michael my mum enjoys reading and my dad enjoys playing chess with my brother ken
| 0
|
97,041
| 16,189,675,039
|
IssuesEvent
|
2021-05-04 06:22:22
|
EmpoHQ/empo.im
|
https://api.github.com/repos/EmpoHQ/empo.im
|
opened
|
CVE-2021-23368 (Medium) detected in postcss-7.0.35.tgz
|
security vulnerability
|
## CVE-2021-23368 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postcss-7.0.35.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.35.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.35.tgz</a></p>
<p>Path to dependency file: empo.im/package.json</p>
<p>Path to vulnerable library: empo.im/node_modules/postcss-minify-params/node_modules/postcss/package.json,empo.im/node_modules/postcss-url/node_modules/postcss/package.json,empo.im/node_modules/postcss-calc/node_modules/postcss/package.json,empo.im/node_modules/postcss-modules-local-by-default/node_modules/postcss/package.json,empo.im/node_modules/cssnano-preset-default/node_modules/postcss/package.json,empo.im/node_modules/postcss-selector-matches/node_modules/postcss/package.json,empo.im/node_modules/postcss-colormin/node_modules/postcss/package.json,empo.im/node_modules/css-loader/node_modules/postcss/package.json,empo.im/node_modules/postcss-pseudo-class-any-link/node_modules/postcss/package.json,empo.im/node_modules/@nuxt/webpack/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-charset/node_modules/postcss/package.json,empo.im/node_modules/postcss-discard-empty/node_modules/postcss/package.json,empo.im/node_modules/postcss-convert-values/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-string/node_modules/postcss/package.json,empo.im/node_modules/postcss-lab-function/node_modules/postcss/package.json,empo.im/node_modules/cssnano-util-raw-cache/node_modules/postcss/package.json,empo.im/node_modules/postcss-focus-within/node_modules/postcss/package.json,empo.im/node_modules/postcss-color-gray/node_modules/postcss/package.json,empo.im/node_modules/postcss-env-function/node_modules/postcss/package.json,empo.im/node_modules/postcss-attribute-case-insensitive/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-url/node_modules/postcss/package.json,empo.im/node_modules/postcss-logical/node_modules/postcss/package.json,empo.im/node_modules/postcss-nesting/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-positions/node_modules/postcss/package.json,empo.im/node_modules/postcss-discard-overridden/node_modules/postcss/package.json,empo.im/node_modules/css-has-pseudo/node_modules/postcss/package.json,empo.im/node_modules/css-declaration-sorter/node_modules/postcss/package.json,empo.im/node_modules/postcss-overflow-shorthand/node_modules/postcss/package.json,empo.im/node_modules/postcss-replace-overflow-wrap/node_modules/postcss/package.json,empo.im/node_modules/postcss-font-variant/node_modules/postcss/package.json,empo.im/node_modules/postcss-reduce-transforms/node_modules/postcss/package.json,empo.im/node_modules/postcss-image-set-function/node_modules/postcss/package.json,empo.im/node_modules/postcss-modules-scope/node_modules/postcss/package.json,empo.im/node_modules/postcss-reduce-initial/node_modules/postcss/package.json,empo.im/node_modules/postcss-discard-duplicates/node_modules/postcss/package.json,empo.im/node_modules/@nuxtjs/tailwindcss/node_modules/postcss-values-parser/node_modules/postcss/package.json,empo.im/node_modules/postcss-dir-pseudo-class/node_modules/postcss/package.json,empo.im/node_modules/postcss-media-minmax/node_modules/postcss/package.json,empo.im/node_modules/postcss-minify-selectors/node_modules/postcss/package.json,empo.im/node_modules/postcss-color-rebeccapurple/node_modules/postcss/package.json,empo.im/node_modules/postcss-svgo/node_modules/postcss/package.json,empo.im/node_modules/postcss-page-break/node_modules/postcss/package.json,empo.im/node_modules/css-prefers-color-scheme/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-display-values/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-timing-functions/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-repeat-style/node_modules/postcss/package.json,empo.im/node_modules/cssnano/node_modules/postcss/package.json,empo.im/node_modules/postcss-double-position-gradients/node_modules/postcss/package.json,empo.im/node_modules/postcss-merge-longhand/node_modules/postcss/package.json,empo.im/node_modules/postcss-loader/node_modules/postcss/package.json,empo.im/node_modules/postcss-discard-comments/node_modules/postcss/package.json,empo.im/node_modules/postcss-custom-selectors/node_modules/postcss/package.json,empo.im/node_modules/postcss-modules-extract-imports/node_modules/postcss/package.json,empo.im/node_modules/postcss-focus-visible/node_modules/postcss/package.json,empo.im/node_modules/postcss-unique-selectors/node_modules/postcss/package.json,empo.im/node_modules/postcss-merge-rules/node_modules/postcss/package.json,empo.im/node_modules/postcss-color-mod-function/node_modules/postcss/package.json,empo.im/node_modules/postcss-minify-gradients/node_modules/postcss/package.json,empo.im/node_modules/autoprefixer/node_modules/postcss/package.json,empo.im/node_modules/postcss-color-hex-alpha/node_modules/postcss/package.json,empo.im/node_modules/postcss-preset-env/node_modules/postcss/package.json,empo.im/node_modules/@vue/component-compiler-utils/node_modules/postcss/package.json,empo.im/node_modules/stylehacks/node_modules/postcss/package.json,empo.im/node_modules/postcss-place/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-whitespace/node_modules/postcss/package.json,empo.im/node_modules/postcss-selector-not/node_modules/postcss/package.json,empo.im/node_modules/postcss-import/node_modules/postcss/package.json,empo.im/node_modules/css-blank-pseudo/node_modules/postcss/package.json,empo.im/node_modules/icss-utils/node_modules/postcss/package.json,empo.im/node_modules/postcss-custom-properties/node_modules/postcss/package.json,empo.im/node_modules/postcss-color-functional-notation/node_modules/postcss/package.json,empo.im/node_modules/postcss-custom-media/node_modules/postcss/package.json,empo.im/node_modules/postcss-gap-properties/node_modules/postcss/package.json,empo.im/node_modules/postcss-initial/node_modules/postcss/package.json,empo.im/node_modules/postcss-modules-values/node_modules/postcss/package.json,empo.im/node_modules/postcss-minify-font-values/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-unicode/node_modules/postcss/package.json,empo.im/node_modules/postcss-ordered-values/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- nuxt-2.15.4.tgz (Root Library)
- webpack-2.15.4.tgz
- postcss-loader-3.0.0.tgz
- :x: **postcss-7.0.35.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/EmpoHQ/empo.im/commit/6a48c9e8bc19dba161bd9bb27023663b562fe4c8">6a48c9e8bc19dba161bd9bb27023663b562fe4c8</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss from 7.0.0 and before 8.2.10 are vulnerable to Regular Expression Denial of Service (ReDoS) during source map parsing.
<p>Publish Date: 2021-04-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23368>CVE-2021-23368</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368</a></p>
<p>Release Date: 2021-04-12</p>
<p>Fix Resolution: postcss -8.2.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23368 (Medium) detected in postcss-7.0.35.tgz - ## CVE-2021-23368 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postcss-7.0.35.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.35.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.35.tgz</a></p>
<p>Path to dependency file: empo.im/package.json</p>
<p>Path to vulnerable library: empo.im/node_modules/postcss-minify-params/node_modules/postcss/package.json,empo.im/node_modules/postcss-url/node_modules/postcss/package.json,empo.im/node_modules/postcss-calc/node_modules/postcss/package.json,empo.im/node_modules/postcss-modules-local-by-default/node_modules/postcss/package.json,empo.im/node_modules/cssnano-preset-default/node_modules/postcss/package.json,empo.im/node_modules/postcss-selector-matches/node_modules/postcss/package.json,empo.im/node_modules/postcss-colormin/node_modules/postcss/package.json,empo.im/node_modules/css-loader/node_modules/postcss/package.json,empo.im/node_modules/postcss-pseudo-class-any-link/node_modules/postcss/package.json,empo.im/node_modules/@nuxt/webpack/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-charset/node_modules/postcss/package.json,empo.im/node_modules/postcss-discard-empty/node_modules/postcss/package.json,empo.im/node_modules/postcss-convert-values/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-string/node_modules/postcss/package.json,empo.im/node_modules/postcss-lab-function/node_modules/postcss/package.json,empo.im/node_modules/cssnano-util-raw-cache/node_modules/postcss/package.json,empo.im/node_modules/postcss-focus-within/node_modules/postcss/package.json,empo.im/node_modules/postcss-color-gray/node_modules/postcss/package.json,empo.im/node_modules/postcss-env-function/node_modules/postcss/package.json,empo.im/node_modules/postcss-attribute-case-insensitive/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-url/node_modules/postcss/package.json,empo.im/node_modules/postcss-logical/node_modules/postcss/package.json,empo.im/node_modules/postcss-nesting/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-positions/node_modules/postcss/package.json,empo.im/node_modules/postcss-discard-overridden/node_modules/postcss/package.json,empo.im/node_modules/css-has-pseudo/node_modules/postcss/package.json,empo.im/node_modules/css-declaration-sorter/node_modules/postcss/package.json,empo.im/node_modules/postcss-overflow-shorthand/node_modules/postcss/package.json,empo.im/node_modules/postcss-replace-overflow-wrap/node_modules/postcss/package.json,empo.im/node_modules/postcss-font-variant/node_modules/postcss/package.json,empo.im/node_modules/postcss-reduce-transforms/node_modules/postcss/package.json,empo.im/node_modules/postcss-image-set-function/node_modules/postcss/package.json,empo.im/node_modules/postcss-modules-scope/node_modules/postcss/package.json,empo.im/node_modules/postcss-reduce-initial/node_modules/postcss/package.json,empo.im/node_modules/postcss-discard-duplicates/node_modules/postcss/package.json,empo.im/node_modules/@nuxtjs/tailwindcss/node_modules/postcss-values-parser/node_modules/postcss/package.json,empo.im/node_modules/postcss-dir-pseudo-class/node_modules/postcss/package.json,empo.im/node_modules/postcss-media-minmax/node_modules/postcss/package.json,empo.im/node_modules/postcss-minify-selectors/node_modules/postcss/package.json,empo.im/node_modules/postcss-color-rebeccapurple/node_modules/postcss/package.json,empo.im/node_modules/postcss-svgo/node_modules/postcss/package.json,empo.im/node_modules/postcss-page-break/node_modules/postcss/package.json,empo.im/node_modules/css-prefers-color-scheme/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-display-values/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-timing-functions/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-repeat-style/node_modules/postcss/package.json,empo.im/node_modules/cssnano/node_modules/postcss/package.json,empo.im/node_modules/postcss-double-position-gradients/node_modules/postcss/package.json,empo.im/node_modules/postcss-merge-longhand/node_modules/postcss/package.json,empo.im/node_modules/postcss-loader/node_modules/postcss/package.json,empo.im/node_modules/postcss-discard-comments/node_modules/postcss/package.json,empo.im/node_modules/postcss-custom-selectors/node_modules/postcss/package.json,empo.im/node_modules/postcss-modules-extract-imports/node_modules/postcss/package.json,empo.im/node_modules/postcss-focus-visible/node_modules/postcss/package.json,empo.im/node_modules/postcss-unique-selectors/node_modules/postcss/package.json,empo.im/node_modules/postcss-merge-rules/node_modules/postcss/package.json,empo.im/node_modules/postcss-color-mod-function/node_modules/postcss/package.json,empo.im/node_modules/postcss-minify-gradients/node_modules/postcss/package.json,empo.im/node_modules/autoprefixer/node_modules/postcss/package.json,empo.im/node_modules/postcss-color-hex-alpha/node_modules/postcss/package.json,empo.im/node_modules/postcss-preset-env/node_modules/postcss/package.json,empo.im/node_modules/@vue/component-compiler-utils/node_modules/postcss/package.json,empo.im/node_modules/stylehacks/node_modules/postcss/package.json,empo.im/node_modules/postcss-place/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-whitespace/node_modules/postcss/package.json,empo.im/node_modules/postcss-selector-not/node_modules/postcss/package.json,empo.im/node_modules/postcss-import/node_modules/postcss/package.json,empo.im/node_modules/css-blank-pseudo/node_modules/postcss/package.json,empo.im/node_modules/icss-utils/node_modules/postcss/package.json,empo.im/node_modules/postcss-custom-properties/node_modules/postcss/package.json,empo.im/node_modules/postcss-color-functional-notation/node_modules/postcss/package.json,empo.im/node_modules/postcss-custom-media/node_modules/postcss/package.json,empo.im/node_modules/postcss-gap-properties/node_modules/postcss/package.json,empo.im/node_modules/postcss-initial/node_modules/postcss/package.json,empo.im/node_modules/postcss-modules-values/node_modules/postcss/package.json,empo.im/node_modules/postcss-minify-font-values/node_modules/postcss/package.json,empo.im/node_modules/postcss-normalize-unicode/node_modules/postcss/package.json,empo.im/node_modules/postcss-ordered-values/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- nuxt-2.15.4.tgz (Root Library)
- webpack-2.15.4.tgz
- postcss-loader-3.0.0.tgz
- :x: **postcss-7.0.35.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/EmpoHQ/empo.im/commit/6a48c9e8bc19dba161bd9bb27023663b562fe4c8">6a48c9e8bc19dba161bd9bb27023663b562fe4c8</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss from 7.0.0 and before 8.2.10 are vulnerable to Regular Expression Denial of Service (ReDoS) during source map parsing.
<p>Publish Date: 2021-04-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23368>CVE-2021-23368</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368</a></p>
<p>Release Date: 2021-04-12</p>
<p>Fix Resolution: postcss -8.2.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in postcss tgz cve medium severity vulnerability vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file empo im package json path to vulnerable library empo im node modules postcss minify params node modules postcss package json empo im node modules postcss url node modules postcss package json empo im node modules postcss calc node modules postcss package json empo im node modules postcss modules local by default node modules postcss package json empo im node modules cssnano preset default node modules postcss package json empo im node modules postcss selector matches node modules postcss package json empo im node modules postcss colormin node modules postcss package json empo im node modules css loader node modules postcss package json empo im node modules postcss pseudo class any link node modules postcss package json empo im node modules nuxt webpack node modules postcss package json empo im node modules postcss normalize charset node modules postcss package json empo im node modules postcss discard empty node modules postcss package json empo im node modules postcss convert values node modules postcss package json empo im node modules postcss normalize string node modules postcss package json empo im node modules postcss lab function node modules postcss package json empo im node modules cssnano util raw cache node modules postcss package json empo im node modules postcss focus within node modules postcss package json empo im node modules postcss color gray node modules postcss package json empo im node modules postcss env function node modules postcss package json empo im node modules postcss attribute case insensitive node modules postcss package json empo im node modules postcss normalize url node modules postcss package json empo im node modules postcss logical node modules postcss package json empo im node modules postcss nesting node modules postcss package json empo im node modules postcss normalize positions node modules postcss package json empo im node modules postcss discard overridden node modules postcss package json empo im node modules css has pseudo node modules postcss package json empo im node modules css declaration sorter node modules postcss package json empo im node modules postcss overflow shorthand node modules postcss package json empo im node modules postcss replace overflow wrap node modules postcss package json empo im node modules postcss font variant node modules postcss package json empo im node modules postcss reduce transforms node modules postcss package json empo im node modules postcss image set function node modules postcss package json empo im node modules postcss modules scope node modules postcss package json empo im node modules postcss reduce initial node modules postcss package json empo im node modules postcss discard duplicates node modules postcss package json empo im node modules nuxtjs tailwindcss node modules postcss values parser node modules postcss package json empo im node modules postcss dir pseudo class node modules postcss package json empo im node modules postcss media minmax node modules postcss package json empo im node modules postcss minify selectors node modules postcss package json empo im node modules postcss color rebeccapurple node modules postcss package json empo im node modules postcss svgo node modules postcss package json empo im node modules postcss page break node modules postcss package json empo im node modules css prefers color scheme node modules postcss package json empo im node modules postcss normalize display values node modules postcss package json empo im node modules postcss normalize timing functions node modules postcss package json empo im node modules postcss normalize repeat style node modules postcss package json empo im node modules cssnano node modules postcss package json empo im node modules postcss double position gradients node modules postcss package json empo im node modules postcss merge longhand node modules postcss package json empo im node modules postcss loader node modules postcss package json empo im node modules postcss discard comments node modules postcss package json empo im node modules postcss custom selectors node modules postcss package json empo im node modules postcss modules extract imports node modules postcss package json empo im node modules postcss focus visible node modules postcss package json empo im node modules postcss unique selectors node modules postcss package json empo im node modules postcss merge rules node modules postcss package json empo im node modules postcss color mod function node modules postcss package json empo im node modules postcss minify gradients node modules postcss package json empo im node modules autoprefixer node modules postcss package json empo im node modules postcss color hex alpha node modules postcss package json empo im node modules postcss preset env node modules postcss package json empo im node modules vue component compiler utils node modules postcss package json empo im node modules stylehacks node modules postcss package json empo im node modules postcss place node modules postcss package json empo im node modules postcss normalize whitespace node modules postcss package json empo im node modules postcss selector not node modules postcss package json empo im node modules postcss import node modules postcss package json empo im node modules css blank pseudo node modules postcss package json empo im node modules icss utils node modules postcss package json empo im node modules postcss custom properties node modules postcss package json empo im node modules postcss color functional notation node modules postcss package json empo im node modules postcss custom media node modules postcss package json empo im node modules postcss gap properties node modules postcss package json empo im node modules postcss initial node modules postcss package json empo im node modules postcss modules values node modules postcss package json empo im node modules postcss minify font values node modules postcss package json empo im node modules postcss normalize unicode node modules postcss package json empo im node modules postcss ordered values node modules postcss package json dependency hierarchy nuxt tgz root library webpack tgz postcss loader tgz x postcss tgz vulnerable library found in head commit a href found in base branch main vulnerability details the package postcss from and before are vulnerable to regular expression denial of service redos during source map parsing publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution postcss step up your open source security game with whitesource
| 0
|
319,541
| 27,381,977,561
|
IssuesEvent
|
2023-02-28 10:38:58
|
marten-voorberg/DD2480-gson
|
https://api.github.com/repos/marten-voorberg/DD2480-gson
|
closed
|
Check Exception message in tests
|
test
|
We have some tests which fail if an Exception is thrown. We should make sure that we check that the thrown exception has the appropriate message.
|
1.0
|
Check Exception message in tests - We have some tests which fail if an Exception is thrown. We should make sure that we check that the thrown exception has the appropriate message.
|
test
|
check exception message in tests we have some tests which fail if an exception is thrown we should make sure that we check that the thrown exception has the appropriate message
| 1
|
43,898
| 2,893,897,972
|
IssuesEvent
|
2015-06-15 20:24:09
|
dkpro/dkpro-tc
|
https://api.github.com/repos/dkpro/dkpro-tc
|
closed
|
Make use of ngramFreqThreshold or remove it
|
Milestone-Release0.8.0 Priority-Medium
|
The ngramFreqThreshold parameter in the NGram module is not used by any (child) feature extractor. That is misleading. Furthermore, its functionality (min. frequency-based ngram selection rather than absolute number) would be helpful.
|
1.0
|
Make use of ngramFreqThreshold or remove it - The ngramFreqThreshold parameter in the NGram module is not used by any (child) feature extractor. That is misleading. Furthermore, its functionality (min. frequency-based ngram selection rather than absolute number) would be helpful.
|
non_test
|
make use of ngramfreqthreshold or remove it the ngramfreqthreshold parameter in the ngram module is not used by any child feature extractor that is misleading furthermore its functionality min frequency based ngram selection rather than absolute number would be helpful
| 0
|
17,095
| 3,593,167,269
|
IssuesEvent
|
2016-02-01 18:41:56
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
VS Integration tests should check for unexpected image loads...
|
Area-IDE Test
|
We shouldn't have to wait for RPS to find out we're loading extra dlls in scenarios where we shouldn't. Specifically, we should ensure that we have coverage for:
- Roslyn loading before a solution is opened
- VB is not loaded in C# projects and vice versa
- Interactive is not loaded if Interactive is not opened
|
1.0
|
VS Integration tests should check for unexpected image loads... - We shouldn't have to wait for RPS to find out we're loading extra dlls in scenarios where we shouldn't. Specifically, we should ensure that we have coverage for:
- Roslyn loading before a solution is opened
- VB is not loaded in C# projects and vice versa
- Interactive is not loaded if Interactive is not opened
|
test
|
vs integration tests should check for unexpected image loads we shouldn t have to wait for rps to find out we re loading extra dlls in scenarios where we shouldn t specifically we should ensure that we have coverage for roslyn loading before a solution is opened vb is not loaded in c projects and vice versa interactive is not loaded if interactive is not opened
| 1
|
85,934
| 8,014,038,837
|
IssuesEvent
|
2018-07-25 04:02:14
|
TEAMMATES/teammates
|
https://api.github.com/repos/TEAMMATES/teammates
|
closed
|
AdminActivityLogPageAction: fix bug in generation of status message
|
a-Testing f-Admin p.Low
|
**Environment**
master branch at efad9e9dc1149779850773b323acaf2639b15c90
**Steps to reproduce**
Go to AdminActivityLog in admin console and type query to filter the log messages.
Currently when the `person:xxx` like query is present, status message is supposed to contain the time in that person's timezone (`xxx in Local Time Zone (8.0).`). But this function gets a huge bug: The query will be processed as lower case. If the person's google id contains upper case, the information could be lost.
Let's use `person:idOfInstructor1OfCourse1` as a example.
**Expected behaviour**
When query the person's timezone in the back-end, the google Id should be `idOfInstructor1OfCourse1`.
**Actual behaviour**
The google Id will become `idofinstructor1ofcourse1`.
**Notes**
The bug is found in #6655 and because of this, the behaviour cannot be tested as our test data got capital letter for googleID.
The bug is not so obvious and not so urgent as most of users' (I guess all?) google id will be in lower case.
If the issue is fixed, remeber to add back the test case in `AdminActivityLogPageActionTest`.
```
verifyLocalTimeInStatusMessage(statusMessage, yesterday, 2);
```
|
1.0
|
AdminActivityLogPageAction: fix bug in generation of status message - **Environment**
master branch at efad9e9dc1149779850773b323acaf2639b15c90
**Steps to reproduce**
Go to AdminActivityLog in admin console and type query to filter the log messages.
Currently when the `person:xxx` like query is present, status message is supposed to contain the time in that person's timezone (`xxx in Local Time Zone (8.0).`). But this function gets a huge bug: The query will be processed as lower case. If the person's google id contains upper case, the information could be lost.
Let's use `person:idOfInstructor1OfCourse1` as a example.
**Expected behaviour**
When query the person's timezone in the back-end, the google Id should be `idOfInstructor1OfCourse1`.
**Actual behaviour**
The google Id will become `idofinstructor1ofcourse1`.
**Notes**
The bug is found in #6655 and because of this, the behaviour cannot be tested as our test data got capital letter for googleID.
The bug is not so obvious and not so urgent as most of users' (I guess all?) google id will be in lower case.
If the issue is fixed, remeber to add back the test case in `AdminActivityLogPageActionTest`.
```
verifyLocalTimeInStatusMessage(statusMessage, yesterday, 2);
```
|
test
|
adminactivitylogpageaction fix bug in generation of status message environment master branch at steps to reproduce go to adminactivitylog in admin console and type query to filter the log messages currently when the person xxx like query is present status message is supposed to contain the time in that person s timezone xxx in local time zone but this function gets a huge bug the query will be processed as lower case if the person s google id contains upper case the information could be lost let s use person as a example expected behaviour when query the person s timezone in the back end the google id should be actual behaviour the google id will become notes the bug is found in and because of this the behaviour cannot be tested as our test data got capital letter for googleid the bug is not so obvious and not so urgent as most of users i guess all google id will be in lower case if the issue is fixed remeber to add back the test case in adminactivitylogpageactiontest verifylocaltimeinstatusmessage statusmessage yesterday
| 1
|
73,849
| 7,360,284,486
|
IssuesEvent
|
2018-03-10 17:01:04
|
istio/istio
|
https://api.github.com/repos/istio/istio
|
closed
|
Postsubmit Failture by replace PROXY_TAG by PROXY_REPO_SHA #4158
|
kind/test-failure
|
Post submit fails after pr #4158 was merged.
https://k8s-gubernator.appspot.com/build/istio-prow/e2e-suite-rbac-no_auth/1008
```
customresourcedefinition "externalservices.networking.istio.io" configured
error when retrieving current configuration of:
&{0xc42022b200 0xc420336bd0 /home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml 0xc4223c7588 0xc4223c7588 false}
from server for: "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": resource name may not be empty
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-pilot-admin-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-sidecar-injector-admin-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-ca-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-ingress-admin-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-sidecar-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-mixer-admin-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-mixer-validator-admin-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (NotFound): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": namespaces "demo-test-4aa5d9c63dac4fde8c15b77e7ab" not found
```
|
1.0
|
Postsubmit Failture by replace PROXY_TAG by PROXY_REPO_SHA #4158 - Post submit fails after pr #4158 was merged.
https://k8s-gubernator.appspot.com/build/istio-prow/e2e-suite-rbac-no_auth/1008
```
customresourcedefinition "externalservices.networking.istio.io" configured
error when retrieving current configuration of:
&{0xc42022b200 0xc420336bd0 /home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml 0xc4223c7588 0xc4223c7588 false}
from server for: "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": resource name may not be empty
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-pilot-admin-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-sidecar-injector-admin-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-ca-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-ingress-admin-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-sidecar-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-mixer-admin-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (Invalid): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": ClusterRoleBinding.rbac.authorization.k8s.io "istio-mixer-validator-admin-role-binding-" is invalid: subjects[0].namespace: Required value
Error from server (NotFound): error when creating "/home/bootstrap/go/src/istio.io/istio/_artifacts/demo-test-4aa5d9c63dac4fde8c15b77e7ab/yaml/istio-one-namespace.yaml": namespaces "demo-test-4aa5d9c63dac4fde8c15b77e7ab" not found
```
|
test
|
postsubmit failture by replace proxy tag by proxy repo sha post submit fails after pr was merged customresourcedefinition externalservices networking istio io configured error when retrieving current configuration of home bootstrap go src istio io istio artifacts demo test yaml istio one namespace yaml false from server for home bootstrap go src istio io istio artifacts demo test yaml istio one namespace yaml resource name may not be empty error from server invalid error when creating home bootstrap go src istio io istio artifacts demo test yaml istio one namespace yaml clusterrolebinding rbac authorization io istio pilot admin role binding is invalid subjects namespace required value error from server invalid error when creating home bootstrap go src istio io istio artifacts demo test yaml istio one namespace yaml clusterrolebinding rbac authorization io istio sidecar injector admin role binding is invalid subjects namespace required value error from server invalid error when creating home bootstrap go src istio io istio artifacts demo test yaml istio one namespace yaml clusterrolebinding rbac authorization io istio ca role binding is invalid subjects namespace required value error from server invalid error when creating home bootstrap go src istio io istio artifacts demo test yaml istio one namespace yaml clusterrolebinding rbac authorization io istio ingress admin role binding is invalid subjects namespace required value error from server invalid error when creating home bootstrap go src istio io istio artifacts demo test yaml istio one namespace yaml clusterrolebinding rbac authorization io istio sidecar role binding is invalid subjects namespace required value error from server invalid error when creating home bootstrap go src istio io istio artifacts demo test yaml istio one namespace yaml clusterrolebinding rbac authorization io istio mixer admin role binding is invalid subjects namespace required value error from server invalid error when creating home bootstrap go src istio io istio artifacts demo test yaml istio one namespace yaml clusterrolebinding rbac authorization io istio mixer validator admin role binding is invalid subjects namespace required value error from server notfound error when creating home bootstrap go src istio io istio artifacts demo test yaml istio one namespace yaml namespaces demo test not found
| 1
|
91,273
| 26,344,281,239
|
IssuesEvent
|
2023-01-10 20:29:24
|
StormSurgeLive/asgs
|
https://api.github.com/repos/StormSurgeLive/asgs
|
closed
|
clean ADCIRC patchset glut
|
IN PROGRESS: development good off-season work adcirc build support
|
**Checklist of patchsets to consolidate:**
- [x] |1 ) latest-stable | Most stable SHA of the upstream development branch (90b399f) |
- [keep ] |2 ) latest-upstream | Latest upstream development HEAD (bleeding edge) |
- [x] |3 ) noaa.stofs.2d.glo.v1.1.0r2 | upstream version tagged, noaa.stofs.2d.glo.v1.1.0r2 |
- [x] |4 ) v53release | latest upstream v53
- [x] |5 ) v53release-adcircpolate | v53release with experimental adaptive mesh refinement (TACC/ND) |
- [skip for now] |6 ) v53release-gfortran-10 | v53 with build support for gfortran 10 |
- [x] (removed) ~~|7 ) v53release-gfortran | v53 for gfortran 9 on desktop |~~
- [x] |8 ) v53release-testsuite | v53 with backported utilities for upstream testsuite |
- [x] |9 ) v55-GLOBAL | publicly available ADCIRC v55 (10.5281/zenodo.3911282) |
- [x] |10) v55.01 | upstream version v55.01 |
- [x] |11) v55.01-5bc04d6 | upstream version v55.01 SHA 5bc04d6 |
- [x] |12) v55release | short lived "v55release" |
- [x] (removed)~~|13) v55release-swan-gfortran | v55release with build support for gfortran < 10 |~~
- [skip for now] |14) v55release-swan-gfortran-10 | v55release with build support for gfortran >= 10 |
**Testing Notes:**
- [x] intel tested, working
- [x] super mike (main development environment, mpif90 linked to `gcc` by default)
- [x] qbc (had to use `module load` to get a non-ancient `gcc`
- [x] lonestar6 (working `gcc`/`gfortran` available in default environment)
**NOTE:**
* unless there is a specific patchset for _gfortran_, it can be assumed the patchset has been verified to work with `gfortran` < v10 (v10 made inverted a flag that's made it difficult to have a generic solution)
|
1.0
|
clean ADCIRC patchset glut - **Checklist of patchsets to consolidate:**
- [x] |1 ) latest-stable | Most stable SHA of the upstream development branch (90b399f) |
- [keep ] |2 ) latest-upstream | Latest upstream development HEAD (bleeding edge) |
- [x] |3 ) noaa.stofs.2d.glo.v1.1.0r2 | upstream version tagged, noaa.stofs.2d.glo.v1.1.0r2 |
- [x] |4 ) v53release | latest upstream v53
- [x] |5 ) v53release-adcircpolate | v53release with experimental adaptive mesh refinement (TACC/ND) |
- [skip for now] |6 ) v53release-gfortran-10 | v53 with build support for gfortran 10 |
- [x] (removed) ~~|7 ) v53release-gfortran | v53 for gfortran 9 on desktop |~~
- [x] |8 ) v53release-testsuite | v53 with backported utilities for upstream testsuite |
- [x] |9 ) v55-GLOBAL | publicly available ADCIRC v55 (10.5281/zenodo.3911282) |
- [x] |10) v55.01 | upstream version v55.01 |
- [x] |11) v55.01-5bc04d6 | upstream version v55.01 SHA 5bc04d6 |
- [x] |12) v55release | short lived "v55release" |
- [x] (removed)~~|13) v55release-swan-gfortran | v55release with build support for gfortran < 10 |~~
- [skip for now] |14) v55release-swan-gfortran-10 | v55release with build support for gfortran >= 10 |
**Testing Notes:**
- [x] intel tested, working
- [x] super mike (main development environment, mpif90 linked to `gcc` by default)
- [x] qbc (had to use `module load` to get a non-ancient `gcc`
- [x] lonestar6 (working `gcc`/`gfortran` available in default environment)
**NOTE:**
* unless there is a specific patchset for _gfortran_, it can be assumed the patchset has been verified to work with `gfortran` < v10 (v10 made inverted a flag that's made it difficult to have a generic solution)
|
non_test
|
clean adcirc patchset glut checklist of patchsets to consolidate latest stable most stable sha of the upstream development branch latest upstream latest upstream development head bleeding edge noaa stofs glo upstream version tagged noaa stofs glo latest upstream adcircpolate with experimental adaptive mesh refinement tacc nd gfortran with build support for gfortran removed gfortran for gfortran on desktop testsuite with backported utilities for upstream testsuite global publicly available adcirc zenodo upstream version upstream version sha short lived removed swan gfortran with build support for gfortran swan gfortran with build support for gfortran testing notes intel tested working super mike main development environment linked to gcc by default qbc had to use module load to get a non ancient gcc working gcc gfortran available in default environment note unless there is a specific patchset for gfortran it can be assumed the patchset has been verified to work with gfortran made inverted a flag that s made it difficult to have a generic solution
| 0
|
34,243
| 4,894,541,901
|
IssuesEvent
|
2016-11-19 10:25:43
|
payara/Payara
|
https://api.github.com/repos/payara/Payara
|
closed
|
JSON wrongly serialise long long values
|
c:PossibleBug Insufficient Detail Test Case Required
|
# Description #
----------
Payara generates wrong value for JSON, not XML.
## Expected Outcome
```java
public class Placement {
@XmlAttribute
private Long id;
}
```
```xml
<placement id="4611686018657026049">
</placement>
```
```json
{
"type": "placement",
"id": 4611686018657026049,
}
```
## Current Outcome
```json
{
"type": "placement",
"id": 4611686018657026000,
}
```
## Steps to reproduce (Only for bug reports)
## Context (Optional)
## Environment ##
- **Payara Version**: 4.1.1.163 #badassfish (build 215)
- **Edition**: Full
- **JDK Version**: 8 u101 - Oracle
- **Operating System**: Linux / Mac
|
1.0
|
JSON wrongly serialise long long values - # Description #
----------
Payara generates wrong value for JSON, not XML.
## Expected Outcome
```java
public class Placement {
@XmlAttribute
private Long id;
}
```
```xml
<placement id="4611686018657026049">
</placement>
```
```json
{
"type": "placement",
"id": 4611686018657026049,
}
```
## Current Outcome
```json
{
"type": "placement",
"id": 4611686018657026000,
}
```
## Steps to reproduce (Only for bug reports)
## Context (Optional)
## Environment ##
- **Payara Version**: 4.1.1.163 #badassfish (build 215)
- **Edition**: Full
- **JDK Version**: 8 u101 - Oracle
- **Operating System**: Linux / Mac
|
test
|
json wrongly serialise long long values description payara generates wrong value for json not xml expected outcome java public class placement xmlattribute private long id xml json type placement id current outcome json type placement id steps to reproduce only for bug reports context optional environment payara version badassfish build edition full jdk version oracle operating system linux mac
| 1
|
233,731
| 19,042,972,271
|
IssuesEvent
|
2021-11-25 01:48:02
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
bench/rttanalysis: TestBenchmarkExpectation failed
|
C-test-failure O-robot T-sql-schema branch-release-21.2
|
bench/rttanalysis.TestBenchmarkExpectation [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3536767&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3536767&tab=artifacts#/) on release-21.2 @ [1e0887f59f117988b9e7a30ae0946d3561247823](https://github.com/cockroachdb/cockroach/commits/1e0887f59f117988b9e7a30ae0946d3561247823):
```
BenchmarkAlterTableAddCheckConstraint/alter_table_add_3_check_constraints
BenchmarkDropTable/drop_2_tables-16 1 576383048 ns/op 27.00 roundtrips
BenchmarkDropTable/drop_3_tables
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkDropDatabase_drop_database_2_tables917572378
test_log_scope.go:80: use -show-logs to present logs inline
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkDropTable_drop_3_tables791856722
test_log_scope.go:80: use -show-logs to present logs inline
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableAddCheckConstraint_alter_table_add_3_check_constraints781278541
test_log_scope.go:80: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:74: -- test log scope end --
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableUnsplit_alter_table_unsplit_at_3_values191990782
test_log_scope.go:80: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:74: -- test log scope end --
BenchmarkCreateRole/create_role_with_1_option-16 1 528724748 ns/op 17.00 roundtrips
BenchmarkCreateRole/create_role_with_2_options
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkCreateRole_create_role_with_2_options186135541
test_log_scope.go:80: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:74: -- test log scope end --
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkSystemDatabaseQueries_select_system.users_with_empty_database_Name075038697
test_log_scope.go:80: use -show-logs to present logs inline
BenchmarkDropView/drop_2_views-16 1 316954533 ns/op 28.00 roundtrips
BenchmarkDropView/drop_3_views
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:74: -- test log scope end --
BenchmarkAlterTableDropConstraint/alter_table_drop_2_check_constraints-16 1 970332886 ns/op 17.00 roundtrips
BenchmarkAlterTableDropConstraint/alter_table_drop_3_check_constraints
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkDropView_drop_3_views868034841
test_log_scope.go:80: use -show-logs to present logs inline
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableDropConstraint_alter_table_drop_3_check_constraints631533040
test_log_scope.go:80: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableConfigureZone_alter_table_configure_zone_ranges958476862
test_log_scope.go:80: use -show-logs to present logs inline
rtt_analysis_bench.go:74: -- test log scope end --
BenchmarkVirtualTableQueries/select_crdb_internal.invalid_objects_with_1_fk-16 1 49760321 ns/op 1.000 roundtrips
=== CONT TestBenchmarkExpectation/BenchmarkTruncate
validate_benchmark_data.go:150:
Error Trace: validate_benchmark_data.go:150
validate_benchmark_data.go:88
Error: Received unexpected error:
exit status 6
Test: TestBenchmarkExpectation/BenchmarkTruncate
Messages: failed to wait for command /go/src/github.com/cockroachdb/cockroach/artifacts/go-build2030725776/b1017/rttanalysis.test --test.run=^$ --test.bench=BenchmarkTruncate --test.benchtime=1x --test.v
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:74: -- test log scope end --
BenchmarkAlterRole/alter_role_with_3_options-16 1 523498396 ns/op 21.00 roundtrips
--- FAIL: TestBenchmarkExpectation/BenchmarkTruncate (17.11s)
```
<details><summary>Reproduce</summary>
<p>
To reproduce, try:
```bash
make stressrace TESTS=TestBenchmarkExpectation PKG=./pkg/bench/rttanalysis TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
Parameters in this failure:
- GOFLAGS=-json
</p>
</details>
/cc @cockroachdb/sql-experience pawalt
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestBenchmarkExpectation.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
bench/rttanalysis: TestBenchmarkExpectation failed - bench/rttanalysis.TestBenchmarkExpectation [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3536767&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3536767&tab=artifacts#/) on release-21.2 @ [1e0887f59f117988b9e7a30ae0946d3561247823](https://github.com/cockroachdb/cockroach/commits/1e0887f59f117988b9e7a30ae0946d3561247823):
```
BenchmarkAlterTableAddCheckConstraint/alter_table_add_3_check_constraints
BenchmarkDropTable/drop_2_tables-16 1 576383048 ns/op 27.00 roundtrips
BenchmarkDropTable/drop_3_tables
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkDropDatabase_drop_database_2_tables917572378
test_log_scope.go:80: use -show-logs to present logs inline
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkDropTable_drop_3_tables791856722
test_log_scope.go:80: use -show-logs to present logs inline
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableAddCheckConstraint_alter_table_add_3_check_constraints781278541
test_log_scope.go:80: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:74: -- test log scope end --
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableUnsplit_alter_table_unsplit_at_3_values191990782
test_log_scope.go:80: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:74: -- test log scope end --
BenchmarkCreateRole/create_role_with_1_option-16 1 528724748 ns/op 17.00 roundtrips
BenchmarkCreateRole/create_role_with_2_options
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkCreateRole_create_role_with_2_options186135541
test_log_scope.go:80: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:74: -- test log scope end --
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkSystemDatabaseQueries_select_system.users_with_empty_database_Name075038697
test_log_scope.go:80: use -show-logs to present logs inline
BenchmarkDropView/drop_2_views-16 1 316954533 ns/op 28.00 roundtrips
BenchmarkDropView/drop_3_views
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:74: -- test log scope end --
BenchmarkAlterTableDropConstraint/alter_table_drop_2_check_constraints-16 1 970332886 ns/op 17.00 roundtrips
BenchmarkAlterTableDropConstraint/alter_table_drop_3_check_constraints
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkDropView_drop_3_views868034841
test_log_scope.go:80: use -show-logs to present logs inline
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableDropConstraint_alter_table_drop_3_check_constraints631533040
test_log_scope.go:80: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableConfigureZone_alter_table_configure_zone_ranges958476862
test_log_scope.go:80: use -show-logs to present logs inline
rtt_analysis_bench.go:74: -- test log scope end --
BenchmarkVirtualTableQueries/select_crdb_internal.invalid_objects_with_1_fk-16 1 49760321 ns/op 1.000 roundtrips
=== CONT TestBenchmarkExpectation/BenchmarkTruncate
validate_benchmark_data.go:150:
Error Trace: validate_benchmark_data.go:150
validate_benchmark_data.go:88
Error: Received unexpected error:
exit status 6
Test: TestBenchmarkExpectation/BenchmarkTruncate
Messages: failed to wait for command /go/src/github.com/cockroachdb/cockroach/artifacts/go-build2030725776/b1017/rttanalysis.test --test.run=^$ --test.bench=BenchmarkTruncate --test.benchtime=1x --test.v
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:74: -- test log scope end --
BenchmarkAlterRole/alter_role_with_3_options-16 1 523498396 ns/op 21.00 roundtrips
--- FAIL: TestBenchmarkExpectation/BenchmarkTruncate (17.11s)
```
<details><summary>Reproduce</summary>
<p>
To reproduce, try:
```bash
make stressrace TESTS=TestBenchmarkExpectation PKG=./pkg/bench/rttanalysis TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
Parameters in this failure:
- GOFLAGS=-json
</p>
</details>
/cc @cockroachdb/sql-experience pawalt
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestBenchmarkExpectation.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
bench rttanalysis testbenchmarkexpectation failed bench rttanalysis testbenchmarkexpectation with on release benchmarkaltertableaddcheckconstraint alter table add check constraints benchmarkdroptable drop tables ns op roundtrips benchmarkdroptable drop tables test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkdropdatabase drop database test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkdroptable drop test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertableaddcheckconstraint alter table add check test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertableunsplit alter table unsplit at test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkcreaterole create role with option ns op roundtrips benchmarkcreaterole create role with options test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkcreaterole create role with test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarksystemdatabasequeries select system users with empty database test log scope go use show logs to present logs inline benchmarkdropview drop views ns op roundtrips benchmarkdropview drop views rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkaltertabledropconstraint alter table drop check constraints ns op roundtrips benchmarkaltertabledropconstraint alter table drop check constraints test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkdropview drop test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertabledropconstraint alter table drop check test log scope go use show logs to present logs inline rtt analysis bench go test log scope end test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertableconfigurezone alter table configure zone test log scope go use show logs to present logs inline rtt analysis bench go test log scope end benchmarkvirtualtablequeries select crdb internal invalid objects with fk ns op roundtrips cont testbenchmarkexpectation benchmarktruncate validate benchmark data go error trace validate benchmark data go validate benchmark data go error received unexpected error exit status test testbenchmarkexpectation benchmarktruncate messages failed to wait for command go src github com cockroachdb cockroach artifacts go rttanalysis test test run test bench benchmarktruncate test benchtime test v rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkalterrole alter role with options ns op roundtrips fail testbenchmarkexpectation benchmarktruncate reproduce to reproduce try bash make stressrace tests testbenchmarkexpectation pkg pkg bench rttanalysis testtimeout stressflags timeout parameters in this failure goflags json cc cockroachdb sql experience pawalt
| 1
|
612,242
| 19,007,945,260
|
IssuesEvent
|
2021-11-23 04:19:38
|
zgzgorg/iam-backend
|
https://api.github.com/repos/zgzgorg/iam-backend
|
opened
|
change commitlintrc.yml to ts or js syntax for better support
|
Type:enhancement Priority:P3
|
We use commitlint for commit check. However, it looks like we don't have fully config support for the yml config file.
I start looking into this due to https://github.com/dependabot/dependabot-core/issues/2445 and try to have `ignores`, but that looks like we need to create a function to do that, but I don't believe yml support that.
Since the origin commitlint is created by javascript, we can use js or ts(TypeScript) for better support.
Outcome:
1. determinate we should use ts or js
2. convert current `commitlintrc.yml` to `commitlintrc.ts` or `commitlintrc.js`
|
1.0
|
change commitlintrc.yml to ts or js syntax for better support - We use commitlint for commit check. However, it looks like we don't have fully config support for the yml config file.
I start looking into this due to https://github.com/dependabot/dependabot-core/issues/2445 and try to have `ignores`, but that looks like we need to create a function to do that, but I don't believe yml support that.
Since the origin commitlint is created by javascript, we can use js or ts(TypeScript) for better support.
Outcome:
1. determinate we should use ts or js
2. convert current `commitlintrc.yml` to `commitlintrc.ts` or `commitlintrc.js`
|
non_test
|
change commitlintrc yml to ts or js syntax for better support we use commitlint for commit check however it looks like we don t have fully config support for the yml config file i start looking into this due to and try to have ignores but that looks like we need to create a function to do that but i don t believe yml support that since the origin commitlint is created by javascript we can use js or ts typescript for better support outcome determinate we should use ts or js convert current commitlintrc yml to commitlintrc ts or commitlintrc js
| 0
|
134,489
| 10,917,735,765
|
IssuesEvent
|
2019-11-21 15:40:14
|
kcigeospatial/Fred_Co_Land-Management
|
https://api.github.com/repos/kcigeospatial/Fred_Co_Land-Management
|
closed
|
Project - Blanket - Assign Reviewers Task Error in PDox
|
Ready for Test Env. Retest
|
Assign Reviewers Task is throwing this error message when trying to assign reviews for BPREVIEW ap type.

|
2.0
|
Project - Blanket - Assign Reviewers Task Error in PDox - Assign Reviewers Task is throwing this error message when trying to assign reviews for BPREVIEW ap type.

|
test
|
project blanket assign reviewers task error in pdox assign reviewers task is throwing this error message when trying to assign reviews for bpreview ap type
| 1
|
109,923
| 9,419,133,932
|
IssuesEvent
|
2019-04-10 21:03:38
|
TMcMeans/election-match
|
https://api.github.com/repos/TMcMeans/election-match
|
closed
|
Test ApiCalls file
|
tests
|
Use jest/enzyme for testing purposes.
- [ ] Test that fetch API has been called with the correct arguments
- [ ] Test that fetch API throws an error if something goes wrong in the fetch call (sad path)
- [ ] Test that async method returns some election results data if response is ok (happy path)
|
1.0
|
Test ApiCalls file - Use jest/enzyme for testing purposes.
- [ ] Test that fetch API has been called with the correct arguments
- [ ] Test that fetch API throws an error if something goes wrong in the fetch call (sad path)
- [ ] Test that async method returns some election results data if response is ok (happy path)
|
test
|
test apicalls file use jest enzyme for testing purposes test that fetch api has been called with the correct arguments test that fetch api throws an error if something goes wrong in the fetch call sad path test that async method returns some election results data if response is ok happy path
| 1
|
43,722
| 5,558,274,329
|
IssuesEvent
|
2017-03-24 14:22:50
|
wp-cli/wp-cli
|
https://api.github.com/repos/wp-cli/wp-cli
|
opened
|
Run PHPCS before Behat
|
scope:testing
|
As both PHPCS and Behat will block a merge on failure, we should run the much faster PHPCS first, and only run the complete Behat test suite only if PHPCS succeeded.
This avoids scenarios where you let the entire test suite to run for an hour, only to then be blocked by a missing space.
|
1.0
|
Run PHPCS before Behat - As both PHPCS and Behat will block a merge on failure, we should run the much faster PHPCS first, and only run the complete Behat test suite only if PHPCS succeeded.
This avoids scenarios where you let the entire test suite to run for an hour, only to then be blocked by a missing space.
|
test
|
run phpcs before behat as both phpcs and behat will block a merge on failure we should run the much faster phpcs first and only run the complete behat test suite only if phpcs succeeded this avoids scenarios where you let the entire test suite to run for an hour only to then be blocked by a missing space
| 1
|
201,619
| 15,215,641,050
|
IssuesEvent
|
2021-02-17 14:38:49
|
WeiXian042901/fyp_repository
|
https://api.github.com/repos/WeiXian042901/fyp_repository
|
opened
|
QA_009-Update Quest Title(invalid)
|
Acceptance Test Admin FAIL Quest
|
Test Case
- Check if the quest title can be updated with invalid inputs
Pre-Conditions
-User must be logged in
-User must be on the quest dashboard page
-User must be an administrator
Test Steps
1. Click on the edit button under the quest ”Upgrade a Tech Gadget”.
2. Leave the quest title input field empty.
Test Data
- n/a
Expected Results
-An error message should show and the quest is not updated.
Actual Results
- The user is shown a success message and directed back to the quest dashboard with the quest updated.
Pass/Fail
- Fail
|
1.0
|
QA_009-Update Quest Title(invalid) - Test Case
- Check if the quest title can be updated with invalid inputs
Pre-Conditions
-User must be logged in
-User must be on the quest dashboard page
-User must be an administrator
Test Steps
1. Click on the edit button under the quest ”Upgrade a Tech Gadget”.
2. Leave the quest title input field empty.
Test Data
- n/a
Expected Results
-An error message should show and the quest is not updated.
Actual Results
- The user is shown a success message and directed back to the quest dashboard with the quest updated.
Pass/Fail
- Fail
|
test
|
qa update quest title invalid test case check if the quest title can be updated with invalid inputs pre conditions user must be logged in user must be on the quest dashboard page user must be an administrator test steps click on the edit button under the quest ”upgrade a tech gadget” leave the quest title input field empty test data n a expected results an error message should show and the quest is not updated actual results the user is shown a success message and directed back to the quest dashboard with the quest updated pass fail fail
| 1
|
591,919
| 17,865,357,823
|
IssuesEvent
|
2021-09-06 08:47:41
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
appleid.apple.com - see bug description
|
status-needsinfo os-ios browser-firefox-ios priority-critical
|
<!-- @browser: Firefox iOS 36.0 -->
<!-- @ua_header: Mozilla/5.0 (iPhone; CPU iPhone OS 14_7_1 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/36.0 Mobile/15E148 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @extra_labels: browser-firefox-ios -->
**URL**: https://appleid.apple.com/manage/section/security/edit
**Browser / Version**: Firefox iOS 36.0
**Operating System**: iOS 14.7.1
**Tested Another Browser**: Yes Safari
**Problem type**: Something else
**Description**: Page in page
**Steps to Reproduce**:
There is always two pages like printing paper
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
appleid.apple.com - see bug description - <!-- @browser: Firefox iOS 36.0 -->
<!-- @ua_header: Mozilla/5.0 (iPhone; CPU iPhone OS 14_7_1 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/36.0 Mobile/15E148 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @extra_labels: browser-firefox-ios -->
**URL**: https://appleid.apple.com/manage/section/security/edit
**Browser / Version**: Firefox iOS 36.0
**Operating System**: iOS 14.7.1
**Tested Another Browser**: Yes Safari
**Problem type**: Something else
**Description**: Page in page
**Steps to Reproduce**:
There is always two pages like printing paper
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
appleid apple com see bug description url browser version firefox ios operating system ios tested another browser yes safari problem type something else description page in page steps to reproduce there is always two pages like printing paper browser configuration none from with ❤️
| 0
|
242,268
| 26,262,078,274
|
IssuesEvent
|
2023-01-06 08:56:32
|
helidon-io/helidon
|
https://api.github.com/repos/helidon-io/helidon
|
closed
|
Support customization of 'logout-uri' in OIDC provider
|
bug security P2
|
[//]: # "Please remove these comments"
[//]: # "Provide a general summary of the issue in the Title above"
## Environment Details
* Helidon Version: 3.0.1
* Helidon SE or Helidon MP : both
* JDK version:
* OS:
* Docker version (if applicable):
----------
## Problem Description
[//]: # "Describe the bug in detail highlighting current behavior vs expected behavior"
[//]: # "State if the problem is easily reproducible or happens intermittently"
[//]: # "Include stack traces or command outputs"
Config option is missing for 'logout-uri' in the OIDC provider. Its always defaulted to "/oidc/logout", which is blocking from supporting logout for multiple application instances in k8s.
There needs a fix in `OidcConfig.Builder::build(Config)` for customizing 'logout-uri'
## Steps to reproduce
[//]: # "Step by step instructions to reproduce the problem"
[//]: # "Provide sample code/application if relevant"
|
True
|
Support customization of 'logout-uri' in OIDC provider - [//]: # "Please remove these comments"
[//]: # "Provide a general summary of the issue in the Title above"
## Environment Details
* Helidon Version: 3.0.1
* Helidon SE or Helidon MP : both
* JDK version:
* OS:
* Docker version (if applicable):
----------
## Problem Description
[//]: # "Describe the bug in detail highlighting current behavior vs expected behavior"
[//]: # "State if the problem is easily reproducible or happens intermittently"
[//]: # "Include stack traces or command outputs"
Config option is missing for 'logout-uri' in the OIDC provider. Its always defaulted to "/oidc/logout", which is blocking from supporting logout for multiple application instances in k8s.
There needs a fix in `OidcConfig.Builder::build(Config)` for customizing 'logout-uri'
## Steps to reproduce
[//]: # "Step by step instructions to reproduce the problem"
[//]: # "Provide sample code/application if relevant"
|
non_test
|
support customization of logout uri in oidc provider please remove these comments provide a general summary of the issue in the title above environment details helidon version helidon se or helidon mp both jdk version os docker version if applicable problem description describe the bug in detail highlighting current behavior vs expected behavior state if the problem is easily reproducible or happens intermittently include stack traces or command outputs config option is missing for logout uri in the oidc provider its always defaulted to oidc logout which is blocking from supporting logout for multiple application instances in there needs a fix in oidcconfig builder build config for customizing logout uri steps to reproduce step by step instructions to reproduce the problem provide sample code application if relevant
| 0
|
680,664
| 23,281,733,635
|
IssuesEvent
|
2022-08-05 12:46:50
|
ever-co/ever-gauzy
|
https://api.github.com/repos/ever-co/ever-gauzy
|
closed
|
Bug: Detect migrations changes every time again and again
|
type: bug :bug: priority: highest
|
We are getting migrations "detection" issue every time again and again even after migrate schema for `contact.entity.ts`.
Below are the steps to reproduce issue in dev environments.
1. Build all packages using `yarn run build:package:all`.
2. Clear all DBs, like Postgres or Sqlite.
3. Run `yarn run migration:run` command from `apps/api` folder.
4. After, that run `yarn run migration:generate --name={migration name}
|
1.0
|
Bug: Detect migrations changes every time again and again - We are getting migrations "detection" issue every time again and again even after migrate schema for `contact.entity.ts`.
Below are the steps to reproduce issue in dev environments.
1. Build all packages using `yarn run build:package:all`.
2. Clear all DBs, like Postgres or Sqlite.
3. Run `yarn run migration:run` command from `apps/api` folder.
4. After, that run `yarn run migration:generate --name={migration name}
|
non_test
|
bug detect migrations changes every time again and again we are getting migrations detection issue every time again and again even after migrate schema for contact entity ts below are the steps to reproduce issue in dev environments build all packages using yarn run build package all clear all dbs like postgres or sqlite run yarn run migration run command from apps api folder after that run yarn run migration generate name migration name
| 0
|
341,248
| 30,575,847,721
|
IssuesEvent
|
2023-07-21 05:18:58
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
[internal/k8s/k8sclient] Flaky Test - TestPodClient_NamespaceToRunningPodNum
|
bug help wanted good first issue Stale priority:p3 flaky test closed as inactive
|
```
=== RUN TestPodClient_NamespaceToRunningPodNum
2022/06/16 20:05:35 NamespaceToRunningPodNum (len=1): {
default: 2
}
pod_test.go:191:
Error Trace: pod_test.go:191
Error: Should be true
Test: TestPodClient_NamespaceToRunningPodNum
--- FAIL: TestPodClient_NamespaceToRunningPodNum (0.00s)
```
Observed here: https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/6925397106?check_suite_focus=true#step:8:538
and here: https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/6925256519?check_suite_focus=true#step:8:539
|
1.0
|
[internal/k8s/k8sclient] Flaky Test - TestPodClient_NamespaceToRunningPodNum - ```
=== RUN TestPodClient_NamespaceToRunningPodNum
2022/06/16 20:05:35 NamespaceToRunningPodNum (len=1): {
default: 2
}
pod_test.go:191:
Error Trace: pod_test.go:191
Error: Should be true
Test: TestPodClient_NamespaceToRunningPodNum
--- FAIL: TestPodClient_NamespaceToRunningPodNum (0.00s)
```
Observed here: https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/6925397106?check_suite_focus=true#step:8:538
and here: https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/6925256519?check_suite_focus=true#step:8:539
|
test
|
flaky test testpodclient namespacetorunningpodnum run testpodclient namespacetorunningpodnum namespacetorunningpodnum len default pod test go error trace pod test go error should be true test testpodclient namespacetorunningpodnum fail testpodclient namespacetorunningpodnum observed here and here
| 1
|
152,732
| 12,124,702,313
|
IssuesEvent
|
2020-04-22 14:32:28
|
scylladb/scylla
|
https://api.github.com/repos/scylladb/scylla
|
closed
|
hintedhandoff_removenode_test Missing: ['hints_manager - Draining for 127.0.9.3: end']
|
bug dtest hinted-handoff
|
Seen in [dtest-debug/314](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-debug/314/testReport/hintedhandoff_additional_test/TestHintedHandoff/hintedhandoff_removenode_test)
```
tacktrace
Traceback (most recent call last):
File "/usr/lib64/python2.7/unittest/case.py", line 367, in run
testMethod()
File "/jenkins/workspace/scylla-master/dtest-debug/scylla-dtest/hintedhandoff_additional_test.py", line 127, in hintedhandoff_removenode_test
node1.watch_log_for(msg, from_mark=from_mark, timeout=timeout)
File "/jenkins/workspace/scylla-master/dtest-debug/scylla-ccm/ccmlib/node.py", line 435, in watch_log_for
[e.pattern for e in tofind]) + ":\n" + reads[:50] + ".....\nSee {} for remainder".format(filename))
TimeoutError: 21 Nov 2019 05:04:14 [node1] Missing: ['hints_manager - Draining for 127.0.9.3: end']:
```
Scylla version 9a686ac55193f0df4fc923b267b7b36684565a9d
Scylla-dtest version 78c93908540fcf5f05ed48d1845005a027b4c904
(That is scylladb/scylla-dtest#1134 fixing #5244)
[node1.log](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-debug/314/artifact/logs-debug.2/1574312654267_hintedhandoff_additional_test.TestHintedHandoff.hintedhandoff_removenode_test/node1.log) contains only:
```
TRACE 2019-11-21 05:03:41,849 [shard 0] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:41,849 [shard 0] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:41,853 [shard 1] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:41,853 [shard 1] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
INFO 2019-11-21 05:03:41,873 [shard 0] gossip - Completing removal of 127.0.9.3
INFO 2019-11-21 05:03:41,874 [shard 0] gossip - Node 127.0.9.3 will be removed from gossip at [2019-11-24 05:03:41]: (expire = 1574571821871, now = 1574312621871, diff = 259200 seconds)
```
While [node2.log](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-debug/314/artifact/logs-debug.2/1574312654267_hintedhandoff_additional_test.TestHintedHandoff.hintedhandoff_removenode_test/node2.log):
```
TRACE 2019-11-21 05:03:42,691 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:42,695 [shard 0] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:42,695 [shard 0] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:42,700 [shard 1] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:42,700 [shard 1] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:42,711 [shard 0] hints_manager - send_one_file(): error while sending hints from /jenkins/workspace/scylla-master/dtest-debug/scylla/.dtest/dtest-H2TyrE/test/node2/hints/0/127.0.9.3/HintsLog-1-331420519.log, last RP is {0, 331420519, 28}
TRACE 2019-11-21 05:03:42,711 [shard 0] hints_manager - send_hints(): we handled 0 segments
INFO 2019-11-21 05:03:42,837 [shard 1] compaction - Compacted 2 sstables to [/jenkins/workspace/scylla-master/dtest-debug/scylla/.dtest/dtest-H2TyrE/test/node2/data/system/peers-37f71aca7dc2383ba70672528af04d4f/mc-15-big-Data.db:level=0, ]. 28613 bytes to 17642 (~61% of original) in 145ms = 0.12MB/s. ~256 total partitions merged to 2.
TRACE 2019-11-21 05:03:43,712 [shard 0] hints_manager - Draining for 127.0.9.3: start
TRACE 2019-11-21 05:03:43,713 [shard 0] hints_manager - send_hints(): going to send hints to 127.0.9.3, we have 1 segment to replay
TRACE 2019-11-21 05:03:43,725 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,728 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,731 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,734 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,736 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,738 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,741 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,744 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,746 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,749 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,752 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,758 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,762 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,765 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,767 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,769 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,772 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,776 [shard 0] hints_manager - send_one_file(): segment /jenkins/workspace/scylla-master/dtest-debug/scylla/.dtest/dtest-H2TyrE/test/node2/hints/0/127.0.9.3/HintsLog-1-331420519.log was sent in full and deleted
TRACE 2019-11-21 05:03:43,776 [shard 0] hints_manager - send_hints(): we handled 1 segments
TRACE 2019-11-21 05:03:43,778 [shard 0] hints_manager - Going to add a store to /jenkins/workspace/scylla-master/dtest-debug/scylla/.dtest/dtest-H2TyrE/test/node2/hints/0/127.0.9.3
TRACE 2019-11-21 05:03:43,786 [shard 0] hints_manager - send_hints(): going to send hints to 127.0.9.3, we have 0 segment to replay
TRACE 2019-11-21 05:03:43,786 [shard 0] hints_manager - send_hints(): we handled 0 segments
TRACE 2019-11-21 05:03:43,786 [shard 0] hints_manager - Draining for 127.0.9.3: end
TRACE 2019-11-21 05:03:43,786 [shard 0] hints_manager - ep_manager(127.0.9.3)::sender: exiting
```
Now, even if node1 had no hints to send, we should still see `ep_manager(127.0.9.3)::sender: exiting` (maybe the test needs to wait for that in the log instead, but this is a different issue)
|
1.0
|
hintedhandoff_removenode_test Missing: ['hints_manager - Draining for 127.0.9.3: end'] - Seen in [dtest-debug/314](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-debug/314/testReport/hintedhandoff_additional_test/TestHintedHandoff/hintedhandoff_removenode_test)
```
tacktrace
Traceback (most recent call last):
File "/usr/lib64/python2.7/unittest/case.py", line 367, in run
testMethod()
File "/jenkins/workspace/scylla-master/dtest-debug/scylla-dtest/hintedhandoff_additional_test.py", line 127, in hintedhandoff_removenode_test
node1.watch_log_for(msg, from_mark=from_mark, timeout=timeout)
File "/jenkins/workspace/scylla-master/dtest-debug/scylla-ccm/ccmlib/node.py", line 435, in watch_log_for
[e.pattern for e in tofind]) + ":\n" + reads[:50] + ".....\nSee {} for remainder".format(filename))
TimeoutError: 21 Nov 2019 05:04:14 [node1] Missing: ['hints_manager - Draining for 127.0.9.3: end']:
```
Scylla version 9a686ac55193f0df4fc923b267b7b36684565a9d
Scylla-dtest version 78c93908540fcf5f05ed48d1845005a027b4c904
(That is scylladb/scylla-dtest#1134 fixing #5244)
[node1.log](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-debug/314/artifact/logs-debug.2/1574312654267_hintedhandoff_additional_test.TestHintedHandoff.hintedhandoff_removenode_test/node1.log) contains only:
```
TRACE 2019-11-21 05:03:41,849 [shard 0] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:41,849 [shard 0] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:41,853 [shard 1] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:41,853 [shard 1] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
INFO 2019-11-21 05:03:41,873 [shard 0] gossip - Completing removal of 127.0.9.3
INFO 2019-11-21 05:03:41,874 [shard 0] gossip - Node 127.0.9.3 will be removed from gossip at [2019-11-24 05:03:41]: (expire = 1574571821871, now = 1574312621871, diff = 259200 seconds)
```
While [node2.log](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-debug/314/artifact/logs-debug.2/1574312654267_hintedhandoff_additional_test.TestHintedHandoff.hintedhandoff_removenode_test/node2.log):
```
TRACE 2019-11-21 05:03:42,691 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:42,695 [shard 0] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:42,695 [shard 0] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:42,700 [shard 1] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:42,700 [shard 1] hints_manager - on_leave_cluster: 127.0.9.3 is removed/decommissioned
TRACE 2019-11-21 05:03:42,711 [shard 0] hints_manager - send_one_file(): error while sending hints from /jenkins/workspace/scylla-master/dtest-debug/scylla/.dtest/dtest-H2TyrE/test/node2/hints/0/127.0.9.3/HintsLog-1-331420519.log, last RP is {0, 331420519, 28}
TRACE 2019-11-21 05:03:42,711 [shard 0] hints_manager - send_hints(): we handled 0 segments
INFO 2019-11-21 05:03:42,837 [shard 1] compaction - Compacted 2 sstables to [/jenkins/workspace/scylla-master/dtest-debug/scylla/.dtest/dtest-H2TyrE/test/node2/data/system/peers-37f71aca7dc2383ba70672528af04d4f/mc-15-big-Data.db:level=0, ]. 28613 bytes to 17642 (~61% of original) in 145ms = 0.12MB/s. ~256 total partitions merged to 2.
TRACE 2019-11-21 05:03:43,712 [shard 0] hints_manager - Draining for 127.0.9.3: start
TRACE 2019-11-21 05:03:43,713 [shard 0] hints_manager - send_hints(): going to send hints to 127.0.9.3, we have 1 segment to replay
TRACE 2019-11-21 05:03:43,725 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,728 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,731 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,734 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,736 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,738 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,741 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,744 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,746 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,749 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,752 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,758 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,762 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,765 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,767 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,769 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,772 [shard 0] hints_manager - Endpoints set has changed and 127.0.9.3 is no longer a replica. Mutating from scratch...
TRACE 2019-11-21 05:03:43,776 [shard 0] hints_manager - send_one_file(): segment /jenkins/workspace/scylla-master/dtest-debug/scylla/.dtest/dtest-H2TyrE/test/node2/hints/0/127.0.9.3/HintsLog-1-331420519.log was sent in full and deleted
TRACE 2019-11-21 05:03:43,776 [shard 0] hints_manager - send_hints(): we handled 1 segments
TRACE 2019-11-21 05:03:43,778 [shard 0] hints_manager - Going to add a store to /jenkins/workspace/scylla-master/dtest-debug/scylla/.dtest/dtest-H2TyrE/test/node2/hints/0/127.0.9.3
TRACE 2019-11-21 05:03:43,786 [shard 0] hints_manager - send_hints(): going to send hints to 127.0.9.3, we have 0 segment to replay
TRACE 2019-11-21 05:03:43,786 [shard 0] hints_manager - send_hints(): we handled 0 segments
TRACE 2019-11-21 05:03:43,786 [shard 0] hints_manager - Draining for 127.0.9.3: end
TRACE 2019-11-21 05:03:43,786 [shard 0] hints_manager - ep_manager(127.0.9.3)::sender: exiting
```
Now, even if node1 had no hints to send, we should still see `ep_manager(127.0.9.3)::sender: exiting` (maybe the test needs to wait for that in the log instead, but this is a different issue)
|
test
|
hintedhandoff removenode test missing seen in tacktrace traceback most recent call last file usr unittest case py line in run testmethod file jenkins workspace scylla master dtest debug scylla dtest hintedhandoff additional test py line in hintedhandoff removenode test watch log for msg from mark from mark timeout timeout file jenkins workspace scylla master dtest debug scylla ccm ccmlib node py line in watch log for n reads nsee for remainder format filename timeouterror nov missing scylla version scylla dtest version that is scylladb scylla dtest fixing contains only trace hints manager on leave cluster is removed decommissioned trace hints manager on leave cluster is removed decommissioned trace hints manager on leave cluster is removed decommissioned trace hints manager on leave cluster is removed decommissioned info gossip completing removal of info gossip node will be removed from gossip at expire now diff seconds while trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager on leave cluster is removed decommissioned trace hints manager on leave cluster is removed decommissioned trace hints manager on leave cluster is removed decommissioned trace hints manager on leave cluster is removed decommissioned trace hints manager send one file error while sending hints from jenkins workspace scylla master dtest debug scylla dtest dtest test hints hintslog log last rp is trace hints manager send hints we handled segments info compaction compacted sstables to bytes to of original in s total partitions merged to trace hints manager draining for start trace hints manager send hints going to send hints to we have segment to replay trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager endpoints set has changed and is no longer a replica mutating from scratch trace hints manager send one file segment jenkins workspace scylla master dtest debug scylla dtest dtest test hints hintslog log was sent in full and deleted trace hints manager send hints we handled segments trace hints manager going to add a store to jenkins workspace scylla master dtest debug scylla dtest dtest test hints trace hints manager send hints going to send hints to we have segment to replay trace hints manager send hints we handled segments trace hints manager draining for end trace hints manager ep manager sender exiting now even if had no hints to send we should still see ep manager sender exiting maybe the test needs to wait for that in the log instead but this is a different issue
| 1
|
686,866
| 23,506,890,137
|
IssuesEvent
|
2022-08-18 13:22:17
|
unfoldingWord/translationCore
|
https://api.github.com/repos/unfoldingWord/translationCore
|
closed
|
Make use of original language resources from unfoldingWord org
|
QA/Pass Priority/High
|
Suggested by @richmahn that we should remove dependency on Door43-Catalog for original languages in the case when we are using resources from other orgs. He suggested that all orgs other than Door43-Catalog should use the original languages in unfoldingWord org. This would protect in future when we stop depending on the catalog.
Maybe future task would be to allow all orgs to have original language resources, with fallback to uW?
|
1.0
|
Make use of original language resources from unfoldingWord org - Suggested by @richmahn that we should remove dependency on Door43-Catalog for original languages in the case when we are using resources from other orgs. He suggested that all orgs other than Door43-Catalog should use the original languages in unfoldingWord org. This would protect in future when we stop depending on the catalog.
Maybe future task would be to allow all orgs to have original language resources, with fallback to uW?
|
non_test
|
make use of original language resources from unfoldingword org suggested by richmahn that we should remove dependency on catalog for original languages in the case when we are using resources from other orgs he suggested that all orgs other than catalog should use the original languages in unfoldingword org this would protect in future when we stop depending on the catalog maybe future task would be to allow all orgs to have original language resources with fallback to uw
| 0
|
1,862
| 3,163,952,155
|
IssuesEvent
|
2015-09-20 19:21:37
|
pydata/pandas
|
https://api.github.com/repos/pydata/pandas
|
closed
|
PERF: infer_datetime_format fails without zero-padding
|
Difficulty Novice Effort Low Performance Timeseries
|
I think this can be fixed by just padding tokens in the inference step. The dates *do* parse, just always fall back to the slow path, even with `infer_datetime_format=True`
In [22]: from pandas.tseries.tools import _guess_datetime_format
In [23]: _guess_datetime_format('02/02/2014')
Out[23]: '%m/%d/%Y'
In [24]: _guess_datetime_format('2/2/2014')
|
True
|
PERF: infer_datetime_format fails without zero-padding - I think this can be fixed by just padding tokens in the inference step. The dates *do* parse, just always fall back to the slow path, even with `infer_datetime_format=True`
In [22]: from pandas.tseries.tools import _guess_datetime_format
In [23]: _guess_datetime_format('02/02/2014')
Out[23]: '%m/%d/%Y'
In [24]: _guess_datetime_format('2/2/2014')
|
non_test
|
perf infer datetime format fails without zero padding i think this can be fixed by just padding tokens in the inference step the dates do parse just always fall back to the slow path even with infer datetime format true in from pandas tseries tools import guess datetime format in guess datetime format out m d y in guess datetime format
| 0
|
350,206
| 31,862,416,021
|
IssuesEvent
|
2023-09-15 11:56:50
|
WPChill/strong-testimonials
|
https://api.github.com/repos/WPChill/strong-testimonials
|
closed
|
dismiss button for "the license has expired" notice
|
enhancement need testing
|
Is your feature request related to a problem? Please describe.
I don't plan on renewing yet but there is a banner message saying it has expired with no way of dismissing it, it shows on every wordpress admin page. How do I remove the message and can you add a dismiss button for future versions, it really is very intrusive.
https://secure.helpscout.net/conversation/2314834383/58689?folderId=4656052
https://github.com/WPChill/strong-testimonials-pro/issues/115
|
1.0
|
dismiss button for "the license has expired" notice - Is your feature request related to a problem? Please describe.
I don't plan on renewing yet but there is a banner message saying it has expired with no way of dismissing it, it shows on every wordpress admin page. How do I remove the message and can you add a dismiss button for future versions, it really is very intrusive.
https://secure.helpscout.net/conversation/2314834383/58689?folderId=4656052
https://github.com/WPChill/strong-testimonials-pro/issues/115
|
test
|
dismiss button for the license has expired notice is your feature request related to a problem please describe i don t plan on renewing yet but there is a banner message saying it has expired with no way of dismissing it it shows on every wordpress admin page how do i remove the message and can you add a dismiss button for future versions it really is very intrusive
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.