Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 4 112 | repo_url stringlengths 33 141 | action stringclasses 3 values | title stringlengths 1 1.02k | labels stringlengths 4 1.54k | body stringlengths 1 262k | index stringclasses 17 values | text_combine stringlengths 95 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
30,113 | 7,163,242,572 | IssuesEvent | 2018-01-29 06:13:53 | cquery-project/cquery | https://api.github.com/repos/cquery-project/cquery | closed | [VSCode] Semantic highlighting does not work for files whose path contains "+" symbol | vscode | Semantic highlighting is working fine in some workspaces, but not at all in others.
In the problematic workspace, I do not see any errors in the output window or the log file, but I do see lines like the following:
```
2018-01-06 19:34:51.858 ( 308.410s) [stdout ] timer.cc:38 0| [e2e] Running $cquery/publishSemanticHighlighting took 260173.259913790ms
```
That duration, if accurate, seems excessively long. Perhaps this is the cause of the problem, and the frontend (I'm using VSCode) is timing out or something? | 1.0 | [VSCode] Semantic highlighting does not work for files whose path contains "+" symbol - Semantic highlighting is working fine in some workspaces, but not at all in others.
In the problematic workspace, I do not see any errors in the output window or the log file, but I do see lines like the following:
```
2018-01-06 19:34:51.858 ( 308.410s) [stdout ] timer.cc:38 0| [e2e] Running $cquery/publishSemanticHighlighting took 260173.259913790ms
```
That duration, if accurate, seems excessively long. Perhaps this is the cause of the problem, and the frontend (I'm using VSCode) is timing out or something? | non_test | semantic highlighting does not work for files whose path contains symbol semantic highlighting is working fine in some workspaces but not at all in others in the problematic workspace i do not see any errors in the output window or the log file but i do see lines like the following timer cc running cquery publishsemantichighlighting took that duration if accurate seems excessively long perhaps this is the cause of the problem and the frontend i m using vscode is timing out or something | 0 |
33,580 | 16,037,617,021 | IssuesEvent | 2021-04-22 00:58:13 | unitystation/unitystation | https://api.github.com/repos/unitystation/unitystation | opened | Garbage reduction and crash fixes bundle B: special, see details | Bounty Type: Performance Type: Refactor |
## Description
**This is a stability related bounty, and is immune to the feature freeze.** This bounty bundle calls for reducing gc, improving performance of the game, and finding and killing crash causes.
This is not a traditional bounty. This bounty need not be assigned, and can be done by anyone, though it is still held to the standard bounty guidelines in the [contribution guide.](https://unitystation.github.io/unitystation/CONTRIBUTING/)
Any contributor may claim a reward for doing the following in a PR:
- substantially reducing gc of a system or method that is utilized often in the game.
- improving performance or speed of a system or that is utilized often in the game.
- fixing any bug that could cause either the host or clients to crash.
- fixing any bug that could prevent the game from properly launching, or initializing in a state that renders the game unplayable.
**the reward for fulfilling this bounty is set at 25.00 USD.**
This bounty is meant to encourage collective action. We want to get **as many contributors as possible** to work on game performance right now so that we may lift the [feature freeze](https://github.com/unitystation/unitystation/issues/6275) as soon as possible.
To give as many contributors the opportunity to get the reward as possible, **EACH CONTRIBUTOR MAY CLAIM A REWARD FROM THIS BOUNTY ONLY ONCE.**
We have allocated 250.00 USD for this bounty, providing **10** individual rewards to be claimed. Each Pull Request will be evaluated on request by @Bod9001 and @corp-0 as to whether the PR qualifies for a reward from this bounty bundle. If you have any questions about what _might_ qualify for a reward, please ask in the #bounties channel in the [unitystation discord.](https://discord.gg/WcF4cbxr)
We are opening this bounty opportunity retroactively to all Pull Requests merged since the feature freeze was put in place (Tuesday, March 23rd). If you would like to claim on a PR that has been merged, you can request an evaluation in a comment under that PR. | True | Garbage reduction and crash fixes bundle B: special, see details -
## Description
**This is a stability related bounty, and is immune to the feature freeze.** This bounty bundle calls for reducing gc, improving performance of the game, and finding and killing crash causes.
This is not a traditional bounty. This bounty need not be assigned, and can be done by anyone, though it is still held to the standard bounty guidelines in the [contribution guide.](https://unitystation.github.io/unitystation/CONTRIBUTING/)
Any contributor may claim a reward for doing the following in a PR:
- substantially reducing gc of a system or method that is utilized often in the game.
- improving performance or speed of a system or that is utilized often in the game.
- fixing any bug that could cause either the host or clients to crash.
- fixing any bug that could prevent the game from properly launching, or initializing in a state that renders the game unplayable.
**the reward for fulfilling this bounty is set at 25.00 USD.**
This bounty is meant to encourage collective action. We want to get **as many contributors as possible** to work on game performance right now so that we may lift the [feature freeze](https://github.com/unitystation/unitystation/issues/6275) as soon as possible.
To give as many contributors the opportunity to get the reward as possible, **EACH CONTRIBUTOR MAY CLAIM A REWARD FROM THIS BOUNTY ONLY ONCE.**
We have allocated 250.00 USD for this bounty, providing **10** individual rewards to be claimed. Each Pull Request will be evaluated on request by @Bod9001 and @corp-0 as to whether the PR qualifies for a reward from this bounty bundle. If you have any questions about what _might_ qualify for a reward, please ask in the #bounties channel in the [unitystation discord.](https://discord.gg/WcF4cbxr)
We are opening this bounty opportunity retroactively to all Pull Requests merged since the feature freeze was put in place (Tuesday, March 23rd). If you would like to claim on a PR that has been merged, you can request an evaluation in a comment under that PR. | non_test | garbage reduction and crash fixes bundle b special see details description this is a stability related bounty and is immune to the feature freeze this bounty bundle calls for reducing gc improving performance of the game and finding and killing crash causes this is not a traditional bounty this bounty need not be assigned and can be done by anyone though it is still held to the standard bounty guidelines in the any contributor may claim a reward for doing the following in a pr substantially reducing gc of a system or method that is utilized often in the game improving performance or speed of a system or that is utilized often in the game fixing any bug that could cause either the host or clients to crash fixing any bug that could prevent the game from properly launching or initializing in a state that renders the game unplayable the reward for fulfilling this bounty is set at usd this bounty is meant to encourage collective action we want to get as many contributors as possible to work on game performance right now so that we may lift the as soon as possible to give as many contributors the opportunity to get the reward as possible each contributor may claim a reward from this bounty only once we have allocated usd for this bounty providing individual rewards to be claimed each pull request will be evaluated on request by and corp as to whether the pr qualifies for a reward from this bounty bundle if you have any questions about what might qualify for a reward please ask in the bounties channel in the we are opening this bounty opportunity retroactively to all pull requests merged since the feature freeze was put in place tuesday march if you would like to claim on a pr that has been merged you can request an evaluation in a comment under that pr | 0 |
326,349 | 9,955,383,143 | IssuesEvent | 2019-07-05 10:52:55 | mozilla/addons-frontend | https://api.github.com/repos/mozilla/addons-frontend | closed | Storybook: do we really need `core/css/inc/lib.scss` in `stories/index.js`? | priority: p3 storybook type: question | The file `core/css/inc/lib.scss` is already imported in `amo/components/App/styles.scss`, which is required in `stories/index.js`.
Do we really need to add it one more time? | 1.0 | Storybook: do we really need `core/css/inc/lib.scss` in `stories/index.js`? - The file `core/css/inc/lib.scss` is already imported in `amo/components/App/styles.scss`, which is required in `stories/index.js`.
Do we really need to add it one more time? | non_test | storybook do we really need core css inc lib scss in stories index js the file core css inc lib scss is already imported in amo components app styles scss which is required in stories index js do we really need to add it one more time | 0 |
247,646 | 20,987,403,850 | IssuesEvent | 2022-03-29 05:40:25 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: follower-reads/survival=zone/locality=regional/reads=exact-staleness failed | C-test-failure O-robot O-roachtest branch-master release-blocker | roachtest.follower-reads/survival=zone/locality=regional/reads=exact-staleness [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713654&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713654&tab=artifacts#/follower-reads/survival=zone/locality=regional/reads=exact-staleness) on master @ [29716850b181718594663889ddb5f479fef7a305](https://github.com/cockroachdb/cockroach/commits/29716850b181718594663889ddb5f479fef7a305):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /artifacts/follower-reads/survival=zone/locality=regional/reads=exact-staleness/run_1
cluster.go:1868,follower_reads.go:64,test_runner.go:875: one or more parallel execution failure
(1) attached stack trace
-- stack trace:
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2042
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:1923
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:167
| github.com/cockroachdb/cockroach/pkg/roachprod.Start
| github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:660
| main.(*clusterImpl).StartE
| main/pkg/cmd/roachtest/cluster.go:1826
| main.(*clusterImpl).Start
| main/pkg/cmd/roachtest/cluster.go:1867
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerFollowerReads.func1.1
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/follower_reads.go:64
| main.(*testRunner).runTest.func2
| main/pkg/cmd/roachtest/test_runner.go:875
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1581
Wraps: (2) one or more parallel execution failure
Error types: (1) *withstack.withStack (2) *errutil.leafError
```
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*follower-reads/survival=zone/locality=regional/reads=exact-staleness.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: follower-reads/survival=zone/locality=regional/reads=exact-staleness failed - roachtest.follower-reads/survival=zone/locality=regional/reads=exact-staleness [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713654&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713654&tab=artifacts#/follower-reads/survival=zone/locality=regional/reads=exact-staleness) on master @ [29716850b181718594663889ddb5f479fef7a305](https://github.com/cockroachdb/cockroach/commits/29716850b181718594663889ddb5f479fef7a305):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /artifacts/follower-reads/survival=zone/locality=regional/reads=exact-staleness/run_1
cluster.go:1868,follower_reads.go:64,test_runner.go:875: one or more parallel execution failure
(1) attached stack trace
-- stack trace:
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2042
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:1923
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:167
| github.com/cockroachdb/cockroach/pkg/roachprod.Start
| github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:660
| main.(*clusterImpl).StartE
| main/pkg/cmd/roachtest/cluster.go:1826
| main.(*clusterImpl).Start
| main/pkg/cmd/roachtest/cluster.go:1867
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerFollowerReads.func1.1
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/follower_reads.go:64
| main.(*testRunner).runTest.func2
| main/pkg/cmd/roachtest/test_runner.go:875
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1581
Wraps: (2) one or more parallel execution failure
Error types: (1) *withstack.withStack (2) *errutil.leafError
```
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*follower-reads/survival=zone/locality=regional/reads=exact-staleness.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| test | roachtest follower reads survival zone locality regional reads exact staleness failed roachtest follower reads survival zone locality regional reads exact staleness with on master the test failed on branch master cloud gce test artifacts and logs in artifacts follower reads survival zone locality regional reads exact staleness run cluster go follower reads go test runner go one or more parallel execution failure attached stack trace stack trace github com cockroachdb cockroach pkg roachprod install syncedcluster parallele github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster parallel github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster start github com cockroachdb cockroach pkg roachprod install cockroach go github com cockroachdb cockroach pkg roachprod start github com cockroachdb cockroach pkg roachprod roachprod go main clusterimpl starte main pkg cmd roachtest cluster go main clusterimpl start main pkg cmd roachtest cluster go github com cockroachdb cockroach pkg cmd roachtest tests registerfollowerreads github com cockroachdb cockroach pkg cmd roachtest tests follower reads go main testrunner runtest main pkg cmd roachtest test runner go runtime goexit goroot src runtime asm s wraps one or more parallel execution failure error types withstack withstack errutil leaferror help see see cc cockroachdb kv triage | 1 |
117,369 | 9,932,812,981 | IssuesEvent | 2019-07-02 10:44:17 | jdev-org/ddv-viewer | https://api.github.com/repos/jdev-org/ddv-viewer | closed | Mettre à jour la couche client avec un fichier csv | To test | En tant qu'utilisateur,
Je souhaite mettre à jour les informations de la carte par l'import d'un fichier CSV qui remplacera les anciennes données
Afin de visualiser les informations clientes à jour et en toute autonomie.
Le fichier CSV remplacera l'ancien dans sa totalité.
Une sauvegarde du fichier M-1 peut être envisagée sur le serveur dans un autre répertoire ou sous un autre nom en cas d'erreur d'import.
Le fichier importer doit comporter le même nom que la couche et que le fichier présent sur le serveur. | 1.0 | Mettre à jour la couche client avec un fichier csv - En tant qu'utilisateur,
Je souhaite mettre à jour les informations de la carte par l'import d'un fichier CSV qui remplacera les anciennes données
Afin de visualiser les informations clientes à jour et en toute autonomie.
Le fichier CSV remplacera l'ancien dans sa totalité.
Une sauvegarde du fichier M-1 peut être envisagée sur le serveur dans un autre répertoire ou sous un autre nom en cas d'erreur d'import.
Le fichier importer doit comporter le même nom que la couche et que le fichier présent sur le serveur. | test | mettre à jour la couche client avec un fichier csv en tant qu utilisateur je souhaite mettre à jour les informations de la carte par l import d un fichier csv qui remplacera les anciennes données afin de visualiser les informations clientes à jour et en toute autonomie le fichier csv remplacera l ancien dans sa totalité une sauvegarde du fichier m peut être envisagée sur le serveur dans un autre répertoire ou sous un autre nom en cas d erreur d import le fichier importer doit comporter le même nom que la couche et que le fichier présent sur le serveur | 1 |
93,144 | 10,764,544,789 | IssuesEvent | 2019-11-01 08:36:58 | Davidcwh/ped | https://api.github.com/repos/Davidcwh/ped | opened | Incorrect delete/remove command in User Gudie | severity.High type.DocumentationBug | User guide states `remove 1` command to delete a food item:

But it is an invalid command:

| 1.0 | Incorrect delete/remove command in User Gudie - User guide states `remove 1` command to delete a food item:

But it is an invalid command:

| non_test | incorrect delete remove command in user gudie user guide states remove command to delete a food item but it is an invalid command | 0 |
638,686 | 20,734,703,234 | IssuesEvent | 2022-03-14 12:44:09 | ballerina-platform/ballerina-dev-website | https://api.github.com/repos/ballerina-platform/ballerina-dev-website | opened | Integrate the React Engine to B.io | Priority/Highest Area/Backend Type/NewFeature | **Description:**
Integrate the React Engine to B.io.
**Describe your problem(s)**
**Describe your solution(s)**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| 1.0 | Integrate the React Engine to B.io - **Description:**
Integrate the React Engine to B.io.
**Describe your problem(s)**
**Describe your solution(s)**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| non_test | integrate the react engine to b io description integrate the react engine to b io describe your problem s describe your solution s related issues optional suggested labels optional suggested assignees optional | 0 |
167,808 | 13,043,419,089 | IssuesEvent | 2020-07-29 01:30:14 | huggingface/transformers | https://api.github.com/repos/huggingface/transformers | closed | CI: run tests against torch=1.6 | Help wanted Tests | Through github or circleci.
If github actions:
copy `.github/self-scheduled.yml` to `.github/torch_future.yml` and modify the install steps. | 1.0 | CI: run tests against torch=1.6 - Through github or circleci.
If github actions:
copy `.github/self-scheduled.yml` to `.github/torch_future.yml` and modify the install steps. | test | ci run tests against torch through github or circleci if github actions copy github self scheduled yml to github torch future yml and modify the install steps | 1 |
147,721 | 11,802,714,446 | IssuesEvent | 2020-03-18 22:11:49 | flutter/flutter | https://api.github.com/repos/flutter/flutter | closed | Enable web golden tests | a: tests team: flakes ☸ platform-web | The tests fail with "Unknown error loading". For example:
```
02:03 +118 ~33 -1: compiling /tmp/flutter sdk/packages/flutter/test/widgets/list_view_vertical_test.dart [E]
Failed to load "/tmp/flutter sdk/packages/flutter/test/widgets/list_view_vertical_test.dart": Unknown error loading http://localhost:43947/packages/flutter/material.ddc.js
```
Seen in the following, but I've seen it multiple times lately.
https://cirrus-ci.com/task/5055207964147712?command=main#L759
https://cirrus-ci.com/task/5899632894279680?command=main#L2370
/cc @jonahwilliams and @yjbanov who may know what's going on. | 1.0 | Enable web golden tests - The tests fail with "Unknown error loading". For example:
```
02:03 +118 ~33 -1: compiling /tmp/flutter sdk/packages/flutter/test/widgets/list_view_vertical_test.dart [E]
Failed to load "/tmp/flutter sdk/packages/flutter/test/widgets/list_view_vertical_test.dart": Unknown error loading http://localhost:43947/packages/flutter/material.ddc.js
```
Seen in the following, but I've seen it multiple times lately.
https://cirrus-ci.com/task/5055207964147712?command=main#L759
https://cirrus-ci.com/task/5899632894279680?command=main#L2370
/cc @jonahwilliams and @yjbanov who may know what's going on. | test | enable web golden tests the tests fail with unknown error loading for example compiling tmp flutter sdk packages flutter test widgets list view vertical test dart failed to load tmp flutter sdk packages flutter test widgets list view vertical test dart unknown error loading seen in the following but i ve seen it multiple times lately cc jonahwilliams and yjbanov who may know what s going on | 1 |
169,323 | 13,133,762,505 | IssuesEvent | 2020-08-06 21:36:53 | ESCOMP/CTSM | https://api.github.com/repos/ESCOMP/CTSM | closed | Failing ISSP585Clm50BgcCrop test for ARCTICGRIS | type: bug type: tests | ### Brief summary of bug
The test SMS.ne0ARCTICGRISne30x8_ne0ARCTICGRISne30x8_mt12.ISSP585Clm50BgcCrop.cheyenne_intel.clm-clm50cam6LndTuningMode
failed in ctsm1.0.dev105. It's a new test added in, and for some reason I missed the fact that it wasn't passing.
### General bug information
**CTSM version you are using:** ctsm1.0.dev105
**Does this bug cause significantly incorrect results in the model's science?** No
**Configurations affected:** Just this test
### Details of bug
I think the walltime is just too short.
| 1.0 | Failing ISSP585Clm50BgcCrop test for ARCTICGRIS - ### Brief summary of bug
The test SMS.ne0ARCTICGRISne30x8_ne0ARCTICGRISne30x8_mt12.ISSP585Clm50BgcCrop.cheyenne_intel.clm-clm50cam6LndTuningMode
failed in ctsm1.0.dev105. It's a new test added in, and for some reason I missed the fact that it wasn't passing.
### General bug information
**CTSM version you are using:** ctsm1.0.dev105
**Does this bug cause significantly incorrect results in the model's science?** No
**Configurations affected:** Just this test
### Details of bug
I think the walltime is just too short.
| test | failing test for arcticgris brief summary of bug the test sms cheyenne intel clm failed in it s a new test added in and for some reason i missed the fact that it wasn t passing general bug information ctsm version you are using does this bug cause significantly incorrect results in the model s science no configurations affected just this test details of bug i think the walltime is just too short | 1 |
626,667 | 19,830,687,810 | IssuesEvent | 2022-01-20 11:39:04 | GoldenSoftwareLtd/gedemin | https://api.github.com/repos/GoldenSoftwareLtd/gedemin | closed | Рапорт о выработке продукции ККЦ печатает не верные данные | Type-Enhancement Priority-Medium Meat | Originally reported on Google Code with ID 2575
```
В колонку Норма выхода попадает значение из предыдущей позиции;
В колонки закладка сырья и закладка специй заполняются с накоплением, т.е. значение
второй позиции содержит значение 1й+2й и так далее...
```
Reported by `stasgm` on 2011-09-12 13:27:20
| 1.0 | Рапорт о выработке продукции ККЦ печатает не верные данные - Originally reported on Google Code with ID 2575
```
В колонку Норма выхода попадает значение из предыдущей позиции;
В колонки закладка сырья и закладка специй заполняются с накоплением, т.е. значение
второй позиции содержит значение 1й+2й и так далее...
```
Reported by `stasgm` on 2011-09-12 13:27:20
| non_test | рапорт о выработке продукции ккц печатает не верные данные originally reported on google code with id в колонку норма выхода попадает значение из предыдущей позиции в колонки закладка сырья и закладка специй заполняются с накоплением т е значение второй позиции содержит значение и так далее reported by stasgm on | 0 |
309,976 | 26,690,439,009 | IssuesEvent | 2023-01-27 04:06:23 | datafuselabs/databend | https://api.github.com/repos/datafuselabs/databend | closed | test: add more sqllogic test for COPY file type | C-testing | **Summary**
There is a bug in the:
https://github.com/datafuselabs/databend/blob/0d32a1936b62ef9d212e6b3177180a1da93a6b97/src/query/ast/src/parser/stage.rs#L75
It should be:
```diff
(TYPE ~ "=" ~ (TSV | CSV | NDJSON | PARQUET | JSON | XML) )
```
We need more SQL logic test to check it. | 1.0 | test: add more sqllogic test for COPY file type - **Summary**
There is a bug in the:
https://github.com/datafuselabs/databend/blob/0d32a1936b62ef9d212e6b3177180a1da93a6b97/src/query/ast/src/parser/stage.rs#L75
It should be:
```diff
(TYPE ~ "=" ~ (TSV | CSV | NDJSON | PARQUET | JSON | XML) )
```
We need more SQL logic test to check it. | test | test add more sqllogic test for copy file type summary there is a bug in the it should be diff type tsv csv ndjson parquet json xml we need more sql logic test to check it | 1 |
88,602 | 10,575,756,063 | IssuesEvent | 2019-10-07 16:21:26 | Varrrro/pay-up | https://api.github.com/repos/Varrrro/pay-up | opened | Redactar README con la descripción del proyecto | documentation | Escribir una pequeña descripción del proyecto en el archivo `README.md` del repositorio. | 1.0 | Redactar README con la descripción del proyecto - Escribir una pequeña descripción del proyecto en el archivo `README.md` del repositorio. | non_test | redactar readme con la descripción del proyecto escribir una pequeña descripción del proyecto en el archivo readme md del repositorio | 0 |
6,134 | 13,771,361,048 | IssuesEvent | 2020-10-07 21:52:14 | DarksunTeam/TaleManager | https://api.github.com/repos/DarksunTeam/TaleManager | closed | Inclusão de Campanha | architecture documentation new feature | **Descreva aqui a sua sugestão**
Será necessária a inclusão de um botão no menu lateral para a chamada da tela de Campanha.
Algumas adaptações se farão necessárias também na modelagem.
**Em que parte do sistema esta funcionalidade entraria**
Menu lateral.
**Como você gostaria**
Um botão a mais no inicio da lista.
**Utilidade**
Realizar a futura chamada à tela de campanha, onde o usuário poderá listar todas suas campanhas.
| 1.0 | Inclusão de Campanha - **Descreva aqui a sua sugestão**
Será necessária a inclusão de um botão no menu lateral para a chamada da tela de Campanha.
Algumas adaptações se farão necessárias também na modelagem.
**Em que parte do sistema esta funcionalidade entraria**
Menu lateral.
**Como você gostaria**
Um botão a mais no inicio da lista.
**Utilidade**
Realizar a futura chamada à tela de campanha, onde o usuário poderá listar todas suas campanhas.
| non_test | inclusão de campanha descreva aqui a sua sugestão será necessária a inclusão de um botão no menu lateral para a chamada da tela de campanha algumas adaptações se farão necessárias também na modelagem em que parte do sistema esta funcionalidade entraria menu lateral como você gostaria um botão a mais no inicio da lista utilidade realizar a futura chamada à tela de campanha onde o usuário poderá listar todas suas campanhas | 0 |
577,762 | 17,119,075,019 | IssuesEvent | 2021-07-12 00:14:21 | Techtonica/curriculum | https://api.github.com/repos/Techtonica/curriculum | closed | Fix JS testing using mocks lesson | MEDIUM good-first-issue pinned priority | Fix the problems below, in the lesson: https://github.com/techtonica/curriculum/blob/main/testing-and-tdd/mocking-and-abstraction.md
## Problems found in the tutorial:
1. Materials:
Example video (10 min) -> This link goes to www.google.com. Its not an actual link to video
2. Other example article(20 min read) -> This article link goes to www.google.com. It does not go to an actual link.
3. **Rewrite the following to help user better understand:**
Challenge
Following example above, try to represent the following scenarios and think about what would happen:
* Call getUser('not-octocat')?
* Change mockObject.id to be 42?
* Change mockObject.name to Techtonica?
**Change the explanation to:**
Challenge
Hope you were able to follow the example in the earlier section. Now, take a look at the following scenarios and see if you can make the changes:
* Call getUser('not-octocat')?
* Change mockObject.id to be 42?
* Change mockObject.name to Techtonica?
4. Well, it's tricky because getTodo is still making an external call to the database which is difficult to handle. -> In this section, its hard to understand which database the author is pointing to. Readers may need more detail on the database and why making the external call is difficult and challenging.
5. our reference TODO project) -> **typo**: our reference TODO project
6. It turns out that when we want to make complex verifications around how a mock is called doing that all manually is a lot of work... that somebody else has done for us. -> **This line is confusing. Need more simpler language explanation about the complex verification.**
8. Independent Practice
It's an interesting task to implement your own mocking and validation code by hand and teaches you a lot of neat tricks. If you're feeling adventurous give that a try!
**The above section is not very helpful, and maybe redundant. Instead we can give link to example problems to practice on.**
| 1.0 | Fix JS testing using mocks lesson - Fix the problems below, in the lesson: https://github.com/techtonica/curriculum/blob/main/testing-and-tdd/mocking-and-abstraction.md
## Problems found in the tutorial:
1. Materials:
Example video (10 min) -> This link goes to www.google.com. Its not an actual link to video
2. Other example article(20 min read) -> This article link goes to www.google.com. It does not go to an actual link.
3. **Rewrite the following to help user better understand:**
Challenge
Following example above, try to represent the following scenarios and think about what would happen:
* Call getUser('not-octocat')?
* Change mockObject.id to be 42?
* Change mockObject.name to Techtonica?
**Change the explanation to:**
Challenge
Hope you were able to follow the example in the earlier section. Now, take a look at the following scenarios and see if you can make the changes:
* Call getUser('not-octocat')?
* Change mockObject.id to be 42?
* Change mockObject.name to Techtonica?
4. Well, it's tricky because getTodo is still making an external call to the database which is difficult to handle. -> In this section, its hard to understand which database the author is pointing to. Readers may need more detail on the database and why making the external call is difficult and challenging.
5. our reference TODO project) -> **typo**: our reference TODO project
6. It turns out that when we want to make complex verifications around how a mock is called doing that all manually is a lot of work... that somebody else has done for us. -> **This line is confusing. Need more simpler language explanation about the complex verification.**
8. Independent Practice
It's an interesting task to implement your own mocking and validation code by hand and teaches you a lot of neat tricks. If you're feeling adventurous give that a try!
**The above section is not very helpful, and maybe redundant. Instead we can give link to example problems to practice on.**
| non_test | fix js testing using mocks lesson fix the problems below in the lesson problems found in the tutorial materials example video min this link goes to its not an actual link to video other example article min read this article link goes to it does not go to an actual link rewrite the following to help user better understand challenge following example above try to represent the following scenarios and think about what would happen call getuser not octocat change mockobject id to be change mockobject name to techtonica change the explanation to challenge hope you were able to follow the example in the earlier section now take a look at the following scenarios and see if you can make the changes call getuser not octocat change mockobject id to be change mockobject name to techtonica well it s tricky because gettodo is still making an external call to the database which is difficult to handle in this section its hard to understand which database the author is pointing to readers may need more detail on the database and why making the external call is difficult and challenging our reference todo project typo our reference todo project it turns out that when we want to make complex verifications around how a mock is called doing that all manually is a lot of work that somebody else has done for us this line is confusing need more simpler language explanation about the complex verification independent practice it s an interesting task to implement your own mocking and validation code by hand and teaches you a lot of neat tricks if you re feeling adventurous give that a try the above section is not very helpful and maybe redundant instead we can give link to example problems to practice on | 0 |
342,083 | 24,728,240,586 | IssuesEvent | 2022-10-20 15:28:42 | nebari-dev/nebari-docs | https://api.github.com/repos/nebari-dev/nebari-docs | closed | [DOC] - Quickstart/Commands page for advanced users | type: enhancement 💅🏼 area: documentation 📖 impact: high | ### Preliminary Checks
- [X] This issue is not a question, feature request, RFC, or anything other than a bug report. Please post those things in GitHub Discussions: https://github.com/nebari-dev/nebari/discussions
### Summary
This page is intended to be a cheat sheet with all the important commands a regualr/savvy user will need to quickly get Nebari up and running.
Inspiration:
- https://cli.github.com/manual/index
- https://spacy.io/usage#quickstart
- https://www.gatsbyjs.com/docs/quick-start/ (best)
This page can live in the Getting started section.
### Steps to Resolve this Issue
n/a | 1.0 | [DOC] - Quickstart/Commands page for advanced users - ### Preliminary Checks
- [X] This issue is not a question, feature request, RFC, or anything other than a bug report. Please post those things in GitHub Discussions: https://github.com/nebari-dev/nebari/discussions
### Summary
This page is intended to be a cheat sheet with all the important commands a regualr/savvy user will need to quickly get Nebari up and running.
Inspiration:
- https://cli.github.com/manual/index
- https://spacy.io/usage#quickstart
- https://www.gatsbyjs.com/docs/quick-start/ (best)
This page can live in the Getting started section.
### Steps to Resolve this Issue
n/a | non_test | quickstart commands page for advanced users preliminary checks this issue is not a question feature request rfc or anything other than a bug report please post those things in github discussions summary this page is intended to be a cheat sheet with all the important commands a regualr savvy user will need to quickly get nebari up and running inspiration best this page can live in the getting started section steps to resolve this issue n a | 0 |
444,332 | 12,810,151,551 | IssuesEvent | 2020-07-03 17:40:34 | zeebe-io/zeebe | https://api.github.com/repos/zeebe-io/zeebe | closed | java client: jobs are getting activated but not coming back to client (intermittent issue) | Impact: Performance Priority: Mid Scope: broker Severity: Mid Status: Ready Type: Bug | **Describe the bug**
I'm noticing that sometime jobs are getting activated but not handled through jobHandler and then after timeout, zeebe again retries the job and then it properly gets processed.
On ES exporter I'm seeing the job intent being "ACTIVATED", then after time out period, its "TIMED_OUT" and another record as "ACTIVATED", then properly handled in jobHandler and then "COMPLETED".
As the jobs are getting activated but timing out, I'm guessing its client issue rather than broker. Client has activated the job and started working on it but has not sent any response within proper time, so broker times out the job and the job is ready for polling again.
We are using java client , maven version 0.22.0-alpha1
As the intent is "ACTIVATED", I'm guessing jobPoller activates this job and it's in its jobList. The next thing it should do is to iterate through the list and submit jobs which eventually calls jobHandler's handle method.
> public void onNext(ActivateJobsResponse activateJobsResponse) {
this.activatedJobs += activateJobsResponse.getJobsCount();
activateJobsResponse.getJobsList().stream().map((job) -> {
return new ActivatedJobImpl(this.objectMapper, job);
}).forEach(this.jobConsumer);
}
I've logger info as the first line in handler method. But this line doesn't get printed in this particular case, neither the warning gets printed which is inside the catch block - this means the handle method dint even gets called.
> private void executeJob(ActivatedJob job, Runnable doneCallback) {
try {
this.handler.handle(this.jobClient, job);
} catch (Exception var10) {
LOG.warn("Worker {} failed to handle job with key {} of type {}, sending fail command to broker", new Object[]{job.getWorker(), job.getKey(), job.getType(), var10});
StringWriter stringWriter = new StringWriter();
PrintWriter printWriter = new PrintWriter(stringWriter);
var10.printStackTrace(printWriter);
String message = stringWriter.toString();
this.jobClient.newFailCommand(job.getKey()).retries(job.getRetries() - 1).errorMessage(message).send();
} finally {
doneCallback.run();
}
}
It seems to me that the thread is getting hanged somehow but not sure how to troubleshoot this particular issue.
**To Reproduce**
This is intermittent. 20% jobs fail on first attempt, but those again pass in second attempt ( ie, after retry)
**Expected behavior**
Jobs should be either fails or passes on first attempt.
| 1.0 | java client: jobs are getting activated but not coming back to client (intermittent issue) - **Describe the bug**
I'm noticing that sometime jobs are getting activated but not handled through jobHandler and then after timeout, zeebe again retries the job and then it properly gets processed.
On ES exporter I'm seeing the job intent being "ACTIVATED", then after time out period, its "TIMED_OUT" and another record as "ACTIVATED", then properly handled in jobHandler and then "COMPLETED".
As the jobs are getting activated but timing out, I'm guessing its client issue rather than broker. Client has activated the job and started working on it but has not sent any response within proper time, so broker times out the job and the job is ready for polling again.
We are using java client , maven version 0.22.0-alpha1
As the intent is "ACTIVATED", I'm guessing jobPoller activates this job and it's in its jobList. The next thing it should do is to iterate through the list and submit jobs which eventually calls jobHandler's handle method.
> public void onNext(ActivateJobsResponse activateJobsResponse) {
this.activatedJobs += activateJobsResponse.getJobsCount();
activateJobsResponse.getJobsList().stream().map((job) -> {
return new ActivatedJobImpl(this.objectMapper, job);
}).forEach(this.jobConsumer);
}
I've logger info as the first line in handler method. But this line doesn't get printed in this particular case, neither the warning gets printed which is inside the catch block - this means the handle method dint even gets called.
> private void executeJob(ActivatedJob job, Runnable doneCallback) {
try {
this.handler.handle(this.jobClient, job);
} catch (Exception var10) {
LOG.warn("Worker {} failed to handle job with key {} of type {}, sending fail command to broker", new Object[]{job.getWorker(), job.getKey(), job.getType(), var10});
StringWriter stringWriter = new StringWriter();
PrintWriter printWriter = new PrintWriter(stringWriter);
var10.printStackTrace(printWriter);
String message = stringWriter.toString();
this.jobClient.newFailCommand(job.getKey()).retries(job.getRetries() - 1).errorMessage(message).send();
} finally {
doneCallback.run();
}
}
It seems to me that the thread is getting hanged somehow but not sure how to troubleshoot this particular issue.
**To Reproduce**
This is intermittent. 20% jobs fail on first attempt, but those again pass in second attempt ( ie, after retry)
**Expected behavior**
Jobs should be either fails or passes on first attempt.
| non_test | java client jobs are getting activated but not coming back to client intermittent issue describe the bug i m noticing that sometime jobs are getting activated but not handled through jobhandler and then after timeout zeebe again retries the job and then it properly gets processed on es exporter i m seeing the job intent being activated then after time out period its timed out and another record as activated then properly handled in jobhandler and then completed as the jobs are getting activated but timing out i m guessing its client issue rather than broker client has activated the job and started working on it but has not sent any response within proper time so broker times out the job and the job is ready for polling again we are using java client maven version as the intent is activated i m guessing jobpoller activates this job and it s in its joblist the next thing it should do is to iterate through the list and submit jobs which eventually calls jobhandler s handle method public void onnext activatejobsresponse activatejobsresponse this activatedjobs activatejobsresponse getjobscount activatejobsresponse getjobslist stream map job return new activatedjobimpl this objectmapper job foreach this jobconsumer i ve logger info as the first line in handler method but this line doesn t get printed in this particular case neither the warning gets printed which is inside the catch block this means the handle method dint even gets called private void executejob activatedjob job runnable donecallback try this handler handle this jobclient job catch exception log warn worker failed to handle job with key of type sending fail command to broker new object job getworker job getkey job gettype stringwriter stringwriter new stringwriter printwriter printwriter new printwriter stringwriter printstacktrace printwriter string message stringwriter tostring this jobclient newfailcommand job getkey retries job getretries errormessage message send finally donecallback run it seems to me that the thread is getting hanged somehow but not sure how to troubleshoot this particular issue to reproduce this is intermittent jobs fail on first attempt but those again pass in second attempt ie after retry expected behavior jobs should be either fails or passes on first attempt | 0 |
49,736 | 6,038,594,377 | IssuesEvent | 2017-06-09 21:55:43 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | GCE 1.5->1.6 upgrade testing: CronJob - the server could not find the expected resource | kind/flake kind/upgrade-test-failure priority/failing-test sig/apps | @soltysh, please take a look ASAP and let me know if this is a 1.6 blocker or not.
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.5-1.6-upgrade-cluster/19#k8sio-cronjob-should-not-emit-unexpected-warnings
On GCE, in the 1.5->1.6 upgrade-cluster job (which creates a 1.5 cluster, upgrades to 1.6, then runs 1.5 E2E tests), the CronJob tests are failing with
```
Expected error:
<*errors.StatusError | 0xc422657180>: {
ErrStatus: {
TypeMeta: {Kind: "", APIVersion: ""},
ListMeta: {SelfLink: "", ResourceVersion: ""},
Status: "Failure",
Message: "the server could not find the requested resource",
Reason: "NotFound",
Details: {Name: "", Group: "", Kind: "", Causes: nil, RetryAfterSeconds: 0},
Code: 404,
},
}
the server could not find the requested resource
not to have occurred
```
| 2.0 | GCE 1.5->1.6 upgrade testing: CronJob - the server could not find the expected resource - @soltysh, please take a look ASAP and let me know if this is a 1.6 blocker or not.
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.5-1.6-upgrade-cluster/19#k8sio-cronjob-should-not-emit-unexpected-warnings
On GCE, in the 1.5->1.6 upgrade-cluster job (which creates a 1.5 cluster, upgrades to 1.6, then runs 1.5 E2E tests), the CronJob tests are failing with
```
Expected error:
<*errors.StatusError | 0xc422657180>: {
ErrStatus: {
TypeMeta: {Kind: "", APIVersion: ""},
ListMeta: {SelfLink: "", ResourceVersion: ""},
Status: "Failure",
Message: "the server could not find the requested resource",
Reason: "NotFound",
Details: {Name: "", Group: "", Kind: "", Causes: nil, RetryAfterSeconds: 0},
Code: 404,
},
}
the server could not find the requested resource
not to have occurred
```
| test | gce upgrade testing cronjob the server could not find the expected resource soltysh please take a look asap and let me know if this is a blocker or not on gce in the upgrade cluster job which creates a cluster upgrades to then runs tests the cronjob tests are failing with expected error errstatus typemeta kind apiversion listmeta selflink resourceversion status failure message the server could not find the requested resource reason notfound details name group kind causes nil retryafterseconds code the server could not find the requested resource not to have occurred | 1 |
287,652 | 8,817,975,661 | IssuesEvent | 2018-12-31 07:36:40 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | hangouts.google.com - site is not usable | browser-firefox priority-critical | <!-- @browser: Firefox 64.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.2; Win64; x64; rv:64.0) Gecko/20100101 Firefox/64.0 -->
<!-- @reported_with: -->
**URL**: https://hangouts.google.com
**Browser / Version**: Firefox 64.0
**Operating System**: Windows 8
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: hangouts chats list ist empty. Seems due to a CSP restriction
**Steps to Reproduce**:
just open the page with a google account with some chats.
[](https://webcompat.com/uploads/2018/12/c72c5af1-5ca7-4163-9222-a47faa2de671.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
Reported by @arneschween
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | hangouts.google.com - site is not usable - <!-- @browser: Firefox 64.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.2; Win64; x64; rv:64.0) Gecko/20100101 Firefox/64.0 -->
<!-- @reported_with: -->
**URL**: https://hangouts.google.com
**Browser / Version**: Firefox 64.0
**Operating System**: Windows 8
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: hangouts chats list ist empty. Seems due to a CSP restriction
**Steps to Reproduce**:
just open the page with a google account with some chats.
[](https://webcompat.com/uploads/2018/12/c72c5af1-5ca7-4163-9222-a47faa2de671.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
Reported by @arneschween
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_test | hangouts google com site is not usable url browser version firefox operating system windows tested another browser yes problem type site is not usable description hangouts chats list ist empty seems due to a csp restriction steps to reproduce just open the page with a google account with some chats browser configuration none reported by arneschween from with ❤️ | 0 |
12,313 | 3,265,785,440 | IssuesEvent | 2015-10-22 17:46:39 | mautic/mautic | https://api.github.com/repos/mautic/mautic | closed | REST API for GET Campaign erroring out | Bug Ready To Test | Hello everyone,
I am trying to perform GET campaign call but getting following error. Anyone has any idea about this issue?
ERROR:
{
"error": {
"message": "Property Mautic\CampaignBundle\Entity\Campaign::$title does not exist",
"code": 0
}
}
REST CALL:
http://192.168.2.185/1.2.0/index.php/api/campaigns/2?access_token=NDYwZZcxZmNjOThmNsEzMjI4ZWRjYjlmMzE5NGI2MTBkMzg4MjAyOTNjNTcyYTllOGRkNjc3NjViNTEyLzdjMg
Error message in log file:
[2015-09-23 11:35:38] mautic.CRITICAL: Uncaught PHP Exception ReflectionException: "Property Mautic\CampaignBundle\Entity\Campaign::$title does not exist" at /var/www/html/1.2.0/vendor/jms/metadata/src/Metadata/PropertyMetadata.php line 40 {"exception":"[object] (ReflectionException(code: 0): Property Mautic\\CampaignBundle\\Entity\\Campaign::$title does not exist at /var/www/html/1.2.0/vendor/jms/metadata/src/Metadata/PropertyMetadata.php:40)"} []
Note: other rest calls like create lead, get lead etc are working just fine. Also GET Campaign is working fine with same request in Mautic 1.1 but not working in Mautic 1.2.
Thanks,
Nimesh
| 1.0 | REST API for GET Campaign erroring out - Hello everyone,
I am trying to perform GET campaign call but getting following error. Anyone has any idea about this issue?
ERROR:
{
"error": {
"message": "Property Mautic\CampaignBundle\Entity\Campaign::$title does not exist",
"code": 0
}
}
REST CALL:
http://192.168.2.185/1.2.0/index.php/api/campaigns/2?access_token=NDYwZZcxZmNjOThmNsEzMjI4ZWRjYjlmMzE5NGI2MTBkMzg4MjAyOTNjNTcyYTllOGRkNjc3NjViNTEyLzdjMg
Error message in log file:
[2015-09-23 11:35:38] mautic.CRITICAL: Uncaught PHP Exception ReflectionException: "Property Mautic\CampaignBundle\Entity\Campaign::$title does not exist" at /var/www/html/1.2.0/vendor/jms/metadata/src/Metadata/PropertyMetadata.php line 40 {"exception":"[object] (ReflectionException(code: 0): Property Mautic\\CampaignBundle\\Entity\\Campaign::$title does not exist at /var/www/html/1.2.0/vendor/jms/metadata/src/Metadata/PropertyMetadata.php:40)"} []
Note: other rest calls like create lead, get lead etc are working just fine. Also GET Campaign is working fine with same request in Mautic 1.1 but not working in Mautic 1.2.
Thanks,
Nimesh
| test | rest api for get campaign erroring out hello everyone i am trying to perform get campaign call but getting following error anyone has any idea about this issue error error message property mautic campaignbundle entity campaign title does not exist code rest call error message in log file mautic critical uncaught php exception reflectionexception property mautic campaignbundle entity campaign title does not exist at var www html vendor jms metadata src metadata propertymetadata php line exception reflectionexception code property mautic campaignbundle entity campaign title does not exist at var www html vendor jms metadata src metadata propertymetadata php note other rest calls like create lead get lead etc are working just fine also get campaign is working fine with same request in mautic but not working in mautic thanks nimesh | 1 |
152,883 | 13,486,580,308 | IssuesEvent | 2020-09-11 09:39:35 | abpframework/abp | https://api.github.com/repos/abpframework/abp | closed | Documentation how to define custom filters | documentation | https://docs.abp.io/en/abp/latest/Data-Filtering
Current document is not complete. For example, need to override `ShouldFilterEntity`. Could we provide more details? | 1.0 | Documentation how to define custom filters - https://docs.abp.io/en/abp/latest/Data-Filtering
Current document is not complete. For example, need to override `ShouldFilterEntity`. Could we provide more details? | non_test | documentation how to define custom filters current document is not complete for example need to override shouldfilterentity could we provide more details | 0 |
203,125 | 15,350,202,248 | IssuesEvent | 2021-03-01 01:43:14 | Azure/azure-sdk-for-net | https://api.github.com/repos/Azure/azure-sdk-for-net | closed | Azure.Communication.Chat Sample issue | Client Communication Docs needs-team-triage test-manual-pass | 1.
Section [link1,](https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/communication/Azure.Communication.Chat/tests/samples/Sample2_MessagingOperations.cs#L33)[link2:](https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/communication/Azure.Communication.Chat/tests/samples/Sample3_MemberOperations.cs#L38)

Suggestion:

Add: `DateTime minDate = DateTime.SpecifyKind(DateTime.MinValue, DateTimeKind.Utc);`
And update `DateTime.MinValue` to `minDate `
@jongio for notification.
| 1.0 | Azure.Communication.Chat Sample issue - 1.
Section [link1,](https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/communication/Azure.Communication.Chat/tests/samples/Sample2_MessagingOperations.cs#L33)[link2:](https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/communication/Azure.Communication.Chat/tests/samples/Sample3_MemberOperations.cs#L38)

Suggestion:

Add: `DateTime minDate = DateTime.SpecifyKind(DateTime.MinValue, DateTimeKind.Utc);`
And update `DateTime.MinValue` to `minDate `
@jongio for notification.
| test | azure communication chat sample issue section suggestion add datetime mindate datetime specifykind datetime minvalue datetimekind utc and update datetime minvalue to mindate jongio for notification | 1 |
152,149 | 12,093,170,369 | IssuesEvent | 2020-04-19 18:31:35 | iqlusioninc/relayer | https://api.github.com/repos/iqlusioninc/relayer | closed | Integration tests | help wanted testing | The current plan is to test the `relayer` package using something like https://github.com/ory/dockertest and a gaia image that allows for passing in addresses to be funded for genesis. Will work on getting this setup. | 1.0 | Integration tests - The current plan is to test the `relayer` package using something like https://github.com/ory/dockertest and a gaia image that allows for passing in addresses to be funded for genesis. Will work on getting this setup. | test | integration tests the current plan is to test the relayer package using something like and a gaia image that allows for passing in addresses to be funded for genesis will work on getting this setup | 1 |
11,290 | 3,197,635,201 | IssuesEvent | 2015-10-01 06:58:24 | uProxy/uproxy | https://api.github.com/repos/uProxy/uproxy | closed | Move freedom mocking out of remote-connection.spec.ts into more obvious file | C:Testing | Right now a number of the tests for uProxy core all use the global storage object, which in unit tests is mocked to use our freedom_mocks.MockStorage class. However the glue to sets freedom['storage'] to freedom_mocks.MockStorage is in remote-connection.spec.ts, despite it being depended on by all our other unit tests in the core. This is not an obvious place for it, rather we should move it to some new file that is included in all our tests. We have a similar situation for MockLoggingController and now MockMetrics | 1.0 | Move freedom mocking out of remote-connection.spec.ts into more obvious file - Right now a number of the tests for uProxy core all use the global storage object, which in unit tests is mocked to use our freedom_mocks.MockStorage class. However the glue to sets freedom['storage'] to freedom_mocks.MockStorage is in remote-connection.spec.ts, despite it being depended on by all our other unit tests in the core. This is not an obvious place for it, rather we should move it to some new file that is included in all our tests. We have a similar situation for MockLoggingController and now MockMetrics | test | move freedom mocking out of remote connection spec ts into more obvious file right now a number of the tests for uproxy core all use the global storage object which in unit tests is mocked to use our freedom mocks mockstorage class however the glue to sets freedom to freedom mocks mockstorage is in remote connection spec ts despite it being depended on by all our other unit tests in the core this is not an obvious place for it rather we should move it to some new file that is included in all our tests we have a similar situation for mockloggingcontroller and now mockmetrics | 1 |
74,037 | 3,427,465,804 | IssuesEvent | 2015-12-10 01:50:55 | OctopusDeploy/Issues | https://api.github.com/repos/OctopusDeploy/Issues | closed | OctopusDeleteScriptsOnCleanup not working in 3.x | bug in progress priority | A customer reported that in 3.x OctopusDeleteScriptsOnCleanup being set to false no longer leaves the ps scripts behind. The variable still exists in the code, so the execution from calamari must be missing. (I have to assume if it was left out on purpose the variable would have been removed).
Source: http://help.octopusdeploy.com/discussions/problems/43117 | 1.0 | OctopusDeleteScriptsOnCleanup not working in 3.x - A customer reported that in 3.x OctopusDeleteScriptsOnCleanup being set to false no longer leaves the ps scripts behind. The variable still exists in the code, so the execution from calamari must be missing. (I have to assume if it was left out on purpose the variable would have been removed).
Source: http://help.octopusdeploy.com/discussions/problems/43117 | non_test | octopusdeletescriptsoncleanup not working in x a customer reported that in x octopusdeletescriptsoncleanup being set to false no longer leaves the ps scripts behind the variable still exists in the code so the execution from calamari must be missing i have to assume if it was left out on purpose the variable would have been removed source | 0 |
13,102 | 3,310,174,308 | IssuesEvent | 2015-11-05 07:13:39 | medic/medic-webapp | https://api.github.com/repos/medic/medic-webapp | closed | Hide head of family from family profile card | 4 - Acceptance testing Feature Request | Continuation of #1422, and similar to #1402, we should hide the "head of family" from the family profile card.
The exception to this is if the "head of family" is **not** in the family list - which is possible if people are reassigned manually. This exception is not likely for the field-test, but theoretically possible. | 1.0 | Hide head of family from family profile card - Continuation of #1422, and similar to #1402, we should hide the "head of family" from the family profile card.
The exception to this is if the "head of family" is **not** in the family list - which is possible if people are reassigned manually. This exception is not likely for the field-test, but theoretically possible. | test | hide head of family from family profile card continuation of and similar to we should hide the head of family from the family profile card the exception to this is if the head of family is not in the family list which is possible if people are reassigned manually this exception is not likely for the field test but theoretically possible | 1 |
4,677 | 7,517,304,707 | IssuesEvent | 2018-04-12 02:48:22 | UnbFeelings/unb-feelings-GQA | https://api.github.com/repos/UnbFeelings/unb-feelings-GQA | closed | Criar templates de documentação | document process wiki | Deve-se identificar quais documentações necessitam de templates para que possam ser criados, como: Resultados das Auditoria, Documentação do Checklist dentro do resultado da Auditoria, Documentação da Entrevista dentro do resultado da Auditoria, entre outros.
[Atividade no processo](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Fluxo-de-Trabalho#17-criar-templates-de-documenta%C3%A7%C3%A3o). | 1.0 | Criar templates de documentação - Deve-se identificar quais documentações necessitam de templates para que possam ser criados, como: Resultados das Auditoria, Documentação do Checklist dentro do resultado da Auditoria, Documentação da Entrevista dentro do resultado da Auditoria, entre outros.
[Atividade no processo](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Fluxo-de-Trabalho#17-criar-templates-de-documenta%C3%A7%C3%A3o). | non_test | criar templates de documentação deve se identificar quais documentações necessitam de templates para que possam ser criados como resultados das auditoria documentação do checklist dentro do resultado da auditoria documentação da entrevista dentro do resultado da auditoria entre outros | 0 |
36,242 | 12,404,341,726 | IssuesEvent | 2020-05-21 15:22:43 | jgeraigery/beaker-notebook | https://api.github.com/repos/jgeraigery/beaker-notebook | opened | CVE-2017-7656 (High) detected in multiple libraries | security vulnerability | ## CVE-2017-7656 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jetty-http-8.1.12.v20130726.jar</b>, <b>jetty-http-8.1.13.v20130916.jar</b>, <b>jetty-server-8.1.13.v20130916.jar</b>, <b>jetty-server-8.1.12.v20130726.jar</b></p></summary>
<p>
<details><summary><b>jetty-http-8.1.12.v20130726.jar</b></p></summary>
<p>Administrative parent pom for Jetty modules</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: /tmp/ws-scm/beaker-notebook/shared/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-http/8.1.12.v20130726/530b3a21d71ac69279bee129869d7eac031e3533/jetty-http-8.1.12.v20130726.jar,/root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-http/8.1.12.v20130726/530b3a21d71ac69279bee129869d7eac031e3533/jetty-http-8.1.12.v20130726.jar</p>
<p>
Dependency Hierarchy:
- cxf-bundle-jaxrs-2.7.7.jar (Root Library)
- :x: **jetty-http-8.1.12.v20130726.jar** (Vulnerable Library)
</details>
<details><summary><b>jetty-http-8.1.13.v20130916.jar</b></p></summary>
<p>Administrative parent pom for Jetty modules</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: /tmp/ws-scm/beaker-notebook/plugin/clojure/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-http/8.1.13.v20130916/6dcf37666815f6d0d90b77a2f5037a9ceaaca968/jetty-http-8.1.13.v20130916.jar,/root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-http/8.1.13.v20130916/6dcf37666815f6d0d90b77a2f5037a9ceaaca968/jetty-http-8.1.13.v20130916.jar</p>
<p>
Dependency Hierarchy:
- cxf-bundle-jaxrs-2.7.7.jar (Root Library)
- :x: **jetty-http-8.1.13.v20130916.jar** (Vulnerable Library)
</details>
<details><summary><b>jetty-server-8.1.13.v20130916.jar</b></p></summary>
<p>The core jetty server artifact.</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-server/8.1.13.v20130916/99d1bf6fb172cecb597f1c029c719c5f878d8405/jetty-server-8.1.13.v20130916.jar,/root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-server/8.1.13.v20130916/99d1bf6fb172cecb597f1c029c719c5f878d8405/jetty-server-8.1.13.v20130916.jar</p>
<p>
Dependency Hierarchy:
- cxf-bundle-jaxrs-2.7.7.jar (Root Library)
- :x: **jetty-server-8.1.13.v20130916.jar** (Vulnerable Library)
</details>
<details><summary><b>jetty-server-8.1.12.v20130726.jar</b></p></summary>
<p>The core jetty server artifact.</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-server/8.1.12.v20130726/e8d89c85edd00680a7b30bf219e6dba181dc4aa1/jetty-server-8.1.12.v20130726.jar,/root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-server/8.1.12.v20130726/e8d89c85edd00680a7b30bf219e6dba181dc4aa1/jetty-server-8.1.12.v20130726.jar</p>
<p>
Dependency Hierarchy:
- cxf-bundle-jaxrs-2.7.7.jar (Root Library)
- :x: **jetty-server-8.1.12.v20130726.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/beaker-notebook/commit/e74341acf643e87bd21b092c7a9e9f6bb96fa7c4">e74341acf643e87bd21b092c7a9e9f6bb96fa7c4</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Eclipse Jetty, versions 9.2.x and older, 9.3.x (all configurations), and 9.4.x (non-default configuration with RFC2616 compliance enabled), HTTP/0.9 is handled poorly. An HTTP/1 style request line (i.e. method space URI space version) that declares a version of HTTP/0.9 was accepted and treated as a 0.9 request. If deployed behind an intermediary that also accepted and passed through the 0.9 version (but did not act on it), then the response sent could be interpreted by the intermediary as HTTP/1 headers. This could be used to poison the cache if the server allowed the origin client to generate arbitrary content in the response.
<p>Publish Date: 2018-06-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-7656>CVE-2017-7656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://www.securitytracker.com/id/1041194">http://www.securitytracker.com/id/1041194</a></p>
<p>Fix Resolution: The vendor has issued a fix (9.4.11.v20180605).
9.2.25.v20180606, 9.3.24.v20180605
The vendor advisory is available at:
http://dev.eclipse.org/mhonarc/lists/jetty-announce/msg00123.html</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-http","packageVersion":"8.1.12.v20130726","isTransitiveDependency":true,"dependencyTree":"org.apache.cxf:cxf-bundle-jaxrs:2.7.7;org.eclipse.jetty:jetty-http:8.1.12.v20130726","isMinimumFixVersionAvailable":false},{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-http","packageVersion":"8.1.13.v20130916","isTransitiveDependency":true,"dependencyTree":"org.apache.cxf:cxf-bundle-jaxrs:2.7.7;org.eclipse.jetty:jetty-http:8.1.13.v20130916","isMinimumFixVersionAvailable":false},{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-server","packageVersion":"8.1.13.v20130916","isTransitiveDependency":true,"dependencyTree":"org.apache.cxf:cxf-bundle-jaxrs:2.7.7;org.eclipse.jetty:jetty-server:8.1.13.v20130916","isMinimumFixVersionAvailable":false},{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-server","packageVersion":"8.1.12.v20130726","isTransitiveDependency":true,"dependencyTree":"org.apache.cxf:cxf-bundle-jaxrs:2.7.7;org.eclipse.jetty:jetty-server:8.1.12.v20130726","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2017-7656","vulnerabilityDetails":"In Eclipse Jetty, versions 9.2.x and older, 9.3.x (all configurations), and 9.4.x (non-default configuration with RFC2616 compliance enabled), HTTP/0.9 is handled poorly. An HTTP/1 style request line (i.e. method space URI space version) that declares a version of HTTP/0.9 was accepted and treated as a 0.9 request. If deployed behind an intermediary that also accepted and passed through the 0.9 version (but did not act on it), then the response sent could be interpreted by the intermediary as HTTP/1 headers. This could be used to poison the cache if the server allowed the origin client to generate arbitrary content in the response.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-7656","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2017-7656 (High) detected in multiple libraries - ## CVE-2017-7656 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jetty-http-8.1.12.v20130726.jar</b>, <b>jetty-http-8.1.13.v20130916.jar</b>, <b>jetty-server-8.1.13.v20130916.jar</b>, <b>jetty-server-8.1.12.v20130726.jar</b></p></summary>
<p>
<details><summary><b>jetty-http-8.1.12.v20130726.jar</b></p></summary>
<p>Administrative parent pom for Jetty modules</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: /tmp/ws-scm/beaker-notebook/shared/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-http/8.1.12.v20130726/530b3a21d71ac69279bee129869d7eac031e3533/jetty-http-8.1.12.v20130726.jar,/root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-http/8.1.12.v20130726/530b3a21d71ac69279bee129869d7eac031e3533/jetty-http-8.1.12.v20130726.jar</p>
<p>
Dependency Hierarchy:
- cxf-bundle-jaxrs-2.7.7.jar (Root Library)
- :x: **jetty-http-8.1.12.v20130726.jar** (Vulnerable Library)
</details>
<details><summary><b>jetty-http-8.1.13.v20130916.jar</b></p></summary>
<p>Administrative parent pom for Jetty modules</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: /tmp/ws-scm/beaker-notebook/plugin/clojure/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-http/8.1.13.v20130916/6dcf37666815f6d0d90b77a2f5037a9ceaaca968/jetty-http-8.1.13.v20130916.jar,/root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-http/8.1.13.v20130916/6dcf37666815f6d0d90b77a2f5037a9ceaaca968/jetty-http-8.1.13.v20130916.jar</p>
<p>
Dependency Hierarchy:
- cxf-bundle-jaxrs-2.7.7.jar (Root Library)
- :x: **jetty-http-8.1.13.v20130916.jar** (Vulnerable Library)
</details>
<details><summary><b>jetty-server-8.1.13.v20130916.jar</b></p></summary>
<p>The core jetty server artifact.</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-server/8.1.13.v20130916/99d1bf6fb172cecb597f1c029c719c5f878d8405/jetty-server-8.1.13.v20130916.jar,/root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-server/8.1.13.v20130916/99d1bf6fb172cecb597f1c029c719c5f878d8405/jetty-server-8.1.13.v20130916.jar</p>
<p>
Dependency Hierarchy:
- cxf-bundle-jaxrs-2.7.7.jar (Root Library)
- :x: **jetty-server-8.1.13.v20130916.jar** (Vulnerable Library)
</details>
<details><summary><b>jetty-server-8.1.12.v20130726.jar</b></p></summary>
<p>The core jetty server artifact.</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-server/8.1.12.v20130726/e8d89c85edd00680a7b30bf219e6dba181dc4aa1/jetty-server-8.1.12.v20130726.jar,/root/.gradle/caches/modules-2/files-2.1/org.eclipse.jetty/jetty-server/8.1.12.v20130726/e8d89c85edd00680a7b30bf219e6dba181dc4aa1/jetty-server-8.1.12.v20130726.jar</p>
<p>
Dependency Hierarchy:
- cxf-bundle-jaxrs-2.7.7.jar (Root Library)
- :x: **jetty-server-8.1.12.v20130726.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/beaker-notebook/commit/e74341acf643e87bd21b092c7a9e9f6bb96fa7c4">e74341acf643e87bd21b092c7a9e9f6bb96fa7c4</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Eclipse Jetty, versions 9.2.x and older, 9.3.x (all configurations), and 9.4.x (non-default configuration with RFC2616 compliance enabled), HTTP/0.9 is handled poorly. An HTTP/1 style request line (i.e. method space URI space version) that declares a version of HTTP/0.9 was accepted and treated as a 0.9 request. If deployed behind an intermediary that also accepted and passed through the 0.9 version (but did not act on it), then the response sent could be interpreted by the intermediary as HTTP/1 headers. This could be used to poison the cache if the server allowed the origin client to generate arbitrary content in the response.
<p>Publish Date: 2018-06-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-7656>CVE-2017-7656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://www.securitytracker.com/id/1041194">http://www.securitytracker.com/id/1041194</a></p>
<p>Fix Resolution: The vendor has issued a fix (9.4.11.v20180605).
9.2.25.v20180606, 9.3.24.v20180605
The vendor advisory is available at:
http://dev.eclipse.org/mhonarc/lists/jetty-announce/msg00123.html</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-http","packageVersion":"8.1.12.v20130726","isTransitiveDependency":true,"dependencyTree":"org.apache.cxf:cxf-bundle-jaxrs:2.7.7;org.eclipse.jetty:jetty-http:8.1.12.v20130726","isMinimumFixVersionAvailable":false},{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-http","packageVersion":"8.1.13.v20130916","isTransitiveDependency":true,"dependencyTree":"org.apache.cxf:cxf-bundle-jaxrs:2.7.7;org.eclipse.jetty:jetty-http:8.1.13.v20130916","isMinimumFixVersionAvailable":false},{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-server","packageVersion":"8.1.13.v20130916","isTransitiveDependency":true,"dependencyTree":"org.apache.cxf:cxf-bundle-jaxrs:2.7.7;org.eclipse.jetty:jetty-server:8.1.13.v20130916","isMinimumFixVersionAvailable":false},{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-server","packageVersion":"8.1.12.v20130726","isTransitiveDependency":true,"dependencyTree":"org.apache.cxf:cxf-bundle-jaxrs:2.7.7;org.eclipse.jetty:jetty-server:8.1.12.v20130726","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2017-7656","vulnerabilityDetails":"In Eclipse Jetty, versions 9.2.x and older, 9.3.x (all configurations), and 9.4.x (non-default configuration with RFC2616 compliance enabled), HTTP/0.9 is handled poorly. An HTTP/1 style request line (i.e. method space URI space version) that declares a version of HTTP/0.9 was accepted and treated as a 0.9 request. If deployed behind an intermediary that also accepted and passed through the 0.9 version (but did not act on it), then the response sent could be interpreted by the intermediary as HTTP/1 headers. This could be used to poison the cache if the server allowed the origin client to generate arbitrary content in the response.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-7656","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_test | cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries jetty http jar jetty http jar jetty server jar jetty server jar jetty http jar administrative parent pom for jetty modules library home page a href path to dependency file tmp ws scm beaker notebook shared build gradle path to vulnerable library root gradle caches modules files org eclipse jetty jetty http jetty http jar root gradle caches modules files org eclipse jetty jetty http jetty http jar dependency hierarchy cxf bundle jaxrs jar root library x jetty http jar vulnerable library jetty http jar administrative parent pom for jetty modules library home page a href path to dependency file tmp ws scm beaker notebook plugin clojure build gradle path to vulnerable library root gradle caches modules files org eclipse jetty jetty http jetty http jar root gradle caches modules files org eclipse jetty jetty http jetty http jar dependency hierarchy cxf bundle jaxrs jar root library x jetty http jar vulnerable library jetty server jar the core jetty server artifact library home page a href path to vulnerable library root gradle caches modules files org eclipse jetty jetty server jetty server jar root gradle caches modules files org eclipse jetty jetty server jetty server jar dependency hierarchy cxf bundle jaxrs jar root library x jetty server jar vulnerable library jetty server jar the core jetty server artifact library home page a href path to vulnerable library root gradle caches modules files org eclipse jetty jetty server jetty server jar root gradle caches modules files org eclipse jetty jetty server jetty server jar dependency hierarchy cxf bundle jaxrs jar root library x jetty server jar vulnerable library found in head commit a href vulnerability details in eclipse jetty versions x and older x all configurations and x non default configuration with compliance enabled http is handled poorly an http style request line i e method space uri space version that declares a version of http was accepted and treated as a request if deployed behind an intermediary that also accepted and passed through the version but did not act on it then the response sent could be interpreted by the intermediary as http headers this could be used to poison the cache if the server allowed the origin client to generate arbitrary content in the response publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href fix resolution the vendor has issued a fix the vendor advisory is available at isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in eclipse jetty versions x and older x all configurations and x non default configuration with compliance enabled http is handled poorly an http style request line i e method space uri space version that declares a version of http was accepted and treated as a request if deployed behind an intermediary that also accepted and passed through the version but did not act on it then the response sent could be interpreted by the intermediary as http headers this could be used to poison the cache if the server allowed the origin client to generate arbitrary content in the response vulnerabilityurl | 0 |
194,468 | 14,679,230,841 | IssuesEvent | 2020-12-31 06:21:37 | github-vet/rangeloop-pointer-findings | https://api.github.com/repos/github-vet/rangeloop-pointer-findings | closed | vmware-tanzu/crash-diagnostics: starlark/copy_from_test.go; 14 LoC | fresh small test |
Found a possible issue in [vmware-tanzu/crash-diagnostics](https://www.github.com/vmware-tanzu/crash-diagnostics) at [starlark/copy_from_test.go](https://github.com/vmware-tanzu/crash-diagnostics/blob/3e9c3f5f9b6009858724cb3e0aa02e4a0ab1ddf3/starlark/copy_from_test.go#L358-L371)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable test used in defer or goroutine at line 363
[Click here to see the code in its original context.](https://github.com/vmware-tanzu/crash-diagnostics/blob/3e9c3f5f9b6009858724cb3e0aa02e4a0ab1ddf3/starlark/copy_from_test.go#L358-L371)
<details>
<summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary>
```go
for _, test := range tests {
for file, content := range test.remoteFiles {
ssh.MakeRemoteTestSSHFile(t, sshArgs, file, content)
}
defer func() {
for file := range test.remoteFiles {
ssh.RemoveRemoteTestSSHFile(t, sshArgs, file)
}
}()
t.Run(test.name, func(t *testing.T) {
test.eval(t, test.script)
})
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 3e9c3f5f9b6009858724cb3e0aa02e4a0ab1ddf3
| 1.0 | vmware-tanzu/crash-diagnostics: starlark/copy_from_test.go; 14 LoC -
Found a possible issue in [vmware-tanzu/crash-diagnostics](https://www.github.com/vmware-tanzu/crash-diagnostics) at [starlark/copy_from_test.go](https://github.com/vmware-tanzu/crash-diagnostics/blob/3e9c3f5f9b6009858724cb3e0aa02e4a0ab1ddf3/starlark/copy_from_test.go#L358-L371)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable test used in defer or goroutine at line 363
[Click here to see the code in its original context.](https://github.com/vmware-tanzu/crash-diagnostics/blob/3e9c3f5f9b6009858724cb3e0aa02e4a0ab1ddf3/starlark/copy_from_test.go#L358-L371)
<details>
<summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary>
```go
for _, test := range tests {
for file, content := range test.remoteFiles {
ssh.MakeRemoteTestSSHFile(t, sshArgs, file, content)
}
defer func() {
for file := range test.remoteFiles {
ssh.RemoveRemoteTestSSHFile(t, sshArgs, file)
}
}()
t.Run(test.name, func(t *testing.T) {
test.eval(t, test.script)
})
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 3e9c3f5f9b6009858724cb3e0aa02e4a0ab1ddf3
| test | vmware tanzu crash diagnostics starlark copy from test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable test used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for test range tests for file content range test remotefiles ssh makeremotetestsshfile t sshargs file content defer func for file range test remotefiles ssh removeremotetestsshfile t sshargs file t run test name func t testing t test eval t test script leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 1 |
142,115 | 11,455,346,600 | IssuesEvent | 2020-02-06 18:54:47 | Libensemble/libensemble | https://api.github.com/repos/Libensemble/libensemble | opened | Try out gitlab CI on Theta | Experimental Testing | Try running tests on Theta via local site gitlab CI at Argonne.
Argonne contact: Kyle Shaver
Documentation at: https://ecp-ci.gitlab.io/
| 1.0 | Try out gitlab CI on Theta - Try running tests on Theta via local site gitlab CI at Argonne.
Argonne contact: Kyle Shaver
Documentation at: https://ecp-ci.gitlab.io/
| test | try out gitlab ci on theta try running tests on theta via local site gitlab ci at argonne argonne contact kyle shaver documentation at | 1 |
282,139 | 30,889,182,999 | IssuesEvent | 2023-08-04 02:21:34 | Nidhi77777/linux-4.1.15 | https://api.github.com/repos/Nidhi77777/linux-4.1.15 | reopened | CVE-2023-0266 (High) detected in linuxlinux-4.6 | Mend: dependency security vulnerability | ## CVE-2023-0266 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A use after free vulnerability exists in the ALSA PCM package in the Linux Kernel. SNDRV_CTL_IOCTL_ELEM_{READ|WRITE}32 is missing locks that can be used in a use-after-free that can result in a priviledge escalation to gain ring0 access from the system user. We recommend upgrading past commit 56b88b50565cd8b946a2d00b0c83927b7ebb055e
<p>Publish Date: 2023-01-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-0266>CVE-2023-0266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-0266">https://www.linuxkernelcves.com/cves/CVE-2023-0266</a></p>
<p>Release Date: 2023-01-30</p>
<p>Fix Resolution: v4.14.303,v4.19.270,v5.4.229,v5.10.163,v5.15.88,v6.1.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-0266 (High) detected in linuxlinux-4.6 - ## CVE-2023-0266 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A use after free vulnerability exists in the ALSA PCM package in the Linux Kernel. SNDRV_CTL_IOCTL_ELEM_{READ|WRITE}32 is missing locks that can be used in a use-after-free that can result in a priviledge escalation to gain ring0 access from the system user. We recommend upgrading past commit 56b88b50565cd8b946a2d00b0c83927b7ebb055e
<p>Publish Date: 2023-01-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-0266>CVE-2023-0266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-0266">https://www.linuxkernelcves.com/cves/CVE-2023-0266</a></p>
<p>Release Date: 2023-01-30</p>
<p>Fix Resolution: v4.14.303,v4.19.270,v5.4.229,v5.10.163,v5.15.88,v6.1.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files vulnerability details a use after free vulnerability exists in the alsa pcm package in the linux kernel sndrv ctl ioctl elem read write is missing locks that can be used in a use after free that can result in a priviledge escalation to gain access from the system user we recommend upgrading past commit publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
119,892 | 10,077,045,179 | IssuesEvent | 2019-07-24 17:45:23 | eclipse/openj9 | https://api.github.com/repos/eclipse/openj9 | closed | getDependencies.pl has hard-coded build number for asmtools.jar | comp:test | TestConfig/scripts/tools/getDependencies.pl:
```
my %asmtools = (
url => 'https://ci.adoptopenjdk.net/view/Dependencies/job/asmtools/107/artifact/asmtools.jar',
fname => 'asmtools.jar',
sha1 => '04cf07c584121c2e5a3d1dad2839fc8ab4828b6d'
);
```
Build 107 no longer exists.
Note that this affects designer testing.
Please update to a newer version and adjust the SHA accordingly.
As this issue is likely to recur, can we use a more durable reference to `asmtools.jar`? | 1.0 | getDependencies.pl has hard-coded build number for asmtools.jar - TestConfig/scripts/tools/getDependencies.pl:
```
my %asmtools = (
url => 'https://ci.adoptopenjdk.net/view/Dependencies/job/asmtools/107/artifact/asmtools.jar',
fname => 'asmtools.jar',
sha1 => '04cf07c584121c2e5a3d1dad2839fc8ab4828b6d'
);
```
Build 107 no longer exists.
Note that this affects designer testing.
Please update to a newer version and adjust the SHA accordingly.
As this issue is likely to recur, can we use a more durable reference to `asmtools.jar`? | test | getdependencies pl has hard coded build number for asmtools jar testconfig scripts tools getdependencies pl my asmtools url fname asmtools jar build no longer exists note that this affects designer testing please update to a newer version and adjust the sha accordingly as this issue is likely to recur can we use a more durable reference to asmtools jar | 1 |
31,902 | 6,658,335,867 | IssuesEvent | 2017-09-30 17:59:28 | MarkSummerville/tvrename | https://api.github.com/repos/MarkSummerville/tvrename | closed | TV Rename returning German TVDB response instead of English for Star Trek (tvdb ID 77526) | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Enter '77526' in Add/Edit show dialogue box.
2. Returns: "Raumschiff Enterprise" rather than Star Trek which is the result
of visiting http://www.thetvdb.com/?tab=series&id=77526
3. Overview text in 'My Shows' is in German.
What is the expected output? What do you see instead?
Expected to see results in English, which is default Language (I've actually
moved German to the bottom of the language list in case this was the cause of
the issue, to no effect)
What version of the product are you using? On what operating system?
TV Rename 2.2.0b7 - same issue happened in 2.2.0b5, on Windows XP 32-bit.
Please provide any additional information below.
```
Original issue reported on code.google.com by `mabsy...@gmail.com` on 23 Jan 2011 at 9:16
| 1.0 | TV Rename returning German TVDB response instead of English for Star Trek (tvdb ID 77526) - ```
What steps will reproduce the problem?
1. Enter '77526' in Add/Edit show dialogue box.
2. Returns: "Raumschiff Enterprise" rather than Star Trek which is the result
of visiting http://www.thetvdb.com/?tab=series&id=77526
3. Overview text in 'My Shows' is in German.
What is the expected output? What do you see instead?
Expected to see results in English, which is default Language (I've actually
moved German to the bottom of the language list in case this was the cause of
the issue, to no effect)
What version of the product are you using? On what operating system?
TV Rename 2.2.0b7 - same issue happened in 2.2.0b5, on Windows XP 32-bit.
Please provide any additional information below.
```
Original issue reported on code.google.com by `mabsy...@gmail.com` on 23 Jan 2011 at 9:16
| non_test | tv rename returning german tvdb response instead of english for star trek tvdb id what steps will reproduce the problem enter in add edit show dialogue box returns raumschiff enterprise rather than star trek which is the result of visiting overview text in my shows is in german what is the expected output what do you see instead expected to see results in english which is default language i ve actually moved german to the bottom of the language list in case this was the cause of the issue to no effect what version of the product are you using on what operating system tv rename same issue happened in on windows xp bit please provide any additional information below original issue reported on code google com by mabsy gmail com on jan at | 0 |
177,639 | 6,586,027,947 | IssuesEvent | 2017-09-13 15:48:27 | sys-bio/roadrunner | https://api.github.com/repos/sys-bio/roadrunner | opened | r.plot(linewidth=2) does not work | bug priority | Some examples in the readthedocs documentation shows plot calls such as:
r.plot(s, loc=None, show=False, color=color, linewidth=2.0)
When I tried:
r,plot (linewidth=2.0)
I get the error message:
plot() got an unexpected keyword argument 'linewidth'
| 1.0 | r.plot(linewidth=2) does not work - Some examples in the readthedocs documentation shows plot calls such as:
r.plot(s, loc=None, show=False, color=color, linewidth=2.0)
When I tried:
r,plot (linewidth=2.0)
I get the error message:
plot() got an unexpected keyword argument 'linewidth'
| non_test | r plot linewidth does not work some examples in the readthedocs documentation shows plot calls such as r plot s loc none show false color color linewidth when i tried r plot linewidth i get the error message plot got an unexpected keyword argument linewidth | 0 |
7,702 | 2,920,706,863 | IssuesEvent | 2015-06-24 20:21:10 | schroet/rdf2salesforce | https://api.github.com/repos/schroet/rdf2salesforce | opened | Moch salesforce rest api | testing | In order to run some basic test for the Services we need to mock the salesforce rest api. | 1.0 | Moch salesforce rest api - In order to run some basic test for the Services we need to mock the salesforce rest api. | test | moch salesforce rest api in order to run some basic test for the services we need to mock the salesforce rest api | 1 |
251,600 | 8,017,843,098 | IssuesEvent | 2018-07-25 17:10:29 | MontrealCorpusTools/polyglot-server | https://api.github.com/repos/MontrealCorpusTools/polyglot-server | opened | Add enrichment for containing certain lower annotations | API UI enrichment low priority | For Menzerath, for instance, it's important to encode on utterances whether they contain any `<unk>` words and filter based on that, since it's unknown how many syllables are in them. | 1.0 | Add enrichment for containing certain lower annotations - For Menzerath, for instance, it's important to encode on utterances whether they contain any `<unk>` words and filter based on that, since it's unknown how many syllables are in them. | non_test | add enrichment for containing certain lower annotations for menzerath for instance it s important to encode on utterances whether they contain any words and filter based on that since it s unknown how many syllables are in them | 0 |
55,106 | 6,426,670,315 | IssuesEvent | 2017-08-09 17:57:31 | Ericsson/codechecker | https://api.github.com/repos/Ericsson/codechecker | opened | Make testing infrastructure more resillient by having the ability to read into the built package's context | CI enhancement test | So things like #786 can be made fancy and without loss of coverage. Perhaps a special CodeChecker subcommand that is only build when `make test` (and likes) are executed, which can spit out the context on demand for the test code. | 1.0 | Make testing infrastructure more resillient by having the ability to read into the built package's context - So things like #786 can be made fancy and without loss of coverage. Perhaps a special CodeChecker subcommand that is only build when `make test` (and likes) are executed, which can spit out the context on demand for the test code. | test | make testing infrastructure more resillient by having the ability to read into the built package s context so things like can be made fancy and without loss of coverage perhaps a special codechecker subcommand that is only build when make test and likes are executed which can spit out the context on demand for the test code | 1 |
210,127 | 16,087,788,828 | IssuesEvent | 2021-04-26 13:24:13 | skyjake/lagrange | https://api.github.com/repos/skyjake/lagrange | closed | HTTP/HTTPS Links with Query Strings | bug needs testing question | When I try opening an http link with a query string, the URL passed to the default web browser is missing the query string.
Copying the link from the gemini capsule will copy the full link, which means the problem is when it passes the link to the default web browser to handle. | 1.0 | HTTP/HTTPS Links with Query Strings - When I try opening an http link with a query string, the URL passed to the default web browser is missing the query string.
Copying the link from the gemini capsule will copy the full link, which means the problem is when it passes the link to the default web browser to handle. | test | http https links with query strings when i try opening an http link with a query string the url passed to the default web browser is missing the query string copying the link from the gemini capsule will copy the full link which means the problem is when it passes the link to the default web browser to handle | 1 |
261,224 | 22,707,014,351 | IssuesEvent | 2022-07-05 15:27:20 | MohistMC/Mohist | https://api.github.com/repos/MohistMC/Mohist | closed | [1.18.2] authlib cant get player name because bmorph refmap and mixin not friends | Wait Needs Testing | <!-- ISSUE_TEMPLATE_1 -> IMPORTANT: DO NOT DELETE THIS LINE.-->
<!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.-->
<!-- If you don't know where to upload your logs and crash reports, you can use these websites : -->
<!-- https://gist.github.com (recommended) -->
<!-- https://mclo.gs -->
<!-- https://haste.mohistmc.com -->
<!-- https://pastebin.com -->
<!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT -->
**Minecraft Version :** 1.18.2
**Mohist Version :** 49
**Operating System :** Ubuntu
**Concerned mod / plugin** : [Bmorph](https://github.com/Budschie/BudschieMorphMod/)
**Logs :** [default logs](https://jelebruh.xyz/files/forge.log)
**Steps to Reproduce :**
1. Install bmorph
2. Start server
3. Say player to join
4. see error
**Description of issue :** Authlib can't get player name and linked to bmorph, in default forge mod working normal (after 1.0.1). Idk what is this, see it for the first time
| 1.0 | [1.18.2] authlib cant get player name because bmorph refmap and mixin not friends - <!-- ISSUE_TEMPLATE_1 -> IMPORTANT: DO NOT DELETE THIS LINE.-->
<!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.-->
<!-- If you don't know where to upload your logs and crash reports, you can use these websites : -->
<!-- https://gist.github.com (recommended) -->
<!-- https://mclo.gs -->
<!-- https://haste.mohistmc.com -->
<!-- https://pastebin.com -->
<!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT -->
**Minecraft Version :** 1.18.2
**Mohist Version :** 49
**Operating System :** Ubuntu
**Concerned mod / plugin** : [Bmorph](https://github.com/Budschie/BudschieMorphMod/)
**Logs :** [default logs](https://jelebruh.xyz/files/forge.log)
**Steps to Reproduce :**
1. Install bmorph
2. Start server
3. Say player to join
4. see error
**Description of issue :** Authlib can't get player name and linked to bmorph, in default forge mod working normal (after 1.0.1). Idk what is this, see it for the first time
| test | authlib cant get player name because bmorph refmap and mixin not friends important do not delete this line minecraft version mohist version operating system ubuntu concerned mod plugin logs steps to reproduce install bmorph start server say player to join see error description of issue authlib can t get player name and linked to bmorph in default forge mod working normal after idk what is this see it for the first time | 1 |
280,030 | 24,274,984,955 | IssuesEvent | 2022-09-28 13:16:16 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | Failing test: Chrome X-Pack UI Functional Tests - ML data_frame_analytics.x-pack/test/functional/apps/ml/data_frame_analytics/classification_creation·ts - machine learning - data frame analytics classification creation bank marketing navigates through the wizard and sets all needed fields | failed-test | A test failed on a tracked branch
```
Error: Timeout of 360000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/var/lib/buildkite-agent/builds/kb-n2-4-spot-e7807e7b94bfe466/elastic/kibana-on-merge/kibana/x-pack/test/functional/apps/ml/data_frame_analytics/classification_creation.ts)
at listOnTimeout (node:internal/timers:559:17)
at processTimers (node:internal/timers:502:7) {
code: 'ERR_MOCHA_TIMEOUT',
timeout: 360000,
file: '/var/lib/buildkite-agent/builds/kb-n2-4-spot-e7807e7b94bfe466/elastic/kibana-on-merge/kibana/x-pack/test/functional/apps/ml/data_frame_analytics/classification_creation.ts'
}
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/21688#018383e3-6d48-4adc-b75e-2d20b0edd52c)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests - ML data_frame_analytics.x-pack/test/functional/apps/ml/data_frame_analytics/classification_creation·ts","test.name":"machine learning - data frame analytics classification creation bank marketing navigates through the wizard and sets all needed fields","test.failCount":1}} --> | 1.0 | Failing test: Chrome X-Pack UI Functional Tests - ML data_frame_analytics.x-pack/test/functional/apps/ml/data_frame_analytics/classification_creation·ts - machine learning - data frame analytics classification creation bank marketing navigates through the wizard and sets all needed fields - A test failed on a tracked branch
```
Error: Timeout of 360000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/var/lib/buildkite-agent/builds/kb-n2-4-spot-e7807e7b94bfe466/elastic/kibana-on-merge/kibana/x-pack/test/functional/apps/ml/data_frame_analytics/classification_creation.ts)
at listOnTimeout (node:internal/timers:559:17)
at processTimers (node:internal/timers:502:7) {
code: 'ERR_MOCHA_TIMEOUT',
timeout: 360000,
file: '/var/lib/buildkite-agent/builds/kb-n2-4-spot-e7807e7b94bfe466/elastic/kibana-on-merge/kibana/x-pack/test/functional/apps/ml/data_frame_analytics/classification_creation.ts'
}
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/21688#018383e3-6d48-4adc-b75e-2d20b0edd52c)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests - ML data_frame_analytics.x-pack/test/functional/apps/ml/data_frame_analytics/classification_creation·ts","test.name":"machine learning - data frame analytics classification creation bank marketing navigates through the wizard and sets all needed fields","test.failCount":1}} --> | test | failing test chrome x pack ui functional tests ml data frame analytics x pack test functional apps ml data frame analytics classification creation·ts machine learning data frame analytics classification creation bank marketing navigates through the wizard and sets all needed fields a test failed on a tracked branch error timeout of exceeded for async tests and hooks ensure done is called if returning a promise ensure it resolves var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack test functional apps ml data frame analytics classification creation ts at listontimeout node internal timers at processtimers node internal timers code err mocha timeout timeout file var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack test functional apps ml data frame analytics classification creation ts first failure | 1 |
60,600 | 6,711,131,936 | IssuesEvent | 2017-10-13 01:42:30 | omegaup/omegaup | https://api.github.com/repos/omegaup/omegaup | closed | Colección de problemas | Infinito omegaUp For Contests | Buscar problemas dentro de omegaup es muy difícil porque lo único por lo que puedes filtrar es por nombre.
Si pudieramos hacer colecciones de problemas facilitaría la organización de los problemas.
Ejemplos:
http://www.cmirg.com/karelotitlan/Pantallas/problemasOMI.aspx
https://lobishomen.wordpress.com/2014/07/23/omi-training-gate-en-omega-up/
https://lobishomen.wordpress.com/2014/07/23/problemas-del-omi-training-gate-en-clasificacion-tematica/
| 1.0 | Colección de problemas - Buscar problemas dentro de omegaup es muy difícil porque lo único por lo que puedes filtrar es por nombre.
Si pudieramos hacer colecciones de problemas facilitaría la organización de los problemas.
Ejemplos:
http://www.cmirg.com/karelotitlan/Pantallas/problemasOMI.aspx
https://lobishomen.wordpress.com/2014/07/23/omi-training-gate-en-omega-up/
https://lobishomen.wordpress.com/2014/07/23/problemas-del-omi-training-gate-en-clasificacion-tematica/
| test | colección de problemas buscar problemas dentro de omegaup es muy difícil porque lo único por lo que puedes filtrar es por nombre si pudieramos hacer colecciones de problemas facilitaría la organización de los problemas ejemplos | 1 |
1,685 | 4,328,565,518 | IssuesEvent | 2016-07-26 14:25:43 | CGAL/cgal | https://api.github.com/repos/CGAL/cgal | closed | Add light display on point sets with normals in Polyhedron demo | CGAL 3D demo feature request Pkg::Point_set_processing | When point sets have normals, a better and clearer display can be done using light (similarly to a polyhedron).
(@maxGimeno If you add this to your todo-list, this is [the commit](https://github.com/CGAL/cgal-dev/commit/455843e1bf01faf87eda7994b4c715d7ea173b6d) where we worked on it together last time, although you can't use it directly because it modifies other unrelated files – sorry about that, I should have separated it.) | 1.0 | Add light display on point sets with normals in Polyhedron demo - When point sets have normals, a better and clearer display can be done using light (similarly to a polyhedron).
(@maxGimeno If you add this to your todo-list, this is [the commit](https://github.com/CGAL/cgal-dev/commit/455843e1bf01faf87eda7994b4c715d7ea173b6d) where we worked on it together last time, although you can't use it directly because it modifies other unrelated files – sorry about that, I should have separated it.) | non_test | add light display on point sets with normals in polyhedron demo when point sets have normals a better and clearer display can be done using light similarly to a polyhedron maxgimeno if you add this to your todo list this is where we worked on it together last time although you can t use it directly because it modifies other unrelated files – sorry about that i should have separated it | 0 |
204,853 | 7,091,540,829 | IssuesEvent | 2018-01-12 13:30:39 | IntegratedAlarmSystem-Group/ias | https://api.github.com/repos/IntegratedAlarmSystem-Group/ias | opened | Review Supervisor, DASU and ASCE constructors | Priority-low enhancement | DASU and Supervisor constructors ignore the DAO and read it again with the passed CDB reader.
We should review and improve DASU and Supervisor constructor and uniform the ASCE constructor | 1.0 | Review Supervisor, DASU and ASCE constructors - DASU and Supervisor constructors ignore the DAO and read it again with the passed CDB reader.
We should review and improve DASU and Supervisor constructor and uniform the ASCE constructor | non_test | review supervisor dasu and asce constructors dasu and supervisor constructors ignore the dao and read it again with the passed cdb reader we should review and improve dasu and supervisor constructor and uniform the asce constructor | 0 |
74,090 | 7,374,574,565 | IssuesEvent | 2018-03-13 20:47:35 | 18F/doi-extractives-data | https://api.github.com/repos/18F/doi-extractives-data | closed | Document the role of Circle, Hound, federalist in workflow | Transition training workflow:testing | Can we talk about the role of Circle, Hound, and federalist checks in the GitHub workflow?
- What role does each play in merging/publishing?
- What are some common issues we might confront with each?
- How do we resolve issues with each when they arise?
| 1.0 | Document the role of Circle, Hound, federalist in workflow - Can we talk about the role of Circle, Hound, and federalist checks in the GitHub workflow?
- What role does each play in merging/publishing?
- What are some common issues we might confront with each?
- How do we resolve issues with each when they arise?
| test | document the role of circle hound federalist in workflow can we talk about the role of circle hound and federalist checks in the github workflow what role does each play in merging publishing what are some common issues we might confront with each how do we resolve issues with each when they arise | 1 |
64,657 | 18,786,900,800 | IssuesEvent | 2021-11-08 13:07:43 | galasa-dev/projectmanagement | https://api.github.com/repos/galasa-dev/projectmanagement | reopened | Fix regression failures for the Local Compilation tests | defect SimBank regression | Both:
dev.galasa.inttests.compilation.simbank.local.online.CompilationLocalJava08Ubuntu
dev.galasa.inttests.compilation.simbank.local.online.CompilationLocalJava11Ubuntu
Are failing and have caught a regression in the levels of org.osgi.service.component that are available. The tests are both online compilation tests which are failing when compiling the simbank manager. The simbank manager requires 1.3.+ of org.osgi:org.osgi.service.component.annotations. However in maven we only have listed 1.5.0 within the metadata thus a match to 1.3+ cannot be found.
Elsewhere we use 1.3.0 which is in maven central (just not the metadata) so the manager should just use that instead | 1.0 | Fix regression failures for the Local Compilation tests - Both:
dev.galasa.inttests.compilation.simbank.local.online.CompilationLocalJava08Ubuntu
dev.galasa.inttests.compilation.simbank.local.online.CompilationLocalJava11Ubuntu
Are failing and have caught a regression in the levels of org.osgi.service.component that are available. The tests are both online compilation tests which are failing when compiling the simbank manager. The simbank manager requires 1.3.+ of org.osgi:org.osgi.service.component.annotations. However in maven we only have listed 1.5.0 within the metadata thus a match to 1.3+ cannot be found.
Elsewhere we use 1.3.0 which is in maven central (just not the metadata) so the manager should just use that instead | non_test | fix regression failures for the local compilation tests both dev galasa inttests compilation simbank local online dev galasa inttests compilation simbank local online are failing and have caught a regression in the levels of org osgi service component that are available the tests are both online compilation tests which are failing when compiling the simbank manager the simbank manager requires of org osgi org osgi service component annotations however in maven we only have listed within the metadata thus a match to cannot be found elsewhere we use which is in maven central just not the metadata so the manager should just use that instead | 0 |
97,433 | 8,655,728,904 | IssuesEvent | 2018-11-27 16:36:39 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | opened | Clean/Dirty Install Smoke Test | Quality: Smoke test | # Clean/dirty install smoke tests
Please run these tests on the compiled package of Mantid; **not a locally compiled version**.
When reporting back please include:
- Your OS
- Your facility/location
- If you have other package versions of Mantid installed (yes/no)
If you have any questions please don't hesitate to contact the creator of this issue.
## Checklist/directions
### Dirty install
* Make sure that you have several versions of Mantid installed
* Last release
* A nightly
* If possible an old release
* Install the latest version of the new Mantid
* Check that it boots up correctly
### Clean install
* Remove all existing Mantid versions and associated files
* Windows:
* Uninstall the program
* Clear shortcuts from desktop
* Clean out the registry
* Load regedit (Command Prompt > regedit)
* Find `HKEY_CURRENT_USER > SOFTWARE > Mantid` delete it
* OSX:
* Remove the application
* Remove the `~/.mantid directory`
* Remove `~/Library/Preferences/org.mantidproject.MantidPlot.plist`
* Linux
* Remove the package: `/opt/Mantid`
* Remove `~/.config/Mantid`
* Remove `~/.mantid/`
* Re-install the latest version of the new Mantid
* Check that it boots up correctly
| 1.0 | Clean/Dirty Install Smoke Test - # Clean/dirty install smoke tests
Please run these tests on the compiled package of Mantid; **not a locally compiled version**.
When reporting back please include:
- Your OS
- Your facility/location
- If you have other package versions of Mantid installed (yes/no)
If you have any questions please don't hesitate to contact the creator of this issue.
## Checklist/directions
### Dirty install
* Make sure that you have several versions of Mantid installed
* Last release
* A nightly
* If possible an old release
* Install the latest version of the new Mantid
* Check that it boots up correctly
### Clean install
* Remove all existing Mantid versions and associated files
* Windows:
* Uninstall the program
* Clear shortcuts from desktop
* Clean out the registry
* Load regedit (Command Prompt > regedit)
* Find `HKEY_CURRENT_USER > SOFTWARE > Mantid` delete it
* OSX:
* Remove the application
* Remove the `~/.mantid directory`
* Remove `~/Library/Preferences/org.mantidproject.MantidPlot.plist`
* Linux
* Remove the package: `/opt/Mantid`
* Remove `~/.config/Mantid`
* Remove `~/.mantid/`
* Re-install the latest version of the new Mantid
* Check that it boots up correctly
| test | clean dirty install smoke test clean dirty install smoke tests please run these tests on the compiled package of mantid not a locally compiled version when reporting back please include your os your facility location if you have other package versions of mantid installed yes no if you have any questions please don t hesitate to contact the creator of this issue checklist directions dirty install make sure that you have several versions of mantid installed last release a nightly if possible an old release install the latest version of the new mantid check that it boots up correctly clean install remove all existing mantid versions and associated files windows uninstall the program clear shortcuts from desktop clean out the registry load regedit command prompt regedit find hkey current user software mantid delete it osx remove the application remove the mantid directory remove library preferences org mantidproject mantidplot plist linux remove the package opt mantid remove config mantid remove mantid re install the latest version of the new mantid check that it boots up correctly | 1 |
137,670 | 11,149,803,518 | IssuesEvent | 2019-12-23 20:01:39 | python-discord/bot | https://api.github.com/repos/python-discord/bot | closed | Write unit tests for `bot/rules/mentions.py` | area: tests priority: 1 - high | Write unit tests for [`bot/rules/mentions.py`](../blob/master/bot/rules/mentions.py).
## Implementation details
Please make sure to read the general information in the [meta issue](553) and the [testing README](../blob/master/tests/README.md). We are aiming for a 100% [branch coverage](https://coverage.readthedocs.io/en/stable/branch.html) for this file, but if you think that is not possible, please discuss that in this issue.
## Additional information
If you want to work on this issue, **please make sure that you get assigned to it** by one of the core devs before starting to work on it. We would like to prevent the situation that multiple people are working on the same issue. To get assigned, leave a comment showing your interesting in tackling this issue.
| 1.0 | Write unit tests for `bot/rules/mentions.py` - Write unit tests for [`bot/rules/mentions.py`](../blob/master/bot/rules/mentions.py).
## Implementation details
Please make sure to read the general information in the [meta issue](553) and the [testing README](../blob/master/tests/README.md). We are aiming for a 100% [branch coverage](https://coverage.readthedocs.io/en/stable/branch.html) for this file, but if you think that is not possible, please discuss that in this issue.
## Additional information
If you want to work on this issue, **please make sure that you get assigned to it** by one of the core devs before starting to work on it. We would like to prevent the situation that multiple people are working on the same issue. To get assigned, leave a comment showing your interesting in tackling this issue.
| test | write unit tests for bot rules mentions py write unit tests for blob master bot rules mentions py implementation details please make sure to read the general information in the and the blob master tests readme md we are aiming for a for this file but if you think that is not possible please discuss that in this issue additional information if you want to work on this issue please make sure that you get assigned to it by one of the core devs before starting to work on it we would like to prevent the situation that multiple people are working on the same issue to get assigned leave a comment showing your interesting in tackling this issue | 1 |
82,986 | 10,314,149,830 | IssuesEvent | 2019-08-30 02:10:37 | readthedocs/readthedocs.org | https://api.github.com/repos/readthedocs/readthedocs.org | opened | Cache in footer view | Improvement Needed: design decision | The footer view is assessed for each page view. We should make it faster :)
Ref: https://github.com/readthedocs/readthedocs.org/pull/6125#discussion_r319328817 | 1.0 | Cache in footer view - The footer view is assessed for each page view. We should make it faster :)
Ref: https://github.com/readthedocs/readthedocs.org/pull/6125#discussion_r319328817 | non_test | cache in footer view the footer view is assessed for each page view we should make it faster ref | 0 |
3,523 | 2,677,708,317 | IssuesEvent | 2015-03-26 02:44:01 | TheSLinux/gs | https://api.github.com/repos/TheSLinux/gs | closed | uim-vi: critical problem with Firefox > 34 | bug testing uim-vi _critical _important | The current `uim-vi` (1.9) can cause Firefox > 34 to crash.
Please take a look and find the cause. The first thing is to test if a similar problem occur with `uim-vi-1.8.7` | 1.0 | uim-vi: critical problem with Firefox > 34 - The current `uim-vi` (1.9) can cause Firefox > 34 to crash.
Please take a look and find the cause. The first thing is to test if a similar problem occur with `uim-vi-1.8.7` | test | uim vi critical problem with firefox the current uim vi can cause firefox to crash please take a look and find the cause the first thing is to test if a similar problem occur with uim vi | 1 |
23,158 | 3,995,398,434 | IssuesEvent | 2016-05-10 15:24:33 | owncloud/client | https://api.github.com/repos/owncloud/client | closed | [Win] [2.2.0 beta1] Server messages in share view too pale | blue-ticket ReadyToTest | @dragotin The error message is nearly invisible:

Text color switches to black after the user clicks to another application window. | 1.0 | [Win] [2.2.0 beta1] Server messages in share view too pale - @dragotin The error message is nearly invisible:

Text color switches to black after the user clicks to another application window. | test | server messages in share view too pale dragotin the error message is nearly invisible text color switches to black after the user clicks to another application window | 1 |
68,233 | 28,302,379,452 | IssuesEvent | 2023-04-10 07:33:30 | microsoft/BotFramework-WebChat | https://api.github.com/repos/microsoft/BotFramework-WebChat | opened | Jittery Speech on iOS | bug customer-reported Bot Services | ### Is it an issue related to Adaptive Cards?
No
### Is this an accessibility issue?
No
### What version of Web Chat are you using?
Latest production
### Which distribution are you using Web Chat from?
NPM
### Which hosting environment does this issue primarily affect?
Web apps
### Which browsers and platforms do the issue happened?
Platform: iOS/iPadOS
### Which area does this issue affect?
Speech
### What is the public URL for the website?
None
### Please describe the bug
The webchat package is being used to facilitate a voice-only journey.
As soon as the conversation is started, the bot sends an activity to the user with the input hint "expectingInput" causing the mic to be opened after the message is read.
Once the mic turns on and user's input is taken, the next utterance from the bot is spoken with a lot of lag (short but many pauses between the synthesized speech, almost as if this is utilizing a lot of resources).
This issue happens on few iOS devices like iPhone 11 (iOS v16.3), iPhone 12 (iOS v16.4.1), while it is working fine on iPhone 14 Pro (iOS v16.4), iPhone 14 Pro Max (iOS v16.4).
### Do you see any errors in console log?
```js
No errors in the console.
```
### How to reproduce the issue?
Using an affected device (listed above)
1. Start a conversation with a speech-enabled bot.
2. The first activity from the bot is spoken properly.
3. Once the mic turns on give any input to the bot.
4. The bug starts here, all activities from the bot will be jittery when played by webchat.
### What do you expect?
All utterances from the bot should be spoken properly on all devices.
### What actually happened?
Spoken activities from the bot are jittery after the mic was turned on.
### Do you have any screenshots or recordings to repro the issue?
Attached is the [audio](https://github.com/microsoft/BotFramework-WebChat/files/11188490/Issue-voice-lag-ios-VID-20230407-WA0009.zip) of this exchange.
### Adaptive Card JSON
_No response_
### Additional context
_No response_ | 1.0 | Jittery Speech on iOS - ### Is it an issue related to Adaptive Cards?
No
### Is this an accessibility issue?
No
### What version of Web Chat are you using?
Latest production
### Which distribution are you using Web Chat from?
NPM
### Which hosting environment does this issue primarily affect?
Web apps
### Which browsers and platforms do the issue happened?
Platform: iOS/iPadOS
### Which area does this issue affect?
Speech
### What is the public URL for the website?
None
### Please describe the bug
The webchat package is being used to facilitate a voice-only journey.
As soon as the conversation is started, the bot sends an activity to the user with the input hint "expectingInput" causing the mic to be opened after the message is read.
Once the mic turns on and user's input is taken, the next utterance from the bot is spoken with a lot of lag (short but many pauses between the synthesized speech, almost as if this is utilizing a lot of resources).
This issue happens on few iOS devices like iPhone 11 (iOS v16.3), iPhone 12 (iOS v16.4.1), while it is working fine on iPhone 14 Pro (iOS v16.4), iPhone 14 Pro Max (iOS v16.4).
### Do you see any errors in console log?
```js
No errors in the console.
```
### How to reproduce the issue?
Using an affected device (listed above)
1. Start a conversation with a speech-enabled bot.
2. The first activity from the bot is spoken properly.
3. Once the mic turns on give any input to the bot.
4. The bug starts here, all activities from the bot will be jittery when played by webchat.
### What do you expect?
All utterances from the bot should be spoken properly on all devices.
### What actually happened?
Spoken activities from the bot are jittery after the mic was turned on.
### Do you have any screenshots or recordings to repro the issue?
Attached is the [audio](https://github.com/microsoft/BotFramework-WebChat/files/11188490/Issue-voice-lag-ios-VID-20230407-WA0009.zip) of this exchange.
### Adaptive Card JSON
_No response_
### Additional context
_No response_ | non_test | jittery speech on ios is it an issue related to adaptive cards no is this an accessibility issue no what version of web chat are you using latest production which distribution are you using web chat from npm which hosting environment does this issue primarily affect web apps which browsers and platforms do the issue happened platform ios ipados which area does this issue affect speech what is the public url for the website none please describe the bug the webchat package is being used to facilitate a voice only journey as soon as the conversation is started the bot sends an activity to the user with the input hint expectinginput causing the mic to be opened after the message is read once the mic turns on and user s input is taken the next utterance from the bot is spoken with a lot of lag short but many pauses between the synthesized speech almost as if this is utilizing a lot of resources this issue happens on few ios devices like iphone ios iphone ios while it is working fine on iphone pro ios iphone pro max ios do you see any errors in console log js no errors in the console how to reproduce the issue using an affected device listed above start a conversation with a speech enabled bot the first activity from the bot is spoken properly once the mic turns on give any input to the bot the bug starts here all activities from the bot will be jittery when played by webchat what do you expect all utterances from the bot should be spoken properly on all devices what actually happened spoken activities from the bot are jittery after the mic was turned on do you have any screenshots or recordings to repro the issue attached is the of this exchange adaptive card json no response additional context no response | 0 |
56,509 | 6,521,197,431 | IssuesEvent | 2017-08-28 19:35:33 | gradle/gradle-native | https://api.github.com/repos/gradle/gradle-native | opened | Improve performance of XCTest execution from CLI | in:xctest roadmap | By minimizing the test execution time, Gradle improves the feedback loop to the developers waiting for his build to complete.
### Expected Behaviors
This epic covers:
- Test filters using `--test` flag
- Parallel test execution
### Test Cases
- [ ] Honor filters passed to `--test` flags
- [ ] Honor wildcard in filters passed to `--test` flags
- [ ] No `--test` filters executes all tests
- [ ] Test are executed in parallel | 1.0 | Improve performance of XCTest execution from CLI - By minimizing the test execution time, Gradle improves the feedback loop to the developers waiting for his build to complete.
### Expected Behaviors
This epic covers:
- Test filters using `--test` flag
- Parallel test execution
### Test Cases
- [ ] Honor filters passed to `--test` flags
- [ ] Honor wildcard in filters passed to `--test` flags
- [ ] No `--test` filters executes all tests
- [ ] Test are executed in parallel | test | improve performance of xctest execution from cli by minimizing the test execution time gradle improves the feedback loop to the developers waiting for his build to complete expected behaviors this epic covers test filters using test flag parallel test execution test cases honor filters passed to test flags honor wildcard in filters passed to test flags no test filters executes all tests test are executed in parallel | 1 |
38,454 | 5,187,827,141 | IssuesEvent | 2017-01-20 18:00:09 | ng2-ui/ng2-datetime-picker | https://api.github.com/repos/ng2-ui/ng2-datetime-picker | closed | ng2-datetime-picker fails to bind to model when set beforehand | Fixed & Waiting For Testing | Getting the standard error: EXCEPTION: Cannot read property 'fullName' of undefined
When I click into the date time picker.
<input id="myDateTime" name="myDateTime" [(ngModel)]="deleteMe" ng2-datetime-picker />
Thanks | 1.0 | ng2-datetime-picker fails to bind to model when set beforehand - Getting the standard error: EXCEPTION: Cannot read property 'fullName' of undefined
When I click into the date time picker.
<input id="myDateTime" name="myDateTime" [(ngModel)]="deleteMe" ng2-datetime-picker />
Thanks | test | datetime picker fails to bind to model when set beforehand getting the standard error exception cannot read property fullname of undefined when i click into the date time picker thanks | 1 |
13,999 | 3,371,696,797 | IssuesEvent | 2015-11-23 20:13:51 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | Test failure in CI build 9545 | test-failure | The following test appears to have failed:
[#9545](https://circleci.com/gh/cockroachdb/cockroach/9545):
```
tfef41e84 20:07:34.564191 399.461µs ·sending RPC kv/dist_sender.go:414
tfef41e84 20:07:34.564727 0 ·error: retry txn "storage/replica_command.go:1281 (*Replica).AdminSplit" id=fef41e84 key=/Local/Range/RangeDescriptor/"j" rw=true pri=99.99928428 iso=SERIALIZABLE stat=PENDING epo=22642 ts=0.000000000,940 orig=0.000000000,940 max=0.000000000,940 kv/txn_coord_sender.go:405
I1123 20:07:34.564929 1018 client/db.go:483 failed batch: retry txn "storage/replica_command.go:1281 (*Replica).AdminSplit" id=fef41e84 key=/Local/Range/RangeDescriptor/"j" rw=true pri=99.99928428 iso=SERIALIZABLE stat=PENDING epo=22642 ts=0.000000000,940 orig=0.000000000,940 max=0.000000000,940
I1123 20:07:34.566188 1018 storage/replica.go:1146 found sequence cache entry for fef41e84@67953
panic: test timed out after 1m10s
goroutine 6602 [running]:
testing.startAlarm.func1()
/usr/local/go/src/testing/testing.go:703 +0x132
created by time.goFunc
/usr/local/go/src/time/sleep.go:129 +0x3a
goroutine 1 [chan receive]:
testing.RunTests(0x163adb0, 0x1c35a20, 0xd8, 0xd8, 0x884201)
/usr/local/go/src/testing/testing.go:562 +0x8ad
--
/go/src/github.com/cockroachdb/cockroach/storage/scanner.go:220 +0x29a
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc8202ce8a0, 0xc8205959c0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:88 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:89 +0x62
FAIL github.com/cockroachdb/cockroach/storage 70.045s
=== RUN TestBatchBasics
I1123 20:06:20.515325 998 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBatchBasics (0.00s)
=== RUN TestBatchGet
I1123 20:06:20.516201 998 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBatchGet (0.00s)
=== RUN TestBatchMerge
I1123 20:06:20.519624 998 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBatchMerge (0.00s)
=== RUN TestBatchProto
```
Please assign, take a look and update the issue accordingly. | 1.0 | Test failure in CI build 9545 - The following test appears to have failed:
[#9545](https://circleci.com/gh/cockroachdb/cockroach/9545):
```
tfef41e84 20:07:34.564191 399.461µs ·sending RPC kv/dist_sender.go:414
tfef41e84 20:07:34.564727 0 ·error: retry txn "storage/replica_command.go:1281 (*Replica).AdminSplit" id=fef41e84 key=/Local/Range/RangeDescriptor/"j" rw=true pri=99.99928428 iso=SERIALIZABLE stat=PENDING epo=22642 ts=0.000000000,940 orig=0.000000000,940 max=0.000000000,940 kv/txn_coord_sender.go:405
I1123 20:07:34.564929 1018 client/db.go:483 failed batch: retry txn "storage/replica_command.go:1281 (*Replica).AdminSplit" id=fef41e84 key=/Local/Range/RangeDescriptor/"j" rw=true pri=99.99928428 iso=SERIALIZABLE stat=PENDING epo=22642 ts=0.000000000,940 orig=0.000000000,940 max=0.000000000,940
I1123 20:07:34.566188 1018 storage/replica.go:1146 found sequence cache entry for fef41e84@67953
panic: test timed out after 1m10s
goroutine 6602 [running]:
testing.startAlarm.func1()
/usr/local/go/src/testing/testing.go:703 +0x132
created by time.goFunc
/usr/local/go/src/time/sleep.go:129 +0x3a
goroutine 1 [chan receive]:
testing.RunTests(0x163adb0, 0x1c35a20, 0xd8, 0xd8, 0x884201)
/usr/local/go/src/testing/testing.go:562 +0x8ad
--
/go/src/github.com/cockroachdb/cockroach/storage/scanner.go:220 +0x29a
github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc8202ce8a0, 0xc8205959c0)
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:88 +0x52
created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:89 +0x62
FAIL github.com/cockroachdb/cockroach/storage 70.045s
=== RUN TestBatchBasics
I1123 20:06:20.515325 998 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBatchBasics (0.00s)
=== RUN TestBatchGet
I1123 20:06:20.516201 998 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBatchGet (0.00s)
=== RUN TestBatchMerge
I1123 20:06:20.519624 998 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBatchMerge (0.00s)
=== RUN TestBatchProto
```
Please assign, take a look and update the issue accordingly. | test | test failure in ci build the following test appears to have failed ·sending rpc kv dist sender go ·error retry txn storage replica command go replica adminsplit id key local range rangedescriptor j rw true pri iso serializable stat pending epo ts orig max kv txn coord sender go client db go failed batch retry txn storage replica command go replica adminsplit id key local range rangedescriptor j rw true pri iso serializable stat pending epo ts orig max storage replica go found sequence cache entry for panic test timed out after goroutine testing startalarm usr local go src testing testing go created by time gofunc usr local go src time sleep go goroutine testing runtests usr local go src testing testing go go src github com cockroachdb cockroach storage scanner go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go fail github com cockroachdb cockroach storage run testbatchbasics storage engine rocksdb go closing in memory rocksdb instance pass testbatchbasics run testbatchget storage engine rocksdb go closing in memory rocksdb instance pass testbatchget run testbatchmerge storage engine rocksdb go closing in memory rocksdb instance pass testbatchmerge run testbatchproto please assign take a look and update the issue accordingly | 1 |
88,357 | 25,382,641,336 | IssuesEvent | 2022-11-21 18:54:06 | microsoft/onnxruntime | https://api.github.com/repos/microsoft/onnxruntime | closed | [Build] No matching distribution found for onnxruntime | build platform:windows | ### Describe the issue
I am installing onnxruntime when facing errors

i am windows 11 with a python 3.11
### Urgency
_No response_
### Target platform
sorry...but i don't know what is Target platform
### Build script
sorry...but i don't know what is Build script
### Error / output

### Visual Studio Version
_No response_
### GCC / Compiler Version
_No response_ | 1.0 | [Build] No matching distribution found for onnxruntime - ### Describe the issue
I am installing onnxruntime when facing errors

i am windows 11 with a python 3.11
### Urgency
_No response_
### Target platform
sorry...but i don't know what is Target platform
### Build script
sorry...but i don't know what is Build script
### Error / output

### Visual Studio Version
_No response_
### GCC / Compiler Version
_No response_ | non_test | no matching distribution found for onnxruntime describe the issue i am installing onnxruntime when facing errors i am windows with a python urgency no response target platform sorry but i don t know what is target platform build script sorry but i don t know what is build script error output visual studio version no response gcc compiler version no response | 0 |
125,430 | 26,657,639,533 | IssuesEvent | 2023-01-25 18:08:40 | pokt-network/pocket | https://api.github.com/repos/pokt-network/pocket | closed | [TECHDEBT] [P2P] Conntype as enum | p2p code health core starter task | ## Objective
There are TODOS where ConnType are booleans that should be updated to be enums so we can have more descriptive values where desired as they relate to P2P network actors.
```go
./p2p/transport.go: switch cfg.IsEmptyConnType() { // TECHDEBT kept in switch format because this should be an enum not a bool
./p2p/types/p2p_config.pb.go: IsEmptyConnectionType bool protobuf:"varint,4,opt,name=is_empty_connection_type,json=isEmptyConnectionType,proto3" json:"is_empty_connection_type,omitempty" // TODO (Drewsky) switch back to enum
./p2p/types/proto/p2p_config.proto: bool is_empty_connection_type = 4; // TODO (Drewsky) switch back to enum
! touches shared/types
```
## Origin Document
[TODO in the code](https://github.com/pokt-network/pocket/blob/main/p2p/transport/transport.go#L17)
## Goals
- Convert P2P booleans to enums
## Deliverable
- [ ] A PR that addresses the above TODOs
## Non-goals / Non-deliverables
- Defining additional enum values
- Updating logging logic
## General issue deliverables
- [ ] Update the appropriate CHANGELOG
- [ ] Update any relevant READMEs (local and/or global)
- [ ] Update any relevant global documentation & references
- [ ] If applicable, update the source code tree explanation
- [ ] If applicable, add or update a state, sequence or flowchart diagram using [mermaid](https://mermaid-js.github.io/mermaid/)
## [Optional] Testing Methodology
- \__REPLACE_ME_: Make sure to update the testing methodology appropriately\_
- **Task specific tests**: `make ...`
- **All tests**: `make test_all`
- **LocalNet**: verify a `LocalNet` is still functioning correctly by following the instructions at [docs/development/README.md](https://github.com/pokt-network/pocket/tree/main/docs/development)
---
**Creator**: @deblasis and @jessicadaugherty
**Co-Owners**: @Olshansk
| 1.0 | [TECHDEBT] [P2P] Conntype as enum - ## Objective
There are TODOS where ConnType are booleans that should be updated to be enums so we can have more descriptive values where desired as they relate to P2P network actors.
```go
./p2p/transport.go: switch cfg.IsEmptyConnType() { // TECHDEBT kept in switch format because this should be an enum not a bool
./p2p/types/p2p_config.pb.go: IsEmptyConnectionType bool protobuf:"varint,4,opt,name=is_empty_connection_type,json=isEmptyConnectionType,proto3" json:"is_empty_connection_type,omitempty" // TODO (Drewsky) switch back to enum
./p2p/types/proto/p2p_config.proto: bool is_empty_connection_type = 4; // TODO (Drewsky) switch back to enum
! touches shared/types
```
## Origin Document
[TODO in the code](https://github.com/pokt-network/pocket/blob/main/p2p/transport/transport.go#L17)
## Goals
- Convert P2P booleans to enums
## Deliverable
- [ ] A PR that addresses the above TODOs
## Non-goals / Non-deliverables
- Defining additional enum values
- Updating logging logic
## General issue deliverables
- [ ] Update the appropriate CHANGELOG
- [ ] Update any relevant READMEs (local and/or global)
- [ ] Update any relevant global documentation & references
- [ ] If applicable, update the source code tree explanation
- [ ] If applicable, add or update a state, sequence or flowchart diagram using [mermaid](https://mermaid-js.github.io/mermaid/)
## [Optional] Testing Methodology
- \__REPLACE_ME_: Make sure to update the testing methodology appropriately\_
- **Task specific tests**: `make ...`
- **All tests**: `make test_all`
- **LocalNet**: verify a `LocalNet` is still functioning correctly by following the instructions at [docs/development/README.md](https://github.com/pokt-network/pocket/tree/main/docs/development)
---
**Creator**: @deblasis and @jessicadaugherty
**Co-Owners**: @Olshansk
| non_test | conntype as enum objective there are todos where conntype are booleans that should be updated to be enums so we can have more descriptive values where desired as they relate to network actors go transport go switch cfg isemptyconntype techdebt kept in switch format because this should be an enum not a bool types config pb go isemptyconnectiontype bool protobuf varint opt name is empty connection type json isemptyconnectiontype json is empty connection type omitempty todo drewsky switch back to enum types proto config proto bool is empty connection type todo drewsky switch back to enum touches shared types origin document goals convert booleans to enums deliverable a pr that addresses the above todos non goals non deliverables defining additional enum values updating logging logic general issue deliverables update the appropriate changelog update any relevant readmes local and or global update any relevant global documentation references if applicable update the source code tree explanation if applicable add or update a state sequence or flowchart diagram using testing methodology replace me make sure to update the testing methodology appropriately task specific tests make all tests make test all localnet verify a localnet is still functioning correctly by following the instructions at creator deblasis and jessicadaugherty co owners olshansk | 0 |
294,171 | 25,350,778,428 | IssuesEvent | 2022-11-19 18:47:58 | GoogleContainerTools/skaffold | https://api.github.com/repos/GoogleContainerTools/skaffold | closed | Flake TestGracefulBuildCancel is flaky | meta/test-flake kind/todo | ### Expected behavior
Run `go test github.com/GoogleContainerTools/skaffold/pkg/skaffold/build/misc` can always give consistent results.
### Actual behavior
Run `make quicktest` multiple times, `github.com/GoogleContainerTools/skaffold/pkg/skaffold/build/misc` sometimes pass and sometimes fail.
### Information
```
=== Failed Tests ===
/pkg/skaffold/build/misc/TestGracefulBuildCancel
--- FAIL: TestGracefulBuildCancel (0.61s)
/pkg/skaffold/build/misc/TestGracefulBuildCancel/terminate_before_timeout
graceful_test.go:64: unexpected error: signal: interrupt
--- FAIL: TestGracefulBuildCancel/terminate_before_timeout (0.20s)
```
- Skaffold version: running against [master](https://github.com/GoogleContainerTools/skaffold/commit/823896bdd7464d79886664b32f56948957ea5f65) | 1.0 | Flake TestGracefulBuildCancel is flaky - ### Expected behavior
Run `go test github.com/GoogleContainerTools/skaffold/pkg/skaffold/build/misc` can always give consistent results.
### Actual behavior
Run `make quicktest` multiple times, `github.com/GoogleContainerTools/skaffold/pkg/skaffold/build/misc` sometimes pass and sometimes fail.
### Information
```
=== Failed Tests ===
/pkg/skaffold/build/misc/TestGracefulBuildCancel
--- FAIL: TestGracefulBuildCancel (0.61s)
/pkg/skaffold/build/misc/TestGracefulBuildCancel/terminate_before_timeout
graceful_test.go:64: unexpected error: signal: interrupt
--- FAIL: TestGracefulBuildCancel/terminate_before_timeout (0.20s)
```
- Skaffold version: running against [master](https://github.com/GoogleContainerTools/skaffold/commit/823896bdd7464d79886664b32f56948957ea5f65) | test | flake testgracefulbuildcancel is flaky expected behavior run go test github com googlecontainertools skaffold pkg skaffold build misc can always give consistent results actual behavior run make quicktest multiple times github com googlecontainertools skaffold pkg skaffold build misc sometimes pass and sometimes fail information failed tests pkg skaffold build misc testgracefulbuildcancel fail testgracefulbuildcancel pkg skaffold build misc testgracefulbuildcancel terminate before timeout graceful test go unexpected error signal interrupt fail testgracefulbuildcancel terminate before timeout skaffold version running against | 1 |
50,209 | 6,336,379,138 | IssuesEvent | 2017-07-26 20:57:54 | oSoc17/datascouts | https://api.github.com/repos/oSoc17/datascouts | closed | Restyle accounts checkboxes to eyes | design | Right now there are just regular checkboxes to be seen, while it should be the same as members. | 1.0 | Restyle accounts checkboxes to eyes - Right now there are just regular checkboxes to be seen, while it should be the same as members. | non_test | restyle accounts checkboxes to eyes right now there are just regular checkboxes to be seen while it should be the same as members | 0 |
354,488 | 10,568,149,828 | IssuesEvent | 2019-10-06 10:52:36 | cu-mkp/m-k-manuscript-data | https://api.github.com/repos/cu-mkp/m-k-manuscript-data | reopened | Decision: Normalisation of i/j and u/v? | consistency low-priority | Should the TC be completely diplomatic regarding i, u, v?
Currently the TC contains non-existend js, and u and v are normalised. We could easily change this back by replacing all js with is and all us with vs.
Question: do we want a completely naked diplomatic transcription, or should the TC have interventions such as normalised i/j and u/v?
Action: talk to other scholars of (digital) editions. Hunt down other DH scholars at RSA?
| 1.0 | Decision: Normalisation of i/j and u/v? - Should the TC be completely diplomatic regarding i, u, v?
Currently the TC contains non-existend js, and u and v are normalised. We could easily change this back by replacing all js with is and all us with vs.
Question: do we want a completely naked diplomatic transcription, or should the TC have interventions such as normalised i/j and u/v?
Action: talk to other scholars of (digital) editions. Hunt down other DH scholars at RSA?
| non_test | decision normalisation of i j and u v should the tc be completely diplomatic regarding i u v currently the tc contains non existend js and u and v are normalised we could easily change this back by replacing all js with is and all us with vs question do we want a completely naked diplomatic transcription or should the tc have interventions such as normalised i j and u v action talk to other scholars of digital editions hunt down other dh scholars at rsa | 0 |
153,489 | 5,893,026,435 | IssuesEvent | 2017-05-17 20:54:58 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio] Renaming a datasource causes the field to disappear from the form | bug Priority: Low | In Crafter 3.0 (probably all versions)
If you change the name for a datasource and then update the field to select the datasource the previous name is stored in the XML and when the form is rendered the field is missing.
```
<property>
<name>datasource</name>
<value>categories,categories_1</value>
<type>datasource:item</type>
</property>
``` | 1.0 | [studio] Renaming a datasource causes the field to disappear from the form - In Crafter 3.0 (probably all versions)
If you change the name for a datasource and then update the field to select the datasource the previous name is stored in the XML and when the form is rendered the field is missing.
```
<property>
<name>datasource</name>
<value>categories,categories_1</value>
<type>datasource:item</type>
</property>
``` | non_test | renaming a datasource causes the field to disappear from the form in crafter probably all versions if you change the name for a datasource and then update the field to select the datasource the previous name is stored in the xml and when the form is rendered the field is missing datasource categories categories datasource item | 0 |
395,122 | 27,055,748,625 | IssuesEvent | 2023-02-13 16:03:50 | oemof/oemof-solph | https://api.github.com/repos/oemof/oemof-solph | opened | Add an example for the OffsetTransformer with two series of coefficients | documentation | Just to show, how that can be used/implemented. | 1.0 | Add an example for the OffsetTransformer with two series of coefficients - Just to show, how that can be used/implemented. | non_test | add an example for the offsettransformer with two series of coefficients just to show how that can be used implemented | 0 |
785,769 | 27,624,569,069 | IssuesEvent | 2023-03-10 05:05:23 | ledd-23/crowdyy | https://api.github.com/repos/ledd-23/crowdyy | closed | [CRD-06] Create a bulk location query on backend based on user's location | enhancement server high priority | **What:** Implement a query for locations in the repository that are in the range of the user's current location using current region data on screen (latitude, longitude, latitudeDelta, longitudeDelta).
**Why:** So we can show the relevant location points to the user.
**AC:** A functioning API endpoint returning relevant data.
| 1.0 | [CRD-06] Create a bulk location query on backend based on user's location - **What:** Implement a query for locations in the repository that are in the range of the user's current location using current region data on screen (latitude, longitude, latitudeDelta, longitudeDelta).
**Why:** So we can show the relevant location points to the user.
**AC:** A functioning API endpoint returning relevant data.
| non_test | create a bulk location query on backend based on user s location what implement a query for locations in the repository that are in the range of the user s current location using current region data on screen latitude longitude latitudedelta longitudedelta why so we can show the relevant location points to the user ac a functioning api endpoint returning relevant data | 0 |
4,437 | 2,724,784,376 | IssuesEvent | 2015-04-14 19:46:13 | socia-platform/htwplus | https://api.github.com/repos/socia-platform/htwplus | closed | Menuitems with name AND icon | enhancement ready to test | (siehe Optionsmenü bspw. in Gruppenübersichtsseite)
Zu ändern sind:
* Navbar
* TabItems | 1.0 | Menuitems with name AND icon - (siehe Optionsmenü bspw. in Gruppenübersichtsseite)
Zu ändern sind:
* Navbar
* TabItems | test | menuitems with name and icon siehe optionsmenü bspw in gruppenübersichtsseite zu ändern sind navbar tabitems | 1 |
165,495 | 12,855,619,168 | IssuesEvent | 2020-07-09 05:46:45 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | CentroidCalculatorTests testPolygonAsPoint failure | :Analytics/Geo >test-failure Team:Analytics v8.0.0 | **Build scan**:
https://gradle-enterprise.elastic.co/s/pdb3fbh5avvl6/tests/:x-pack:plugin:spatial:test/org.elasticsearch.xpack.spatial.index.fielddata.CentroidCalculatorTests/testPolygonAsPoint#1
**Repro line**:
./gradlew ':x-pack:plugin:spatial:test' --tests "org.elasticsearch.xpack.spatial.index.fielddata.CentroidCalculatorTests.testPolygonAsPoint" -Dtests.seed=B380D8C378A207ED -Dtests.security.manager=true -Dtests.locale=sr-Latn-BA -Dtests.timezone=America/Bahia_Banderas -Druntime.java=11
**Reproduces locally?**:
Yes
**Applicable branches**:
8.0.0
**Failure history**:
Failed my PR build.
I also reproduced the failure on the head of master b5490f20d04a2815d3251ae1ad4a08b6d79a7cdd
**Failure excerpt**:
Expected: <4.6058851069340926E-229>
but: was <4.605885106934093E-229>
at __randomizedtesting.SeedInfo.seed([B380D8C378A207ED:BD448C9983ECDE9E]:0)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.xpack.spatial.index.fielddata.CentroidCalculatorTests.testPolygonAsPoint(CentroidCalculatorTests.java:259)
| 1.0 | CentroidCalculatorTests testPolygonAsPoint failure - **Build scan**:
https://gradle-enterprise.elastic.co/s/pdb3fbh5avvl6/tests/:x-pack:plugin:spatial:test/org.elasticsearch.xpack.spatial.index.fielddata.CentroidCalculatorTests/testPolygonAsPoint#1
**Repro line**:
./gradlew ':x-pack:plugin:spatial:test' --tests "org.elasticsearch.xpack.spatial.index.fielddata.CentroidCalculatorTests.testPolygonAsPoint" -Dtests.seed=B380D8C378A207ED -Dtests.security.manager=true -Dtests.locale=sr-Latn-BA -Dtests.timezone=America/Bahia_Banderas -Druntime.java=11
**Reproduces locally?**:
Yes
**Applicable branches**:
8.0.0
**Failure history**:
Failed my PR build.
I also reproduced the failure on the head of master b5490f20d04a2815d3251ae1ad4a08b6d79a7cdd
**Failure excerpt**:
Expected: <4.6058851069340926E-229>
but: was <4.605885106934093E-229>
at __randomizedtesting.SeedInfo.seed([B380D8C378A207ED:BD448C9983ECDE9E]:0)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.xpack.spatial.index.fielddata.CentroidCalculatorTests.testPolygonAsPoint(CentroidCalculatorTests.java:259)
| test | centroidcalculatortests testpolygonaspoint failure build scan repro line gradlew x pack plugin spatial test tests org elasticsearch xpack spatial index fielddata centroidcalculatortests testpolygonaspoint dtests seed dtests security manager true dtests locale sr latn ba dtests timezone america bahia banderas druntime java reproduces locally yes applicable branches failure history failed my pr build i also reproduced the failure on the head of master failure excerpt expected but was at randomizedtesting seedinfo seed at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at org elasticsearch xpack spatial index fielddata centroidcalculatortests testpolygonaspoint centroidcalculatortests java | 1 |
43,574 | 2,889,862,348 | IssuesEvent | 2015-06-13 20:44:27 | damonkohler/sl4a | https://api.github.com/repos/damonkohler/sl4a | opened | Use gestures to enter ()s and []s | auto-migrated Priority-Medium Type-Enhancement | _From @GoogleCodeExporter on May 31, 2015 11:25_
```
It would be pretty sweet if you could just draw an arc or bracket on the
screen to insert ()s or []s. Another gesture for = would be great too. This
would be helpful because the normal keyboard makes it hard to get to special
symbols that are common in scripts.
```
Original issue reported on code.google.com by `damonkoh...@gmail.com` on 9 Mar 2010 at 10:26
_Copied from original issue: damonkohler/android-scripting#219_ | 1.0 | Use gestures to enter ()s and []s - _From @GoogleCodeExporter on May 31, 2015 11:25_
```
It would be pretty sweet if you could just draw an arc or bracket on the
screen to insert ()s or []s. Another gesture for = would be great too. This
would be helpful because the normal keyboard makes it hard to get to special
symbols that are common in scripts.
```
Original issue reported on code.google.com by `damonkoh...@gmail.com` on 9 Mar 2010 at 10:26
_Copied from original issue: damonkohler/android-scripting#219_ | non_test | use gestures to enter s and s from googlecodeexporter on may it would be pretty sweet if you could just draw an arc or bracket on the screen to insert s or s another gesture for would be great too this would be helpful because the normal keyboard makes it hard to get to special symbols that are common in scripts original issue reported on code google com by damonkoh gmail com on mar at copied from original issue damonkohler android scripting | 0 |
398,925 | 27,217,091,350 | IssuesEvent | 2023-02-20 23:24:08 | nix-community/lanzaboote | https://api.github.com/repos/nix-community/lanzaboote | closed | Add step of adding lanzaboote flake input to quick start guide | documentation enhancement | In the quick start guide, we should explain that you have to add the lanzaboote flake input and how to do it. | 1.0 | Add step of adding lanzaboote flake input to quick start guide - In the quick start guide, we should explain that you have to add the lanzaboote flake input and how to do it. | non_test | add step of adding lanzaboote flake input to quick start guide in the quick start guide we should explain that you have to add the lanzaboote flake input and how to do it | 0 |
738,936 | 25,574,587,429 | IssuesEvent | 2022-11-30 20:49:48 | GoogleCloudPlatform/python-docs-samples | https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples | closed | asset.snippets.quickstart_getfeed_test: test_get_feed failed | priority: p1 type: bug flakybot: issue flakybot: flaky | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 07e8b8145670fc1f9f9ed99e348d8fe28ca2ca7e
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/e3c01052-a4f3-49b3-b93e-6123c271a1b0), [Sponge](http://sponge2/e3c01052-a4f3-49b3-b93e-6123c271a1b0)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable
return callable_(*args, **kwargs)
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 946, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking
raise _InactiveRpcError(state)
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.UNAUTHENTICATED
details = "Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project."
debug_error_string = "UNKNOWN:Error received from peer ipv4:74.125.197.95:443 {created_time:"2022-11-30T11:14:46.152776564+00:00", grpc_status:16, grpc_message:"Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project."}"
>
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/asset/snippets/conftest.py", line 72, in test_feed
feed = create_feed()
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/backoff/_sync.py", line 105, in retry
ret = target(*args, **kwargs)
File "/workspace/asset/snippets/conftest.py", line 69, in create_feed
PROJECT, feed_id, [asset_name], test_topic.name, asset_v1.ContentType.RESOURCE
File "/workspace/asset/snippets/quickstart_createfeed.py", line 38, in create_feed
request={"parent": parent, "feed_id": feed_id, "feed": feed}
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/cloud/asset_v1/services/asset_service/client.py", line 961, in create_feed
metadata=metadata,
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 154, in __call__
return wrapped_func(*args, **kwargs)
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable
raise exceptions.from_grpc_error(exc) from exc
google.api_core.exceptions.Unauthenticated: 401 Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.</pre></details> | 1.0 | asset.snippets.quickstart_getfeed_test: test_get_feed failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 07e8b8145670fc1f9f9ed99e348d8fe28ca2ca7e
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/e3c01052-a4f3-49b3-b93e-6123c271a1b0), [Sponge](http://sponge2/e3c01052-a4f3-49b3-b93e-6123c271a1b0)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable
return callable_(*args, **kwargs)
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 946, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking
raise _InactiveRpcError(state)
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.UNAUTHENTICATED
details = "Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project."
debug_error_string = "UNKNOWN:Error received from peer ipv4:74.125.197.95:443 {created_time:"2022-11-30T11:14:46.152776564+00:00", grpc_status:16, grpc_message:"Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project."}"
>
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/asset/snippets/conftest.py", line 72, in test_feed
feed = create_feed()
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/backoff/_sync.py", line 105, in retry
ret = target(*args, **kwargs)
File "/workspace/asset/snippets/conftest.py", line 69, in create_feed
PROJECT, feed_id, [asset_name], test_topic.name, asset_v1.ContentType.RESOURCE
File "/workspace/asset/snippets/quickstart_createfeed.py", line 38, in create_feed
request={"parent": parent, "feed_id": feed_id, "feed": feed}
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/cloud/asset_v1/services/asset_service/client.py", line 961, in create_feed
metadata=metadata,
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 154, in __call__
return wrapped_func(*args, **kwargs)
File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable
raise exceptions.from_grpc_error(exc) from exc
google.api_core.exceptions.Unauthenticated: 401 Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.</pre></details> | non_test | asset snippets quickstart getfeed test test get feed failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output traceback most recent call last file workspace asset snippets nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace asset snippets nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace asset snippets nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode unauthenticated details request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see debug error string unknown error received from peer created time grpc status grpc message request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see the above exception was the direct cause of the following exception traceback most recent call last file workspace asset snippets conftest py line in test feed feed create feed file workspace asset snippets nox py lib site packages backoff sync py line in retry ret target args kwargs file workspace asset snippets conftest py line in create feed project feed id test topic name asset contenttype resource file workspace asset snippets quickstart createfeed py line in create feed request parent parent feed id feed id feed feed file workspace asset snippets nox py lib site packages google cloud asset services asset service client py line in create feed metadata metadata file workspace asset snippets nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace asset snippets nox py lib site packages google api core grpc helpers py line in error remapped callable raise exceptions from grpc error exc from exc google api core exceptions unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see | 0 |
341,184 | 30,571,515,608 | IssuesEvent | 2023-07-20 22:52:23 | nightingaleproject/canary | https://api.github.com/repos/nightingaleproject/canary | closed | Alias message generation should be a supported message type for generation | enhancement ForTestingEvent | Currently, Canary supports message generation for submission, update, and void message types, but not for the alias message type. This should be added to the UI and backend. | 1.0 | Alias message generation should be a supported message type for generation - Currently, Canary supports message generation for submission, update, and void message types, but not for the alias message type. This should be added to the UI and backend. | test | alias message generation should be a supported message type for generation currently canary supports message generation for submission update and void message types but not for the alias message type this should be added to the ui and backend | 1 |
37,379 | 5,114,684,815 | IssuesEvent | 2017-01-06 19:18:36 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Test failure: System.Xml.Tests.XmlWriterTests_InvalidSurrogate/XmlWriterChecksUpperBoundOfLowerSurrogate_newBehavior | area-System.Xml test bug test-run-desktop | Opened on behalf of @jiangzeng
The test `System.Xml.Tests.XmlWriterTests_InvalidSurrogate/XmlWriterChecksUpperBoundOfLowerSurrogate_newBehavior` has failed.
Assert.Throws() Failure\r
Expected: typeof(System.ArgumentException)\r
Actual: (No exception was thrown)
Stack Trace:
at System.Xml.Tests.XmlWriterTests_InvalidSurrogate.XmlWriterChecksUpperBoundOfLowerSurrogate_newBehavior() in D:\A\_work\32\s\corefx\src\System.Private.Xml\tests\XmlWriter\WriteWithInvalidSurrogate.cs:line 52
Build : Master - 20161102.03 (Full Framework Tests)
Failing configurations:
- Windows.10.Amd64
- AnyCPU-Debug
- AnyCPU-Release
Details:
https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fdesktop~2Fcli~2F/build/20161102.03/workItem/System.Xml.RW.XmlWriter.Tests/analysis/xunit/System.Xml.Tests.XmlWriterTests_InvalidSurrogate~2FXmlWriterChecksUpperBoundOfLowerSurrogate_newBehavior | 2.0 | Test failure: System.Xml.Tests.XmlWriterTests_InvalidSurrogate/XmlWriterChecksUpperBoundOfLowerSurrogate_newBehavior - Opened on behalf of @jiangzeng
The test `System.Xml.Tests.XmlWriterTests_InvalidSurrogate/XmlWriterChecksUpperBoundOfLowerSurrogate_newBehavior` has failed.
Assert.Throws() Failure\r
Expected: typeof(System.ArgumentException)\r
Actual: (No exception was thrown)
Stack Trace:
at System.Xml.Tests.XmlWriterTests_InvalidSurrogate.XmlWriterChecksUpperBoundOfLowerSurrogate_newBehavior() in D:\A\_work\32\s\corefx\src\System.Private.Xml\tests\XmlWriter\WriteWithInvalidSurrogate.cs:line 52
Build : Master - 20161102.03 (Full Framework Tests)
Failing configurations:
- Windows.10.Amd64
- AnyCPU-Debug
- AnyCPU-Release
Details:
https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fdesktop~2Fcli~2F/build/20161102.03/workItem/System.Xml.RW.XmlWriter.Tests/analysis/xunit/System.Xml.Tests.XmlWriterTests_InvalidSurrogate~2FXmlWriterChecksUpperBoundOfLowerSurrogate_newBehavior | test | test failure system xml tests xmlwritertests invalidsurrogate xmlwriterchecksupperboundoflowersurrogate newbehavior opened on behalf of jiangzeng the test system xml tests xmlwritertests invalidsurrogate xmlwriterchecksupperboundoflowersurrogate newbehavior has failed assert throws failure r expected typeof system argumentexception r actual no exception was thrown stack trace at system xml tests xmlwritertests invalidsurrogate xmlwriterchecksupperboundoflowersurrogate newbehavior in d a work s corefx src system private xml tests xmlwriter writewithinvalidsurrogate cs line build master full framework tests failing configurations windows anycpu debug anycpu release details | 1 |
110,545 | 9,460,288,162 | IssuesEvent | 2019-04-17 10:33:56 | status-im/status-react | https://api.github.com/repos/status-im/status-react | opened | "Device2" and "Device1" links are set randomly in TestRail | tests | On TestRail links to Sauce Labs to `Device1` and `Device2` can be mixed up.
Expected that names are matched links.
Example of test result - https://ethstatus.testrail.net/index.php?/tests/view/583778

| 1.0 | "Device2" and "Device1" links are set randomly in TestRail - On TestRail links to Sauce Labs to `Device1` and `Device2` can be mixed up.
Expected that names are matched links.
Example of test result - https://ethstatus.testrail.net/index.php?/tests/view/583778

| test | and links are set randomly in testrail on testrail links to sauce labs to and can be mixed up expected that names are matched links example of test result | 1 |
140,235 | 31,861,979,903 | IssuesEvent | 2023-09-15 11:37:23 | kamilsk/dotfiles | https://api.github.com/repos/kamilsk/dotfiles | closed | command: add books as the same as obsidian | type: feature scope: code impact: medium effort: easy | **Motivation:** review disk usage, now I have 27G audiobooks.
```bash
$ books
# cd ~/Library/Containers/com.apple.BKAgentService/Data/Documents/iBooks
# du -h * | sort -h
``` | 1.0 | command: add books as the same as obsidian - **Motivation:** review disk usage, now I have 27G audiobooks.
```bash
$ books
# cd ~/Library/Containers/com.apple.BKAgentService/Data/Documents/iBooks
# du -h * | sort -h
``` | non_test | command add books as the same as obsidian motivation review disk usage now i have audiobooks bash books cd library containers com apple bkagentservice data documents ibooks du h sort h | 0 |
70,372 | 7,186,894,721 | IssuesEvent | 2018-02-02 01:42:13 | WormBase/website | https://api.github.com/repos/WormBase/website | closed | Interactions Cytoscape graph defaults to Physical interactions even if there are none | Under testing Webteam feature request | @sibyl229
First, the new interactions Cytoscape display is great!
Second: we had discussed a way of pruning the graph by default to only include physical interactions when there were possibly too many to display at once, but I found an instance of a gene that doesn't have any physical interactions, itr-1:
http://www.wormbase.org/species/c_elegans/gene/WBGene00002173#08--10
and what happens is that when the interaction network graph is expanded, nothing is rendered because the interaction type has been restricted to physical interaction types, even though there are none for this gene. Could we do a little calculation to only show physical interactions if there are any, and then if not default to regulatory interactions, and if not default to genetic interactions, and if not default to predicted interactions? Or something like that? | 1.0 | Interactions Cytoscape graph defaults to Physical interactions even if there are none - @sibyl229
First, the new interactions Cytoscape display is great!
Second: we had discussed a way of pruning the graph by default to only include physical interactions when there were possibly too many to display at once, but I found an instance of a gene that doesn't have any physical interactions, itr-1:
http://www.wormbase.org/species/c_elegans/gene/WBGene00002173#08--10
and what happens is that when the interaction network graph is expanded, nothing is rendered because the interaction type has been restricted to physical interaction types, even though there are none for this gene. Could we do a little calculation to only show physical interactions if there are any, and then if not default to regulatory interactions, and if not default to genetic interactions, and if not default to predicted interactions? Or something like that? | test | interactions cytoscape graph defaults to physical interactions even if there are none first the new interactions cytoscape display is great second we had discussed a way of pruning the graph by default to only include physical interactions when there were possibly too many to display at once but i found an instance of a gene that doesn t have any physical interactions itr and what happens is that when the interaction network graph is expanded nothing is rendered because the interaction type has been restricted to physical interaction types even though there are none for this gene could we do a little calculation to only show physical interactions if there are any and then if not default to regulatory interactions and if not default to genetic interactions and if not default to predicted interactions or something like that | 1 |
475,766 | 13,725,525,845 | IssuesEvent | 2020-10-03 18:53:44 | NCIEVS/nci-protege5 | https://api.github.com/repos/NCIEVS/nci-protege5 | closed | lucene search fails with bad listener | bug low priority needs investigation ready for QA | From [TQ, 6/4/18].
"I have done a simple search for full syn contains Terminology Subset and it just keeps searching without returning results."
"There is a bad listener in here…. Is that relevant?"
Snippet of console output with the listener exception is below:
Loaded ontology, now updating from server
[34mClientSessionChangeEvent: org.protege.editor.owl.client.event.ClientSessionChangeEvent@7c3de2c7[0;39m
[34mInstantiating node LuceneIndexPreferences[0;39m
[34mInstantiating node KEY:ecd96c9c-1faa-40b6-a69f-912a54d1d3b1[0;39m
**[31mException thrown by listener: edu.stanford.protege.search.lucene.tab.engine.SearchTabManager$$Lambda$45/1208043143. Detatching bad listener.[0;39m**
[34mClientSessionChangeEvent: org.protege.editor.owl.client.event.ClientSessionChangeEvent@1a3bcb81[0;39m
Voila
| 1.0 | lucene search fails with bad listener - From [TQ, 6/4/18].
"I have done a simple search for full syn contains Terminology Subset and it just keeps searching without returning results."
"There is a bad listener in here…. Is that relevant?"
Snippet of console output with the listener exception is below:
Loaded ontology, now updating from server
[34mClientSessionChangeEvent: org.protege.editor.owl.client.event.ClientSessionChangeEvent@7c3de2c7[0;39m
[34mInstantiating node LuceneIndexPreferences[0;39m
[34mInstantiating node KEY:ecd96c9c-1faa-40b6-a69f-912a54d1d3b1[0;39m
**[31mException thrown by listener: edu.stanford.protege.search.lucene.tab.engine.SearchTabManager$$Lambda$45/1208043143. Detatching bad listener.[0;39m**
[34mClientSessionChangeEvent: org.protege.editor.owl.client.event.ClientSessionChangeEvent@1a3bcb81[0;39m
Voila
| non_test | lucene search fails with bad listener from i have done a simple search for full syn contains terminology subset and it just keeps searching without returning results there is a bad listener in here… is that relevant snippet of console output with the listener exception is below loaded ontology now updating from server org protege editor owl client event clientsessionchangeevent node luceneindexpreferences node key thrown by listener edu stanford protege search lucene tab engine searchtabmanager lambda detatching bad listener org protege editor owl client event clientsessionchangeevent voila | 0 |
287,620 | 24,846,668,200 | IssuesEvent | 2022-10-26 16:23:56 | jdi-testing/jdi-light | https://api.github.com/repos/jdi-testing/jdi-light | closed | Update test-site: element "alerts" | TestSite Vuetify | Add alerts with following properties:
- [x] rounded
- [x] tile
- [x] height/width | 1.0 | Update test-site: element "alerts" - Add alerts with following properties:
- [x] rounded
- [x] tile
- [x] height/width | test | update test site element alerts add alerts with following properties rounded tile height width | 1 |
299,999 | 25,941,645,442 | IssuesEvent | 2022-12-16 19:02:48 | EmotionCognitionLab/pvs | https://api.github.com/repos/EmotionCognitionLab/pvs | closed | Lumosity message | enhancement needs testing | The message at the top ("After completing the Lumosity segment, please click here") may be confusing because participants may not know that they are seeing Lumosity below. It resembles a pop up ad that I usually quickly close to maximize view of the screen. I'm not sure if we should change the color or add an icon to distinguish it. Or maybe change the text to something like: "Please begin your Lumosity game session below. When you complete all of the games, please click here." | 1.0 | Lumosity message - The message at the top ("After completing the Lumosity segment, please click here") may be confusing because participants may not know that they are seeing Lumosity below. It resembles a pop up ad that I usually quickly close to maximize view of the screen. I'm not sure if we should change the color or add an icon to distinguish it. Or maybe change the text to something like: "Please begin your Lumosity game session below. When you complete all of the games, please click here." | test | lumosity message the message at the top after completing the lumosity segment please click here may be confusing because participants may not know that they are seeing lumosity below it resembles a pop up ad that i usually quickly close to maximize view of the screen i m not sure if we should change the color or add an icon to distinguish it or maybe change the text to something like please begin your lumosity game session below when you complete all of the games please click here | 1 |
59,977 | 24,943,873,304 | IssuesEvent | 2022-10-31 21:29:55 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | Issue with Search Functionality On the Flashing Beacon Page | Workgroup: AMD Type: Bug Report Impact: 3-Minor Service: Apps Product: AMD Data Tracker | <!-- Email -->
<!-- Brian.vandewalle@austintexas.gov -->
> What application are you using?
Data Tracker
> Describe the problem.
On the Flashing Beacon page, I try to search for the term "RRFB" and get zero results. However, if I manually sort them I find that there are 17 entries in the database where the Flasher Type is "RRFB".
> Website Address
https://atd.knack.com/amd#flashing-beacons/?view_1454_sort=field_1697|asc&view_1454_page=1
Internet Browser: Chrome
> Requested By
Brian V.
Request ID: DTS22-105355
| 1.0 | Issue with Search Functionality On the Flashing Beacon Page - <!-- Email -->
<!-- Brian.vandewalle@austintexas.gov -->
> What application are you using?
Data Tracker
> Describe the problem.
On the Flashing Beacon page, I try to search for the term "RRFB" and get zero results. However, if I manually sort them I find that there are 17 entries in the database where the Flasher Type is "RRFB".
> Website Address
https://atd.knack.com/amd#flashing-beacons/?view_1454_sort=field_1697|asc&view_1454_page=1
Internet Browser: Chrome
> Requested By
Brian V.
Request ID: DTS22-105355
| non_test | issue with search functionality on the flashing beacon page what application are you using data tracker describe the problem on the flashing beacon page i try to search for the term rrfb and get zero results however if i manually sort them i find that there are entries in the database where the flasher type is rrfb website address internet browser chrome requested by brian v request id | 0 |
342,495 | 30,624,382,206 | IssuesEvent | 2023-07-24 10:25:02 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | c2c/admission: hit `observed raft log position less than per-stream lower bound` panic during race run of TestTenantStreamingMultipleNodes | C-bug C-test-failure A-admission-control T-kv | UPDATE: I have repro'd the failure on a cleaner branch [here](https://github.com/cockroachdb/cockroach/pull/106675). See newer CI failure [here](https://teamcity.cockroachdb.com/viewLog.html?buildId=10883373&buildTypeId=Cockroach_BazelExtendedCi).
ORIGINAL MESSAGE: See the CI failure [here](https://teamcity.cockroachdb.com/viewLog.html?buildId=10857713&buildTypeId=Cockroach_BazelExtendedCi) in a [draft PR ](https://github.com/cockroachdb/cockroach/pull/106560) that attempts to unskip `TestTenantStreamingMultiNodes` under stress race. Full test log for this run [here](https://drive.google.com/drive/u/0/folders/1QUwKan142_cT0aea_4t2S8ttmu6-6A9M). I have seen this stack trace on master Extended CI, but have yet to reproduce this locally on my gce worker.
Full stack trace:
```
[TestTenantStreamingMultipleNodes] === RUN TestTenantStreamingMultipleNodes
test_log_scope.go:167: test logs captured to: /artifacts/tmp/_tmp/90c1b75d835f45b8488807abb5b1092d/logTestTenantStreamingMultipleNodes1823897493
test_log_scope.go:81: use -show-logs to present logs inline
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 observed raft log position less than per-stream lower bound (log-position=0/0 <= log-position=9/6600)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !goroutine 255859 [running]:
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !runtime/debug.Stack()
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! GOROOT/src/runtime/debug/stack.go:24 +0x72
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/util/log.(*loggerT).outputLogEntry(0xc010fa16e0, {{{0xc010f040c0, 0x24}, {0x94dc09b, 0x1}, {0x94dc09a, 0x1}, {0x94dc09b, 0x1}}, 0x1770ab40ea3045ae, ...})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/util/log/clog.go:261 +0xb7
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/util/log.logfDepthInternal({0xd499400, 0xc011b828a0}, 0x2, 0x4, 0x0, 0x0, {0x9603daa, 0x46}, {0xc039c01720, 0x2, ...})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/util/log/channels.go:106 +0x6e6
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/util/log.logfDepth(...)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/util/log/channels.go:39
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/util/log.Fatalf({0xd499400, 0xc011b828a0}, {0x9603daa, 0x46}, {0xc039c01720, 0x2, 0x2})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/bazel-out/k8-fastbuild/bin/pkg/util/log/log_channels_generated.go:848 +0x9d
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver/kvflowcontrol/kvflowtokentracker.(*Tracker).Track(0xc04b7bade0, {0xd499400, 0xc011b828a0}, 0x0, 0xbd, {0xc0154167a0?, 0x4?})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 +0x85b
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver/kvflowcontrol/kvflowhandle.(*Handle).deductTokensForInner(0xc02ddcfc00, {0xd499400, 0xc011b828a0}, 0x0?, {0x1350ce80?, 0xc0fb967b30?}, 0xbd)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/kvflowcontrol/kvflowhandle/kvflowhandle.go:162 +0x394
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver/kvflowcontrol/kvflowhandle.(*Handle).DeductTokensFor(0xc02ddcfc00, {0xd499400, 0xc011b828a0}, 0x7d?, {0x2c?, 0xc015416880?}, 0x1?)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/kvflowcontrol/kvflowhandle/kvflowhandle.go:141 +0x72
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.maybeDeductFlowTokens({0xd499400, 0xc0f155cf60}, {0xd4df900, 0xc02ddcfc00}, {0xc05ce6e2d0, 0x3, 0x3}, {0xc037f647e0, 0x3, 0x3})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_proposal_buf.go:1051 +0x2e5
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.proposeBatch({0xd499400, 0xc0f155cf60}, {0x7fcccd588ca0, 0xc0efe8a880}, {0xd4b9258, 0xc11e1c02d0}, {0xc037f647e0, 0x3, 0x3}, {0xc05ce6e2d0, ...}, ...)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_proposal_buf.go:1028 +0x24f
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*propBuf).FlushLockedWithRaftGroup(0xc0efe8b0d0, {0xd499400, 0xc0f155cf60}, {0xd4b9258, 0xc11e1c02d0})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_proposal_buf.go:621 +0x1c97
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*Replica).handleRaftReadyRaftMuLocked.func2(0x0?)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:804 +0xfd
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*Replica).withRaftGroupLocked.func1(...)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:2042
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*Replica).withRaftGroupLocked(0xc0efe8a880, 0x1, 0xc015418760)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:2043 +0x3ae
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*Replica).handleRaftReadyRaftMuLocked(_, {_, _}, {{0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...}, ...})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:801 +0x3a5
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*Replica).handleRaftReady(_, {_, _}, {{0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...}, ...})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:757 +0x210
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*Store).processReady(0xc00396aa80, 0xd499400?)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/store_raft.go:660 +0x1d0
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*raftSchedulerShard).worker(0xc007028e70, {0xd499400, 0xc006dd95f0}, {0xd47cc40, 0xc00396aa80}, 0xc004303500)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/scheduler.go:418 +0x2f6
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*raftScheduler).Start.func2({0xd499400, 0xc006dd95f0})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/scheduler.go:321 +0x9b
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx.func2()
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:484 +0x1f7
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:475 +0x61a
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !For more context, check log files in: /artifacts/tmp/_tmp/90c1b75d835f45b8488807abb5b1092d/logTestTenantStreamingMultipleNodes1823897493
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !****************************************************************************
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !This node experienced a fatal error (printed above), and as a result the
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !process is terminating.
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !Fatal errors can occur due to faulty hardware (disks, memory, clocks) or a
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !problem in CockroachDB. With your help, the support team at Cockroach Labs
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !will try to determine the root cause, recommend next steps, and we can
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !improve CockroachDB based on your report.
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !Please submit a crash report by following the instructions here:
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! https://github.com/cockroachdb/cockroach/issues/new/choose
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !If you would rather not post publicly, please contact us directly at:
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! support@cockroachlabs.com
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !The Cockroach Labs team appreciates your feedback.
```
Jira issue: CRDB-29634 | 1.0 | c2c/admission: hit `observed raft log position less than per-stream lower bound` panic during race run of TestTenantStreamingMultipleNodes - UPDATE: I have repro'd the failure on a cleaner branch [here](https://github.com/cockroachdb/cockroach/pull/106675). See newer CI failure [here](https://teamcity.cockroachdb.com/viewLog.html?buildId=10883373&buildTypeId=Cockroach_BazelExtendedCi).
ORIGINAL MESSAGE: See the CI failure [here](https://teamcity.cockroachdb.com/viewLog.html?buildId=10857713&buildTypeId=Cockroach_BazelExtendedCi) in a [draft PR ](https://github.com/cockroachdb/cockroach/pull/106560) that attempts to unskip `TestTenantStreamingMultiNodes` under stress race. Full test log for this run [here](https://drive.google.com/drive/u/0/folders/1QUwKan142_cT0aea_4t2S8ttmu6-6A9M). I have seen this stack trace on master Extended CI, but have yet to reproduce this locally on my gce worker.
Full stack trace:
```
[TestTenantStreamingMultipleNodes] === RUN TestTenantStreamingMultipleNodes
test_log_scope.go:167: test logs captured to: /artifacts/tmp/_tmp/90c1b75d835f45b8488807abb5b1092d/logTestTenantStreamingMultipleNodes1823897493
test_log_scope.go:81: use -show-logs to present logs inline
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 observed raft log position less than per-stream lower bound (log-position=0/0 <= log-position=9/6600)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !goroutine 255859 [running]:
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !runtime/debug.Stack()
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! GOROOT/src/runtime/debug/stack.go:24 +0x72
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/util/log.(*loggerT).outputLogEntry(0xc010fa16e0, {{{0xc010f040c0, 0x24}, {0x94dc09b, 0x1}, {0x94dc09a, 0x1}, {0x94dc09b, 0x1}}, 0x1770ab40ea3045ae, ...})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/util/log/clog.go:261 +0xb7
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/util/log.logfDepthInternal({0xd499400, 0xc011b828a0}, 0x2, 0x4, 0x0, 0x0, {0x9603daa, 0x46}, {0xc039c01720, 0x2, ...})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/util/log/channels.go:106 +0x6e6
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/util/log.logfDepth(...)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/util/log/channels.go:39
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/util/log.Fatalf({0xd499400, 0xc011b828a0}, {0x9603daa, 0x46}, {0xc039c01720, 0x2, 0x2})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/bazel-out/k8-fastbuild/bin/pkg/util/log/log_channels_generated.go:848 +0x9d
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver/kvflowcontrol/kvflowtokentracker.(*Tracker).Track(0xc04b7bade0, {0xd499400, 0xc011b828a0}, 0x0, 0xbd, {0xc0154167a0?, 0x4?})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 +0x85b
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver/kvflowcontrol/kvflowhandle.(*Handle).deductTokensForInner(0xc02ddcfc00, {0xd499400, 0xc011b828a0}, 0x0?, {0x1350ce80?, 0xc0fb967b30?}, 0xbd)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/kvflowcontrol/kvflowhandle/kvflowhandle.go:162 +0x394
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver/kvflowcontrol/kvflowhandle.(*Handle).DeductTokensFor(0xc02ddcfc00, {0xd499400, 0xc011b828a0}, 0x7d?, {0x2c?, 0xc015416880?}, 0x1?)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/kvflowcontrol/kvflowhandle/kvflowhandle.go:141 +0x72
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.maybeDeductFlowTokens({0xd499400, 0xc0f155cf60}, {0xd4df900, 0xc02ddcfc00}, {0xc05ce6e2d0, 0x3, 0x3}, {0xc037f647e0, 0x3, 0x3})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_proposal_buf.go:1051 +0x2e5
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.proposeBatch({0xd499400, 0xc0f155cf60}, {0x7fcccd588ca0, 0xc0efe8a880}, {0xd4b9258, 0xc11e1c02d0}, {0xc037f647e0, 0x3, 0x3}, {0xc05ce6e2d0, ...}, ...)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_proposal_buf.go:1028 +0x24f
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*propBuf).FlushLockedWithRaftGroup(0xc0efe8b0d0, {0xd499400, 0xc0f155cf60}, {0xd4b9258, 0xc11e1c02d0})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_proposal_buf.go:621 +0x1c97
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*Replica).handleRaftReadyRaftMuLocked.func2(0x0?)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:804 +0xfd
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*Replica).withRaftGroupLocked.func1(...)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:2042
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*Replica).withRaftGroupLocked(0xc0efe8a880, 0x1, 0xc015418760)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:2043 +0x3ae
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*Replica).handleRaftReadyRaftMuLocked(_, {_, _}, {{0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...}, ...})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:801 +0x3a5
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*Replica).handleRaftReady(_, {_, _}, {{0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...}, ...})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/replica_raft.go:757 +0x210
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*Store).processReady(0xc00396aa80, 0xd499400?)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/store_raft.go:660 +0x1d0
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*raftSchedulerShard).worker(0xc007028e70, {0xd499400, 0xc006dd95f0}, {0xd47cc40, 0xc00396aa80}, 0xc004303500)
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/scheduler.go:418 +0x2f6
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/kv/kvserver.(*raftScheduler).Start.func2({0xd499400, 0xc006dd95f0})
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/kv/kvserver/pkg/kv/kvserver/scheduler.go:321 +0x9b
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx.func2()
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:484 +0x1f7
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:475 +0x61a
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !For more context, check log files in: /artifacts/tmp/_tmp/90c1b75d835f45b8488807abb5b1092d/logTestTenantStreamingMultipleNodes1823897493
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !****************************************************************************
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !This node experienced a fatal error (printed above), and as a result the
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !process is terminating.
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !Fatal errors can occur due to faulty hardware (disks, memory, clocks) or a
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !problem in CockroachDB. With your help, the support team at Cockroach Labs
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !will try to determine the root cause, recommend next steps, and we can
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !improve CockroachDB based on your report.
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !Please submit a crash report by following the instructions here:
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! https://github.com/cockroachdb/cockroach/issues/new/choose
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !If you would rather not post publicly, please contact us directly at:
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 ! support@cockroachlabs.com
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !
F230711 01:15:55.559224 255859 kv/kvserver/kvflowcontrol/kvflowtokentracker/tracker.go:95 [T1,n2,tenant=10,s2,r76/5:/Tenant/10/Table/1{1/2…-5/4…}] 1 !The Cockroach Labs team appreciates your feedback.
```
Jira issue: CRDB-29634 | test | admission hit observed raft log position less than per stream lower bound panic during race run of testtenantstreamingmultiplenodes update i have repro d the failure on a cleaner branch see newer ci failure original message see the ci failure in a that attempts to unskip testtenantstreamingmultinodes under stress race full test log for this run i have seen this stack trace on master extended ci but have yet to reproduce this locally on my gce worker full stack trace run testtenantstreamingmultiplenodes test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline kv kvserver kvflowcontrol kvflowtokentracker tracker go observed raft log position less than per stream lower bound log position log position kv kvserver kvflowcontrol kvflowtokentracker tracker go goroutine kv kvserver kvflowcontrol kvflowtokentracker tracker go runtime debug stack kv kvserver kvflowcontrol kvflowtokentracker tracker go goroot src runtime debug stack go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg util log loggert outputlogentry kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg util log clog go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg util log logfdepthinternal kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg util log channels go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg util log logfdepth kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg util log channels go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg util log fatalf kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach bazel out fastbuild bin pkg util log log channels generated go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver kvflowcontrol kvflowtokentracker tracker track kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver kvflowcontrol kvflowhandle handle deducttokensforinner kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver kvflowcontrol kvflowhandle kvflowhandle go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver kvflowcontrol kvflowhandle handle deducttokensfor kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver kvflowcontrol kvflowhandle kvflowhandle go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver maybedeductflowtokens kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica proposal buf go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver proposebatch kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica proposal buf go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver propbuf flushlockedwithraftgroup kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica proposal buf go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver replica handleraftreadyraftmulocked kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica raft go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver replica withraftgrouplocked kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica raft go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver replica withraftgrouplocked kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica raft go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver replica handleraftreadyraftmulocked kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica raft go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver replica handleraftready kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver replica raft go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver store processready kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver store raft go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver raftschedulershard worker kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver scheduler go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver raftscheduler start kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg kv kvserver pkg kv kvserver scheduler go kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg util stop stopper runasynctaskex kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg util stop stopper go kv kvserver kvflowcontrol kvflowtokentracker tracker go created by github com cockroachdb cockroach pkg util stop stopper runasynctaskex kv kvserver kvflowcontrol kvflowtokentracker tracker go github com cockroachdb cockroach pkg util stop stopper go kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go for more context check log files in artifacts tmp tmp kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go this node experienced a fatal error printed above and as a result the kv kvserver kvflowcontrol kvflowtokentracker tracker go process is terminating kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go fatal errors can occur due to faulty hardware disks memory clocks or a kv kvserver kvflowcontrol kvflowtokentracker tracker go problem in cockroachdb with your help the support team at cockroach labs kv kvserver kvflowcontrol kvflowtokentracker tracker go will try to determine the root cause recommend next steps and we can kv kvserver kvflowcontrol kvflowtokentracker tracker go improve cockroachdb based on your report kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go please submit a crash report by following the instructions here kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go if you would rather not post publicly please contact us directly at kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go support cockroachlabs com kv kvserver kvflowcontrol kvflowtokentracker tracker go kv kvserver kvflowcontrol kvflowtokentracker tracker go the cockroach labs team appreciates your feedback jira issue crdb | 1 |
66,784 | 7,018,317,607 | IssuesEvent | 2017-12-21 13:16:02 | mautic/mautic | https://api.github.com/repos/mautic/mautic | closed | Users with restricted permissions cannot create landing pages | Bug Ready To Test | What type of report is this:
| Q | A
| ---| ---
| Bug report? | x
| Feature request? |
| Enhancement? |
## Description:
I have an instance set up with a demo user, who has the following permissions on the landing page bundle:

Note specifically the create permission.
When the user logs in, they are not able to create a landing page - the blue flash across the top of the screen indicates something starts, but nothing completes - video explanation here https://youtu.be/AFrowWrIUoA.
As an administrator, I can complete the action as expected.
## If a bug:
| Q | A
| --- | ---
| Mautic version | 2.11.0
| PHP version | 7.0.15
### Steps to reproduce:
1. Create a role with restricted permissions on landing pages
2. Log in and try to create the landing page
### Log errors:
No logs reported in error log
| 1.0 | Users with restricted permissions cannot create landing pages - What type of report is this:
| Q | A
| ---| ---
| Bug report? | x
| Feature request? |
| Enhancement? |
## Description:
I have an instance set up with a demo user, who has the following permissions on the landing page bundle:

Note specifically the create permission.
When the user logs in, they are not able to create a landing page - the blue flash across the top of the screen indicates something starts, but nothing completes - video explanation here https://youtu.be/AFrowWrIUoA.
As an administrator, I can complete the action as expected.
## If a bug:
| Q | A
| --- | ---
| Mautic version | 2.11.0
| PHP version | 7.0.15
### Steps to reproduce:
1. Create a role with restricted permissions on landing pages
2. Log in and try to create the landing page
### Log errors:
No logs reported in error log
| test | users with restricted permissions cannot create landing pages what type of report is this q a bug report x feature request enhancement description i have an instance set up with a demo user who has the following permissions on the landing page bundle note specifically the create permission when the user logs in they are not able to create a landing page the blue flash across the top of the screen indicates something starts but nothing completes video explanation here as an administrator i can complete the action as expected if a bug q a mautic version php version steps to reproduce create a role with restricted permissions on landing pages log in and try to create the landing page log errors no logs reported in error log | 1 |
187,221 | 14,427,236,257 | IssuesEvent | 2020-12-06 02:45:40 | kalexmills/github-vet-tests-dec2020 | https://api.github.com/repos/kalexmills/github-vet-tests-dec2020 | closed | fraibacas/zing-injector: vendor/cloud.google.com/go/spanner/session_test.go; 121 LoC | fresh large test vendored |
Found a possible issue in [fraibacas/zing-injector](https://www.github.com/fraibacas/zing-injector) at [vendor/cloud.google.com/go/spanner/session_test.go](https://github.com/fraibacas/zing-injector/blob/d8119599ee022f33d4f3faf29c90de0d81346c15/vendor/cloud.google.com/go/spanner/session_test.go#L659-L779)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable ti used in defer or goroutine at line 701
[Click here to see the code in its original context.](https://github.com/fraibacas/zing-injector/blob/d8119599ee022f33d4f3faf29c90de0d81346c15/vendor/cloud.google.com/go/spanner/session_test.go#L659-L779)
<details>
<summary>Click here to show the 121 line(s) of Go which triggered the analyzer.</summary>
```go
for ti, cfg := range []SessionPoolConfig{
SessionPoolConfig{},
SessionPoolConfig{MinOpened: 10, MaxOpened: 100},
SessionPoolConfig{MaxBurst: 50},
SessionPoolConfig{MinOpened: 10, MaxOpened: 200, MaxBurst: 5},
SessionPoolConfig{MinOpened: 10, MaxOpened: 200, MaxBurst: 5, WriteSessions: 0.2},
} {
var wg sync.WaitGroup
// Create a more aggressive session healthchecker to increase test concurrency.
cfg.HealthCheckInterval = 50 * time.Millisecond
cfg.healthCheckSampleInterval = 10 * time.Millisecond
cfg.HealthCheckWorkers = 50
sc := testutil.NewMockCloudSpannerClient(t)
sc.MakeNice()
cfg.getRPCClient = func() (sppb.SpannerClient, error) {
return sc, nil
}
sp, _ := newSessionPool("mockdb", cfg, nil)
for i := 0; i < 100; i++ {
wg.Add(1)
// Schedule a test worker.
go func(idx int, pool *sessionPool, client sppb.SpannerClient) {
defer wg.Done()
// Test worker iterates 1K times and tries different session / session pool operations.
for j := 0; j < 1000; j++ {
if idx%10 == 0 && j >= 900 {
// Close the pool in selected set of workers during the middle of the test.
pool.close()
}
// Take a write sessions ~ 20% of the times.
takeWrite := rand.Intn(5) == 4
var (
sh *sessionHandle
gotErr error
)
if takeWrite {
sh, gotErr = pool.takeWriteSession(context.Background())
} else {
sh, gotErr = pool.take(context.Background())
}
if gotErr != nil {
if pool.isValid() {
t.Errorf("%v.%v: pool.take returns error when pool is still valid: %v", ti, idx, gotErr)
}
if wantErr := errInvalidSessionPool(); !testEqual(gotErr, wantErr) {
t.Errorf("%v.%v: got error when pool is closed: %v, want %v", ti, idx, gotErr, wantErr)
}
continue
}
// Verify if session is valid when session pool is valid. Note that if session pool is invalid after sh is taken,
// then sh might be invalidated by healthcheck workers.
if (sh.getID() == "" || sh.session == nil || !sh.session.isValid()) && pool.isValid() {
t.Errorf("%v.%v.%v: pool.take returns invalid session %v", ti, idx, takeWrite, sh.session)
}
if takeWrite && sh.getTransactionID() == nil {
t.Errorf("%v.%v: pool.takeWriteSession returns session %v without transaction", ti, idx, sh.session)
}
if rand.Intn(100) < idx {
// Random sleep before destroying/recycling the session, to give healthcheck worker a chance to step in.
<-time.After(time.Duration(rand.Int63n(int64(cfg.HealthCheckInterval))))
}
if rand.Intn(100) < idx {
// destroy the session.
sh.destroy()
continue
}
// recycle the session.
sh.recycle()
}
}(i, sp, sc)
}
wg.Wait()
sp.hc.close()
// Here the states of healthchecker, session pool and mockclient are stable.
idleSessions := map[string]bool{}
hcSessions := map[string]bool{}
mockSessions := sc.DumpSessions()
// Dump session pool's idle list.
for sl := sp.idleList.Front(); sl != nil; sl = sl.Next() {
s := sl.Value.(*session)
if idleSessions[s.getID()] {
t.Errorf("%v: found duplicated session in idle list: %v", ti, s.getID())
}
idleSessions[s.getID()] = true
}
for sl := sp.idleWriteList.Front(); sl != nil; sl = sl.Next() {
s := sl.Value.(*session)
if idleSessions[s.getID()] {
t.Errorf("%v: found duplicated session in idle write list: %v", ti, s.getID())
}
idleSessions[s.getID()] = true
}
sp.mu.Lock()
if int(sp.numOpened) != len(idleSessions) {
t.Errorf("%v: number of opened sessions (%v) != number of idle sessions (%v)", ti, sp.numOpened, len(idleSessions))
}
if sp.createReqs != 0 {
t.Errorf("%v: number of pending session creations = %v, want 0", ti, sp.createReqs)
}
// Dump healthcheck queue.
for _, s := range sp.hc.queue.sessions {
if hcSessions[s.getID()] {
t.Errorf("%v: found duplicated session in healthcheck queue: %v", ti, s.getID())
}
hcSessions[s.getID()] = true
}
sp.mu.Unlock()
// Verify that idleSessions == hcSessions == mockSessions.
if !testEqual(idleSessions, hcSessions) {
t.Errorf("%v: sessions in idle list (%v) != sessions in healthcheck queue (%v)", ti, idleSessions, hcSessions)
}
if !testEqual(hcSessions, mockSessions) {
t.Errorf("%v: sessions in healthcheck queue (%v) != sessions in mockclient (%v)", ti, hcSessions, mockSessions)
}
sp.close()
mockSessions = sc.DumpSessions()
if len(mockSessions) != 0 {
t.Errorf("Found live sessions: %v", mockSessions)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: d8119599ee022f33d4f3faf29c90de0d81346c15
| 1.0 | fraibacas/zing-injector: vendor/cloud.google.com/go/spanner/session_test.go; 121 LoC -
Found a possible issue in [fraibacas/zing-injector](https://www.github.com/fraibacas/zing-injector) at [vendor/cloud.google.com/go/spanner/session_test.go](https://github.com/fraibacas/zing-injector/blob/d8119599ee022f33d4f3faf29c90de0d81346c15/vendor/cloud.google.com/go/spanner/session_test.go#L659-L779)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable ti used in defer or goroutine at line 701
[Click here to see the code in its original context.](https://github.com/fraibacas/zing-injector/blob/d8119599ee022f33d4f3faf29c90de0d81346c15/vendor/cloud.google.com/go/spanner/session_test.go#L659-L779)
<details>
<summary>Click here to show the 121 line(s) of Go which triggered the analyzer.</summary>
```go
for ti, cfg := range []SessionPoolConfig{
SessionPoolConfig{},
SessionPoolConfig{MinOpened: 10, MaxOpened: 100},
SessionPoolConfig{MaxBurst: 50},
SessionPoolConfig{MinOpened: 10, MaxOpened: 200, MaxBurst: 5},
SessionPoolConfig{MinOpened: 10, MaxOpened: 200, MaxBurst: 5, WriteSessions: 0.2},
} {
var wg sync.WaitGroup
// Create a more aggressive session healthchecker to increase test concurrency.
cfg.HealthCheckInterval = 50 * time.Millisecond
cfg.healthCheckSampleInterval = 10 * time.Millisecond
cfg.HealthCheckWorkers = 50
sc := testutil.NewMockCloudSpannerClient(t)
sc.MakeNice()
cfg.getRPCClient = func() (sppb.SpannerClient, error) {
return sc, nil
}
sp, _ := newSessionPool("mockdb", cfg, nil)
for i := 0; i < 100; i++ {
wg.Add(1)
// Schedule a test worker.
go func(idx int, pool *sessionPool, client sppb.SpannerClient) {
defer wg.Done()
// Test worker iterates 1K times and tries different session / session pool operations.
for j := 0; j < 1000; j++ {
if idx%10 == 0 && j >= 900 {
// Close the pool in selected set of workers during the middle of the test.
pool.close()
}
// Take a write sessions ~ 20% of the times.
takeWrite := rand.Intn(5) == 4
var (
sh *sessionHandle
gotErr error
)
if takeWrite {
sh, gotErr = pool.takeWriteSession(context.Background())
} else {
sh, gotErr = pool.take(context.Background())
}
if gotErr != nil {
if pool.isValid() {
t.Errorf("%v.%v: pool.take returns error when pool is still valid: %v", ti, idx, gotErr)
}
if wantErr := errInvalidSessionPool(); !testEqual(gotErr, wantErr) {
t.Errorf("%v.%v: got error when pool is closed: %v, want %v", ti, idx, gotErr, wantErr)
}
continue
}
// Verify if session is valid when session pool is valid. Note that if session pool is invalid after sh is taken,
// then sh might be invalidated by healthcheck workers.
if (sh.getID() == "" || sh.session == nil || !sh.session.isValid()) && pool.isValid() {
t.Errorf("%v.%v.%v: pool.take returns invalid session %v", ti, idx, takeWrite, sh.session)
}
if takeWrite && sh.getTransactionID() == nil {
t.Errorf("%v.%v: pool.takeWriteSession returns session %v without transaction", ti, idx, sh.session)
}
if rand.Intn(100) < idx {
// Random sleep before destroying/recycling the session, to give healthcheck worker a chance to step in.
<-time.After(time.Duration(rand.Int63n(int64(cfg.HealthCheckInterval))))
}
if rand.Intn(100) < idx {
// destroy the session.
sh.destroy()
continue
}
// recycle the session.
sh.recycle()
}
}(i, sp, sc)
}
wg.Wait()
sp.hc.close()
// Here the states of healthchecker, session pool and mockclient are stable.
idleSessions := map[string]bool{}
hcSessions := map[string]bool{}
mockSessions := sc.DumpSessions()
// Dump session pool's idle list.
for sl := sp.idleList.Front(); sl != nil; sl = sl.Next() {
s := sl.Value.(*session)
if idleSessions[s.getID()] {
t.Errorf("%v: found duplicated session in idle list: %v", ti, s.getID())
}
idleSessions[s.getID()] = true
}
for sl := sp.idleWriteList.Front(); sl != nil; sl = sl.Next() {
s := sl.Value.(*session)
if idleSessions[s.getID()] {
t.Errorf("%v: found duplicated session in idle write list: %v", ti, s.getID())
}
idleSessions[s.getID()] = true
}
sp.mu.Lock()
if int(sp.numOpened) != len(idleSessions) {
t.Errorf("%v: number of opened sessions (%v) != number of idle sessions (%v)", ti, sp.numOpened, len(idleSessions))
}
if sp.createReqs != 0 {
t.Errorf("%v: number of pending session creations = %v, want 0", ti, sp.createReqs)
}
// Dump healthcheck queue.
for _, s := range sp.hc.queue.sessions {
if hcSessions[s.getID()] {
t.Errorf("%v: found duplicated session in healthcheck queue: %v", ti, s.getID())
}
hcSessions[s.getID()] = true
}
sp.mu.Unlock()
// Verify that idleSessions == hcSessions == mockSessions.
if !testEqual(idleSessions, hcSessions) {
t.Errorf("%v: sessions in idle list (%v) != sessions in healthcheck queue (%v)", ti, idleSessions, hcSessions)
}
if !testEqual(hcSessions, mockSessions) {
t.Errorf("%v: sessions in healthcheck queue (%v) != sessions in mockclient (%v)", ti, hcSessions, mockSessions)
}
sp.close()
mockSessions = sc.DumpSessions()
if len(mockSessions) != 0 {
t.Errorf("Found live sessions: %v", mockSessions)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: d8119599ee022f33d4f3faf29c90de0d81346c15
| test | fraibacas zing injector vendor cloud google com go spanner session test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable ti used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for ti cfg range sessionpoolconfig sessionpoolconfig sessionpoolconfig minopened maxopened sessionpoolconfig maxburst sessionpoolconfig minopened maxopened maxburst sessionpoolconfig minopened maxopened maxburst writesessions var wg sync waitgroup create a more aggressive session healthchecker to increase test concurrency cfg healthcheckinterval time millisecond cfg healthchecksampleinterval time millisecond cfg healthcheckworkers sc testutil newmockcloudspannerclient t sc makenice cfg getrpcclient func sppb spannerclient error return sc nil sp newsessionpool mockdb cfg nil for i i i wg add schedule a test worker go func idx int pool sessionpool client sppb spannerclient defer wg done test worker iterates times and tries different session session pool operations for j j j if idx j close the pool in selected set of workers during the middle of the test pool close take a write sessions of the times takewrite rand intn var sh sessionhandle goterr error if takewrite sh goterr pool takewritesession context background else sh goterr pool take context background if goterr nil if pool isvalid t errorf v v pool take returns error when pool is still valid v ti idx goterr if wanterr errinvalidsessionpool testequal goterr wanterr t errorf v v got error when pool is closed v want v ti idx goterr wanterr continue verify if session is valid when session pool is valid note that if session pool is invalid after sh is taken then sh might be invalidated by healthcheck workers if sh getid sh session nil sh session isvalid pool isvalid t errorf v v v pool take returns invalid session v ti idx takewrite sh session if takewrite sh gettransactionid nil t errorf v v pool takewritesession returns session v without transaction ti idx sh session if rand intn idx random sleep before destroying recycling the session to give healthcheck worker a chance to step in time after time duration rand cfg healthcheckinterval if rand intn idx destroy the session sh destroy continue recycle the session sh recycle i sp sc wg wait sp hc close here the states of healthchecker session pool and mockclient are stable idlesessions map bool hcsessions map bool mocksessions sc dumpsessions dump session pool s idle list for sl sp idlelist front sl nil sl sl next s sl value session if idlesessions t errorf v found duplicated session in idle list v ti s getid idlesessions true for sl sp idlewritelist front sl nil sl sl next s sl value session if idlesessions t errorf v found duplicated session in idle write list v ti s getid idlesessions true sp mu lock if int sp numopened len idlesessions t errorf v number of opened sessions v number of idle sessions v ti sp numopened len idlesessions if sp createreqs t errorf v number of pending session creations v want ti sp createreqs dump healthcheck queue for s range sp hc queue sessions if hcsessions t errorf v found duplicated session in healthcheck queue v ti s getid hcsessions true sp mu unlock verify that idlesessions hcsessions mocksessions if testequal idlesessions hcsessions t errorf v sessions in idle list v sessions in healthcheck queue v ti idlesessions hcsessions if testequal hcsessions mocksessions t errorf v sessions in healthcheck queue v sessions in mockclient v ti hcsessions mocksessions sp close mocksessions sc dumpsessions if len mocksessions t errorf found live sessions v mocksessions leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 1 |
29,305 | 2,714,582,938 | IssuesEvent | 2015-04-10 05:46:35 | OCHA-DAP/hdx-ckan | https://api.github.com/repos/OCHA-DAP/hdx-ckan | closed | Custom org page: improvements for logo1 and logo2 | bug Custom org page Priority-High | 1. User should see some information about each "browse" for logo1 and logo2 such as:
- logo square (75x75)
- logo rectangulare ...
2. When edit an organization user has to re-upload again the logos otherwise it will loose the settings.
3. What is happening with uploaded logos when a new logo is uploaded. I think we should discuss about this as we don't want to keep old logos in system and to make it easy for user. | 1.0 | Custom org page: improvements for logo1 and logo2 - 1. User should see some information about each "browse" for logo1 and logo2 such as:
- logo square (75x75)
- logo rectangulare ...
2. When edit an organization user has to re-upload again the logos otherwise it will loose the settings.
3. What is happening with uploaded logos when a new logo is uploaded. I think we should discuss about this as we don't want to keep old logos in system and to make it easy for user. | non_test | custom org page improvements for and user should see some information about each browse for and such as logo square logo rectangulare when edit an organization user has to re upload again the logos otherwise it will loose the settings what is happening with uploaded logos when a new logo is uploaded i think we should discuss about this as we don t want to keep old logos in system and to make it easy for user | 0 |
143,263 | 11,531,574,507 | IssuesEvent | 2020-02-17 01:25:38 | timeforcamp/time-for-camp | https://api.github.com/repos/timeforcamp/time-for-camp | closed | CAMP TRANSACTIONS: Payment Method when a Store Item Fee is refunded | PRIORITY 2 UI Modification ready to test | When a store item is refunded, we return the previously deducted item fee to the camp, but right now, there is no payment method listed. We need to apply a payment method of ADD (I guess) since that is the opposite of deduct.


| 1.0 | CAMP TRANSACTIONS: Payment Method when a Store Item Fee is refunded - When a store item is refunded, we return the previously deducted item fee to the camp, but right now, there is no payment method listed. We need to apply a payment method of ADD (I guess) since that is the opposite of deduct.


| test | camp transactions payment method when a store item fee is refunded when a store item is refunded we return the previously deducted item fee to the camp but right now there is no payment method listed we need to apply a payment method of add i guess since that is the opposite of deduct | 1 |
127,470 | 17,273,076,267 | IssuesEvent | 2021-07-22 23:18:59 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | Use predefined search queries in links to the docs instead of specific doc URLs | Team:Docs Team:Kibana-Design discuss docs | **Describe the feature:**
A key feature of the new doc system is a much-improved search experience that will make it easier to discover and navigate through our ever-growing network of topics. Leveraging that system instead of hard-coding links to specific doc topics has a number of advantages:
- Links would always direct users to the most relevant, up-to-date content. No more link rot.
- Links (essentially query terms) would only need to be updated if the UI/functionality changed significantly.
- We can curate the results via App Search independent of the release cycle.
- Users would see *all* of the relevant content that's available, not just a single topic that might not answer their question.
- It would reinforce using our improved search capabilities to discover and navigate through our content.
The downside is that it's always two clicks to get to actual info. However, giving people a tailored access point to the docs fits our "Learn more" model of linking to docs. If there's a specific piece of information that we expect people to want/need, the various forms of in-product assistance are a better choice than sending people out of Kibana.
@gchaps @goodroot | 1.0 | Use predefined search queries in links to the docs instead of specific doc URLs - **Describe the feature:**
A key feature of the new doc system is a much-improved search experience that will make it easier to discover and navigate through our ever-growing network of topics. Leveraging that system instead of hard-coding links to specific doc topics has a number of advantages:
- Links would always direct users to the most relevant, up-to-date content. No more link rot.
- Links (essentially query terms) would only need to be updated if the UI/functionality changed significantly.
- We can curate the results via App Search independent of the release cycle.
- Users would see *all* of the relevant content that's available, not just a single topic that might not answer their question.
- It would reinforce using our improved search capabilities to discover and navigate through our content.
The downside is that it's always two clicks to get to actual info. However, giving people a tailored access point to the docs fits our "Learn more" model of linking to docs. If there's a specific piece of information that we expect people to want/need, the various forms of in-product assistance are a better choice than sending people out of Kibana.
@gchaps @goodroot | non_test | use predefined search queries in links to the docs instead of specific doc urls describe the feature a key feature of the new doc system is a much improved search experience that will make it easier to discover and navigate through our ever growing network of topics leveraging that system instead of hard coding links to specific doc topics has a number of advantages links would always direct users to the most relevant up to date content no more link rot links essentially query terms would only need to be updated if the ui functionality changed significantly we can curate the results via app search independent of the release cycle users would see all of the relevant content that s available not just a single topic that might not answer their question it would reinforce using our improved search capabilities to discover and navigate through our content the downside is that it s always two clicks to get to actual info however giving people a tailored access point to the docs fits our learn more model of linking to docs if there s a specific piece of information that we expect people to want need the various forms of in product assistance are a better choice than sending people out of kibana gchaps goodroot | 0 |
359,113 | 25,220,364,234 | IssuesEvent | 2022-11-14 12:19:55 | ToolJet/ToolJet | https://api.github.com/repos/ToolJet/ToolJet | closed | [docs]: Adding Restack as a way to deploy ToolJet on k8s on the users' AWS infrastructure. | documentation | ### Summary
I wish to add [Restack](https://www.restack.io) as a way to deploy Tooljet on kubernetes on the user's AWS infrastructure.
I propose to add it on the same level as the option with Heroku under [Setup](https://docs.tooljet.com/docs/setup/).
Please let me know if you have any suggestions.
### Issue Type
Documentation feature/update
### The entire URL of the documentation with the issue
https://docs.tooljet.com/docs/setup/
### Steps to reproduce the issue
_No response_
### Additional Information
_No response_
### If the issue is confirmed, would you be willing to submit a pull request?
Yes
### Code of Conduct
- [X] I agree to follow the ToolJet Code of Conduct | 1.0 | [docs]: Adding Restack as a way to deploy ToolJet on k8s on the users' AWS infrastructure. - ### Summary
I wish to add [Restack](https://www.restack.io) as a way to deploy Tooljet on kubernetes on the user's AWS infrastructure.
I propose to add it on the same level as the option with Heroku under [Setup](https://docs.tooljet.com/docs/setup/).
Please let me know if you have any suggestions.
### Issue Type
Documentation feature/update
### The entire URL of the documentation with the issue
https://docs.tooljet.com/docs/setup/
### Steps to reproduce the issue
_No response_
### Additional Information
_No response_
### If the issue is confirmed, would you be willing to submit a pull request?
Yes
### Code of Conduct
- [X] I agree to follow the ToolJet Code of Conduct | non_test | adding restack as a way to deploy tooljet on on the users aws infrastructure summary i wish to add as a way to deploy tooljet on kubernetes on the user s aws infrastructure i propose to add it on the same level as the option with heroku under please let me know if you have any suggestions issue type documentation feature update the entire url of the documentation with the issue steps to reproduce the issue no response additional information no response if the issue is confirmed would you be willing to submit a pull request yes code of conduct i agree to follow the tooljet code of conduct | 0 |
297,710 | 25,758,133,805 | IssuesEvent | 2022-12-08 18:02:49 | johnpaulrusso/svelte-text-logger | https://api.github.com/repos/johnpaulrusso/svelte-text-logger | closed | The logger should auto-scroll to the latest log entry. | enhancement R - Ready For Test | > estimate 2
The logger should auto-scroll to the latest log entry. | 1.0 | The logger should auto-scroll to the latest log entry. - > estimate 2
The logger should auto-scroll to the latest log entry. | test | the logger should auto scroll to the latest log entry estimate the logger should auto scroll to the latest log entry | 1 |
90,200 | 26,006,219,625 | IssuesEvent | 2022-12-20 19:39:14 | MicrosoftDocs/azure-devops-docs | https://api.github.com/repos/MicrosoftDocs/azure-devops-docs | closed | There used to be a section in this documentation that explained how to refer to these variables in tasks and scripts | doc-enhancement devops/prod Pri2 devops-cicd-build/tech | Like in a powershell script, one would use `$env:BUILD_BUILD_NUMBER` or something like that, or in a shell script it would be something different. That article content needs to be restored or made clearly available to a doc page like this one.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 11d3469c-83c2-c2f6-41d0-08f82c330bbc
* Version Independent ID: a373c48d-6395-3002-d2d8-bbc1fea2e48e
* Content: [Predefined variables - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/build/variables?view=azure-devops&tabs=classic)
* Content Source: [docs/pipelines/build/variables.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/build/variables.md)
* Product: **devops**
* Technology: **devops-cicd-build**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | 1.0 | There used to be a section in this documentation that explained how to refer to these variables in tasks and scripts - Like in a powershell script, one would use `$env:BUILD_BUILD_NUMBER` or something like that, or in a shell script it would be something different. That article content needs to be restored or made clearly available to a doc page like this one.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 11d3469c-83c2-c2f6-41d0-08f82c330bbc
* Version Independent ID: a373c48d-6395-3002-d2d8-bbc1fea2e48e
* Content: [Predefined variables - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/build/variables?view=azure-devops&tabs=classic)
* Content Source: [docs/pipelines/build/variables.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/build/variables.md)
* Product: **devops**
* Technology: **devops-cicd-build**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | non_test | there used to be a section in this documentation that explained how to refer to these variables in tasks and scripts like in a powershell script one would use env build build number or something like that or in a shell script it would be something different that article content needs to be restored or made clearly available to a doc page like this one document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd build github login juliakm microsoft alias jukullam | 0 |
129,071 | 10,561,696,303 | IssuesEvent | 2019-10-04 16:28:00 | knative/serving | https://api.github.com/repos/knative/serving | closed | Run unit tests locally | area/test-and-release kind/bug kind/question | /area test-and-release
Classify what kind of issue this is:
/kind question
/kind bug
## Expected Behavior
1) git clone https://github.com/knative/serving.git && cd serving
2) go test ./... should be pass (as it is passing in CI)
## Actual Behavior
1) git clone url && cd serving
2) go test ./... is faling
Note: I am in correct GOPATH(`$(go env GOPATH)/src/github.com/knative.dev`) I tried after `dep ensure` also. Still tests are failing | 1.0 | Run unit tests locally - /area test-and-release
Classify what kind of issue this is:
/kind question
/kind bug
## Expected Behavior
1) git clone https://github.com/knative/serving.git && cd serving
2) go test ./... should be pass (as it is passing in CI)
## Actual Behavior
1) git clone url && cd serving
2) go test ./... is faling
Note: I am in correct GOPATH(`$(go env GOPATH)/src/github.com/knative.dev`) I tried after `dep ensure` also. Still tests are failing | test | run unit tests locally area test and release classify what kind of issue this is kind question kind bug expected behavior git clone cd serving go test should be pass as it is passing in ci actual behavior git clone url cd serving go test is faling note i am in correct gopath go env gopath src github com knative dev i tried after dep ensure also still tests are failing | 1 |
254,806 | 21,877,701,510 | IssuesEvent | 2022-05-19 11:45:53 | status-im/status-desktop | https://api.github.com/repos/status-im/status-desktop | closed | should not be possible to edit stickers | bug ui Chat tested priority 3: low | # Bug Report
1. send a sticker
2. right click and press edit
result:

| 1.0 | should not be possible to edit stickers - # Bug Report
1. send a sticker
2. right click and press edit
result:

| test | should not be possible to edit stickers bug report send a sticker right click and press edit result | 1 |
23,915 | 3,871,282,201 | IssuesEvent | 2016-04-11 09:10:39 | katsube/WingPHP | https://api.github.com/repos/katsube/WingPHP | closed | PHP5.4 | auto-migrated Priority-Medium Type-Defect | ```
PHP5.4での動作チェック。
```
Original issue reported on code.google.com by `katsubemakito` on 14 Apr 2012 at 7:25 | 1.0 | PHP5.4 - ```
PHP5.4での動作チェック。
```
Original issue reported on code.google.com by `katsubemakito` on 14 Apr 2012 at 7:25 | non_test | 。 original issue reported on code google com by katsubemakito on apr at | 0 |
652,596 | 21,556,582,638 | IssuesEvent | 2022-04-30 14:23:27 | pystardust/ani-cli | https://api.github.com/repos/pystardust/ani-cli | closed | Let's make a pull request template | type: feature request priority 4: wishlist category: qa | **Is your feature request related to a problem? Please describe.**
Currently our PR workflow feels kinda incomplete / awkward, a template could enrich it.
**Describe the solution you'd like**
Just have a checklist as detailed in the Contributing.md under Merge checklist
**Describe alternatives you've considered**
Adding the releases checklist seems like a good idea if we can communicate that it's not needed for every PR.
Additionally we could add a modified metadata section from the bug report template.
**Additional context**
I've had the idea for several weeks now.
A short ~google~ duckduckgo search gave me the following promising [article](https://embeddedartistry.com/blog/2017/08/04/a-github-pull-request-template-for-your-projects/
) | 1.0 | Let's make a pull request template - **Is your feature request related to a problem? Please describe.**
Currently our PR workflow feels kinda incomplete / awkward, a template could enrich it.
**Describe the solution you'd like**
Just have a checklist as detailed in the Contributing.md under Merge checklist
**Describe alternatives you've considered**
Adding the releases checklist seems like a good idea if we can communicate that it's not needed for every PR.
Additionally we could add a modified metadata section from the bug report template.
**Additional context**
I've had the idea for several weeks now.
A short ~google~ duckduckgo search gave me the following promising [article](https://embeddedartistry.com/blog/2017/08/04/a-github-pull-request-template-for-your-projects/
) | non_test | let s make a pull request template is your feature request related to a problem please describe currently our pr workflow feels kinda incomplete awkward a template could enrich it describe the solution you d like just have a checklist as detailed in the contributing md under merge checklist describe alternatives you ve considered adding the releases checklist seems like a good idea if we can communicate that it s not needed for every pr additionally we could add a modified metadata section from the bug report template additional context i ve had the idea for several weeks now a short google duckduckgo search gave me the following promising | 0 |
151,762 | 12,057,567,144 | IssuesEvent | 2020-04-15 16:02:05 | Azure/azure-iot-sdk-csharp | https://api.github.com/repos/Azure/azure-iot-sdk-csharp | closed | ProvisioningServiceClient_v2: Add unit tests for ProductInfo in common | IoTSDK test bug | Need to add unit tests for the ProductInfo class in commons | 1.0 | ProvisioningServiceClient_v2: Add unit tests for ProductInfo in common - Need to add unit tests for the ProductInfo class in commons | test | provisioningserviceclient add unit tests for productinfo in common need to add unit tests for the productinfo class in commons | 1 |
267,186 | 8,380,101,739 | IssuesEvent | 2018-10-07 11:14:23 | CS2103-AY1819S1-W12-1/main | https://api.github.com/repos/CS2103-AY1819S1-W12-1/main | closed | Tests failing due to global name change | priority.medium severity.low type.bug | See #29. Tests are getting screwed, `checkstyleMain` in particular. @plty clean your own mess after rushing for the deadline you dimwit :( | 1.0 | Tests failing due to global name change - See #29. Tests are getting screwed, `checkstyleMain` in particular. @plty clean your own mess after rushing for the deadline you dimwit :( | non_test | tests failing due to global name change see tests are getting screwed checkstylemain in particular plty clean your own mess after rushing for the deadline you dimwit | 0 |
4,388 | 2,851,635,229 | IssuesEvent | 2015-06-01 08:24:14 | arduino/Arduino | https://api.github.com/repos/arduino/Arduino | closed | Document the WiFi status() function. | Component: Documentation Library: Wifi | It's used in the examples but I don't see it in the reference. We should document its possible return values too. | 1.0 | Document the WiFi status() function. - It's used in the examples but I don't see it in the reference. We should document its possible return values too. | non_test | document the wifi status function it s used in the examples but i don t see it in the reference we should document its possible return values too | 0 |
1,734 | 2,570,986,004 | IssuesEvent | 2015-02-10 13:54:14 | photonstorm/phaser | https://api.github.com/repos/photonstorm/phaser | closed | require('phaser') with Browserify throws error: Error: Cannot find module './Scalar' | please test | I've also got these errors:
- Error: Cannot find module './Scalar'
- Error: Cannot find module './Line' | 1.0 | require('phaser') with Browserify throws error: Error: Cannot find module './Scalar' - I've also got these errors:
- Error: Cannot find module './Scalar'
- Error: Cannot find module './Line' | test | require phaser with browserify throws error error cannot find module scalar i ve also got these errors error cannot find module scalar error cannot find module line | 1 |
159,352 | 12,474,470,844 | IssuesEvent | 2020-05-29 09:42:42 | aliasrobotics/RVD | https://api.github.com/repos/aliasrobotics/RVD | opened | Using xml., /opt/ros_noetic_ws/src/rqt/rqt_gui/src/rqt_gui/ros_plugin_provider.py:123 | bandit bug static analysis testing triage | ```yaml
{
"id": 1,
"title": "Using xml., /opt/ros_noetic_ws/src/rqt/rqt_gui/src/rqt_gui/ros_plugin_provider.py:123",
"type": "bug",
"description": "HIGH confidence of MEDIUM severity bug. Using xml.etree.ElementTree.parse to parse untrusted XML data is known to be vulnerable to XML attacks. Replace xml.etree.ElementTree.parse with its defusedxml equivalent function or make sure defusedxml.defuse_stdlib() is called at /opt/ros_noetic_ws/src/rqt/rqt_gui/src/rqt_gui/ros_plugin_provider.py:123 See links for more info on the bug.",
"cwe": "None",
"cve": "None",
"keywords": [
"bandit",
"bug",
"static analysis",
"testing",
"triage",
"bug"
],
"system": "",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "subject-specific",
"architectural-location": "application-specific",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-05-29 (09:42)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-05-29 (09:42)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "/opt/ros_noetic_ws/src/rqt/rqt_gui/src/rqt_gui/ros_plugin_provider.py:123",
"reproduction": "See artifacts below (if available)",
"reproduction-image": ""
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
``` | 1.0 | Using xml., /opt/ros_noetic_ws/src/rqt/rqt_gui/src/rqt_gui/ros_plugin_provider.py:123 - ```yaml
{
"id": 1,
"title": "Using xml., /opt/ros_noetic_ws/src/rqt/rqt_gui/src/rqt_gui/ros_plugin_provider.py:123",
"type": "bug",
"description": "HIGH confidence of MEDIUM severity bug. Using xml.etree.ElementTree.parse to parse untrusted XML data is known to be vulnerable to XML attacks. Replace xml.etree.ElementTree.parse with its defusedxml equivalent function or make sure defusedxml.defuse_stdlib() is called at /opt/ros_noetic_ws/src/rqt/rqt_gui/src/rqt_gui/ros_plugin_provider.py:123 See links for more info on the bug.",
"cwe": "None",
"cve": "None",
"keywords": [
"bandit",
"bug",
"static analysis",
"testing",
"triage",
"bug"
],
"system": "",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "subject-specific",
"architectural-location": "application-specific",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-05-29 (09:42)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-05-29 (09:42)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "/opt/ros_noetic_ws/src/rqt/rqt_gui/src/rqt_gui/ros_plugin_provider.py:123",
"reproduction": "See artifacts below (if available)",
"reproduction-image": ""
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
``` | test | using xml opt ros noetic ws src rqt rqt gui src rqt gui ros plugin provider py yaml id title using xml opt ros noetic ws src rqt rqt gui src rqt gui ros plugin provider py type bug description high confidence of medium severity bug using xml etree elementtree parse to parse untrusted xml data is known to be vulnerable to xml attacks replace xml etree elementtree parse with its defusedxml equivalent function or make sure defusedxml defuse stdlib is called at opt ros noetic ws src rqt rqt gui src rqt gui ros plugin provider py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace opt ros noetic ws src rqt rqt gui src rqt gui ros plugin provider py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation | 1 |
341,049 | 30,565,125,691 | IssuesEvent | 2023-07-20 17:10:45 | jon-harper/OmniBox | https://api.github.com/repos/jon-harper/OmniBox | closed | Fit Test: New RSP-500 PSU tray | PSU Fit Test Needed | This version is modified to make room for M4 socket head cap screws.
[Files are in the `unstable` branch.](https://github.com/jon-harper/OmniBox/tree/unstable/Trays/PSU/Mean%20Well%20RSP-500) | 1.0 | Fit Test: New RSP-500 PSU tray - This version is modified to make room for M4 socket head cap screws.
[Files are in the `unstable` branch.](https://github.com/jon-harper/OmniBox/tree/unstable/Trays/PSU/Mean%20Well%20RSP-500) | test | fit test new rsp psu tray this version is modified to make room for socket head cap screws | 1 |
15,119 | 3,441,444,192 | IssuesEvent | 2015-12-14 18:29:57 | DMOJ/site | https://api.github.com/repos/DMOJ/site | closed | Disable ability to view others' submissions in contest, even if you have solved the problem | contest duplicate enhancement | Why is this even a thing? | 1.0 | Disable ability to view others' submissions in contest, even if you have solved the problem - Why is this even a thing? | test | disable ability to view others submissions in contest even if you have solved the problem why is this even a thing | 1 |
270,325 | 23,501,400,153 | IssuesEvent | 2022-08-18 08:44:58 | wazuh/wazuh | https://api.github.com/repos/wazuh/wazuh | closed | Release 4.3.7 - Revision 1 - Release Candidate RC1 - Footprint Metrics - LOGCOLLECTOR,SCA (4h) | release test/4.3.7 | ## Footprint metrics information
| | |
|---------------------------------|--------------------------------------------|
| **Main release candidate issue #** | # |
| **Main footprint metrics issue #** | #14554 |
| **Version** | 4.3.7 |
| **Release candidate #** | RC1 |
| **Tag** | https://github.com/wazuh/wazuh/tree/4.3.7-rc1 |
## Stress test documentation
### Packages used
- Repository: `packages-dev.wazuh.com`
- Package path: `pre-release`
- Package revision: `1`
- **Jenkins build**: https://ci.wazuh.info/job/Test_stress/3523/
---
<details><summary>Manager</summary>
+ <details><summary>Plots</summary>
















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3523_manager_2022-08-16.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_manager_centos/logs/ossec_Test_stress_B3523_manager_2022-08-16.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-manager-Test_stress_B3523_manager-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_manager_centos/data/monitor-manager-Test_stress_B3523_manager-pre-release.csv)
[Test_stress_B3523_manager_analysisd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_manager_centos/data/Test_stress_B3523_manager_analysisd_state.csv)
[Test_stress_B3523_manager_remoted_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_manager_centos/data/Test_stress_B3523_manager_remoted_state.csv)
</details>
</details>
<details><summary>Centos agent</summary>
+ <details><summary>Plots</summary>

















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3523_centos_2022-08-16.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_centos/logs/ossec_Test_stress_B3523_centos_2022-08-16.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-agent-Test_stress_B3523_centos-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_centos/data/monitor-agent-Test_stress_B3523_centos-pre-release.csv)
[Test_stress_B3523_centos_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_centos/data/Test_stress_B3523_centos_agentd_state.csv)
</details>
</details>
<details><summary>Ubuntu agent</summary>
+ <details><summary>Plots</summary>

















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3523_ubuntu_2022-08-16.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_ubuntu/logs/ossec_Test_stress_B3523_ubuntu_2022-08-16.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-agent-Test_stress_B3523_ubuntu-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_ubuntu/data/monitor-agent-Test_stress_B3523_ubuntu-pre-release.csv)
[Test_stress_B3523_ubuntu_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_ubuntu/data/Test_stress_B3523_ubuntu_agentd_state.csv)
</details>
</details>
<details><summary>Windows agent</summary>
+ <details><summary>Plots</summary>















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3523_windows_2022-08-16.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_windows/logs/ossec_Test_stress_B3523_windows_2022-08-16.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-winagent-Test_stress_B3523_windows-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_windows/data/monitor-winagent-Test_stress_B3523_windows-pre-release.csv)
[Test_stress_B3523_windows_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_windows/data/Test_stress_B3523_windows_agentd_state.csv)
</details>
</details>
<details><summary>macOS agent</summary>
+ <details><summary>Plots</summary>
</details>
+ <details><summary>Logs and configuration</summary>
</details>
+ <details><summary>CSV</summary>
</details>
</details>
<details><summary>Solaris agent</summary>
+ <details><summary>Plots</summary>
</details>
+ <details><summary>Logs and configuration</summary>
</details>
+ <details><summary>CSV</summary>
</details>
</details> | 1.0 | Release 4.3.7 - Revision 1 - Release Candidate RC1 - Footprint Metrics - LOGCOLLECTOR,SCA (4h) - ## Footprint metrics information
| | |
|---------------------------------|--------------------------------------------|
| **Main release candidate issue #** | # |
| **Main footprint metrics issue #** | #14554 |
| **Version** | 4.3.7 |
| **Release candidate #** | RC1 |
| **Tag** | https://github.com/wazuh/wazuh/tree/4.3.7-rc1 |
## Stress test documentation
### Packages used
- Repository: `packages-dev.wazuh.com`
- Package path: `pre-release`
- Package revision: `1`
- **Jenkins build**: https://ci.wazuh.info/job/Test_stress/3523/
---
<details><summary>Manager</summary>
+ <details><summary>Plots</summary>
















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3523_manager_2022-08-16.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_manager_centos/logs/ossec_Test_stress_B3523_manager_2022-08-16.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-manager-Test_stress_B3523_manager-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_manager_centos/data/monitor-manager-Test_stress_B3523_manager-pre-release.csv)
[Test_stress_B3523_manager_analysisd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_manager_centos/data/Test_stress_B3523_manager_analysisd_state.csv)
[Test_stress_B3523_manager_remoted_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_manager_centos/data/Test_stress_B3523_manager_remoted_state.csv)
</details>
</details>
<details><summary>Centos agent</summary>
+ <details><summary>Plots</summary>

















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3523_centos_2022-08-16.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_centos/logs/ossec_Test_stress_B3523_centos_2022-08-16.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-agent-Test_stress_B3523_centos-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_centos/data/monitor-agent-Test_stress_B3523_centos-pre-release.csv)
[Test_stress_B3523_centos_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_centos/data/Test_stress_B3523_centos_agentd_state.csv)
</details>
</details>
<details><summary>Ubuntu agent</summary>
+ <details><summary>Plots</summary>

















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3523_ubuntu_2022-08-16.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_ubuntu/logs/ossec_Test_stress_B3523_ubuntu_2022-08-16.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-agent-Test_stress_B3523_ubuntu-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_ubuntu/data/monitor-agent-Test_stress_B3523_ubuntu-pre-release.csv)
[Test_stress_B3523_ubuntu_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_ubuntu/data/Test_stress_B3523_ubuntu_agentd_state.csv)
</details>
</details>
<details><summary>Windows agent</summary>
+ <details><summary>Plots</summary>















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3523_windows_2022-08-16.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_windows/logs/ossec_Test_stress_B3523_windows_2022-08-16.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-winagent-Test_stress_B3523_windows-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_windows/data/monitor-winagent-Test_stress_B3523_windows-pre-release.csv)
[Test_stress_B3523_windows_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.7/B3523-240m/B3523_agent_windows/data/Test_stress_B3523_windows_agentd_state.csv)
</details>
</details>
<details><summary>macOS agent</summary>
+ <details><summary>Plots</summary>
</details>
+ <details><summary>Logs and configuration</summary>
</details>
+ <details><summary>CSV</summary>
</details>
</details>
<details><summary>Solaris agent</summary>
+ <details><summary>Plots</summary>
</details>
+ <details><summary>Logs and configuration</summary>
</details>
+ <details><summary>CSV</summary>
</details>
</details> | test | release revision release candidate footprint metrics logcollector sca footprint metrics information main release candidate issue main footprint metrics issue version release candidate tag stress test documentation packages used repository packages dev wazuh com package path pre release package revision jenkins build manager plots logs and configuration csv centos agent plots logs and configuration csv ubuntu agent plots logs and configuration csv windows agent plots logs and configuration csv macos agent plots logs and configuration csv solaris agent plots logs and configuration csv | 1 |
24,360 | 17,138,885,985 | IssuesEvent | 2021-07-13 07:19:55 | YAPP-18th/Android-Team-1-Backend | https://api.github.com/repos/YAPP-18th/Android-Team-1-Backend | closed | [수정] 로깅 관련 조정 | application enhancement infrastructure | 현재 로그 레벨이 error 로 되어 있어서, 예외 발생시에만 에러 수집이 되고 있다.
운영상 이슈를 점검하기 위한 적절한 로그 레벨 검토 필요
또한, GKE 의 Ingress 상에서 요청 / 응답 데이터를 뜰 수 있을까? 불가능한 경우 WAS 상에서 구현해야 할 듯 하다. | 1.0 | [수정] 로깅 관련 조정 - 현재 로그 레벨이 error 로 되어 있어서, 예외 발생시에만 에러 수집이 되고 있다.
운영상 이슈를 점검하기 위한 적절한 로그 레벨 검토 필요
또한, GKE 의 Ingress 상에서 요청 / 응답 데이터를 뜰 수 있을까? 불가능한 경우 WAS 상에서 구현해야 할 듯 하다. | non_test | 로깅 관련 조정 현재 로그 레벨이 error 로 되어 있어서 예외 발생시에만 에러 수집이 되고 있다 운영상 이슈를 점검하기 위한 적절한 로그 레벨 검토 필요 또한 gke 의 ingress 상에서 요청 응답 데이터를 뜰 수 있을까 불가능한 경우 was 상에서 구현해야 할 듯 하다 | 0 |
166,482 | 6,305,585,988 | IssuesEvent | 2017-07-21 18:44:04 | reymesson1/primerejercicio | https://api.github.com/repos/reymesson1/primerejercicio | closed | Create microservices | In Progress Priority: Middle Requeriments | ## The notifications
* E-mails should be HTML formatted and should have the same look & feel as your site.
* A micro-service should send notifications for active orders placed between the past 3 and 5 minutes. (This rule should randomize the criteria)
* Once an e-mail is sent, the order should be marked as Delivered.
| 1.0 | Create microservices - ## The notifications
* E-mails should be HTML formatted and should have the same look & feel as your site.
* A micro-service should send notifications for active orders placed between the past 3 and 5 minutes. (This rule should randomize the criteria)
* Once an e-mail is sent, the order should be marked as Delivered.
| non_test | create microservices the notifications e mails should be html formatted and should have the same look feel as your site a micro service should send notifications for active orders placed between the past and minutes this rule should randomize the criteria once an e mail is sent the order should be marked as delivered | 0 |
93,794 | 8,444,718,527 | IssuesEvent | 2018-10-18 19:16:17 | steve8x8/geotoad | https://api.github.com/repos/steve8x8/geotoad | closed | Logbook (json) comes gzipped | Priority-Low bug patch to test | Starting around September 1, logbook entries (for log counts > 10 only) are delivered in gzipped form, but the decompression part was still missing from the code.
The patch is simple: just comment out the "Accept-Encoding" setting in lib/logbook.rb - a better solution (adding decompression to lib/shadowget.rb) is being worked on.
[issue360.patch.txt](https://github.com/steve8x8/geotoad/files/2343208/issue360.patch.txt)
| 1.0 | Logbook (json) comes gzipped - Starting around September 1, logbook entries (for log counts > 10 only) are delivered in gzipped form, but the decompression part was still missing from the code.
The patch is simple: just comment out the "Accept-Encoding" setting in lib/logbook.rb - a better solution (adding decompression to lib/shadowget.rb) is being worked on.
[issue360.patch.txt](https://github.com/steve8x8/geotoad/files/2343208/issue360.patch.txt)
| test | logbook json comes gzipped starting around september logbook entries for log counts only are delivered in gzipped form but the decompression part was still missing from the code the patch is simple just comment out the accept encoding setting in lib logbook rb a better solution adding decompression to lib shadowget rb is being worked on | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.