Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
288,673
| 24,926,004,878
|
IssuesEvent
|
2022-10-31 07:25:47
|
kubernetes-sigs/kubespray
|
https://api.github.com/repos/kubernetes-sigs/kubespray
|
opened
|
Vagrant jobs are broken due to OS detection logic
|
kind/failing-test
|
<!-- Please only use this template for submitting reports about failing tests in Kubespray CI jobs -->
**Which jobs are failing**:
Sample: https://gitlab.com/kargo-ci/kubernetes-sigs-kubespray/-/jobs/3227705561
**Which test(s) are failing**:
Vagrant tests:
* vagrant_fedora35-kube-router
* vagrant_ubuntu18-calico-dual-stack
* vagrant_ubuntu18-flannel
* vagrant_ubuntu20-flannel
**Since when has it been failing**:
Likely the breakage was introduced by https://github.com/kubernetes-sigs/kubespray/pull/9416 and perpetuated by https://github.com/kubernetes-sigs/kubespray/pull/9432
**Testgrid link**:
N/A
**Reason for failure**:
Overriding ansible detected variables could have unforeseen side effects and should never be done. This looks like we need a better way to detect and account for operating system differences.
**Anything else we need to know**:
To fix the CI I propose we temporarily revert the two changes while the developers figure out a better solution for the problems they were trying to address with the changes they proposed there.
|
1.0
|
Vagrant jobs are broken due to OS detection logic - <!-- Please only use this template for submitting reports about failing tests in Kubespray CI jobs -->
**Which jobs are failing**:
Sample: https://gitlab.com/kargo-ci/kubernetes-sigs-kubespray/-/jobs/3227705561
**Which test(s) are failing**:
Vagrant tests:
* vagrant_fedora35-kube-router
* vagrant_ubuntu18-calico-dual-stack
* vagrant_ubuntu18-flannel
* vagrant_ubuntu20-flannel
**Since when has it been failing**:
Likely the breakage was introduced by https://github.com/kubernetes-sigs/kubespray/pull/9416 and perpetuated by https://github.com/kubernetes-sigs/kubespray/pull/9432
**Testgrid link**:
N/A
**Reason for failure**:
Overriding ansible detected variables could have unforeseen side effects and should never be done. This looks like we need a better way to detect and account for operating system differences.
**Anything else we need to know**:
To fix the CI I propose we temporarily revert the two changes while the developers figure out a better solution for the problems they were trying to address with the changes they proposed there.
|
test
|
vagrant jobs are broken due to os detection logic which jobs are failing sample which test s are failing vagrant tests vagrant kube router vagrant calico dual stack vagrant flannel vagrant flannel since when has it been failing likely the breakage was introduced by and perpetuated by testgrid link n a reason for failure overriding ansible detected variables could have unforeseen side effects and should never be done this looks like we need a better way to detect and account for operating system differences anything else we need to know to fix the ci i propose we temporarily revert the two changes while the developers figure out a better solution for the problems they were trying to address with the changes they proposed there
| 1
|
65,355
| 14,713,728,852
|
IssuesEvent
|
2021-01-05 10:49:03
|
sourcegraph/sourcegraph
|
https://api.github.com/repos/sourcegraph/sourcegraph
|
closed
|
Triage and fix vulnerabilities, ensure important vulnerabilities fixed
|
pre-cloud-GA-scan team/security
|
This should be graduated into a tracking issue once scans are implemented.
|
True
|
Triage and fix vulnerabilities, ensure important vulnerabilities fixed - This should be graduated into a tracking issue once scans are implemented.
|
non_test
|
triage and fix vulnerabilities ensure important vulnerabilities fixed this should be graduated into a tracking issue once scans are implemented
| 0
|
336,862
| 30,225,074,602
|
IssuesEvent
|
2023-07-05 23:16:15
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix jax_nn_activations.test_jax_nn_swish
|
JAX Frontend Sub Task Failing Test
|
| | |
|---|---|
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5442531781"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5371448716/jobs/9744217013"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5371448716/jobs/9744217013"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5427202833"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5469565117"><img src=https://img.shields.io/badge/-failure-red></a>
|
1.0
|
Fix jax_nn_activations.test_jax_nn_swish - | | |
|---|---|
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5442531781"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5371448716/jobs/9744217013"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5371448716/jobs/9744217013"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5427202833"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5469565117"><img src=https://img.shields.io/badge/-failure-red></a>
|
test
|
fix jax nn activations test jax nn swish jax a href src numpy a href src tensorflow a href src torch a href src paddle a href src
| 1
|
66,681
| 8,038,361,397
|
IssuesEvent
|
2018-07-30 15:11:47
|
Opentrons/opentrons
|
https://api.github.com/repos/Opentrons/opentrons
|
opened
|
300 uL tip racks are referred to as 200 uL tip racks
|
bug protocol designer
|
Currently we are referring to ur 300 uL tip racks as 200 uL tip racks in both run app and PD.
|
1.0
|
300 uL tip racks are referred to as 200 uL tip racks - Currently we are referring to ur 300 uL tip racks as 200 uL tip racks in both run app and PD.
|
non_test
|
ul tip racks are referred to as ul tip racks currently we are referring to ur ul tip racks as ul tip racks in both run app and pd
| 0
|
220,609
| 17,210,984,937
|
IssuesEvent
|
2021-07-19 04:20:04
|
microsoft/BotFramework-Composer
|
https://api.github.com/repos/microsoft/BotFramework-Composer
|
closed
|
Problems pane defaults settings should be to show both errors and warning and all bots
|
P0 R14 Bugbash Testing & Debugging Type: Bug
|
<!-- Please search for your feature request before creating a new one. >
<!-- Complete the necessary portions of this template and delete the rest. -->
## Describe the bug
The Problems pane should be not remember settings when you switch bots. It should reset to show all errors, all warnings, and for all bots each time you open a bot in Composer.
## Version
2.0.0-nightly.258070.287a4dc
## Browser
<!-- What browser are you using? -->
- [x] Electron distribution
- [ ] Chrome
- [ ] Safari
- [ ] Firefox
- [ ] Edge
## OS
<!-- What operating system are you using? -->
- [x] macOS
- [ ] Windows
- [ ] Ubuntu
## To Reproduce
Steps to reproduce the behavior:
1. Open any bot that does not have multiple bots.
2. Go back to the home page.
4. Open an enterprise assistant bot.
5. Open Problems pane and see that the pane is filtered to just the root bot.
## Expected behavior
The Problems pane defaults should be reset to show everything (errors, warning, all bots) whenever I (re-)open a bot.
## Screenshots
<!-- If applicable, add screenshots/gif/video to help explain your problem. -->
https://user-images.githubusercontent.com/39317542/125354199-5fbeeb80-e318-11eb-9e6c-fed556645038.mov
|
1.0
|
Problems pane defaults settings should be to show both errors and warning and all bots - <!-- Please search for your feature request before creating a new one. >
<!-- Complete the necessary portions of this template and delete the rest. -->
## Describe the bug
The Problems pane should be not remember settings when you switch bots. It should reset to show all errors, all warnings, and for all bots each time you open a bot in Composer.
## Version
2.0.0-nightly.258070.287a4dc
## Browser
<!-- What browser are you using? -->
- [x] Electron distribution
- [ ] Chrome
- [ ] Safari
- [ ] Firefox
- [ ] Edge
## OS
<!-- What operating system are you using? -->
- [x] macOS
- [ ] Windows
- [ ] Ubuntu
## To Reproduce
Steps to reproduce the behavior:
1. Open any bot that does not have multiple bots.
2. Go back to the home page.
4. Open an enterprise assistant bot.
5. Open Problems pane and see that the pane is filtered to just the root bot.
## Expected behavior
The Problems pane defaults should be reset to show everything (errors, warning, all bots) whenever I (re-)open a bot.
## Screenshots
<!-- If applicable, add screenshots/gif/video to help explain your problem. -->
https://user-images.githubusercontent.com/39317542/125354199-5fbeeb80-e318-11eb-9e6c-fed556645038.mov
|
test
|
problems pane defaults settings should be to show both errors and warning and all bots describe the bug the problems pane should be not remember settings when you switch bots it should reset to show all errors all warnings and for all bots each time you open a bot in composer version nightly browser electron distribution chrome safari firefox edge os macos windows ubuntu to reproduce steps to reproduce the behavior open any bot that does not have multiple bots go back to the home page open an enterprise assistant bot open problems pane and see that the pane is filtered to just the root bot expected behavior the problems pane defaults should be reset to show everything errors warning all bots whenever i re open a bot screenshots
| 1
|
18,909
| 24,847,873,808
|
IssuesEvent
|
2022-10-26 17:21:42
|
rladstaetter/LogoRRR
|
https://api.github.com/repos/rladstaetter/LogoRRR
|
closed
|
Linux Support / Build
|
release process
|
Build LogoRRR on Linux and solve possible problems. Integrate Linux as supported Platform in Release.
|
1.0
|
Linux Support / Build - Build LogoRRR on Linux and solve possible problems. Integrate Linux as supported Platform in Release.
|
non_test
|
linux support build build logorrr on linux and solve possible problems integrate linux as supported platform in release
| 0
|
40,491
| 6,826,842,274
|
IssuesEvent
|
2017-11-08 15:20:49
|
openmpf/openmpf
|
https://api.github.com/repos/openmpf/openmpf
|
closed
|
Update Component API docs with Feed Forward for Generic Detections
|
documentation
|
Related to #326.
Update Component API docs with Feed Forward for Generic Detections.
|
1.0
|
Update Component API docs with Feed Forward for Generic Detections - Related to #326.
Update Component API docs with Feed Forward for Generic Detections.
|
non_test
|
update component api docs with feed forward for generic detections related to update component api docs with feed forward for generic detections
| 0
|
218,636
| 17,012,228,606
|
IssuesEvent
|
2021-07-02 07:01:21
|
mozilla-mobile/focus-android
|
https://api.github.com/repos/mozilla-mobile/focus-android
|
closed
|
Selecting 'Find in page' does not show the dialog when the dialog is
|
P3 bug size L testing
|
### Steps to reproduce
- Open webpage, open find in page dialog
- Scroll up, so the find in page dialog disappears from screen
- Go to menu, and select find in page menu item again
### Expected behavior
- Find in page dialog is in focus
### Actual behavior
- Find in page dialog is not shown on screen
### Device information
- Nexus 4, API 23
|
1.0
|
Selecting 'Find in page' does not show the dialog when the dialog is -
### Steps to reproduce
- Open webpage, open find in page dialog
- Scroll up, so the find in page dialog disappears from screen
- Go to menu, and select find in page menu item again
### Expected behavior
- Find in page dialog is in focus
### Actual behavior
- Find in page dialog is not shown on screen
### Device information
- Nexus 4, API 23
|
test
|
selecting find in page does not show the dialog when the dialog is steps to reproduce open webpage open find in page dialog scroll up so the find in page dialog disappears from screen go to menu and select find in page menu item again expected behavior find in page dialog is in focus actual behavior find in page dialog is not shown on screen device information nexus api
| 1
|
288,065
| 24,882,269,760
|
IssuesEvent
|
2022-10-28 03:01:05
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Orçamento - Legislação - Campestre
|
generalization test development template - Memory (66) tag - Orçamento subtag - Legislação
|
DoD: Realizar o teste de Generalização do validador da tag Orçamento - Legislação para o Município de Campestre.
|
1.0
|
Teste de generalizacao para a tag Orçamento - Legislação - Campestre - DoD: Realizar o teste de Generalização do validador da tag Orçamento - Legislação para o Município de Campestre.
|
test
|
teste de generalizacao para a tag orçamento legislação campestre dod realizar o teste de generalização do validador da tag orçamento legislação para o município de campestre
| 1
|
132,832
| 12,519,537,617
|
IssuesEvent
|
2020-06-03 14:35:05
|
GluuFederation/community-edition-setup
|
https://api.github.com/repos/GluuFederation/community-edition-setup
|
closed
|
Migrate existing u2f enrollments to the fido2 branch
|
Needs Documentation Needs QA enhancement
|
Consolidate u2f and fido2 enrollments in a single place.
That way admins can employ a single custom script (fido2) so that old u2f credentials will still work
This will also add some usability benefits for casa.
|
1.0
|
Migrate existing u2f enrollments to the fido2 branch - Consolidate u2f and fido2 enrollments in a single place.
That way admins can employ a single custom script (fido2) so that old u2f credentials will still work
This will also add some usability benefits for casa.
|
non_test
|
migrate existing enrollments to the branch consolidate and enrollments in a single place that way admins can employ a single custom script so that old credentials will still work this will also add some usability benefits for casa
| 0
|
20,775
| 3,634,163,958
|
IssuesEvent
|
2016-02-11 17:00:51
|
SIB-Colombia/cygnus
|
https://api.github.com/repos/SIB-Colombia/cygnus
|
closed
|
Eliminar el marco polaroid
|
design
|
Las imágenes del preview de las fichas poseen un marco que no va con el diseño acordado. Hay que eliminarlo.

|
1.0
|
Eliminar el marco polaroid - Las imágenes del preview de las fichas poseen un marco que no va con el diseño acordado. Hay que eliminarlo.

|
non_test
|
eliminar el marco polaroid las imágenes del preview de las fichas poseen un marco que no va con el diseño acordado hay que eliminarlo
| 0
|
108,985
| 23,691,457,590
|
IssuesEvent
|
2022-08-29 11:12:15
|
pkgjs/parseargs
|
https://api.github.com/repos/pkgjs/parseargs
|
closed
|
How might I add my own support for negative numbers?
|
bring-your-own-code
|
(This is a bring-your-own-feature experiment, not a prototype of parseArgs implementation.)
Inspired by discussion in #62. How could I add my own support for options and positionals treating negative numbers as ordinary args and not options?
```js
const { parseArgs } = require('@pkgjs/parseArgs');
const kNegativePrefix = 'parseArgs.NEGATIVE:';
// preprocess
const rawArgs = process.argv.slice(2);
const preparedArgs = rawArgs.map(arg => /^-[0-9]+/.test(arg) ? kNegativePrefix.concat(arg) : arg);
const result = parseArgs({
args: preparedArgs, {
options: { profit: { type: 'string' }}
});
// postprocess
const stripPrefix = (arg) => arg.startsWith(kNegativePrefix) ? arg.slice(kNegativePrefix.length) : arg;
result.positionals = result.positionals.map(arg => stripPrefix(arg));
Object.entries(result.values).forEach(([key, value]) => {
if (typeof value === 'string')
result.values[key] = stripPrefix(value);
});
console.log(result);
```
```
% node index.js --profit 33 44
{
flags: { profit: true },
values: { profit: '33' },
positionals: [ '44' ]
}
% node index.js --profit -33 -44
{
flags: { profit: true },
values: { profit: '-33' },
positionals: [ '-44' ]
}
```
|
1.0
|
How might I add my own support for negative numbers? - (This is a bring-your-own-feature experiment, not a prototype of parseArgs implementation.)
Inspired by discussion in #62. How could I add my own support for options and positionals treating negative numbers as ordinary args and not options?
```js
const { parseArgs } = require('@pkgjs/parseArgs');
const kNegativePrefix = 'parseArgs.NEGATIVE:';
// preprocess
const rawArgs = process.argv.slice(2);
const preparedArgs = rawArgs.map(arg => /^-[0-9]+/.test(arg) ? kNegativePrefix.concat(arg) : arg);
const result = parseArgs({
args: preparedArgs, {
options: { profit: { type: 'string' }}
});
// postprocess
const stripPrefix = (arg) => arg.startsWith(kNegativePrefix) ? arg.slice(kNegativePrefix.length) : arg;
result.positionals = result.positionals.map(arg => stripPrefix(arg));
Object.entries(result.values).forEach(([key, value]) => {
if (typeof value === 'string')
result.values[key] = stripPrefix(value);
});
console.log(result);
```
```
% node index.js --profit 33 44
{
flags: { profit: true },
values: { profit: '33' },
positionals: [ '44' ]
}
% node index.js --profit -33 -44
{
flags: { profit: true },
values: { profit: '-33' },
positionals: [ '-44' ]
}
```
|
non_test
|
how might i add my own support for negative numbers this is a bring your own feature experiment not a prototype of parseargs implementation inspired by discussion in how could i add my own support for options and positionals treating negative numbers as ordinary args and not options js const parseargs require pkgjs parseargs const knegativeprefix parseargs negative preprocess const rawargs process argv slice const preparedargs rawargs map arg test arg knegativeprefix concat arg arg const result parseargs args preparedargs options profit type string postprocess const stripprefix arg arg startswith knegativeprefix arg slice knegativeprefix length arg result positionals result positionals map arg stripprefix arg object entries result values foreach if typeof value string result values stripprefix value console log result node index js profit flags profit true values profit positionals node index js profit flags profit true values profit positionals
| 0
|
11,997
| 3,249,234,336
|
IssuesEvent
|
2015-10-18 00:24:10
|
jQueryGeo/geo
|
https://api.github.com/repos/jQueryGeo/geo
|
closed
|
find should return shapes in reverse order
|
2 - Medium Test With Latest
|
Shapes are drawn in FIFO meaning shapes appended later are drawn on top of shapes appended before. However, the find method searches the same way, finding a large shape first if it was appended first; which it would have been because you'd have to append the smaller shape second to see it on top.
|
2.0
|
find should return shapes in reverse order - Shapes are drawn in FIFO meaning shapes appended later are drawn on top of shapes appended before. However, the find method searches the same way, finding a large shape first if it was appended first; which it would have been because you'd have to append the smaller shape second to see it on top.
|
test
|
find should return shapes in reverse order shapes are drawn in fifo meaning shapes appended later are drawn on top of shapes appended before however the find method searches the same way finding a large shape first if it was appended first which it would have been because you d have to append the smaller shape second to see it on top
| 1
|
669,970
| 22,647,861,459
|
IssuesEvent
|
2022-07-01 10:28:11
|
theAsmodai/metamod-r
|
https://api.github.com/repos/theAsmodai/metamod-r
|
closed
|
Problem compiling with ICC 19
|
Type: Bug Priority: Medium Status: Available OS: Independent
|
https://github.com/theAsmodai/metamod-r/blob/0cf2f709dbeae18ca84d2fafd4481ffbba06ad0c/metamod/src/sys_module.cpp#L145
Is it ok to use "false" here? Looks like ICC don't like that.
/home/user/build/metamod-r/metamod/src/sys_module.cpp(145): error: return value type does not match the function type
return false;
^
|
1.0
|
Problem compiling with ICC 19 - https://github.com/theAsmodai/metamod-r/blob/0cf2f709dbeae18ca84d2fafd4481ffbba06ad0c/metamod/src/sys_module.cpp#L145
Is it ok to use "false" here? Looks like ICC don't like that.
/home/user/build/metamod-r/metamod/src/sys_module.cpp(145): error: return value type does not match the function type
return false;
^
|
non_test
|
problem compiling with icc is it ok to use false here looks like icc don t like that home user build metamod r metamod src sys module cpp error return value type does not match the function type return false
| 0
|
539,424
| 15,788,379,009
|
IssuesEvent
|
2021-04-01 20:42:09
|
CanberraOceanRacingClub/namadgi3
|
https://api.github.com/repos/CanberraOceanRacingClub/namadgi3
|
opened
|
Docking at Holme Port
|
Notice priority 2: Medium
|
Skippers are asked to consider the following procedure when docking at Holme Port:
While **slowly reversing** into the pen:
1. Attach ``Stern Spring`` to Namadgi's midships cleat -- continue reversing to tension the line
2. Attach bow line to dock cleat and secure after adjusting to length
3. Attach port stern line to dock cleat
4. Attach starboard stern line to dock cleat
When departing **PLEASE leave the ``Stern Spring`` on the dock**, attached to it's cleat, ready for reuse.

|
1.0
|
Docking at Holme Port - Skippers are asked to consider the following procedure when docking at Holme Port:
While **slowly reversing** into the pen:
1. Attach ``Stern Spring`` to Namadgi's midships cleat -- continue reversing to tension the line
2. Attach bow line to dock cleat and secure after adjusting to length
3. Attach port stern line to dock cleat
4. Attach starboard stern line to dock cleat
When departing **PLEASE leave the ``Stern Spring`` on the dock**, attached to it's cleat, ready for reuse.

|
non_test
|
docking at holme port skippers are asked to consider the following procedure when docking at holme port while slowly reversing into the pen attach stern spring to namadgi s midships cleat continue reversing to tension the line attach bow line to dock cleat and secure after adjusting to length attach port stern line to dock cleat attach starboard stern line to dock cleat when departing please leave the stern spring on the dock attached to it s cleat ready for reuse
| 0
|
258,557
| 22,327,855,473
|
IssuesEvent
|
2022-06-14 12:16:57
|
Uuvana-Studios/longvinter-windows-client
|
https://api.github.com/repos/Uuvana-Studios/longvinter-windows-client
|
closed
|
When Opening chests, cooked fish looses its energy symbol
|
Bug Low Priority Tested
|
**Describe the bug**
In the open chest menu, the symbol for cooked fish and raw fish is the same. The energy symbol added on the cooked fish disappears. You need to hover over the fish to find out if it;'s cooked or not
**To Reproduce**
Steps to reproduce the behavior:
1. Have cooked fish in inventory'
2. Open a chest
3. the inventory menu for both character bag and chest does not show the energy symbol for cooked fish
**Expected behavior**
the energy symbol for cooked fish should be retained when opening chests to easily differentiate between cooked and raw
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Desktop (please complete the following information):**
- OS: Windows 11
- Game Version Beta 1.0.0
- Steam Version v020
**Additional context**
Add any other context about the problem here.
|
1.0
|
When Opening chests, cooked fish looses its energy symbol - **Describe the bug**
In the open chest menu, the symbol for cooked fish and raw fish is the same. The energy symbol added on the cooked fish disappears. You need to hover over the fish to find out if it;'s cooked or not
**To Reproduce**
Steps to reproduce the behavior:
1. Have cooked fish in inventory'
2. Open a chest
3. the inventory menu for both character bag and chest does not show the energy symbol for cooked fish
**Expected behavior**
the energy symbol for cooked fish should be retained when opening chests to easily differentiate between cooked and raw
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Desktop (please complete the following information):**
- OS: Windows 11
- Game Version Beta 1.0.0
- Steam Version v020
**Additional context**
Add any other context about the problem here.
|
test
|
when opening chests cooked fish looses its energy symbol describe the bug in the open chest menu the symbol for cooked fish and raw fish is the same the energy symbol added on the cooked fish disappears you need to hover over the fish to find out if it s cooked or not to reproduce steps to reproduce the behavior have cooked fish in inventory open a chest the inventory menu for both character bag and chest does not show the energy symbol for cooked fish expected behavior the energy symbol for cooked fish should be retained when opening chests to easily differentiate between cooked and raw screenshots if applicable add screenshots to help explain your problem desktop please complete the following information os windows game version beta steam version additional context add any other context about the problem here
| 1
|
239,318
| 19,846,918,560
|
IssuesEvent
|
2022-01-21 07:49:33
|
Oldes/Rebol-issues
|
https://api.github.com/repos/Oldes/Rebol-issues
|
closed
|
TRIM /head and /tail on block
|
Test.written Type.bug Datatype: block! CC.resolved
|
_Submitted by:_ **Carl**
Gives an error, but should be valid.
``` rebol
>> trim/head reduce [none 1 none 2]
== [1 none 2]
>> trim/tail reduce [1 none 2 none]
== [1 none 2]
```
---
<sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=825)** [ Version: alpha 54 Type: Bug Platform: All Category: n/a Reproduce: Always Fixed-in:alpha 55 ]</sup>
<sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/825</sup>
Comments:
---
---
> **Rebolbot** added the **Type.bug** on Jan 12, 2016
---
|
1.0
|
TRIM /head and /tail on block - _Submitted by:_ **Carl**
Gives an error, but should be valid.
``` rebol
>> trim/head reduce [none 1 none 2]
== [1 none 2]
>> trim/tail reduce [1 none 2 none]
== [1 none 2]
```
---
<sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=825)** [ Version: alpha 54 Type: Bug Platform: All Category: n/a Reproduce: Always Fixed-in:alpha 55 ]</sup>
<sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/825</sup>
Comments:
---
---
> **Rebolbot** added the **Type.bug** on Jan 12, 2016
---
|
test
|
trim head and tail on block submitted by carl gives an error but should be valid rebol trim head reduce trim tail reduce imported from imported from comments rebolbot added the type bug on jan
| 1
|
305,540
| 26,391,923,255
|
IssuesEvent
|
2023-01-12 16:16:43
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix reduction_ops.test_torch_prod
|
PyTorch Frontend Sub Task Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_reduction_ops.py::test_torch_prod[cpu-ivy.functional.backends.numpy-False-False]</summary>
2023-01-12T03:47:11.4120487Z E AssertionError: the return with a torch backend produced data type of float64, while the return with a numpy backend returned a data type of float32.
2023-01-12T03:47:11.4120954Z E Falsifying example: test_torch_prod(
2023-01-12T03:47:11.4121740Z E dtype_x_axis=(['float64'], [array([-1.])], 0),
2023-01-12T03:47:11.4122064Z E dtype=[None],
2023-01-12T03:47:11.4122317Z E keepdims=False,
2023-01-12T03:47:11.4122585Z E as_variable=[False],
2023-01-12T03:47:11.4123810Z E num_positional_args=0,
2023-01-12T03:47:11.4124153Z E native_array=[False],
2023-01-12T03:47:11.4124473Z E with_out=False,
2023-01-12T03:47:11.4124942Z E fn_tree='ivy.functional.frontends.torch.prod',
2023-01-12T03:47:11.4125324Z E on_device='cpu',
2023-01-12T03:47:11.4125646Z E frontend='torch',
2023-01-12T03:47:11.4125887Z E )
2023-01-12T03:47:11.4126093Z E
2023-01-12T03:47:11.4126734Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2JkAAMoBaTZIQwAAK4ADQ==') as a decorator on your test case
</details>
|
1.0
|
Fix reduction_ops.test_torch_prod - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_reduction_ops.py::test_torch_prod[cpu-ivy.functional.backends.numpy-False-False]</summary>
2023-01-12T03:47:11.4120487Z E AssertionError: the return with a torch backend produced data type of float64, while the return with a numpy backend returned a data type of float32.
2023-01-12T03:47:11.4120954Z E Falsifying example: test_torch_prod(
2023-01-12T03:47:11.4121740Z E dtype_x_axis=(['float64'], [array([-1.])], 0),
2023-01-12T03:47:11.4122064Z E dtype=[None],
2023-01-12T03:47:11.4122317Z E keepdims=False,
2023-01-12T03:47:11.4122585Z E as_variable=[False],
2023-01-12T03:47:11.4123810Z E num_positional_args=0,
2023-01-12T03:47:11.4124153Z E native_array=[False],
2023-01-12T03:47:11.4124473Z E with_out=False,
2023-01-12T03:47:11.4124942Z E fn_tree='ivy.functional.frontends.torch.prod',
2023-01-12T03:47:11.4125324Z E on_device='cpu',
2023-01-12T03:47:11.4125646Z E frontend='torch',
2023-01-12T03:47:11.4125887Z E )
2023-01-12T03:47:11.4126093Z E
2023-01-12T03:47:11.4126734Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2JkAAMoBaTZIQwAAK4ADQ==') as a decorator on your test case
</details>
|
test
|
fix reduction ops test torch prod tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test torch test reduction ops py test torch prod e assertionerror the return with a torch backend produced data type of while the return with a numpy backend returned a data type of e falsifying example test torch prod e dtype x axis e dtype e keepdims false e as variable e num positional args e native array e with out false e fn tree ivy functional frontends torch prod e on device cpu e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
| 1
|
73,103
| 3,407,210,493
|
IssuesEvent
|
2015-12-04 01:11:14
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
Possible regression in waiting for DOMContentLoaded
|
area-dart2js area-html Priority-Medium
|
Opening a new bug for possible regression reported in https://github.com/dart-lang/sdk/issues/885#issuecomment-161689055
@vsmenon did the original fix.
Comment from there by @joseluis is
On version 1.13.0 this doesn't seem true anymore.
Minimal example: querySelector('#someid'); returns null when the element is placed after the script tag that loads the Dart script compiled with dart2js. In order to make it work I must use the script defer attribute, like this:
```html
<html>
<head>
<meta charset="utf-8">
</head>
<body>
<script src="dart-program.js" defer></script>
<script id="someid"></script>
</body>
</html>
```
|
1.0
|
Possible regression in waiting for DOMContentLoaded - Opening a new bug for possible regression reported in https://github.com/dart-lang/sdk/issues/885#issuecomment-161689055
@vsmenon did the original fix.
Comment from there by @joseluis is
On version 1.13.0 this doesn't seem true anymore.
Minimal example: querySelector('#someid'); returns null when the element is placed after the script tag that loads the Dart script compiled with dart2js. In order to make it work I must use the script defer attribute, like this:
```html
<html>
<head>
<meta charset="utf-8">
</head>
<body>
<script src="dart-program.js" defer></script>
<script id="someid"></script>
</body>
</html>
```
|
non_test
|
possible regression in waiting for domcontentloaded opening a new bug for possible regression reported in vsmenon did the original fix comment from there by joseluis is on version this doesn t seem true anymore minimal example queryselector someid returns null when the element is placed after the script tag that loads the dart script compiled with in order to make it work i must use the script defer attribute like this html
| 0
|
253,312
| 21,673,814,904
|
IssuesEvent
|
2022-05-08 11:44:42
|
mennaelkashef/eShop
|
https://api.github.com/repos/mennaelkashef/eShop
|
opened
|
this is a message to test encryption time
|
Hello! RULE-GOT-APPLIED DOES-NOT-CONTAIN-STRING Rule-works-on-convert-to-bug test instabug
|
# :clipboard: Bug Details
>this is a message to test encryption time
key | value
--|--
Reported At | 2022-04-28 13:14:52 UTC
Email | abarakat@gmail.com
Categories | Report a bug
Tags | test, Hello!, RULE-GOT-APPLIED, DOES-NOT-CONTAIN-STRING, Rule-works-on-convert-to-bug, instabug
App Version | 1.1 (1)
Session Duration | 154
Device | OnePlus KB2001, OS Level 30
Display | 1080x2400 (xhdpi)
Location | Giza, Egypt (en)
## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?utm_source=github&utm_medium=integrations) :point_left:
___
# :iphone: View Hierarchy
This bug was reported from **com.example.app.complexviews.ComplexViewsFragment**
Find its interactive view hierarchy with all its subviews here: :point_right: **[Check View Hierarchy](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-hierarchy-view=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :chart_with_downwards_trend: Session Profiler
Here is what the app was doing right before the bug was reported:
Key | Value
--|--
Used Memory | 58.1% - 4.27/7.34 GB
Used Storage | 7.1% - 7.48/105.28 GB
Connectivity | no_connection
Battery | 100% - plugged
Orientation | portrait
Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :bust_in_silhouette: User Info
### User Attributes
```
key_name -643333045: key value bla bla bla la
key_name -1391770268: key value bla bla bla la
key_name -1233063091: key value bla bla bla la
key_name 12389232: key value bla bla bla la
key_name -1400931524: key value bla bla bla la
key_name 1545921878: key value bla bla bla la
key_name -778916395: key value bla bla bla la
key_name -1822943291: key value bla bla bla la
key_name 1537783730: key value bla bla bla la
key_name -2119608186: key value bla bla bla la
key_name -683370323: key value bla bla bla la
key_name -1977430918: key value bla bla bla la
key_name -223322809: key value bla bla bla la
```
___
# :mag_right: Logs
### User Steps
Here are the last 10 steps done by the user right before the bug was reported:
```
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 com.example.app.developerOption.list.PhotosActivity was paused.
13:14:52 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
```
Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left:
### Console Log
Here are the last 10 console logs logged right before the bug was reported:
```
13:14:56 D/StrictMode(17648): at android.view.View.access$3600(View.java:826)
13:14:56 D/StrictMode(17648): at android.view.View$PerformClick.run(View.java:28555)
13:14:56 D/StrictMode(17648): at android.os.Handler.handleCallback(Handler.java:938)
13:14:56 D/StrictMode(17648): at android.os.Handler.dispatchMessage(Handler.java:99)
13:14:56 D/StrictMode(17648): at android.os.Looper.loop(Looper.java:233)
13:14:56 D/StrictMode(17648): at android.app.ActivityThread.main(ActivityThread.java:8030)
13:14:56 D/StrictMode(17648): at java.lang.reflect.Method.invoke(Native Method)
13:14:56 D/StrictMode(17648): at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:631)
13:14:56 D/StrictMode(17648): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:978)
13:14:56 D/skia (17648): got procName com.example.app for pid 17648
```
Find all the logged console logs throughout the session here: :point_right: **[View All Console Log](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-logs=console_log&utm_source=github&utm_medium=integrations)** :point_left:
___
# :camera: Images
[](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/18282355/a308575a5f1a44fd0352ae352e28eb6e_original/25943016/bug_1651151691655_.jpg?Expires=4807683881&Signature=QOE3w0fF16ZCCYrKWHy0i-patAuo9V9lUuq2VBKREUCCFwb58gsQZGSEWdJ~jhg6ILNEBOejsCXAUpsjHrvQ4ShFtYpAhEmARJ3soeZQ0lCZfoZE9XypuunJud5IyvolXDd8GH2uBI5t1ECPKrUU7YTGqlUuFsExzBx-n2LWDHvXB8fTAvDpTRSylBYh9fIHqafB9Z2exl7tU67Yacva5Nop7IrfaK3mrlzLTrE0AKBUPc9eKIFwoG7t-Gwp2WMC-WgDAd8uZJ0vKHOiwc3HwBgzdeJVmRKUSG8yg1eEhVWDsxjBZFal42OYN81cxpwHULcDNx-nU1T5sKMQreOq7w__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)
___
# :warning: Looking for More Details?
1. **Network Log**: we are unable to capture your network requests automatically. If you are using HttpUrlConnection or Okhttp requests, [**check the details mentioned here**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations#section-network-logs).
2. **User Events**: start capturing custom User Events to send them along with each report. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations).
3. **Instabug Log**: start adding Instabug logs to see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations).
|
1.0
|
this is a message to test encryption time - # :clipboard: Bug Details
>this is a message to test encryption time
key | value
--|--
Reported At | 2022-04-28 13:14:52 UTC
Email | abarakat@gmail.com
Categories | Report a bug
Tags | test, Hello!, RULE-GOT-APPLIED, DOES-NOT-CONTAIN-STRING, Rule-works-on-convert-to-bug, instabug
App Version | 1.1 (1)
Session Duration | 154
Device | OnePlus KB2001, OS Level 30
Display | 1080x2400 (xhdpi)
Location | Giza, Egypt (en)
## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?utm_source=github&utm_medium=integrations) :point_left:
___
# :iphone: View Hierarchy
This bug was reported from **com.example.app.complexviews.ComplexViewsFragment**
Find its interactive view hierarchy with all its subviews here: :point_right: **[Check View Hierarchy](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-hierarchy-view=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :chart_with_downwards_trend: Session Profiler
Here is what the app was doing right before the bug was reported:
Key | Value
--|--
Used Memory | 58.1% - 4.27/7.34 GB
Used Storage | 7.1% - 7.48/105.28 GB
Connectivity | no_connection
Battery | 100% - plugged
Orientation | portrait
Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :bust_in_silhouette: User Info
### User Attributes
```
key_name -643333045: key value bla bla bla la
key_name -1391770268: key value bla bla bla la
key_name -1233063091: key value bla bla bla la
key_name 12389232: key value bla bla bla la
key_name -1400931524: key value bla bla bla la
key_name 1545921878: key value bla bla bla la
key_name -778916395: key value bla bla bla la
key_name -1822943291: key value bla bla bla la
key_name 1537783730: key value bla bla bla la
key_name -2119608186: key value bla bla bla la
key_name -683370323: key value bla bla bla la
key_name -1977430918: key value bla bla bla la
key_name -223322809: key value bla bla bla la
```
___
# :mag_right: Logs
### User Steps
Here are the last 10 steps done by the user right before the bug was reported:
```
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
13:14:51 com.example.app.developerOption.list.PhotosActivity was paused.
13:14:52 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity"
```
Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left:
### Console Log
Here are the last 10 console logs logged right before the bug was reported:
```
13:14:56 D/StrictMode(17648): at android.view.View.access$3600(View.java:826)
13:14:56 D/StrictMode(17648): at android.view.View$PerformClick.run(View.java:28555)
13:14:56 D/StrictMode(17648): at android.os.Handler.handleCallback(Handler.java:938)
13:14:56 D/StrictMode(17648): at android.os.Handler.dispatchMessage(Handler.java:99)
13:14:56 D/StrictMode(17648): at android.os.Looper.loop(Looper.java:233)
13:14:56 D/StrictMode(17648): at android.app.ActivityThread.main(ActivityThread.java:8030)
13:14:56 D/StrictMode(17648): at java.lang.reflect.Method.invoke(Native Method)
13:14:56 D/StrictMode(17648): at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:631)
13:14:56 D/StrictMode(17648): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:978)
13:14:56 D/skia (17648): got procName com.example.app for pid 17648
```
Find all the logged console logs throughout the session here: :point_right: **[View All Console Log](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-logs=console_log&utm_source=github&utm_medium=integrations)** :point_left:
___
# :camera: Images
[](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/18282355/a308575a5f1a44fd0352ae352e28eb6e_original/25943016/bug_1651151691655_.jpg?Expires=4807683881&Signature=QOE3w0fF16ZCCYrKWHy0i-patAuo9V9lUuq2VBKREUCCFwb58gsQZGSEWdJ~jhg6ILNEBOejsCXAUpsjHrvQ4ShFtYpAhEmARJ3soeZQ0lCZfoZE9XypuunJud5IyvolXDd8GH2uBI5t1ECPKrUU7YTGqlUuFsExzBx-n2LWDHvXB8fTAvDpTRSylBYh9fIHqafB9Z2exl7tU67Yacva5Nop7IrfaK3mrlzLTrE0AKBUPc9eKIFwoG7t-Gwp2WMC-WgDAd8uZJ0vKHOiwc3HwBgzdeJVmRKUSG8yg1eEhVWDsxjBZFal42OYN81cxpwHULcDNx-nU1T5sKMQreOq7w__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)
___
# :warning: Looking for More Details?
1. **Network Log**: we are unable to capture your network requests automatically. If you are using HttpUrlConnection or Okhttp requests, [**check the details mentioned here**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations#section-network-logs).
2. **User Events**: start capturing custom User Events to send them along with each report. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations).
3. **Instabug Log**: start adding Instabug logs to see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations).
|
test
|
this is a message to test encryption time clipboard bug details this is a message to test encryption time key value reported at utc email abarakat gmail com categories report a bug tags test hello rule got applied does not contain string rule works on convert to bug instabug app version session duration device oneplus os level display xhdpi location giza egypt en point right point left iphone view hierarchy this bug was reported from com example app complexviews complexviewsfragment find its interactive view hierarchy with all its subviews here point right point left chart with downwards trend session profiler here is what the app was doing right before the bug was reported key value used memory gb used storage gb connectivity no connection battery plugged orientation portrait find all the changes that happened in the parameters mentioned above during the last seconds before the bug was reported here point right point left bust in silhouette user info user attributes key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la mag right logs user steps here are the last steps done by the user right before the bug was reported tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity com example app developeroption list photosactivity was paused tap in android widget linearlayout in com example app developeroption list photosactivity find all the user steps done by the user throughout the session here point right point left console log here are the last console logs logged right before the bug was reported d strictmode at android view view access view java d strictmode at android view view performclick run view java d strictmode at android os handler handlecallback handler java d strictmode at android os handler dispatchmessage handler java d strictmode at android os looper loop looper java d strictmode at android app activitythread main activitythread java d strictmode at java lang reflect method invoke native method d strictmode at com android internal os runtimeinit methodandargscaller run runtimeinit java d strictmode at com android internal os zygoteinit main zygoteinit java d skia got procname com example app for pid find all the logged console logs throughout the session here point right point left camera images warning looking for more details network log we are unable to capture your network requests automatically if you are using httpurlconnection or okhttp requests user events start capturing custom user events to send them along with each report instabug log start adding instabug logs to see them right inside each report you receive
| 1
|
123,967
| 17,772,422,600
|
IssuesEvent
|
2021-08-30 15:03:52
|
kapseliboi/mimic
|
https://api.github.com/repos/kapseliboi/mimic
|
opened
|
CVE-2017-16137 (Medium) detected in multiple libraries
|
security vulnerability
|
## CVE-2017-16137 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>debug-2.6.7.tgz</b>, <b>debug-2.2.0.tgz</b>, <b>debug-2.3.3.tgz</b>, <b>debug-2.6.8.tgz</b></p></summary>
<p>
<details><summary><b>debug-2.6.7.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.7.tgz">https://registry.npmjs.org/debug/-/debug-2.6.7.tgz</a></p>
<p>Path to dependency file: mimic/package.json</p>
<p>Path to vulnerable library: mimic/node_modules/express/node_modules/debug/package.json,mimic/node_modules/body-parser/node_modules/debug/package.json,mimic/node_modules/connect/node_modules/debug/package.json,mimic/node_modules/send/node_modules/debug/package.json,mimic/node_modules/finalhandler/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.0.tgz (Root Library)
- connect-3.6.2.tgz
- :x: **debug-2.6.7.tgz** (Vulnerable Library)
</details>
<details><summary><b>debug-2.2.0.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p>
<p>Path to dependency file: mimic/package.json</p>
<p>Path to vulnerable library: mimic/node_modules/socket.io-parser/node_modules/debug/package.json,mimic/node_modules/mocha-nightwatch/node_modules/debug/package.json,mimic/node_modules/compression/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.0.tgz (Root Library)
- socket.io-1.7.3.tgz
- socket.io-parser-2.3.1.tgz
- :x: **debug-2.2.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>debug-2.3.3.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.3.3.tgz">https://registry.npmjs.org/debug/-/debug-2.3.3.tgz</a></p>
<p>Path to dependency file: mimic/package.json</p>
<p>Path to vulnerable library: mimic/node_modules/engine.io/node_modules/debug/package.json,mimic/node_modules/socket.io/node_modules/debug/package.json,mimic/node_modules/socket.io-client/node_modules/debug/package.json,mimic/node_modules/socket.io-adapter/node_modules/debug/package.json,mimic/node_modules/engine.io-client/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.0.tgz (Root Library)
- socket.io-1.7.3.tgz
- engine.io-1.8.3.tgz
- :x: **debug-2.3.3.tgz** (Vulnerable Library)
</details>
<details><summary><b>debug-2.6.8.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.8.tgz">https://registry.npmjs.org/debug/-/debug-2.6.8.tgz</a></p>
<p>Path to dependency file: mimic/package.json</p>
<p>Path to vulnerable library: mimic/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- babel-core-6.25.0.tgz (Root Library)
- :x: **debug-2.6.8.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/mimic/commit/1b3df8b0fb992059d70591ed86206dbf53aec93d">1b3df8b0fb992059d70591ed86206dbf53aec93d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 2.6.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-16137 (Medium) detected in multiple libraries - ## CVE-2017-16137 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>debug-2.6.7.tgz</b>, <b>debug-2.2.0.tgz</b>, <b>debug-2.3.3.tgz</b>, <b>debug-2.6.8.tgz</b></p></summary>
<p>
<details><summary><b>debug-2.6.7.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.7.tgz">https://registry.npmjs.org/debug/-/debug-2.6.7.tgz</a></p>
<p>Path to dependency file: mimic/package.json</p>
<p>Path to vulnerable library: mimic/node_modules/express/node_modules/debug/package.json,mimic/node_modules/body-parser/node_modules/debug/package.json,mimic/node_modules/connect/node_modules/debug/package.json,mimic/node_modules/send/node_modules/debug/package.json,mimic/node_modules/finalhandler/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.0.tgz (Root Library)
- connect-3.6.2.tgz
- :x: **debug-2.6.7.tgz** (Vulnerable Library)
</details>
<details><summary><b>debug-2.2.0.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p>
<p>Path to dependency file: mimic/package.json</p>
<p>Path to vulnerable library: mimic/node_modules/socket.io-parser/node_modules/debug/package.json,mimic/node_modules/mocha-nightwatch/node_modules/debug/package.json,mimic/node_modules/compression/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.0.tgz (Root Library)
- socket.io-1.7.3.tgz
- socket.io-parser-2.3.1.tgz
- :x: **debug-2.2.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>debug-2.3.3.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.3.3.tgz">https://registry.npmjs.org/debug/-/debug-2.3.3.tgz</a></p>
<p>Path to dependency file: mimic/package.json</p>
<p>Path to vulnerable library: mimic/node_modules/engine.io/node_modules/debug/package.json,mimic/node_modules/socket.io/node_modules/debug/package.json,mimic/node_modules/socket.io-client/node_modules/debug/package.json,mimic/node_modules/socket.io-adapter/node_modules/debug/package.json,mimic/node_modules/engine.io-client/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.0.tgz (Root Library)
- socket.io-1.7.3.tgz
- engine.io-1.8.3.tgz
- :x: **debug-2.3.3.tgz** (Vulnerable Library)
</details>
<details><summary><b>debug-2.6.8.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.8.tgz">https://registry.npmjs.org/debug/-/debug-2.6.8.tgz</a></p>
<p>Path to dependency file: mimic/package.json</p>
<p>Path to vulnerable library: mimic/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- babel-core-6.25.0.tgz (Root Library)
- :x: **debug-2.6.8.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/mimic/commit/1b3df8b0fb992059d70591ed86206dbf53aec93d">1b3df8b0fb992059d70591ed86206dbf53aec93d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 2.6.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries debug tgz debug tgz debug tgz debug tgz debug tgz small debugging utility library home page a href path to dependency file mimic package json path to vulnerable library mimic node modules express node modules debug package json mimic node modules body parser node modules debug package json mimic node modules connect node modules debug package json mimic node modules send node modules debug package json mimic node modules finalhandler node modules debug package json dependency hierarchy karma tgz root library connect tgz x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file mimic package json path to vulnerable library mimic node modules socket io parser node modules debug package json mimic node modules mocha nightwatch node modules debug package json mimic node modules compression node modules debug package json dependency hierarchy karma tgz root library socket io tgz socket io parser tgz x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file mimic package json path to vulnerable library mimic node modules engine io node modules debug package json mimic node modules socket io node modules debug package json mimic node modules socket io client node modules debug package json mimic node modules socket io adapter node modules debug package json mimic node modules engine io client node modules debug package json dependency hierarchy karma tgz root library socket io tgz engine io tgz x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file mimic package json path to vulnerable library mimic node modules debug package json dependency hierarchy babel core tgz root library x debug tgz vulnerable library found in head commit a href found in base branch master vulnerability details the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
22,517
| 3,787,948,146
|
IssuesEvent
|
2016-03-21 13:01:07
|
studentinsights/studentinsights
|
https://api.github.com/repos/studentinsights/studentinsights
|
opened
|
Full case history - section breaks across school years
|
design enhancement help wanted
|
This is with demo data, so it's not the best example, but with all the absence and tardy data it's hard to see milestone dates like the beginning of a new school year.
<img width="600" alt="screen shot 2016-03-21 at 8 59 19 am" src="https://cloud.githubusercontent.com/assets/1056957/13919335/36248ca4-ef43-11e5-825b-b7d2f14acd71.png">
One idea might just be adding a heading for the start of each school year, or perhaps using the box styling for sections that @bw120 is awesomely trying out in https://github.com/studentinsights/studentinsights/pull/199.
|
1.0
|
Full case history - section breaks across school years - This is with demo data, so it's not the best example, but with all the absence and tardy data it's hard to see milestone dates like the beginning of a new school year.
<img width="600" alt="screen shot 2016-03-21 at 8 59 19 am" src="https://cloud.githubusercontent.com/assets/1056957/13919335/36248ca4-ef43-11e5-825b-b7d2f14acd71.png">
One idea might just be adding a heading for the start of each school year, or perhaps using the box styling for sections that @bw120 is awesomely trying out in https://github.com/studentinsights/studentinsights/pull/199.
|
non_test
|
full case history section breaks across school years this is with demo data so it s not the best example but with all the absence and tardy data it s hard to see milestone dates like the beginning of a new school year img width alt screen shot at am src one idea might just be adding a heading for the start of each school year or perhaps using the box styling for sections that is awesomely trying out in
| 0
|
381,508
| 11,276,550,586
|
IssuesEvent
|
2020-01-14 23:35:39
|
googleapis/google-api-java-client-services
|
https://api.github.com/repos/googleapis/google-api-java-client-services
|
closed
|
Synthesis failed for redis
|
autosynth failure priority: p1 type: bug
|
Hello! Autosynth couldn't regenerate redis. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Checking out files: 24% (16109/65361)
Checking out files: 25% (16341/65361)
Checking out files: 26% (16994/65361)
Checking out files: 27% (17648/65361)
Checking out files: 28% (18302/65361)
Checking out files: 29% (18955/65361)
Checking out files: 30% (19609/65361)
Checking out files: 31% (20262/65361)
Checking out files: 32% (20916/65361)
Checking out files: 33% (21570/65361)
Checking out files: 34% (22223/65361)
Checking out files: 35% (22877/65361)
Checking out files: 36% (23530/65361)
Checking out files: 37% (24184/65361)
Checking out files: 38% (24838/65361)
Checking out files: 39% (25491/65361)
Checking out files: 40% (26145/65361)
Checking out files: 41% (26799/65361)
Checking out files: 42% (27452/65361)
Checking out files: 43% (28106/65361)
Checking out files: 44% (28759/65361)
Checking out files: 45% (29413/65361)
Checking out files: 46% (30067/65361)
Checking out files: 47% (30720/65361)
Checking out files: 48% (31374/65361)
Checking out files: 49% (32027/65361)
Checking out files: 50% (32681/65361)
Checking out files: 51% (33335/65361)
Checking out files: 52% (33988/65361)
Checking out files: 52% (34477/65361)
Checking out files: 53% (34642/65361)
Checking out files: 54% (35295/65361)
Checking out files: 55% (35949/65361)
Checking out files: 56% (36603/65361)
Checking out files: 57% (37256/65361)
Checking out files: 58% (37910/65361)
Checking out files: 59% (38563/65361)
Checking out files: 60% (39217/65361)
Checking out files: 61% (39871/65361)
Checking out files: 62% (40524/65361)
Checking out files: 63% (41178/65361)
Checking out files: 64% (41832/65361)
Checking out files: 65% (42485/65361)
Checking out files: 66% (43139/65361)
Checking out files: 67% (43792/65361)
Checking out files: 68% (44446/65361)
Checking out files: 69% (45100/65361)
Checking out files: 70% (45753/65361)
Checking out files: 71% (46407/65361)
Checking out files: 72% (47060/65361)
Checking out files: 73% (47714/65361)
Checking out files: 74% (48368/65361)
Checking out files: 75% (49021/65361)
Checking out files: 76% (49675/65361)
Checking out files: 77% (50328/65361)
Checking out files: 78% (50982/65361)
Checking out files: 78% (51514/65361)
Checking out files: 79% (51636/65361)
Checking out files: 80% (52289/65361)
Checking out files: 81% (52943/65361)
Checking out files: 82% (53597/65361)
Checking out files: 83% (54250/65361)
Checking out files: 84% (54904/65361)
Checking out files: 85% (55557/65361)
Checking out files: 86% (56211/65361)
Checking out files: 87% (56865/65361)
Checking out files: 88% (57518/65361)
Checking out files: 89% (58172/65361)
Checking out files: 90% (58825/65361)
Checking out files: 91% (59479/65361)
Checking out files: 92% (60133/65361)
Checking out files: 93% (60786/65361)
Checking out files: 94% (61440/65361)
Checking out files: 95% (62093/65361)
Checking out files: 96% (62747/65361)
Checking out files: 97% (63401/65361)
Checking out files: 98% (64054/65361)
Checking out files: 99% (64708/65361)
Checking out files: 100% (65361/65361)
Checking out files: 100% (65361/65361), done.
Switched to branch 'autosynth-redis'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main
last_synth_commit_hash = get_last_metadata_commit(args.metadata_path)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit
text=True,
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run
with Popen(*popenargs, **kwargs) as process:
TypeError: __init__() got an unexpected keyword argument 'text'
```
Google internal developers can see the full log [here](https://sponge/40f694d4-43de-41f0-b993-f4694e4a45de).
|
1.0
|
Synthesis failed for redis - Hello! Autosynth couldn't regenerate redis. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Checking out files: 24% (16109/65361)
Checking out files: 25% (16341/65361)
Checking out files: 26% (16994/65361)
Checking out files: 27% (17648/65361)
Checking out files: 28% (18302/65361)
Checking out files: 29% (18955/65361)
Checking out files: 30% (19609/65361)
Checking out files: 31% (20262/65361)
Checking out files: 32% (20916/65361)
Checking out files: 33% (21570/65361)
Checking out files: 34% (22223/65361)
Checking out files: 35% (22877/65361)
Checking out files: 36% (23530/65361)
Checking out files: 37% (24184/65361)
Checking out files: 38% (24838/65361)
Checking out files: 39% (25491/65361)
Checking out files: 40% (26145/65361)
Checking out files: 41% (26799/65361)
Checking out files: 42% (27452/65361)
Checking out files: 43% (28106/65361)
Checking out files: 44% (28759/65361)
Checking out files: 45% (29413/65361)
Checking out files: 46% (30067/65361)
Checking out files: 47% (30720/65361)
Checking out files: 48% (31374/65361)
Checking out files: 49% (32027/65361)
Checking out files: 50% (32681/65361)
Checking out files: 51% (33335/65361)
Checking out files: 52% (33988/65361)
Checking out files: 52% (34477/65361)
Checking out files: 53% (34642/65361)
Checking out files: 54% (35295/65361)
Checking out files: 55% (35949/65361)
Checking out files: 56% (36603/65361)
Checking out files: 57% (37256/65361)
Checking out files: 58% (37910/65361)
Checking out files: 59% (38563/65361)
Checking out files: 60% (39217/65361)
Checking out files: 61% (39871/65361)
Checking out files: 62% (40524/65361)
Checking out files: 63% (41178/65361)
Checking out files: 64% (41832/65361)
Checking out files: 65% (42485/65361)
Checking out files: 66% (43139/65361)
Checking out files: 67% (43792/65361)
Checking out files: 68% (44446/65361)
Checking out files: 69% (45100/65361)
Checking out files: 70% (45753/65361)
Checking out files: 71% (46407/65361)
Checking out files: 72% (47060/65361)
Checking out files: 73% (47714/65361)
Checking out files: 74% (48368/65361)
Checking out files: 75% (49021/65361)
Checking out files: 76% (49675/65361)
Checking out files: 77% (50328/65361)
Checking out files: 78% (50982/65361)
Checking out files: 78% (51514/65361)
Checking out files: 79% (51636/65361)
Checking out files: 80% (52289/65361)
Checking out files: 81% (52943/65361)
Checking out files: 82% (53597/65361)
Checking out files: 83% (54250/65361)
Checking out files: 84% (54904/65361)
Checking out files: 85% (55557/65361)
Checking out files: 86% (56211/65361)
Checking out files: 87% (56865/65361)
Checking out files: 88% (57518/65361)
Checking out files: 89% (58172/65361)
Checking out files: 90% (58825/65361)
Checking out files: 91% (59479/65361)
Checking out files: 92% (60133/65361)
Checking out files: 93% (60786/65361)
Checking out files: 94% (61440/65361)
Checking out files: 95% (62093/65361)
Checking out files: 96% (62747/65361)
Checking out files: 97% (63401/65361)
Checking out files: 98% (64054/65361)
Checking out files: 99% (64708/65361)
Checking out files: 100% (65361/65361)
Checking out files: 100% (65361/65361), done.
Switched to branch 'autosynth-redis'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main
last_synth_commit_hash = get_last_metadata_commit(args.metadata_path)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit
text=True,
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run
with Popen(*popenargs, **kwargs) as process:
TypeError: __init__() got an unexpected keyword argument 'text'
```
Google internal developers can see the full log [here](https://sponge/40f694d4-43de-41f0-b993-f4694e4a45de).
|
non_test
|
synthesis failed for redis hello autosynth couldn t regenerate redis broken heart here s the output from running synth py cloning into working repo checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files done switched to branch autosynth redis traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main last synth commit hash get last metadata commit args metadata path file tmpfs src git autosynth autosynth synth py line in get last metadata commit text true file home kbuilder pyenv versions lib subprocess py line in run with popen popenargs kwargs as process typeerror init got an unexpected keyword argument text google internal developers can see the full log
| 0
|
4,190
| 2,713,029,424
|
IssuesEvent
|
2015-04-09 16:56:16
|
rssidlowski/Pollution_Source_Tracking
|
https://api.github.com/repos/rssidlowski/Pollution_Source_Tracking
|
closed
|
Add CSR layer to PST
|
COBDev Ready for Testing enhancement moderate priority
|
Add CSR layer, allow user to toggle on and off and by default have turned off.
|
1.0
|
Add CSR layer to PST - Add CSR layer, allow user to toggle on and off and by default have turned off.
|
test
|
add csr layer to pst add csr layer allow user to toggle on and off and by default have turned off
| 1
|
224,515
| 17,191,420,796
|
IssuesEvent
|
2021-07-16 11:33:13
|
eficode-academy/SimpleFortuneCookie
|
https://api.github.com/repos/eficode-academy/SimpleFortuneCookie
|
opened
|
Add new steps to push to Kubernetes
|
documentation enhancement
|
Problem: getting access to the K8S cluster in AWS/GCP from GH actions.
One solution is to make people run self-hosted runners on their instance:
Link: https://github.com/actions/runner
|
1.0
|
Add new steps to push to Kubernetes - Problem: getting access to the K8S cluster in AWS/GCP from GH actions.
One solution is to make people run self-hosted runners on their instance:
Link: https://github.com/actions/runner
|
non_test
|
add new steps to push to kubernetes problem getting access to the cluster in aws gcp from gh actions one solution is to make people run self hosted runners on their instance link
| 0
|
72,702
| 7,309,292,937
|
IssuesEvent
|
2018-02-28 11:15:01
|
KirillVladimirov/python-messenger
|
https://api.github.com/repos/KirillVladimirov/python-messenger
|
closed
|
lesson_1
|
doc lesson tests
|
1. Функционал
Первая часть домашнего задания будет заключаться в реализации простого клиент-серверного взаимодействия по протоколу JIM (JSON instant messaging):
клиент отправляет запрос серверу;
сервер отвечает соответствующим кодом результата.
Клиент и сервер должны быть реализованы в виде отдельных скриптов, содержащих соответствующие функции.
Функции клиента:
- сформировать presence-сообщение;
- отправить сообщение серверу;
- получить ответ сервера;
- разобрать сообщение сервера;
- параметры командной строки скрипта client.py <addr> [<port>]:
addr - ip-адрес сервера
port - tcp-порт на сервере, по умолчанию 7777
Функции сервера:
- принимает сообщение клиента;
- формирует ответ клиенту;
- отправляет ответ клиенту;
- имеет параметры командной строки:
-p <port> - TCP-порт для работы (по умолчанию использует порт 7777)
-a <addr> - IP-адрес для прослушивания (по умолчанию слушает все доступные адреса)
2. Тесты
Для всех функций необходимо написать тесты с использованием doctest (небольшие тесты в документации функций), unittest или py.test (в дальнейшем упор будет делаться на библиотеку py.test). Тесты должны быть оформлены в отдельных скриптах с префиксом test_ в имени файла (например, test_client.py).
3. Дополнительно
В качестве практики написания тестов напишите тесты для домашних работ курса Python-1.
|
1.0
|
lesson_1 - 1. Функционал
Первая часть домашнего задания будет заключаться в реализации простого клиент-серверного взаимодействия по протоколу JIM (JSON instant messaging):
клиент отправляет запрос серверу;
сервер отвечает соответствующим кодом результата.
Клиент и сервер должны быть реализованы в виде отдельных скриптов, содержащих соответствующие функции.
Функции клиента:
- сформировать presence-сообщение;
- отправить сообщение серверу;
- получить ответ сервера;
- разобрать сообщение сервера;
- параметры командной строки скрипта client.py <addr> [<port>]:
addr - ip-адрес сервера
port - tcp-порт на сервере, по умолчанию 7777
Функции сервера:
- принимает сообщение клиента;
- формирует ответ клиенту;
- отправляет ответ клиенту;
- имеет параметры командной строки:
-p <port> - TCP-порт для работы (по умолчанию использует порт 7777)
-a <addr> - IP-адрес для прослушивания (по умолчанию слушает все доступные адреса)
2. Тесты
Для всех функций необходимо написать тесты с использованием doctest (небольшие тесты в документации функций), unittest или py.test (в дальнейшем упор будет делаться на библиотеку py.test). Тесты должны быть оформлены в отдельных скриптах с префиксом test_ в имени файла (например, test_client.py).
3. Дополнительно
В качестве практики написания тестов напишите тесты для домашних работ курса Python-1.
|
test
|
lesson функционал первая часть домашнего задания будет заключаться в реализации простого клиент серверного взаимодействия по протоколу jim json instant messaging клиент отправляет запрос серверу сервер отвечает соответствующим кодом результата клиент и сервер должны быть реализованы в виде отдельных скриптов содержащих соответствующие функции функции клиента сформировать presence сообщение отправить сообщение серверу получить ответ сервера разобрать сообщение сервера параметры командной строки скрипта client py addr ip адрес сервера port tcp порт на сервере по умолчанию функции сервера принимает сообщение клиента формирует ответ клиенту отправляет ответ клиенту имеет параметры командной строки p tcp порт для работы по умолчанию использует порт a ip адрес для прослушивания по умолчанию слушает все доступные адреса тесты для всех функций необходимо написать тесты с использованием doctest небольшие тесты в документации функций unittest или py test в дальнейшем упор будет делаться на библиотеку py test тесты должны быть оформлены в отдельных скриптах с префиксом test в имени файла например test client py дополнительно в качестве практики написания тестов напишите тесты для домашних работ курса python
| 1
|
299,812
| 25,928,576,294
|
IssuesEvent
|
2022-12-16 07:51:36
|
saleor/saleor-dashboard
|
https://api.github.com/repos/saleor/saleor-dashboard
|
closed
|
Cypress test fail: should create customer. TC: SALEOR_1201
|
tests
|
**Known bug for versions:**
v39: true
**Additional Info:**
Spec: Tests for customer
|
1.0
|
Cypress test fail: should create customer. TC: SALEOR_1201 - **Known bug for versions:**
v39: true
**Additional Info:**
Spec: Tests for customer
|
test
|
cypress test fail should create customer tc saleor known bug for versions true additional info spec tests for customer
| 1
|
811,207
| 30,278,884,234
|
IssuesEvent
|
2023-07-07 23:06:24
|
microsoft/PowerToys
|
https://api.github.com/repos/microsoft/PowerToys
|
closed
|
Update notification doesn't open PowerToys
|
Issue-Bug Product-Settings Area-Runner Priority-1
|
### Description of the new feature / enhancement
When I click the Windows notification that lets me know there's an update for PowerToys I expected the PowerToys main window to at least open so I could click on Update.
### Scenario when this would be used?
It speeds up the update process. Now I have to dismiss the notification, then open PowerToys, then click on Update.
### Supporting information
Not sure if this is a bug but it happens on both my PCs, on Windows 10 and 11.
|
1.0
|
Update notification doesn't open PowerToys - ### Description of the new feature / enhancement
When I click the Windows notification that lets me know there's an update for PowerToys I expected the PowerToys main window to at least open so I could click on Update.
### Scenario when this would be used?
It speeds up the update process. Now I have to dismiss the notification, then open PowerToys, then click on Update.
### Supporting information
Not sure if this is a bug but it happens on both my PCs, on Windows 10 and 11.
|
non_test
|
update notification doesn t open powertoys description of the new feature enhancement when i click the windows notification that lets me know there s an update for powertoys i expected the powertoys main window to at least open so i could click on update scenario when this would be used it speeds up the update process now i have to dismiss the notification then open powertoys then click on update supporting information not sure if this is a bug but it happens on both my pcs on windows and
| 0
|
4,194
| 4,876,306,550
|
IssuesEvent
|
2016-11-16 12:25:23
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
Full framework test csproj referencing netstandard csproj using HttpClient fails at runtime
|
Infrastructure question
|
I have a net461 test project that uses a netstandard 1.3 csproj. The netstandard project uses HttpClient.
When I run the tests, some with HTTP gets and some with HTTPS, they fail with this:
```
System.IO.FileLoadException : Could not load file or assembly 'System.Security.Cryptography.X509Certificates, Version=4.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a' or one of its dependencies. The located assembly's manifest definition does not match the assembly reference. (Exception from HRESULT: 0x80131040)
at System.Net.Http.WinHttpHandler.SendAsync(HttpRequestMessage request, CancellationToken cancellationToken)
at System.Net.Http.HttpClientHandler.SendAsync(HttpRequestMessage request, CancellationToken cancellationToken)
at System.Net.Http.HttpMessageInvoker.SendAsync(HttpRequestMessage request, CancellationToken cancellationToken)
at System.Net.Http.HttpClient.SendAsync(HttpRequestMessage request, HttpCompletionOption completionOption, CancellationToken cancellationToken)
at System.Net.Http.HttpClient.GetAsync(Uri requestUri, HttpCompletionOption completionOption, CancellationToken cancellationToken)
```
What can I do about this? There are no warnings or other hints to point me in any direction, and Google searches have left me more confused than ever.
Windows 10 x64 AU
Microsoft Visual Studio Professional 2015
Version 14.0.25431.01 Update 3
Microsoft .NET Framework
Version 4.6.01586
Microsoft .NET Core Tools (Preview 2) 14.1.20907.0
Don't know if this matters, dotnet:
Version : 1.0.1
Build : cee57bf6c981237d80aa1631cfe83cb9ba329f12
|
1.0
|
Full framework test csproj referencing netstandard csproj using HttpClient fails at runtime - I have a net461 test project that uses a netstandard 1.3 csproj. The netstandard project uses HttpClient.
When I run the tests, some with HTTP gets and some with HTTPS, they fail with this:
```
System.IO.FileLoadException : Could not load file or assembly 'System.Security.Cryptography.X509Certificates, Version=4.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a' or one of its dependencies. The located assembly's manifest definition does not match the assembly reference. (Exception from HRESULT: 0x80131040)
at System.Net.Http.WinHttpHandler.SendAsync(HttpRequestMessage request, CancellationToken cancellationToken)
at System.Net.Http.HttpClientHandler.SendAsync(HttpRequestMessage request, CancellationToken cancellationToken)
at System.Net.Http.HttpMessageInvoker.SendAsync(HttpRequestMessage request, CancellationToken cancellationToken)
at System.Net.Http.HttpClient.SendAsync(HttpRequestMessage request, HttpCompletionOption completionOption, CancellationToken cancellationToken)
at System.Net.Http.HttpClient.GetAsync(Uri requestUri, HttpCompletionOption completionOption, CancellationToken cancellationToken)
```
What can I do about this? There are no warnings or other hints to point me in any direction, and Google searches have left me more confused than ever.
Windows 10 x64 AU
Microsoft Visual Studio Professional 2015
Version 14.0.25431.01 Update 3
Microsoft .NET Framework
Version 4.6.01586
Microsoft .NET Core Tools (Preview 2) 14.1.20907.0
Don't know if this matters, dotnet:
Version : 1.0.1
Build : cee57bf6c981237d80aa1631cfe83cb9ba329f12
|
non_test
|
full framework test csproj referencing netstandard csproj using httpclient fails at runtime i have a test project that uses a netstandard csproj the netstandard project uses httpclient when i run the tests some with http gets and some with https they fail with this system io fileloadexception could not load file or assembly system security cryptography version culture neutral publickeytoken or one of its dependencies the located assembly s manifest definition does not match the assembly reference exception from hresult at system net http winhttphandler sendasync httprequestmessage request cancellationtoken cancellationtoken at system net http httpclienthandler sendasync httprequestmessage request cancellationtoken cancellationtoken at system net http httpmessageinvoker sendasync httprequestmessage request cancellationtoken cancellationtoken at system net http httpclient sendasync httprequestmessage request httpcompletionoption completionoption cancellationtoken cancellationtoken at system net http httpclient getasync uri requesturi httpcompletionoption completionoption cancellationtoken cancellationtoken what can i do about this there are no warnings or other hints to point me in any direction and google searches have left me more confused than ever windows au microsoft visual studio professional version update microsoft net framework version microsoft net core tools preview don t know if this matters dotnet version build
| 0
|
283,804
| 24,563,241,467
|
IssuesEvent
|
2022-10-12 22:47:10
|
systemd/systemd
|
https://api.github.com/repos/systemd/systemd
|
closed
|
pid1: simultaneously starting multiple services with same image may trigger conflict
|
bug 🐛 pid1 tests portable
|
### systemd version the issue has been seen with
HEAD
### Used distribution
Ubuntu focal
### Linux kernel version used
5.4.0-125-generic
### CPU architectures issue was seen on
x86_64
### Component
systemd
### Expected behaviour you didn't see
TEST-29-PORTABLE succeeds.
### Unexpected behaviour you saw
TEST-29-PORTABLE fails:
https://autopkgtest.ubuntu.com/results/autopkgtest-focal-upstream-systemd-ci-systemd-ci/focal/amd64/s/systemd-upstream/20220908_232004_2ec87@/log.gz
See [TEST-29-PORTABLE.journal.tar.gz](https://github.com/systemd/systemd/files/9535214/TEST-29-PORTABLE.journal.tar.gz).
### Steps to reproduce the problem
Run the test.
### Additional program output to the terminal or log subsystem illustrating the issue
```
Sep 09 07:33:18 systemd[452]: Opened '/usr/share/minimal_1.raw' in O_RDONLY access mode, with O_DIRECT enabled.
Sep 09 07:33:18 systemd[452]: Successfully acquired /dev/loop0, devno=7:0, nr=0, diskseq=0
Sep 09 07:33:18 systemd[452]: Allocating context for crypt device /usr/share/minimal_1.verity.
Sep 09 07:33:18 systemd[452]: Trying to open and read device /usr/share/minimal_1.verity with direct-io.
Sep 09 07:33:18 systemd[452]: Crypto backend (OpenSSL 1.1.1f 31 Mar 2020) initialized in cryptsetup library version 2.2.2.
Sep 09 07:33:18 systemd[452]: Detected kernel Linux 5.4.0-125-generic x86_64.
Sep 09 07:33:18 systemd[452]: Reading VERITY header of size 512 on device /usr/share/minimal_1.verity, offset 0.
Sep 09 07:33:18 systemd[454]: Opened '/usr/share/minimal_1.raw' in O_RDONLY access mode, with O_DIRECT enabled.
Sep 09 07:33:18 systemd[452]: Setting ciphertext data device to /dev/loop0.
Sep 09 07:33:18 systemd[452]: Trying to open and read device /dev/loop0 with direct-io.
Sep 09 07:33:18 systemd[452]: Activating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity by volume key.
Sep 09 07:33:18 systemd[452]: Trying to activate VERITY device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity using hash sha256.
Sep 09 07:33:18 systemd[452]: Allocating a free loop device.
Sep 09 07:33:18 systemd[454]: Successfully acquired /dev/loop1, devno=7:1, nr=1, diskseq=0
Sep 09 07:33:18 systemd[452]: Trying to open and read device /dev/loop5 with direct-io.
Sep 09 07:33:18 systemd[454]: Allocating context for crypt device /usr/share/minimal_1.verity.
Sep 09 07:33:18 systemd[454]: Trying to open and read device /usr/share/minimal_1.verity with direct-io.
Sep 09 07:33:18 systemd[454]: Trying to load VERITY crypt type from device /usr/share/minimal_1.verity.
Sep 09 07:33:18 systemd[454]: Reading VERITY header of size 512 on device /usr/share/minimal_1.verity, offset 0.
Sep 09 07:33:18 systemd[454]: Setting ciphertext data device to /dev/loop1.
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop1 with direct-io.
Sep 09 07:33:18 systemd[454]: Activating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity by volume key.
Sep 09 07:33:18 systemd[454]: Device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity already exists. <-- in LOG_ERR
Sep 09 07:33:18 systemd[454]: Allocating crypt device context by device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity.
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- What??
Sep 09 07:33:18 systemd[454]: Allocating context for crypt device /dev/loop0. <-- Hm??
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- Huh??
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- again??
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop5 with direct-io.
Sep 09 07:33:18 systemd[454]: Error opening verity device, crypt_volume_key_get failed: Invalid argument
Sep 09 07:33:18 systemd[454]: Releasing crypt device /dev/loop5 context.
Sep 09 07:33:18 systemd[454]: Allocating context for crypt device /usr/share/minimal_1.verity.
Sep 09 07:33:18 systemd[454]: Trying to open and read device /usr/share/minimal_1.verity with direct-io.
Sep 09 07:33:18 systemd[454]: Trying to load VERITY crypt type from device /usr/share/minimal_1.verity.
Sep 09 07:33:18 systemd[454]: Reading VERITY header of size 512 on device /usr/share/minimal_1.verity, offset 0.
Sep 09 07:33:18 systemd[454]: Setting ciphertext data device to /dev/loop1.
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop1 with direct-io.
Sep 09 07:33:18 systemd[454]: Activating volume loop1-verity by volume key.
Sep 09 07:33:18 systemd[454]: Trying to activate VERITY device loop1-verity using hash sha256.
Sep 09 07:33:18 systemd[454]: Allocating a free loop device.
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop6 with direct-io.
Sep 09 07:33:18 systemd[454]: Calculated device size is 5928 sectors (RO), offset 0.
(udev processing many events)
Sep 09 07:33:21 kernel: /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity: Can't open blockdev <-- LOG_ERR
(snip)
Sep 09 07:33:21 systemd[452]: Udev cookie 0xd4dedbd (semid 10) destroyed
Sep 09 07:33:21 systemd[452]: Verity volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity status is V.
Sep 09 07:33:21 systemd[454]: Udev cookie 0xd4d0212 (semid 11) destroyed
Sep 09 07:33:21 systemd[454]: Verity volume loop1-verity status is V.
Sep 09 07:33:21 systemd[454]: Deactivating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity. <-- ??
Sep 09 07:33:21 systemd[454]: Allocating crypt device context by device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity.
Sep 09 07:33:21 systemd[452]: Probed fstype 'squashfs' on partition /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity.
Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- Oh, again??
Sep 09 07:33:21 systemd[454]: Allocating context for crypt device /dev/loop0.
Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io.
Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io.
Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop5 with direct-io.
Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io.
Sep 09 07:33:21 systemd[454]: Verity volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity status is V.
Sep 09 07:33:21 systemd[454]: dm remove dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity [ opencount flush retryremove deferredremove ] [16384] (*1) <-- Wait, wait, it is still used by PID=452...
Sep 09 07:33:21 systemd[452]: Mounting /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity (squashfs) on /run/systemd/unit-root (MS_RDONLY|MS_NODEV "")...
Sep 09 07:33:21 systemd-udevd[281]: dm-0: Device is queued (SEQNUM=3921, ACTION=remove)
Sep 09 07:33:21 systemd[452]: Failed to mount /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity (type squashfs) on /run/systemd/unit-root (MS_RDONLY|MS_NODEV ""): No such file or directory
Sep 09 07:33:21 systemd[452]: Failed to mount root image: No such file or directory
Sep 09 07:33:21 systemd[452]: Deactivating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity.
Sep 09 07:33:21 systemd[452]: Device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity is not active.
Sep 09 07:33:21 systemd[452]: Failed to deactivate encrypted partition dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity <-- yeah, it is already removed by PID=454...
Sep 09 07:33:21 systemd[452]: Releasing crypt device /usr/share/minimal_1.verity context.
Sep 09 07:33:21 systemd[452]: Releasing device-mapper backend.
Sep 09 07:33:21 systemd[452]: Closing read only fd for /usr/share/minimal_1.verity.
Sep 09 07:33:21 systemd[452]: Closed loop /dev/loop5 (/usr/share/minimal_1.verity).
Sep 09 07:33:21 systemd[452]: minimal-app0-bar.service: Failed to set up mount namespacing: No such file or directory
Sep 09 07:33:21 systemd[452]: minimal-app0-bar.service: Failed at step NAMESPACE spawning cat: No such file or directory
Sep 09 07:33:21 systemd[1]: Received SIGCHLD from PID 452 ((cat)).
Sep 09 07:33:21 systemd[1]: Child 452 ((cat)) died (code=exited, status=226/NAMESPACE)
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Child 452 belongs to minimal-app0-bar.service.
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Control process exited, code=exited, status=226/NAMESPACE
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Got final SIGCHLD for state start-pre.
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Failed with result 'exit-code'.
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Service will not restart (restart setting)
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Changed start-pre -> failed
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Job 486 minimal-app0-bar.service/start finished, result=failed
Sep 09 07:33:21 systemd[1]: Failed to start minimal-app0-bar.service.
```
|
1.0
|
pid1: simultaneously starting multiple services with same image may trigger conflict - ### systemd version the issue has been seen with
HEAD
### Used distribution
Ubuntu focal
### Linux kernel version used
5.4.0-125-generic
### CPU architectures issue was seen on
x86_64
### Component
systemd
### Expected behaviour you didn't see
TEST-29-PORTABLE succeeds.
### Unexpected behaviour you saw
TEST-29-PORTABLE fails:
https://autopkgtest.ubuntu.com/results/autopkgtest-focal-upstream-systemd-ci-systemd-ci/focal/amd64/s/systemd-upstream/20220908_232004_2ec87@/log.gz
See [TEST-29-PORTABLE.journal.tar.gz](https://github.com/systemd/systemd/files/9535214/TEST-29-PORTABLE.journal.tar.gz).
### Steps to reproduce the problem
Run the test.
### Additional program output to the terminal or log subsystem illustrating the issue
```
Sep 09 07:33:18 systemd[452]: Opened '/usr/share/minimal_1.raw' in O_RDONLY access mode, with O_DIRECT enabled.
Sep 09 07:33:18 systemd[452]: Successfully acquired /dev/loop0, devno=7:0, nr=0, diskseq=0
Sep 09 07:33:18 systemd[452]: Allocating context for crypt device /usr/share/minimal_1.verity.
Sep 09 07:33:18 systemd[452]: Trying to open and read device /usr/share/minimal_1.verity with direct-io.
Sep 09 07:33:18 systemd[452]: Crypto backend (OpenSSL 1.1.1f 31 Mar 2020) initialized in cryptsetup library version 2.2.2.
Sep 09 07:33:18 systemd[452]: Detected kernel Linux 5.4.0-125-generic x86_64.
Sep 09 07:33:18 systemd[452]: Reading VERITY header of size 512 on device /usr/share/minimal_1.verity, offset 0.
Sep 09 07:33:18 systemd[454]: Opened '/usr/share/minimal_1.raw' in O_RDONLY access mode, with O_DIRECT enabled.
Sep 09 07:33:18 systemd[452]: Setting ciphertext data device to /dev/loop0.
Sep 09 07:33:18 systemd[452]: Trying to open and read device /dev/loop0 with direct-io.
Sep 09 07:33:18 systemd[452]: Activating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity by volume key.
Sep 09 07:33:18 systemd[452]: Trying to activate VERITY device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity using hash sha256.
Sep 09 07:33:18 systemd[452]: Allocating a free loop device.
Sep 09 07:33:18 systemd[454]: Successfully acquired /dev/loop1, devno=7:1, nr=1, diskseq=0
Sep 09 07:33:18 systemd[452]: Trying to open and read device /dev/loop5 with direct-io.
Sep 09 07:33:18 systemd[454]: Allocating context for crypt device /usr/share/minimal_1.verity.
Sep 09 07:33:18 systemd[454]: Trying to open and read device /usr/share/minimal_1.verity with direct-io.
Sep 09 07:33:18 systemd[454]: Trying to load VERITY crypt type from device /usr/share/minimal_1.verity.
Sep 09 07:33:18 systemd[454]: Reading VERITY header of size 512 on device /usr/share/minimal_1.verity, offset 0.
Sep 09 07:33:18 systemd[454]: Setting ciphertext data device to /dev/loop1.
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop1 with direct-io.
Sep 09 07:33:18 systemd[454]: Activating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity by volume key.
Sep 09 07:33:18 systemd[454]: Device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity already exists. <-- in LOG_ERR
Sep 09 07:33:18 systemd[454]: Allocating crypt device context by device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity.
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- What??
Sep 09 07:33:18 systemd[454]: Allocating context for crypt device /dev/loop0. <-- Hm??
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- Huh??
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- again??
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop5 with direct-io.
Sep 09 07:33:18 systemd[454]: Error opening verity device, crypt_volume_key_get failed: Invalid argument
Sep 09 07:33:18 systemd[454]: Releasing crypt device /dev/loop5 context.
Sep 09 07:33:18 systemd[454]: Allocating context for crypt device /usr/share/minimal_1.verity.
Sep 09 07:33:18 systemd[454]: Trying to open and read device /usr/share/minimal_1.verity with direct-io.
Sep 09 07:33:18 systemd[454]: Trying to load VERITY crypt type from device /usr/share/minimal_1.verity.
Sep 09 07:33:18 systemd[454]: Reading VERITY header of size 512 on device /usr/share/minimal_1.verity, offset 0.
Sep 09 07:33:18 systemd[454]: Setting ciphertext data device to /dev/loop1.
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop1 with direct-io.
Sep 09 07:33:18 systemd[454]: Activating volume loop1-verity by volume key.
Sep 09 07:33:18 systemd[454]: Trying to activate VERITY device loop1-verity using hash sha256.
Sep 09 07:33:18 systemd[454]: Allocating a free loop device.
Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop6 with direct-io.
Sep 09 07:33:18 systemd[454]: Calculated device size is 5928 sectors (RO), offset 0.
(udev processing many events)
Sep 09 07:33:21 kernel: /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity: Can't open blockdev <-- LOG_ERR
(snip)
Sep 09 07:33:21 systemd[452]: Udev cookie 0xd4dedbd (semid 10) destroyed
Sep 09 07:33:21 systemd[452]: Verity volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity status is V.
Sep 09 07:33:21 systemd[454]: Udev cookie 0xd4d0212 (semid 11) destroyed
Sep 09 07:33:21 systemd[454]: Verity volume loop1-verity status is V.
Sep 09 07:33:21 systemd[454]: Deactivating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity. <-- ??
Sep 09 07:33:21 systemd[454]: Allocating crypt device context by device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity.
Sep 09 07:33:21 systemd[452]: Probed fstype 'squashfs' on partition /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity.
Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- Oh, again??
Sep 09 07:33:21 systemd[454]: Allocating context for crypt device /dev/loop0.
Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io.
Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io.
Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop5 with direct-io.
Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io.
Sep 09 07:33:21 systemd[454]: Verity volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity status is V.
Sep 09 07:33:21 systemd[454]: dm remove dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity [ opencount flush retryremove deferredremove ] [16384] (*1) <-- Wait, wait, it is still used by PID=452...
Sep 09 07:33:21 systemd[452]: Mounting /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity (squashfs) on /run/systemd/unit-root (MS_RDONLY|MS_NODEV "")...
Sep 09 07:33:21 systemd-udevd[281]: dm-0: Device is queued (SEQNUM=3921, ACTION=remove)
Sep 09 07:33:21 systemd[452]: Failed to mount /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity (type squashfs) on /run/systemd/unit-root (MS_RDONLY|MS_NODEV ""): No such file or directory
Sep 09 07:33:21 systemd[452]: Failed to mount root image: No such file or directory
Sep 09 07:33:21 systemd[452]: Deactivating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity.
Sep 09 07:33:21 systemd[452]: Device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity is not active.
Sep 09 07:33:21 systemd[452]: Failed to deactivate encrypted partition dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity <-- yeah, it is already removed by PID=454...
Sep 09 07:33:21 systemd[452]: Releasing crypt device /usr/share/minimal_1.verity context.
Sep 09 07:33:21 systemd[452]: Releasing device-mapper backend.
Sep 09 07:33:21 systemd[452]: Closing read only fd for /usr/share/minimal_1.verity.
Sep 09 07:33:21 systemd[452]: Closed loop /dev/loop5 (/usr/share/minimal_1.verity).
Sep 09 07:33:21 systemd[452]: minimal-app0-bar.service: Failed to set up mount namespacing: No such file or directory
Sep 09 07:33:21 systemd[452]: minimal-app0-bar.service: Failed at step NAMESPACE spawning cat: No such file or directory
Sep 09 07:33:21 systemd[1]: Received SIGCHLD from PID 452 ((cat)).
Sep 09 07:33:21 systemd[1]: Child 452 ((cat)) died (code=exited, status=226/NAMESPACE)
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Child 452 belongs to minimal-app0-bar.service.
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Control process exited, code=exited, status=226/NAMESPACE
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Got final SIGCHLD for state start-pre.
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Failed with result 'exit-code'.
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Service will not restart (restart setting)
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Changed start-pre -> failed
Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Job 486 minimal-app0-bar.service/start finished, result=failed
Sep 09 07:33:21 systemd[1]: Failed to start minimal-app0-bar.service.
```
|
test
|
simultaneously starting multiple services with same image may trigger conflict systemd version the issue has been seen with head used distribution ubuntu focal linux kernel version used generic cpu architectures issue was seen on component systemd expected behaviour you didn t see test portable succeeds unexpected behaviour you saw test portable fails see steps to reproduce the problem run the test additional program output to the terminal or log subsystem illustrating the issue sep systemd opened usr share minimal raw in o rdonly access mode with o direct enabled sep systemd successfully acquired dev devno nr diskseq sep systemd allocating context for crypt device usr share minimal verity sep systemd trying to open and read device usr share minimal verity with direct io sep systemd crypto backend openssl mar initialized in cryptsetup library version sep systemd detected kernel linux generic sep systemd reading verity header of size on device usr share minimal verity offset sep systemd opened usr share minimal raw in o rdonly access mode with o direct enabled sep systemd setting ciphertext data device to dev sep systemd trying to open and read device dev with direct io sep systemd activating volume verity by volume key sep systemd trying to activate verity device verity using hash sep systemd allocating a free loop device sep systemd successfully acquired dev devno nr diskseq sep systemd trying to open and read device dev with direct io sep systemd allocating context for crypt device usr share minimal verity sep systemd trying to open and read device usr share minimal verity with direct io sep systemd trying to load verity crypt type from device usr share minimal verity sep systemd reading verity header of size on device usr share minimal verity offset sep systemd setting ciphertext data device to dev sep systemd trying to open and read device dev with direct io sep systemd activating volume verity by volume key sep systemd device verity already exists in log err sep systemd allocating crypt device context by device verity sep systemd trying to open and read device dev with direct io what sep systemd allocating context for crypt device dev hm sep systemd trying to open and read device dev with direct io huh sep systemd trying to open and read device dev with direct io again sep systemd trying to open and read device dev with direct io sep systemd error opening verity device crypt volume key get failed invalid argument sep systemd releasing crypt device dev context sep systemd allocating context for crypt device usr share minimal verity sep systemd trying to open and read device usr share minimal verity with direct io sep systemd trying to load verity crypt type from device usr share minimal verity sep systemd reading verity header of size on device usr share minimal verity offset sep systemd setting ciphertext data device to dev sep systemd trying to open and read device dev with direct io sep systemd activating volume verity by volume key sep systemd trying to activate verity device verity using hash sep systemd allocating a free loop device sep systemd trying to open and read device dev with direct io sep systemd calculated device size is sectors ro offset udev processing many events sep kernel dev mapper verity can t open blockdev log err snip sep systemd udev cookie semid destroyed sep systemd verity volume verity status is v sep systemd udev cookie semid destroyed sep systemd verity volume verity status is v sep systemd deactivating volume verity sep systemd allocating crypt device context by device verity sep systemd probed fstype squashfs on partition dev mapper verity sep systemd trying to open and read device dev with direct io oh again sep systemd allocating context for crypt device dev sep systemd trying to open and read device dev with direct io sep systemd trying to open and read device dev with direct io sep systemd trying to open and read device dev with direct io sep systemd trying to open and read device dev with direct io sep systemd verity volume verity status is v sep systemd dm remove verity wait wait it is still used by pid sep systemd mounting dev mapper verity squashfs on run systemd unit root ms rdonly ms nodev sep systemd udevd dm device is queued seqnum action remove sep systemd failed to mount dev mapper verity type squashfs on run systemd unit root ms rdonly ms nodev no such file or directory sep systemd failed to mount root image no such file or directory sep systemd deactivating volume verity sep systemd device verity is not active sep systemd failed to deactivate encrypted partition verity yeah it is already removed by pid sep systemd releasing crypt device usr share minimal verity context sep systemd releasing device mapper backend sep systemd closing read only fd for usr share minimal verity sep systemd closed loop dev usr share minimal verity sep systemd minimal bar service failed to set up mount namespacing no such file or directory sep systemd minimal bar service failed at step namespace spawning cat no such file or directory sep systemd received sigchld from pid cat sep systemd child cat died code exited status namespace sep systemd minimal bar service child belongs to minimal bar service sep systemd minimal bar service control process exited code exited status namespace sep systemd minimal bar service got final sigchld for state start pre sep systemd minimal bar service failed with result exit code sep systemd minimal bar service service will not restart restart setting sep systemd minimal bar service changed start pre failed sep systemd minimal bar service job minimal bar service start finished result failed sep systemd failed to start minimal bar service
| 1
|
654,089
| 21,637,123,199
|
IssuesEvent
|
2022-05-05 15:07:48
|
redwoodjs/redwood
|
https://api.github.com/repos/redwoodjs/redwood
|
closed
|
Scary and duplicated output when generating sdl
|
help wanted v1/priority
|
I just created a new RW project and upgraded to 44.1-canary.26
First thing I did was to add this to `prisma.schema`
```
model Product {
id String @id @default(uuid())
createdAt DateTime @default(now())
updatedAt DateTime @default(now()) @updatedAt
name String
}
```
And then when I try to generate the sdl for it I get this output:
```
$ yarn rw g sdl --crud product
yarn run v1.22.17
$ /Users/tobbe/tmp/uniquets/node_modules/.bin/rw g sdl --crud product
(node:34268) ExperimentalWarning: stream/web is an experimental feature. This feature could change at any time
(Use `node --trace-warnings ...` to show where the warning was created)
✔ Generating SDL files...
✔ Successfully wrote file `./api/src/graphql/products.sdl.ts`
✔ Successfully wrote file `./api/src/services/products/products.scenarios.ts`
✔ Successfully wrote file `./api/src/services/products/products.test.ts`
✔ Successfully wrote file `./api/src/services/products/products.ts`
⠹ Generating types ...
Unable to find any GraphQL type definitions for the following pointers:
- ./web/src/**/!(*.d).{ts,tsx,js,jsx}
Unable to find any GraphQL type definitions for the following pointers:
✔ Generating SDL files...
✔ Successfully wrote file `./api/src/graphql/products.sdl.ts`
✔ Successfully wrote file `./api/src/services/products/products.scenarios.ts`
✔ Successfully wrote file `./api/src/services/products/products.test.ts`
✔ Successfully wrote file `./api/src/services/products/products.ts`
✔ Generating types ...
✨ Done in 2.68s.
```
The "Unable to find any GraphQL..." message looks scary. Is it an error? A warning?
Also, why does it say "Successfully wrote..." for all files twice?
|
1.0
|
Scary and duplicated output when generating sdl - I just created a new RW project and upgraded to 44.1-canary.26
First thing I did was to add this to `prisma.schema`
```
model Product {
id String @id @default(uuid())
createdAt DateTime @default(now())
updatedAt DateTime @default(now()) @updatedAt
name String
}
```
And then when I try to generate the sdl for it I get this output:
```
$ yarn rw g sdl --crud product
yarn run v1.22.17
$ /Users/tobbe/tmp/uniquets/node_modules/.bin/rw g sdl --crud product
(node:34268) ExperimentalWarning: stream/web is an experimental feature. This feature could change at any time
(Use `node --trace-warnings ...` to show where the warning was created)
✔ Generating SDL files...
✔ Successfully wrote file `./api/src/graphql/products.sdl.ts`
✔ Successfully wrote file `./api/src/services/products/products.scenarios.ts`
✔ Successfully wrote file `./api/src/services/products/products.test.ts`
✔ Successfully wrote file `./api/src/services/products/products.ts`
⠹ Generating types ...
Unable to find any GraphQL type definitions for the following pointers:
- ./web/src/**/!(*.d).{ts,tsx,js,jsx}
Unable to find any GraphQL type definitions for the following pointers:
✔ Generating SDL files...
✔ Successfully wrote file `./api/src/graphql/products.sdl.ts`
✔ Successfully wrote file `./api/src/services/products/products.scenarios.ts`
✔ Successfully wrote file `./api/src/services/products/products.test.ts`
✔ Successfully wrote file `./api/src/services/products/products.ts`
✔ Generating types ...
✨ Done in 2.68s.
```
The "Unable to find any GraphQL..." message looks scary. Is it an error? A warning?
Also, why does it say "Successfully wrote..." for all files twice?
|
non_test
|
scary and duplicated output when generating sdl i just created a new rw project and upgraded to canary first thing i did was to add this to prisma schema model product id string id default uuid createdat datetime default now updatedat datetime default now updatedat name string and then when i try to generate the sdl for it i get this output yarn rw g sdl crud product yarn run users tobbe tmp uniquets node modules bin rw g sdl crud product node experimentalwarning stream web is an experimental feature this feature could change at any time use node trace warnings to show where the warning was created ✔ generating sdl files ✔ successfully wrote file api src graphql products sdl ts ✔ successfully wrote file api src services products products scenarios ts ✔ successfully wrote file api src services products products test ts ✔ successfully wrote file api src services products products ts ⠹ generating types unable to find any graphql type definitions for the following pointers web src d ts tsx js jsx unable to find any graphql type definitions for the following pointers ✔ generating sdl files ✔ successfully wrote file api src graphql products sdl ts ✔ successfully wrote file api src services products products scenarios ts ✔ successfully wrote file api src services products products test ts ✔ successfully wrote file api src services products products ts ✔ generating types ✨ done in the unable to find any graphql message looks scary is it an error a warning also why does it say successfully wrote for all files twice
| 0
|
373,775
| 26,084,208,625
|
IssuesEvent
|
2022-12-25 21:52:11
|
tfiers/PkgGraph.jl
|
https://api.github.com/repos/tfiers/PkgGraph.jl
|
closed
|
readme-in-docs: preprocess further
|
documentation
|
- [x] Global custom subs (`replace`)
- [ ] rm `<details>` surround (regex → subs w/ matched content)
- [x] replace "Reference section of [the docs]" with link to Internals
- [x] rm docs link in H1
- [ ] transform CommonMark.jl output
- [x] specify sections (w/ regex mayb) to keep / exclude
Exclude: Dev, Versions (instead, #22 :))
|
1.0
|
readme-in-docs: preprocess further - - [x] Global custom subs (`replace`)
- [ ] rm `<details>` surround (regex → subs w/ matched content)
- [x] replace "Reference section of [the docs]" with link to Internals
- [x] rm docs link in H1
- [ ] transform CommonMark.jl output
- [x] specify sections (w/ regex mayb) to keep / exclude
Exclude: Dev, Versions (instead, #22 :))
|
non_test
|
readme in docs preprocess further global custom subs replace rm surround regex → subs w matched content replace reference section of with link to internals rm docs link in transform commonmark jl output specify sections w regex mayb to keep exclude exclude dev versions instead
| 0
|
92,672
| 8,375,540,611
|
IssuesEvent
|
2018-10-05 16:44:37
|
blackbaud/skyux2
|
https://api.github.com/repos/blackbaud/skyux2
|
reopened
|
Uncaught errors in logs when running npm run watch
|
Priority: Critical Status: Ready to merge Type: Bug testing
|
When running `npm run watch`, we're seeing a lot of uncaught errors in the logs. For example:
```
ERROR: 'ERROR', TypeError{line: 49079, column: 17082, sourceURL: 'http://localhost:9876/absolute/home/travis/build/blackbaud/skyux2/config/utils/spec-bundle.js?b65bce9e10e00c33526065336acf99271e9affd8', ngDebugContext: DebugContext_{view: Object{def: ..., parent: ..., viewContainerParent: ..., parentNodeDef: ..., context: ..., component: ..., nodes: ..., state: ..., root: ..., renderer: ..., oldValues: ..., disposables: ...}, nodeIndex: 1, nodeDef: Object{index: ..., parent: ..., renderParent: ..., bindingIndex: ..., outputIndex: ..., flags: ..., childFlags: ..., directChildFlags: ..., childMatchedQueries: ..., matchedQueries: ..., matchedQueryIds: ..., references: ..., ngContentIndex: ..., childCount: ..., bindings: ..., bindingFlags: ..., outputs: ..., element: ..., provider: ..., text: ..., query: ..., ngContent: ...}, elDef: Object{index: ..., parent: ..., renderParent: ..., bindingIndex: ..., outputIndex: ..., flags: ..., childFlags: ..., directChildFlags: ..., childMatchedQueries: ..., matchedQueries: ..., matchedQueryIds: ..., references: ..., ngContentIndex: ..., childCount: ..., bindings: ..., bindingFlags: ..., outputs: ..., element: ..., provider: ..., text: ..., query: ..., ngContent: ...}, elView: Object{def: ..., parent: ..., viewContainerParent: ..., parentNodeDef: ..., context: ..., component: ..., nodes: ..., state: ..., root: ..., renderer: ..., oldValues: ..., disposables: ...}}, ngErrorLogger: function () { ... }}
```
Refer to log to see more: https://travis-ci.org/blackbaud/skyux2/jobs/431683334
|
1.0
|
Uncaught errors in logs when running npm run watch - When running `npm run watch`, we're seeing a lot of uncaught errors in the logs. For example:
```
ERROR: 'ERROR', TypeError{line: 49079, column: 17082, sourceURL: 'http://localhost:9876/absolute/home/travis/build/blackbaud/skyux2/config/utils/spec-bundle.js?b65bce9e10e00c33526065336acf99271e9affd8', ngDebugContext: DebugContext_{view: Object{def: ..., parent: ..., viewContainerParent: ..., parentNodeDef: ..., context: ..., component: ..., nodes: ..., state: ..., root: ..., renderer: ..., oldValues: ..., disposables: ...}, nodeIndex: 1, nodeDef: Object{index: ..., parent: ..., renderParent: ..., bindingIndex: ..., outputIndex: ..., flags: ..., childFlags: ..., directChildFlags: ..., childMatchedQueries: ..., matchedQueries: ..., matchedQueryIds: ..., references: ..., ngContentIndex: ..., childCount: ..., bindings: ..., bindingFlags: ..., outputs: ..., element: ..., provider: ..., text: ..., query: ..., ngContent: ...}, elDef: Object{index: ..., parent: ..., renderParent: ..., bindingIndex: ..., outputIndex: ..., flags: ..., childFlags: ..., directChildFlags: ..., childMatchedQueries: ..., matchedQueries: ..., matchedQueryIds: ..., references: ..., ngContentIndex: ..., childCount: ..., bindings: ..., bindingFlags: ..., outputs: ..., element: ..., provider: ..., text: ..., query: ..., ngContent: ...}, elView: Object{def: ..., parent: ..., viewContainerParent: ..., parentNodeDef: ..., context: ..., component: ..., nodes: ..., state: ..., root: ..., renderer: ..., oldValues: ..., disposables: ...}}, ngErrorLogger: function () { ... }}
```
Refer to log to see more: https://travis-ci.org/blackbaud/skyux2/jobs/431683334
|
test
|
uncaught errors in logs when running npm run watch when running npm run watch we re seeing a lot of uncaught errors in the logs for example error error typeerror line column sourceurl ngdebugcontext debugcontext view object def parent viewcontainerparent parentnodedef context component nodes state root renderer oldvalues disposables nodeindex nodedef object index parent renderparent bindingindex outputindex flags childflags directchildflags childmatchedqueries matchedqueries matchedqueryids references ngcontentindex childcount bindings bindingflags outputs element provider text query ngcontent eldef object index parent renderparent bindingindex outputindex flags childflags directchildflags childmatchedqueries matchedqueries matchedqueryids references ngcontentindex childcount bindings bindingflags outputs element provider text query ngcontent elview object def parent viewcontainerparent parentnodedef context component nodes state root renderer oldvalues disposables ngerrorlogger function refer to log to see more
| 1
|
64,966
| 6,927,426,852
|
IssuesEvent
|
2017-11-30 22:49:48
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
System.Net.Security.Tests.dll FailFast on Debian
|
area-System.Net.Security os-linux test-run-core
|
This comes just after "starting KDC" from `src\System.Net.Security\tests\Scripts\Unix\setup-kdc.sh`
https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20170601.02/workItem/System.Net.Security.Tests/wilogs
```
Starting KDC..
FailFast: Invalid context passed to SafeDeleteNegoContext
at System.Diagnostics.Debug.Assert(Boolean condition, String message, String detailMessage)
at System.Net.Security.SafeDeleteNegoContext.SetGssContext(SafeGssContextHandle context) in /root/corefx/src/Common/src/System/Net/Security/Unix/SafeDeleteNegoContext.cs:line 52
at System.Net.Security.NegotiateStreamPal.EstablishSecurityContext(SafeFreeNegoCredentials credential, SafeDeleteContext& context, String targetName, ContextFlagsPal inFlags, SecurityBuffer inputBuffer, SecurityBuffer outputBuffer, ContextFlagsPal& outFlags) in /root/corefx/src/Common/src/System/Net/Security/NegotiateStreamPal.Unix.cs:line 197
at System.Net.Security.NegotiateStreamPal.InitializeSecurityContext(SafeFreeCredentials credentialsHandle, SafeDeleteContext& securityContext, String spn, ContextFlagsPal requestedContextFlags, SecurityBuffer[] inSecurityBufferArray, SecurityBuffer outSecurityBuffer, ContextFlagsPal& contextFlags) in /root/corefx/src/Common/src/System/Net/Security/NegotiateStreamPal.Unix.cs:line 240
at System.Net.NTAuthentication.GetOutgoingBlob(Byte[] incomingBlob, Boolean throwOnError, SecurityStatusPal& statusCode) in /root/corefx/src/Common/src/System/Net/NTAuthentication.Common.cs:line 243
at System.Net.Security.NegoState.GetOutgoingBlob(Byte[] incomingBlob, Exception& e) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 778
at System.Net.Security.NegoState.StartSendBlob(Byte[] message, LazyAsyncResult lazyResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 444
at System.Net.Security.NegoState.CheckCompletionBeforeNextSend(Byte[] message, LazyAsyncResult lazyResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 648
at System.Net.Security.NegoState.ProcessReceivedBlob(Byte[] message, LazyAsyncResult lazyResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 624
at System.Net.Security.NegoState.ReadCallback(IAsyncResult transportResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 756
at System.Net.LazyAsyncResult.Complete(IntPtr userToken) in /root/corefx/src/Common/src/System/Net/LazyAsyncResult.cs:line 428
at System.Net.LazyAsyncResult.ProtectedInvokeCallback(Object result, IntPtr userToken) in /root/corefx/src/Common/src/System/Net/LazyAsyncResult.cs:line 382
at System.Net.LazyAsyncResult.InvokeCallback(Object result) in /root/corefx/src/Common/src/System/Net/LazyAsyncResult.cs:line 389
at System.Net.StreamFramer.ReadFrameComplete(IAsyncResult transportResult) in /root/corefx/src/System.Net.Security/src/System/Net/StreamFramer.cs:line 270
at System.Net.StreamFramer.ReadFrameCallback(IAsyncResult transportResult) in /root/corefx/src/System.Net.Security/src/System/Net/StreamFramer.cs:line 168
at System.Threading.Tasks.TaskToApm.<>c__DisplayClass3_0.<InvokeCallbackWhenTaskCompletes>b__0() in /root/corefx/src/Common/src/System/Threading/Tasks/TaskToApm.cs:line 133
at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state)
at System.Threading.Tasks.AwaitTaskContinuation.RunCallback(ContextCallback callback, Object state, Task& currentTask)
at System.Threading.Tasks.Task.RunContinuations(Object continuationObject)
at System.Threading.Tasks.Task.ExecuteWithThreadLocal(Task& currentTaskSlot)
at System.Threading.ThreadPoolWorkQueue.Dispatch()
at System.Environment.FailFast(System.String, System.Exception)
at System.Net.Security.SafeDeleteNegoContext.SetGssContext(Microsoft.Win32.SafeHandles.SafeGssContextHandle)
at System.Net.Security.NegotiateStreamPal.EstablishSecurityContext(System.Net.Security.SafeFreeNegoCredentials, System.Net.Security.SafeDeleteContext ByRef, System.String, System.Net.ContextFlagsPal, System.Net.Security.SecurityBuffer, System.Net.Security.SecurityBuffer, System.Net.ContextFlagsPal ByRef)
at System.Net.Security.NegotiateStreamPal.InitializeSecurityContext(System.Net.Security.SafeFreeCredentials, System.Net.Security.SafeDeleteContext ByRef, System.String, System.Net.ContextFlagsPal, System.Net.Security.SecurityBuffer[], System.Net.Security.SecurityBuffer, System.Net.ContextFlagsPal ByRef)
at System.Net.NTAuthentication.GetOutgoingBlob(Byte[], Boolean, System.Net.SecurityStatusPal ByRef)
at System.Net.Security.NegoState.GetOutgoingBlob(Byte[], System.Exception ByRef)
at System.Net.Security.NegoState.StartSendBlob(Byte[], System.Net.LazyAsyncResult)
at System.Net.Security.NegoState.CheckCompletionBeforeNextSend(Byte[], System.Net.LazyAsyncResult)
at System.Net.Security.NegoState.ProcessReceivedBlob(Byte[], System.Net.LazyAsyncResult)
at System.Net.Security.NegoState.ReadCallback(System.IAsyncResult)
at System.Net.LazyAsyncResult.Complete(IntPtr)
at System.Net.LazyAsyncResult.ProtectedInvokeCallback(System.Object, IntPtr)
at System.Net.LazyAsyncResult.InvokeCallback(System.Object)
at System.Net.StreamFramer.ReadFrameComplete(System.IAsyncResult)
at System.Net.StreamFramer.ReadFrameCallback(System.IAsyncResult)
at System.Threading.Tasks.TaskToApm+<>c__DisplayClass3_0.<InvokeCallbackWhenTaskCompletes>b__0()
at System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object)
at System.Threading.Tasks.AwaitTaskContinuation.RunCallback(System.Threading.ContextCallback, System.Object, System.Threading.Tasks.Task ByRef)
at System.Threading.Tasks.Task.RunContinuations(System.Object)
at System.Threading.Tasks.Task.ExecuteWithThreadLocal(System.Threading.Tasks.Task ByRef)
at System.Threading.ThreadPoolWorkQueue.Dispatch()
/home/helixbot/dotnetbuild/work/a3f18f62-593d-45f2-985d-8cbdf5c68e39/Work/e26112cb-e993-41f5-aa16-81e08426a81c/Unzip/RunTests.sh: line 89: 4484 Aborted (core dumped) $RUNTIME_PATH/dotnet xunit.console.netcore.exe System.Net.Security.Tests.dll -xml testResults.xml -notrait category=nonnetcoreapptests -notrait category=nonlinuxtests -notrait category=failing
command exited with ExitCode: 134
```
|
1.0
|
System.Net.Security.Tests.dll FailFast on Debian - This comes just after "starting KDC" from `src\System.Net.Security\tests\Scripts\Unix\setup-kdc.sh`
https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20170601.02/workItem/System.Net.Security.Tests/wilogs
```
Starting KDC..
FailFast: Invalid context passed to SafeDeleteNegoContext
at System.Diagnostics.Debug.Assert(Boolean condition, String message, String detailMessage)
at System.Net.Security.SafeDeleteNegoContext.SetGssContext(SafeGssContextHandle context) in /root/corefx/src/Common/src/System/Net/Security/Unix/SafeDeleteNegoContext.cs:line 52
at System.Net.Security.NegotiateStreamPal.EstablishSecurityContext(SafeFreeNegoCredentials credential, SafeDeleteContext& context, String targetName, ContextFlagsPal inFlags, SecurityBuffer inputBuffer, SecurityBuffer outputBuffer, ContextFlagsPal& outFlags) in /root/corefx/src/Common/src/System/Net/Security/NegotiateStreamPal.Unix.cs:line 197
at System.Net.Security.NegotiateStreamPal.InitializeSecurityContext(SafeFreeCredentials credentialsHandle, SafeDeleteContext& securityContext, String spn, ContextFlagsPal requestedContextFlags, SecurityBuffer[] inSecurityBufferArray, SecurityBuffer outSecurityBuffer, ContextFlagsPal& contextFlags) in /root/corefx/src/Common/src/System/Net/Security/NegotiateStreamPal.Unix.cs:line 240
at System.Net.NTAuthentication.GetOutgoingBlob(Byte[] incomingBlob, Boolean throwOnError, SecurityStatusPal& statusCode) in /root/corefx/src/Common/src/System/Net/NTAuthentication.Common.cs:line 243
at System.Net.Security.NegoState.GetOutgoingBlob(Byte[] incomingBlob, Exception& e) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 778
at System.Net.Security.NegoState.StartSendBlob(Byte[] message, LazyAsyncResult lazyResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 444
at System.Net.Security.NegoState.CheckCompletionBeforeNextSend(Byte[] message, LazyAsyncResult lazyResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 648
at System.Net.Security.NegoState.ProcessReceivedBlob(Byte[] message, LazyAsyncResult lazyResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 624
at System.Net.Security.NegoState.ReadCallback(IAsyncResult transportResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 756
at System.Net.LazyAsyncResult.Complete(IntPtr userToken) in /root/corefx/src/Common/src/System/Net/LazyAsyncResult.cs:line 428
at System.Net.LazyAsyncResult.ProtectedInvokeCallback(Object result, IntPtr userToken) in /root/corefx/src/Common/src/System/Net/LazyAsyncResult.cs:line 382
at System.Net.LazyAsyncResult.InvokeCallback(Object result) in /root/corefx/src/Common/src/System/Net/LazyAsyncResult.cs:line 389
at System.Net.StreamFramer.ReadFrameComplete(IAsyncResult transportResult) in /root/corefx/src/System.Net.Security/src/System/Net/StreamFramer.cs:line 270
at System.Net.StreamFramer.ReadFrameCallback(IAsyncResult transportResult) in /root/corefx/src/System.Net.Security/src/System/Net/StreamFramer.cs:line 168
at System.Threading.Tasks.TaskToApm.<>c__DisplayClass3_0.<InvokeCallbackWhenTaskCompletes>b__0() in /root/corefx/src/Common/src/System/Threading/Tasks/TaskToApm.cs:line 133
at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state)
at System.Threading.Tasks.AwaitTaskContinuation.RunCallback(ContextCallback callback, Object state, Task& currentTask)
at System.Threading.Tasks.Task.RunContinuations(Object continuationObject)
at System.Threading.Tasks.Task.ExecuteWithThreadLocal(Task& currentTaskSlot)
at System.Threading.ThreadPoolWorkQueue.Dispatch()
at System.Environment.FailFast(System.String, System.Exception)
at System.Net.Security.SafeDeleteNegoContext.SetGssContext(Microsoft.Win32.SafeHandles.SafeGssContextHandle)
at System.Net.Security.NegotiateStreamPal.EstablishSecurityContext(System.Net.Security.SafeFreeNegoCredentials, System.Net.Security.SafeDeleteContext ByRef, System.String, System.Net.ContextFlagsPal, System.Net.Security.SecurityBuffer, System.Net.Security.SecurityBuffer, System.Net.ContextFlagsPal ByRef)
at System.Net.Security.NegotiateStreamPal.InitializeSecurityContext(System.Net.Security.SafeFreeCredentials, System.Net.Security.SafeDeleteContext ByRef, System.String, System.Net.ContextFlagsPal, System.Net.Security.SecurityBuffer[], System.Net.Security.SecurityBuffer, System.Net.ContextFlagsPal ByRef)
at System.Net.NTAuthentication.GetOutgoingBlob(Byte[], Boolean, System.Net.SecurityStatusPal ByRef)
at System.Net.Security.NegoState.GetOutgoingBlob(Byte[], System.Exception ByRef)
at System.Net.Security.NegoState.StartSendBlob(Byte[], System.Net.LazyAsyncResult)
at System.Net.Security.NegoState.CheckCompletionBeforeNextSend(Byte[], System.Net.LazyAsyncResult)
at System.Net.Security.NegoState.ProcessReceivedBlob(Byte[], System.Net.LazyAsyncResult)
at System.Net.Security.NegoState.ReadCallback(System.IAsyncResult)
at System.Net.LazyAsyncResult.Complete(IntPtr)
at System.Net.LazyAsyncResult.ProtectedInvokeCallback(System.Object, IntPtr)
at System.Net.LazyAsyncResult.InvokeCallback(System.Object)
at System.Net.StreamFramer.ReadFrameComplete(System.IAsyncResult)
at System.Net.StreamFramer.ReadFrameCallback(System.IAsyncResult)
at System.Threading.Tasks.TaskToApm+<>c__DisplayClass3_0.<InvokeCallbackWhenTaskCompletes>b__0()
at System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object)
at System.Threading.Tasks.AwaitTaskContinuation.RunCallback(System.Threading.ContextCallback, System.Object, System.Threading.Tasks.Task ByRef)
at System.Threading.Tasks.Task.RunContinuations(System.Object)
at System.Threading.Tasks.Task.ExecuteWithThreadLocal(System.Threading.Tasks.Task ByRef)
at System.Threading.ThreadPoolWorkQueue.Dispatch()
/home/helixbot/dotnetbuild/work/a3f18f62-593d-45f2-985d-8cbdf5c68e39/Work/e26112cb-e993-41f5-aa16-81e08426a81c/Unzip/RunTests.sh: line 89: 4484 Aborted (core dumped) $RUNTIME_PATH/dotnet xunit.console.netcore.exe System.Net.Security.Tests.dll -xml testResults.xml -notrait category=nonnetcoreapptests -notrait category=nonlinuxtests -notrait category=failing
command exited with ExitCode: 134
```
|
test
|
system net security tests dll failfast on debian this comes just after starting kdc from src system net security tests scripts unix setup kdc sh starting kdc failfast invalid context passed to safedeletenegocontext at system diagnostics debug assert boolean condition string message string detailmessage at system net security safedeletenegocontext setgsscontext safegsscontexthandle context in root corefx src common src system net security unix safedeletenegocontext cs line at system net security negotiatestreampal establishsecuritycontext safefreenegocredentials credential safedeletecontext context string targetname contextflagspal inflags securitybuffer inputbuffer securitybuffer outputbuffer contextflagspal outflags in root corefx src common src system net security negotiatestreampal unix cs line at system net security negotiatestreampal initializesecuritycontext safefreecredentials credentialshandle safedeletecontext securitycontext string spn contextflagspal requestedcontextflags securitybuffer insecuritybufferarray securitybuffer outsecuritybuffer contextflagspal contextflags in root corefx src common src system net security negotiatestreampal unix cs line at system net ntauthentication getoutgoingblob byte incomingblob boolean throwonerror securitystatuspal statuscode in root corefx src common src system net ntauthentication common cs line at system net security negostate getoutgoingblob byte incomingblob exception e in root corefx src system net security src system net security negostate cs line at system net security negostate startsendblob byte message lazyasyncresult lazyresult in root corefx src system net security src system net security negostate cs line at system net security negostate checkcompletionbeforenextsend byte message lazyasyncresult lazyresult in root corefx src system net security src system net security negostate cs line at system net security negostate processreceivedblob byte message lazyasyncresult lazyresult in root corefx src system net security src system net security negostate cs line at system net security negostate readcallback iasyncresult transportresult in root corefx src system net security src system net security negostate cs line at system net lazyasyncresult complete intptr usertoken in root corefx src common src system net lazyasyncresult cs line at system net lazyasyncresult protectedinvokecallback object result intptr usertoken in root corefx src common src system net lazyasyncresult cs line at system net lazyasyncresult invokecallback object result in root corefx src common src system net lazyasyncresult cs line at system net streamframer readframecomplete iasyncresult transportresult in root corefx src system net security src system net streamframer cs line at system net streamframer readframecallback iasyncresult transportresult in root corefx src system net security src system net streamframer cs line at system threading tasks tasktoapm c b in root corefx src common src system threading tasks tasktoapm cs line at system threading executioncontext run executioncontext executioncontext contextcallback callback object state at system threading tasks awaittaskcontinuation runcallback contextcallback callback object state task currenttask at system threading tasks task runcontinuations object continuationobject at system threading tasks task executewiththreadlocal task currenttaskslot at system threading threadpoolworkqueue dispatch at system environment failfast system string system exception at system net security safedeletenegocontext setgsscontext microsoft safehandles safegsscontexthandle at system net security negotiatestreampal establishsecuritycontext system net security safefreenegocredentials system net security safedeletecontext byref system string system net contextflagspal system net security securitybuffer system net security securitybuffer system net contextflagspal byref at system net security negotiatestreampal initializesecuritycontext system net security safefreecredentials system net security safedeletecontext byref system string system net contextflagspal system net security securitybuffer system net security securitybuffer system net contextflagspal byref at system net ntauthentication getoutgoingblob byte boolean system net securitystatuspal byref at system net security negostate getoutgoingblob byte system exception byref at system net security negostate startsendblob byte system net lazyasyncresult at system net security negostate checkcompletionbeforenextsend byte system net lazyasyncresult at system net security negostate processreceivedblob byte system net lazyasyncresult at system net security negostate readcallback system iasyncresult at system net lazyasyncresult complete intptr at system net lazyasyncresult protectedinvokecallback system object intptr at system net lazyasyncresult invokecallback system object at system net streamframer readframecomplete system iasyncresult at system net streamframer readframecallback system iasyncresult at system threading tasks tasktoapm c b at system threading executioncontext run system threading executioncontext system threading contextcallback system object at system threading tasks awaittaskcontinuation runcallback system threading contextcallback system object system threading tasks task byref at system threading tasks task runcontinuations system object at system threading tasks task executewiththreadlocal system threading tasks task byref at system threading threadpoolworkqueue dispatch home helixbot dotnetbuild work work unzip runtests sh line aborted core dumped runtime path dotnet xunit console netcore exe system net security tests dll xml testresults xml notrait category nonnetcoreapptests notrait category nonlinuxtests notrait category failing command exited with exitcode
| 1
|
74,347
| 20,142,018,073
|
IssuesEvent
|
2022-02-09 00:54:32
|
microsoft/PowerToys
|
https://api.github.com/repos/microsoft/PowerToys
|
closed
|
[Build] Build failed: Path to exceeds max length
|
Issue-Bug Product-PowerToys Run Area-Build Needs-Triage
|
### Microsoft PowerToys version
f2a3fa5ec68d1b07ab347beadc8f6e9160069cce
### Running as admin
- [ ] Yes
### Area(s) with issue?
PowerToys Run
### Steps to reproduce
Build Launcher.
### ✔️ Expected Behavior
Build works.
### ❌ Actual Behavior
```
Severity Code Description Project File Line Suppression State
Error MSB3491 Could not write lines to file "obj\x64\Debug\net5.0-windows10.0.18362.0\Microsoft.PowerToys.Run.Plugin.WindowsTerminal.UnitTests.GeneratedMSBuildEditorConfig.editorconfig". Path: obj\x64\Debug\net5.0-windows10.0.18362.0\Microsoft.PowerToys.Run.Plugin.WindowsTerminal.UnitTests.GeneratedMSBuildEditorConfig.editorconfig exceeds the OS max path limit. The fully qualified file name must be less than 260 characters. Microsoft.PowerToys.Run.Plugin.WindowsTerminal.UnitTests C:\Program Files (x86)\Microsoft Visual Studio\2019\Community\MSBuild\Current\Bin\Roslyn\Microsoft.Managed.Core.targets 150
```
### Other Software
_No response_
|
1.0
|
[Build] Build failed: Path to exceeds max length - ### Microsoft PowerToys version
f2a3fa5ec68d1b07ab347beadc8f6e9160069cce
### Running as admin
- [ ] Yes
### Area(s) with issue?
PowerToys Run
### Steps to reproduce
Build Launcher.
### ✔️ Expected Behavior
Build works.
### ❌ Actual Behavior
```
Severity Code Description Project File Line Suppression State
Error MSB3491 Could not write lines to file "obj\x64\Debug\net5.0-windows10.0.18362.0\Microsoft.PowerToys.Run.Plugin.WindowsTerminal.UnitTests.GeneratedMSBuildEditorConfig.editorconfig". Path: obj\x64\Debug\net5.0-windows10.0.18362.0\Microsoft.PowerToys.Run.Plugin.WindowsTerminal.UnitTests.GeneratedMSBuildEditorConfig.editorconfig exceeds the OS max path limit. The fully qualified file name must be less than 260 characters. Microsoft.PowerToys.Run.Plugin.WindowsTerminal.UnitTests C:\Program Files (x86)\Microsoft Visual Studio\2019\Community\MSBuild\Current\Bin\Roslyn\Microsoft.Managed.Core.targets 150
```
### Other Software
_No response_
|
non_test
|
build failed path to exceeds max length microsoft powertoys version running as admin yes area s with issue powertoys run steps to reproduce build launcher ✔️ expected behavior build works ❌ actual behavior severity code description project file line suppression state error could not write lines to file obj debug microsoft powertoys run plugin windowsterminal unittests generatedmsbuildeditorconfig editorconfig path obj debug microsoft powertoys run plugin windowsterminal unittests generatedmsbuildeditorconfig editorconfig exceeds the os max path limit the fully qualified file name must be less than characters microsoft powertoys run plugin windowsterminal unittests c program files microsoft visual studio community msbuild current bin roslyn microsoft managed core targets other software no response
| 0
|
90,310
| 8,233,447,323
|
IssuesEvent
|
2018-09-08 01:02:05
|
knative/serving
|
https://api.github.com/repos/knative/serving
|
closed
|
We should return route domain for getRouteDomain() in route_test
|
area/test-and-release kind/bug
| ERROR: type should be string, got "\r\nhttps://github.com/knative/serving/pull/1894 updated the getRouteDomain() method to poll for route state, but does not return domain. \r\n\r\nhttps://github.com/knative/serving/blob/master/test/conformance/route_test.go#L130"
|
1.0
|
We should return route domain for getRouteDomain() in route_test -
https://github.com/knative/serving/pull/1894 updated the getRouteDomain() method to poll for route state, but does not return domain.
https://github.com/knative/serving/blob/master/test/conformance/route_test.go#L130
|
test
|
we should return route domain for getroutedomain in route test updated the getroutedomain method to poll for route state but does not return domain
| 1
|
123,653
| 12,214,970,041
|
IssuesEvent
|
2020-05-01 11:32:35
|
hwanglight/ImageAugmentation
|
https://api.github.com/repos/hwanglight/ImageAugmentation
|
opened
|
Написать Readme
|
documentation
|
Readme должен содержать
1. Пререквизиты (библиотеки, которые нужно поставить, что собрать приложение)
2. Инструкция по сборке (какие команды выполнить в терминале, чтобы собрать приложение)
3. Инструкция по запуску и работе с консольной версией: формат параметров командной строки, формат конфигурационного файла (и закоммитить пример конфига)
4. Инструкция по работе с графической версией
|
1.0
|
Написать Readme - Readme должен содержать
1. Пререквизиты (библиотеки, которые нужно поставить, что собрать приложение)
2. Инструкция по сборке (какие команды выполнить в терминале, чтобы собрать приложение)
3. Инструкция по запуску и работе с консольной версией: формат параметров командной строки, формат конфигурационного файла (и закоммитить пример конфига)
4. Инструкция по работе с графической версией
|
non_test
|
написать readme readme должен содержать пререквизиты библиотеки которые нужно поставить что собрать приложение инструкция по сборке какие команды выполнить в терминале чтобы собрать приложение инструкция по запуску и работе с консольной версией формат параметров командной строки формат конфигурационного файла и закоммитить пример конфига инструкция по работе с графической версией
| 0
|
766,330
| 26,879,211,720
|
IssuesEvent
|
2023-02-05 12:32:06
|
herdstat/herdstat
|
https://api.github.com/repos/herdstat/herdstat
|
closed
|
Support more than commits as contributions
|
kind/feature priority/important-soon triage/accepted
|
The GitHub contribution graph includes much more than commits when computing contributions (see [here](https://docs.github.com/articles/why-are-my-contributions-not-showing-up-on-my-profile)). This would also be very cool for the `herdstat`contribution graph as these other things are also very important types of contributions.
As a first step, we will include opened issues and PRs.
|
1.0
|
Support more than commits as contributions - The GitHub contribution graph includes much more than commits when computing contributions (see [here](https://docs.github.com/articles/why-are-my-contributions-not-showing-up-on-my-profile)). This would also be very cool for the `herdstat`contribution graph as these other things are also very important types of contributions.
As a first step, we will include opened issues and PRs.
|
non_test
|
support more than commits as contributions the github contribution graph includes much more than commits when computing contributions see this would also be very cool for the herdstat contribution graph as these other things are also very important types of contributions as a first step we will include opened issues and prs
| 0
|
111,795
| 9,541,712,640
|
IssuesEvent
|
2019-04-30 23:30:36
|
Greenstand/treetracker-android
|
https://api.github.com/repos/Greenstand/treetracker-android
|
closed
|
Add Swahili Translation
|
android enhancement in testing
|
Update--boresha
Upload --pakia
Uploaded-imepakiwa
GPS--GPS
Accuracy --usahihi
GPS accuracy --Usahihi wa GPS
Meters--mita
Sync-sawazisha
To sync--yakusawazisha.
Resume--endeleza
Pause--katisha
Data-data
About--Kuhusu
Map--ramani
Your note--maoni yako
New tree--mti mpya
Insufficient--haitoshi
Insufficient GPS accuracy-- GPS sio sahihi
Save--hifadhi.
Information--maelekezo
The following is the translation of the information page
Kuhusu.
Tree Tracker planter app ni kifaa cha kufuatilia miti, kilichotengenezwa ili kuwawezesha watu na mashirika kupanda, kufuatilia na kulinda miti.
mpangilio wa kifaa chako unahitaji usahihi wa GPS wa (mita10).
Ili kufahamu kwa undani zaidi ni kwa jinsi gani kifaa hichi kitanufaisha mradi wako wa upandaji miti, tembelea tovuti ifuatayo.
www.Greenstand.org
kifaa hichi bado kipo kwenye matengenezo. kama una maombi yeyote au kama kifaa chako kimeshindwa kufanya kazi vizuri, toa taarifa kupitia e-mail ifuatayo.
info@greenstand.org
Treetracker(kwa majaribio)1:2:1
awamu ya 31
|
1.0
|
Add Swahili Translation - Update--boresha
Upload --pakia
Uploaded-imepakiwa
GPS--GPS
Accuracy --usahihi
GPS accuracy --Usahihi wa GPS
Meters--mita
Sync-sawazisha
To sync--yakusawazisha.
Resume--endeleza
Pause--katisha
Data-data
About--Kuhusu
Map--ramani
Your note--maoni yako
New tree--mti mpya
Insufficient--haitoshi
Insufficient GPS accuracy-- GPS sio sahihi
Save--hifadhi.
Information--maelekezo
The following is the translation of the information page
Kuhusu.
Tree Tracker planter app ni kifaa cha kufuatilia miti, kilichotengenezwa ili kuwawezesha watu na mashirika kupanda, kufuatilia na kulinda miti.
mpangilio wa kifaa chako unahitaji usahihi wa GPS wa (mita10).
Ili kufahamu kwa undani zaidi ni kwa jinsi gani kifaa hichi kitanufaisha mradi wako wa upandaji miti, tembelea tovuti ifuatayo.
www.Greenstand.org
kifaa hichi bado kipo kwenye matengenezo. kama una maombi yeyote au kama kifaa chako kimeshindwa kufanya kazi vizuri, toa taarifa kupitia e-mail ifuatayo.
info@greenstand.org
Treetracker(kwa majaribio)1:2:1
awamu ya 31
|
test
|
add swahili translation update boresha upload pakia uploaded imepakiwa gps gps accuracy usahihi gps accuracy usahihi wa gps meters mita sync sawazisha to sync yakusawazisha resume endeleza pause katisha data data about kuhusu map ramani your note maoni yako new tree mti mpya insufficient haitoshi insufficient gps accuracy gps sio sahihi save hifadhi information maelekezo the following is the translation of the information page kuhusu tree tracker planter app ni kifaa cha kufuatilia miti kilichotengenezwa ili kuwawezesha watu na mashirika kupanda kufuatilia na kulinda miti mpangilio wa kifaa chako unahitaji usahihi wa gps wa ili kufahamu kwa undani zaidi ni kwa jinsi gani kifaa hichi kitanufaisha mradi wako wa upandaji miti tembelea tovuti ifuatayo kifaa hichi bado kipo kwenye matengenezo kama una maombi yeyote au kama kifaa chako kimeshindwa kufanya kazi vizuri toa taarifa kupitia e mail ifuatayo info greenstand org treetracker kwa majaribio awamu ya
| 1
|
243,611
| 18,719,295,598
|
IssuesEvent
|
2021-11-03 09:55:33
|
ebi-ait/hca-ebi-wrangler-central
|
https://api.github.com/repos/ebi-ait/hca-ebi-wrangler-central
|
closed
|
Review HCA_to_scea documentation
|
documentation operations
|
We need to review https://ebi-ait.github.io/hca-ebi-wrangler-central/SOPs/hca_to_scea_tools_SOP.html and make sure that the whole process can be run without previous knowledge of the tools.
Identified missing bits:
- How to assign a HCAD accession
- How to send fastq files
- How and when to update the ticket with information about process in SCEA
- How to get updates about process on the SCEA dataset
- How to check a dataset has been successfully brokered into SCEA
**Acceptability criteria**:
- [x] The documentation has been reviewed and modified
- [x] A tool-agnostic wrangler reviews the SOP after doing a full conversion of a dataset and lists anything missing in this ticket
- [x] Ami will again update the documentation to cover any missing or unclear information in a PR
- [x] Another tool-agnostic wrangler reviews the SOP and if needed, suggests further changes, or closes this ticket to reflect it is done.
|
1.0
|
Review HCA_to_scea documentation - We need to review https://ebi-ait.github.io/hca-ebi-wrangler-central/SOPs/hca_to_scea_tools_SOP.html and make sure that the whole process can be run without previous knowledge of the tools.
Identified missing bits:
- How to assign a HCAD accession
- How to send fastq files
- How and when to update the ticket with information about process in SCEA
- How to get updates about process on the SCEA dataset
- How to check a dataset has been successfully brokered into SCEA
**Acceptability criteria**:
- [x] The documentation has been reviewed and modified
- [x] A tool-agnostic wrangler reviews the SOP after doing a full conversion of a dataset and lists anything missing in this ticket
- [x] Ami will again update the documentation to cover any missing or unclear information in a PR
- [x] Another tool-agnostic wrangler reviews the SOP and if needed, suggests further changes, or closes this ticket to reflect it is done.
|
non_test
|
review hca to scea documentation we need to review and make sure that the whole process can be run without previous knowledge of the tools identified missing bits how to assign a hcad accession how to send fastq files how and when to update the ticket with information about process in scea how to get updates about process on the scea dataset how to check a dataset has been successfully brokered into scea acceptability criteria the documentation has been reviewed and modified a tool agnostic wrangler reviews the sop after doing a full conversion of a dataset and lists anything missing in this ticket ami will again update the documentation to cover any missing or unclear information in a pr another tool agnostic wrangler reviews the sop and if needed suggests further changes or closes this ticket to reflect it is done
| 0
|
253,893
| 21,712,543,502
|
IssuesEvent
|
2022-05-10 14:57:42
|
aldrichtr/infraspective
|
https://api.github.com/repos/aldrichtr/infraspective
|
opened
|
Refactor unit tests so they do not rely on the module being loaded
|
>refactor @tests
|
The Pester unit tests should load the file that we want to test, and mock any supporting functions so that they are truly _unit_ tests. If the module is loaded first, then most of the functions will be coupled with their supporting functions. Those types of tests should be in **integration** not **unit** tests.
|
1.0
|
Refactor unit tests so they do not rely on the module being loaded - The Pester unit tests should load the file that we want to test, and mock any supporting functions so that they are truly _unit_ tests. If the module is loaded first, then most of the functions will be coupled with their supporting functions. Those types of tests should be in **integration** not **unit** tests.
|
test
|
refactor unit tests so they do not rely on the module being loaded the pester unit tests should load the file that we want to test and mock any supporting functions so that they are truly unit tests if the module is loaded first then most of the functions will be coupled with their supporting functions those types of tests should be in integration not unit tests
| 1
|
261,299
| 22,717,766,912
|
IssuesEvent
|
2022-07-06 05:00:28
|
Merck/metalite.ae
|
https://api.github.com/repos/Merck/metalite.ae
|
closed
|
Independent Testing of fmt_ci.R
|
independent test
|
Test plan of `fmt_ci`:
- fmt_ci(-1.96,1.96, digits = 1) -> (-2.0, 2.0)
|
1.0
|
Independent Testing of fmt_ci.R - Test plan of `fmt_ci`:
- fmt_ci(-1.96,1.96, digits = 1) -> (-2.0, 2.0)
|
test
|
independent testing of fmt ci r test plan of fmt ci fmt ci digits
| 1
|
101,934
| 12,731,490,828
|
IssuesEvent
|
2020-06-25 08:57:47
|
dambem/ClimateMonitorV2
|
https://api.github.com/repos/dambem/ClimateMonitorV2
|
closed
|
LuftDaten Custom Map Markers
|
Design Improvements
|
## Description
As we're gonna start using weather company data in conjunction with luftdaten info, the current circular map markers won't do! Instead we need to make some custom markers so that it's clear it's luftdaten info.
## Steps to reproduce
## The actual vs expected behaviour
## Additional Information
|
1.0
|
LuftDaten Custom Map Markers - ## Description
As we're gonna start using weather company data in conjunction with luftdaten info, the current circular map markers won't do! Instead we need to make some custom markers so that it's clear it's luftdaten info.
## Steps to reproduce
## The actual vs expected behaviour
## Additional Information
|
non_test
|
luftdaten custom map markers description as we re gonna start using weather company data in conjunction with luftdaten info the current circular map markers won t do instead we need to make some custom markers so that it s clear it s luftdaten info steps to reproduce the actual vs expected behaviour additional information
| 0
|
98,439
| 8,677,233,636
|
IssuesEvent
|
2018-11-30 16:13:40
|
cerner/terra-framework
|
https://api.github.com/repos/cerner/terra-framework
|
closed
|
Use themeCombinationOfCustomProperties
|
Orion Reviewed beginner-friendly testing
|
# Feature Request
Some component WDIO tests use `themeEachCustomProperty` to test theme variables. We should evaluate if component tests can use `themeCombinationOfCustomProperties`. This will drastically speed up visually regression tests by taking a single screenshot for a set of theme variables, versus a screenshot for each theme variable.
|
1.0
|
Use themeCombinationOfCustomProperties - # Feature Request
Some component WDIO tests use `themeEachCustomProperty` to test theme variables. We should evaluate if component tests can use `themeCombinationOfCustomProperties`. This will drastically speed up visually regression tests by taking a single screenshot for a set of theme variables, versus a screenshot for each theme variable.
|
test
|
use themecombinationofcustomproperties feature request some component wdio tests use themeeachcustomproperty to test theme variables we should evaluate if component tests can use themecombinationofcustomproperties this will drastically speed up visually regression tests by taking a single screenshot for a set of theme variables versus a screenshot for each theme variable
| 1
|
430,830
| 12,466,664,405
|
IssuesEvent
|
2020-05-28 15:49:06
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
IPv6 prefix could be added multiple times to prefix timer list
|
area: Networking bug priority: medium
|
**Describe the bug**
We could add same IPv6 prefix multiple times to prefix timer list. This can lead to denial-of-service issue if we receive suitable crafted IPv6 RA packets.
|
1.0
|
IPv6 prefix could be added multiple times to prefix timer list - **Describe the bug**
We could add same IPv6 prefix multiple times to prefix timer list. This can lead to denial-of-service issue if we receive suitable crafted IPv6 RA packets.
|
non_test
|
prefix could be added multiple times to prefix timer list describe the bug we could add same prefix multiple times to prefix timer list this can lead to denial of service issue if we receive suitable crafted ra packets
| 0
|
130,108
| 10,596,758,032
|
IssuesEvent
|
2019-10-09 22:06:33
|
inspec/inspec
|
https://api.github.com/repos/inspec/inspec
|
closed
|
Improve integration testing of databases
|
Aspect: Testing Platform: Databases Type: Enhancement
|
### 🎛 Description
🙋 feature request
For database-targeted resources such as postgres_session, oracledb_session, etc , no formal integration testing exists. Some unit testing exists, but unit testing simulates responses from outside systems. Many recent issues have been reported which would have been caught by testing with real databases, such as:
* [ ] #3027
* [ ] #3659
* [ ] #3177
* [ ] #3680
* [ ] #3255
* [ ] #3594
Fixing these issues is also difficult since we don't have a repeatable way of creating a test fixture database. It also prevents regression testing.
### 🌍 InSpec and Platform Version
3.1.3
### 💁 Possible Solutions
It should be exposed through Rake, but not be a default test job. `rake test:db` seems reasonable.
Ideally, we'd have something (vagrant, RDS, docker?) that would be able to standup one system of each of the 4 we support; possibly running on multiple platforms (oracle on windows and linux, for example) or using different connection techniques (different CLI clients vs internal ruby API clients).
We'd like to to be able to isolate a setup, as well.
It doesn't need to be fast, but repeatability and maintainability are important.
|
1.0
|
Improve integration testing of databases - ### 🎛 Description
🙋 feature request
For database-targeted resources such as postgres_session, oracledb_session, etc , no formal integration testing exists. Some unit testing exists, but unit testing simulates responses from outside systems. Many recent issues have been reported which would have been caught by testing with real databases, such as:
* [ ] #3027
* [ ] #3659
* [ ] #3177
* [ ] #3680
* [ ] #3255
* [ ] #3594
Fixing these issues is also difficult since we don't have a repeatable way of creating a test fixture database. It also prevents regression testing.
### 🌍 InSpec and Platform Version
3.1.3
### 💁 Possible Solutions
It should be exposed through Rake, but not be a default test job. `rake test:db` seems reasonable.
Ideally, we'd have something (vagrant, RDS, docker?) that would be able to standup one system of each of the 4 we support; possibly running on multiple platforms (oracle on windows and linux, for example) or using different connection techniques (different CLI clients vs internal ruby API clients).
We'd like to to be able to isolate a setup, as well.
It doesn't need to be fast, but repeatability and maintainability are important.
|
test
|
improve integration testing of databases 🎛 description 🙋 feature request for database targeted resources such as postgres session oracledb session etc no formal integration testing exists some unit testing exists but unit testing simulates responses from outside systems many recent issues have been reported which would have been caught by testing with real databases such as fixing these issues is also difficult since we don t have a repeatable way of creating a test fixture database it also prevents regression testing 🌍 inspec and platform version 💁 possible solutions it should be exposed through rake but not be a default test job rake test db seems reasonable ideally we d have something vagrant rds docker that would be able to standup one system of each of the we support possibly running on multiple platforms oracle on windows and linux for example or using different connection techniques different cli clients vs internal ruby api clients we d like to to be able to isolate a setup as well it doesn t need to be fast but repeatability and maintainability are important
| 1
|
414,828
| 28,005,942,718
|
IssuesEvent
|
2023-03-27 15:15:51
|
csia-pme/a-guide-to-mlops
|
https://api.github.com/repos/csia-pme/a-guide-to-mlops
|
closed
|
Rewrite the guide in a more formal way
|
documentation enhancement
|
As discussed with @rmarquis, it could be a good idea to rewrite some parts of the guide in a more formal way.
## Examples
### Chapter 2: Share your ML experiment code with Git
"Instead of relying on ZIP archives, **we'll** create a Git repository to enable easy collaboration with the rest of the team"
"Instead of relying on ZIP archives, **we will** create a Git repository to enable easy collaboration with the rest of the team"
### Chapter 5: Track model evolutions with DVC
"Once this stage is created, **you'll** be able to change our model's configuration, evaluate the new configuration and compare its performance with the last commited ones."
"Once this stage is created, **you will** be able to change our model's configuration, evaluate the new configuration and compare its performance with the last commited ones."
|
1.0
|
Rewrite the guide in a more formal way - As discussed with @rmarquis, it could be a good idea to rewrite some parts of the guide in a more formal way.
## Examples
### Chapter 2: Share your ML experiment code with Git
"Instead of relying on ZIP archives, **we'll** create a Git repository to enable easy collaboration with the rest of the team"
"Instead of relying on ZIP archives, **we will** create a Git repository to enable easy collaboration with the rest of the team"
### Chapter 5: Track model evolutions with DVC
"Once this stage is created, **you'll** be able to change our model's configuration, evaluate the new configuration and compare its performance with the last commited ones."
"Once this stage is created, **you will** be able to change our model's configuration, evaluate the new configuration and compare its performance with the last commited ones."
|
non_test
|
rewrite the guide in a more formal way as discussed with rmarquis it could be a good idea to rewrite some parts of the guide in a more formal way examples chapter share your ml experiment code with git instead of relying on zip archives we ll create a git repository to enable easy collaboration with the rest of the team instead of relying on zip archives we will create a git repository to enable easy collaboration with the rest of the team chapter track model evolutions with dvc once this stage is created you ll be able to change our model s configuration evaluate the new configuration and compare its performance with the last commited ones once this stage is created you will be able to change our model s configuration evaluate the new configuration and compare its performance with the last commited ones
| 0
|
58,926
| 11,911,642,753
|
IssuesEvent
|
2020-03-31 08:58:04
|
ModellingWebLab/weblab-fc
|
https://api.github.com/repos/ModellingWebLab/weblab-fc
|
closed
|
Cut dependency on pycml
|
code-and-design install
|
At present running an experiment requires pycml to generate the manipulated model. This will change to use cellmlmanip & fccodegen.
- [x] As a first step, put a manually generated (using fccodegen) model in the tests and see if we can just simulate it.
- [x] Then call this automatically rather than calling pycml.
- [x] Then iteratively implement model manipulations. See #23.
|
1.0
|
Cut dependency on pycml - At present running an experiment requires pycml to generate the manipulated model. This will change to use cellmlmanip & fccodegen.
- [x] As a first step, put a manually generated (using fccodegen) model in the tests and see if we can just simulate it.
- [x] Then call this automatically rather than calling pycml.
- [x] Then iteratively implement model manipulations. See #23.
|
non_test
|
cut dependency on pycml at present running an experiment requires pycml to generate the manipulated model this will change to use cellmlmanip fccodegen as a first step put a manually generated using fccodegen model in the tests and see if we can just simulate it then call this automatically rather than calling pycml then iteratively implement model manipulations see
| 0
|
4,842
| 7,325,304,980
|
IssuesEvent
|
2018-03-03 07:03:45
|
renovateapp/renovate
|
https://api.github.com/repos/renovateapp/renovate
|
closed
|
bug: Branch prefix not working with gitLab platform
|
needs-requirements pri4-low
|
I have just realized that even through I was always setting the branchPrefix option in the config.js, all dependency level branches were still using the default /renovate/eslint... format.
It looks like the branchPrefix config option is never read or it gets overwritten somewhere.
`config.branchPrefix` returns undefined.
Setting in config.js, for example:
`branchPrefix: "chore/WQ-0_renovate_",`
Results in default branch format, such as: renovate/eslint-4.x
The only way to get around it is to actually overwrite the branchName in the config:
`branchName: "chore/WQ-0_renovate_{{depName}}-{{newVersionMajor}}.x",`
This will work and correct branch names are used.
|
1.0
|
bug: Branch prefix not working with gitLab platform - I have just realized that even through I was always setting the branchPrefix option in the config.js, all dependency level branches were still using the default /renovate/eslint... format.
It looks like the branchPrefix config option is never read or it gets overwritten somewhere.
`config.branchPrefix` returns undefined.
Setting in config.js, for example:
`branchPrefix: "chore/WQ-0_renovate_",`
Results in default branch format, such as: renovate/eslint-4.x
The only way to get around it is to actually overwrite the branchName in the config:
`branchName: "chore/WQ-0_renovate_{{depName}}-{{newVersionMajor}}.x",`
This will work and correct branch names are used.
|
non_test
|
bug branch prefix not working with gitlab platform i have just realized that even through i was always setting the branchprefix option in the config js all dependency level branches were still using the default renovate eslint format it looks like the branchprefix config option is never read or it gets overwritten somewhere config branchprefix returns undefined setting in config js for example branchprefix chore wq renovate results in default branch format such as renovate eslint x the only way to get around it is to actually overwrite the branchname in the config branchname chore wq renovate depname newversionmajor x this will work and correct branch names are used
| 0
|
48,346
| 7,422,610,780
|
IssuesEvent
|
2018-03-23 00:15:57
|
OneDrive/onedrive-api-docs
|
https://api.github.com/repos/OneDrive/onedrive-api-docs
|
closed
|
Webhooks on items (not "/me/drive/root")
|
documentation question
|
I notice that the Update webhooks has a `PATCH /drive/items/{item-id}/subscriptions/{id}` request where the response has a value for resource of ` "resource": "/drives/0/items/57705F13F13C3C0C!104",`. Does this mean that webhooks can be created for specific item IDs?
When I try to send a create request with a specific item-id, I get this:
```
{
"error": {
"code": "InvalidRequest",
"message": "resource '/drive/items/{item-id}/' is not supported.",
"innerError": {
"request-id": "793ba53f-a055-4000-aa69-25e775483d26",
"date": "2017-08-21T17:04:25"
}
}
}
```
I'm only able to subscribe to the root folder (`"/me/drive/root"`).
|
1.0
|
Webhooks on items (not "/me/drive/root") - I notice that the Update webhooks has a `PATCH /drive/items/{item-id}/subscriptions/{id}` request where the response has a value for resource of ` "resource": "/drives/0/items/57705F13F13C3C0C!104",`. Does this mean that webhooks can be created for specific item IDs?
When I try to send a create request with a specific item-id, I get this:
```
{
"error": {
"code": "InvalidRequest",
"message": "resource '/drive/items/{item-id}/' is not supported.",
"innerError": {
"request-id": "793ba53f-a055-4000-aa69-25e775483d26",
"date": "2017-08-21T17:04:25"
}
}
}
```
I'm only able to subscribe to the root folder (`"/me/drive/root"`).
|
non_test
|
webhooks on items not me drive root i notice that the update webhooks has a patch drive items item id subscriptions id request where the response has a value for resource of resource drives items does this mean that webhooks can be created for specific item ids when i try to send a create request with a specific item id i get this error code invalidrequest message resource drive items item id is not supported innererror request id date i m only able to subscribe to the root folder me drive root
| 0
|
203,147
| 15,352,105,944
|
IssuesEvent
|
2021-03-01 06:25:10
|
apache/apisix-dashboard
|
https://api.github.com/repos/apache/apisix-dashboard
|
closed
|
test: add e2e test for dubbo
|
testcase
|
after PR https://github.com/apache/apisix/pull/3224 merged, we should add some test cases for it. and we need build a test env for it.
|
1.0
|
test: add e2e test for dubbo - after PR https://github.com/apache/apisix/pull/3224 merged, we should add some test cases for it. and we need build a test env for it.
|
test
|
test add test for dubbo after pr merged we should add some test cases for it and we need build a test env for it
| 1
|
339,960
| 30,488,783,072
|
IssuesEvent
|
2023-07-18 05:55:49
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: alterpk-bank failed
|
C-test-failure O-robot O-roachtest branch-master release-blocker T-sql-foundations
|
roachtest.alterpk-bank [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=artifacts#/alterpk-bank) on master @ [7675ca4998134028f0623e04737b5cb69fcc33a9](https://github.com/cockroachdb/cockroach/commits/7675ca4998134028f0623e04737b5cb69fcc33a9):
```
(cluster.go:2180).Start: ~ COCKROACH_CONNECT_TIMEOUT=1200 ./cockroach sql --url 'postgres://root@localhost:26257?sslmode=disable' -e "CREATE SCHEDULE IF NOT EXISTS test_only_backup FOR BACKUP INTO 'gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-06-n4cpu4/1689659690804569106?AUTH=implicit' RECURRING '*/15 * * * *' FULL BACKUP '@hourly' WITH SCHEDULE OPTIONS first_run = 'now'"
ERROR: unexpected error occurred when checking for existing backups in gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-06-n4cpu4/1689659690804569106?AUTH=implicit: unable to list files in gcs bucket: googleapi: Error 403: 21965078311-compute@developer.gserviceaccount.com does not have storage.objects.list access to the Google Cloud Storage bucket. Permission 'storage.objects.list' denied on resource (or it may not exist).
SQLSTATE: 58030
Failed running "sql": COMMAND_PROBLEM: exit status 1
test artifacts and logs in: /artifacts/alterpk-bank/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-foundations
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*alterpk-bank.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
2.0
|
roachtest: alterpk-bank failed - roachtest.alterpk-bank [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=artifacts#/alterpk-bank) on master @ [7675ca4998134028f0623e04737b5cb69fcc33a9](https://github.com/cockroachdb/cockroach/commits/7675ca4998134028f0623e04737b5cb69fcc33a9):
```
(cluster.go:2180).Start: ~ COCKROACH_CONNECT_TIMEOUT=1200 ./cockroach sql --url 'postgres://root@localhost:26257?sslmode=disable' -e "CREATE SCHEDULE IF NOT EXISTS test_only_backup FOR BACKUP INTO 'gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-06-n4cpu4/1689659690804569106?AUTH=implicit' RECURRING '*/15 * * * *' FULL BACKUP '@hourly' WITH SCHEDULE OPTIONS first_run = 'now'"
ERROR: unexpected error occurred when checking for existing backups in gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-06-n4cpu4/1689659690804569106?AUTH=implicit: unable to list files in gcs bucket: googleapi: Error 403: 21965078311-compute@developer.gserviceaccount.com does not have storage.objects.list access to the Google Cloud Storage bucket. Permission 'storage.objects.list' denied on resource (or it may not exist).
SQLSTATE: 58030
Failed running "sql": COMMAND_PROBLEM: exit status 1
test artifacts and logs in: /artifacts/alterpk-bank/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-foundations
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*alterpk-bank.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
roachtest alterpk bank failed roachtest alterpk bank with on master cluster go start cockroach connect timeout cockroach sql url postgres root localhost sslmode disable e create schedule if not exists test only backup for backup into gs cockroach backup testing private roachprod scheduled backups teamcity auth implicit recurring full backup hourly with schedule options first run now error unexpected error occurred when checking for existing backups in gs cockroach backup testing private roachprod scheduled backups teamcity auth implicit unable to list files in gcs bucket googleapi error compute developer gserviceaccount com does not have storage objects list access to the google cloud storage bucket permission storage objects list denied on resource or it may not exist sqlstate failed running sql command problem exit status test artifacts and logs in artifacts alterpk bank run parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb sql foundations
| 1
|
107,468
| 23,418,927,187
|
IssuesEvent
|
2022-08-13 11:42:03
|
FerretDB/FerretDB
|
https://api.github.com/repos/FerretDB/FerretDB
|
opened
|
Enrich contribution guide
|
code/enhancement
|
I started my contribution by looking at https://github.com/FerretDB/FerretDB/blob/main/CONTRIBUTING.md and moving step by step. And first step `task init` is failing by default.
The problem is with a command `git describe --tags --dirty > gen/version.txt` git describe return error on forks with default mark `Copy the main branch only`. I think would be helpful to describe this behavior in CONTRIBUTING.md, so that new contributors will not face that issues
|
1.0
|
Enrich contribution guide - I started my contribution by looking at https://github.com/FerretDB/FerretDB/blob/main/CONTRIBUTING.md and moving step by step. And first step `task init` is failing by default.
The problem is with a command `git describe --tags --dirty > gen/version.txt` git describe return error on forks with default mark `Copy the main branch only`. I think would be helpful to describe this behavior in CONTRIBUTING.md, so that new contributors will not face that issues
|
non_test
|
enrich contribution guide i started my contribution by looking at and moving step by step and first step task init is failing by default the problem is with a command git describe tags dirty gen version txt git describe return error on forks with default mark copy the main branch only i think would be helpful to describe this behavior in contributing md so that new contributors will not face that issues
| 0
|
721,110
| 24,818,267,057
|
IssuesEvent
|
2022-10-25 14:36:40
|
ArctosDB/arctos
|
https://api.github.com/repos/ArctosDB/arctos
|
closed
|
Geography request for BELL Museum, part 2
|
Priority-High (Needed for work)
|
Hi, I just pre-bulkloaded a bunch of data and it spit back some higher geography terms that need fixing or that I found a bit confusing. I've attached a CSV with my notes on these. Three entries were just missing from higher geography where they shouldn't be; I have noted this in the CSV under NOTE, and request they be added. The others need clarification for me or correction in the database:
1) Kenya does not appear to have a continent? I input "Africa, Kenya" but that does not seem to exist at the moment, although I was grooming my data with a list previously provided by @dustymc.
2) Our data has an entry for "North America, United States, Virginia, Washington County", but Arctos recognizes "North America, United States, Virginia, Virginia, Washington County". Any idea what's going on there?
3) Vanuatu appears to be missing an ocean, requiring use of "Vanuatu" rather than "Pacific Ocean, Vanuatu". Is this correct?
BTW, will I have to restart the entire bulk upload if the missing geography terms are added? Or can I just fill in the added term in 'SHOULDBE' and it will be okay?
Thanks!
[geography_additions_queries_10_13_2022.csv](https://github.com/ArctosDB/arctos/files/9781470/geography_additions_queries_10_13_2022.csv)
|
1.0
|
Geography request for BELL Museum, part 2 - Hi, I just pre-bulkloaded a bunch of data and it spit back some higher geography terms that need fixing or that I found a bit confusing. I've attached a CSV with my notes on these. Three entries were just missing from higher geography where they shouldn't be; I have noted this in the CSV under NOTE, and request they be added. The others need clarification for me or correction in the database:
1) Kenya does not appear to have a continent? I input "Africa, Kenya" but that does not seem to exist at the moment, although I was grooming my data with a list previously provided by @dustymc.
2) Our data has an entry for "North America, United States, Virginia, Washington County", but Arctos recognizes "North America, United States, Virginia, Virginia, Washington County". Any idea what's going on there?
3) Vanuatu appears to be missing an ocean, requiring use of "Vanuatu" rather than "Pacific Ocean, Vanuatu". Is this correct?
BTW, will I have to restart the entire bulk upload if the missing geography terms are added? Or can I just fill in the added term in 'SHOULDBE' and it will be okay?
Thanks!
[geography_additions_queries_10_13_2022.csv](https://github.com/ArctosDB/arctos/files/9781470/geography_additions_queries_10_13_2022.csv)
|
non_test
|
geography request for bell museum part hi i just pre bulkloaded a bunch of data and it spit back some higher geography terms that need fixing or that i found a bit confusing i ve attached a csv with my notes on these three entries were just missing from higher geography where they shouldn t be i have noted this in the csv under note and request they be added the others need clarification for me or correction in the database kenya does not appear to have a continent i input africa kenya but that does not seem to exist at the moment although i was grooming my data with a list previously provided by dustymc our data has an entry for north america united states virginia washington county but arctos recognizes north america united states virginia virginia washington county any idea what s going on there vanuatu appears to be missing an ocean requiring use of vanuatu rather than pacific ocean vanuatu is this correct btw will i have to restart the entire bulk upload if the missing geography terms are added or can i just fill in the added term in shouldbe and it will be okay thanks
| 0
|
319,589
| 27,387,196,494
|
IssuesEvent
|
2023-02-28 14:06:06
|
splendo/kaluga
|
https://api.github.com/repos/splendo/kaluga
|
closed
|
Unstable test testScanning in com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest (android integration test)
|
🤖android 1.0.0 ❄ flaky test
|
currently in https://github.com/splendo/kaluga/tree/feature/performance-and-tests only
```
java.lang.AssertionError: Expected Exactly(times=1) but got 0 times
at com.splendo.kaluga.test.base.mock.MockMethodKt.fail(MockMethod.kt:35)
at com.splendo.kaluga.test.base.mock.MockMethodKt.expect(MockMethod.kt:32)
at com.splendo.kaluga.test.base.mock.MockMethodKt.access$expect(MockMethod.kt:1)
at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:166)
at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:158)
at com.splendo.kaluga.test.base.mock.VerifyKt.verify(verify.kt:36)
at com.splendo.kaluga.test.base.mock.VerifyKt.verify$default(verify.kt:18)
at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invokeSuspend(ScanningStateRepoTest.kt:181)
at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:8)
at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:4)
at com.splendo.kaluga.test.base.BaseFlowTest$mainAction$3.invokeSuspend(FlowTest.kt:264)
at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33)
at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106)
at android.os.Handler.handleCallback(Handler.java:883)
at android.os.Handler.dispatchMessage(Handler.java:100)
at android.os.Looper.loop(Looper.java:214)
at android.app.ActivityThread.main(ActivityThread.java:7356)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:492)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:930)
```
```
07-20 01:11:31.058: I/TestRunner(21080): started: testScanning(com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest)
07-20 01:11:31.096: V/LogKt$debug(21080): first test offered, starting collection
07-20 01:11:31.096: V/LogKt$debug(21080): launch flow scope...
07-20 01:11:31.097: V/LogKt$debug(21080): wait for main thread to be launched in StandaloneCoroutine{Active}@ef49f3e
07-20 01:11:31.100: V/LogKt$debug(21080): main scope launched, about to flow, test channel empty
07-20 01:11:31.100: V/LogKt$debug(21080): waited for main thread to be launched
07-20 01:11:31.100: V/LogKt$debug(21080): 1 in collection (including this one), offering
07-20 01:11:31.101: V/LogKt$debug(21080): start mainAction
07-20 01:11:31.101: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds
07-20 01:11:31.120: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@f17714a], test channel not empty "
07-20 01:11:31.122: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@abbf8bb)
07-20 01:11:31.126: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@abbf8bb)
07-20 01:11:31.134: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@abbf8bb)
07-20 01:11:31.135: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@f17714a]
07-20 01:11:31.135: V/LogKt$debug(21080): in main scope for mainAction
07-20 01:11:31.138: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@37411d8], test channel empty "
07-20 01:11:31.138: V/LogKt$debug(21080): did mainAction
07-20 01:11:31.139: V/LogKt$debug(21080): 1 in collection (including this one), offering
07-20 01:11:31.139: V/LogKt$debug(21080): start mainAction
07-20 01:11:31.140: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds
07-20 01:11:31.153: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@aeb9931)
07-20 01:11:31.157: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@aeb9931)
07-20 01:11:31.159: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@aeb9931)
07-20 01:11:31.159: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@37411d8]
07-20 01:11:31.159: V/LogKt$debug(21080): in main scope for mainAction
07-20 01:11:31.160: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@be35816], test channel empty "
07-20 01:11:31.160: V/LogKt$debug(21080): did mainAction
07-20 01:11:31.160: V/LogKt$debug(21080): 1 in collection (including this one), offering
07-20 01:11:31.161: V/LogKt$debug(21080): start mainAction
07-20 01:11:31.161: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds
07-20 01:11:31.163: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@ba76597)
07-20 01:11:31.167: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@ba76597)
07-20 01:11:31.169: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@ba76597)
07-20 01:11:31.170: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@be35816]
07-20 01:11:31.170: V/LogKt$debug(21080): in main scope for mainAction
07-20 01:11:31.172: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@1047384], test channel empty "
07-20 01:11:31.172: V/LogKt$debug(21080): did mainAction
07-20 01:11:31.173: V/LogKt$debug(21080): 1 in collection (including this one), offering
07-20 01:11:31.173: V/LogKt$debug(21080): start action
07-20 01:11:31.174: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds
07-20 01:11:31.176: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@62f976d)
07-20 01:11:31.178: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@62f976d)
07-20 01:11:31.179: V/LogKt$debug(21080): await all test blocks, but none found, skip waiting
07-20 01:11:31.179: V/LogKt$debug(21080): job: StandaloneCoroutine{Active}@ef49f3e
07-20 01:11:31.179: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@62f976d)
07-20 01:11:31.179: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@1047384]
07-20 01:11:31.181: V/LogKt$debug(21080): Ending flow, job canceled
07-20 01:11:31.181: V/LogKt$debug(21080): test channel closed
07-20 01:11:31.181: V/LogKt$debug(21080): did action
07-20 01:11:31.182: V/LogKt$debug(21080): start mainAction
07-20 01:11:31.182: V/LogKt$debug(21080): await all test blocks, but none found, skip waiting
07-20 01:11:31.183: V/LogKt$debug(21080): in main scope for mainAction
07-20 01:11:31.191: E/TestRunner(21080): failed: testScanning(com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest)
07-20 01:11:31.191: E/TestRunner(21080): ----- begin exception -----
07-20 01:11:31.192: I/link_layer_controller(1331): Sending scan response
07-20 01:11:31.193: E/TestRunner(21080): java.lang.AssertionError: Expected Exactly(times=1) but got 0 times
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.MockMethodKt.fail(MockMethod.kt:35)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.MockMethodKt.expect(MockMethod.kt:32)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.MockMethodKt.access$expect(MockMethod.kt:1)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:166)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:158)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.VerifyKt.verify(verify.kt:36)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.VerifyKt.verify$default(verify.kt:18)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invokeSuspend(ScanningStateRepoTest.kt:182)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:8)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:4)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.BaseFlowTest$mainAction$3.invokeSuspend(FlowTest.kt:264)
07-20 01:11:31.193: E/TestRunner(21080): at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33)
07-20 01:11:31.193: E/TestRunner(21080): at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106)
07-20 01:11:31.193: E/TestRunner(21080): at android.os.Handler.handleCallback(Handler.java:883)
07-20 01:11:31.193: E/TestRunner(21080): at android.os.Handler.dispatchMessage(Handler.java:100)
07-20 01:11:31.193: E/TestRunner(21080): at android.os.Looper.loop(Looper.java:214)
07-20 01:11:31.193: E/TestRunner(21080): at android.app.ActivityThread.main(ActivityThread.java:7356)
07-20 01:11:31.193: E/TestRunner(21080): at java.lang.reflect.Method.invoke(Native Method)
07-20 01:11:31.193: E/TestRunner(21080): at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:492)
07-20 01:11:31.193: E/TestRunner(21080): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:930)
07-20 01:11:31.193: E/TestRunner(21080): ----- end exception -----
07-20 01:11:31.195: I/TestRunner(21080): finished: testScanning(com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest)
```
|
1.0
|
Unstable test testScanning in com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest (android integration test) - currently in https://github.com/splendo/kaluga/tree/feature/performance-and-tests only
```
java.lang.AssertionError: Expected Exactly(times=1) but got 0 times
at com.splendo.kaluga.test.base.mock.MockMethodKt.fail(MockMethod.kt:35)
at com.splendo.kaluga.test.base.mock.MockMethodKt.expect(MockMethod.kt:32)
at com.splendo.kaluga.test.base.mock.MockMethodKt.access$expect(MockMethod.kt:1)
at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:166)
at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:158)
at com.splendo.kaluga.test.base.mock.VerifyKt.verify(verify.kt:36)
at com.splendo.kaluga.test.base.mock.VerifyKt.verify$default(verify.kt:18)
at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invokeSuspend(ScanningStateRepoTest.kt:181)
at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:8)
at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:4)
at com.splendo.kaluga.test.base.BaseFlowTest$mainAction$3.invokeSuspend(FlowTest.kt:264)
at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33)
at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106)
at android.os.Handler.handleCallback(Handler.java:883)
at android.os.Handler.dispatchMessage(Handler.java:100)
at android.os.Looper.loop(Looper.java:214)
at android.app.ActivityThread.main(ActivityThread.java:7356)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:492)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:930)
```
```
07-20 01:11:31.058: I/TestRunner(21080): started: testScanning(com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest)
07-20 01:11:31.096: V/LogKt$debug(21080): first test offered, starting collection
07-20 01:11:31.096: V/LogKt$debug(21080): launch flow scope...
07-20 01:11:31.097: V/LogKt$debug(21080): wait for main thread to be launched in StandaloneCoroutine{Active}@ef49f3e
07-20 01:11:31.100: V/LogKt$debug(21080): main scope launched, about to flow, test channel empty
07-20 01:11:31.100: V/LogKt$debug(21080): waited for main thread to be launched
07-20 01:11:31.100: V/LogKt$debug(21080): 1 in collection (including this one), offering
07-20 01:11:31.101: V/LogKt$debug(21080): start mainAction
07-20 01:11:31.101: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds
07-20 01:11:31.120: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@f17714a], test channel not empty "
07-20 01:11:31.122: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@abbf8bb)
07-20 01:11:31.126: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@abbf8bb)
07-20 01:11:31.134: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@abbf8bb)
07-20 01:11:31.135: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@f17714a]
07-20 01:11:31.135: V/LogKt$debug(21080): in main scope for mainAction
07-20 01:11:31.138: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@37411d8], test channel empty "
07-20 01:11:31.138: V/LogKt$debug(21080): did mainAction
07-20 01:11:31.139: V/LogKt$debug(21080): 1 in collection (including this one), offering
07-20 01:11:31.139: V/LogKt$debug(21080): start mainAction
07-20 01:11:31.140: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds
07-20 01:11:31.153: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@aeb9931)
07-20 01:11:31.157: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@aeb9931)
07-20 01:11:31.159: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@aeb9931)
07-20 01:11:31.159: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@37411d8]
07-20 01:11:31.159: V/LogKt$debug(21080): in main scope for mainAction
07-20 01:11:31.160: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@be35816], test channel empty "
07-20 01:11:31.160: V/LogKt$debug(21080): did mainAction
07-20 01:11:31.160: V/LogKt$debug(21080): 1 in collection (including this one), offering
07-20 01:11:31.161: V/LogKt$debug(21080): start mainAction
07-20 01:11:31.161: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds
07-20 01:11:31.163: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@ba76597)
07-20 01:11:31.167: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@ba76597)
07-20 01:11:31.169: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@ba76597)
07-20 01:11:31.170: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@be35816]
07-20 01:11:31.170: V/LogKt$debug(21080): in main scope for mainAction
07-20 01:11:31.172: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@1047384], test channel empty "
07-20 01:11:31.172: V/LogKt$debug(21080): did mainAction
07-20 01:11:31.173: V/LogKt$debug(21080): 1 in collection (including this one), offering
07-20 01:11:31.173: V/LogKt$debug(21080): start action
07-20 01:11:31.174: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds
07-20 01:11:31.176: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@62f976d)
07-20 01:11:31.178: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@62f976d)
07-20 01:11:31.179: V/LogKt$debug(21080): await all test blocks, but none found, skip waiting
07-20 01:11:31.179: V/LogKt$debug(21080): job: StandaloneCoroutine{Active}@ef49f3e
07-20 01:11:31.179: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@62f976d)
07-20 01:11:31.179: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@1047384]
07-20 01:11:31.181: V/LogKt$debug(21080): Ending flow, job canceled
07-20 01:11:31.181: V/LogKt$debug(21080): test channel closed
07-20 01:11:31.181: V/LogKt$debug(21080): did action
07-20 01:11:31.182: V/LogKt$debug(21080): start mainAction
07-20 01:11:31.182: V/LogKt$debug(21080): await all test blocks, but none found, skip waiting
07-20 01:11:31.183: V/LogKt$debug(21080): in main scope for mainAction
07-20 01:11:31.191: E/TestRunner(21080): failed: testScanning(com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest)
07-20 01:11:31.191: E/TestRunner(21080): ----- begin exception -----
07-20 01:11:31.192: I/link_layer_controller(1331): Sending scan response
07-20 01:11:31.193: E/TestRunner(21080): java.lang.AssertionError: Expected Exactly(times=1) but got 0 times
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.MockMethodKt.fail(MockMethod.kt:35)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.MockMethodKt.expect(MockMethod.kt:32)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.MockMethodKt.access$expect(MockMethod.kt:1)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:166)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:158)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.VerifyKt.verify(verify.kt:36)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.VerifyKt.verify$default(verify.kt:18)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invokeSuspend(ScanningStateRepoTest.kt:182)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:8)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:4)
07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.BaseFlowTest$mainAction$3.invokeSuspend(FlowTest.kt:264)
07-20 01:11:31.193: E/TestRunner(21080): at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33)
07-20 01:11:31.193: E/TestRunner(21080): at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106)
07-20 01:11:31.193: E/TestRunner(21080): at android.os.Handler.handleCallback(Handler.java:883)
07-20 01:11:31.193: E/TestRunner(21080): at android.os.Handler.dispatchMessage(Handler.java:100)
07-20 01:11:31.193: E/TestRunner(21080): at android.os.Looper.loop(Looper.java:214)
07-20 01:11:31.193: E/TestRunner(21080): at android.app.ActivityThread.main(ActivityThread.java:7356)
07-20 01:11:31.193: E/TestRunner(21080): at java.lang.reflect.Method.invoke(Native Method)
07-20 01:11:31.193: E/TestRunner(21080): at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:492)
07-20 01:11:31.193: E/TestRunner(21080): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:930)
07-20 01:11:31.193: E/TestRunner(21080): ----- end exception -----
07-20 01:11:31.195: I/TestRunner(21080): finished: testScanning(com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest)
```
|
test
|
unstable test testscanning in com splendo kaluga bluetooth scanner scanningstaterepotest android integration test currently in only java lang assertionerror expected exactly times but got times at com splendo kaluga test base mock mockmethodkt fail mockmethod kt at com splendo kaluga test base mock mockmethodkt expect mockmethod kt at com splendo kaluga test base mock mockmethodkt access expect mockmethod kt at com splendo kaluga test base mock basemethodmock verify test utils base debug mockmethod kt at com splendo kaluga test base mock basemethodmock verify test utils base debug mockmethod kt at com splendo kaluga test base mock verifykt verify verify kt at com splendo kaluga test base mock verifykt verify default verify kt at com splendo kaluga bluetooth scanner scanningstaterepotest testscanning invokesuspend scanningstaterepotest kt at com splendo kaluga bluetooth scanner scanningstaterepotest testscanning invoke unknown source at com splendo kaluga bluetooth scanner scanningstaterepotest testscanning invoke unknown source at com splendo kaluga test base baseflowtest mainaction invokesuspend flowtest kt at kotlin coroutines jvm internal basecontinuationimpl resumewith continuationimpl kt at kotlinx coroutines dispatchedtask run dispatchedtask kt at android os handler handlecallback handler java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at com android internal os runtimeinit methodandargscaller run runtimeinit java at com android internal os zygoteinit main zygoteinit java i testrunner started testscanning com splendo kaluga bluetooth scanner scanningstaterepotest v logkt debug first test offered starting collection v logkt debug launch flow scope v logkt debug wait for main thread to be launched in standalonecoroutine active v logkt debug main scope launched about to flow test channel empty v logkt debug waited for main thread to be launched v logkt debug in collection including this one offering v logkt debug start mainaction v logkt debug await all test blocks give it milliseconds v logkt debug in flow received test channel not empty v logkt debug received test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug ran test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug completed com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl completed v logkt debug handeling value completed v logkt debug in main scope for mainaction v logkt debug in flow received test channel empty v logkt debug did mainaction v logkt debug in collection including this one offering v logkt debug start mainaction v logkt debug await all test blocks give it milliseconds v logkt debug received test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug ran test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug completed com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl completed v logkt debug handeling value completed v logkt debug in main scope for mainaction v logkt debug in flow received test channel empty v logkt debug did mainaction v logkt debug in collection including this one offering v logkt debug start mainaction v logkt debug await all test blocks give it milliseconds v logkt debug received test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug ran test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug completed com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl completed v logkt debug handeling value completed v logkt debug in main scope for mainaction v logkt debug in flow received test channel empty v logkt debug did mainaction v logkt debug in collection including this one offering v logkt debug start action v logkt debug await all test blocks give it milliseconds v logkt debug received test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug ran test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug await all test blocks but none found skip waiting v logkt debug job standalonecoroutine active v logkt debug completed com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl completed v logkt debug handeling value completed v logkt debug ending flow job canceled v logkt debug test channel closed v logkt debug did action v logkt debug start mainaction v logkt debug await all test blocks but none found skip waiting v logkt debug in main scope for mainaction e testrunner failed testscanning com splendo kaluga bluetooth scanner scanningstaterepotest e testrunner begin exception i link layer controller sending scan response e testrunner java lang assertionerror expected exactly times but got times e testrunner at com splendo kaluga test base mock mockmethodkt fail mockmethod kt e testrunner at com splendo kaluga test base mock mockmethodkt expect mockmethod kt e testrunner at com splendo kaluga test base mock mockmethodkt access expect mockmethod kt e testrunner at com splendo kaluga test base mock basemethodmock verify test utils base debug mockmethod kt e testrunner at com splendo kaluga test base mock basemethodmock verify test utils base debug mockmethod kt e testrunner at com splendo kaluga test base mock verifykt verify verify kt e testrunner at com splendo kaluga test base mock verifykt verify default verify kt e testrunner at com splendo kaluga bluetooth scanner scanningstaterepotest testscanning invokesuspend scanningstaterepotest kt e testrunner at com splendo kaluga bluetooth scanner scanningstaterepotest testscanning invoke unknown source e testrunner at com splendo kaluga bluetooth scanner scanningstaterepotest testscanning invoke unknown source e testrunner at com splendo kaluga test base baseflowtest mainaction invokesuspend flowtest kt e testrunner at kotlin coroutines jvm internal basecontinuationimpl resumewith continuationimpl kt e testrunner at kotlinx coroutines dispatchedtask run dispatchedtask kt e testrunner at android os handler handlecallback handler java e testrunner at android os handler dispatchmessage handler java e testrunner at android os looper loop looper java e testrunner at android app activitythread main activitythread java e testrunner at java lang reflect method invoke native method e testrunner at com android internal os runtimeinit methodandargscaller run runtimeinit java e testrunner at com android internal os zygoteinit main zygoteinit java e testrunner end exception i testrunner finished testscanning com splendo kaluga bluetooth scanner scanningstaterepotest
| 1
|
265,676
| 23,188,710,710
|
IssuesEvent
|
2022-08-01 10:40:16
|
redpanda-data/redpanda
|
https://api.github.com/repos/redpanda-data/redpanda
|
closed
|
SchemaRegistryTest.test_protobuf
|
kind/bug area/schema_registry area/tests ci-failure
|
https://buildkite.com/redpanda/redpanda/builds/13300#018248a7-cb3a-4031-8a0e-8f4c7187645e
```
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/ducktape/tests/runner_client.py", line 135, in run
data = self.run_test()
File "/usr/local/lib/python3.10/dist-packages/ducktape/tests/runner_client.py", line 227, in run_test
return self.test_context.function(self.test)
File "/root/tests/rptest/services/cluster.py", line 35, in wrapped
r = f(self, *args, **kwargs)
File "/root/tests/rptest/tests/schema_registry_test.py", line 999, in test_protobuf
assert result_raw.status_code == requests.codes.unprocessable_entity
AssertionError
```
|
1.0
|
SchemaRegistryTest.test_protobuf - https://buildkite.com/redpanda/redpanda/builds/13300#018248a7-cb3a-4031-8a0e-8f4c7187645e
```
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/ducktape/tests/runner_client.py", line 135, in run
data = self.run_test()
File "/usr/local/lib/python3.10/dist-packages/ducktape/tests/runner_client.py", line 227, in run_test
return self.test_context.function(self.test)
File "/root/tests/rptest/services/cluster.py", line 35, in wrapped
r = f(self, *args, **kwargs)
File "/root/tests/rptest/tests/schema_registry_test.py", line 999, in test_protobuf
assert result_raw.status_code == requests.codes.unprocessable_entity
AssertionError
```
|
test
|
schemaregistrytest test protobuf traceback most recent call last file usr local lib dist packages ducktape tests runner client py line in run data self run test file usr local lib dist packages ducktape tests runner client py line in run test return self test context function self test file root tests rptest services cluster py line in wrapped r f self args kwargs file root tests rptest tests schema registry test py line in test protobuf assert result raw status code requests codes unprocessable entity assertionerror
| 1
|
352,082
| 32,044,242,484
|
IssuesEvent
|
2023-09-22 22:44:37
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
sql/tests: TestRandomSyntaxSchemaChangeColumn failed
|
C-test-failure O-robot branch-master T-sql-foundations
|
sql/tests.TestRandomSyntaxSchemaChangeColumn [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/11880882?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/11880882?buildTab=artifacts#/) on master @ [6a2097e4c2b6537be84f265bd55ac23d2cf83351](https://github.com/cockroachdb/cockroach/commits/6a2097e4c2b6537be84f265bd55ac23d2cf83351):
Random syntax error:
```
rsg_test.go:889: Crash detected: server panic: statement exec timeout
```
Query:
```
ALTER TABLE ident.ident ADD CONSTRAINT IF NOT EXISTS TIES PRIMARY KEY ( DEFAULTS ident DESC NULLS LAST );
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #111020 sql/tests: TestRandomSyntaxSchemaChangeColumn failed [C-test-failure O-robot T-sql-foundations branch-release-22.2]
</p>
</details>
/cc @cockroachdb/sql-foundations
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRandomSyntaxSchemaChangeColumn.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-31759
|
1.0
|
sql/tests: TestRandomSyntaxSchemaChangeColumn failed - sql/tests.TestRandomSyntaxSchemaChangeColumn [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/11880882?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/11880882?buildTab=artifacts#/) on master @ [6a2097e4c2b6537be84f265bd55ac23d2cf83351](https://github.com/cockroachdb/cockroach/commits/6a2097e4c2b6537be84f265bd55ac23d2cf83351):
Random syntax error:
```
rsg_test.go:889: Crash detected: server panic: statement exec timeout
```
Query:
```
ALTER TABLE ident.ident ADD CONSTRAINT IF NOT EXISTS TIES PRIMARY KEY ( DEFAULTS ident DESC NULLS LAST );
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #111020 sql/tests: TestRandomSyntaxSchemaChangeColumn failed [C-test-failure O-robot T-sql-foundations branch-release-22.2]
</p>
</details>
/cc @cockroachdb/sql-foundations
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRandomSyntaxSchemaChangeColumn.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-31759
|
test
|
sql tests testrandomsyntaxschemachangecolumn failed sql tests testrandomsyntaxschemachangecolumn with on master random syntax error rsg test go crash detected server panic statement exec timeout query alter table ident ident add constraint if not exists ties primary key defaults ident desc nulls last help see also same failure on other branches sql tests testrandomsyntaxschemachangecolumn failed cc cockroachdb sql foundations jira issue crdb
| 1
|
25,233
| 12,522,073,015
|
IssuesEvent
|
2020-06-03 18:30:34
|
OpenNeuroOrg/openneuro
|
https://api.github.com/repos/OpenNeuroOrg/openneuro
|
closed
|
Migrate existing git trees out of infrequent access tier storage
|
performance
|
The fastest way to do this would be a maintenance window to rsync them to another volume and back. This is a small subset of data, so the window could be fairly short.
See #1458 for fixing this going forward.
|
True
|
Migrate existing git trees out of infrequent access tier storage - The fastest way to do this would be a maintenance window to rsync them to another volume and back. This is a small subset of data, so the window could be fairly short.
See #1458 for fixing this going forward.
|
non_test
|
migrate existing git trees out of infrequent access tier storage the fastest way to do this would be a maintenance window to rsync them to another volume and back this is a small subset of data so the window could be fairly short see for fixing this going forward
| 0
|
802,641
| 29,042,869,449
|
IssuesEvent
|
2023-05-13 06:50:59
|
IdkwhatImD0ing/SlugLoop
|
https://api.github.com/repos/IdkwhatImD0ing/SlugLoop
|
closed
|
Implement Route Toggle Feature to Display Buses and Route Lines for Selected Routes
|
wontfix low-priority long-term
|
# Description:
Building upon the previous issue of adding a JSON file for common bus routes, we need to enhance the web app by providing users with the ability to toggle between different routes. This feature should display only the buses corresponding to the selected route and show a line on the map representing the route path. We can utilize the JSON file provided in issue #41 to achieve this.
# Issue Details:
1. No route toggle functionality: Users currently cannot switch between different bus routes and view the corresponding buses in the web app.
2. Missing route lines: The web app does not display a visual representation of the bus routes on the map.
# Proposed Solution:
1. Implement route toggle feature: Add a user interface element, such as a dropdown menu or buttons, that allows users to select a specific bus route.
2. Filter displayed buses: When a route is selected, update the web app to show only the buses corresponding to the chosen route.
3. Display route lines: Upon selecting a route, display a line on the map that represents the route path, providing users with a visual guide of the bus journey.
4. Utilize JSON file from issue #41: Integrate the JSON file provided in issue #41 to store the route data, including the bus stops and route coordinates for displaying route lines.
# Acceptance Criteria:
1. Users can toggle between different bus routes using a dedicated interface element in the web app.
2. When a route is selected, only the buses corresponding to that route are displayed.
3. A line representing the chosen bus route is shown on the map, providing users with a visual understanding of the route path.
4. The JSON file from issue #41 is effectively utilized for storing and managing route data.
|
1.0
|
Implement Route Toggle Feature to Display Buses and Route Lines for Selected Routes - # Description:
Building upon the previous issue of adding a JSON file for common bus routes, we need to enhance the web app by providing users with the ability to toggle between different routes. This feature should display only the buses corresponding to the selected route and show a line on the map representing the route path. We can utilize the JSON file provided in issue #41 to achieve this.
# Issue Details:
1. No route toggle functionality: Users currently cannot switch between different bus routes and view the corresponding buses in the web app.
2. Missing route lines: The web app does not display a visual representation of the bus routes on the map.
# Proposed Solution:
1. Implement route toggle feature: Add a user interface element, such as a dropdown menu or buttons, that allows users to select a specific bus route.
2. Filter displayed buses: When a route is selected, update the web app to show only the buses corresponding to the chosen route.
3. Display route lines: Upon selecting a route, display a line on the map that represents the route path, providing users with a visual guide of the bus journey.
4. Utilize JSON file from issue #41: Integrate the JSON file provided in issue #41 to store the route data, including the bus stops and route coordinates for displaying route lines.
# Acceptance Criteria:
1. Users can toggle between different bus routes using a dedicated interface element in the web app.
2. When a route is selected, only the buses corresponding to that route are displayed.
3. A line representing the chosen bus route is shown on the map, providing users with a visual understanding of the route path.
4. The JSON file from issue #41 is effectively utilized for storing and managing route data.
|
non_test
|
implement route toggle feature to display buses and route lines for selected routes description building upon the previous issue of adding a json file for common bus routes we need to enhance the web app by providing users with the ability to toggle between different routes this feature should display only the buses corresponding to the selected route and show a line on the map representing the route path we can utilize the json file provided in issue to achieve this issue details no route toggle functionality users currently cannot switch between different bus routes and view the corresponding buses in the web app missing route lines the web app does not display a visual representation of the bus routes on the map proposed solution implement route toggle feature add a user interface element such as a dropdown menu or buttons that allows users to select a specific bus route filter displayed buses when a route is selected update the web app to show only the buses corresponding to the chosen route display route lines upon selecting a route display a line on the map that represents the route path providing users with a visual guide of the bus journey utilize json file from issue integrate the json file provided in issue to store the route data including the bus stops and route coordinates for displaying route lines acceptance criteria users can toggle between different bus routes using a dedicated interface element in the web app when a route is selected only the buses corresponding to that route are displayed a line representing the chosen bus route is shown on the map providing users with a visual understanding of the route path the json file from issue is effectively utilized for storing and managing route data
| 0
|
151,106
| 13,390,775,279
|
IssuesEvent
|
2020-09-02 21:10:09
|
google/generic-webdriver-server
|
https://api.github.com/repos/google/generic-webdriver-server
|
opened
|
Document requirements for iframe embedding for Chromecast receiver
|
documentation
|
Sites can prevent iframe-embedding with the [X-Frame-Options](https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-Frame-Options) header. Though this should not be an issue for a test runner, one of the first things a person might try with the command-line tools is:
```sh
chromecast-webdriver-cli --hostname=chromecast-hostname --url=https://www.google.com/
```
However, google.com prevents iframe embedding, which means it cannot be shown in our Chromecast receiver. The user would see a blank screen instead. This should be clearly documented to avoid confusion.
|
1.0
|
Document requirements for iframe embedding for Chromecast receiver - Sites can prevent iframe-embedding with the [X-Frame-Options](https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-Frame-Options) header. Though this should not be an issue for a test runner, one of the first things a person might try with the command-line tools is:
```sh
chromecast-webdriver-cli --hostname=chromecast-hostname --url=https://www.google.com/
```
However, google.com prevents iframe embedding, which means it cannot be shown in our Chromecast receiver. The user would see a blank screen instead. This should be clearly documented to avoid confusion.
|
non_test
|
document requirements for iframe embedding for chromecast receiver sites can prevent iframe embedding with the header though this should not be an issue for a test runner one of the first things a person might try with the command line tools is sh chromecast webdriver cli hostname chromecast hostname url however google com prevents iframe embedding which means it cannot be shown in our chromecast receiver the user would see a blank screen instead this should be clearly documented to avoid confusion
| 0
|
307,189
| 26,518,717,896
|
IssuesEvent
|
2023-01-18 23:32:55
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
DISABLED test_write_to_closures_in_inlining_dynamic_shapes (torch._dynamo.testing.make_test_cls_with_patches.<locals>.DummyTestClass)
|
module: flaky-tests skipped module: unknown
|
Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_write_to_closures_in_inlining_dynamic_shapes) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/10711572477).
Over the past 72 hours, it has flakily failed in 4 workflow(s).
**Debugging instructions (after clicking on the recent samples link):**
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Grep for `test_write_to_closures_in_inlining_dynamic_shapes`
Error retrieving /opt/conda/lib/python3.10/site-packages/torch/_dynamo/testing.py: Error: Statuscode 301
|
1.0
|
DISABLED test_write_to_closures_in_inlining_dynamic_shapes (torch._dynamo.testing.make_test_cls_with_patches.<locals>.DummyTestClass) - Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_write_to_closures_in_inlining_dynamic_shapes) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/10711572477).
Over the past 72 hours, it has flakily failed in 4 workflow(s).
**Debugging instructions (after clicking on the recent samples link):**
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Grep for `test_write_to_closures_in_inlining_dynamic_shapes`
Error retrieving /opt/conda/lib/python3.10/site-packages/torch/_dynamo/testing.py: Error: Statuscode 301
|
test
|
disabled test write to closures in inlining dynamic shapes torch dynamo testing make test cls with patches dummytestclass platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has flakily failed in workflow s debugging instructions after clicking on the recent samples link to find relevant log snippets click on the workflow logs linked above grep for test write to closures in inlining dynamic shapes error retrieving opt conda lib site packages torch dynamo testing py error statuscode
| 1
|
516,188
| 14,976,763,065
|
IssuesEvent
|
2021-01-28 08:33:41
|
magento/magento2
|
https://api.github.com/repos/magento/magento2
|
closed
|
Config data is re-encrypted when saved in the back office on Magento 2210
|
Component: Config Issue: Confirmed Issue: Format is valid Issue: Ready for Work Issue: needs update Priority: P2 Progress: dev in progress Reproduced on 2.4.x Severity: S2 Triage: Dev.Experience
|
### Preconditions
1. Magento 2.4-develop
### Steps to reproduce
1. Create a module with this a back office form, with such a field (system.xml):
```
[...]
<field id="hmackey" translate="label comment" type="text" sortOrder="50" showInDefault="1"
showInWebsite="1" showInStore="1">
<label>HMAC</label>
<comment><![CDATA[...]]></comment>
<validate>required-entry</validate>
<config_path>foo/bar/hmackey</config_path>
<backend_model>Magento\Config\Model\Config\Backend\Encrypted</backend_model>
</field>
[...]
```
2. Go to the admin on the form page
3. The data is correctly shown, save the form
### Expected result
1. Data is correctly encrypted and saved in the database
2. De-crypted data is shown in the admin form
### Actual result
1. Data is correctly encrypted and saved in the database
2. EN-crypted data is shown in the admin form
### Notes
I've tried the same configuration on previous versions and can't reproduce this issue, only on 2210.
I can't see anything in the changelog between 2200 and 2210 that would explain such a behavior.
Any ideas?
|
1.0
|
Config data is re-encrypted when saved in the back office on Magento 2210 - ### Preconditions
1. Magento 2.4-develop
### Steps to reproduce
1. Create a module with this a back office form, with such a field (system.xml):
```
[...]
<field id="hmackey" translate="label comment" type="text" sortOrder="50" showInDefault="1"
showInWebsite="1" showInStore="1">
<label>HMAC</label>
<comment><![CDATA[...]]></comment>
<validate>required-entry</validate>
<config_path>foo/bar/hmackey</config_path>
<backend_model>Magento\Config\Model\Config\Backend\Encrypted</backend_model>
</field>
[...]
```
2. Go to the admin on the form page
3. The data is correctly shown, save the form
### Expected result
1. Data is correctly encrypted and saved in the database
2. De-crypted data is shown in the admin form
### Actual result
1. Data is correctly encrypted and saved in the database
2. EN-crypted data is shown in the admin form
### Notes
I've tried the same configuration on previous versions and can't reproduce this issue, only on 2210.
I can't see anything in the changelog between 2200 and 2210 that would explain such a behavior.
Any ideas?
|
non_test
|
config data is re encrypted when saved in the back office on magento preconditions magento develop steps to reproduce create a module with this a back office form with such a field system xml field id hmackey translate label comment type text sortorder showindefault showinwebsite showinstore hmac required entry foo bar hmackey magento config model config backend encrypted go to the admin on the form page the data is correctly shown save the form expected result data is correctly encrypted and saved in the database de crypted data is shown in the admin form actual result data is correctly encrypted and saved in the database en crypted data is shown in the admin form notes i ve tried the same configuration on previous versions and can t reproduce this issue only on i can t see anything in the changelog between and that would explain such a behavior any ideas
| 0
|
188,101
| 22,046,151,458
|
IssuesEvent
|
2022-05-30 02:06:25
|
maddyCode23/linux-4.1.15
|
https://api.github.com/repos/maddyCode23/linux-4.1.15
|
closed
|
CVE-2019-15291 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed
|
security vulnerability
|
## CVE-2019-15291 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/usb/b2c2/flexcop-usb.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/usb/b2c2/flexcop-usb.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel through 5.2.9. There is a NULL pointer dereference caused by a malicious USB device in the flexcop_usb_probe function in the drivers/media/usb/b2c2/flexcop-usb.c driver.
<p>Publish Date: 2019-08-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15291>CVE-2019-15291</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2019-15291">https://www.linuxkernelcves.com/cves/CVE-2019-15291</a></p>
<p>Release Date: 2019-09-06</p>
<p>Fix Resolution: v5.5-rc1,v3.16.79,v4.14.157,v4.19.87,v4.4.204,v4.9.204,v5.3.14,v5.4.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-15291 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed - ## CVE-2019-15291 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/usb/b2c2/flexcop-usb.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/usb/b2c2/flexcop-usb.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel through 5.2.9. There is a NULL pointer dereference caused by a malicious USB device in the flexcop_usb_probe function in the drivers/media/usb/b2c2/flexcop-usb.c driver.
<p>Publish Date: 2019-08-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15291>CVE-2019-15291</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2019-15291">https://www.linuxkernelcves.com/cves/CVE-2019-15291</a></p>
<p>Release Date: 2019-09-06</p>
<p>Fix Resolution: v5.5-rc1,v3.16.79,v4.14.157,v4.19.87,v4.4.204,v4.9.204,v5.3.14,v5.4.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers media usb flexcop usb c drivers media usb flexcop usb c vulnerability details an issue was discovered in the linux kernel through there is a null pointer dereference caused by a malicious usb device in the flexcop usb probe function in the drivers media usb flexcop usb c driver publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
527,309
| 15,339,629,750
|
IssuesEvent
|
2021-02-27 02:49:46
|
RubyMoney/money
|
https://api.github.com/repos/RubyMoney/money
|
closed
|
Explicit +/- change breaks existing API
|
priority:low
|
After https://github.com/RubyMoney/money/pull/852 has been merged, classes not inheriting Money cannot be handled anymore. I've got a codebase with the following class:
```
class Pricing
class NullPrice < SimpleDelegator
def initialize(item_id=nil)
super(Money.zero)
LogContext.add(null_priced_item_id: item_id)
end
def format(*_rules)
"$N/A"
end
alias_method :to_s, :format
alias_method :centless_format, :format
def inspect
"#<#{self.class.name}>"
end
end
end
```
Which worked with the version `6.13.2`, but breaks in `6.13.3` when testing +/-:
```
Failure/Error: expect(Money.new(1_00) - subject).to be_eql Money.new(1_00)
Unsupported argument type: Pricing::NullPrice
```
While this is possible to patch to work with the new update, I don't think a point release should break existing code.
I think the new approach in general has an issue where the new object is constructed via `self.class.new` which gives inconsistent results: `Foo.new + Money.new -> Foo`, but `Money.new + Foo.new -> Money`
|
1.0
|
Explicit +/- change breaks existing API - After https://github.com/RubyMoney/money/pull/852 has been merged, classes not inheriting Money cannot be handled anymore. I've got a codebase with the following class:
```
class Pricing
class NullPrice < SimpleDelegator
def initialize(item_id=nil)
super(Money.zero)
LogContext.add(null_priced_item_id: item_id)
end
def format(*_rules)
"$N/A"
end
alias_method :to_s, :format
alias_method :centless_format, :format
def inspect
"#<#{self.class.name}>"
end
end
end
```
Which worked with the version `6.13.2`, but breaks in `6.13.3` when testing +/-:
```
Failure/Error: expect(Money.new(1_00) - subject).to be_eql Money.new(1_00)
Unsupported argument type: Pricing::NullPrice
```
While this is possible to patch to work with the new update, I don't think a point release should break existing code.
I think the new approach in general has an issue where the new object is constructed via `self.class.new` which gives inconsistent results: `Foo.new + Money.new -> Foo`, but `Money.new + Foo.new -> Money`
|
non_test
|
explicit change breaks existing api after has been merged classes not inheriting money cannot be handled anymore i ve got a codebase with the following class class pricing class nullprice simpledelegator def initialize item id nil super money zero logcontext add null priced item id item id end def format rules n a end alias method to s format alias method centless format format def inspect end end end which worked with the version but breaks in when testing failure error expect money new subject to be eql money new unsupported argument type pricing nullprice while this is possible to patch to work with the new update i don t think a point release should break existing code i think the new approach in general has an issue where the new object is constructed via self class new which gives inconsistent results foo new money new foo but money new foo new money
| 0
|
432,034
| 30,264,328,470
|
IssuesEvent
|
2023-07-07 10:35:11
|
Eu4ng/TIL
|
https://api.github.com/repos/Eu4ng/TIL
|
opened
|
[UE5] 데이터 애셋을 할당하거나 인스턴스를 생성하는 예제
|
documentation
|
```cpp
// Fill out your copyright notice in the Description page of Project Settings.
#pragma once
#include "CoreMinimal.h"
#include "Engine/DataAsset.h"
#include "DataAsset_CharacterAction.generated.h"
USTRUCT(BlueprintType)
struct FCharacterAction
{
GENERATED_BODY()
UPROPERTY(EditDefaultsOnly)
bool bUseInstance;
UPROPERTY(EditDefaultsOnly, meta=(EditCondition="!bUseInstance", EditConditionHides))
UDataAsset_CharacterAction* DataAsset;
UPROPERTY(EditDefaultsOnly, Instanced, meta=(EditCondition="bUseInstance", EditConditionHides))
UDataAsset_CharacterAction* Instance;
};
/**
* 무기 종류에 따른 애님 몽타주를 설정하는 데이터 애셋
*/
UCLASS(EditInlineNew, DisplayName="CharacterAction")
class RPGCORE_API UDataAsset_CharacterAction : public UDataAsset
{
GENERATED_BODY()
UPROPERTY(EditDefaultsOnly, BlueprintGetter=GetSkeleton)
USkeleton* Skeleton;
UPROPERTY(EditDefaultsOnly, BlueprintGetter=GetAttackMontage)
UAnimMontage* AttackMontage;
public:
UFUNCTION(BlueprintGetter)
FORCEINLINE USkeleton* GetSkeleton() const { return Skeleton; }
UFUNCTION(BlueprintGetter)
FORCEINLINE UAnimMontage* GetAttackMontage() const { return AttackMontage; }
UFUNCTION(BlueprintPure)
static FORCEINLINE UDataAsset_CharacterAction* Get(const FCharacterAction& CharacterAction) { return CharacterAction.bUseInstance ? CharacterAction.Instance : CharacterAction.DataAsset; }
};
```
|
1.0
|
[UE5] 데이터 애셋을 할당하거나 인스턴스를 생성하는 예제 - ```cpp
// Fill out your copyright notice in the Description page of Project Settings.
#pragma once
#include "CoreMinimal.h"
#include "Engine/DataAsset.h"
#include "DataAsset_CharacterAction.generated.h"
USTRUCT(BlueprintType)
struct FCharacterAction
{
GENERATED_BODY()
UPROPERTY(EditDefaultsOnly)
bool bUseInstance;
UPROPERTY(EditDefaultsOnly, meta=(EditCondition="!bUseInstance", EditConditionHides))
UDataAsset_CharacterAction* DataAsset;
UPROPERTY(EditDefaultsOnly, Instanced, meta=(EditCondition="bUseInstance", EditConditionHides))
UDataAsset_CharacterAction* Instance;
};
/**
* 무기 종류에 따른 애님 몽타주를 설정하는 데이터 애셋
*/
UCLASS(EditInlineNew, DisplayName="CharacterAction")
class RPGCORE_API UDataAsset_CharacterAction : public UDataAsset
{
GENERATED_BODY()
UPROPERTY(EditDefaultsOnly, BlueprintGetter=GetSkeleton)
USkeleton* Skeleton;
UPROPERTY(EditDefaultsOnly, BlueprintGetter=GetAttackMontage)
UAnimMontage* AttackMontage;
public:
UFUNCTION(BlueprintGetter)
FORCEINLINE USkeleton* GetSkeleton() const { return Skeleton; }
UFUNCTION(BlueprintGetter)
FORCEINLINE UAnimMontage* GetAttackMontage() const { return AttackMontage; }
UFUNCTION(BlueprintPure)
static FORCEINLINE UDataAsset_CharacterAction* Get(const FCharacterAction& CharacterAction) { return CharacterAction.bUseInstance ? CharacterAction.Instance : CharacterAction.DataAsset; }
};
```
|
non_test
|
데이터 애셋을 할당하거나 인스턴스를 생성하는 예제 cpp fill out your copyright notice in the description page of project settings pragma once include coreminimal h include engine dataasset h include dataasset characteraction generated h ustruct blueprinttype struct fcharacteraction generated body uproperty editdefaultsonly bool buseinstance uproperty editdefaultsonly meta editcondition buseinstance editconditionhides udataasset characteraction dataasset uproperty editdefaultsonly instanced meta editcondition buseinstance editconditionhides udataasset characteraction instance 무기 종류에 따른 애님 몽타주를 설정하는 데이터 애셋 uclass editinlinenew displayname characteraction class rpgcore api udataasset characteraction public udataasset generated body uproperty editdefaultsonly blueprintgetter getskeleton uskeleton skeleton uproperty editdefaultsonly blueprintgetter getattackmontage uanimmontage attackmontage public ufunction blueprintgetter forceinline uskeleton getskeleton const return skeleton ufunction blueprintgetter forceinline uanimmontage getattackmontage const return attackmontage ufunction blueprintpure static forceinline udataasset characteraction get const fcharacteraction characteraction return characteraction buseinstance characteraction instance characteraction dataasset
| 0
|
34,455
| 4,927,210,621
|
IssuesEvent
|
2016-11-26 16:16:51
|
khartec/waltz
|
https://api.github.com/repos/khartec/waltz
|
closed
|
High Phys Flows: Lineage Edit page needlessly retrieves all phys flows for an app
|
fixed (test & close) performance
|
In order to render the candidate sources, all flows either sourcing from or to an application are retrieved. This is unnecessary as we only need the flows incoming into the application.
The api calls used are:
http://localhost:8443/api/physical-specification/application/APPLICATION/1746
http://localhost:8443/api/physical-flow/entity/APPLICATION/1746 (this one is expensive when lots of flows in system)
|
1.0
|
High Phys Flows: Lineage Edit page needlessly retrieves all phys flows for an app - In order to render the candidate sources, all flows either sourcing from or to an application are retrieved. This is unnecessary as we only need the flows incoming into the application.
The api calls used are:
http://localhost:8443/api/physical-specification/application/APPLICATION/1746
http://localhost:8443/api/physical-flow/entity/APPLICATION/1746 (this one is expensive when lots of flows in system)
|
test
|
high phys flows lineage edit page needlessly retrieves all phys flows for an app in order to render the candidate sources all flows either sourcing from or to an application are retrieved this is unnecessary as we only need the flows incoming into the application the api calls used are this one is expensive when lots of flows in system
| 1
|
87,739
| 8,120,627,907
|
IssuesEvent
|
2018-08-16 03:57:08
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
teamcity: failed tests on master: testrace/TestShowSessions, lint/TestLint
|
C-test-failure O-robot
|
The following tests appear to have failed:
[#839592](https://teamcity.cockroachdb.com/viewLog.html?buildId=839592):
```
--- FAIL: testrace/TestShowSessions (2.450s)
show_test.go:684: unexpected number of running sessions: 2, expected 1.
Active sessions (results might have changed since the test checked):
app: "$ internal-log-range-event", query: "INSERT INTO system.public.rangelog(\"timestamp\", \"rangeID\", \"storeID\", \"eventType\", \"otherRangeID\", info) VALUES ($1, $2, $3, $4, $5, $6)", last query: app: "", query: "SELECT active_queries, last_active_query, application_name FROM [SHOW CLUSTER SESSIONS]", last query: SELECT node_id, (now() - session_start)::FLOAT FROM [SHOW CLUSTER SESSIONS] WHERE application_name NOT LIKE 'internal-%'
------- Stdout: -------
W180815 20:40:34.423397 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180815 20:40:34.450676 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180815 20:40:34.451091 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:34.451159 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:34.455752 158099 server/config.go:496 [n?] 1 storage engine initialized
I180815 20:40:34.456002 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180815 20:40:34.456043 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B
I180815 20:40:34.514781 158099 server/node.go:373 [n?] **** cluster 6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 has been created
I180815 20:40:34.515045 158099 server/server.go:1398 [n?] **** add additional nodes by specifying --join=127.0.0.1:41091
I180815 20:40:34.516516 158099 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41091" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365634515938360
I180815 20:40:34.534941 158099 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180815 20:40:34.537635 158099 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180815 20:40:34.538394 158099 storage/stores.go:222 [n1] read 0 node addresses from persistent storage
I180815 20:40:34.539206 158099 server/node.go:697 [n1] connecting to gossip network to verify cluster ID...
I180815 20:40:34.539582 158099 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102"
I180815 20:40:34.539943 158099 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180815 20:40:34.541685 158099 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:34.541909 158099 server/server.go:1833 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180815 20:40:34.543736 158325 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:34.545405 158099 server/server.go:1535 [n1] starting https server at 127.0.0.1:34275 (use: 127.0.0.1:34275)
I180815 20:40:34.545680 158099 server/server.go:1537 [n1] starting grpc/postgres server at 127.0.0.1:41091
I180815 20:40:34.545753 158099 server/server.go:1538 [n1] advertising CockroachDB node at 127.0.0.1:41091
I180815 20:40:34.566927 158335 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180815 20:40:34.719270 158304 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180815 20:40:34.817666 158087 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180815 20:40:34.840768 158068 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180815 20:40:34.933807 158160 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180815 20:40:35.038918 158386 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
I180815 20:40:35.116602 158341 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root}
W180815 20:40:35.128647 158374 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=5ddfb628 key=/Local/Range/System/tsd/RangeDescriptor rw=true pri=0.02614628 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534365635.039580425,0 orig=1534365635.039580425,0 max=1534365635.039580425,0 wto=false rop=false seq=1
I180815 20:40:35.147031 158074 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180815 20:40:35.241602 158379 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180815 20:40:35.285924 158289 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180815 20:40:35.328157 158390 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180815 20:40:35.401113 158408 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180815 20:40:35.470572 158398 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180815 20:40:35.530967 158452 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180815 20:40:35.548516 158385 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root}
I180815 20:40:35.603367 158447 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180815 20:40:35.633708 158442 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established
I180815 20:40:35.654186 158514 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180815 20:40:35.677178 158463 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180815 20:40:35.720541 158458 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180815 20:40:35.744993 158521 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180815 20:40:35.765024 158099 server/server.go:1620 [n1] done ensuring all necessary migrations have run
I180815 20:40:35.765364 158099 server/server.go:1623 [n1] serving sql connections
I180815 20:40:35.811858 158481 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41091} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365634515938360 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365634515938360 LastUp:1534365634515938360}
I180815 20:40:35.813481 158479 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180815 20:40:35.820374 158564 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180815 20:40:35.889664 158497 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180815 20:40:35.951092 158556 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180815 20:40:36.009944 158598 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180815 20:40:36.059947 158619 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
I180815 20:40:36.119763 158606 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180815 20:40:36.175659 158630 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
W180815 20:40:36.287477 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180815 20:40:36.315831 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180815 20:40:36.316449 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:36.316518 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:36.321488 158099 server/config.go:496 [n?] 1 storage engine initialized
I180815 20:40:36.321673 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180815 20:40:36.321701 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B
W180815 20:40:36.321999 158099 gossip/gossip.go:1351 [n?] no incoming or outgoing connections
I180815 20:40:36.322565 158099 server/server.go:1400 [n?] no stores bootstrapped and --join flag specified, awaiting init command.
I180815 20:40:36.398063 158665 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:41091
I180815 20:40:36.399347 158763 gossip/server.go:219 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:32789}
I180815 20:40:36.404521 158099 server/node.go:697 [n?] connecting to gossip network to verify cluster ID...
I180815 20:40:36.405108 158099 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102"
I180815 20:40:36.408670 158717 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:40:36.415122 158716 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:40:36.422331 158099 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:40:36.431430 158099 server/node.go:428 [n?] new node allocated ID 2
I180815 20:40:36.432094 158099 gossip/gossip.go:383 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:32789" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365636431659703
I180815 20:40:36.433009 158099 storage/stores.go:222 [n2] read 0 node addresses from persistent storage
I180815 20:40:36.433961 158099 storage/stores.go:241 [n2] wrote 1 node addresses to persistent storage
I180815 20:40:36.436697 158720 storage/stores.go:241 [n1] wrote 1 node addresses to persistent storage
I180815 20:40:36.457872 158099 server/node.go:672 [n2] bootstrapped store [n2,s2]
I180815 20:40:36.460045 158099 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes []
I180815 20:40:36.461355 158099 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:36.461544 158099 server/server.go:1833 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180815 20:40:36.465410 158099 server/server.go:1535 [n2] starting https server at 127.0.0.1:36335 (use: 127.0.0.1:36335)
I180815 20:40:36.465633 158099 server/server.go:1537 [n2] starting grpc/postgres server at 127.0.0.1:32789
I180815 20:40:36.465700 158099 server/server.go:1538 [n2] advertising CockroachDB node at 127.0.0.1:32789
I180815 20:40:36.468619 158867 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:36.476134 158099 server/server.go:1620 [n2] done ensuring all necessary migrations have run
I180815 20:40:36.476492 158099 server/server.go:1623 [n2] serving sql connections
I180815 20:40:36.576176 158869 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version
I180815 20:40:36.577336 158871 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:32789} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365636431659703 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365636431659703 LastUp:1534365636431659703}
I180815 20:40:36.617219 158877 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established
I180815 20:40:36.623403 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root}
I180815 20:40:36.624884 158702 storage/replica_command.go:282 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
I180815 20:40:36.685520 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_table", target: 53, info: {TableName:test.public.t Statement:CREATE TABLE test.public.t (num INT) User:root}
I180815 20:40:36.689918 158686 storage/replica_command.go:282 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
I180815 20:40:36.729831 158942 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] transport racer
1 [async] storage.split: processing replica
1 [async] kv.TxnCoordSender: heartbeat loop
W180815 20:40:36.732234 158686 internal/client/txn.go:514 [split,n1,s1,r23/1:/{Table/52-Max}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer
E180815 20:40:36.733096 158686 storage/queue.go:788 [split,n1,s1,r23/1:/{Table/52-Max}] unable to split [n1,s1,r23/1:/{Table/52-Max}] at key "/Table/53": split at key /Table/53 failed: node unavailable; try another peer
I180815 20:40:36.733743 158942 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] transport racer
I180815 20:40:36.789547 158280 kv/transport_race.go:67 transport race promotion: ran 39 iterations on up to 877 requests
--- FAIL: lint/TestLint (146.790s)
--- FAIL: lint/TestLint: TestLint/TestHelpURLs (41.520s)
lint_test.go:945: https://www.cockroachlabs.com/docs/v2.1/set-vars.html : Head https://www.cockroachlabs.com/docs/v2.1/set-vars.html: EOF
324: https://www.cockroachlabs.com/docs/v2.1/set-vars.html
990: https://www.cockroachlabs.com/docs/v2.1/set-vars.html
1245:https://www.cockroachlabs.com/docs/v2.1/set-vars.html
https://www.cockroachlabs.com/docs/v2.1/create-table-as.html : Head https://www.cockroachlabs.com/docs/v2.1/create-table-as.html: EOF
241:https://www.cockroachlabs.com/docs/v2.1/create-table-as.html
2 errors
------- Stdout: -------
=== PAUSE TestLint/TestHelpURLs
--- FAIL: lint/TestLint (146.790s)
--- FAIL: lint/TestLint: TestLint/TestHelpURLs (41.520s)
lint_test.go:945: https://www.cockroachlabs.com/docs/v2.1/set-vars.html : Head https://www.cockroachlabs.com/docs/v2.1/set-vars.html: EOF
324: https://www.cockroachlabs.com/docs/v2.1/set-vars.html
990: https://www.cockroachlabs.com/docs/v2.1/set-vars.html
1245:https://www.cockroachlabs.com/docs/v2.1/set-vars.html
https://www.cockroachlabs.com/docs/v2.1/create-table-as.html : Head https://www.cockroachlabs.com/docs/v2.1/create-table-as.html: EOF
241:https://www.cockroachlabs.com/docs/v2.1/create-table-as.html
2 errors
------- Stdout: -------
=== PAUSE TestLint/TestHelpURLs
--- FAIL: testrace/TestShowSessions (3.280s)
show_test.go:684: unexpected number of running sessions: 2, expected 1.
Active sessions (results might have changed since the test checked):
app: "$ internal-log-range-event", query: "INSERT INTO system.public.rangelog(\"timestamp\", \"rangeID\", \"storeID\", \"eventType\", \"otherRangeID\", info) VALUES ($1, $2, $3, $4, $5, $6)", last query: app: "", query: "SELECT active_queries, last_active_query, application_name FROM [SHOW CLUSTER SESSIONS]", last query: SELECT node_id, (now() - session_start)::FLOAT FROM [SHOW CLUSTER SESSIONS] WHERE application_name NOT LIKE 'internal-%'
------- Stdout: -------
W180815 20:40:34.423397 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180815 20:40:34.450676 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180815 20:40:34.451091 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:34.451159 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:34.455752 158099 server/config.go:496 [n?] 1 storage engine initialized
I180815 20:40:34.456002 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180815 20:40:34.456043 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B
I180815 20:40:34.514781 158099 server/node.go:373 [n?] **** cluster 6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 has been created
I180815 20:40:34.515045 158099 server/server.go:1398 [n?] **** add additional nodes by specifying --join=127.0.0.1:41091
I180815 20:40:34.516516 158099 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41091" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365634515938360
I180815 20:40:34.534941 158099 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180815 20:40:34.537635 158099 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180815 20:40:34.538394 158099 storage/stores.go:222 [n1] read 0 node addresses from persistent storage
I180815 20:40:34.539206 158099 server/node.go:697 [n1] connecting to gossip network to verify cluster ID...
I180815 20:40:34.539582 158099 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102"
I180815 20:40:34.539943 158099 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180815 20:40:34.541685 158099 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:34.541909 158099 server/server.go:1833 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180815 20:40:34.543736 158325 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:34.545405 158099 server/server.go:1535 [n1] starting https server at 127.0.0.1:34275 (use: 127.0.0.1:34275)
I180815 20:40:34.545680 158099 server/server.go:1537 [n1] starting grpc/postgres server at 127.0.0.1:41091
I180815 20:40:34.545753 158099 server/server.go:1538 [n1] advertising CockroachDB node at 127.0.0.1:41091
I180815 20:40:34.566927 158335 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180815 20:40:34.719270 158304 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180815 20:40:34.817666 158087 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180815 20:40:34.840768 158068 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180815 20:40:34.933807 158160 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180815 20:40:35.038918 158386 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
I180815 20:40:35.116602 158341 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root}
W180815 20:40:35.128647 158374 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=5ddfb628 key=/Local/Range/System/tsd/RangeDescriptor rw=true pri=0.02614628 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534365635.039580425,0 orig=1534365635.039580425,0 max=1534365635.039580425,0 wto=false rop=false seq=1
I180815 20:40:35.147031 158074 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180815 20:40:35.241602 158379 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180815 20:40:35.285924 158289 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180815 20:40:35.328157 158390 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180815 20:40:35.401113 158408 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180815 20:40:35.470572 158398 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180815 20:40:35.530967 158452 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180815 20:40:35.548516 158385 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root}
I180815 20:40:35.603367 158447 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180815 20:40:35.633708 158442 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established
I180815 20:40:35.654186 158514 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180815 20:40:35.677178 158463 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180815 20:40:35.720541 158458 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180815 20:40:35.744993 158521 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180815 20:40:35.765024 158099 server/server.go:1620 [n1] done ensuring all necessary migrations have run
I180815 20:40:35.765364 158099 server/server.go:1623 [n1] serving sql connections
I180815 20:40:35.811858 158481 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41091} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365634515938360 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365634515938360 LastUp:1534365634515938360}
I180815 20:40:35.813481 158479 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180815 20:40:35.820374 158564 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180815 20:40:35.889664 158497 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180815 20:40:35.951092 158556 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180815 20:40:36.009944 158598 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180815 20:40:36.059947 158619 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
I180815 20:40:36.119763 158606 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180815 20:40:36.175659 158630 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
W180815 20:40:36.287477 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180815 20:40:36.315831 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180815 20:40:36.316449 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:36.316518 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:36.321488 158099 server/config.go:496 [n?] 1 storage engine initialized
I180815 20:40:36.321673 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180815 20:40:36.321701 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B
W180815 20:40:36.321999 158099 gossip/gossip.go:1351 [n?] no incoming or outgoing connections
I180815 20:40:36.322565 158099 server/server.go:1400 [n?] no stores bootstrapped and --join flag specified, awaiting init command.
I180815 20:40:36.398063 158665 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:41091
I180815 20:40:36.399347 158763 gossip/server.go:219 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:32789}
I180815 20:40:36.404521 158099 server/node.go:697 [n?] connecting to gossip network to verify cluster ID...
I180815 20:40:36.405108 158099 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102"
I180815 20:40:36.408670 158717 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:40:36.415122 158716 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:40:36.422331 158099 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:40:36.431430 158099 server/node.go:428 [n?] new node allocated ID 2
I180815 20:40:36.432094 158099 gossip/gossip.go:383 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:32789" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365636431659703
I180815 20:40:36.433009 158099 storage/stores.go:222 [n2] read 0 node addresses from persistent storage
I180815 20:40:36.433961 158099 storage/stores.go:241 [n2] wrote 1 node addresses to persistent storage
I180815 20:40:36.436697 158720 storage/stores.go:241 [n1] wrote 1 node addresses to persistent storage
I180815 20:40:36.457872 158099 server/node.go:672 [n2] bootstrapped store [n2,s2]
I180815 20:40:36.460045 158099 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes []
I180815 20:40:36.461355 158099 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:36.461544 158099 server/server.go:1833 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180815 20:40:36.465410 158099 server/server.go:1535 [n2] starting https server at 127.0.0.1:36335 (use: 127.0.0.1:36335)
I180815 20:40:36.465633 158099 server/server.go:1537 [n2] starting grpc/postgres server at 127.0.0.1:32789
I180815 20:40:36.465700 158099 server/server.go:1538 [n2] advertising CockroachDB node at 127.0.0.1:32789
I180815 20:40:36.468619 158867 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:36.476134 158099 server/server.go:1620 [n2] done ensuring all necessary migrations have run
I180815 20:40:36.476492 158099 server/server.go:1623 [n2] serving sql connections
I180815 20:40:36.576176 158869 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version
I180815 20:40:36.577336 158871 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:32789} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365636431659703 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365636431659703 LastUp:1534365636431659703}
I180815 20:40:36.617219 158877 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established
I180815 20:40:36.623403 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root}
I180815 20:40:36.624884 158702 storage/replica_command.go:282 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
I180815 20:40:36.685520 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_table", target: 53, info: {TableName:test.public.t Statement:CREATE TABLE test.public.t (num INT) User:root}
I180815 20:40:36.689918 158686 storage/replica_command.go:282 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
I180815 20:40:36.729831 158942 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] transport racer
1 [async] storage.split: processing replica
1 [async] kv.TxnCoordSender: heartbeat loop
W180815 20:40:36.732234 158686 internal/client/txn.go:514 [split,n1,s1,r23/1:/{Table/52-Max}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer
E180815 20:40:36.733096 158686 storage/queue.go:788 [split,n1,s1,r23/1:/{Table/52-Max}] unable to split [n1,s1,r23/1:/{Table/52-Max}] at key "/Table/53": split at key /Table/53 failed: node unavailable; try another peer
I180815 20:40:36.733743 158942 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] transport racer
I180815 20:40:36.789547 158280 kv/transport_race.go:67 transport race promotion: ran 39 iterations on up to 877 requests
------- Stdout: -------
W180815 20:20:04.735629 158449 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180815 20:20:04.756225 158449 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180815 20:20:04.756547 158449 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:20:04.756590 158449 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:20:04.770676 158449 server/config.go:496 [n?] 1 storage engine initialized
I180815 20:20:04.770775 158449 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180815 20:20:04.770795 158449 server/config.go:499 [n?] store 0: in-memory, size 0 B
I180815 20:20:04.774753 158449 server/node.go:373 [n?] **** cluster 897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a has been created
I180815 20:20:04.774783 158449 server/server.go:1398 [n?] **** add additional nodes by specifying --join=127.0.0.1:35641
I180815 20:20:04.774934 158449 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:35641" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534364404774867605
I180815 20:20:04.788664 158449 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180815 20:20:04.788825 158449 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180815 20:20:04.788933 158449 storage/stores.go:222 [n1] read 0 node addresses from persistent storage
I180815 20:20:04.789122 158449 server/node.go:697 [n1] connecting to gossip network to verify cluster ID...
I180815 20:20:04.792799 158449 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a"
I180815 20:20:04.792912 158449 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180815 20:20:04.794388 158449 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:20:04.794466 158449 server/server.go:1833 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180815 20:20:04.794618 158449 server/server.go:1535 [n1] starting https server at 127.0.0.1:35159 (use: 127.0.0.1:35159)
I180815 20:20:04.800514 158449 server/server.go:1537 [n1] starting grpc/postgres server at 127.0.0.1:35641
I180815 20:20:04.800600 158449 server/server.go:1538 [n1] advertising CockroachDB node at 127.0.0.1:35641
I180815 20:20:04.795116 158433 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180815 20:20:04.795346 158681 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
E180815 20:20:04.816788 158723 storage/queue.go:788 [replicate,n1,s1,r1/1:/{Min-System/}] range requires a replication change, but lacks a quorum of live replicas (0/1)
I180815 20:20:04.817580 158712 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180815 20:20:04.995544 158709 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180815 20:20:04.997521 158738 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180815 20:20:05.014104 158744 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180815 20:20:05.031468 158456 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root}
I180815 20:20:05.039572 158756 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
I180815 20:20:05.057269 158729 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180815 20:20:05.087390 158775 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root}
I180815 20:20:05.090490 158462 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180815 20:20:05.100042 158794 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180815 20:20:05.103041 158804 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180815 20:20:05.108352 158799 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180815 20:20:05.110062 158381 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180815 20:20:05.114153 158449 server/server.go:1620 [n1] done ensuring all necessary migrations have run
I180815 20:20:05.114189 158449 server/server.go:1623 [n1] serving sql connections
I180815 20:20:05.117362 158850 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:35641} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534364404774867605 LocalityAddress:[]} ClusterID:897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a StartedAt:1534364404774867605 LastUp:1534364404774867605}
I180815 20:20:05.117710 158736 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180815 20:20:05.120942 158749 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180815 20:20:05.126759 158782 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180815 20:20:05.133741 158869 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180815 20:20:05.148799 158890 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180815 20:20:05.155296 158901 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180815 20:20:05.177525 158811 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180815 20:20:05.184671 158931 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180815 20:20:05.193079 158916 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180815 20:20:05.213605 158926 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180815 20:20:05.220306 158986 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180815 20:20:05.226131 159001 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
W180815 20:20:05.232969 159014 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=530964b3 key=/Local/Range/Table/22/RangeDescriptor rw=true pri=0.02693795 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534364405.227916665,1 orig=1534364405.226151335,0 max=1534364405.226151335,0 wto=false rop=false seq=1
I180815 20:20:05.235499 158952 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180815 20:20:05.245022 159005 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
W180815 20:20:05.303250 158449 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180815 20:20:05.310656 158449 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180815 20:20:05.315402 158449 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:20:05.315480 158449 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:20:05.330854 158449 server/config.go:496 [n?] 1 storage engine initialized
I180815 20:20:05.330964 158449 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180815 20:20:05.330983 158449 server/config.go:499 [n?] store 0: in-memory, size 0 B
W180815 20:20:05.331113 158449 gossip/gossip.go:1351 [n?] no incoming or outgoing connections
I180815 20:20:05.331177 158449 server/server.go:1400 [n?] no stores bootstrapped and --join flag specified, awaiting init command.
I180815 20:20:05.344590 159092 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:35641
I180815 20:20:05.345720 159045 gossip/server.go:219 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:45155}
I180815 20:20:05.360942 158449 server/node.go:697 [n?] connecting to gossip network to verify cluster ID...
I180815 20:20:05.360981 158449 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a"
I180815 20:20:05.361374 159100 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:20:05.365029 159099 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:20:05.366271 158449 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:20:05.367766 158449 server/node.go:428 [n?] new node allocated ID 2
I180815 20:20:05.367934 158449 gossip/gossip.go:383 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:45155" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534364405367814511
I180815 20:20:05.368073 158449 storage/stores.go:222 [n2] read 0 node addresses from persistent storage
I180815 20:20:05.368143 158449 storage/stores.go:241 [n2] wrote 1 node addresses to persistent storage
I180815 20:20:05.369673 158824 storage/stores.go:241 [n1] wrote 1 node addresses to persistent storage
I180815 20:20:05.371353 158449 server/node.go:672 [n2] bootstrapped store [n2,s2]
I180815 20:20:05.374076 158449 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes []
I180815 20:20:05.374378 158449 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:20:05.374427 158449 server/server.go:1833 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180815 20:20:05.374526 158449 server/server.go:1535 [n2] starting https server at 127.0.0.1:43861 (use: 127.0.0.1:43861)
I180815 20:20:05.374557 158449 server/server.go:1537 [n2] starting grpc/postgres server at 127.0.0.1:45155
I180815 20:20:05.374579 158449 server/server.go:1538 [n2] advertising CockroachDB node at 127.0.0.1:45155
I180815 20:20:05.377097 159053 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:20:05.379673 158449 server/server.go:1620 [n2] done ensuring all necessary migrations have run
I180815 20:20:05.397824 158449 server/server.go:1623 [n2] serving sql connections
I180815 20:20:05.424198 159220 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established
I180815 20:20:05.447142 159221 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version
I180815 20:20:05.458760 159223 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:45155} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534364405367814511 LocalityAddress:[]} ClusterID:897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a StartedAt:1534364405367814511 LastUp:1534364405367814511}
I180815 20:20:05.461728 159262 sql/event_log.go:126 [n1,client=127.0.0.1:39484,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root}
I180815 20:20:05.466851 159264 storage/replica_command.go:282 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
I180815 20:20:05.468550 159262 sql/event_log.go:126 [n1,client=127.0.0.1:39484,user=root] Event: "create_table", target: 53, info: {TableName:test.public.t Statement:CREATE TABLE test.public.t (num INT) User:root}
I180815 20:20:05.495103 159292 storage/replica_command.go:282 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
W180815 20:20:05.516400 159262 sql/crdb_internal.go:900 [n1,client=127.0.0.1:39484,user=root] failed to dial into node 2 (LIVE): initial connection heartbeat failed: rpc error: code = Unavailable desc = all SubConns are in TransientFailure, latest connection error: connection error: desc = "transport: Error while dialing dial tcp 127.0.0.1:45155: connect: connection refused"
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
teamcity: failed tests on master: testrace/TestShowSessions, lint/TestLint - The following tests appear to have failed:
[#839592](https://teamcity.cockroachdb.com/viewLog.html?buildId=839592):
```
--- FAIL: testrace/TestShowSessions (2.450s)
show_test.go:684: unexpected number of running sessions: 2, expected 1.
Active sessions (results might have changed since the test checked):
app: "$ internal-log-range-event", query: "INSERT INTO system.public.rangelog(\"timestamp\", \"rangeID\", \"storeID\", \"eventType\", \"otherRangeID\", info) VALUES ($1, $2, $3, $4, $5, $6)", last query: app: "", query: "SELECT active_queries, last_active_query, application_name FROM [SHOW CLUSTER SESSIONS]", last query: SELECT node_id, (now() - session_start)::FLOAT FROM [SHOW CLUSTER SESSIONS] WHERE application_name NOT LIKE 'internal-%'
------- Stdout: -------
W180815 20:40:34.423397 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180815 20:40:34.450676 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180815 20:40:34.451091 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:34.451159 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:34.455752 158099 server/config.go:496 [n?] 1 storage engine initialized
I180815 20:40:34.456002 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180815 20:40:34.456043 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B
I180815 20:40:34.514781 158099 server/node.go:373 [n?] **** cluster 6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 has been created
I180815 20:40:34.515045 158099 server/server.go:1398 [n?] **** add additional nodes by specifying --join=127.0.0.1:41091
I180815 20:40:34.516516 158099 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41091" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365634515938360
I180815 20:40:34.534941 158099 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180815 20:40:34.537635 158099 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180815 20:40:34.538394 158099 storage/stores.go:222 [n1] read 0 node addresses from persistent storage
I180815 20:40:34.539206 158099 server/node.go:697 [n1] connecting to gossip network to verify cluster ID...
I180815 20:40:34.539582 158099 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102"
I180815 20:40:34.539943 158099 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180815 20:40:34.541685 158099 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:34.541909 158099 server/server.go:1833 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180815 20:40:34.543736 158325 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:34.545405 158099 server/server.go:1535 [n1] starting https server at 127.0.0.1:34275 (use: 127.0.0.1:34275)
I180815 20:40:34.545680 158099 server/server.go:1537 [n1] starting grpc/postgres server at 127.0.0.1:41091
I180815 20:40:34.545753 158099 server/server.go:1538 [n1] advertising CockroachDB node at 127.0.0.1:41091
I180815 20:40:34.566927 158335 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180815 20:40:34.719270 158304 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180815 20:40:34.817666 158087 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180815 20:40:34.840768 158068 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180815 20:40:34.933807 158160 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180815 20:40:35.038918 158386 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
I180815 20:40:35.116602 158341 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root}
W180815 20:40:35.128647 158374 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=5ddfb628 key=/Local/Range/System/tsd/RangeDescriptor rw=true pri=0.02614628 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534365635.039580425,0 orig=1534365635.039580425,0 max=1534365635.039580425,0 wto=false rop=false seq=1
I180815 20:40:35.147031 158074 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180815 20:40:35.241602 158379 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180815 20:40:35.285924 158289 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180815 20:40:35.328157 158390 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180815 20:40:35.401113 158408 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180815 20:40:35.470572 158398 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180815 20:40:35.530967 158452 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180815 20:40:35.548516 158385 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root}
I180815 20:40:35.603367 158447 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180815 20:40:35.633708 158442 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established
I180815 20:40:35.654186 158514 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180815 20:40:35.677178 158463 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180815 20:40:35.720541 158458 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180815 20:40:35.744993 158521 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180815 20:40:35.765024 158099 server/server.go:1620 [n1] done ensuring all necessary migrations have run
I180815 20:40:35.765364 158099 server/server.go:1623 [n1] serving sql connections
I180815 20:40:35.811858 158481 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41091} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365634515938360 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365634515938360 LastUp:1534365634515938360}
I180815 20:40:35.813481 158479 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180815 20:40:35.820374 158564 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180815 20:40:35.889664 158497 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180815 20:40:35.951092 158556 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180815 20:40:36.009944 158598 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180815 20:40:36.059947 158619 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
I180815 20:40:36.119763 158606 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180815 20:40:36.175659 158630 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
W180815 20:40:36.287477 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180815 20:40:36.315831 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180815 20:40:36.316449 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:36.316518 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:36.321488 158099 server/config.go:496 [n?] 1 storage engine initialized
I180815 20:40:36.321673 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180815 20:40:36.321701 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B
W180815 20:40:36.321999 158099 gossip/gossip.go:1351 [n?] no incoming or outgoing connections
I180815 20:40:36.322565 158099 server/server.go:1400 [n?] no stores bootstrapped and --join flag specified, awaiting init command.
I180815 20:40:36.398063 158665 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:41091
I180815 20:40:36.399347 158763 gossip/server.go:219 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:32789}
I180815 20:40:36.404521 158099 server/node.go:697 [n?] connecting to gossip network to verify cluster ID...
I180815 20:40:36.405108 158099 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102"
I180815 20:40:36.408670 158717 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:40:36.415122 158716 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:40:36.422331 158099 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:40:36.431430 158099 server/node.go:428 [n?] new node allocated ID 2
I180815 20:40:36.432094 158099 gossip/gossip.go:383 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:32789" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365636431659703
I180815 20:40:36.433009 158099 storage/stores.go:222 [n2] read 0 node addresses from persistent storage
I180815 20:40:36.433961 158099 storage/stores.go:241 [n2] wrote 1 node addresses to persistent storage
I180815 20:40:36.436697 158720 storage/stores.go:241 [n1] wrote 1 node addresses to persistent storage
I180815 20:40:36.457872 158099 server/node.go:672 [n2] bootstrapped store [n2,s2]
I180815 20:40:36.460045 158099 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes []
I180815 20:40:36.461355 158099 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:36.461544 158099 server/server.go:1833 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180815 20:40:36.465410 158099 server/server.go:1535 [n2] starting https server at 127.0.0.1:36335 (use: 127.0.0.1:36335)
I180815 20:40:36.465633 158099 server/server.go:1537 [n2] starting grpc/postgres server at 127.0.0.1:32789
I180815 20:40:36.465700 158099 server/server.go:1538 [n2] advertising CockroachDB node at 127.0.0.1:32789
I180815 20:40:36.468619 158867 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:36.476134 158099 server/server.go:1620 [n2] done ensuring all necessary migrations have run
I180815 20:40:36.476492 158099 server/server.go:1623 [n2] serving sql connections
I180815 20:40:36.576176 158869 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version
I180815 20:40:36.577336 158871 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:32789} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365636431659703 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365636431659703 LastUp:1534365636431659703}
I180815 20:40:36.617219 158877 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established
I180815 20:40:36.623403 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root}
I180815 20:40:36.624884 158702 storage/replica_command.go:282 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
I180815 20:40:36.685520 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_table", target: 53, info: {TableName:test.public.t Statement:CREATE TABLE test.public.t (num INT) User:root}
I180815 20:40:36.689918 158686 storage/replica_command.go:282 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
I180815 20:40:36.729831 158942 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] transport racer
1 [async] storage.split: processing replica
1 [async] kv.TxnCoordSender: heartbeat loop
W180815 20:40:36.732234 158686 internal/client/txn.go:514 [split,n1,s1,r23/1:/{Table/52-Max}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer
E180815 20:40:36.733096 158686 storage/queue.go:788 [split,n1,s1,r23/1:/{Table/52-Max}] unable to split [n1,s1,r23/1:/{Table/52-Max}] at key "/Table/53": split at key /Table/53 failed: node unavailable; try another peer
I180815 20:40:36.733743 158942 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] transport racer
I180815 20:40:36.789547 158280 kv/transport_race.go:67 transport race promotion: ran 39 iterations on up to 877 requests
--- FAIL: lint/TestLint (146.790s)
--- FAIL: lint/TestLint: TestLint/TestHelpURLs (41.520s)
lint_test.go:945: https://www.cockroachlabs.com/docs/v2.1/set-vars.html : Head https://www.cockroachlabs.com/docs/v2.1/set-vars.html: EOF
324: https://www.cockroachlabs.com/docs/v2.1/set-vars.html
990: https://www.cockroachlabs.com/docs/v2.1/set-vars.html
1245:https://www.cockroachlabs.com/docs/v2.1/set-vars.html
https://www.cockroachlabs.com/docs/v2.1/create-table-as.html : Head https://www.cockroachlabs.com/docs/v2.1/create-table-as.html: EOF
241:https://www.cockroachlabs.com/docs/v2.1/create-table-as.html
2 errors
------- Stdout: -------
=== PAUSE TestLint/TestHelpURLs
--- FAIL: lint/TestLint (146.790s)
--- FAIL: lint/TestLint: TestLint/TestHelpURLs (41.520s)
lint_test.go:945: https://www.cockroachlabs.com/docs/v2.1/set-vars.html : Head https://www.cockroachlabs.com/docs/v2.1/set-vars.html: EOF
324: https://www.cockroachlabs.com/docs/v2.1/set-vars.html
990: https://www.cockroachlabs.com/docs/v2.1/set-vars.html
1245:https://www.cockroachlabs.com/docs/v2.1/set-vars.html
https://www.cockroachlabs.com/docs/v2.1/create-table-as.html : Head https://www.cockroachlabs.com/docs/v2.1/create-table-as.html: EOF
241:https://www.cockroachlabs.com/docs/v2.1/create-table-as.html
2 errors
------- Stdout: -------
=== PAUSE TestLint/TestHelpURLs
--- FAIL: testrace/TestShowSessions (3.280s)
show_test.go:684: unexpected number of running sessions: 2, expected 1.
Active sessions (results might have changed since the test checked):
app: "$ internal-log-range-event", query: "INSERT INTO system.public.rangelog(\"timestamp\", \"rangeID\", \"storeID\", \"eventType\", \"otherRangeID\", info) VALUES ($1, $2, $3, $4, $5, $6)", last query: app: "", query: "SELECT active_queries, last_active_query, application_name FROM [SHOW CLUSTER SESSIONS]", last query: SELECT node_id, (now() - session_start)::FLOAT FROM [SHOW CLUSTER SESSIONS] WHERE application_name NOT LIKE 'internal-%'
------- Stdout: -------
W180815 20:40:34.423397 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180815 20:40:34.450676 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180815 20:40:34.451091 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:34.451159 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:34.455752 158099 server/config.go:496 [n?] 1 storage engine initialized
I180815 20:40:34.456002 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180815 20:40:34.456043 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B
I180815 20:40:34.514781 158099 server/node.go:373 [n?] **** cluster 6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 has been created
I180815 20:40:34.515045 158099 server/server.go:1398 [n?] **** add additional nodes by specifying --join=127.0.0.1:41091
I180815 20:40:34.516516 158099 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41091" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365634515938360
I180815 20:40:34.534941 158099 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180815 20:40:34.537635 158099 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180815 20:40:34.538394 158099 storage/stores.go:222 [n1] read 0 node addresses from persistent storage
I180815 20:40:34.539206 158099 server/node.go:697 [n1] connecting to gossip network to verify cluster ID...
I180815 20:40:34.539582 158099 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102"
I180815 20:40:34.539943 158099 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180815 20:40:34.541685 158099 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:34.541909 158099 server/server.go:1833 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180815 20:40:34.543736 158325 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:34.545405 158099 server/server.go:1535 [n1] starting https server at 127.0.0.1:34275 (use: 127.0.0.1:34275)
I180815 20:40:34.545680 158099 server/server.go:1537 [n1] starting grpc/postgres server at 127.0.0.1:41091
I180815 20:40:34.545753 158099 server/server.go:1538 [n1] advertising CockroachDB node at 127.0.0.1:41091
I180815 20:40:34.566927 158335 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180815 20:40:34.719270 158304 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180815 20:40:34.817666 158087 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180815 20:40:34.840768 158068 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180815 20:40:34.933807 158160 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180815 20:40:35.038918 158386 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
I180815 20:40:35.116602 158341 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root}
W180815 20:40:35.128647 158374 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=5ddfb628 key=/Local/Range/System/tsd/RangeDescriptor rw=true pri=0.02614628 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534365635.039580425,0 orig=1534365635.039580425,0 max=1534365635.039580425,0 wto=false rop=false seq=1
I180815 20:40:35.147031 158074 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180815 20:40:35.241602 158379 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180815 20:40:35.285924 158289 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180815 20:40:35.328157 158390 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180815 20:40:35.401113 158408 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180815 20:40:35.470572 158398 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180815 20:40:35.530967 158452 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180815 20:40:35.548516 158385 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root}
I180815 20:40:35.603367 158447 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180815 20:40:35.633708 158442 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established
I180815 20:40:35.654186 158514 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180815 20:40:35.677178 158463 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180815 20:40:35.720541 158458 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180815 20:40:35.744993 158521 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180815 20:40:35.765024 158099 server/server.go:1620 [n1] done ensuring all necessary migrations have run
I180815 20:40:35.765364 158099 server/server.go:1623 [n1] serving sql connections
I180815 20:40:35.811858 158481 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41091} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365634515938360 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365634515938360 LastUp:1534365634515938360}
I180815 20:40:35.813481 158479 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180815 20:40:35.820374 158564 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180815 20:40:35.889664 158497 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180815 20:40:35.951092 158556 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180815 20:40:36.009944 158598 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180815 20:40:36.059947 158619 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
I180815 20:40:36.119763 158606 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180815 20:40:36.175659 158630 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
W180815 20:40:36.287477 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180815 20:40:36.315831 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180815 20:40:36.316449 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:36.316518 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:40:36.321488 158099 server/config.go:496 [n?] 1 storage engine initialized
I180815 20:40:36.321673 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180815 20:40:36.321701 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B
W180815 20:40:36.321999 158099 gossip/gossip.go:1351 [n?] no incoming or outgoing connections
I180815 20:40:36.322565 158099 server/server.go:1400 [n?] no stores bootstrapped and --join flag specified, awaiting init command.
I180815 20:40:36.398063 158665 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:41091
I180815 20:40:36.399347 158763 gossip/server.go:219 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:32789}
I180815 20:40:36.404521 158099 server/node.go:697 [n?] connecting to gossip network to verify cluster ID...
I180815 20:40:36.405108 158099 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102"
I180815 20:40:36.408670 158717 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:40:36.415122 158716 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:40:36.422331 158099 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:40:36.431430 158099 server/node.go:428 [n?] new node allocated ID 2
I180815 20:40:36.432094 158099 gossip/gossip.go:383 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:32789" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365636431659703
I180815 20:40:36.433009 158099 storage/stores.go:222 [n2] read 0 node addresses from persistent storage
I180815 20:40:36.433961 158099 storage/stores.go:241 [n2] wrote 1 node addresses to persistent storage
I180815 20:40:36.436697 158720 storage/stores.go:241 [n1] wrote 1 node addresses to persistent storage
I180815 20:40:36.457872 158099 server/node.go:672 [n2] bootstrapped store [n2,s2]
I180815 20:40:36.460045 158099 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes []
I180815 20:40:36.461355 158099 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:36.461544 158099 server/server.go:1833 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180815 20:40:36.465410 158099 server/server.go:1535 [n2] starting https server at 127.0.0.1:36335 (use: 127.0.0.1:36335)
I180815 20:40:36.465633 158099 server/server.go:1537 [n2] starting grpc/postgres server at 127.0.0.1:32789
I180815 20:40:36.465700 158099 server/server.go:1538 [n2] advertising CockroachDB node at 127.0.0.1:32789
I180815 20:40:36.468619 158867 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:40:36.476134 158099 server/server.go:1620 [n2] done ensuring all necessary migrations have run
I180815 20:40:36.476492 158099 server/server.go:1623 [n2] serving sql connections
I180815 20:40:36.576176 158869 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version
I180815 20:40:36.577336 158871 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:32789} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365636431659703 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365636431659703 LastUp:1534365636431659703}
I180815 20:40:36.617219 158877 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established
I180815 20:40:36.623403 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root}
I180815 20:40:36.624884 158702 storage/replica_command.go:282 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
I180815 20:40:36.685520 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_table", target: 53, info: {TableName:test.public.t Statement:CREATE TABLE test.public.t (num INT) User:root}
I180815 20:40:36.689918 158686 storage/replica_command.go:282 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
I180815 20:40:36.729831 158942 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] transport racer
1 [async] storage.split: processing replica
1 [async] kv.TxnCoordSender: heartbeat loop
W180815 20:40:36.732234 158686 internal/client/txn.go:514 [split,n1,s1,r23/1:/{Table/52-Max}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer
E180815 20:40:36.733096 158686 storage/queue.go:788 [split,n1,s1,r23/1:/{Table/52-Max}] unable to split [n1,s1,r23/1:/{Table/52-Max}] at key "/Table/53": split at key /Table/53 failed: node unavailable; try another peer
I180815 20:40:36.733743 158942 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] transport racer
I180815 20:40:36.789547 158280 kv/transport_race.go:67 transport race promotion: ran 39 iterations on up to 877 requests
------- Stdout: -------
W180815 20:20:04.735629 158449 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180815 20:20:04.756225 158449 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180815 20:20:04.756547 158449 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:20:04.756590 158449 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:20:04.770676 158449 server/config.go:496 [n?] 1 storage engine initialized
I180815 20:20:04.770775 158449 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180815 20:20:04.770795 158449 server/config.go:499 [n?] store 0: in-memory, size 0 B
I180815 20:20:04.774753 158449 server/node.go:373 [n?] **** cluster 897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a has been created
I180815 20:20:04.774783 158449 server/server.go:1398 [n?] **** add additional nodes by specifying --join=127.0.0.1:35641
I180815 20:20:04.774934 158449 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:35641" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534364404774867605
I180815 20:20:04.788664 158449 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180815 20:20:04.788825 158449 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180815 20:20:04.788933 158449 storage/stores.go:222 [n1] read 0 node addresses from persistent storage
I180815 20:20:04.789122 158449 server/node.go:697 [n1] connecting to gossip network to verify cluster ID...
I180815 20:20:04.792799 158449 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a"
I180815 20:20:04.792912 158449 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180815 20:20:04.794388 158449 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:20:04.794466 158449 server/server.go:1833 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180815 20:20:04.794618 158449 server/server.go:1535 [n1] starting https server at 127.0.0.1:35159 (use: 127.0.0.1:35159)
I180815 20:20:04.800514 158449 server/server.go:1537 [n1] starting grpc/postgres server at 127.0.0.1:35641
I180815 20:20:04.800600 158449 server/server.go:1538 [n1] advertising CockroachDB node at 127.0.0.1:35641
I180815 20:20:04.795116 158433 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180815 20:20:04.795346 158681 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
E180815 20:20:04.816788 158723 storage/queue.go:788 [replicate,n1,s1,r1/1:/{Min-System/}] range requires a replication change, but lacks a quorum of live replicas (0/1)
I180815 20:20:04.817580 158712 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180815 20:20:04.995544 158709 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180815 20:20:04.997521 158738 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180815 20:20:05.014104 158744 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180815 20:20:05.031468 158456 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root}
I180815 20:20:05.039572 158756 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
I180815 20:20:05.057269 158729 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180815 20:20:05.087390 158775 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root}
I180815 20:20:05.090490 158462 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180815 20:20:05.100042 158794 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180815 20:20:05.103041 158804 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180815 20:20:05.108352 158799 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180815 20:20:05.110062 158381 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180815 20:20:05.114153 158449 server/server.go:1620 [n1] done ensuring all necessary migrations have run
I180815 20:20:05.114189 158449 server/server.go:1623 [n1] serving sql connections
I180815 20:20:05.117362 158850 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:35641} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534364404774867605 LocalityAddress:[]} ClusterID:897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a StartedAt:1534364404774867605 LastUp:1534364404774867605}
I180815 20:20:05.117710 158736 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180815 20:20:05.120942 158749 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180815 20:20:05.126759 158782 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180815 20:20:05.133741 158869 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180815 20:20:05.148799 158890 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180815 20:20:05.155296 158901 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180815 20:20:05.177525 158811 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180815 20:20:05.184671 158931 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180815 20:20:05.193079 158916 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180815 20:20:05.213605 158926 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180815 20:20:05.220306 158986 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180815 20:20:05.226131 159001 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
W180815 20:20:05.232969 159014 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=530964b3 key=/Local/Range/Table/22/RangeDescriptor rw=true pri=0.02693795 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534364405.227916665,1 orig=1534364405.226151335,0 max=1534364405.226151335,0 wto=false rop=false seq=1
I180815 20:20:05.235499 158952 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180815 20:20:05.245022 159005 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
W180815 20:20:05.303250 158449 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180815 20:20:05.310656 158449 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180815 20:20:05.315402 158449 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:20:05.315480 158449 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180815 20:20:05.330854 158449 server/config.go:496 [n?] 1 storage engine initialized
I180815 20:20:05.330964 158449 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180815 20:20:05.330983 158449 server/config.go:499 [n?] store 0: in-memory, size 0 B
W180815 20:20:05.331113 158449 gossip/gossip.go:1351 [n?] no incoming or outgoing connections
I180815 20:20:05.331177 158449 server/server.go:1400 [n?] no stores bootstrapped and --join flag specified, awaiting init command.
I180815 20:20:05.344590 159092 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:35641
I180815 20:20:05.345720 159045 gossip/server.go:219 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:45155}
I180815 20:20:05.360942 158449 server/node.go:697 [n?] connecting to gossip network to verify cluster ID...
I180815 20:20:05.360981 158449 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a"
I180815 20:20:05.361374 159100 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:20:05.365029 159099 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:20:05.366271 158449 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I180815 20:20:05.367766 158449 server/node.go:428 [n?] new node allocated ID 2
I180815 20:20:05.367934 158449 gossip/gossip.go:383 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:45155" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534364405367814511
I180815 20:20:05.368073 158449 storage/stores.go:222 [n2] read 0 node addresses from persistent storage
I180815 20:20:05.368143 158449 storage/stores.go:241 [n2] wrote 1 node addresses to persistent storage
I180815 20:20:05.369673 158824 storage/stores.go:241 [n1] wrote 1 node addresses to persistent storage
I180815 20:20:05.371353 158449 server/node.go:672 [n2] bootstrapped store [n2,s2]
I180815 20:20:05.374076 158449 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes []
I180815 20:20:05.374378 158449 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:20:05.374427 158449 server/server.go:1833 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180815 20:20:05.374526 158449 server/server.go:1535 [n2] starting https server at 127.0.0.1:43861 (use: 127.0.0.1:43861)
I180815 20:20:05.374557 158449 server/server.go:1537 [n2] starting grpc/postgres server at 127.0.0.1:45155
I180815 20:20:05.374579 158449 server/server.go:1538 [n2] advertising CockroachDB node at 127.0.0.1:45155
I180815 20:20:05.377097 159053 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180815 20:20:05.379673 158449 server/server.go:1620 [n2] done ensuring all necessary migrations have run
I180815 20:20:05.397824 158449 server/server.go:1623 [n2] serving sql connections
I180815 20:20:05.424198 159220 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established
I180815 20:20:05.447142 159221 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version
I180815 20:20:05.458760 159223 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:45155} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534364405367814511 LocalityAddress:[]} ClusterID:897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a StartedAt:1534364405367814511 LastUp:1534364405367814511}
I180815 20:20:05.461728 159262 sql/event_log.go:126 [n1,client=127.0.0.1:39484,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root}
I180815 20:20:05.466851 159264 storage/replica_command.go:282 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
I180815 20:20:05.468550 159262 sql/event_log.go:126 [n1,client=127.0.0.1:39484,user=root] Event: "create_table", target: 53, info: {TableName:test.public.t Statement:CREATE TABLE test.public.t (num INT) User:root}
I180815 20:20:05.495103 159292 storage/replica_command.go:282 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
W180815 20:20:05.516400 159262 sql/crdb_internal.go:900 [n1,client=127.0.0.1:39484,user=root] failed to dial into node 2 (LIVE): initial connection heartbeat failed: rpc error: code = Unavailable desc = all SubConns are in TransientFailure, latest connection error: connection error: desc = "transport: Error while dialing dial tcp 127.0.0.1:45155: connect: connection refused"
```
Please assign, take a look and update the issue accordingly.
|
test
|
teamcity failed tests on master testrace testshowsessions lint testlint the following tests appear to have failed fail testrace testshowsessions show test go unexpected number of running sessions expected active sessions results might have changed since the test checked app internal log range event query insert into system public rangelog timestamp rangeid storeid eventtype otherrangeid info values last query app query select active queries last active query application name from last query select node id now session start float from where application name not like internal stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases queries writes bytesperreplica pmax writesperreplica pmax storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at storage replica command go initiating a split of this range at key system storage replica command go initiating a split of this range at key system nodeliveness storage replica command go initiating a split of this range at key system nodelivenessmax sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user root storage replica command go initiating a split of this range at key system tsd storage replica command go initiating a split of this range at key system tse sql event log go event set cluster setting target info settingname version value user root storage intent resolver go failed to push during intent resolution failed to push split id key local range system tsd rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table systemconfigspan start storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname trace debug enable value false user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname cluster secret value gen random uuid string user root storage replica command go initiating a split of this range at key table rpc nodedialer nodedialer go connection to established sql event log go event create database target info databasename defaultdb statement create database if not exists defaultdb user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename postgres statement create database if not exists postgres user root storage replica command go initiating a split of this range at key table server server go done ensuring all necessary migrations have run server server go serving sql connections sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup server server update go no need to upgrade cluster already at the newest version storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b gossip gossip go no incoming or outgoing connections server server go no stores bootstrapped and join flag specified awaiting init command gossip client go started gossip client to gossip server go received initial cluster verification connection from tcp server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping server node go new node allocated id gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage stores go read node addresses from persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage server node go bootstrapped store server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go done ensuring all necessary migrations have run server server go serving sql connections server server update go no need to upgrade cluster already at the newest version sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup rpc nodedialer nodedialer go connection to established sql event log go event create database target info databasename test statement create database if not exists test user root storage replica command go initiating a split of this range at key table sql event log go event create table target info tablename test public t statement create table test public t num int user root storage replica command go initiating a split of this range at key table util stop stopper go quiescing tasks left transport racer storage split processing replica kv txncoordsender heartbeat loop internal client txn go failure aborting transaction node unavailable try another peer abort caused by node unavailable try another peer storage queue go unable to split at key table split at key table failed node unavailable try another peer util stop stopper go quiescing tasks left transport racer kv transport race go transport race promotion ran iterations on up to requests fail lint testlint fail lint testlint testlint testhelpurls lint test go head eof head eof errors stdout pause testlint testhelpurls fail lint testlint fail lint testlint testlint testhelpurls lint test go head eof head eof errors stdout pause testlint testhelpurls fail testrace testshowsessions show test go unexpected number of running sessions expected active sessions results might have changed since the test checked app internal log range event query insert into system public rangelog timestamp rangeid storeid eventtype otherrangeid info values last query app query select active queries last active query application name from last query select node id now session start float from where application name not like internal stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases queries writes bytesperreplica pmax writesperreplica pmax storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at storage replica command go initiating a split of this range at key system storage replica command go initiating a split of this range at key system nodeliveness storage replica command go initiating a split of this range at key system nodelivenessmax sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user root storage replica command go initiating a split of this range at key system tsd storage replica command go initiating a split of this range at key system tse sql event log go event set cluster setting target info settingname version value user root storage intent resolver go failed to push during intent resolution failed to push split id key local range system tsd rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table systemconfigspan start storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname trace debug enable value false user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname cluster secret value gen random uuid string user root storage replica command go initiating a split of this range at key table rpc nodedialer nodedialer go connection to established sql event log go event create database target info databasename defaultdb statement create database if not exists defaultdb user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename postgres statement create database if not exists postgres user root storage replica command go initiating a split of this range at key table server server go done ensuring all necessary migrations have run server server go serving sql connections sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup server server update go no need to upgrade cluster already at the newest version storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b gossip gossip go no incoming or outgoing connections server server go no stores bootstrapped and join flag specified awaiting init command gossip client go started gossip client to gossip server go received initial cluster verification connection from tcp server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping server node go new node allocated id gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage stores go read node addresses from persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage server node go bootstrapped store server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go done ensuring all necessary migrations have run server server go serving sql connections server server update go no need to upgrade cluster already at the newest version sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup rpc nodedialer nodedialer go connection to established sql event log go event create database target info databasename test statement create database if not exists test user root storage replica command go initiating a split of this range at key table sql event log go event create table target info tablename test public t statement create table test public t num int user root storage replica command go initiating a split of this range at key table util stop stopper go quiescing tasks left transport racer storage split processing replica kv txncoordsender heartbeat loop internal client txn go failure aborting transaction node unavailable try another peer abort caused by node unavailable try another peer storage queue go unable to split at key table split at key table failed node unavailable try another peer util stop stopper go quiescing tasks left transport racer kv transport race go transport race promotion ran iterations on up to requests stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases queries writes bytesperreplica pmax writesperreplica pmax storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at storage replica command go initiating a split of this range at key system server status recorder go available memory from cgroups eib exceeds system memory gib using system memory storage queue go range requires a replication change but lacks a quorum of live replicas storage replica command go initiating a split of this range at key system nodeliveness sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user root storage replica command go initiating a split of this range at key system nodelivenessmax storage replica command go initiating a split of this range at key system tsd sql event log go event set cluster setting target info settingname version value user root storage replica command go initiating a split of this range at key system tse sql event log go event set cluster setting target info settingname trace debug enable value false user root sql event log go event set cluster setting target info settingname cluster secret value gen random uuid string user root storage replica command go initiating a split of this range at key table systemconfigspan start sql event log go event create database target info databasename defaultdb statement create database if not exists defaultdb user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename postgres statement create database if not exists postgres user root storage replica command go initiating a split of this range at key table server server go done ensuring all necessary migrations have run server server go serving sql connections sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup server server update go no need to upgrade cluster already at the newest version storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b gossip gossip go no incoming or outgoing connections server server go no stores bootstrapped and join flag specified awaiting init command gossip client go started gossip client to gossip server go received initial cluster verification connection from tcp server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping server node go new node allocated id gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage stores go read node addresses from persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage server node go bootstrapped store server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go done ensuring all necessary migrations have run server server go serving sql connections rpc nodedialer nodedialer go connection to established server server update go no need to upgrade cluster already at the newest version sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup sql event log go event create database target info databasename test statement create database if not exists test user root storage replica command go initiating a split of this range at key table sql event log go event create table target info tablename test public t statement create table test public t num int user root storage replica command go initiating a split of this range at key table sql crdb internal go failed to dial into node live initial connection heartbeat failed rpc error code unavailable desc all subconns are in transientfailure latest connection error connection error desc transport error while dialing dial tcp connect connection refused please assign take a look and update the issue accordingly
| 1
|
304,090
| 26,251,701,804
|
IssuesEvent
|
2023-01-05 19:57:12
|
MetaMask/metamask-extension
|
https://api.github.com/repos/MetaMask/metamask-extension
|
closed
|
Write tests for Infura network client
|
area-testSuite type-refactor
|
As part of an effort to [merge the NetworkController that exists in this repo with the NetworkController that exists in the `controllers` repo](https://github.com/MetaMask/controllers/issues/753), we need to write tests for the JSON-RPC layer, and all of its custom behavior, that the NetworkController exposes.
We have two "network clients" that handle JSON-RPC network requests, one for Infura and one for custom networks. The custom network client tests are covered under issue #16041. This issue is for testing the network client that is specific to Infura.
|
1.0
|
Write tests for Infura network client - As part of an effort to [merge the NetworkController that exists in this repo with the NetworkController that exists in the `controllers` repo](https://github.com/MetaMask/controllers/issues/753), we need to write tests for the JSON-RPC layer, and all of its custom behavior, that the NetworkController exposes.
We have two "network clients" that handle JSON-RPC network requests, one for Infura and one for custom networks. The custom network client tests are covered under issue #16041. This issue is for testing the network client that is specific to Infura.
|
test
|
write tests for infura network client as part of an effort to we need to write tests for the json rpc layer and all of its custom behavior that the networkcontroller exposes we have two network clients that handle json rpc network requests one for infura and one for custom networks the custom network client tests are covered under issue this issue is for testing the network client that is specific to infura
| 1
|
273,022
| 23,722,526,481
|
IssuesEvent
|
2022-08-30 16:29:53
|
k3s-io/k3s
|
https://api.github.com/repos/k3s-io/k3s
|
opened
|
Add Integration Test for common/uncommon server/agent flag options
|
kind/task kind/test
|
Similar to https://github.com/k3s-io/k3s/issues/6056. This should add tests for different options that we see. We still need to define what is common vs. uncommon, but for example, a lot of users deploy k3s with traefik and svclb disabled, so that is one configuration to have in the tests here.
Feel free to add configurations as comments to this issue for tracking and to make it easier to add tests for them!
To start:
```
k3s server --disable=traefik,servicelb,metrics-server,local-storage,coredns --disable-network-policy \
--disable-agent --disable-cloud-controller --flannel-backend=none --disable-scheduler \
--kube-controller-manager-arg=controllers=*,-nodeipam,-nodelifecycle,-persistentvolume-binder,-attachdetach,-persistentvolume-expander,-cloud-node-lifecycle \
--kube-apiserver-arg=endpoint-reconciler-type=none
```
|
1.0
|
Add Integration Test for common/uncommon server/agent flag options - Similar to https://github.com/k3s-io/k3s/issues/6056. This should add tests for different options that we see. We still need to define what is common vs. uncommon, but for example, a lot of users deploy k3s with traefik and svclb disabled, so that is one configuration to have in the tests here.
Feel free to add configurations as comments to this issue for tracking and to make it easier to add tests for them!
To start:
```
k3s server --disable=traefik,servicelb,metrics-server,local-storage,coredns --disable-network-policy \
--disable-agent --disable-cloud-controller --flannel-backend=none --disable-scheduler \
--kube-controller-manager-arg=controllers=*,-nodeipam,-nodelifecycle,-persistentvolume-binder,-attachdetach,-persistentvolume-expander,-cloud-node-lifecycle \
--kube-apiserver-arg=endpoint-reconciler-type=none
```
|
test
|
add integration test for common uncommon server agent flag options similar to this should add tests for different options that we see we still need to define what is common vs uncommon but for example a lot of users deploy with traefik and svclb disabled so that is one configuration to have in the tests here feel free to add configurations as comments to this issue for tracking and to make it easier to add tests for them to start server disable traefik servicelb metrics server local storage coredns disable network policy disable agent disable cloud controller flannel backend none disable scheduler kube controller manager arg controllers nodeipam nodelifecycle persistentvolume binder attachdetach persistentvolume expander cloud node lifecycle kube apiserver arg endpoint reconciler type none
| 1
|
401,894
| 27,340,994,440
|
IssuesEvent
|
2023-02-26 19:50:48
|
chbackstrom/Assignment-5
|
https://api.github.com/repos/chbackstrom/Assignment-5
|
closed
|
Assignment 5: Completed for Review
|
documentation
|
My Assignment-5 repository includes a .Rmd file (in the "figures" folder, as instructed) including two exploratory iterations of a length vs. weight figure (by fish sex) and an expository figure enhancing the presentation according to concepts covered in lecture. Please let me know if you have any questions or issues accessing the file(s).
|
1.0
|
Assignment 5: Completed for Review - My Assignment-5 repository includes a .Rmd file (in the "figures" folder, as instructed) including two exploratory iterations of a length vs. weight figure (by fish sex) and an expository figure enhancing the presentation according to concepts covered in lecture. Please let me know if you have any questions or issues accessing the file(s).
|
non_test
|
assignment completed for review my assignment repository includes a rmd file in the figures folder as instructed including two exploratory iterations of a length vs weight figure by fish sex and an expository figure enhancing the presentation according to concepts covered in lecture please let me know if you have any questions or issues accessing the file s
| 0
|
324,519
| 9,904,974,012
|
IssuesEvent
|
2019-06-27 10:23:46
|
huridocs/uwazi
|
https://api.github.com/repos/huridocs/uwazi
|
closed
|
Menu translations get deleted
|
Bug Priority: Critical Status: Sprint
|
When editing settings > collection > custom CSS, Menu translations get deleted.
1. Add a menu to a blank Uwazi with at least 2 languages.
2. The translations are created in settings > translations > menu.
3. Go to settings > collection > custom css
4. Edit and save.
5. Go to settings > translations > menu. The translations have disappeared.
|
1.0
|
Menu translations get deleted - When editing settings > collection > custom CSS, Menu translations get deleted.
1. Add a menu to a blank Uwazi with at least 2 languages.
2. The translations are created in settings > translations > menu.
3. Go to settings > collection > custom css
4. Edit and save.
5. Go to settings > translations > menu. The translations have disappeared.
|
non_test
|
menu translations get deleted when editing settings collection custom css menu translations get deleted add a menu to a blank uwazi with at least languages the translations are created in settings translations menu go to settings collection custom css edit and save go to settings translations menu the translations have disappeared
| 0
|
38,845
| 5,201,415,356
|
IssuesEvent
|
2017-01-24 04:36:48
|
ElucidataInc/ElMaven
|
https://api.github.com/repos/ElucidataInc/ElMaven
|
closed
|
allow .txt extension for tab-delimited files
|
enhancement fixed medium tested
|
Maven always wants .tab for tab-delimited input files (like compound database) which is just mildly annoying since most people use Excel to edit the file, and Excel only saves as .txt . Let's just allow .txt extension and assume it's tab-delimited.
(maybe at some point in the future we can do what spotfire and excel do and guess the format/delimiter, but this is extremely low priority).
|
1.0
|
allow .txt extension for tab-delimited files - Maven always wants .tab for tab-delimited input files (like compound database) which is just mildly annoying since most people use Excel to edit the file, and Excel only saves as .txt . Let's just allow .txt extension and assume it's tab-delimited.
(maybe at some point in the future we can do what spotfire and excel do and guess the format/delimiter, but this is extremely low priority).
|
test
|
allow txt extension for tab delimited files maven always wants tab for tab delimited input files like compound database which is just mildly annoying since most people use excel to edit the file and excel only saves as txt let s just allow txt extension and assume it s tab delimited maybe at some point in the future we can do what spotfire and excel do and guess the format delimiter but this is extremely low priority
| 1
|
91,412
| 8,304,750,049
|
IssuesEvent
|
2018-09-21 22:39:40
|
excellaco/play-slick-postgres.g8
|
https://api.github.com/repos/excellaco/play-slick-postgres.g8
|
opened
|
Add Evolutions Tests
|
nice-to-have test-coverage
|
Though we can't see it in Statement Coverage report, it would be nice to include a test for applyEvolutions using H2. This can be helpful to warn of an evolution that will break when Play applies it (except of course in cases of merge conflicts).
|
1.0
|
Add Evolutions Tests - Though we can't see it in Statement Coverage report, it would be nice to include a test for applyEvolutions using H2. This can be helpful to warn of an evolution that will break when Play applies it (except of course in cases of merge conflicts).
|
test
|
add evolutions tests though we can t see it in statement coverage report it would be nice to include a test for applyevolutions using this can be helpful to warn of an evolution that will break when play applies it except of course in cases of merge conflicts
| 1
|
269,931
| 8,444,584,923
|
IssuesEvent
|
2018-10-18 18:52:27
|
ampproject/amphtml
|
https://api.github.com/repos/ampproject/amphtml
|
closed
|
amp-youtube autoplay does not work unless initially in viewport
|
Category: Audio&Video P1: High Priority Type: Bug
|
not a regression of our code but it used to work before a month ago or so.
Go to https://ampbyexample.com/components/amp-youtube/
Scroll down
See the autoplaying embed has started playing but is paused.
something with youtube embed may have changed.
/cc @kevinkassimo looks similar to the issue you found where messages where not received if user scrolls to amp-youtube.
|
1.0
|
amp-youtube autoplay does not work unless initially in viewport - not a regression of our code but it used to work before a month ago or so.
Go to https://ampbyexample.com/components/amp-youtube/
Scroll down
See the autoplaying embed has started playing but is paused.
something with youtube embed may have changed.
/cc @kevinkassimo looks similar to the issue you found where messages where not received if user scrolls to amp-youtube.
|
non_test
|
amp youtube autoplay does not work unless initially in viewport not a regression of our code but it used to work before a month ago or so go to scroll down see the autoplaying embed has started playing but is paused something with youtube embed may have changed cc kevinkassimo looks similar to the issue you found where messages where not received if user scrolls to amp youtube
| 0
|
471,387
| 13,565,846,606
|
IssuesEvent
|
2020-09-18 12:23:50
|
bc-ticketing/host-client
|
https://api.github.com/repos/bc-ticketing/host-client
|
opened
|
Export and Import Seating Plan as JSON
|
event feature frontend medium priority
|
A host should be able to export and import a json of a seating plan. To avoid a host creating the exact same seating plan every time a new event is planned, the host should be able to export the seating plan in a json. The next time an event is created, the host can just import this json file again, set the prices etc. for each type and then create the types without creating the whole plan from scratch.
|
1.0
|
Export and Import Seating Plan as JSON - A host should be able to export and import a json of a seating plan. To avoid a host creating the exact same seating plan every time a new event is planned, the host should be able to export the seating plan in a json. The next time an event is created, the host can just import this json file again, set the prices etc. for each type and then create the types without creating the whole plan from scratch.
|
non_test
|
export and import seating plan as json a host should be able to export and import a json of a seating plan to avoid a host creating the exact same seating plan every time a new event is planned the host should be able to export the seating plan in a json the next time an event is created the host can just import this json file again set the prices etc for each type and then create the types without creating the whole plan from scratch
| 0
|
146,192
| 13,173,470,756
|
IssuesEvent
|
2020-08-11 20:22:59
|
leomaurodesenv/data-science-api-framework
|
https://api.github.com/repos/leomaurodesenv/data-science-api-framework
|
closed
|
swagger integration
|
documentation enhancement
|
Automatize the [Swagger](https://swagger.io/) documentation to generate a simple documentation based on the flask.
Additional documentation:
- https://pypi.org/project/flask-swagger/
- https://flask-restplus.readthedocs.io/en/stable/swagger.html
- http://michal.karzynski.pl/blog/2016/06/19/building-beautiful-restful-apis-using-flask-swagger-ui-flask-restplus/
|
1.0
|
swagger integration - Automatize the [Swagger](https://swagger.io/) documentation to generate a simple documentation based on the flask.
Additional documentation:
- https://pypi.org/project/flask-swagger/
- https://flask-restplus.readthedocs.io/en/stable/swagger.html
- http://michal.karzynski.pl/blog/2016/06/19/building-beautiful-restful-apis-using-flask-swagger-ui-flask-restplus/
|
non_test
|
swagger integration automatize the documentation to generate a simple documentation based on the flask additional documentation
| 0
|
8,694
| 7,570,597,083
|
IssuesEvent
|
2018-04-23 09:29:33
|
matthiasbeyer/imag
|
https://api.github.com/repos/matthiasbeyer/imag
|
opened
|
Idea: libimagentrydatapoint
|
complexity/high kind/enhancement kind/infrastructure meta/importance/medium
|
One idea I had: A library for writing data points to entries. The library would not only handle unit conversions (whereas the unit is a thing which can be defined by the user of the library) but also aggregation, statistics and possibly even anomaly detection.
---
This library can then be used for other tools, for example for a health tracker, a crypto currency price tracker etc etc
---
Details have to be worked out of course.
|
1.0
|
Idea: libimagentrydatapoint - One idea I had: A library for writing data points to entries. The library would not only handle unit conversions (whereas the unit is a thing which can be defined by the user of the library) but also aggregation, statistics and possibly even anomaly detection.
---
This library can then be used for other tools, for example for a health tracker, a crypto currency price tracker etc etc
---
Details have to be worked out of course.
|
non_test
|
idea libimagentrydatapoint one idea i had a library for writing data points to entries the library would not only handle unit conversions whereas the unit is a thing which can be defined by the user of the library but also aggregation statistics and possibly even anomaly detection this library can then be used for other tools for example for a health tracker a crypto currency price tracker etc etc details have to be worked out of course
| 0
|
201,838
| 15,226,848,239
|
IssuesEvent
|
2021-02-18 09:26:12
|
WeiXian042901/fyp_repository
|
https://api.github.com/repos/WeiXian042901/fyp_repository
|
opened
|
FU_041_Quiz Play Page Two Options(Display Only Two Options)
|
Acceptance Test Quiz User
|
**Test Scenario**
- User has chosen to attempt the Testing Title(Two Options) quiz
**Test Case**
- Check that the system displays only two available options for the users to choose from
**Pre-Conditions**
User has successfully entered the application
User clicked on the “Quizzes” Option
User selected the “Testing title(two options)” quiz option
User clicked on the “Start Quiz” button.
**Test-Steps**
**Test Data**
**Expected Results**
- The system should only displays two available options for the users to chooses from
**Actual Results**
- The system only displays two available options for the users to chooses from
**Pass/Fail**
- Pass
**Date Tested**
- 10th February 2021
**Tested By**
- Zachary Tan
|
1.0
|
FU_041_Quiz Play Page Two Options(Display Only Two Options) - **Test Scenario**
- User has chosen to attempt the Testing Title(Two Options) quiz
**Test Case**
- Check that the system displays only two available options for the users to choose from
**Pre-Conditions**
User has successfully entered the application
User clicked on the “Quizzes” Option
User selected the “Testing title(two options)” quiz option
User clicked on the “Start Quiz” button.
**Test-Steps**
**Test Data**
**Expected Results**
- The system should only displays two available options for the users to chooses from
**Actual Results**
- The system only displays two available options for the users to chooses from
**Pass/Fail**
- Pass
**Date Tested**
- 10th February 2021
**Tested By**
- Zachary Tan
|
test
|
fu quiz play page two options display only two options test scenario user has chosen to attempt the testing title two options quiz test case check that the system displays only two available options for the users to choose from pre conditions user has successfully entered the application user clicked on the “quizzes” option user selected the “testing title two options ” quiz option user clicked on the “start quiz” button test steps test data expected results the system should only displays two available options for the users to chooses from actual results the system only displays two available options for the users to chooses from pass fail pass date tested february tested by zachary tan
| 1
|
284,223
| 24,584,201,248
|
IssuesEvent
|
2022-10-13 18:10:09
|
lightningnetwork/lnd
|
https://api.github.com/repos/lightningnetwork/lnd
|
closed
|
tests: race in TestBlockCacheMutexes tests
|
tests data-race
|
Saw a test run fail w/ a race, looks like it's in the mock backend itself and not the actual code:
```
==================
WARNING: DATA RACE
Read at 0x00c0001ca2d8 by goroutine 108:
github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:33 +0xda
github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock-fm()
<autogenerated>:1 +0x44
github.com/lightningnetwork/lnd/blockcache.(*BlockCache).GetBlock()
/home/runner/work/lnd/lnd/blockcache/blockcache.go:52 +0x3b6
github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func1()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:1[77](https://github.com/lightningnetwork/lnd/actions/runs/3011169890/jobs/4838020176#step:6:78) +0xb0
github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func2()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:182 +0x47
Previous write at 0x00c0001ca2d8 by goroutine 51:
github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:33 +0xf8
github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock-fm()
<autogenerated>:1 +0x44
github.com/lightningnetwork/lnd/blockcache.(*BlockCache).GetBlock()
/home/runner/work/lnd/lnd/blockcache/blockcache.go:52 +0x3b6
github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func1()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:174 +0x7b
github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func2()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:1[82](https://github.com/lightningnetwork/lnd/actions/runs/3011169890/jobs/4838020176#step:6:83) +0x47
Goroutine 108 (running) created at:
github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:172 +0x824
testing.tRunner()
/opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:1446 +0x216
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:14[93](https://github.com/lightningnetwork/lnd/actions/runs/3011169890/jobs/4838020176#step:6:94) +0x47
Goroutine 51 (finished) created at:
github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:172 +0x824
testing.tRunner()
/opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:1446 +0x216
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:1493 +0x47
==================
--- FAIL: TestBlockCacheMutexes (0.01s)
```
|
1.0
|
tests: race in TestBlockCacheMutexes tests - Saw a test run fail w/ a race, looks like it's in the mock backend itself and not the actual code:
```
==================
WARNING: DATA RACE
Read at 0x00c0001ca2d8 by goroutine 108:
github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:33 +0xda
github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock-fm()
<autogenerated>:1 +0x44
github.com/lightningnetwork/lnd/blockcache.(*BlockCache).GetBlock()
/home/runner/work/lnd/lnd/blockcache/blockcache.go:52 +0x3b6
github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func1()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:1[77](https://github.com/lightningnetwork/lnd/actions/runs/3011169890/jobs/4838020176#step:6:78) +0xb0
github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func2()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:182 +0x47
Previous write at 0x00c0001ca2d8 by goroutine 51:
github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:33 +0xf8
github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock-fm()
<autogenerated>:1 +0x44
github.com/lightningnetwork/lnd/blockcache.(*BlockCache).GetBlock()
/home/runner/work/lnd/lnd/blockcache/blockcache.go:52 +0x3b6
github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func1()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:174 +0x7b
github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func2()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:1[82](https://github.com/lightningnetwork/lnd/actions/runs/3011169890/jobs/4838020176#step:6:83) +0x47
Goroutine 108 (running) created at:
github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:172 +0x824
testing.tRunner()
/opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:1446 +0x216
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:14[93](https://github.com/lightningnetwork/lnd/actions/runs/3011169890/jobs/4838020176#step:6:94) +0x47
Goroutine 51 (finished) created at:
github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes()
/home/runner/work/lnd/lnd/blockcache/blockcache_test.go:172 +0x824
testing.tRunner()
/opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:1446 +0x216
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:1493 +0x47
==================
--- FAIL: TestBlockCacheMutexes (0.01s)
```
|
test
|
tests race in testblockcachemutexes tests saw a test run fail w a race looks like it s in the mock backend itself and not the actual code warning data race read at by goroutine github com lightningnetwork lnd blockcache mockchainbackend getblock home runner work lnd lnd blockcache blockcache test go github com lightningnetwork lnd blockcache mockchainbackend getblock fm github com lightningnetwork lnd blockcache blockcache getblock home runner work lnd lnd blockcache blockcache go github com lightningnetwork lnd blockcache testblockcachemutexes home runner work lnd lnd blockcache blockcache test go github com lightningnetwork lnd blockcache testblockcachemutexes home runner work lnd lnd blockcache blockcache test go previous write at by goroutine github com lightningnetwork lnd blockcache mockchainbackend getblock home runner work lnd lnd blockcache blockcache test go github com lightningnetwork lnd blockcache mockchainbackend getblock fm github com lightningnetwork lnd blockcache blockcache getblock home runner work lnd lnd blockcache blockcache go github com lightningnetwork lnd blockcache testblockcachemutexes home runner work lnd lnd blockcache blockcache test go github com lightningnetwork lnd blockcache testblockcachemutexes home runner work lnd lnd blockcache blockcache test go goroutine running created at github com lightningnetwork lnd blockcache testblockcachemutexes home runner work lnd lnd blockcache blockcache test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go goroutine finished created at github com lightningnetwork lnd blockcache testblockcachemutexes home runner work lnd lnd blockcache blockcache test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go fail testblockcachemutexes
| 1
|
302,759
| 26,160,950,366
|
IssuesEvent
|
2022-12-31 14:16:55
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: tlp failed
|
C-test-failure O-robot O-roachtest release-blocker branch-release-22.2
|
roachtest.tlp [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=artifacts#/tlp) on release-22.2 @ [07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1](https://github.com/cockroachdb/cockroach/commits/07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1):
```
test artifacts and logs in: /artifacts/tlp/run_1
(test_impl.go:286).Fatal: pq: Use of partitions requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out.
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #94551 roachtest: tlp failed [C-test-failure O-roachtest O-robot T-sql-queries branch-master release-blocker]
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*tlp.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
2.0
|
roachtest: tlp failed - roachtest.tlp [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=artifacts#/tlp) on release-22.2 @ [07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1](https://github.com/cockroachdb/cockroach/commits/07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1):
```
test artifacts and logs in: /artifacts/tlp/run_1
(test_impl.go:286).Fatal: pq: Use of partitions requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out.
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #94551 roachtest: tlp failed [C-test-failure O-roachtest O-robot T-sql-queries branch-master release-blocker]
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*tlp.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
roachtest tlp failed roachtest tlp with on release test artifacts and logs in artifacts tlp run test impl go fatal pq use of partitions requires an enterprise license your evaluation license expired on december if you re interested in getting a new license please contact subscriptions cockroachlabs com and we can help you out parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see same failure on other branches roachtest tlp failed cc cockroachdb sql queries
| 1
|
543,523
| 15,883,376,446
|
IssuesEvent
|
2021-04-09 17:18:06
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
Ask password option: Provide explanation and image of management console config
|
Affected/5.4.0-Update1 Priority/Highest docs
|
In [1] the management console config section does not contain an image where the expected configs are displayed. Please provide it. Furthermore in management console under User on boarding section there are other configs as well. Please explain the usage of them and mention if they are not relevant to this particular case,
[1] https://docs.wso2.com/display/IS540/Creating+Users+using+the+Ask+Password+Option
|
1.0
|
Ask password option: Provide explanation and image of management console config - In [1] the management console config section does not contain an image where the expected configs are displayed. Please provide it. Furthermore in management console under User on boarding section there are other configs as well. Please explain the usage of them and mention if they are not relevant to this particular case,
[1] https://docs.wso2.com/display/IS540/Creating+Users+using+the+Ask+Password+Option
|
non_test
|
ask password option provide explanation and image of management console config in the management console config section does not contain an image where the expected configs are displayed please provide it furthermore in management console under user on boarding section there are other configs as well please explain the usage of them and mention if they are not relevant to this particular case
| 0
|
111,877
| 17,038,617,138
|
IssuesEvent
|
2021-07-05 10:23:45
|
tamirdahan/NodeTestBench
|
https://api.github.com/repos/tamirdahan/NodeTestBench
|
opened
|
CVE-2021-23343 (High) detected in path-parse-1.0.6.tgz
|
security vulnerability
|
## CVE-2021-23343 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>path-parse-1.0.6.tgz</b></p></summary>
<p>Node.js path.parse() ponyfill</p>
<p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz</a></p>
<p>Path to dependency file: NodeTestBench/package.json</p>
<p>Path to vulnerable library: NodeTestBench/node_modules/path-parse/package.json</p>
<p>
Dependency Hierarchy:
- eslint-config-1.0.3.tgz (Root Library)
- eslint-plugin-node-9.1.0.tgz
- resolve-1.12.0.tgz
- :x: **path-parse-1.0.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tamirdahan/NodeTestBench/commit/20c4644edbd7cd6f64e400c703d1594170fe90df">20c4644edbd7cd6f64e400c703d1594170fe90df</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jbgutierrez/path-parse/issues/8">https://github.com/jbgutierrez/path-parse/issues/8</a></p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution: path-parse - 1.0.7</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"path-parse","packageVersion":"1.0.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"@contrast/eslint-config:1.0.3;eslint-plugin-node:9.1.0;resolve:1.12.0;path-parse:1.0.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"path-parse - 1.0.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23343","vulnerabilityDetails":"All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-23343 (High) detected in path-parse-1.0.6.tgz - ## CVE-2021-23343 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>path-parse-1.0.6.tgz</b></p></summary>
<p>Node.js path.parse() ponyfill</p>
<p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz</a></p>
<p>Path to dependency file: NodeTestBench/package.json</p>
<p>Path to vulnerable library: NodeTestBench/node_modules/path-parse/package.json</p>
<p>
Dependency Hierarchy:
- eslint-config-1.0.3.tgz (Root Library)
- eslint-plugin-node-9.1.0.tgz
- resolve-1.12.0.tgz
- :x: **path-parse-1.0.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tamirdahan/NodeTestBench/commit/20c4644edbd7cd6f64e400c703d1594170fe90df">20c4644edbd7cd6f64e400c703d1594170fe90df</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jbgutierrez/path-parse/issues/8">https://github.com/jbgutierrez/path-parse/issues/8</a></p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution: path-parse - 1.0.7</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"path-parse","packageVersion":"1.0.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"@contrast/eslint-config:1.0.3;eslint-plugin-node:9.1.0;resolve:1.12.0;path-parse:1.0.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"path-parse - 1.0.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23343","vulnerabilityDetails":"All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve high detected in path parse tgz cve high severity vulnerability vulnerable library path parse tgz node js path parse ponyfill library home page a href path to dependency file nodetestbench package json path to vulnerable library nodetestbench node modules path parse package json dependency hierarchy eslint config tgz root library eslint plugin node tgz resolve tgz x path parse tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of package path parse are vulnerable to regular expression denial of service redos via splitdevicere splittailre and splitpathre regular expressions redos exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution path parse isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree contrast eslint config eslint plugin node resolve path parse isminimumfixversionavailable true minimumfixversion path parse basebranches vulnerabilityidentifier cve vulnerabilitydetails all versions of package path parse are vulnerable to regular expression denial of service redos via splitdevicere splittailre and splitpathre regular expressions redos exhibits polynomial worst case time complexity vulnerabilityurl
| 0
|
167,122
| 14,102,269,171
|
IssuesEvent
|
2020-11-06 08:28:45
|
lensapp/lens
|
https://api.github.com/repos/lensapp/lens
|
opened
|
Documentation for UI components
|
area/documentation area/ui
|
We need to document inner Lens UI components using one of the utilities:
* Storybook (https://storybook.js.org/)
* Docz (https://www.docz.site/)
* Blueprint (https://blueprintjs.com/)
Each documented component should have at least:
* Brief description (if needed)
* Its visual representation
* List of props
|
1.0
|
Documentation for UI components - We need to document inner Lens UI components using one of the utilities:
* Storybook (https://storybook.js.org/)
* Docz (https://www.docz.site/)
* Blueprint (https://blueprintjs.com/)
Each documented component should have at least:
* Brief description (if needed)
* Its visual representation
* List of props
|
non_test
|
documentation for ui components we need to document inner lens ui components using one of the utilities storybook docz blueprint each documented component should have at least brief description if needed its visual representation list of props
| 0
|
90,032
| 25,953,823,886
|
IssuesEvent
|
2022-12-18 00:14:31
|
QubesOS/updates-status
|
https://api.github.com/repos/QubesOS/updates-status
|
closed
|
desktop-linux-manager v4.2.0 (r4.2)
|
r4.2-vm-bookworm-building r4.2-host-cur-test r4.2-vm-bullseye-cur-test r4.2-vm-fc37-cur-test r4.2-vm-fc36-cur-test r4.2-vm-centos-stream8-cur-test
|
Update of desktop-linux-manager to v4.2.0 for Qubes r4.2, see comments below for details and build status.
From commit: https://github.com/QubesOS/qubes-desktop-linux-manager/commit/f5f5a00a41217dd2c036843fd669968ee3e2248c
[Changes since previous version](https://github.com/QubesOS/qubes-desktop-linux-manager/compare/v4.1.15...v4.2.0):
QubesOS/qubes-desktop-linux-manager@f5f5a00 version 4.2.0
QubesOS/qubes-desktop-linux-manager@e116e35 Add detecting changes in raw policy text
QubesOS/qubes-desktop-linux-manager@24b20bf Fix updating initial value in text dropdowns
QubesOS/qubes-desktop-linux-manager@16badf1 Improve formatting
QubesOS/qubes-desktop-linux-manager@68aa035 Remove a superfluous variable
QubesOS/qubes-desktop-linux-manager@1bdc940 Refactor raw policy handler as a separate handler
QubesOS/qubes-desktop-linux-manager@a6d431f Improve error handling in policy config
QubesOS/qubes-desktop-linux-manager@e85099f Change Open File/URL policy to more reasonable version
QubesOS/qubes-desktop-linux-manager@5c9bd66 Refactor ExceptionsList to a separate files
QubesOS/qubes-desktop-linux-manager@8a4ec98 Improve AdminVM approach: remove dom0 from most policy configs
QubesOS/qubes-desktop-linux-manager@e4dba08 Use --get-default-whitelist qubes-appmenus option for Create New Qube
QubesOS/qubes-desktop-linux-manager@6f7ddab Add 'copy system info to global clipboard' button
QubesOS/qubes-desktop-linux-manager@bde5175 Show errors on loading policy rules
QubesOS/qubes-desktop-linux-manager@ea1a5e3 Add gitlab CI integration
QubesOS/qubes-desktop-linux-manager@8abee61 New Config Tool and new New Qube tool
Referenced issues:
If you're release manager, you can issue GPG-inline signed command:
* `Upload-component r4.2 desktop-linux-manager f5f5a00a41217dd2c036843fd669968ee3e2248c current all` (available 5 days from now)
* `Upload-component r4.2 desktop-linux-manager f5f5a00a41217dd2c036843fd669968ee3e2248c security-testing`
You can choose subset of distributions like:
* `Upload-component r4.2 desktop-linux-manager f5f5a00a41217dd2c036843fd669968ee3e2248c current vm-bookworm,vm-fc37` (available 5 days from now)
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
|
1.0
|
desktop-linux-manager v4.2.0 (r4.2) - Update of desktop-linux-manager to v4.2.0 for Qubes r4.2, see comments below for details and build status.
From commit: https://github.com/QubesOS/qubes-desktop-linux-manager/commit/f5f5a00a41217dd2c036843fd669968ee3e2248c
[Changes since previous version](https://github.com/QubesOS/qubes-desktop-linux-manager/compare/v4.1.15...v4.2.0):
QubesOS/qubes-desktop-linux-manager@f5f5a00 version 4.2.0
QubesOS/qubes-desktop-linux-manager@e116e35 Add detecting changes in raw policy text
QubesOS/qubes-desktop-linux-manager@24b20bf Fix updating initial value in text dropdowns
QubesOS/qubes-desktop-linux-manager@16badf1 Improve formatting
QubesOS/qubes-desktop-linux-manager@68aa035 Remove a superfluous variable
QubesOS/qubes-desktop-linux-manager@1bdc940 Refactor raw policy handler as a separate handler
QubesOS/qubes-desktop-linux-manager@a6d431f Improve error handling in policy config
QubesOS/qubes-desktop-linux-manager@e85099f Change Open File/URL policy to more reasonable version
QubesOS/qubes-desktop-linux-manager@5c9bd66 Refactor ExceptionsList to a separate files
QubesOS/qubes-desktop-linux-manager@8a4ec98 Improve AdminVM approach: remove dom0 from most policy configs
QubesOS/qubes-desktop-linux-manager@e4dba08 Use --get-default-whitelist qubes-appmenus option for Create New Qube
QubesOS/qubes-desktop-linux-manager@6f7ddab Add 'copy system info to global clipboard' button
QubesOS/qubes-desktop-linux-manager@bde5175 Show errors on loading policy rules
QubesOS/qubes-desktop-linux-manager@ea1a5e3 Add gitlab CI integration
QubesOS/qubes-desktop-linux-manager@8abee61 New Config Tool and new New Qube tool
Referenced issues:
If you're release manager, you can issue GPG-inline signed command:
* `Upload-component r4.2 desktop-linux-manager f5f5a00a41217dd2c036843fd669968ee3e2248c current all` (available 5 days from now)
* `Upload-component r4.2 desktop-linux-manager f5f5a00a41217dd2c036843fd669968ee3e2248c security-testing`
You can choose subset of distributions like:
* `Upload-component r4.2 desktop-linux-manager f5f5a00a41217dd2c036843fd669968ee3e2248c current vm-bookworm,vm-fc37` (available 5 days from now)
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
|
non_test
|
desktop linux manager update of desktop linux manager to for qubes see comments below for details and build status from commit qubesos qubes desktop linux manager version qubesos qubes desktop linux manager add detecting changes in raw policy text qubesos qubes desktop linux manager fix updating initial value in text dropdowns qubesos qubes desktop linux manager improve formatting qubesos qubes desktop linux manager remove a superfluous variable qubesos qubes desktop linux manager refactor raw policy handler as a separate handler qubesos qubes desktop linux manager improve error handling in policy config qubesos qubes desktop linux manager change open file url policy to more reasonable version qubesos qubes desktop linux manager refactor exceptionslist to a separate files qubesos qubes desktop linux manager improve adminvm approach remove from most policy configs qubesos qubes desktop linux manager use get default whitelist qubes appmenus option for create new qube qubesos qubes desktop linux manager add copy system info to global clipboard button qubesos qubes desktop linux manager show errors on loading policy rules qubesos qubes desktop linux manager add gitlab ci integration qubesos qubes desktop linux manager new config tool and new new qube tool referenced issues if you re release manager you can issue gpg inline signed command upload component desktop linux manager current all available days from now upload component desktop linux manager security testing you can choose subset of distributions like upload component desktop linux manager current vm bookworm vm available days from now above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it for more information on how to test this update please take a look at
| 0
|
83,277
| 7,867,928,542
|
IssuesEvent
|
2018-06-23 15:06:27
|
jbeard4/SCION
|
https://api.github.com/repos/jbeard4/SCION
|
closed
|
fail test/scxml-test-framework/test/w3c-ecma/test187.txml.scxml
|
2.0.0 Node v0.10.24 Tests fail feature:invoke
|
[https://github.com/jbeard4/scxml-test-framework/tree/master/test/scxml-test-framework/test/w3c-ecma/test187.txml.scxml](https://github.com/jbeard4/scxml-test-framework/tree/master/test/scxml-test-framework/test/w3c-ecma/test187.txml.scxml)
**Error** <code><pre>{
"name": "AssertionError",
"actual": [
"subFinal"
],
"expected": [
"pass"
],
"operator": "deepEqual",
"message": "[\"subFinal\"] deepEqual [\"pass\"]"
}</code></pre>
**Data**: <code><pre>{
"sessionToken": 551,
"nextConfiguration": [
"subFinal"
]
}</code></pre>
**scxml**:
``` xml
<?xml version="1.0" encoding="UTF-8"?>
<!-- we test that delayed <send> is not sent if the sending session terminates. In this case,
a subscript is invoked which sends the event childToParent delayed by 1 second, and then terminates. The
parent session, should not receive childToParent. If it does, we fail. Otherwise the
10 sec timer expires and we pass -->
<scxml xmlns="http://www.w3.org/2005/07/scxml" xmlns:conf="http://www.w3.org/2005/scxml-conformance" initial="s0" version="1.0" datamodel="ecmascript">
<state id="s0">
<onentry>
<send event="timeout" delay="10s"/>
</onentry>
<invoke type="scxml">
<content>
<!-- exit before the delayed send can execute -->
<scxml initial="sub0" version="1.0" datamodel="ecmascript">
<state id="sub0">
<onentry>
<send event="childToParent" target="#_parent" delay="1s"/>
</onentry>
<transition target="subFinal"/>
</state>
<final id="subFinal"/>
</scxml>
</content>
</invoke>
<transition event="childToParent" target="fail"/>
<transition event="timeout" target="pass"/>
</state>
<final id="pass"><onentry><log label="Outcome" expr="'pass'"/></onentry></final>
<final id="fail"><onentry><log label="Outcome" expr="'fail'"/></onentry></final>
</scxml>
```
**JSON**: <code><pre>{
"initialConfiguration": [
"pass"
],
"events": []
}</code></pre>
|
1.0
|
fail test/scxml-test-framework/test/w3c-ecma/test187.txml.scxml - [https://github.com/jbeard4/scxml-test-framework/tree/master/test/scxml-test-framework/test/w3c-ecma/test187.txml.scxml](https://github.com/jbeard4/scxml-test-framework/tree/master/test/scxml-test-framework/test/w3c-ecma/test187.txml.scxml)
**Error** <code><pre>{
"name": "AssertionError",
"actual": [
"subFinal"
],
"expected": [
"pass"
],
"operator": "deepEqual",
"message": "[\"subFinal\"] deepEqual [\"pass\"]"
}</code></pre>
**Data**: <code><pre>{
"sessionToken": 551,
"nextConfiguration": [
"subFinal"
]
}</code></pre>
**scxml**:
``` xml
<?xml version="1.0" encoding="UTF-8"?>
<!-- we test that delayed <send> is not sent if the sending session terminates. In this case,
a subscript is invoked which sends the event childToParent delayed by 1 second, and then terminates. The
parent session, should not receive childToParent. If it does, we fail. Otherwise the
10 sec timer expires and we pass -->
<scxml xmlns="http://www.w3.org/2005/07/scxml" xmlns:conf="http://www.w3.org/2005/scxml-conformance" initial="s0" version="1.0" datamodel="ecmascript">
<state id="s0">
<onentry>
<send event="timeout" delay="10s"/>
</onentry>
<invoke type="scxml">
<content>
<!-- exit before the delayed send can execute -->
<scxml initial="sub0" version="1.0" datamodel="ecmascript">
<state id="sub0">
<onentry>
<send event="childToParent" target="#_parent" delay="1s"/>
</onentry>
<transition target="subFinal"/>
</state>
<final id="subFinal"/>
</scxml>
</content>
</invoke>
<transition event="childToParent" target="fail"/>
<transition event="timeout" target="pass"/>
</state>
<final id="pass"><onentry><log label="Outcome" expr="'pass'"/></onentry></final>
<final id="fail"><onentry><log label="Outcome" expr="'fail'"/></onentry></final>
</scxml>
```
**JSON**: <code><pre>{
"initialConfiguration": [
"pass"
],
"events": []
}</code></pre>
|
test
|
fail test scxml test framework test ecma txml scxml error name assertionerror actual subfinal expected pass operator deepequal message deepequal data sessiontoken nextconfiguration subfinal scxml xml is not sent if the sending session terminates in this case a subscript is invoked which sends the event childtoparent delayed by second and then terminates the parent session should not receive childtoparent if it does we fail otherwise the sec timer expires and we pass json initialconfiguration pass events
| 1
|
4,240
| 6,493,246,173
|
IssuesEvent
|
2017-08-21 16:14:00
|
Microsoft/UWPCommunityToolkit
|
https://api.github.com/repos/Microsoft/UWPCommunityToolkit
|
closed
|
Resolve service dependencies on obsolete HttpHelper class
|
help wanted PR in progress services
|
HttpHelper will become obsolete in v2.0 of the toolkit. We need to remove service dependencies from HttpHelper to HttpClient to prevent a large number of warnings upon building.
|
1.0
|
Resolve service dependencies on obsolete HttpHelper class - HttpHelper will become obsolete in v2.0 of the toolkit. We need to remove service dependencies from HttpHelper to HttpClient to prevent a large number of warnings upon building.
|
non_test
|
resolve service dependencies on obsolete httphelper class httphelper will become obsolete in of the toolkit we need to remove service dependencies from httphelper to httpclient to prevent a large number of warnings upon building
| 0
|
381,297
| 26,445,626,310
|
IssuesEvent
|
2023-01-16 06:58:22
|
GameCache/CodeTemplate
|
https://api.github.com/repos/GameCache/CodeTemplate
|
opened
|
Add Initial Structure
|
documentation enhancement
|
<!--- Please search issues before creating a new one. --->
<!--- This is just a template, feel free to customize if prudent. --->
## Wanted Behavior
<!--- Describe the wanted feature. --->
Ability to use this repository as a seed for others.
## Possible Solution
<!--- Optional implementation suggestion. --->
Add a default implementation with the common structure to use for other projects.
## Additional Information
<!--- For example, why do you want the feature. --->
Also provides a way to test new language enhancements for already created projects.
|
1.0
|
Add Initial Structure - <!--- Please search issues before creating a new one. --->
<!--- This is just a template, feel free to customize if prudent. --->
## Wanted Behavior
<!--- Describe the wanted feature. --->
Ability to use this repository as a seed for others.
## Possible Solution
<!--- Optional implementation suggestion. --->
Add a default implementation with the common structure to use for other projects.
## Additional Information
<!--- For example, why do you want the feature. --->
Also provides a way to test new language enhancements for already created projects.
|
non_test
|
add initial structure wanted behavior ability to use this repository as a seed for others possible solution add a default implementation with the common structure to use for other projects additional information also provides a way to test new language enhancements for already created projects
| 0
|
47,945
| 19,842,709,587
|
IssuesEvent
|
2022-01-21 00:17:19
|
Azure/azure-sdk-for-python
|
https://api.github.com/repos/Azure/azure-sdk-for-python
|
opened
|
cognitiveservices/azure-cognitiveservices-knowledge-qnamaker - cspell found spelling errors in public API surface
|
Cognitive Services Client
|
Spell check scanning of package at `sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker` detected spelling errors in the public API surface. This directory is opted out of PR spell checking in PR #22199 to keep PRs unblocked.
## What to do
1. Ensure Node.js is installed (https://nodejs.org/en/download/).
1. Delete the entry in `.vscode/cspell.json`'s `ignorePaths` field. It will look like: `sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/**`. You need to do this to enable checking the files.
1. From the root of the repo run spell check using `./eng/common/spelling/Invoke-Cspell.ps1 -ScanGlobs "sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/**"`
1. Fix detections according to http://aka.ms/azsdk/engsys/spellcheck use the "False positives" section to fix false positives
1. Check in changes (including the change to `.vscode/cspell.json` where the `ignorePaths` is updated to remove the entry for this service.). You may need to run `git add -f .vscode/cspell.json` to force adding the changes to the file in git.
## Spell checking output
```
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/__init__.py:12:10 - Unknown word (amaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:29:14 - Unknown word (adto)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:29:40 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:30:14 - Unknown word (adto)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:30:32 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:33:14 - Unknown word (adocuments)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:54:14 - Unknown word (asearch)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:55:14 - Unknown word (asearch)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:56:14 - Unknown word (asearch)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:57:14 - Unknown word (asearch)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:79:14 - Unknown word (adto)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:79:36 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:80:14 - Unknown word (adto)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:80:28 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:83:14 - Unknown word (adocuments)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:112:10 - Unknown word (amaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/context_dto_py3.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/context_dto.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:22:37 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:25:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:25:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:30:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:62:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:22:37 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:25:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:25:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:30:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:62:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:21:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:21:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:30:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:21:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:21:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:30:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/endpoint_settings_dto_py3.py:20:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/endpoint_settings_dto.py:20:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_py3.py:28:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_py3.py:36:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_py3.py:40:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error_py3.py:27:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error_py3.py:35:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error_py3.py:39:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error.py:27:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error.py:35:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error.py:39:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_py3.py:22:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response.py:22:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error.py:28:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error.py:36:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error.py:40:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/feedback_records_dto_py3.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/feedback_records_dto.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/inner_error_model_py3.py:26:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/inner_error_model.py:26:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/knowledgebases_dto_py3.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/knowledgebases_dto.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation_py3.py:21:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation_py3.py:36:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation.py:21:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation.py:36:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:22:8 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:24:19 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:24:44 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:26:51 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:12:10 - Unknown word (adto)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:12:28 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:15:22 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:16:10 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:16:35 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:31:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:34:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:12:10 - Unknown word (adto)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:12:24 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:15:22 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:16:10 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:16:35 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:31:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:34:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:22:8 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:24:19 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:24:44 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:26:51 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:16:18 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:20:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:24:62 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:16:18 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:20:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:24:62 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context_py3.py:15:9 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context_py3.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context_py3.py:39:17 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context.py:15:9 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context.py:39:17 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:15:9 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:31:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:34:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:34:59 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:52:50 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:57:17 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:15:9 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:31:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:34:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:34:59 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:52:50 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:57:17 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_context_py3.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_context.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_list_py3.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_list.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_py3.py:30:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_py3.py:33:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_py3.py:37:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result.py:30:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result.py:33:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result.py:37:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:33:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:40:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:45:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:48:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:33:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:40:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:45:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:48:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:20:15 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:20:47 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:22:14 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:23:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:27:13 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:31:13 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:31:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:20:15 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:20:47 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:22:14 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:23:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:27:13 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:31:13 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:31:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_context_dto_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_context_dto.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_contents_dto_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_contents_dto.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:21:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:21:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:30:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:21:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:21:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:30:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_py3.py:20:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_py3.py:23:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_py3.py:26:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_update_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_update.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto.py:20:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto.py:23:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto.py:26:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto_py3.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_context_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_context.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata_py3.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_py3.py:27:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_py3.py:30:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_py3.py:33:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto.py:27:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto.py:30:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto.py:33:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/word_alterations_dto_py3.py:22:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/word_alterations_dto.py:22:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:47:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:50:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:93:50 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:102:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:47:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:50:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:100:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:103:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:47:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:50:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:93:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:102:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:47:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:50:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:100:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:103:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:231:29 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:236:19 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:236:41 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:239:18 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:240:68 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:251:45 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/operations.py:48:52 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/operations.py:51:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:29:6 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:43:32 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:47:64 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:60:67 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/CHANGELOG.md:14:76 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/CHANGELOG.md:18:87 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/CHANGELOG.md:18:135 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/sdk_packaging.toml:2:51 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/setup.py:15:51 - Unknown word (qnamaker)
```
|
1.0
|
cognitiveservices/azure-cognitiveservices-knowledge-qnamaker - cspell found spelling errors in public API surface - Spell check scanning of package at `sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker` detected spelling errors in the public API surface. This directory is opted out of PR spell checking in PR #22199 to keep PRs unblocked.
## What to do
1. Ensure Node.js is installed (https://nodejs.org/en/download/).
1. Delete the entry in `.vscode/cspell.json`'s `ignorePaths` field. It will look like: `sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/**`. You need to do this to enable checking the files.
1. From the root of the repo run spell check using `./eng/common/spelling/Invoke-Cspell.ps1 -ScanGlobs "sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/**"`
1. Fix detections according to http://aka.ms/azsdk/engsys/spellcheck use the "False positives" section to fix false positives
1. Check in changes (including the change to `.vscode/cspell.json` where the `ignorePaths` is updated to remove the entry for this service.). You may need to run `git add -f .vscode/cspell.json` to force adding the changes to the file in git.
## Spell checking output
```
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/__init__.py:12:10 - Unknown word (amaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:29:14 - Unknown word (adto)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:29:40 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:30:14 - Unknown word (adto)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:30:32 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:33:14 - Unknown word (adocuments)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:54:14 - Unknown word (asearch)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:55:14 - Unknown word (asearch)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:56:14 - Unknown word (asearch)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:57:14 - Unknown word (asearch)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:79:14 - Unknown word (adto)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:79:36 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:80:14 - Unknown word (adto)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:80:28 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:83:14 - Unknown word (adocuments)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:112:10 - Unknown word (amaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/context_dto_py3.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/context_dto.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:22:37 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:25:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:25:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:30:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:62:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:22:37 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:25:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:25:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:30:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:62:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:21:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:21:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:30:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:21:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:21:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:30:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/endpoint_settings_dto_py3.py:20:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/endpoint_settings_dto.py:20:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_py3.py:28:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_py3.py:36:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_py3.py:40:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error_py3.py:27:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error_py3.py:35:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error_py3.py:39:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error.py:27:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error.py:35:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error.py:39:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_py3.py:22:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response.py:22:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error.py:28:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error.py:36:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error.py:40:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/feedback_records_dto_py3.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/feedback_records_dto.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/inner_error_model_py3.py:26:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/inner_error_model.py:26:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/knowledgebases_dto_py3.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/knowledgebases_dto.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation_py3.py:21:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation_py3.py:36:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation.py:21:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation.py:36:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:22:8 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:24:19 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:24:44 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:26:51 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:12:10 - Unknown word (adto)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:12:28 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:15:22 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:16:10 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:16:35 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:31:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:34:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:12:10 - Unknown word (adto)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:12:24 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:15:22 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:16:10 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:16:35 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:31:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:34:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:22:8 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:24:19 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:24:44 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:26:51 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:16:18 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:20:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:24:62 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:16:18 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:20:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:24:62 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context_py3.py:15:9 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context_py3.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context_py3.py:39:17 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context.py:15:9 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context.py:39:17 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:15:9 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:31:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:34:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:34:59 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:52:50 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:57:17 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:15:9 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:31:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:34:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:34:59 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:52:50 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:57:17 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_context_py3.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_context.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_list_py3.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_list.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_py3.py:30:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_py3.py:33:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_py3.py:37:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result.py:30:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result.py:33:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result.py:37:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:33:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:40:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:45:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:48:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:33:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:40:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:45:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:48:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:20:15 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:20:47 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:22:14 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:23:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:27:13 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:31:13 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:31:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:20:15 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:20:47 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:22:14 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:23:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:27:13 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:31:13 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:31:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_context_dto_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_context_dto.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_contents_dto_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_contents_dto.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:21:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:21:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:30:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:21:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:21:64 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:26:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:30:52 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_py3.py:20:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_py3.py:23:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_py3.py:26:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_update_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_update.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto.py:20:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto.py:23:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto.py:26:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto_py3.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_context_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_context.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata_py3.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata_py3.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata.py:20:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata.py:23:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_py3.py:27:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_py3.py:30:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_py3.py:33:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto.py:27:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto.py:30:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto.py:33:41 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/word_alterations_dto_py3.py:22:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/word_alterations_dto.py:22:46 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:47:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:50:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:93:50 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:102:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:47:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:50:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:100:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:103:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:47:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:50:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:93:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:102:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:47:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:50:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:100:45 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:103:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:231:29 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:236:19 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:236:41 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:239:18 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:240:68 - Unknown word (ADTO)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:251:45 - Unknown word (alist)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/operations.py:48:52 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/operations.py:51:75 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:29:6 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:43:32 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:47:64 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:60:67 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/CHANGELOG.md:14:76 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/CHANGELOG.md:18:87 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/CHANGELOG.md:18:135 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/sdk_packaging.toml:2:51 - Unknown word (qnamaker)
./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/setup.py:15:51 - Unknown word (qnamaker)
```
|
non_test
|
cognitiveservices azure cognitiveservices knowledge qnamaker cspell found spelling errors in public api surface spell check scanning of package at sdk cognitiveservices azure cognitiveservices knowledge qnamaker detected spelling errors in the public api surface this directory is opted out of pr spell checking in pr to keep prs unblocked what to do ensure node js is installed delete the entry in vscode cspell json s ignorepaths field it will look like sdk cognitiveservices azure cognitiveservices knowledge qnamaker you need to do this to enable checking the files from the root of the repo run spell check using eng common spelling invoke cspell scanglobs sdk cognitiveservices azure cognitiveservices knowledge qnamaker fix detections according to use the false positives section to fix false positives check in changes including the change to vscode cspell json where the ignorepaths is updated to remove the entry for this service you may need to run git add f vscode cspell json to force adding the changes to the file in git spell checking output sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker init py unknown word amaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adocuments sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word asearch sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word asearch sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word asearch sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word asearch sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adocuments sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word amaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models context dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models context dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models endpoint settings dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models endpoint settings dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models feedback records dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models feedback records dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models inner error model py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models inner error model py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models knowledgebases dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models knowledgebases dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models operation py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models operation py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models operation py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models operation py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto context py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto context py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto context py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto context py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto context py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto context py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result context py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result context py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result list py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result list py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update context dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update context dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb contents dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb contents dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto update py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto update py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update metadata dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update metadata dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update metadata dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update metadata dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto context py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto context py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto metadata py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto metadata py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto metadata py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto metadata py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models word alterations dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models word alterations dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations alterations operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations alterations operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations alterations operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations alterations operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint keys operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint keys operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint keys operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint keys operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint settings operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint settings operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint settings operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint settings operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker qn amaker client py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker qn amaker client py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker qn amaker client py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker qn amaker client py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker changelog md unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker changelog md unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker changelog md unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker sdk packaging toml unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker setup py unknown word qnamaker
| 0
|
114,738
| 11,855,220,791
|
IssuesEvent
|
2020-03-25 03:30:09
|
kotest/kotest
|
https://api.github.com/repos/kotest/kotest
|
closed
|
Example of property-based testing does not compile in most recent version (3.4.2)
|
documentation
|
In the main readme, this section of code won't compile with the latest version of the library:
```
class PropertyExample: StringSpec({
"String size" {
checkAll<String, String> { a, b ->
(a + b) should haveLength(a.length + b.length)
}
}
})
```
But the version here does work (https://github.com/kotest/kotest/blob/master/doc/property_testing.md):
```
class PropertyExample: StringSpec() {
init {
"String size" {
assertAll({ a: String, b: String ->
(a + b).length shouldBe a.length + b.length
})
}
}
}
```
|
1.0
|
Example of property-based testing does not compile in most recent version (3.4.2) - In the main readme, this section of code won't compile with the latest version of the library:
```
class PropertyExample: StringSpec({
"String size" {
checkAll<String, String> { a, b ->
(a + b) should haveLength(a.length + b.length)
}
}
})
```
But the version here does work (https://github.com/kotest/kotest/blob/master/doc/property_testing.md):
```
class PropertyExample: StringSpec() {
init {
"String size" {
assertAll({ a: String, b: String ->
(a + b).length shouldBe a.length + b.length
})
}
}
}
```
|
non_test
|
example of property based testing does not compile in most recent version in the main readme this section of code won t compile with the latest version of the library class propertyexample stringspec string size checkall a b a b should havelength a length b length but the version here does work class propertyexample stringspec init string size assertall a string b string a b length shouldbe a length b length
| 0
|
120,887
| 4,796,091,952
|
IssuesEvent
|
2016-11-01 05:34:24
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
kubectl apply doesn't work with thirdparty resource
|
area/app-config-deployment component/kubectl priority/P2 team/CSI-API Machinery SIG team/ux
|
This is to keep track of known issues of thirdparty resource.
cc @brendandburns
|
1.0
|
kubectl apply doesn't work with thirdparty resource - This is to keep track of known issues of thirdparty resource.
cc @brendandburns
|
non_test
|
kubectl apply doesn t work with thirdparty resource this is to keep track of known issues of thirdparty resource cc brendandburns
| 0
|
799,757
| 28,313,502,353
|
IssuesEvent
|
2023-04-10 17:29:14
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
Python: 3.6 EOL. Drop 3.6 support?
|
lang/Python kind/internal cleanup priority/P2
|
Python 3.6 has been EOLed. Should we drop 3.6 support?
Thanks!
|
1.0
|
Python: 3.6 EOL. Drop 3.6 support? - Python 3.6 has been EOLed. Should we drop 3.6 support?
Thanks!
|
non_test
|
python eol drop support python has been eoled should we drop support thanks
| 0
|
78,629
| 7,656,275,666
|
IssuesEvent
|
2018-05-10 15:46:30
|
italia/spid
|
https://api.github.com/repos/italia/spid
|
closed
|
Richiesta aggiornamento metadati - Comune di Carpenedolo
|
aggiornamento md test metadata
|
Buongiorno,
si richiede per conto del Comune di Carpenedolo l'aggiornamento dei metadati SPID deployati. I nuovi metadati sono disponibili all'endpoint https://sportellotelematico.comune.carpenedolo.bs.it/sites/default/files/spid-metadata/carpenedolo-signed.xml
Grazie
|
1.0
|
Richiesta aggiornamento metadati - Comune di Carpenedolo - Buongiorno,
si richiede per conto del Comune di Carpenedolo l'aggiornamento dei metadati SPID deployati. I nuovi metadati sono disponibili all'endpoint https://sportellotelematico.comune.carpenedolo.bs.it/sites/default/files/spid-metadata/carpenedolo-signed.xml
Grazie
|
test
|
richiesta aggiornamento metadati comune di carpenedolo buongiorno si richiede per conto del comune di carpenedolo l aggiornamento dei metadati spid deployati i nuovi metadati sono disponibili all endpoint grazie
| 1
|
282,808
| 24,498,024,755
|
IssuesEvent
|
2022-10-10 10:21:33
|
wazuh/wazuh-qa
|
https://api.github.com/repos/wazuh/wazuh-qa
|
opened
|
Ubuntu Linux 22.04 SCA Policy - Update and rework - checks 5 to 5.2.9
|
team/qa type/qa-testing status/not-tracked
|
| Target version | Related issue | Related PR |
| -------------- | ------------- | ----------------------------------------- |
| 4.4.0 | #3391 | https://github.com/wazuh/wazuh/pull/10487 |
| Check ID | Check Name | Implemented | Ready for review | QA review |
| -------- | --------------------------------------------------------------------------- | ----------- | ---------------- | --------- |
| 5 | Access, Authentication and Authorization | | | |
| 5.1 | Configure time-based job schedulers | | | |
| 5.1.1 | Ensure cron daemon is enabled and running (Automated) | | | |
| 5.1.2 | Ensure permissions on /etc/crontab are configured (Automated) | | | |
| 5.1.3 | Ensure permissions on /etc/cron.hourly are configured (Automated) | | | |
| 5.1.4 | Ensure permissions on /etc/cron.daily are configured (Automated) | | | |
| 5.1.5 | Ensure permissions on /etc/cron.weekly are configured (Automated) | | | |
| 5.1.6 | Ensure permissions on /etc/cron.monthly are configured (Automated) | | | |
| 5.1.7 | Ensure permissions on /etc/cron.d are configured (Automated) | | | |
| 5.1.8 | Ensure cron is restricted to authorized users (Automated) | | | |
| 5.1.9 | Ensure at is restricted to authorized users (Automated) | | | |
| 5.2 | Configure SSH Server | | | |
| 5.2.1 | Ensure permissions on /etc/ssh/sshd_config are configured (Automated) | | | |
| 5.2.2 | Ensure permissions on SSH private host key files are configured (Automated) | | | |
| 5.2.3 | Ensure permissions on SSH public host key files are configured (Automated) | | | |
| 5.2.4 | Ensure SSH access is limited (Automated) | | | |
| 5.2.5 | Ensure SSH LogLevel is appropriate (Automated) | | | |
| 5.2.6 | Ensure SSH PAM is enabled (Automated) | | | |
| 5.2.7 | Ensure SSH root login is disabled (Automated) | | | |
| 5.2.8 | Ensure SSH HostbasedAuthentication is disabled (Automated) | | | |
| 5.2.9 | Ensure SSH PermitEmptyPasswords is disabled (Automated) | | | |
|
1.0
|
Ubuntu Linux 22.04 SCA Policy - Update and rework - checks 5 to 5.2.9 - | Target version | Related issue | Related PR |
| -------------- | ------------- | ----------------------------------------- |
| 4.4.0 | #3391 | https://github.com/wazuh/wazuh/pull/10487 |
| Check ID | Check Name | Implemented | Ready for review | QA review |
| -------- | --------------------------------------------------------------------------- | ----------- | ---------------- | --------- |
| 5 | Access, Authentication and Authorization | | | |
| 5.1 | Configure time-based job schedulers | | | |
| 5.1.1 | Ensure cron daemon is enabled and running (Automated) | | | |
| 5.1.2 | Ensure permissions on /etc/crontab are configured (Automated) | | | |
| 5.1.3 | Ensure permissions on /etc/cron.hourly are configured (Automated) | | | |
| 5.1.4 | Ensure permissions on /etc/cron.daily are configured (Automated) | | | |
| 5.1.5 | Ensure permissions on /etc/cron.weekly are configured (Automated) | | | |
| 5.1.6 | Ensure permissions on /etc/cron.monthly are configured (Automated) | | | |
| 5.1.7 | Ensure permissions on /etc/cron.d are configured (Automated) | | | |
| 5.1.8 | Ensure cron is restricted to authorized users (Automated) | | | |
| 5.1.9 | Ensure at is restricted to authorized users (Automated) | | | |
| 5.2 | Configure SSH Server | | | |
| 5.2.1 | Ensure permissions on /etc/ssh/sshd_config are configured (Automated) | | | |
| 5.2.2 | Ensure permissions on SSH private host key files are configured (Automated) | | | |
| 5.2.3 | Ensure permissions on SSH public host key files are configured (Automated) | | | |
| 5.2.4 | Ensure SSH access is limited (Automated) | | | |
| 5.2.5 | Ensure SSH LogLevel is appropriate (Automated) | | | |
| 5.2.6 | Ensure SSH PAM is enabled (Automated) | | | |
| 5.2.7 | Ensure SSH root login is disabled (Automated) | | | |
| 5.2.8 | Ensure SSH HostbasedAuthentication is disabled (Automated) | | | |
| 5.2.9 | Ensure SSH PermitEmptyPasswords is disabled (Automated) | | | |
|
test
|
ubuntu linux sca policy update and rework checks to target version related issue related pr check id check name implemented ready for review qa review access authentication and authorization configure time based job schedulers ensure cron daemon is enabled and running automated ensure permissions on etc crontab are configured automated ensure permissions on etc cron hourly are configured automated ensure permissions on etc cron daily are configured automated ensure permissions on etc cron weekly are configured automated ensure permissions on etc cron monthly are configured automated ensure permissions on etc cron d are configured automated ensure cron is restricted to authorized users automated ensure at is restricted to authorized users automated configure ssh server ensure permissions on etc ssh sshd config are configured automated ensure permissions on ssh private host key files are configured automated ensure permissions on ssh public host key files are configured automated ensure ssh access is limited automated ensure ssh loglevel is appropriate automated ensure ssh pam is enabled automated ensure ssh root login is disabled automated ensure ssh hostbasedauthentication is disabled automated ensure ssh permitemptypasswords is disabled automated
| 1
|
171,685
| 13,244,139,480
|
IssuesEvent
|
2020-08-19 12:36:51
|
MarcoConsiglio/shopping-list
|
https://api.github.com/repos/MarcoConsiglio/shopping-list
|
opened
|
Segnare i prodotti già presi
|
browser-test miglioria estetica
|
Quando un prodotto ha valorizzata la quantità nel carrello, allora segna il prodotto in modo tale da far capire che è stato già preso.
|
1.0
|
Segnare i prodotti già presi - Quando un prodotto ha valorizzata la quantità nel carrello, allora segna il prodotto in modo tale da far capire che è stato già preso.
|
test
|
segnare i prodotti già presi quando un prodotto ha valorizzata la quantità nel carrello allora segna il prodotto in modo tale da far capire che è stato già preso
| 1
|
302,369
| 26,140,936,855
|
IssuesEvent
|
2022-12-29 18:20:05
|
Hamlib/Hamlib
|
https://api.github.com/repos/Hamlib/Hamlib
|
closed
|
QRP-Labs QDX IF command parsing buggy
|
bug needs test
|
I'm using a QRP-Labs QDX which emulates TS-440. When requesting PTT state from rigctl (sending command 't') it returns 0 (RX) even when still transmitting.
It appears that the 440 protocol returns accurate PTT state from 'P8' of the 'IF' command as per the spec. I'm going to suggest that hamlib reads PTT state from this, rather than return 0 as it does now.
Discussion here: https://groups.io/g/QRPLabs/topic/95922381?p=Created,,,20,1,0,0::recentpostdate/sticky,,,20,2,0,95922381,previd=1672274690459582723,nextid=1672093401297648786
I might be wrong but this is as far as I've gotten in terms of debug.
|
1.0
|
QRP-Labs QDX IF command parsing buggy - I'm using a QRP-Labs QDX which emulates TS-440. When requesting PTT state from rigctl (sending command 't') it returns 0 (RX) even when still transmitting.
It appears that the 440 protocol returns accurate PTT state from 'P8' of the 'IF' command as per the spec. I'm going to suggest that hamlib reads PTT state from this, rather than return 0 as it does now.
Discussion here: https://groups.io/g/QRPLabs/topic/95922381?p=Created,,,20,1,0,0::recentpostdate/sticky,,,20,2,0,95922381,previd=1672274690459582723,nextid=1672093401297648786
I might be wrong but this is as far as I've gotten in terms of debug.
|
test
|
qrp labs qdx if command parsing buggy i m using a qrp labs qdx which emulates ts when requesting ptt state from rigctl sending command t it returns rx even when still transmitting it appears that the protocol returns accurate ptt state from of the if command as per the spec i m going to suggest that hamlib reads ptt state from this rather than return as it does now discussion here i might be wrong but this is as far as i ve gotten in terms of debug
| 1
|
60,385
| 6,689,743,960
|
IssuesEvent
|
2017-10-09 05:02:26
|
Kademi/kademi-dev
|
https://api.github.com/repos/Kademi/kademi-dev
|
closed
|
Improve UI of switch repo
|
Ready to Test QA
|
Instead of showing in dropdown list, we can show in combo with chosen plugin
|
1.0
|
Improve UI of switch repo - Instead of showing in dropdown list, we can show in combo with chosen plugin
|
test
|
improve ui of switch repo instead of showing in dropdown list we can show in combo with chosen plugin
| 1
|
99,122
| 16,430,784,870
|
IssuesEvent
|
2021-05-20 01:03:37
|
DavidSpek/pipelines
|
https://api.github.com/repos/DavidSpek/pipelines
|
opened
|
CVE-2021-29515 (Low) detected in tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl
|
security vulnerability
|
## CVE-2021-29515 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt</p>
<p>Path to vulnerable library: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt,pipelines/samples/core/ai_platform/training</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an end-to-end open source platform for machine learning. The implementation of `MatrixDiag*` operations(https://github.com/tensorflow/tensorflow/blob/4c4f420e68f1cfaf8f4b6e8e3eb857e9e4c3ff33/tensorflow/core/kernels/linalg/matrix_diag_op.cc#L195-L197) does not validate that the tensor arguments are non-empty. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-05-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29515>CVE-2021-29515</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-hc6c-75p4-hmq4">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-hc6c-75p4-hmq4</a></p>
<p>Release Date: 2021-05-14</p>
<p>Fix Resolution: tensorflow - 2.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-29515 (Low) detected in tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl - ## CVE-2021-29515 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt</p>
<p>Path to vulnerable library: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt,pipelines/samples/core/ai_platform/training</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an end-to-end open source platform for machine learning. The implementation of `MatrixDiag*` operations(https://github.com/tensorflow/tensorflow/blob/4c4f420e68f1cfaf8f4b6e8e3eb857e9e4c3ff33/tensorflow/core/kernels/linalg/matrix_diag_op.cc#L195-L197) does not validate that the tensor arguments are non-empty. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-05-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29515>CVE-2021-29515</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-hc6c-75p4-hmq4">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-hc6c-75p4-hmq4</a></p>
<p>Release Date: 2021-05-14</p>
<p>Fix Resolution: tensorflow - 2.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve low detected in tensorflow whl cve low severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file pipelines contrib components openvino ovms deployer containers requirements txt path to vulnerable library pipelines contrib components openvino ovms deployer containers requirements txt pipelines samples core ai platform training dependency hierarchy x tensorflow whl vulnerable library found in base branch master vulnerability details tensorflow is an end to end open source platform for machine learning the implementation of matrixdiag operations does not validate that the tensor arguments are non empty the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow step up your open source security game with whitesource
| 0
|
159,423
| 12,475,129,603
|
IssuesEvent
|
2020-05-29 10:57:53
|
tactilenews/100eyes
|
https://api.github.com/repos/tactilenews/100eyes
|
opened
|
Revise function "Hinweise anhängen"
|
Ergebnis Usertest
|
As an editor, the "attach notes"("Hinweise anhängen") function bothers me because it is very prominent, I don't understand it and it keeps me from asking a question.
|
1.0
|
Revise function "Hinweise anhängen" - As an editor, the "attach notes"("Hinweise anhängen") function bothers me because it is very prominent, I don't understand it and it keeps me from asking a question.
|
test
|
revise function hinweise anhängen as an editor the attach notes hinweise anhängen function bothers me because it is very prominent i don t understand it and it keeps me from asking a question
| 1
|
62,268
| 6,792,565,838
|
IssuesEvent
|
2017-11-01 01:17:16
|
googlefonts/fontbakery
|
https://api.github.com/repos/googlefonts/fontbakery
|
closed
|
Upstream UFO Check: Validate GLIF
|
testing
|
- [ ] Add an upstream UFO check that uses https://github.com/davelab6/glif-schema and http://lxml.de/validation.html to validate GLIF files
|
1.0
|
Upstream UFO Check: Validate GLIF - - [ ] Add an upstream UFO check that uses https://github.com/davelab6/glif-schema and http://lxml.de/validation.html to validate GLIF files
|
test
|
upstream ufo check validate glif add an upstream ufo check that uses and to validate glif files
| 1
|
52,440
| 12,965,507,827
|
IssuesEvent
|
2020-07-20 22:30:58
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
nixos.iso_minimal has a nondeterministic disk guid
|
0.kind: bug 6.topic: reproducible builds
|
**Describe the bug**
When building `nixos.iso_minimal.x86_64-linux` twice, the GPT partition has a different GUID.
**To Reproduce**
Steps to reproduce the behavior:
1. `nix-build ./nixos/release-combined.nix -A nixos.iso_minimal.x86_64-linux`
2. `nix-build ./nixos/release-combined.nix -A nixos.iso_minimal.x86_64-linux --check --keep-failed`
3. use `gdisk` to see the GUID is not consistent (or `diffoscope` at `0x00000238`, https://en.wikipedia.org/wiki/GUID_Partition_Table)
**Expected behavior**
A deterministic GUID
**Additional context**
For additional background on why it's helpful for builds to be reproducible, see https://reproducible-builds.org/
I'd be happy to help hunt down and fix this problem further, but this is rather new territory for me, so I'd appreciate any help
|
1.0
|
nixos.iso_minimal has a nondeterministic disk guid - **Describe the bug**
When building `nixos.iso_minimal.x86_64-linux` twice, the GPT partition has a different GUID.
**To Reproduce**
Steps to reproduce the behavior:
1. `nix-build ./nixos/release-combined.nix -A nixos.iso_minimal.x86_64-linux`
2. `nix-build ./nixos/release-combined.nix -A nixos.iso_minimal.x86_64-linux --check --keep-failed`
3. use `gdisk` to see the GUID is not consistent (or `diffoscope` at `0x00000238`, https://en.wikipedia.org/wiki/GUID_Partition_Table)
**Expected behavior**
A deterministic GUID
**Additional context**
For additional background on why it's helpful for builds to be reproducible, see https://reproducible-builds.org/
I'd be happy to help hunt down and fix this problem further, but this is rather new territory for me, so I'd appreciate any help
|
non_test
|
nixos iso minimal has a nondeterministic disk guid describe the bug when building nixos iso minimal linux twice the gpt partition has a different guid to reproduce steps to reproduce the behavior nix build nixos release combined nix a nixos iso minimal linux nix build nixos release combined nix a nixos iso minimal linux check keep failed use gdisk to see the guid is not consistent or diffoscope at expected behavior a deterministic guid additional context for additional background on why it s helpful for builds to be reproducible see i d be happy to help hunt down and fix this problem further but this is rather new territory for me so i d appreciate any help
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.