Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
288,673
24,926,004,878
IssuesEvent
2022-10-31 07:25:47
kubernetes-sigs/kubespray
https://api.github.com/repos/kubernetes-sigs/kubespray
opened
Vagrant jobs are broken due to OS detection logic
kind/failing-test
<!-- Please only use this template for submitting reports about failing tests in Kubespray CI jobs --> **Which jobs are failing**: Sample: https://gitlab.com/kargo-ci/kubernetes-sigs-kubespray/-/jobs/3227705561 **Which test(s) are failing**: Vagrant tests: * vagrant_fedora35-kube-router * vagrant_ubuntu18-calico-dual-stack * vagrant_ubuntu18-flannel * vagrant_ubuntu20-flannel **Since when has it been failing**: Likely the breakage was introduced by https://github.com/kubernetes-sigs/kubespray/pull/9416 and perpetuated by https://github.com/kubernetes-sigs/kubespray/pull/9432 **Testgrid link**: N/A **Reason for failure**: Overriding ansible detected variables could have unforeseen side effects and should never be done. This looks like we need a better way to detect and account for operating system differences. **Anything else we need to know**: To fix the CI I propose we temporarily revert the two changes while the developers figure out a better solution for the problems they were trying to address with the changes they proposed there.
1.0
Vagrant jobs are broken due to OS detection logic - <!-- Please only use this template for submitting reports about failing tests in Kubespray CI jobs --> **Which jobs are failing**: Sample: https://gitlab.com/kargo-ci/kubernetes-sigs-kubespray/-/jobs/3227705561 **Which test(s) are failing**: Vagrant tests: * vagrant_fedora35-kube-router * vagrant_ubuntu18-calico-dual-stack * vagrant_ubuntu18-flannel * vagrant_ubuntu20-flannel **Since when has it been failing**: Likely the breakage was introduced by https://github.com/kubernetes-sigs/kubespray/pull/9416 and perpetuated by https://github.com/kubernetes-sigs/kubespray/pull/9432 **Testgrid link**: N/A **Reason for failure**: Overriding ansible detected variables could have unforeseen side effects and should never be done. This looks like we need a better way to detect and account for operating system differences. **Anything else we need to know**: To fix the CI I propose we temporarily revert the two changes while the developers figure out a better solution for the problems they were trying to address with the changes they proposed there.
test
vagrant jobs are broken due to os detection logic which jobs are failing sample which test s are failing vagrant tests vagrant kube router vagrant calico dual stack vagrant flannel vagrant flannel since when has it been failing likely the breakage was introduced by and perpetuated by testgrid link n a reason for failure overriding ansible detected variables could have unforeseen side effects and should never be done this looks like we need a better way to detect and account for operating system differences anything else we need to know to fix the ci i propose we temporarily revert the two changes while the developers figure out a better solution for the problems they were trying to address with the changes they proposed there
1
65,355
14,713,728,852
IssuesEvent
2021-01-05 10:49:03
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
closed
Triage and fix vulnerabilities, ensure important vulnerabilities fixed
pre-cloud-GA-scan team/security
This should be graduated into a tracking issue once scans are implemented.
True
Triage and fix vulnerabilities, ensure important vulnerabilities fixed - This should be graduated into a tracking issue once scans are implemented.
non_test
triage and fix vulnerabilities ensure important vulnerabilities fixed this should be graduated into a tracking issue once scans are implemented
0
336,862
30,225,074,602
IssuesEvent
2023-07-05 23:16:15
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix jax_nn_activations.test_jax_nn_swish
JAX Frontend Sub Task Failing Test
| | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5442531781"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5371448716/jobs/9744217013"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5371448716/jobs/9744217013"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5427202833"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5469565117"><img src=https://img.shields.io/badge/-failure-red></a>
1.0
Fix jax_nn_activations.test_jax_nn_swish - | | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5442531781"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5371448716/jobs/9744217013"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5371448716/jobs/9744217013"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5427202833"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5469565117"><img src=https://img.shields.io/badge/-failure-red></a>
test
fix jax nn activations test jax nn swish jax a href src numpy a href src tensorflow a href src torch a href src paddle a href src
1
66,681
8,038,361,397
IssuesEvent
2018-07-30 15:11:47
Opentrons/opentrons
https://api.github.com/repos/Opentrons/opentrons
opened
300 uL tip racks are referred to as 200 uL tip racks
bug protocol designer
Currently we are referring to ur 300 uL tip racks as 200 uL tip racks in both run app and PD.
1.0
300 uL tip racks are referred to as 200 uL tip racks - Currently we are referring to ur 300 uL tip racks as 200 uL tip racks in both run app and PD.
non_test
ul tip racks are referred to as ul tip racks currently we are referring to ur ul tip racks as ul tip racks in both run app and pd
0
220,609
17,210,984,937
IssuesEvent
2021-07-19 04:20:04
microsoft/BotFramework-Composer
https://api.github.com/repos/microsoft/BotFramework-Composer
closed
Problems pane defaults settings should be to show both errors and warning and all bots
P0 R14 Bugbash Testing & Debugging Type: Bug
<!-- Please search for your feature request before creating a new one. > <!-- Complete the necessary portions of this template and delete the rest. --> ## Describe the bug The Problems pane should be not remember settings when you switch bots. It should reset to show all errors, all warnings, and for all bots each time you open a bot in Composer. ## Version 2.0.0-nightly.258070.287a4dc ## Browser <!-- What browser are you using? --> - [x] Electron distribution - [ ] Chrome - [ ] Safari - [ ] Firefox - [ ] Edge ## OS <!-- What operating system are you using? --> - [x] macOS - [ ] Windows - [ ] Ubuntu ## To Reproduce Steps to reproduce the behavior: 1. Open any bot that does not have multiple bots. 2. Go back to the home page. 4. Open an enterprise assistant bot. 5. Open Problems pane and see that the pane is filtered to just the root bot. ## Expected behavior The Problems pane defaults should be reset to show everything (errors, warning, all bots) whenever I (re-)open a bot. ## Screenshots <!-- If applicable, add screenshots/gif/video to help explain your problem. --> https://user-images.githubusercontent.com/39317542/125354199-5fbeeb80-e318-11eb-9e6c-fed556645038.mov
1.0
Problems pane defaults settings should be to show both errors and warning and all bots - <!-- Please search for your feature request before creating a new one. > <!-- Complete the necessary portions of this template and delete the rest. --> ## Describe the bug The Problems pane should be not remember settings when you switch bots. It should reset to show all errors, all warnings, and for all bots each time you open a bot in Composer. ## Version 2.0.0-nightly.258070.287a4dc ## Browser <!-- What browser are you using? --> - [x] Electron distribution - [ ] Chrome - [ ] Safari - [ ] Firefox - [ ] Edge ## OS <!-- What operating system are you using? --> - [x] macOS - [ ] Windows - [ ] Ubuntu ## To Reproduce Steps to reproduce the behavior: 1. Open any bot that does not have multiple bots. 2. Go back to the home page. 4. Open an enterprise assistant bot. 5. Open Problems pane and see that the pane is filtered to just the root bot. ## Expected behavior The Problems pane defaults should be reset to show everything (errors, warning, all bots) whenever I (re-)open a bot. ## Screenshots <!-- If applicable, add screenshots/gif/video to help explain your problem. --> https://user-images.githubusercontent.com/39317542/125354199-5fbeeb80-e318-11eb-9e6c-fed556645038.mov
test
problems pane defaults settings should be to show both errors and warning and all bots describe the bug the problems pane should be not remember settings when you switch bots it should reset to show all errors all warnings and for all bots each time you open a bot in composer version nightly browser electron distribution chrome safari firefox edge os macos windows ubuntu to reproduce steps to reproduce the behavior open any bot that does not have multiple bots go back to the home page open an enterprise assistant bot open problems pane and see that the pane is filtered to just the root bot expected behavior the problems pane defaults should be reset to show everything errors warning all bots whenever i re open a bot screenshots
1
18,909
24,847,873,808
IssuesEvent
2022-10-26 17:21:42
rladstaetter/LogoRRR
https://api.github.com/repos/rladstaetter/LogoRRR
closed
Linux Support / Build
release process
Build LogoRRR on Linux and solve possible problems. Integrate Linux as supported Platform in Release.
1.0
Linux Support / Build - Build LogoRRR on Linux and solve possible problems. Integrate Linux as supported Platform in Release.
non_test
linux support build build logorrr on linux and solve possible problems integrate linux as supported platform in release
0
40,491
6,826,842,274
IssuesEvent
2017-11-08 15:20:49
openmpf/openmpf
https://api.github.com/repos/openmpf/openmpf
closed
Update Component API docs with Feed Forward for Generic Detections
documentation
Related to #326. Update Component API docs with Feed Forward for Generic Detections.
1.0
Update Component API docs with Feed Forward for Generic Detections - Related to #326. Update Component API docs with Feed Forward for Generic Detections.
non_test
update component api docs with feed forward for generic detections related to update component api docs with feed forward for generic detections
0
218,636
17,012,228,606
IssuesEvent
2021-07-02 07:01:21
mozilla-mobile/focus-android
https://api.github.com/repos/mozilla-mobile/focus-android
closed
Selecting 'Find in page' does not show the dialog when the dialog is
P3 bug size L testing
### Steps to reproduce - Open webpage, open find in page dialog - Scroll up, so the find in page dialog disappears from screen - Go to menu, and select find in page menu item again ### Expected behavior - Find in page dialog is in focus ### Actual behavior - Find in page dialog is not shown on screen ### Device information - Nexus 4, API 23
1.0
Selecting 'Find in page' does not show the dialog when the dialog is - ### Steps to reproduce - Open webpage, open find in page dialog - Scroll up, so the find in page dialog disappears from screen - Go to menu, and select find in page menu item again ### Expected behavior - Find in page dialog is in focus ### Actual behavior - Find in page dialog is not shown on screen ### Device information - Nexus 4, API 23
test
selecting find in page does not show the dialog when the dialog is steps to reproduce open webpage open find in page dialog scroll up so the find in page dialog disappears from screen go to menu and select find in page menu item again expected behavior find in page dialog is in focus actual behavior find in page dialog is not shown on screen device information nexus api
1
288,065
24,882,269,760
IssuesEvent
2022-10-28 03:01:05
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Orçamento - Legislação - Campestre
generalization test development template - Memory (66) tag - Orçamento subtag - Legislação
DoD: Realizar o teste de Generalização do validador da tag Orçamento - Legislação para o Município de Campestre.
1.0
Teste de generalizacao para a tag Orçamento - Legislação - Campestre - DoD: Realizar o teste de Generalização do validador da tag Orçamento - Legislação para o Município de Campestre.
test
teste de generalizacao para a tag orçamento legislação campestre dod realizar o teste de generalização do validador da tag orçamento legislação para o município de campestre
1
132,832
12,519,537,617
IssuesEvent
2020-06-03 14:35:05
GluuFederation/community-edition-setup
https://api.github.com/repos/GluuFederation/community-edition-setup
closed
Migrate existing u2f enrollments to the fido2 branch
Needs Documentation Needs QA enhancement
Consolidate u2f and fido2 enrollments in a single place. That way admins can employ a single custom script (fido2) so that old u2f credentials will still work This will also add some usability benefits for casa.
1.0
Migrate existing u2f enrollments to the fido2 branch - Consolidate u2f and fido2 enrollments in a single place. That way admins can employ a single custom script (fido2) so that old u2f credentials will still work This will also add some usability benefits for casa.
non_test
migrate existing enrollments to the branch consolidate and enrollments in a single place that way admins can employ a single custom script so that old credentials will still work this will also add some usability benefits for casa
0
20,775
3,634,163,958
IssuesEvent
2016-02-11 17:00:51
SIB-Colombia/cygnus
https://api.github.com/repos/SIB-Colombia/cygnus
closed
Eliminar el marco polaroid
design
Las imágenes del preview de las fichas poseen un marco que no va con el diseño acordado. Hay que eliminarlo. ![captura de pantalla 2016-02-11 a las 10 42 22 a m](https://cloud.githubusercontent.com/assets/11927544/12980846/3685d262-d0ac-11e5-9783-e3abe1bfd09d.png)
1.0
Eliminar el marco polaroid - Las imágenes del preview de las fichas poseen un marco que no va con el diseño acordado. Hay que eliminarlo. ![captura de pantalla 2016-02-11 a las 10 42 22 a m](https://cloud.githubusercontent.com/assets/11927544/12980846/3685d262-d0ac-11e5-9783-e3abe1bfd09d.png)
non_test
eliminar el marco polaroid las imágenes del preview de las fichas poseen un marco que no va con el diseño acordado hay que eliminarlo
0
108,985
23,691,457,590
IssuesEvent
2022-08-29 11:12:15
pkgjs/parseargs
https://api.github.com/repos/pkgjs/parseargs
closed
How might I add my own support for negative numbers?
bring-your-own-code
(This is a bring-your-own-feature experiment, not a prototype of parseArgs implementation.) Inspired by discussion in #62. How could I add my own support for options and positionals treating negative numbers as ordinary args and not options? ```js const { parseArgs } = require('@pkgjs/parseArgs'); const kNegativePrefix = 'parseArgs.NEGATIVE:'; // preprocess const rawArgs = process.argv.slice(2); const preparedArgs = rawArgs.map(arg => /^-[0-9]+/.test(arg) ? kNegativePrefix.concat(arg) : arg); const result = parseArgs({ args: preparedArgs, { options: { profit: { type: 'string' }} }); // postprocess const stripPrefix = (arg) => arg.startsWith(kNegativePrefix) ? arg.slice(kNegativePrefix.length) : arg; result.positionals = result.positionals.map(arg => stripPrefix(arg)); Object.entries(result.values).forEach(([key, value]) => { if (typeof value === 'string') result.values[key] = stripPrefix(value); }); console.log(result); ``` ``` % node index.js --profit 33 44 { flags: { profit: true }, values: { profit: '33' }, positionals: [ '44' ] } % node index.js --profit -33 -44 { flags: { profit: true }, values: { profit: '-33' }, positionals: [ '-44' ] } ```
1.0
How might I add my own support for negative numbers? - (This is a bring-your-own-feature experiment, not a prototype of parseArgs implementation.) Inspired by discussion in #62. How could I add my own support for options and positionals treating negative numbers as ordinary args and not options? ```js const { parseArgs } = require('@pkgjs/parseArgs'); const kNegativePrefix = 'parseArgs.NEGATIVE:'; // preprocess const rawArgs = process.argv.slice(2); const preparedArgs = rawArgs.map(arg => /^-[0-9]+/.test(arg) ? kNegativePrefix.concat(arg) : arg); const result = parseArgs({ args: preparedArgs, { options: { profit: { type: 'string' }} }); // postprocess const stripPrefix = (arg) => arg.startsWith(kNegativePrefix) ? arg.slice(kNegativePrefix.length) : arg; result.positionals = result.positionals.map(arg => stripPrefix(arg)); Object.entries(result.values).forEach(([key, value]) => { if (typeof value === 'string') result.values[key] = stripPrefix(value); }); console.log(result); ``` ``` % node index.js --profit 33 44 { flags: { profit: true }, values: { profit: '33' }, positionals: [ '44' ] } % node index.js --profit -33 -44 { flags: { profit: true }, values: { profit: '-33' }, positionals: [ '-44' ] } ```
non_test
how might i add my own support for negative numbers this is a bring your own feature experiment not a prototype of parseargs implementation inspired by discussion in how could i add my own support for options and positionals treating negative numbers as ordinary args and not options js const parseargs require pkgjs parseargs const knegativeprefix parseargs negative preprocess const rawargs process argv slice const preparedargs rawargs map arg test arg knegativeprefix concat arg arg const result parseargs args preparedargs options profit type string postprocess const stripprefix arg arg startswith knegativeprefix arg slice knegativeprefix length arg result positionals result positionals map arg stripprefix arg object entries result values foreach if typeof value string result values stripprefix value console log result node index js profit flags profit true values profit positionals node index js profit flags profit true values profit positionals
0
11,997
3,249,234,336
IssuesEvent
2015-10-18 00:24:10
jQueryGeo/geo
https://api.github.com/repos/jQueryGeo/geo
closed
find should return shapes in reverse order
2 - Medium Test With Latest
Shapes are drawn in FIFO meaning shapes appended later are drawn on top of shapes appended before. However, the find method searches the same way, finding a large shape first if it was appended first; which it would have been because you'd have to append the smaller shape second to see it on top.
2.0
find should return shapes in reverse order - Shapes are drawn in FIFO meaning shapes appended later are drawn on top of shapes appended before. However, the find method searches the same way, finding a large shape first if it was appended first; which it would have been because you'd have to append the smaller shape second to see it on top.
test
find should return shapes in reverse order shapes are drawn in fifo meaning shapes appended later are drawn on top of shapes appended before however the find method searches the same way finding a large shape first if it was appended first which it would have been because you d have to append the smaller shape second to see it on top
1
669,970
22,647,861,459
IssuesEvent
2022-07-01 10:28:11
theAsmodai/metamod-r
https://api.github.com/repos/theAsmodai/metamod-r
closed
Problem compiling with ICC 19
Type: Bug Priority: Medium Status: Available OS: Independent
https://github.com/theAsmodai/metamod-r/blob/0cf2f709dbeae18ca84d2fafd4481ffbba06ad0c/metamod/src/sys_module.cpp#L145 Is it ok to use "false" here? Looks like ICC don't like that. /home/user/build/metamod-r/metamod/src/sys_module.cpp(145): error: return value type does not match the function type return false; ^
1.0
Problem compiling with ICC 19 - https://github.com/theAsmodai/metamod-r/blob/0cf2f709dbeae18ca84d2fafd4481ffbba06ad0c/metamod/src/sys_module.cpp#L145 Is it ok to use "false" here? Looks like ICC don't like that. /home/user/build/metamod-r/metamod/src/sys_module.cpp(145): error: return value type does not match the function type return false; ^
non_test
problem compiling with icc is it ok to use false here looks like icc don t like that home user build metamod r metamod src sys module cpp error return value type does not match the function type return false
0
539,424
15,788,379,009
IssuesEvent
2021-04-01 20:42:09
CanberraOceanRacingClub/namadgi3
https://api.github.com/repos/CanberraOceanRacingClub/namadgi3
opened
Docking at Holme Port
Notice priority 2: Medium
Skippers are asked to consider the following procedure when docking at Holme Port: While **slowly reversing** into the pen: 1. Attach ``Stern Spring`` to Namadgi's midships cleat -- continue reversing to tension the line 2. Attach bow line to dock cleat and secure after adjusting to length 3. Attach port stern line to dock cleat 4. Attach starboard stern line to dock cleat When departing **PLEASE leave the ``Stern Spring`` on the dock**, attached to it's cleat, ready for reuse. ![Docking diagram](https://github.com/CanberraOceanRacingClub/namadgi3/raw/master/docs/holmeport_docking.png)
1.0
Docking at Holme Port - Skippers are asked to consider the following procedure when docking at Holme Port: While **slowly reversing** into the pen: 1. Attach ``Stern Spring`` to Namadgi's midships cleat -- continue reversing to tension the line 2. Attach bow line to dock cleat and secure after adjusting to length 3. Attach port stern line to dock cleat 4. Attach starboard stern line to dock cleat When departing **PLEASE leave the ``Stern Spring`` on the dock**, attached to it's cleat, ready for reuse. ![Docking diagram](https://github.com/CanberraOceanRacingClub/namadgi3/raw/master/docs/holmeport_docking.png)
non_test
docking at holme port skippers are asked to consider the following procedure when docking at holme port while slowly reversing into the pen attach stern spring to namadgi s midships cleat continue reversing to tension the line attach bow line to dock cleat and secure after adjusting to length attach port stern line to dock cleat attach starboard stern line to dock cleat when departing please leave the stern spring on the dock attached to it s cleat ready for reuse
0
258,557
22,327,855,473
IssuesEvent
2022-06-14 12:16:57
Uuvana-Studios/longvinter-windows-client
https://api.github.com/repos/Uuvana-Studios/longvinter-windows-client
closed
When Opening chests, cooked fish looses its energy symbol
Bug Low Priority Tested
**Describe the bug** In the open chest menu, the symbol for cooked fish and raw fish is the same. The energy symbol added on the cooked fish disappears. You need to hover over the fish to find out if it;'s cooked or not **To Reproduce** Steps to reproduce the behavior: 1. Have cooked fish in inventory' 2. Open a chest 3. the inventory menu for both character bag and chest does not show the energy symbol for cooked fish **Expected behavior** the energy symbol for cooked fish should be retained when opening chests to easily differentiate between cooked and raw **Screenshots** If applicable, add screenshots to help explain your problem. **Desktop (please complete the following information):** - OS: Windows 11 - Game Version Beta 1.0.0 - Steam Version v020 **Additional context** Add any other context about the problem here.
1.0
When Opening chests, cooked fish looses its energy symbol - **Describe the bug** In the open chest menu, the symbol for cooked fish and raw fish is the same. The energy symbol added on the cooked fish disappears. You need to hover over the fish to find out if it;'s cooked or not **To Reproduce** Steps to reproduce the behavior: 1. Have cooked fish in inventory' 2. Open a chest 3. the inventory menu for both character bag and chest does not show the energy symbol for cooked fish **Expected behavior** the energy symbol for cooked fish should be retained when opening chests to easily differentiate between cooked and raw **Screenshots** If applicable, add screenshots to help explain your problem. **Desktop (please complete the following information):** - OS: Windows 11 - Game Version Beta 1.0.0 - Steam Version v020 **Additional context** Add any other context about the problem here.
test
when opening chests cooked fish looses its energy symbol describe the bug in the open chest menu the symbol for cooked fish and raw fish is the same the energy symbol added on the cooked fish disappears you need to hover over the fish to find out if it s cooked or not to reproduce steps to reproduce the behavior have cooked fish in inventory open a chest the inventory menu for both character bag and chest does not show the energy symbol for cooked fish expected behavior the energy symbol for cooked fish should be retained when opening chests to easily differentiate between cooked and raw screenshots if applicable add screenshots to help explain your problem desktop please complete the following information os windows game version beta steam version additional context add any other context about the problem here
1
239,318
19,846,918,560
IssuesEvent
2022-01-21 07:49:33
Oldes/Rebol-issues
https://api.github.com/repos/Oldes/Rebol-issues
closed
TRIM /head and /tail on block
Test.written Type.bug Datatype: block! CC.resolved
_Submitted by:_ **Carl** Gives an error, but should be valid. ``` rebol >> trim/head reduce [none 1 none 2] == [1 none 2] >> trim/tail reduce [1 none 2 none] == [1 none 2] ``` --- <sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=825)** [ Version: alpha 54 Type: Bug Platform: All Category: n/a Reproduce: Always Fixed-in:alpha 55 ]</sup> <sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/825</sup> Comments: --- --- > **Rebolbot** added the **Type.bug** on Jan 12, 2016 ---
1.0
TRIM /head and /tail on block - _Submitted by:_ **Carl** Gives an error, but should be valid. ``` rebol >> trim/head reduce [none 1 none 2] == [1 none 2] >> trim/tail reduce [1 none 2 none] == [1 none 2] ``` --- <sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=825)** [ Version: alpha 54 Type: Bug Platform: All Category: n/a Reproduce: Always Fixed-in:alpha 55 ]</sup> <sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/825</sup> Comments: --- --- > **Rebolbot** added the **Type.bug** on Jan 12, 2016 ---
test
trim head and tail on block submitted by carl gives an error but should be valid rebol trim head reduce trim tail reduce imported from imported from comments rebolbot added the type bug on jan
1
305,540
26,391,923,255
IssuesEvent
2023-01-12 16:16:43
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix reduction_ops.test_torch_prod
PyTorch Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_reduction_ops.py::test_torch_prod[cpu-ivy.functional.backends.numpy-False-False]</summary> 2023-01-12T03:47:11.4120487Z E AssertionError: the return with a torch backend produced data type of float64, while the return with a numpy backend returned a data type of float32. 2023-01-12T03:47:11.4120954Z E Falsifying example: test_torch_prod( 2023-01-12T03:47:11.4121740Z E dtype_x_axis=(['float64'], [array([-1.])], 0), 2023-01-12T03:47:11.4122064Z E dtype=[None], 2023-01-12T03:47:11.4122317Z E keepdims=False, 2023-01-12T03:47:11.4122585Z E as_variable=[False], 2023-01-12T03:47:11.4123810Z E num_positional_args=0, 2023-01-12T03:47:11.4124153Z E native_array=[False], 2023-01-12T03:47:11.4124473Z E with_out=False, 2023-01-12T03:47:11.4124942Z E fn_tree='ivy.functional.frontends.torch.prod', 2023-01-12T03:47:11.4125324Z E on_device='cpu', 2023-01-12T03:47:11.4125646Z E frontend='torch', 2023-01-12T03:47:11.4125887Z E ) 2023-01-12T03:47:11.4126093Z E 2023-01-12T03:47:11.4126734Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2JkAAMoBaTZIQwAAK4ADQ==') as a decorator on your test case </details>
1.0
Fix reduction_ops.test_torch_prod - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3898903911/jobs/6658070666" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_reduction_ops.py::test_torch_prod[cpu-ivy.functional.backends.numpy-False-False]</summary> 2023-01-12T03:47:11.4120487Z E AssertionError: the return with a torch backend produced data type of float64, while the return with a numpy backend returned a data type of float32. 2023-01-12T03:47:11.4120954Z E Falsifying example: test_torch_prod( 2023-01-12T03:47:11.4121740Z E dtype_x_axis=(['float64'], [array([-1.])], 0), 2023-01-12T03:47:11.4122064Z E dtype=[None], 2023-01-12T03:47:11.4122317Z E keepdims=False, 2023-01-12T03:47:11.4122585Z E as_variable=[False], 2023-01-12T03:47:11.4123810Z E num_positional_args=0, 2023-01-12T03:47:11.4124153Z E native_array=[False], 2023-01-12T03:47:11.4124473Z E with_out=False, 2023-01-12T03:47:11.4124942Z E fn_tree='ivy.functional.frontends.torch.prod', 2023-01-12T03:47:11.4125324Z E on_device='cpu', 2023-01-12T03:47:11.4125646Z E frontend='torch', 2023-01-12T03:47:11.4125887Z E ) 2023-01-12T03:47:11.4126093Z E 2023-01-12T03:47:11.4126734Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2JkAAMoBaTZIQwAAK4ADQ==') as a decorator on your test case </details>
test
fix reduction ops test torch prod tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test torch test reduction ops py test torch prod e assertionerror the return with a torch backend produced data type of while the return with a numpy backend returned a data type of e falsifying example test torch prod e dtype x axis e dtype e keepdims false e as variable e num positional args e native array e with out false e fn tree ivy functional frontends torch prod e on device cpu e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
1
73,103
3,407,210,493
IssuesEvent
2015-12-04 01:11:14
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
Possible regression in waiting for DOMContentLoaded
area-dart2js area-html Priority-Medium
Opening a new bug for possible regression reported in https://github.com/dart-lang/sdk/issues/885#issuecomment-161689055 @vsmenon did the original fix. Comment from there by @joseluis is On version 1.13.0 this doesn't seem true anymore. Minimal example: querySelector('#someid'); returns null when the element is placed after the script tag that loads the Dart script compiled with dart2js. In order to make it work I must use the script defer attribute, like this: ```html <html> <head> <meta charset="utf-8"> </head> <body> <script src="dart-program.js" defer></script> <script id="someid"></script> </body> </html> ```
1.0
Possible regression in waiting for DOMContentLoaded - Opening a new bug for possible regression reported in https://github.com/dart-lang/sdk/issues/885#issuecomment-161689055 @vsmenon did the original fix. Comment from there by @joseluis is On version 1.13.0 this doesn't seem true anymore. Minimal example: querySelector('#someid'); returns null when the element is placed after the script tag that loads the Dart script compiled with dart2js. In order to make it work I must use the script defer attribute, like this: ```html <html> <head> <meta charset="utf-8"> </head> <body> <script src="dart-program.js" defer></script> <script id="someid"></script> </body> </html> ```
non_test
possible regression in waiting for domcontentloaded opening a new bug for possible regression reported in vsmenon did the original fix comment from there by joseluis is on version this doesn t seem true anymore minimal example queryselector someid returns null when the element is placed after the script tag that loads the dart script compiled with in order to make it work i must use the script defer attribute like this html
0
253,312
21,673,814,904
IssuesEvent
2022-05-08 11:44:42
mennaelkashef/eShop
https://api.github.com/repos/mennaelkashef/eShop
opened
this is a message to test encryption time
Hello! RULE-GOT-APPLIED DOES-NOT-CONTAIN-STRING Rule-works-on-convert-to-bug test instabug
# :clipboard: Bug Details >this is a message to test encryption time key | value --|-- Reported At | 2022-04-28 13:14:52 UTC Email | abarakat@gmail.com Categories | Report a bug Tags | test, Hello!, RULE-GOT-APPLIED, DOES-NOT-CONTAIN-STRING, Rule-works-on-convert-to-bug, instabug App Version | 1.1 (1) Session Duration | 154 Device | OnePlus KB2001, OS Level 30 Display | 1080x2400 (xhdpi) Location | Giza, Egypt (en) ## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?utm_source=github&utm_medium=integrations) :point_left: ___ # :iphone: View Hierarchy This bug was reported from **com.example.app.complexviews.ComplexViewsFragment** Find its interactive view hierarchy with all its subviews here: :point_right: **[Check View Hierarchy](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-hierarchy-view=true&utm_source=github&utm_medium=integrations)** :point_left: ___ # :chart_with_downwards_trend: Session Profiler Here is what the app was doing right before the bug was reported: Key | Value --|-- Used Memory | 58.1% - 4.27/7.34 GB Used Storage | 7.1% - 7.48/105.28 GB Connectivity | no_connection Battery | 100% - plugged Orientation | portrait Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left: ___ # :bust_in_silhouette: User Info ### User Attributes ``` key_name -643333045: key value bla bla bla la key_name -1391770268: key value bla bla bla la key_name -1233063091: key value bla bla bla la key_name 12389232: key value bla bla bla la key_name -1400931524: key value bla bla bla la key_name 1545921878: key value bla bla bla la key_name -778916395: key value bla bla bla la key_name -1822943291: key value bla bla bla la key_name 1537783730: key value bla bla bla la key_name -2119608186: key value bla bla bla la key_name -683370323: key value bla bla bla la key_name -1977430918: key value bla bla bla la key_name -223322809: key value bla bla bla la ``` ___ # :mag_right: Logs ### User Steps Here are the last 10 steps done by the user right before the bug was reported: ``` 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 com.example.app.developerOption.list.PhotosActivity was paused. 13:14:52 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" ``` Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left: ### Console Log Here are the last 10 console logs logged right before the bug was reported: ``` 13:14:56 D/StrictMode(17648): at android.view.View.access$3600(View.java:826) 13:14:56 D/StrictMode(17648): at android.view.View$PerformClick.run(View.java:28555) 13:14:56 D/StrictMode(17648): at android.os.Handler.handleCallback(Handler.java:938) 13:14:56 D/StrictMode(17648): at android.os.Handler.dispatchMessage(Handler.java:99) 13:14:56 D/StrictMode(17648): at android.os.Looper.loop(Looper.java:233) 13:14:56 D/StrictMode(17648): at android.app.ActivityThread.main(ActivityThread.java:8030) 13:14:56 D/StrictMode(17648): at java.lang.reflect.Method.invoke(Native Method) 13:14:56 D/StrictMode(17648): at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:631) 13:14:56 D/StrictMode(17648): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:978) 13:14:56 D/skia (17648): got procName com.example.app for pid 17648 ``` Find all the logged console logs throughout the session here: :point_right: **[View All Console Log](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-logs=console_log&utm_source=github&utm_medium=integrations)** :point_left: ___ # :camera: Images [![image attachment](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/18282355/a308575a5f1a44fd0352ae352e28eb6e_one_bug_thumb/25943016/bug_1651151691655_.jpg?Expires=4807683881&Signature=hG-MqWGO1T-I9dscfnESlYG3VAzP96oS9~~LeB4aIZkOsZD79hNiYP~kDltITIHORFTnJN0ltkoQWxkUX3ItkYf6ZKnfrtss~UsmaSBL7PvJIWc6DB7XbgSfamw6SIrrwP1XxkbWiuIxyMRrebH8hdkoTTD77l9KROHsqFpPl5K2U~P1wzDID79ott94x8wbDxZjfe6BlWELXRT6F4VwtCxn6O9Eu4C7g8Lrho4M6SlmNEAJT58cRw8dCsaat0TdPHfXQQ9RfvoYH2J-tiY2pUhS8BZw2MF4u-uxtnFu6a0wA6~P1CYArYQHgcoYw259fdYuvClEClFVpSUi0ejd-g__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/18282355/a308575a5f1a44fd0352ae352e28eb6e_original/25943016/bug_1651151691655_.jpg?Expires=4807683881&Signature=QOE3w0fF16ZCCYrKWHy0i-patAuo9V9lUuq2VBKREUCCFwb58gsQZGSEWdJ~jhg6ILNEBOejsCXAUpsjHrvQ4ShFtYpAhEmARJ3soeZQ0lCZfoZE9XypuunJud5IyvolXDd8GH2uBI5t1ECPKrUU7YTGqlUuFsExzBx-n2LWDHvXB8fTAvDpTRSylBYh9fIHqafB9Z2exl7tU67Yacva5Nop7IrfaK3mrlzLTrE0AKBUPc9eKIFwoG7t-Gwp2WMC-WgDAd8uZJ0vKHOiwc3HwBgzdeJVmRKUSG8yg1eEhVWDsxjBZFal42OYN81cxpwHULcDNx-nU1T5sKMQreOq7w__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ) ___ # :warning: Looking for More Details? 1. **Network Log**: we are unable to capture your network requests automatically. If you are using HttpUrlConnection or Okhttp requests, [**check the details mentioned here**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations#section-network-logs). 2. **User Events**: start capturing custom User Events to send them along with each report. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations). 3. **Instabug Log**: start adding Instabug logs to see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations).
1.0
this is a message to test encryption time - # :clipboard: Bug Details >this is a message to test encryption time key | value --|-- Reported At | 2022-04-28 13:14:52 UTC Email | abarakat@gmail.com Categories | Report a bug Tags | test, Hello!, RULE-GOT-APPLIED, DOES-NOT-CONTAIN-STRING, Rule-works-on-convert-to-bug, instabug App Version | 1.1 (1) Session Duration | 154 Device | OnePlus KB2001, OS Level 30 Display | 1080x2400 (xhdpi) Location | Giza, Egypt (en) ## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?utm_source=github&utm_medium=integrations) :point_left: ___ # :iphone: View Hierarchy This bug was reported from **com.example.app.complexviews.ComplexViewsFragment** Find its interactive view hierarchy with all its subviews here: :point_right: **[Check View Hierarchy](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-hierarchy-view=true&utm_source=github&utm_medium=integrations)** :point_left: ___ # :chart_with_downwards_trend: Session Profiler Here is what the app was doing right before the bug was reported: Key | Value --|-- Used Memory | 58.1% - 4.27/7.34 GB Used Storage | 7.1% - 7.48/105.28 GB Connectivity | no_connection Battery | 100% - plugged Orientation | portrait Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left: ___ # :bust_in_silhouette: User Info ### User Attributes ``` key_name -643333045: key value bla bla bla la key_name -1391770268: key value bla bla bla la key_name -1233063091: key value bla bla bla la key_name 12389232: key value bla bla bla la key_name -1400931524: key value bla bla bla la key_name 1545921878: key value bla bla bla la key_name -778916395: key value bla bla bla la key_name -1822943291: key value bla bla bla la key_name 1537783730: key value bla bla bla la key_name -2119608186: key value bla bla bla la key_name -683370323: key value bla bla bla la key_name -1977430918: key value bla bla bla la key_name -223322809: key value bla bla bla la ``` ___ # :mag_right: Logs ### User Steps Here are the last 10 steps done by the user right before the bug was reported: ``` 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" 13:14:51 com.example.app.developerOption.list.PhotosActivity was paused. 13:14:52 Tap in "android.widget.LinearLayout" in "com.example.app.developerOption.list.PhotosActivity" ``` Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left: ### Console Log Here are the last 10 console logs logged right before the bug was reported: ``` 13:14:56 D/StrictMode(17648): at android.view.View.access$3600(View.java:826) 13:14:56 D/StrictMode(17648): at android.view.View$PerformClick.run(View.java:28555) 13:14:56 D/StrictMode(17648): at android.os.Handler.handleCallback(Handler.java:938) 13:14:56 D/StrictMode(17648): at android.os.Handler.dispatchMessage(Handler.java:99) 13:14:56 D/StrictMode(17648): at android.os.Looper.loop(Looper.java:233) 13:14:56 D/StrictMode(17648): at android.app.ActivityThread.main(ActivityThread.java:8030) 13:14:56 D/StrictMode(17648): at java.lang.reflect.Method.invoke(Native Method) 13:14:56 D/StrictMode(17648): at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:631) 13:14:56 D/StrictMode(17648): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:978) 13:14:56 D/skia (17648): got procName com.example.app for pid 17648 ``` Find all the logged console logs throughout the session here: :point_right: **[View All Console Log](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8711?show-logs=console_log&utm_source=github&utm_medium=integrations)** :point_left: ___ # :camera: Images [![image attachment](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/18282355/a308575a5f1a44fd0352ae352e28eb6e_one_bug_thumb/25943016/bug_1651151691655_.jpg?Expires=4807683881&Signature=hG-MqWGO1T-I9dscfnESlYG3VAzP96oS9~~LeB4aIZkOsZD79hNiYP~kDltITIHORFTnJN0ltkoQWxkUX3ItkYf6ZKnfrtss~UsmaSBL7PvJIWc6DB7XbgSfamw6SIrrwP1XxkbWiuIxyMRrebH8hdkoTTD77l9KROHsqFpPl5K2U~P1wzDID79ott94x8wbDxZjfe6BlWELXRT6F4VwtCxn6O9Eu4C7g8Lrho4M6SlmNEAJT58cRw8dCsaat0TdPHfXQQ9RfvoYH2J-tiY2pUhS8BZw2MF4u-uxtnFu6a0wA6~P1CYArYQHgcoYw259fdYuvClEClFVpSUi0ejd-g__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/18282355/a308575a5f1a44fd0352ae352e28eb6e_original/25943016/bug_1651151691655_.jpg?Expires=4807683881&Signature=QOE3w0fF16ZCCYrKWHy0i-patAuo9V9lUuq2VBKREUCCFwb58gsQZGSEWdJ~jhg6ILNEBOejsCXAUpsjHrvQ4ShFtYpAhEmARJ3soeZQ0lCZfoZE9XypuunJud5IyvolXDd8GH2uBI5t1ECPKrUU7YTGqlUuFsExzBx-n2LWDHvXB8fTAvDpTRSylBYh9fIHqafB9Z2exl7tU67Yacva5Nop7IrfaK3mrlzLTrE0AKBUPc9eKIFwoG7t-Gwp2WMC-WgDAd8uZJ0vKHOiwc3HwBgzdeJVmRKUSG8yg1eEhVWDsxjBZFal42OYN81cxpwHULcDNx-nU1T5sKMQreOq7w__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ) ___ # :warning: Looking for More Details? 1. **Network Log**: we are unable to capture your network requests automatically. If you are using HttpUrlConnection or Okhttp requests, [**check the details mentioned here**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations#section-network-logs). 2. **User Events**: start capturing custom User Events to send them along with each report. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations). 3. **Instabug Log**: start adding Instabug logs to see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations).
test
this is a message to test encryption time clipboard bug details this is a message to test encryption time key value reported at utc email abarakat gmail com categories report a bug tags test hello rule got applied does not contain string rule works on convert to bug instabug app version session duration device oneplus os level display xhdpi location giza egypt en point right point left iphone view hierarchy this bug was reported from com example app complexviews complexviewsfragment find its interactive view hierarchy with all its subviews here point right point left chart with downwards trend session profiler here is what the app was doing right before the bug was reported key value used memory gb used storage gb connectivity no connection battery plugged orientation portrait find all the changes that happened in the parameters mentioned above during the last seconds before the bug was reported here point right point left bust in silhouette user info user attributes key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la mag right logs user steps here are the last steps done by the user right before the bug was reported tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity tap in android widget linearlayout in com example app developeroption list photosactivity com example app developeroption list photosactivity was paused tap in android widget linearlayout in com example app developeroption list photosactivity find all the user steps done by the user throughout the session here point right point left console log here are the last console logs logged right before the bug was reported d strictmode at android view view access view java d strictmode at android view view performclick run view java d strictmode at android os handler handlecallback handler java d strictmode at android os handler dispatchmessage handler java d strictmode at android os looper loop looper java d strictmode at android app activitythread main activitythread java d strictmode at java lang reflect method invoke native method d strictmode at com android internal os runtimeinit methodandargscaller run runtimeinit java d strictmode at com android internal os zygoteinit main zygoteinit java d skia got procname com example app for pid find all the logged console logs throughout the session here point right point left camera images warning looking for more details network log we are unable to capture your network requests automatically if you are using httpurlconnection or okhttp requests user events start capturing custom user events to send them along with each report instabug log start adding instabug logs to see them right inside each report you receive
1
123,967
17,772,422,600
IssuesEvent
2021-08-30 15:03:52
kapseliboi/mimic
https://api.github.com/repos/kapseliboi/mimic
opened
CVE-2017-16137 (Medium) detected in multiple libraries
security vulnerability
## CVE-2017-16137 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>debug-2.6.7.tgz</b>, <b>debug-2.2.0.tgz</b>, <b>debug-2.3.3.tgz</b>, <b>debug-2.6.8.tgz</b></p></summary> <p> <details><summary><b>debug-2.6.7.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.7.tgz">https://registry.npmjs.org/debug/-/debug-2.6.7.tgz</a></p> <p>Path to dependency file: mimic/package.json</p> <p>Path to vulnerable library: mimic/node_modules/express/node_modules/debug/package.json,mimic/node_modules/body-parser/node_modules/debug/package.json,mimic/node_modules/connect/node_modules/debug/package.json,mimic/node_modules/send/node_modules/debug/package.json,mimic/node_modules/finalhandler/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - karma-1.7.0.tgz (Root Library) - connect-3.6.2.tgz - :x: **debug-2.6.7.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.2.0.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p> <p>Path to dependency file: mimic/package.json</p> <p>Path to vulnerable library: mimic/node_modules/socket.io-parser/node_modules/debug/package.json,mimic/node_modules/mocha-nightwatch/node_modules/debug/package.json,mimic/node_modules/compression/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - karma-1.7.0.tgz (Root Library) - socket.io-1.7.3.tgz - socket.io-parser-2.3.1.tgz - :x: **debug-2.2.0.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.3.3.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.3.3.tgz">https://registry.npmjs.org/debug/-/debug-2.3.3.tgz</a></p> <p>Path to dependency file: mimic/package.json</p> <p>Path to vulnerable library: mimic/node_modules/engine.io/node_modules/debug/package.json,mimic/node_modules/socket.io/node_modules/debug/package.json,mimic/node_modules/socket.io-client/node_modules/debug/package.json,mimic/node_modules/socket.io-adapter/node_modules/debug/package.json,mimic/node_modules/engine.io-client/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - karma-1.7.0.tgz (Root Library) - socket.io-1.7.3.tgz - engine.io-1.8.3.tgz - :x: **debug-2.3.3.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.6.8.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.8.tgz">https://registry.npmjs.org/debug/-/debug-2.6.8.tgz</a></p> <p>Path to dependency file: mimic/package.json</p> <p>Path to vulnerable library: mimic/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - babel-core-6.25.0.tgz (Root Library) - :x: **debug-2.6.8.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/mimic/commit/1b3df8b0fb992059d70591ed86206dbf53aec93d">1b3df8b0fb992059d70591ed86206dbf53aec93d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 2.6.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-16137 (Medium) detected in multiple libraries - ## CVE-2017-16137 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>debug-2.6.7.tgz</b>, <b>debug-2.2.0.tgz</b>, <b>debug-2.3.3.tgz</b>, <b>debug-2.6.8.tgz</b></p></summary> <p> <details><summary><b>debug-2.6.7.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.7.tgz">https://registry.npmjs.org/debug/-/debug-2.6.7.tgz</a></p> <p>Path to dependency file: mimic/package.json</p> <p>Path to vulnerable library: mimic/node_modules/express/node_modules/debug/package.json,mimic/node_modules/body-parser/node_modules/debug/package.json,mimic/node_modules/connect/node_modules/debug/package.json,mimic/node_modules/send/node_modules/debug/package.json,mimic/node_modules/finalhandler/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - karma-1.7.0.tgz (Root Library) - connect-3.6.2.tgz - :x: **debug-2.6.7.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.2.0.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p> <p>Path to dependency file: mimic/package.json</p> <p>Path to vulnerable library: mimic/node_modules/socket.io-parser/node_modules/debug/package.json,mimic/node_modules/mocha-nightwatch/node_modules/debug/package.json,mimic/node_modules/compression/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - karma-1.7.0.tgz (Root Library) - socket.io-1.7.3.tgz - socket.io-parser-2.3.1.tgz - :x: **debug-2.2.0.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.3.3.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.3.3.tgz">https://registry.npmjs.org/debug/-/debug-2.3.3.tgz</a></p> <p>Path to dependency file: mimic/package.json</p> <p>Path to vulnerable library: mimic/node_modules/engine.io/node_modules/debug/package.json,mimic/node_modules/socket.io/node_modules/debug/package.json,mimic/node_modules/socket.io-client/node_modules/debug/package.json,mimic/node_modules/socket.io-adapter/node_modules/debug/package.json,mimic/node_modules/engine.io-client/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - karma-1.7.0.tgz (Root Library) - socket.io-1.7.3.tgz - engine.io-1.8.3.tgz - :x: **debug-2.3.3.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.6.8.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.8.tgz">https://registry.npmjs.org/debug/-/debug-2.6.8.tgz</a></p> <p>Path to dependency file: mimic/package.json</p> <p>Path to vulnerable library: mimic/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - babel-core-6.25.0.tgz (Root Library) - :x: **debug-2.6.8.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/mimic/commit/1b3df8b0fb992059d70591ed86206dbf53aec93d">1b3df8b0fb992059d70591ed86206dbf53aec93d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 2.6.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries debug tgz debug tgz debug tgz debug tgz debug tgz small debugging utility library home page a href path to dependency file mimic package json path to vulnerable library mimic node modules express node modules debug package json mimic node modules body parser node modules debug package json mimic node modules connect node modules debug package json mimic node modules send node modules debug package json mimic node modules finalhandler node modules debug package json dependency hierarchy karma tgz root library connect tgz x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file mimic package json path to vulnerable library mimic node modules socket io parser node modules debug package json mimic node modules mocha nightwatch node modules debug package json mimic node modules compression node modules debug package json dependency hierarchy karma tgz root library socket io tgz socket io parser tgz x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file mimic package json path to vulnerable library mimic node modules engine io node modules debug package json mimic node modules socket io node modules debug package json mimic node modules socket io client node modules debug package json mimic node modules socket io adapter node modules debug package json mimic node modules engine io client node modules debug package json dependency hierarchy karma tgz root library socket io tgz engine io tgz x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file mimic package json path to vulnerable library mimic node modules debug package json dependency hierarchy babel core tgz root library x debug tgz vulnerable library found in head commit a href found in base branch master vulnerability details the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
22,517
3,787,948,146
IssuesEvent
2016-03-21 13:01:07
studentinsights/studentinsights
https://api.github.com/repos/studentinsights/studentinsights
opened
Full case history - section breaks across school years
design enhancement help wanted
This is with demo data, so it's not the best example, but with all the absence and tardy data it's hard to see milestone dates like the beginning of a new school year. <img width="600" alt="screen shot 2016-03-21 at 8 59 19 am" src="https://cloud.githubusercontent.com/assets/1056957/13919335/36248ca4-ef43-11e5-825b-b7d2f14acd71.png"> One idea might just be adding a heading for the start of each school year, or perhaps using the box styling for sections that @bw120 is awesomely trying out in https://github.com/studentinsights/studentinsights/pull/199.
1.0
Full case history - section breaks across school years - This is with demo data, so it's not the best example, but with all the absence and tardy data it's hard to see milestone dates like the beginning of a new school year. <img width="600" alt="screen shot 2016-03-21 at 8 59 19 am" src="https://cloud.githubusercontent.com/assets/1056957/13919335/36248ca4-ef43-11e5-825b-b7d2f14acd71.png"> One idea might just be adding a heading for the start of each school year, or perhaps using the box styling for sections that @bw120 is awesomely trying out in https://github.com/studentinsights/studentinsights/pull/199.
non_test
full case history section breaks across school years this is with demo data so it s not the best example but with all the absence and tardy data it s hard to see milestone dates like the beginning of a new school year img width alt screen shot at am src one idea might just be adding a heading for the start of each school year or perhaps using the box styling for sections that is awesomely trying out in
0
381,508
11,276,550,586
IssuesEvent
2020-01-14 23:35:39
googleapis/google-api-java-client-services
https://api.github.com/repos/googleapis/google-api-java-client-services
closed
Synthesis failed for redis
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate redis. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Checking out files: 24% (16109/65361) Checking out files: 25% (16341/65361) Checking out files: 26% (16994/65361) Checking out files: 27% (17648/65361) Checking out files: 28% (18302/65361) Checking out files: 29% (18955/65361) Checking out files: 30% (19609/65361) Checking out files: 31% (20262/65361) Checking out files: 32% (20916/65361) Checking out files: 33% (21570/65361) Checking out files: 34% (22223/65361) Checking out files: 35% (22877/65361) Checking out files: 36% (23530/65361) Checking out files: 37% (24184/65361) Checking out files: 38% (24838/65361) Checking out files: 39% (25491/65361) Checking out files: 40% (26145/65361) Checking out files: 41% (26799/65361) Checking out files: 42% (27452/65361) Checking out files: 43% (28106/65361) Checking out files: 44% (28759/65361) Checking out files: 45% (29413/65361) Checking out files: 46% (30067/65361) Checking out files: 47% (30720/65361) Checking out files: 48% (31374/65361) Checking out files: 49% (32027/65361) Checking out files: 50% (32681/65361) Checking out files: 51% (33335/65361) Checking out files: 52% (33988/65361) Checking out files: 52% (34477/65361) Checking out files: 53% (34642/65361) Checking out files: 54% (35295/65361) Checking out files: 55% (35949/65361) Checking out files: 56% (36603/65361) Checking out files: 57% (37256/65361) Checking out files: 58% (37910/65361) Checking out files: 59% (38563/65361) Checking out files: 60% (39217/65361) Checking out files: 61% (39871/65361) Checking out files: 62% (40524/65361) Checking out files: 63% (41178/65361) Checking out files: 64% (41832/65361) Checking out files: 65% (42485/65361) Checking out files: 66% (43139/65361) Checking out files: 67% (43792/65361) Checking out files: 68% (44446/65361) Checking out files: 69% (45100/65361) Checking out files: 70% (45753/65361) Checking out files: 71% (46407/65361) Checking out files: 72% (47060/65361) Checking out files: 73% (47714/65361) Checking out files: 74% (48368/65361) Checking out files: 75% (49021/65361) Checking out files: 76% (49675/65361) Checking out files: 77% (50328/65361) Checking out files: 78% (50982/65361) Checking out files: 78% (51514/65361) Checking out files: 79% (51636/65361) Checking out files: 80% (52289/65361) Checking out files: 81% (52943/65361) Checking out files: 82% (53597/65361) Checking out files: 83% (54250/65361) Checking out files: 84% (54904/65361) Checking out files: 85% (55557/65361) Checking out files: 86% (56211/65361) Checking out files: 87% (56865/65361) Checking out files: 88% (57518/65361) Checking out files: 89% (58172/65361) Checking out files: 90% (58825/65361) Checking out files: 91% (59479/65361) Checking out files: 92% (60133/65361) Checking out files: 93% (60786/65361) Checking out files: 94% (61440/65361) Checking out files: 95% (62093/65361) Checking out files: 96% (62747/65361) Checking out files: 97% (63401/65361) Checking out files: 98% (64054/65361) Checking out files: 99% (64708/65361) Checking out files: 100% (65361/65361) Checking out files: 100% (65361/65361), done. Switched to branch 'autosynth-redis' Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main last_synth_commit_hash = get_last_metadata_commit(args.metadata_path) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit text=True, File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run with Popen(*popenargs, **kwargs) as process: TypeError: __init__() got an unexpected keyword argument 'text' ``` Google internal developers can see the full log [here](https://sponge/40f694d4-43de-41f0-b993-f4694e4a45de).
1.0
Synthesis failed for redis - Hello! Autosynth couldn't regenerate redis. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Checking out files: 24% (16109/65361) Checking out files: 25% (16341/65361) Checking out files: 26% (16994/65361) Checking out files: 27% (17648/65361) Checking out files: 28% (18302/65361) Checking out files: 29% (18955/65361) Checking out files: 30% (19609/65361) Checking out files: 31% (20262/65361) Checking out files: 32% (20916/65361) Checking out files: 33% (21570/65361) Checking out files: 34% (22223/65361) Checking out files: 35% (22877/65361) Checking out files: 36% (23530/65361) Checking out files: 37% (24184/65361) Checking out files: 38% (24838/65361) Checking out files: 39% (25491/65361) Checking out files: 40% (26145/65361) Checking out files: 41% (26799/65361) Checking out files: 42% (27452/65361) Checking out files: 43% (28106/65361) Checking out files: 44% (28759/65361) Checking out files: 45% (29413/65361) Checking out files: 46% (30067/65361) Checking out files: 47% (30720/65361) Checking out files: 48% (31374/65361) Checking out files: 49% (32027/65361) Checking out files: 50% (32681/65361) Checking out files: 51% (33335/65361) Checking out files: 52% (33988/65361) Checking out files: 52% (34477/65361) Checking out files: 53% (34642/65361) Checking out files: 54% (35295/65361) Checking out files: 55% (35949/65361) Checking out files: 56% (36603/65361) Checking out files: 57% (37256/65361) Checking out files: 58% (37910/65361) Checking out files: 59% (38563/65361) Checking out files: 60% (39217/65361) Checking out files: 61% (39871/65361) Checking out files: 62% (40524/65361) Checking out files: 63% (41178/65361) Checking out files: 64% (41832/65361) Checking out files: 65% (42485/65361) Checking out files: 66% (43139/65361) Checking out files: 67% (43792/65361) Checking out files: 68% (44446/65361) Checking out files: 69% (45100/65361) Checking out files: 70% (45753/65361) Checking out files: 71% (46407/65361) Checking out files: 72% (47060/65361) Checking out files: 73% (47714/65361) Checking out files: 74% (48368/65361) Checking out files: 75% (49021/65361) Checking out files: 76% (49675/65361) Checking out files: 77% (50328/65361) Checking out files: 78% (50982/65361) Checking out files: 78% (51514/65361) Checking out files: 79% (51636/65361) Checking out files: 80% (52289/65361) Checking out files: 81% (52943/65361) Checking out files: 82% (53597/65361) Checking out files: 83% (54250/65361) Checking out files: 84% (54904/65361) Checking out files: 85% (55557/65361) Checking out files: 86% (56211/65361) Checking out files: 87% (56865/65361) Checking out files: 88% (57518/65361) Checking out files: 89% (58172/65361) Checking out files: 90% (58825/65361) Checking out files: 91% (59479/65361) Checking out files: 92% (60133/65361) Checking out files: 93% (60786/65361) Checking out files: 94% (61440/65361) Checking out files: 95% (62093/65361) Checking out files: 96% (62747/65361) Checking out files: 97% (63401/65361) Checking out files: 98% (64054/65361) Checking out files: 99% (64708/65361) Checking out files: 100% (65361/65361) Checking out files: 100% (65361/65361), done. Switched to branch 'autosynth-redis' Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main last_synth_commit_hash = get_last_metadata_commit(args.metadata_path) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit text=True, File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run with Popen(*popenargs, **kwargs) as process: TypeError: __init__() got an unexpected keyword argument 'text' ``` Google internal developers can see the full log [here](https://sponge/40f694d4-43de-41f0-b993-f4694e4a45de).
non_test
synthesis failed for redis hello autosynth couldn t regenerate redis broken heart here s the output from running synth py cloning into working repo checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files done switched to branch autosynth redis traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main last synth commit hash get last metadata commit args metadata path file tmpfs src git autosynth autosynth synth py line in get last metadata commit text true file home kbuilder pyenv versions lib subprocess py line in run with popen popenargs kwargs as process typeerror init got an unexpected keyword argument text google internal developers can see the full log
0
4,190
2,713,029,424
IssuesEvent
2015-04-09 16:56:16
rssidlowski/Pollution_Source_Tracking
https://api.github.com/repos/rssidlowski/Pollution_Source_Tracking
closed
Add CSR layer to PST
COBDev Ready for Testing enhancement moderate priority
Add CSR layer, allow user to toggle on and off and by default have turned off.
1.0
Add CSR layer to PST - Add CSR layer, allow user to toggle on and off and by default have turned off.
test
add csr layer to pst add csr layer allow user to toggle on and off and by default have turned off
1
224,515
17,191,420,796
IssuesEvent
2021-07-16 11:33:13
eficode-academy/SimpleFortuneCookie
https://api.github.com/repos/eficode-academy/SimpleFortuneCookie
opened
Add new steps to push to Kubernetes
documentation enhancement
Problem: getting access to the K8S cluster in AWS/GCP from GH actions. One solution is to make people run self-hosted runners on their instance: Link: https://github.com/actions/runner
1.0
Add new steps to push to Kubernetes - Problem: getting access to the K8S cluster in AWS/GCP from GH actions. One solution is to make people run self-hosted runners on their instance: Link: https://github.com/actions/runner
non_test
add new steps to push to kubernetes problem getting access to the cluster in aws gcp from gh actions one solution is to make people run self hosted runners on their instance link
0
72,702
7,309,292,937
IssuesEvent
2018-02-28 11:15:01
KirillVladimirov/python-messenger
https://api.github.com/repos/KirillVladimirov/python-messenger
closed
lesson_1
doc lesson tests
1. Функционал Первая часть домашнего задания будет заключаться в реализации простого клиент-серверного взаимодействия по протоколу JIM (JSON instant messaging): клиент отправляет запрос серверу; сервер отвечает соответствующим кодом результата. Клиент и сервер должны быть реализованы в виде отдельных скриптов, содержащих соответствующие функции. Функции клиента: - сформировать presence-сообщение; - отправить сообщение серверу; - получить ответ сервера; - разобрать сообщение сервера; - параметры командной строки скрипта client.py <addr> [<port>]: addr - ip-адрес сервера port - tcp-порт на сервере, по умолчанию 7777 Функции сервера: - принимает сообщение клиента; - формирует ответ клиенту; - отправляет ответ клиенту; - имеет параметры командной строки: -p <port> - TCP-порт для работы (по умолчанию использует порт 7777) -a <addr> - IP-адрес для прослушивания (по умолчанию слушает все доступные адреса) 2. Тесты Для всех функций необходимо написать тесты с использованием doctest (небольшие тесты в документации функций), unittest или py.test (в дальнейшем упор будет делаться на библиотеку py.test). Тесты должны быть оформлены в отдельных скриптах с префиксом test_ в имени файла (например, test_client.py). 3. Дополнительно В качестве практики написания тестов напишите тесты для домашних работ курса Python-1.
1.0
lesson_1 - 1. Функционал Первая часть домашнего задания будет заключаться в реализации простого клиент-серверного взаимодействия по протоколу JIM (JSON instant messaging): клиент отправляет запрос серверу; сервер отвечает соответствующим кодом результата. Клиент и сервер должны быть реализованы в виде отдельных скриптов, содержащих соответствующие функции. Функции клиента: - сформировать presence-сообщение; - отправить сообщение серверу; - получить ответ сервера; - разобрать сообщение сервера; - параметры командной строки скрипта client.py <addr> [<port>]: addr - ip-адрес сервера port - tcp-порт на сервере, по умолчанию 7777 Функции сервера: - принимает сообщение клиента; - формирует ответ клиенту; - отправляет ответ клиенту; - имеет параметры командной строки: -p <port> - TCP-порт для работы (по умолчанию использует порт 7777) -a <addr> - IP-адрес для прослушивания (по умолчанию слушает все доступные адреса) 2. Тесты Для всех функций необходимо написать тесты с использованием doctest (небольшие тесты в документации функций), unittest или py.test (в дальнейшем упор будет делаться на библиотеку py.test). Тесты должны быть оформлены в отдельных скриптах с префиксом test_ в имени файла (например, test_client.py). 3. Дополнительно В качестве практики написания тестов напишите тесты для домашних работ курса Python-1.
test
lesson функционал первая часть домашнего задания будет заключаться в реализации простого клиент серверного взаимодействия по протоколу jim json instant messaging клиент отправляет запрос серверу сервер отвечает соответствующим кодом результата клиент и сервер должны быть реализованы в виде отдельных скриптов содержащих соответствующие функции функции клиента сформировать presence сообщение отправить сообщение серверу получить ответ сервера разобрать сообщение сервера параметры командной строки скрипта client py addr ip адрес сервера port tcp порт на сервере по умолчанию функции сервера принимает сообщение клиента формирует ответ клиенту отправляет ответ клиенту имеет параметры командной строки p tcp порт для работы по умолчанию использует порт a ip адрес для прослушивания по умолчанию слушает все доступные адреса тесты для всех функций необходимо написать тесты с использованием doctest небольшие тесты в документации функций unittest или py test в дальнейшем упор будет делаться на библиотеку py test тесты должны быть оформлены в отдельных скриптах с префиксом test в имени файла например test client py дополнительно в качестве практики написания тестов напишите тесты для домашних работ курса python
1
299,812
25,928,576,294
IssuesEvent
2022-12-16 07:51:36
saleor/saleor-dashboard
https://api.github.com/repos/saleor/saleor-dashboard
closed
Cypress test fail: should create customer. TC: SALEOR_1201
tests
**Known bug for versions:** v39: true **Additional Info:** Spec: Tests for customer
1.0
Cypress test fail: should create customer. TC: SALEOR_1201 - **Known bug for versions:** v39: true **Additional Info:** Spec: Tests for customer
test
cypress test fail should create customer tc saleor known bug for versions true additional info spec tests for customer
1
811,207
30,278,884,234
IssuesEvent
2023-07-07 23:06:24
microsoft/PowerToys
https://api.github.com/repos/microsoft/PowerToys
closed
Update notification doesn't open PowerToys
Issue-Bug Product-Settings Area-Runner Priority-1
### Description of the new feature / enhancement When I click the Windows notification that lets me know there's an update for PowerToys I expected the PowerToys main window to at least open so I could click on Update. ### Scenario when this would be used? It speeds up the update process. Now I have to dismiss the notification, then open PowerToys, then click on Update. ### Supporting information Not sure if this is a bug but it happens on both my PCs, on Windows 10 and 11.
1.0
Update notification doesn't open PowerToys - ### Description of the new feature / enhancement When I click the Windows notification that lets me know there's an update for PowerToys I expected the PowerToys main window to at least open so I could click on Update. ### Scenario when this would be used? It speeds up the update process. Now I have to dismiss the notification, then open PowerToys, then click on Update. ### Supporting information Not sure if this is a bug but it happens on both my PCs, on Windows 10 and 11.
non_test
update notification doesn t open powertoys description of the new feature enhancement when i click the windows notification that lets me know there s an update for powertoys i expected the powertoys main window to at least open so i could click on update scenario when this would be used it speeds up the update process now i have to dismiss the notification then open powertoys then click on update supporting information not sure if this is a bug but it happens on both my pcs on windows and
0
4,194
4,876,306,550
IssuesEvent
2016-11-16 12:25:23
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Full framework test csproj referencing netstandard csproj using HttpClient fails at runtime
Infrastructure question
I have a net461 test project that uses a netstandard 1.3 csproj. The netstandard project uses HttpClient. When I run the tests, some with HTTP gets and some with HTTPS, they fail with this: ``` System.IO.FileLoadException : Could not load file or assembly 'System.Security.Cryptography.X509Certificates, Version=4.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a' or one of its dependencies. The located assembly's manifest definition does not match the assembly reference. (Exception from HRESULT: 0x80131040) at System.Net.Http.WinHttpHandler.SendAsync(HttpRequestMessage request, CancellationToken cancellationToken) at System.Net.Http.HttpClientHandler.SendAsync(HttpRequestMessage request, CancellationToken cancellationToken) at System.Net.Http.HttpMessageInvoker.SendAsync(HttpRequestMessage request, CancellationToken cancellationToken) at System.Net.Http.HttpClient.SendAsync(HttpRequestMessage request, HttpCompletionOption completionOption, CancellationToken cancellationToken) at System.Net.Http.HttpClient.GetAsync(Uri requestUri, HttpCompletionOption completionOption, CancellationToken cancellationToken) ``` What can I do about this? There are no warnings or other hints to point me in any direction, and Google searches have left me more confused than ever. Windows 10 x64 AU Microsoft Visual Studio Professional 2015 Version 14.0.25431.01 Update 3 Microsoft .NET Framework Version 4.6.01586 Microsoft .NET Core Tools (Preview 2) 14.1.20907.0 Don't know if this matters, dotnet: Version : 1.0.1 Build : cee57bf6c981237d80aa1631cfe83cb9ba329f12
1.0
Full framework test csproj referencing netstandard csproj using HttpClient fails at runtime - I have a net461 test project that uses a netstandard 1.3 csproj. The netstandard project uses HttpClient. When I run the tests, some with HTTP gets and some with HTTPS, they fail with this: ``` System.IO.FileLoadException : Could not load file or assembly 'System.Security.Cryptography.X509Certificates, Version=4.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a' or one of its dependencies. The located assembly's manifest definition does not match the assembly reference. (Exception from HRESULT: 0x80131040) at System.Net.Http.WinHttpHandler.SendAsync(HttpRequestMessage request, CancellationToken cancellationToken) at System.Net.Http.HttpClientHandler.SendAsync(HttpRequestMessage request, CancellationToken cancellationToken) at System.Net.Http.HttpMessageInvoker.SendAsync(HttpRequestMessage request, CancellationToken cancellationToken) at System.Net.Http.HttpClient.SendAsync(HttpRequestMessage request, HttpCompletionOption completionOption, CancellationToken cancellationToken) at System.Net.Http.HttpClient.GetAsync(Uri requestUri, HttpCompletionOption completionOption, CancellationToken cancellationToken) ``` What can I do about this? There are no warnings or other hints to point me in any direction, and Google searches have left me more confused than ever. Windows 10 x64 AU Microsoft Visual Studio Professional 2015 Version 14.0.25431.01 Update 3 Microsoft .NET Framework Version 4.6.01586 Microsoft .NET Core Tools (Preview 2) 14.1.20907.0 Don't know if this matters, dotnet: Version : 1.0.1 Build : cee57bf6c981237d80aa1631cfe83cb9ba329f12
non_test
full framework test csproj referencing netstandard csproj using httpclient fails at runtime i have a test project that uses a netstandard csproj the netstandard project uses httpclient when i run the tests some with http gets and some with https they fail with this system io fileloadexception could not load file or assembly system security cryptography version culture neutral publickeytoken or one of its dependencies the located assembly s manifest definition does not match the assembly reference exception from hresult at system net http winhttphandler sendasync httprequestmessage request cancellationtoken cancellationtoken at system net http httpclienthandler sendasync httprequestmessage request cancellationtoken cancellationtoken at system net http httpmessageinvoker sendasync httprequestmessage request cancellationtoken cancellationtoken at system net http httpclient sendasync httprequestmessage request httpcompletionoption completionoption cancellationtoken cancellationtoken at system net http httpclient getasync uri requesturi httpcompletionoption completionoption cancellationtoken cancellationtoken what can i do about this there are no warnings or other hints to point me in any direction and google searches have left me more confused than ever windows au microsoft visual studio professional version update microsoft net framework version microsoft net core tools preview don t know if this matters dotnet version build
0
283,804
24,563,241,467
IssuesEvent
2022-10-12 22:47:10
systemd/systemd
https://api.github.com/repos/systemd/systemd
closed
pid1: simultaneously starting multiple services with same image may trigger conflict
bug 🐛 pid1 tests portable
### systemd version the issue has been seen with HEAD ### Used distribution Ubuntu focal ### Linux kernel version used 5.4.0-125-generic ### CPU architectures issue was seen on x86_64 ### Component systemd ### Expected behaviour you didn't see TEST-29-PORTABLE succeeds. ### Unexpected behaviour you saw TEST-29-PORTABLE fails: https://autopkgtest.ubuntu.com/results/autopkgtest-focal-upstream-systemd-ci-systemd-ci/focal/amd64/s/systemd-upstream/20220908_232004_2ec87@/log.gz See [TEST-29-PORTABLE.journal.tar.gz](https://github.com/systemd/systemd/files/9535214/TEST-29-PORTABLE.journal.tar.gz). ### Steps to reproduce the problem Run the test. ### Additional program output to the terminal or log subsystem illustrating the issue ``` Sep 09 07:33:18 systemd[452]: Opened '/usr/share/minimal_1.raw' in O_RDONLY access mode, with O_DIRECT enabled. Sep 09 07:33:18 systemd[452]: Successfully acquired /dev/loop0, devno=7:0, nr=0, diskseq=0 Sep 09 07:33:18 systemd[452]: Allocating context for crypt device /usr/share/minimal_1.verity. Sep 09 07:33:18 systemd[452]: Trying to open and read device /usr/share/minimal_1.verity with direct-io. Sep 09 07:33:18 systemd[452]: Crypto backend (OpenSSL 1.1.1f 31 Mar 2020) initialized in cryptsetup library version 2.2.2. Sep 09 07:33:18 systemd[452]: Detected kernel Linux 5.4.0-125-generic x86_64. Sep 09 07:33:18 systemd[452]: Reading VERITY header of size 512 on device /usr/share/minimal_1.verity, offset 0. Sep 09 07:33:18 systemd[454]: Opened '/usr/share/minimal_1.raw' in O_RDONLY access mode, with O_DIRECT enabled. Sep 09 07:33:18 systemd[452]: Setting ciphertext data device to /dev/loop0. Sep 09 07:33:18 systemd[452]: Trying to open and read device /dev/loop0 with direct-io. Sep 09 07:33:18 systemd[452]: Activating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity by volume key. Sep 09 07:33:18 systemd[452]: Trying to activate VERITY device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity using hash sha256. Sep 09 07:33:18 systemd[452]: Allocating a free loop device. Sep 09 07:33:18 systemd[454]: Successfully acquired /dev/loop1, devno=7:1, nr=1, diskseq=0 Sep 09 07:33:18 systemd[452]: Trying to open and read device /dev/loop5 with direct-io. Sep 09 07:33:18 systemd[454]: Allocating context for crypt device /usr/share/minimal_1.verity. Sep 09 07:33:18 systemd[454]: Trying to open and read device /usr/share/minimal_1.verity with direct-io. Sep 09 07:33:18 systemd[454]: Trying to load VERITY crypt type from device /usr/share/minimal_1.verity. Sep 09 07:33:18 systemd[454]: Reading VERITY header of size 512 on device /usr/share/minimal_1.verity, offset 0. Sep 09 07:33:18 systemd[454]: Setting ciphertext data device to /dev/loop1. Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop1 with direct-io. Sep 09 07:33:18 systemd[454]: Activating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity by volume key. Sep 09 07:33:18 systemd[454]: Device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity already exists. <-- in LOG_ERR Sep 09 07:33:18 systemd[454]: Allocating crypt device context by device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity. Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- What?? Sep 09 07:33:18 systemd[454]: Allocating context for crypt device /dev/loop0. <-- Hm?? Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- Huh?? Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- again?? Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop5 with direct-io. Sep 09 07:33:18 systemd[454]: Error opening verity device, crypt_volume_key_get failed: Invalid argument Sep 09 07:33:18 systemd[454]: Releasing crypt device /dev/loop5 context. Sep 09 07:33:18 systemd[454]: Allocating context for crypt device /usr/share/minimal_1.verity. Sep 09 07:33:18 systemd[454]: Trying to open and read device /usr/share/minimal_1.verity with direct-io. Sep 09 07:33:18 systemd[454]: Trying to load VERITY crypt type from device /usr/share/minimal_1.verity. Sep 09 07:33:18 systemd[454]: Reading VERITY header of size 512 on device /usr/share/minimal_1.verity, offset 0. Sep 09 07:33:18 systemd[454]: Setting ciphertext data device to /dev/loop1. Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop1 with direct-io. Sep 09 07:33:18 systemd[454]: Activating volume loop1-verity by volume key. Sep 09 07:33:18 systemd[454]: Trying to activate VERITY device loop1-verity using hash sha256. Sep 09 07:33:18 systemd[454]: Allocating a free loop device. Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop6 with direct-io. Sep 09 07:33:18 systemd[454]: Calculated device size is 5928 sectors (RO), offset 0. (udev processing many events) Sep 09 07:33:21 kernel: /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity: Can't open blockdev <-- LOG_ERR (snip) Sep 09 07:33:21 systemd[452]: Udev cookie 0xd4dedbd (semid 10) destroyed Sep 09 07:33:21 systemd[452]: Verity volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity status is V. Sep 09 07:33:21 systemd[454]: Udev cookie 0xd4d0212 (semid 11) destroyed Sep 09 07:33:21 systemd[454]: Verity volume loop1-verity status is V. Sep 09 07:33:21 systemd[454]: Deactivating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity. <-- ?? Sep 09 07:33:21 systemd[454]: Allocating crypt device context by device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity. Sep 09 07:33:21 systemd[452]: Probed fstype 'squashfs' on partition /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity. Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- Oh, again?? Sep 09 07:33:21 systemd[454]: Allocating context for crypt device /dev/loop0. Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop5 with direct-io. Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. Sep 09 07:33:21 systemd[454]: Verity volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity status is V. Sep 09 07:33:21 systemd[454]: dm remove dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity [ opencount flush retryremove deferredremove ] [16384] (*1) <-- Wait, wait, it is still used by PID=452... Sep 09 07:33:21 systemd[452]: Mounting /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity (squashfs) on /run/systemd/unit-root (MS_RDONLY|MS_NODEV "")... Sep 09 07:33:21 systemd-udevd[281]: dm-0: Device is queued (SEQNUM=3921, ACTION=remove) Sep 09 07:33:21 systemd[452]: Failed to mount /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity (type squashfs) on /run/systemd/unit-root (MS_RDONLY|MS_NODEV ""): No such file or directory Sep 09 07:33:21 systemd[452]: Failed to mount root image: No such file or directory Sep 09 07:33:21 systemd[452]: Deactivating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity. Sep 09 07:33:21 systemd[452]: Device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity is not active. Sep 09 07:33:21 systemd[452]: Failed to deactivate encrypted partition dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity <-- yeah, it is already removed by PID=454... Sep 09 07:33:21 systemd[452]: Releasing crypt device /usr/share/minimal_1.verity context. Sep 09 07:33:21 systemd[452]: Releasing device-mapper backend. Sep 09 07:33:21 systemd[452]: Closing read only fd for /usr/share/minimal_1.verity. Sep 09 07:33:21 systemd[452]: Closed loop /dev/loop5 (/usr/share/minimal_1.verity). Sep 09 07:33:21 systemd[452]: minimal-app0-bar.service: Failed to set up mount namespacing: No such file or directory Sep 09 07:33:21 systemd[452]: minimal-app0-bar.service: Failed at step NAMESPACE spawning cat: No such file or directory Sep 09 07:33:21 systemd[1]: Received SIGCHLD from PID 452 ((cat)). Sep 09 07:33:21 systemd[1]: Child 452 ((cat)) died (code=exited, status=226/NAMESPACE) Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Child 452 belongs to minimal-app0-bar.service. Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Control process exited, code=exited, status=226/NAMESPACE Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Got final SIGCHLD for state start-pre. Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Failed with result 'exit-code'. Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Service will not restart (restart setting) Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Changed start-pre -> failed Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Job 486 minimal-app0-bar.service/start finished, result=failed Sep 09 07:33:21 systemd[1]: Failed to start minimal-app0-bar.service. ```
1.0
pid1: simultaneously starting multiple services with same image may trigger conflict - ### systemd version the issue has been seen with HEAD ### Used distribution Ubuntu focal ### Linux kernel version used 5.4.0-125-generic ### CPU architectures issue was seen on x86_64 ### Component systemd ### Expected behaviour you didn't see TEST-29-PORTABLE succeeds. ### Unexpected behaviour you saw TEST-29-PORTABLE fails: https://autopkgtest.ubuntu.com/results/autopkgtest-focal-upstream-systemd-ci-systemd-ci/focal/amd64/s/systemd-upstream/20220908_232004_2ec87@/log.gz See [TEST-29-PORTABLE.journal.tar.gz](https://github.com/systemd/systemd/files/9535214/TEST-29-PORTABLE.journal.tar.gz). ### Steps to reproduce the problem Run the test. ### Additional program output to the terminal or log subsystem illustrating the issue ``` Sep 09 07:33:18 systemd[452]: Opened '/usr/share/minimal_1.raw' in O_RDONLY access mode, with O_DIRECT enabled. Sep 09 07:33:18 systemd[452]: Successfully acquired /dev/loop0, devno=7:0, nr=0, diskseq=0 Sep 09 07:33:18 systemd[452]: Allocating context for crypt device /usr/share/minimal_1.verity. Sep 09 07:33:18 systemd[452]: Trying to open and read device /usr/share/minimal_1.verity with direct-io. Sep 09 07:33:18 systemd[452]: Crypto backend (OpenSSL 1.1.1f 31 Mar 2020) initialized in cryptsetup library version 2.2.2. Sep 09 07:33:18 systemd[452]: Detected kernel Linux 5.4.0-125-generic x86_64. Sep 09 07:33:18 systemd[452]: Reading VERITY header of size 512 on device /usr/share/minimal_1.verity, offset 0. Sep 09 07:33:18 systemd[454]: Opened '/usr/share/minimal_1.raw' in O_RDONLY access mode, with O_DIRECT enabled. Sep 09 07:33:18 systemd[452]: Setting ciphertext data device to /dev/loop0. Sep 09 07:33:18 systemd[452]: Trying to open and read device /dev/loop0 with direct-io. Sep 09 07:33:18 systemd[452]: Activating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity by volume key. Sep 09 07:33:18 systemd[452]: Trying to activate VERITY device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity using hash sha256. Sep 09 07:33:18 systemd[452]: Allocating a free loop device. Sep 09 07:33:18 systemd[454]: Successfully acquired /dev/loop1, devno=7:1, nr=1, diskseq=0 Sep 09 07:33:18 systemd[452]: Trying to open and read device /dev/loop5 with direct-io. Sep 09 07:33:18 systemd[454]: Allocating context for crypt device /usr/share/minimal_1.verity. Sep 09 07:33:18 systemd[454]: Trying to open and read device /usr/share/minimal_1.verity with direct-io. Sep 09 07:33:18 systemd[454]: Trying to load VERITY crypt type from device /usr/share/minimal_1.verity. Sep 09 07:33:18 systemd[454]: Reading VERITY header of size 512 on device /usr/share/minimal_1.verity, offset 0. Sep 09 07:33:18 systemd[454]: Setting ciphertext data device to /dev/loop1. Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop1 with direct-io. Sep 09 07:33:18 systemd[454]: Activating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity by volume key. Sep 09 07:33:18 systemd[454]: Device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity already exists. <-- in LOG_ERR Sep 09 07:33:18 systemd[454]: Allocating crypt device context by device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity. Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- What?? Sep 09 07:33:18 systemd[454]: Allocating context for crypt device /dev/loop0. <-- Hm?? Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- Huh?? Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- again?? Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop5 with direct-io. Sep 09 07:33:18 systemd[454]: Error opening verity device, crypt_volume_key_get failed: Invalid argument Sep 09 07:33:18 systemd[454]: Releasing crypt device /dev/loop5 context. Sep 09 07:33:18 systemd[454]: Allocating context for crypt device /usr/share/minimal_1.verity. Sep 09 07:33:18 systemd[454]: Trying to open and read device /usr/share/minimal_1.verity with direct-io. Sep 09 07:33:18 systemd[454]: Trying to load VERITY crypt type from device /usr/share/minimal_1.verity. Sep 09 07:33:18 systemd[454]: Reading VERITY header of size 512 on device /usr/share/minimal_1.verity, offset 0. Sep 09 07:33:18 systemd[454]: Setting ciphertext data device to /dev/loop1. Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop1 with direct-io. Sep 09 07:33:18 systemd[454]: Activating volume loop1-verity by volume key. Sep 09 07:33:18 systemd[454]: Trying to activate VERITY device loop1-verity using hash sha256. Sep 09 07:33:18 systemd[454]: Allocating a free loop device. Sep 09 07:33:18 systemd[454]: Trying to open and read device /dev/loop6 with direct-io. Sep 09 07:33:18 systemd[454]: Calculated device size is 5928 sectors (RO), offset 0. (udev processing many events) Sep 09 07:33:21 kernel: /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity: Can't open blockdev <-- LOG_ERR (snip) Sep 09 07:33:21 systemd[452]: Udev cookie 0xd4dedbd (semid 10) destroyed Sep 09 07:33:21 systemd[452]: Verity volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity status is V. Sep 09 07:33:21 systemd[454]: Udev cookie 0xd4d0212 (semid 11) destroyed Sep 09 07:33:21 systemd[454]: Verity volume loop1-verity status is V. Sep 09 07:33:21 systemd[454]: Deactivating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity. <-- ?? Sep 09 07:33:21 systemd[454]: Allocating crypt device context by device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity. Sep 09 07:33:21 systemd[452]: Probed fstype 'squashfs' on partition /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity. Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. <-- Oh, again?? Sep 09 07:33:21 systemd[454]: Allocating context for crypt device /dev/loop0. Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop5 with direct-io. Sep 09 07:33:21 systemd[454]: Trying to open and read device /dev/loop0 with direct-io. Sep 09 07:33:21 systemd[454]: Verity volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity status is V. Sep 09 07:33:21 systemd[454]: dm remove dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity [ opencount flush retryremove deferredremove ] [16384] (*1) <-- Wait, wait, it is still used by PID=452... Sep 09 07:33:21 systemd[452]: Mounting /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity (squashfs) on /run/systemd/unit-root (MS_RDONLY|MS_NODEV "")... Sep 09 07:33:21 systemd-udevd[281]: dm-0: Device is queued (SEQNUM=3921, ACTION=remove) Sep 09 07:33:21 systemd[452]: Failed to mount /dev/mapper/dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity (type squashfs) on /run/systemd/unit-root (MS_RDONLY|MS_NODEV ""): No such file or directory Sep 09 07:33:21 systemd[452]: Failed to mount root image: No such file or directory Sep 09 07:33:21 systemd[452]: Deactivating volume dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity. Sep 09 07:33:21 systemd[452]: Device dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity is not active. Sep 09 07:33:21 systemd[452]: Failed to deactivate encrypted partition dfbc0b7c6019e90420b41209c8f54b867537b6cbf64de5e7fafa098947484c90-verity <-- yeah, it is already removed by PID=454... Sep 09 07:33:21 systemd[452]: Releasing crypt device /usr/share/minimal_1.verity context. Sep 09 07:33:21 systemd[452]: Releasing device-mapper backend. Sep 09 07:33:21 systemd[452]: Closing read only fd for /usr/share/minimal_1.verity. Sep 09 07:33:21 systemd[452]: Closed loop /dev/loop5 (/usr/share/minimal_1.verity). Sep 09 07:33:21 systemd[452]: minimal-app0-bar.service: Failed to set up mount namespacing: No such file or directory Sep 09 07:33:21 systemd[452]: minimal-app0-bar.service: Failed at step NAMESPACE spawning cat: No such file or directory Sep 09 07:33:21 systemd[1]: Received SIGCHLD from PID 452 ((cat)). Sep 09 07:33:21 systemd[1]: Child 452 ((cat)) died (code=exited, status=226/NAMESPACE) Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Child 452 belongs to minimal-app0-bar.service. Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Control process exited, code=exited, status=226/NAMESPACE Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Got final SIGCHLD for state start-pre. Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Failed with result 'exit-code'. Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Service will not restart (restart setting) Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Changed start-pre -> failed Sep 09 07:33:21 systemd[1]: minimal-app0-bar.service: Job 486 minimal-app0-bar.service/start finished, result=failed Sep 09 07:33:21 systemd[1]: Failed to start minimal-app0-bar.service. ```
test
simultaneously starting multiple services with same image may trigger conflict systemd version the issue has been seen with head used distribution ubuntu focal linux kernel version used generic cpu architectures issue was seen on component systemd expected behaviour you didn t see test portable succeeds unexpected behaviour you saw test portable fails see steps to reproduce the problem run the test additional program output to the terminal or log subsystem illustrating the issue sep systemd opened usr share minimal raw in o rdonly access mode with o direct enabled sep systemd successfully acquired dev devno nr diskseq sep systemd allocating context for crypt device usr share minimal verity sep systemd trying to open and read device usr share minimal verity with direct io sep systemd crypto backend openssl mar initialized in cryptsetup library version sep systemd detected kernel linux generic sep systemd reading verity header of size on device usr share minimal verity offset sep systemd opened usr share minimal raw in o rdonly access mode with o direct enabled sep systemd setting ciphertext data device to dev sep systemd trying to open and read device dev with direct io sep systemd activating volume verity by volume key sep systemd trying to activate verity device verity using hash sep systemd allocating a free loop device sep systemd successfully acquired dev devno nr diskseq sep systemd trying to open and read device dev with direct io sep systemd allocating context for crypt device usr share minimal verity sep systemd trying to open and read device usr share minimal verity with direct io sep systemd trying to load verity crypt type from device usr share minimal verity sep systemd reading verity header of size on device usr share minimal verity offset sep systemd setting ciphertext data device to dev sep systemd trying to open and read device dev with direct io sep systemd activating volume verity by volume key sep systemd device verity already exists in log err sep systemd allocating crypt device context by device verity sep systemd trying to open and read device dev with direct io what sep systemd allocating context for crypt device dev hm sep systemd trying to open and read device dev with direct io huh sep systemd trying to open and read device dev with direct io again sep systemd trying to open and read device dev with direct io sep systemd error opening verity device crypt volume key get failed invalid argument sep systemd releasing crypt device dev context sep systemd allocating context for crypt device usr share minimal verity sep systemd trying to open and read device usr share minimal verity with direct io sep systemd trying to load verity crypt type from device usr share minimal verity sep systemd reading verity header of size on device usr share minimal verity offset sep systemd setting ciphertext data device to dev sep systemd trying to open and read device dev with direct io sep systemd activating volume verity by volume key sep systemd trying to activate verity device verity using hash sep systemd allocating a free loop device sep systemd trying to open and read device dev with direct io sep systemd calculated device size is sectors ro offset udev processing many events sep kernel dev mapper verity can t open blockdev log err snip sep systemd udev cookie semid destroyed sep systemd verity volume verity status is v sep systemd udev cookie semid destroyed sep systemd verity volume verity status is v sep systemd deactivating volume verity sep systemd allocating crypt device context by device verity sep systemd probed fstype squashfs on partition dev mapper verity sep systemd trying to open and read device dev with direct io oh again sep systemd allocating context for crypt device dev sep systemd trying to open and read device dev with direct io sep systemd trying to open and read device dev with direct io sep systemd trying to open and read device dev with direct io sep systemd trying to open and read device dev with direct io sep systemd verity volume verity status is v sep systemd dm remove verity wait wait it is still used by pid sep systemd mounting dev mapper verity squashfs on run systemd unit root ms rdonly ms nodev sep systemd udevd dm device is queued seqnum action remove sep systemd failed to mount dev mapper verity type squashfs on run systemd unit root ms rdonly ms nodev no such file or directory sep systemd failed to mount root image no such file or directory sep systemd deactivating volume verity sep systemd device verity is not active sep systemd failed to deactivate encrypted partition verity yeah it is already removed by pid sep systemd releasing crypt device usr share minimal verity context sep systemd releasing device mapper backend sep systemd closing read only fd for usr share minimal verity sep systemd closed loop dev usr share minimal verity sep systemd minimal bar service failed to set up mount namespacing no such file or directory sep systemd minimal bar service failed at step namespace spawning cat no such file or directory sep systemd received sigchld from pid cat sep systemd child cat died code exited status namespace sep systemd minimal bar service child belongs to minimal bar service sep systemd minimal bar service control process exited code exited status namespace sep systemd minimal bar service got final sigchld for state start pre sep systemd minimal bar service failed with result exit code sep systemd minimal bar service service will not restart restart setting sep systemd minimal bar service changed start pre failed sep systemd minimal bar service job minimal bar service start finished result failed sep systemd failed to start minimal bar service
1
654,089
21,637,123,199
IssuesEvent
2022-05-05 15:07:48
redwoodjs/redwood
https://api.github.com/repos/redwoodjs/redwood
closed
Scary and duplicated output when generating sdl
help wanted v1/priority
I just created a new RW project and upgraded to 44.1-canary.26 First thing I did was to add this to `prisma.schema` ``` model Product { id String @id @default(uuid()) createdAt DateTime @default(now()) updatedAt DateTime @default(now()) @updatedAt name String } ``` And then when I try to generate the sdl for it I get this output: ``` $ yarn rw g sdl --crud product yarn run v1.22.17 $ /Users/tobbe/tmp/uniquets/node_modules/.bin/rw g sdl --crud product (node:34268) ExperimentalWarning: stream/web is an experimental feature. This feature could change at any time (Use `node --trace-warnings ...` to show where the warning was created) ✔ Generating SDL files... ✔ Successfully wrote file `./api/src/graphql/products.sdl.ts` ✔ Successfully wrote file `./api/src/services/products/products.scenarios.ts` ✔ Successfully wrote file `./api/src/services/products/products.test.ts` ✔ Successfully wrote file `./api/src/services/products/products.ts` ⠹ Generating types ... Unable to find any GraphQL type definitions for the following pointers: - ./web/src/**/!(*.d).{ts,tsx,js,jsx} Unable to find any GraphQL type definitions for the following pointers: ✔ Generating SDL files... ✔ Successfully wrote file `./api/src/graphql/products.sdl.ts` ✔ Successfully wrote file `./api/src/services/products/products.scenarios.ts` ✔ Successfully wrote file `./api/src/services/products/products.test.ts` ✔ Successfully wrote file `./api/src/services/products/products.ts` ✔ Generating types ... ✨ Done in 2.68s. ``` The "Unable to find any GraphQL..." message looks scary. Is it an error? A warning? Also, why does it say "Successfully wrote..." for all files twice?
1.0
Scary and duplicated output when generating sdl - I just created a new RW project and upgraded to 44.1-canary.26 First thing I did was to add this to `prisma.schema` ``` model Product { id String @id @default(uuid()) createdAt DateTime @default(now()) updatedAt DateTime @default(now()) @updatedAt name String } ``` And then when I try to generate the sdl for it I get this output: ``` $ yarn rw g sdl --crud product yarn run v1.22.17 $ /Users/tobbe/tmp/uniquets/node_modules/.bin/rw g sdl --crud product (node:34268) ExperimentalWarning: stream/web is an experimental feature. This feature could change at any time (Use `node --trace-warnings ...` to show where the warning was created) ✔ Generating SDL files... ✔ Successfully wrote file `./api/src/graphql/products.sdl.ts` ✔ Successfully wrote file `./api/src/services/products/products.scenarios.ts` ✔ Successfully wrote file `./api/src/services/products/products.test.ts` ✔ Successfully wrote file `./api/src/services/products/products.ts` ⠹ Generating types ... Unable to find any GraphQL type definitions for the following pointers: - ./web/src/**/!(*.d).{ts,tsx,js,jsx} Unable to find any GraphQL type definitions for the following pointers: ✔ Generating SDL files... ✔ Successfully wrote file `./api/src/graphql/products.sdl.ts` ✔ Successfully wrote file `./api/src/services/products/products.scenarios.ts` ✔ Successfully wrote file `./api/src/services/products/products.test.ts` ✔ Successfully wrote file `./api/src/services/products/products.ts` ✔ Generating types ... ✨ Done in 2.68s. ``` The "Unable to find any GraphQL..." message looks scary. Is it an error? A warning? Also, why does it say "Successfully wrote..." for all files twice?
non_test
scary and duplicated output when generating sdl i just created a new rw project and upgraded to canary first thing i did was to add this to prisma schema model product id string id default uuid createdat datetime default now updatedat datetime default now updatedat name string and then when i try to generate the sdl for it i get this output yarn rw g sdl crud product yarn run users tobbe tmp uniquets node modules bin rw g sdl crud product node experimentalwarning stream web is an experimental feature this feature could change at any time use node trace warnings to show where the warning was created ✔ generating sdl files ✔ successfully wrote file api src graphql products sdl ts ✔ successfully wrote file api src services products products scenarios ts ✔ successfully wrote file api src services products products test ts ✔ successfully wrote file api src services products products ts ⠹ generating types unable to find any graphql type definitions for the following pointers web src d ts tsx js jsx unable to find any graphql type definitions for the following pointers ✔ generating sdl files ✔ successfully wrote file api src graphql products sdl ts ✔ successfully wrote file api src services products products scenarios ts ✔ successfully wrote file api src services products products test ts ✔ successfully wrote file api src services products products ts ✔ generating types ✨ done in the unable to find any graphql message looks scary is it an error a warning also why does it say successfully wrote for all files twice
0
373,775
26,084,208,625
IssuesEvent
2022-12-25 21:52:11
tfiers/PkgGraph.jl
https://api.github.com/repos/tfiers/PkgGraph.jl
closed
readme-in-docs: preprocess further
documentation
- [x] Global custom subs (`replace`) - [ ] rm `<details>` surround (regex → subs w/ matched content) - [x] replace "Reference section of [the docs]" with link to Internals - [x] rm docs link in H1 - [ ] transform CommonMark.jl output - [x] specify sections (w/ regex mayb) to keep / exclude Exclude: Dev, Versions (instead, #22 :))
1.0
readme-in-docs: preprocess further - - [x] Global custom subs (`replace`) - [ ] rm `<details>` surround (regex → subs w/ matched content) - [x] replace "Reference section of [the docs]" with link to Internals - [x] rm docs link in H1 - [ ] transform CommonMark.jl output - [x] specify sections (w/ regex mayb) to keep / exclude Exclude: Dev, Versions (instead, #22 :))
non_test
readme in docs preprocess further global custom subs replace rm surround regex → subs w matched content replace reference section of with link to internals rm docs link in transform commonmark jl output specify sections w regex mayb to keep exclude exclude dev versions instead
0
92,672
8,375,540,611
IssuesEvent
2018-10-05 16:44:37
blackbaud/skyux2
https://api.github.com/repos/blackbaud/skyux2
reopened
Uncaught errors in logs when running npm run watch
Priority: Critical Status: Ready to merge Type: Bug testing
When running `npm run watch`, we're seeing a lot of uncaught errors in the logs. For example: ``` ERROR: 'ERROR', TypeError{line: 49079, column: 17082, sourceURL: 'http://localhost:9876/absolute/home/travis/build/blackbaud/skyux2/config/utils/spec-bundle.js?b65bce9e10e00c33526065336acf99271e9affd8', ngDebugContext: DebugContext_{view: Object{def: ..., parent: ..., viewContainerParent: ..., parentNodeDef: ..., context: ..., component: ..., nodes: ..., state: ..., root: ..., renderer: ..., oldValues: ..., disposables: ...}, nodeIndex: 1, nodeDef: Object{index: ..., parent: ..., renderParent: ..., bindingIndex: ..., outputIndex: ..., flags: ..., childFlags: ..., directChildFlags: ..., childMatchedQueries: ..., matchedQueries: ..., matchedQueryIds: ..., references: ..., ngContentIndex: ..., childCount: ..., bindings: ..., bindingFlags: ..., outputs: ..., element: ..., provider: ..., text: ..., query: ..., ngContent: ...}, elDef: Object{index: ..., parent: ..., renderParent: ..., bindingIndex: ..., outputIndex: ..., flags: ..., childFlags: ..., directChildFlags: ..., childMatchedQueries: ..., matchedQueries: ..., matchedQueryIds: ..., references: ..., ngContentIndex: ..., childCount: ..., bindings: ..., bindingFlags: ..., outputs: ..., element: ..., provider: ..., text: ..., query: ..., ngContent: ...}, elView: Object{def: ..., parent: ..., viewContainerParent: ..., parentNodeDef: ..., context: ..., component: ..., nodes: ..., state: ..., root: ..., renderer: ..., oldValues: ..., disposables: ...}}, ngErrorLogger: function () { ... }} ``` Refer to log to see more: https://travis-ci.org/blackbaud/skyux2/jobs/431683334
1.0
Uncaught errors in logs when running npm run watch - When running `npm run watch`, we're seeing a lot of uncaught errors in the logs. For example: ``` ERROR: 'ERROR', TypeError{line: 49079, column: 17082, sourceURL: 'http://localhost:9876/absolute/home/travis/build/blackbaud/skyux2/config/utils/spec-bundle.js?b65bce9e10e00c33526065336acf99271e9affd8', ngDebugContext: DebugContext_{view: Object{def: ..., parent: ..., viewContainerParent: ..., parentNodeDef: ..., context: ..., component: ..., nodes: ..., state: ..., root: ..., renderer: ..., oldValues: ..., disposables: ...}, nodeIndex: 1, nodeDef: Object{index: ..., parent: ..., renderParent: ..., bindingIndex: ..., outputIndex: ..., flags: ..., childFlags: ..., directChildFlags: ..., childMatchedQueries: ..., matchedQueries: ..., matchedQueryIds: ..., references: ..., ngContentIndex: ..., childCount: ..., bindings: ..., bindingFlags: ..., outputs: ..., element: ..., provider: ..., text: ..., query: ..., ngContent: ...}, elDef: Object{index: ..., parent: ..., renderParent: ..., bindingIndex: ..., outputIndex: ..., flags: ..., childFlags: ..., directChildFlags: ..., childMatchedQueries: ..., matchedQueries: ..., matchedQueryIds: ..., references: ..., ngContentIndex: ..., childCount: ..., bindings: ..., bindingFlags: ..., outputs: ..., element: ..., provider: ..., text: ..., query: ..., ngContent: ...}, elView: Object{def: ..., parent: ..., viewContainerParent: ..., parentNodeDef: ..., context: ..., component: ..., nodes: ..., state: ..., root: ..., renderer: ..., oldValues: ..., disposables: ...}}, ngErrorLogger: function () { ... }} ``` Refer to log to see more: https://travis-ci.org/blackbaud/skyux2/jobs/431683334
test
uncaught errors in logs when running npm run watch when running npm run watch we re seeing a lot of uncaught errors in the logs for example error error typeerror line column sourceurl ngdebugcontext debugcontext view object def parent viewcontainerparent parentnodedef context component nodes state root renderer oldvalues disposables nodeindex nodedef object index parent renderparent bindingindex outputindex flags childflags directchildflags childmatchedqueries matchedqueries matchedqueryids references ngcontentindex childcount bindings bindingflags outputs element provider text query ngcontent eldef object index parent renderparent bindingindex outputindex flags childflags directchildflags childmatchedqueries matchedqueries matchedqueryids references ngcontentindex childcount bindings bindingflags outputs element provider text query ngcontent elview object def parent viewcontainerparent parentnodedef context component nodes state root renderer oldvalues disposables ngerrorlogger function refer to log to see more
1
64,966
6,927,426,852
IssuesEvent
2017-11-30 22:49:48
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
System.Net.Security.Tests.dll FailFast on Debian
area-System.Net.Security os-linux test-run-core
This comes just after "starting KDC" from `src\System.Net.Security\tests\Scripts\Unix\setup-kdc.sh` https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20170601.02/workItem/System.Net.Security.Tests/wilogs ``` Starting KDC.. FailFast: Invalid context passed to SafeDeleteNegoContext at System.Diagnostics.Debug.Assert(Boolean condition, String message, String detailMessage) at System.Net.Security.SafeDeleteNegoContext.SetGssContext(SafeGssContextHandle context) in /root/corefx/src/Common/src/System/Net/Security/Unix/SafeDeleteNegoContext.cs:line 52 at System.Net.Security.NegotiateStreamPal.EstablishSecurityContext(SafeFreeNegoCredentials credential, SafeDeleteContext& context, String targetName, ContextFlagsPal inFlags, SecurityBuffer inputBuffer, SecurityBuffer outputBuffer, ContextFlagsPal& outFlags) in /root/corefx/src/Common/src/System/Net/Security/NegotiateStreamPal.Unix.cs:line 197 at System.Net.Security.NegotiateStreamPal.InitializeSecurityContext(SafeFreeCredentials credentialsHandle, SafeDeleteContext& securityContext, String spn, ContextFlagsPal requestedContextFlags, SecurityBuffer[] inSecurityBufferArray, SecurityBuffer outSecurityBuffer, ContextFlagsPal& contextFlags) in /root/corefx/src/Common/src/System/Net/Security/NegotiateStreamPal.Unix.cs:line 240 at System.Net.NTAuthentication.GetOutgoingBlob(Byte[] incomingBlob, Boolean throwOnError, SecurityStatusPal& statusCode) in /root/corefx/src/Common/src/System/Net/NTAuthentication.Common.cs:line 243 at System.Net.Security.NegoState.GetOutgoingBlob(Byte[] incomingBlob, Exception& e) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 778 at System.Net.Security.NegoState.StartSendBlob(Byte[] message, LazyAsyncResult lazyResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 444 at System.Net.Security.NegoState.CheckCompletionBeforeNextSend(Byte[] message, LazyAsyncResult lazyResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 648 at System.Net.Security.NegoState.ProcessReceivedBlob(Byte[] message, LazyAsyncResult lazyResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 624 at System.Net.Security.NegoState.ReadCallback(IAsyncResult transportResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 756 at System.Net.LazyAsyncResult.Complete(IntPtr userToken) in /root/corefx/src/Common/src/System/Net/LazyAsyncResult.cs:line 428 at System.Net.LazyAsyncResult.ProtectedInvokeCallback(Object result, IntPtr userToken) in /root/corefx/src/Common/src/System/Net/LazyAsyncResult.cs:line 382 at System.Net.LazyAsyncResult.InvokeCallback(Object result) in /root/corefx/src/Common/src/System/Net/LazyAsyncResult.cs:line 389 at System.Net.StreamFramer.ReadFrameComplete(IAsyncResult transportResult) in /root/corefx/src/System.Net.Security/src/System/Net/StreamFramer.cs:line 270 at System.Net.StreamFramer.ReadFrameCallback(IAsyncResult transportResult) in /root/corefx/src/System.Net.Security/src/System/Net/StreamFramer.cs:line 168 at System.Threading.Tasks.TaskToApm.<>c__DisplayClass3_0.<InvokeCallbackWhenTaskCompletes>b__0() in /root/corefx/src/Common/src/System/Threading/Tasks/TaskToApm.cs:line 133 at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state) at System.Threading.Tasks.AwaitTaskContinuation.RunCallback(ContextCallback callback, Object state, Task& currentTask) at System.Threading.Tasks.Task.RunContinuations(Object continuationObject) at System.Threading.Tasks.Task.ExecuteWithThreadLocal(Task& currentTaskSlot) at System.Threading.ThreadPoolWorkQueue.Dispatch() at System.Environment.FailFast(System.String, System.Exception) at System.Net.Security.SafeDeleteNegoContext.SetGssContext(Microsoft.Win32.SafeHandles.SafeGssContextHandle) at System.Net.Security.NegotiateStreamPal.EstablishSecurityContext(System.Net.Security.SafeFreeNegoCredentials, System.Net.Security.SafeDeleteContext ByRef, System.String, System.Net.ContextFlagsPal, System.Net.Security.SecurityBuffer, System.Net.Security.SecurityBuffer, System.Net.ContextFlagsPal ByRef) at System.Net.Security.NegotiateStreamPal.InitializeSecurityContext(System.Net.Security.SafeFreeCredentials, System.Net.Security.SafeDeleteContext ByRef, System.String, System.Net.ContextFlagsPal, System.Net.Security.SecurityBuffer[], System.Net.Security.SecurityBuffer, System.Net.ContextFlagsPal ByRef) at System.Net.NTAuthentication.GetOutgoingBlob(Byte[], Boolean, System.Net.SecurityStatusPal ByRef) at System.Net.Security.NegoState.GetOutgoingBlob(Byte[], System.Exception ByRef) at System.Net.Security.NegoState.StartSendBlob(Byte[], System.Net.LazyAsyncResult) at System.Net.Security.NegoState.CheckCompletionBeforeNextSend(Byte[], System.Net.LazyAsyncResult) at System.Net.Security.NegoState.ProcessReceivedBlob(Byte[], System.Net.LazyAsyncResult) at System.Net.Security.NegoState.ReadCallback(System.IAsyncResult) at System.Net.LazyAsyncResult.Complete(IntPtr) at System.Net.LazyAsyncResult.ProtectedInvokeCallback(System.Object, IntPtr) at System.Net.LazyAsyncResult.InvokeCallback(System.Object) at System.Net.StreamFramer.ReadFrameComplete(System.IAsyncResult) at System.Net.StreamFramer.ReadFrameCallback(System.IAsyncResult) at System.Threading.Tasks.TaskToApm+<>c__DisplayClass3_0.<InvokeCallbackWhenTaskCompletes>b__0() at System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object) at System.Threading.Tasks.AwaitTaskContinuation.RunCallback(System.Threading.ContextCallback, System.Object, System.Threading.Tasks.Task ByRef) at System.Threading.Tasks.Task.RunContinuations(System.Object) at System.Threading.Tasks.Task.ExecuteWithThreadLocal(System.Threading.Tasks.Task ByRef) at System.Threading.ThreadPoolWorkQueue.Dispatch() /home/helixbot/dotnetbuild/work/a3f18f62-593d-45f2-985d-8cbdf5c68e39/Work/e26112cb-e993-41f5-aa16-81e08426a81c/Unzip/RunTests.sh: line 89: 4484 Aborted (core dumped) $RUNTIME_PATH/dotnet xunit.console.netcore.exe System.Net.Security.Tests.dll -xml testResults.xml -notrait category=nonnetcoreapptests -notrait category=nonlinuxtests -notrait category=failing command exited with ExitCode: 134 ```
1.0
System.Net.Security.Tests.dll FailFast on Debian - This comes just after "starting KDC" from `src\System.Net.Security\tests\Scripts\Unix\setup-kdc.sh` https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20170601.02/workItem/System.Net.Security.Tests/wilogs ``` Starting KDC.. FailFast: Invalid context passed to SafeDeleteNegoContext at System.Diagnostics.Debug.Assert(Boolean condition, String message, String detailMessage) at System.Net.Security.SafeDeleteNegoContext.SetGssContext(SafeGssContextHandle context) in /root/corefx/src/Common/src/System/Net/Security/Unix/SafeDeleteNegoContext.cs:line 52 at System.Net.Security.NegotiateStreamPal.EstablishSecurityContext(SafeFreeNegoCredentials credential, SafeDeleteContext& context, String targetName, ContextFlagsPal inFlags, SecurityBuffer inputBuffer, SecurityBuffer outputBuffer, ContextFlagsPal& outFlags) in /root/corefx/src/Common/src/System/Net/Security/NegotiateStreamPal.Unix.cs:line 197 at System.Net.Security.NegotiateStreamPal.InitializeSecurityContext(SafeFreeCredentials credentialsHandle, SafeDeleteContext& securityContext, String spn, ContextFlagsPal requestedContextFlags, SecurityBuffer[] inSecurityBufferArray, SecurityBuffer outSecurityBuffer, ContextFlagsPal& contextFlags) in /root/corefx/src/Common/src/System/Net/Security/NegotiateStreamPal.Unix.cs:line 240 at System.Net.NTAuthentication.GetOutgoingBlob(Byte[] incomingBlob, Boolean throwOnError, SecurityStatusPal& statusCode) in /root/corefx/src/Common/src/System/Net/NTAuthentication.Common.cs:line 243 at System.Net.Security.NegoState.GetOutgoingBlob(Byte[] incomingBlob, Exception& e) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 778 at System.Net.Security.NegoState.StartSendBlob(Byte[] message, LazyAsyncResult lazyResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 444 at System.Net.Security.NegoState.CheckCompletionBeforeNextSend(Byte[] message, LazyAsyncResult lazyResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 648 at System.Net.Security.NegoState.ProcessReceivedBlob(Byte[] message, LazyAsyncResult lazyResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 624 at System.Net.Security.NegoState.ReadCallback(IAsyncResult transportResult) in /root/corefx/src/System.Net.Security/src/System/Net/Security/NegoState.cs:line 756 at System.Net.LazyAsyncResult.Complete(IntPtr userToken) in /root/corefx/src/Common/src/System/Net/LazyAsyncResult.cs:line 428 at System.Net.LazyAsyncResult.ProtectedInvokeCallback(Object result, IntPtr userToken) in /root/corefx/src/Common/src/System/Net/LazyAsyncResult.cs:line 382 at System.Net.LazyAsyncResult.InvokeCallback(Object result) in /root/corefx/src/Common/src/System/Net/LazyAsyncResult.cs:line 389 at System.Net.StreamFramer.ReadFrameComplete(IAsyncResult transportResult) in /root/corefx/src/System.Net.Security/src/System/Net/StreamFramer.cs:line 270 at System.Net.StreamFramer.ReadFrameCallback(IAsyncResult transportResult) in /root/corefx/src/System.Net.Security/src/System/Net/StreamFramer.cs:line 168 at System.Threading.Tasks.TaskToApm.<>c__DisplayClass3_0.<InvokeCallbackWhenTaskCompletes>b__0() in /root/corefx/src/Common/src/System/Threading/Tasks/TaskToApm.cs:line 133 at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state) at System.Threading.Tasks.AwaitTaskContinuation.RunCallback(ContextCallback callback, Object state, Task& currentTask) at System.Threading.Tasks.Task.RunContinuations(Object continuationObject) at System.Threading.Tasks.Task.ExecuteWithThreadLocal(Task& currentTaskSlot) at System.Threading.ThreadPoolWorkQueue.Dispatch() at System.Environment.FailFast(System.String, System.Exception) at System.Net.Security.SafeDeleteNegoContext.SetGssContext(Microsoft.Win32.SafeHandles.SafeGssContextHandle) at System.Net.Security.NegotiateStreamPal.EstablishSecurityContext(System.Net.Security.SafeFreeNegoCredentials, System.Net.Security.SafeDeleteContext ByRef, System.String, System.Net.ContextFlagsPal, System.Net.Security.SecurityBuffer, System.Net.Security.SecurityBuffer, System.Net.ContextFlagsPal ByRef) at System.Net.Security.NegotiateStreamPal.InitializeSecurityContext(System.Net.Security.SafeFreeCredentials, System.Net.Security.SafeDeleteContext ByRef, System.String, System.Net.ContextFlagsPal, System.Net.Security.SecurityBuffer[], System.Net.Security.SecurityBuffer, System.Net.ContextFlagsPal ByRef) at System.Net.NTAuthentication.GetOutgoingBlob(Byte[], Boolean, System.Net.SecurityStatusPal ByRef) at System.Net.Security.NegoState.GetOutgoingBlob(Byte[], System.Exception ByRef) at System.Net.Security.NegoState.StartSendBlob(Byte[], System.Net.LazyAsyncResult) at System.Net.Security.NegoState.CheckCompletionBeforeNextSend(Byte[], System.Net.LazyAsyncResult) at System.Net.Security.NegoState.ProcessReceivedBlob(Byte[], System.Net.LazyAsyncResult) at System.Net.Security.NegoState.ReadCallback(System.IAsyncResult) at System.Net.LazyAsyncResult.Complete(IntPtr) at System.Net.LazyAsyncResult.ProtectedInvokeCallback(System.Object, IntPtr) at System.Net.LazyAsyncResult.InvokeCallback(System.Object) at System.Net.StreamFramer.ReadFrameComplete(System.IAsyncResult) at System.Net.StreamFramer.ReadFrameCallback(System.IAsyncResult) at System.Threading.Tasks.TaskToApm+<>c__DisplayClass3_0.<InvokeCallbackWhenTaskCompletes>b__0() at System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext, System.Threading.ContextCallback, System.Object) at System.Threading.Tasks.AwaitTaskContinuation.RunCallback(System.Threading.ContextCallback, System.Object, System.Threading.Tasks.Task ByRef) at System.Threading.Tasks.Task.RunContinuations(System.Object) at System.Threading.Tasks.Task.ExecuteWithThreadLocal(System.Threading.Tasks.Task ByRef) at System.Threading.ThreadPoolWorkQueue.Dispatch() /home/helixbot/dotnetbuild/work/a3f18f62-593d-45f2-985d-8cbdf5c68e39/Work/e26112cb-e993-41f5-aa16-81e08426a81c/Unzip/RunTests.sh: line 89: 4484 Aborted (core dumped) $RUNTIME_PATH/dotnet xunit.console.netcore.exe System.Net.Security.Tests.dll -xml testResults.xml -notrait category=nonnetcoreapptests -notrait category=nonlinuxtests -notrait category=failing command exited with ExitCode: 134 ```
test
system net security tests dll failfast on debian this comes just after starting kdc from src system net security tests scripts unix setup kdc sh starting kdc failfast invalid context passed to safedeletenegocontext at system diagnostics debug assert boolean condition string message string detailmessage at system net security safedeletenegocontext setgsscontext safegsscontexthandle context in root corefx src common src system net security unix safedeletenegocontext cs line at system net security negotiatestreampal establishsecuritycontext safefreenegocredentials credential safedeletecontext context string targetname contextflagspal inflags securitybuffer inputbuffer securitybuffer outputbuffer contextflagspal outflags in root corefx src common src system net security negotiatestreampal unix cs line at system net security negotiatestreampal initializesecuritycontext safefreecredentials credentialshandle safedeletecontext securitycontext string spn contextflagspal requestedcontextflags securitybuffer insecuritybufferarray securitybuffer outsecuritybuffer contextflagspal contextflags in root corefx src common src system net security negotiatestreampal unix cs line at system net ntauthentication getoutgoingblob byte incomingblob boolean throwonerror securitystatuspal statuscode in root corefx src common src system net ntauthentication common cs line at system net security negostate getoutgoingblob byte incomingblob exception e in root corefx src system net security src system net security negostate cs line at system net security negostate startsendblob byte message lazyasyncresult lazyresult in root corefx src system net security src system net security negostate cs line at system net security negostate checkcompletionbeforenextsend byte message lazyasyncresult lazyresult in root corefx src system net security src system net security negostate cs line at system net security negostate processreceivedblob byte message lazyasyncresult lazyresult in root corefx src system net security src system net security negostate cs line at system net security negostate readcallback iasyncresult transportresult in root corefx src system net security src system net security negostate cs line at system net lazyasyncresult complete intptr usertoken in root corefx src common src system net lazyasyncresult cs line at system net lazyasyncresult protectedinvokecallback object result intptr usertoken in root corefx src common src system net lazyasyncresult cs line at system net lazyasyncresult invokecallback object result in root corefx src common src system net lazyasyncresult cs line at system net streamframer readframecomplete iasyncresult transportresult in root corefx src system net security src system net streamframer cs line at system net streamframer readframecallback iasyncresult transportresult in root corefx src system net security src system net streamframer cs line at system threading tasks tasktoapm c b in root corefx src common src system threading tasks tasktoapm cs line at system threading executioncontext run executioncontext executioncontext contextcallback callback object state at system threading tasks awaittaskcontinuation runcallback contextcallback callback object state task currenttask at system threading tasks task runcontinuations object continuationobject at system threading tasks task executewiththreadlocal task currenttaskslot at system threading threadpoolworkqueue dispatch at system environment failfast system string system exception at system net security safedeletenegocontext setgsscontext microsoft safehandles safegsscontexthandle at system net security negotiatestreampal establishsecuritycontext system net security safefreenegocredentials system net security safedeletecontext byref system string system net contextflagspal system net security securitybuffer system net security securitybuffer system net contextflagspal byref at system net security negotiatestreampal initializesecuritycontext system net security safefreecredentials system net security safedeletecontext byref system string system net contextflagspal system net security securitybuffer system net security securitybuffer system net contextflagspal byref at system net ntauthentication getoutgoingblob byte boolean system net securitystatuspal byref at system net security negostate getoutgoingblob byte system exception byref at system net security negostate startsendblob byte system net lazyasyncresult at system net security negostate checkcompletionbeforenextsend byte system net lazyasyncresult at system net security negostate processreceivedblob byte system net lazyasyncresult at system net security negostate readcallback system iasyncresult at system net lazyasyncresult complete intptr at system net lazyasyncresult protectedinvokecallback system object intptr at system net lazyasyncresult invokecallback system object at system net streamframer readframecomplete system iasyncresult at system net streamframer readframecallback system iasyncresult at system threading tasks tasktoapm c b at system threading executioncontext run system threading executioncontext system threading contextcallback system object at system threading tasks awaittaskcontinuation runcallback system threading contextcallback system object system threading tasks task byref at system threading tasks task runcontinuations system object at system threading tasks task executewiththreadlocal system threading tasks task byref at system threading threadpoolworkqueue dispatch home helixbot dotnetbuild work work unzip runtests sh line aborted core dumped runtime path dotnet xunit console netcore exe system net security tests dll xml testresults xml notrait category nonnetcoreapptests notrait category nonlinuxtests notrait category failing command exited with exitcode
1
74,347
20,142,018,073
IssuesEvent
2022-02-09 00:54:32
microsoft/PowerToys
https://api.github.com/repos/microsoft/PowerToys
closed
[Build] Build failed: Path to exceeds max length
Issue-Bug Product-PowerToys Run Area-Build Needs-Triage
### Microsoft PowerToys version f2a3fa5ec68d1b07ab347beadc8f6e9160069cce ### Running as admin - [ ] Yes ### Area(s) with issue? PowerToys Run ### Steps to reproduce Build Launcher. ### ✔️ Expected Behavior Build works. ### ❌ Actual Behavior ``` Severity Code Description Project File Line Suppression State Error MSB3491 Could not write lines to file "obj\x64\Debug\net5.0-windows10.0.18362.0\Microsoft.PowerToys.Run.Plugin.WindowsTerminal.UnitTests.GeneratedMSBuildEditorConfig.editorconfig". Path: obj\x64\Debug\net5.0-windows10.0.18362.0\Microsoft.PowerToys.Run.Plugin.WindowsTerminal.UnitTests.GeneratedMSBuildEditorConfig.editorconfig exceeds the OS max path limit. The fully qualified file name must be less than 260 characters. Microsoft.PowerToys.Run.Plugin.WindowsTerminal.UnitTests C:\Program Files (x86)\Microsoft Visual Studio\2019\Community\MSBuild\Current\Bin\Roslyn\Microsoft.Managed.Core.targets 150 ``` ### Other Software _No response_
1.0
[Build] Build failed: Path to exceeds max length - ### Microsoft PowerToys version f2a3fa5ec68d1b07ab347beadc8f6e9160069cce ### Running as admin - [ ] Yes ### Area(s) with issue? PowerToys Run ### Steps to reproduce Build Launcher. ### ✔️ Expected Behavior Build works. ### ❌ Actual Behavior ``` Severity Code Description Project File Line Suppression State Error MSB3491 Could not write lines to file "obj\x64\Debug\net5.0-windows10.0.18362.0\Microsoft.PowerToys.Run.Plugin.WindowsTerminal.UnitTests.GeneratedMSBuildEditorConfig.editorconfig". Path: obj\x64\Debug\net5.0-windows10.0.18362.0\Microsoft.PowerToys.Run.Plugin.WindowsTerminal.UnitTests.GeneratedMSBuildEditorConfig.editorconfig exceeds the OS max path limit. The fully qualified file name must be less than 260 characters. Microsoft.PowerToys.Run.Plugin.WindowsTerminal.UnitTests C:\Program Files (x86)\Microsoft Visual Studio\2019\Community\MSBuild\Current\Bin\Roslyn\Microsoft.Managed.Core.targets 150 ``` ### Other Software _No response_
non_test
build failed path to exceeds max length microsoft powertoys version running as admin yes area s with issue powertoys run steps to reproduce build launcher ✔️ expected behavior build works ❌ actual behavior severity code description project file line suppression state error could not write lines to file obj debug microsoft powertoys run plugin windowsterminal unittests generatedmsbuildeditorconfig editorconfig path obj debug microsoft powertoys run plugin windowsterminal unittests generatedmsbuildeditorconfig editorconfig exceeds the os max path limit the fully qualified file name must be less than characters microsoft powertoys run plugin windowsterminal unittests c program files microsoft visual studio community msbuild current bin roslyn microsoft managed core targets other software no response
0
90,310
8,233,447,323
IssuesEvent
2018-09-08 01:02:05
knative/serving
https://api.github.com/repos/knative/serving
closed
We should return route domain for getRouteDomain() in route_test
area/test-and-release kind/bug
ERROR: type should be string, got "\r\nhttps://github.com/knative/serving/pull/1894 updated the getRouteDomain() method to poll for route state, but does not return domain. \r\n\r\nhttps://github.com/knative/serving/blob/master/test/conformance/route_test.go#L130"
1.0
We should return route domain for getRouteDomain() in route_test - https://github.com/knative/serving/pull/1894 updated the getRouteDomain() method to poll for route state, but does not return domain. https://github.com/knative/serving/blob/master/test/conformance/route_test.go#L130
test
we should return route domain for getroutedomain in route test updated the getroutedomain method to poll for route state but does not return domain
1
123,653
12,214,970,041
IssuesEvent
2020-05-01 11:32:35
hwanglight/ImageAugmentation
https://api.github.com/repos/hwanglight/ImageAugmentation
opened
Написать Readme
documentation
Readme должен содержать 1. Пререквизиты (библиотеки, которые нужно поставить, что собрать приложение) 2. Инструкция по сборке (какие команды выполнить в терминале, чтобы собрать приложение) 3. Инструкция по запуску и работе с консольной версией: формат параметров командной строки, формат конфигурационного файла (и закоммитить пример конфига) 4. Инструкция по работе с графической версией
1.0
Написать Readme - Readme должен содержать 1. Пререквизиты (библиотеки, которые нужно поставить, что собрать приложение) 2. Инструкция по сборке (какие команды выполнить в терминале, чтобы собрать приложение) 3. Инструкция по запуску и работе с консольной версией: формат параметров командной строки, формат конфигурационного файла (и закоммитить пример конфига) 4. Инструкция по работе с графической версией
non_test
написать readme readme должен содержать пререквизиты библиотеки которые нужно поставить что собрать приложение инструкция по сборке какие команды выполнить в терминале чтобы собрать приложение инструкция по запуску и работе с консольной версией формат параметров командной строки формат конфигурационного файла и закоммитить пример конфига инструкция по работе с графической версией
0
766,330
26,879,211,720
IssuesEvent
2023-02-05 12:32:06
herdstat/herdstat
https://api.github.com/repos/herdstat/herdstat
closed
Support more than commits as contributions
kind/feature priority/important-soon triage/accepted
The GitHub contribution graph includes much more than commits when computing contributions (see [here](https://docs.github.com/articles/why-are-my-contributions-not-showing-up-on-my-profile)). This would also be very cool for the `herdstat`contribution graph as these other things are also very important types of contributions. As a first step, we will include opened issues and PRs.
1.0
Support more than commits as contributions - The GitHub contribution graph includes much more than commits when computing contributions (see [here](https://docs.github.com/articles/why-are-my-contributions-not-showing-up-on-my-profile)). This would also be very cool for the `herdstat`contribution graph as these other things are also very important types of contributions. As a first step, we will include opened issues and PRs.
non_test
support more than commits as contributions the github contribution graph includes much more than commits when computing contributions see this would also be very cool for the herdstat contribution graph as these other things are also very important types of contributions as a first step we will include opened issues and prs
0
111,795
9,541,712,640
IssuesEvent
2019-04-30 23:30:36
Greenstand/treetracker-android
https://api.github.com/repos/Greenstand/treetracker-android
closed
Add Swahili Translation
android enhancement in testing
Update--boresha Upload --pakia Uploaded-imepakiwa GPS--GPS Accuracy --usahihi GPS accuracy --Usahihi wa GPS Meters--mita Sync-sawazisha To sync--yakusawazisha. Resume--endeleza Pause--katisha Data-data About--Kuhusu Map--ramani Your note--maoni yako New tree--mti mpya Insufficient--haitoshi Insufficient GPS accuracy-- GPS sio sahihi Save--hifadhi. Information--maelekezo The following is the translation of the information page Kuhusu. Tree Tracker planter app ni kifaa cha kufuatilia miti, kilichotengenezwa ili kuwawezesha watu na mashirika kupanda, kufuatilia na kulinda miti. mpangilio wa kifaa chako unahitaji usahihi wa GPS wa (mita10). Ili kufahamu kwa undani zaidi ni kwa jinsi gani kifaa hichi kitanufaisha mradi wako wa upandaji miti, tembelea tovuti ifuatayo. www.Greenstand.org kifaa hichi bado kipo kwenye matengenezo. kama una maombi yeyote au kama kifaa chako kimeshindwa kufanya kazi vizuri, toa taarifa kupitia e-mail ifuatayo. info@greenstand.org Treetracker(kwa majaribio)1:2:1 awamu ya 31
1.0
Add Swahili Translation - Update--boresha Upload --pakia Uploaded-imepakiwa GPS--GPS Accuracy --usahihi GPS accuracy --Usahihi wa GPS Meters--mita Sync-sawazisha To sync--yakusawazisha. Resume--endeleza Pause--katisha Data-data About--Kuhusu Map--ramani Your note--maoni yako New tree--mti mpya Insufficient--haitoshi Insufficient GPS accuracy-- GPS sio sahihi Save--hifadhi. Information--maelekezo The following is the translation of the information page Kuhusu. Tree Tracker planter app ni kifaa cha kufuatilia miti, kilichotengenezwa ili kuwawezesha watu na mashirika kupanda, kufuatilia na kulinda miti. mpangilio wa kifaa chako unahitaji usahihi wa GPS wa (mita10). Ili kufahamu kwa undani zaidi ni kwa jinsi gani kifaa hichi kitanufaisha mradi wako wa upandaji miti, tembelea tovuti ifuatayo. www.Greenstand.org kifaa hichi bado kipo kwenye matengenezo. kama una maombi yeyote au kama kifaa chako kimeshindwa kufanya kazi vizuri, toa taarifa kupitia e-mail ifuatayo. info@greenstand.org Treetracker(kwa majaribio)1:2:1 awamu ya 31
test
add swahili translation update boresha upload pakia uploaded imepakiwa gps gps accuracy usahihi gps accuracy usahihi wa gps meters mita sync sawazisha to sync yakusawazisha resume endeleza pause katisha data data about kuhusu map ramani your note maoni yako new tree mti mpya insufficient haitoshi insufficient gps accuracy gps sio sahihi save hifadhi information maelekezo the following is the translation of the information page kuhusu tree tracker planter app ni kifaa cha kufuatilia miti kilichotengenezwa ili kuwawezesha watu na mashirika kupanda kufuatilia na kulinda miti mpangilio wa kifaa chako unahitaji usahihi wa gps wa ili kufahamu kwa undani zaidi ni kwa jinsi gani kifaa hichi kitanufaisha mradi wako wa upandaji miti tembelea tovuti ifuatayo kifaa hichi bado kipo kwenye matengenezo kama una maombi yeyote au kama kifaa chako kimeshindwa kufanya kazi vizuri toa taarifa kupitia e mail ifuatayo info greenstand org treetracker kwa majaribio awamu ya
1
243,611
18,719,295,598
IssuesEvent
2021-11-03 09:55:33
ebi-ait/hca-ebi-wrangler-central
https://api.github.com/repos/ebi-ait/hca-ebi-wrangler-central
closed
Review HCA_to_scea documentation
documentation operations
We need to review https://ebi-ait.github.io/hca-ebi-wrangler-central/SOPs/hca_to_scea_tools_SOP.html and make sure that the whole process can be run without previous knowledge of the tools. Identified missing bits: - How to assign a HCAD accession - How to send fastq files - How and when to update the ticket with information about process in SCEA - How to get updates about process on the SCEA dataset - How to check a dataset has been successfully brokered into SCEA **Acceptability criteria**: - [x] The documentation has been reviewed and modified - [x] A tool-agnostic wrangler reviews the SOP after doing a full conversion of a dataset and lists anything missing in this ticket - [x] Ami will again update the documentation to cover any missing or unclear information in a PR - [x] Another tool-agnostic wrangler reviews the SOP and if needed, suggests further changes, or closes this ticket to reflect it is done.
1.0
Review HCA_to_scea documentation - We need to review https://ebi-ait.github.io/hca-ebi-wrangler-central/SOPs/hca_to_scea_tools_SOP.html and make sure that the whole process can be run without previous knowledge of the tools. Identified missing bits: - How to assign a HCAD accession - How to send fastq files - How and when to update the ticket with information about process in SCEA - How to get updates about process on the SCEA dataset - How to check a dataset has been successfully brokered into SCEA **Acceptability criteria**: - [x] The documentation has been reviewed and modified - [x] A tool-agnostic wrangler reviews the SOP after doing a full conversion of a dataset and lists anything missing in this ticket - [x] Ami will again update the documentation to cover any missing or unclear information in a PR - [x] Another tool-agnostic wrangler reviews the SOP and if needed, suggests further changes, or closes this ticket to reflect it is done.
non_test
review hca to scea documentation we need to review and make sure that the whole process can be run without previous knowledge of the tools identified missing bits how to assign a hcad accession how to send fastq files how and when to update the ticket with information about process in scea how to get updates about process on the scea dataset how to check a dataset has been successfully brokered into scea acceptability criteria the documentation has been reviewed and modified a tool agnostic wrangler reviews the sop after doing a full conversion of a dataset and lists anything missing in this ticket ami will again update the documentation to cover any missing or unclear information in a pr another tool agnostic wrangler reviews the sop and if needed suggests further changes or closes this ticket to reflect it is done
0
253,893
21,712,543,502
IssuesEvent
2022-05-10 14:57:42
aldrichtr/infraspective
https://api.github.com/repos/aldrichtr/infraspective
opened
Refactor unit tests so they do not rely on the module being loaded
>refactor @tests
The Pester unit tests should load the file that we want to test, and mock any supporting functions so that they are truly _unit_ tests. If the module is loaded first, then most of the functions will be coupled with their supporting functions. Those types of tests should be in **integration** not **unit** tests.
1.0
Refactor unit tests so they do not rely on the module being loaded - The Pester unit tests should load the file that we want to test, and mock any supporting functions so that they are truly _unit_ tests. If the module is loaded first, then most of the functions will be coupled with their supporting functions. Those types of tests should be in **integration** not **unit** tests.
test
refactor unit tests so they do not rely on the module being loaded the pester unit tests should load the file that we want to test and mock any supporting functions so that they are truly unit tests if the module is loaded first then most of the functions will be coupled with their supporting functions those types of tests should be in integration not unit tests
1
261,299
22,717,766,912
IssuesEvent
2022-07-06 05:00:28
Merck/metalite.ae
https://api.github.com/repos/Merck/metalite.ae
closed
Independent Testing of fmt_ci.R
independent test
Test plan of `fmt_ci`: - fmt_ci(-1.96,1.96, digits = 1) -> (-2.0, 2.0)
1.0
Independent Testing of fmt_ci.R - Test plan of `fmt_ci`: - fmt_ci(-1.96,1.96, digits = 1) -> (-2.0, 2.0)
test
independent testing of fmt ci r test plan of fmt ci fmt ci digits
1
101,934
12,731,490,828
IssuesEvent
2020-06-25 08:57:47
dambem/ClimateMonitorV2
https://api.github.com/repos/dambem/ClimateMonitorV2
closed
LuftDaten Custom Map Markers
Design Improvements
## Description As we're gonna start using weather company data in conjunction with luftdaten info, the current circular map markers won't do! Instead we need to make some custom markers so that it's clear it's luftdaten info. ## Steps to reproduce ## The actual vs expected behaviour ## Additional Information
1.0
LuftDaten Custom Map Markers - ## Description As we're gonna start using weather company data in conjunction with luftdaten info, the current circular map markers won't do! Instead we need to make some custom markers so that it's clear it's luftdaten info. ## Steps to reproduce ## The actual vs expected behaviour ## Additional Information
non_test
luftdaten custom map markers description as we re gonna start using weather company data in conjunction with luftdaten info the current circular map markers won t do instead we need to make some custom markers so that it s clear it s luftdaten info steps to reproduce the actual vs expected behaviour additional information
0
98,439
8,677,233,636
IssuesEvent
2018-11-30 16:13:40
cerner/terra-framework
https://api.github.com/repos/cerner/terra-framework
closed
Use themeCombinationOfCustomProperties
Orion Reviewed beginner-friendly testing
# Feature Request Some component WDIO tests use `themeEachCustomProperty` to test theme variables. We should evaluate if component tests can use `themeCombinationOfCustomProperties`. This will drastically speed up visually regression tests by taking a single screenshot for a set of theme variables, versus a screenshot for each theme variable.
1.0
Use themeCombinationOfCustomProperties - # Feature Request Some component WDIO tests use `themeEachCustomProperty` to test theme variables. We should evaluate if component tests can use `themeCombinationOfCustomProperties`. This will drastically speed up visually regression tests by taking a single screenshot for a set of theme variables, versus a screenshot for each theme variable.
test
use themecombinationofcustomproperties feature request some component wdio tests use themeeachcustomproperty to test theme variables we should evaluate if component tests can use themecombinationofcustomproperties this will drastically speed up visually regression tests by taking a single screenshot for a set of theme variables versus a screenshot for each theme variable
1
430,830
12,466,664,405
IssuesEvent
2020-05-28 15:49:06
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
IPv6 prefix could be added multiple times to prefix timer list
area: Networking bug priority: medium
**Describe the bug** We could add same IPv6 prefix multiple times to prefix timer list. This can lead to denial-of-service issue if we receive suitable crafted IPv6 RA packets.
1.0
IPv6 prefix could be added multiple times to prefix timer list - **Describe the bug** We could add same IPv6 prefix multiple times to prefix timer list. This can lead to denial-of-service issue if we receive suitable crafted IPv6 RA packets.
non_test
prefix could be added multiple times to prefix timer list describe the bug we could add same prefix multiple times to prefix timer list this can lead to denial of service issue if we receive suitable crafted ra packets
0
130,108
10,596,758,032
IssuesEvent
2019-10-09 22:06:33
inspec/inspec
https://api.github.com/repos/inspec/inspec
closed
Improve integration testing of databases
Aspect: Testing Platform: Databases Type: Enhancement
### 🎛 Description 🙋 feature request For database-targeted resources such as postgres_session, oracledb_session, etc , no formal integration testing exists. Some unit testing exists, but unit testing simulates responses from outside systems. Many recent issues have been reported which would have been caught by testing with real databases, such as: * [ ] #3027 * [ ] #3659 * [ ] #3177 * [ ] #3680 * [ ] #3255 * [ ] #3594 Fixing these issues is also difficult since we don't have a repeatable way of creating a test fixture database. It also prevents regression testing. ### 🌍 InSpec and Platform Version 3.1.3 ### 💁 Possible Solutions It should be exposed through Rake, but not be a default test job. `rake test:db` seems reasonable. Ideally, we'd have something (vagrant, RDS, docker?) that would be able to standup one system of each of the 4 we support; possibly running on multiple platforms (oracle on windows and linux, for example) or using different connection techniques (different CLI clients vs internal ruby API clients). We'd like to to be able to isolate a setup, as well. It doesn't need to be fast, but repeatability and maintainability are important.
1.0
Improve integration testing of databases - ### 🎛 Description 🙋 feature request For database-targeted resources such as postgres_session, oracledb_session, etc , no formal integration testing exists. Some unit testing exists, but unit testing simulates responses from outside systems. Many recent issues have been reported which would have been caught by testing with real databases, such as: * [ ] #3027 * [ ] #3659 * [ ] #3177 * [ ] #3680 * [ ] #3255 * [ ] #3594 Fixing these issues is also difficult since we don't have a repeatable way of creating a test fixture database. It also prevents regression testing. ### 🌍 InSpec and Platform Version 3.1.3 ### 💁 Possible Solutions It should be exposed through Rake, but not be a default test job. `rake test:db` seems reasonable. Ideally, we'd have something (vagrant, RDS, docker?) that would be able to standup one system of each of the 4 we support; possibly running on multiple platforms (oracle on windows and linux, for example) or using different connection techniques (different CLI clients vs internal ruby API clients). We'd like to to be able to isolate a setup, as well. It doesn't need to be fast, but repeatability and maintainability are important.
test
improve integration testing of databases 🎛 description 🙋 feature request for database targeted resources such as postgres session oracledb session etc no formal integration testing exists some unit testing exists but unit testing simulates responses from outside systems many recent issues have been reported which would have been caught by testing with real databases such as fixing these issues is also difficult since we don t have a repeatable way of creating a test fixture database it also prevents regression testing 🌍 inspec and platform version 💁 possible solutions it should be exposed through rake but not be a default test job rake test db seems reasonable ideally we d have something vagrant rds docker that would be able to standup one system of each of the we support possibly running on multiple platforms oracle on windows and linux for example or using different connection techniques different cli clients vs internal ruby api clients we d like to to be able to isolate a setup as well it doesn t need to be fast but repeatability and maintainability are important
1
414,828
28,005,942,718
IssuesEvent
2023-03-27 15:15:51
csia-pme/a-guide-to-mlops
https://api.github.com/repos/csia-pme/a-guide-to-mlops
closed
Rewrite the guide in a more formal way
documentation enhancement
As discussed with @rmarquis, it could be a good idea to rewrite some parts of the guide in a more formal way. ## Examples ### Chapter 2: Share your ML experiment code with Git "Instead of relying on ZIP archives, **we'll** create a Git repository to enable easy collaboration with the rest of the team" "Instead of relying on ZIP archives, **we will** create a Git repository to enable easy collaboration with the rest of the team" ### Chapter 5: Track model evolutions with DVC "Once this stage is created, **you'll** be able to change our model's configuration, evaluate the new configuration and compare its performance with the last commited ones." "Once this stage is created, **you will** be able to change our model's configuration, evaluate the new configuration and compare its performance with the last commited ones."
1.0
Rewrite the guide in a more formal way - As discussed with @rmarquis, it could be a good idea to rewrite some parts of the guide in a more formal way. ## Examples ### Chapter 2: Share your ML experiment code with Git "Instead of relying on ZIP archives, **we'll** create a Git repository to enable easy collaboration with the rest of the team" "Instead of relying on ZIP archives, **we will** create a Git repository to enable easy collaboration with the rest of the team" ### Chapter 5: Track model evolutions with DVC "Once this stage is created, **you'll** be able to change our model's configuration, evaluate the new configuration and compare its performance with the last commited ones." "Once this stage is created, **you will** be able to change our model's configuration, evaluate the new configuration and compare its performance with the last commited ones."
non_test
rewrite the guide in a more formal way as discussed with rmarquis it could be a good idea to rewrite some parts of the guide in a more formal way examples chapter share your ml experiment code with git instead of relying on zip archives we ll create a git repository to enable easy collaboration with the rest of the team instead of relying on zip archives we will create a git repository to enable easy collaboration with the rest of the team chapter track model evolutions with dvc once this stage is created you ll be able to change our model s configuration evaluate the new configuration and compare its performance with the last commited ones once this stage is created you will be able to change our model s configuration evaluate the new configuration and compare its performance with the last commited ones
0
58,926
11,911,642,753
IssuesEvent
2020-03-31 08:58:04
ModellingWebLab/weblab-fc
https://api.github.com/repos/ModellingWebLab/weblab-fc
closed
Cut dependency on pycml
code-and-design install
At present running an experiment requires pycml to generate the manipulated model. This will change to use cellmlmanip & fccodegen. - [x] As a first step, put a manually generated (using fccodegen) model in the tests and see if we can just simulate it. - [x] Then call this automatically rather than calling pycml. - [x] Then iteratively implement model manipulations. See #23.
1.0
Cut dependency on pycml - At present running an experiment requires pycml to generate the manipulated model. This will change to use cellmlmanip & fccodegen. - [x] As a first step, put a manually generated (using fccodegen) model in the tests and see if we can just simulate it. - [x] Then call this automatically rather than calling pycml. - [x] Then iteratively implement model manipulations. See #23.
non_test
cut dependency on pycml at present running an experiment requires pycml to generate the manipulated model this will change to use cellmlmanip fccodegen as a first step put a manually generated using fccodegen model in the tests and see if we can just simulate it then call this automatically rather than calling pycml then iteratively implement model manipulations see
0
4,842
7,325,304,980
IssuesEvent
2018-03-03 07:03:45
renovateapp/renovate
https://api.github.com/repos/renovateapp/renovate
closed
bug: Branch prefix not working with gitLab platform
needs-requirements pri4-low
I have just realized that even through I was always setting the branchPrefix option in the config.js, all dependency level branches were still using the default /renovate/eslint... format. It looks like the branchPrefix config option is never read or it gets overwritten somewhere. `config.branchPrefix` returns undefined. Setting in config.js, for example: `branchPrefix: "chore/WQ-0_renovate_",` Results in default branch format, such as: renovate/eslint-4.x The only way to get around it is to actually overwrite the branchName in the config: `branchName: "chore/WQ-0_renovate_{{depName}}-{{newVersionMajor}}.x",` This will work and correct branch names are used.
1.0
bug: Branch prefix not working with gitLab platform - I have just realized that even through I was always setting the branchPrefix option in the config.js, all dependency level branches were still using the default /renovate/eslint... format. It looks like the branchPrefix config option is never read or it gets overwritten somewhere. `config.branchPrefix` returns undefined. Setting in config.js, for example: `branchPrefix: "chore/WQ-0_renovate_",` Results in default branch format, such as: renovate/eslint-4.x The only way to get around it is to actually overwrite the branchName in the config: `branchName: "chore/WQ-0_renovate_{{depName}}-{{newVersionMajor}}.x",` This will work and correct branch names are used.
non_test
bug branch prefix not working with gitlab platform i have just realized that even through i was always setting the branchprefix option in the config js all dependency level branches were still using the default renovate eslint format it looks like the branchprefix config option is never read or it gets overwritten somewhere config branchprefix returns undefined setting in config js for example branchprefix chore wq renovate results in default branch format such as renovate eslint x the only way to get around it is to actually overwrite the branchname in the config branchname chore wq renovate depname newversionmajor x this will work and correct branch names are used
0
48,346
7,422,610,780
IssuesEvent
2018-03-23 00:15:57
OneDrive/onedrive-api-docs
https://api.github.com/repos/OneDrive/onedrive-api-docs
closed
Webhooks on items (not "/me/drive/root")
documentation question
I notice that the Update webhooks has a `PATCH /drive/items/{item-id}/subscriptions/{id}` request where the response has a value for resource of ` "resource": "/drives/0/items/57705F13F13C3C0C!104",`. Does this mean that webhooks can be created for specific item IDs? When I try to send a create request with a specific item-id, I get this: ``` { "error": { "code": "InvalidRequest", "message": "resource '/drive/items/{item-id}/' is not supported.", "innerError": { "request-id": "793ba53f-a055-4000-aa69-25e775483d26", "date": "2017-08-21T17:04:25" } } } ``` I'm only able to subscribe to the root folder (`"/me/drive/root"`).
1.0
Webhooks on items (not "/me/drive/root") - I notice that the Update webhooks has a `PATCH /drive/items/{item-id}/subscriptions/{id}` request where the response has a value for resource of ` "resource": "/drives/0/items/57705F13F13C3C0C!104",`. Does this mean that webhooks can be created for specific item IDs? When I try to send a create request with a specific item-id, I get this: ``` { "error": { "code": "InvalidRequest", "message": "resource '/drive/items/{item-id}/' is not supported.", "innerError": { "request-id": "793ba53f-a055-4000-aa69-25e775483d26", "date": "2017-08-21T17:04:25" } } } ``` I'm only able to subscribe to the root folder (`"/me/drive/root"`).
non_test
webhooks on items not me drive root i notice that the update webhooks has a patch drive items item id subscriptions id request where the response has a value for resource of resource drives items does this mean that webhooks can be created for specific item ids when i try to send a create request with a specific item id i get this error code invalidrequest message resource drive items item id is not supported innererror request id date i m only able to subscribe to the root folder me drive root
0
203,147
15,352,105,944
IssuesEvent
2021-03-01 06:25:10
apache/apisix-dashboard
https://api.github.com/repos/apache/apisix-dashboard
closed
test: add e2e test for dubbo
testcase
after PR https://github.com/apache/apisix/pull/3224 merged, we should add some test cases for it. and we need build a test env for it.
1.0
test: add e2e test for dubbo - after PR https://github.com/apache/apisix/pull/3224 merged, we should add some test cases for it. and we need build a test env for it.
test
test add test for dubbo after pr merged we should add some test cases for it and we need build a test env for it
1
339,960
30,488,783,072
IssuesEvent
2023-07-18 05:55:49
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: alterpk-bank failed
C-test-failure O-robot O-roachtest branch-master release-blocker T-sql-foundations
roachtest.alterpk-bank [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=artifacts#/alterpk-bank) on master @ [7675ca4998134028f0623e04737b5cb69fcc33a9](https://github.com/cockroachdb/cockroach/commits/7675ca4998134028f0623e04737b5cb69fcc33a9): ``` (cluster.go:2180).Start: ~ COCKROACH_CONNECT_TIMEOUT=1200 ./cockroach sql --url 'postgres://root@localhost:26257?sslmode=disable' -e "CREATE SCHEDULE IF NOT EXISTS test_only_backup FOR BACKUP INTO 'gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-06-n4cpu4/1689659690804569106?AUTH=implicit' RECURRING '*/15 * * * *' FULL BACKUP '@hourly' WITH SCHEDULE OPTIONS first_run = 'now'" ERROR: unexpected error occurred when checking for existing backups in gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-06-n4cpu4/1689659690804569106?AUTH=implicit: unable to list files in gcs bucket: googleapi: Error 403: 21965078311-compute@developer.gserviceaccount.com does not have storage.objects.list access to the Google Cloud Storage bucket. Permission 'storage.objects.list' denied on resource (or it may not exist). SQLSTATE: 58030 Failed running "sql": COMMAND_PROBLEM: exit status 1 test artifacts and logs in: /artifacts/alterpk-bank/run_1 ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-foundations <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*alterpk-bank.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: alterpk-bank failed - roachtest.alterpk-bank [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=artifacts#/alterpk-bank) on master @ [7675ca4998134028f0623e04737b5cb69fcc33a9](https://github.com/cockroachdb/cockroach/commits/7675ca4998134028f0623e04737b5cb69fcc33a9): ``` (cluster.go:2180).Start: ~ COCKROACH_CONNECT_TIMEOUT=1200 ./cockroach sql --url 'postgres://root@localhost:26257?sslmode=disable' -e "CREATE SCHEDULE IF NOT EXISTS test_only_backup FOR BACKUP INTO 'gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-06-n4cpu4/1689659690804569106?AUTH=implicit' RECURRING '*/15 * * * *' FULL BACKUP '@hourly' WITH SCHEDULE OPTIONS first_run = 'now'" ERROR: unexpected error occurred when checking for existing backups in gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-06-n4cpu4/1689659690804569106?AUTH=implicit: unable to list files in gcs bucket: googleapi: Error 403: 21965078311-compute@developer.gserviceaccount.com does not have storage.objects.list access to the Google Cloud Storage bucket. Permission 'storage.objects.list' denied on resource (or it may not exist). SQLSTATE: 58030 Failed running "sql": COMMAND_PROBLEM: exit status 1 test artifacts and logs in: /artifacts/alterpk-bank/run_1 ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-foundations <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*alterpk-bank.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest alterpk bank failed roachtest alterpk bank with on master cluster go start cockroach connect timeout cockroach sql url postgres root localhost sslmode disable e create schedule if not exists test only backup for backup into gs cockroach backup testing private roachprod scheduled backups teamcity auth implicit recurring full backup hourly with schedule options first run now error unexpected error occurred when checking for existing backups in gs cockroach backup testing private roachprod scheduled backups teamcity auth implicit unable to list files in gcs bucket googleapi error compute developer gserviceaccount com does not have storage objects list access to the google cloud storage bucket permission storage objects list denied on resource or it may not exist sqlstate failed running sql command problem exit status test artifacts and logs in artifacts alterpk bank run parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb sql foundations
1
107,468
23,418,927,187
IssuesEvent
2022-08-13 11:42:03
FerretDB/FerretDB
https://api.github.com/repos/FerretDB/FerretDB
opened
Enrich contribution guide
code/enhancement
I started my contribution by looking at https://github.com/FerretDB/FerretDB/blob/main/CONTRIBUTING.md and moving step by step. And first step `task init` is failing by default. The problem is with a command `git describe --tags --dirty > gen/version.txt` git describe return error on forks with default mark `Copy the main branch only`. I think would be helpful to describe this behavior in CONTRIBUTING.md, so that new contributors will not face that issues
1.0
Enrich contribution guide - I started my contribution by looking at https://github.com/FerretDB/FerretDB/blob/main/CONTRIBUTING.md and moving step by step. And first step `task init` is failing by default. The problem is with a command `git describe --tags --dirty > gen/version.txt` git describe return error on forks with default mark `Copy the main branch only`. I think would be helpful to describe this behavior in CONTRIBUTING.md, so that new contributors will not face that issues
non_test
enrich contribution guide i started my contribution by looking at and moving step by step and first step task init is failing by default the problem is with a command git describe tags dirty gen version txt git describe return error on forks with default mark copy the main branch only i think would be helpful to describe this behavior in contributing md so that new contributors will not face that issues
0
721,110
24,818,267,057
IssuesEvent
2022-10-25 14:36:40
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
closed
Geography request for BELL Museum, part 2
Priority-High (Needed for work)
Hi, I just pre-bulkloaded a bunch of data and it spit back some higher geography terms that need fixing or that I found a bit confusing. I've attached a CSV with my notes on these. Three entries were just missing from higher geography where they shouldn't be; I have noted this in the CSV under NOTE, and request they be added. The others need clarification for me or correction in the database: 1) Kenya does not appear to have a continent? I input "Africa, Kenya" but that does not seem to exist at the moment, although I was grooming my data with a list previously provided by @dustymc. 2) Our data has an entry for "North America, United States, Virginia, Washington County", but Arctos recognizes "North America, United States, Virginia, Virginia, Washington County". Any idea what's going on there? 3) Vanuatu appears to be missing an ocean, requiring use of "Vanuatu" rather than "Pacific Ocean, Vanuatu". Is this correct? BTW, will I have to restart the entire bulk upload if the missing geography terms are added? Or can I just fill in the added term in 'SHOULDBE' and it will be okay? Thanks! [geography_additions_queries_10_13_2022.csv](https://github.com/ArctosDB/arctos/files/9781470/geography_additions_queries_10_13_2022.csv)
1.0
Geography request for BELL Museum, part 2 - Hi, I just pre-bulkloaded a bunch of data and it spit back some higher geography terms that need fixing or that I found a bit confusing. I've attached a CSV with my notes on these. Three entries were just missing from higher geography where they shouldn't be; I have noted this in the CSV under NOTE, and request they be added. The others need clarification for me or correction in the database: 1) Kenya does not appear to have a continent? I input "Africa, Kenya" but that does not seem to exist at the moment, although I was grooming my data with a list previously provided by @dustymc. 2) Our data has an entry for "North America, United States, Virginia, Washington County", but Arctos recognizes "North America, United States, Virginia, Virginia, Washington County". Any idea what's going on there? 3) Vanuatu appears to be missing an ocean, requiring use of "Vanuatu" rather than "Pacific Ocean, Vanuatu". Is this correct? BTW, will I have to restart the entire bulk upload if the missing geography terms are added? Or can I just fill in the added term in 'SHOULDBE' and it will be okay? Thanks! [geography_additions_queries_10_13_2022.csv](https://github.com/ArctosDB/arctos/files/9781470/geography_additions_queries_10_13_2022.csv)
non_test
geography request for bell museum part hi i just pre bulkloaded a bunch of data and it spit back some higher geography terms that need fixing or that i found a bit confusing i ve attached a csv with my notes on these three entries were just missing from higher geography where they shouldn t be i have noted this in the csv under note and request they be added the others need clarification for me or correction in the database kenya does not appear to have a continent i input africa kenya but that does not seem to exist at the moment although i was grooming my data with a list previously provided by dustymc our data has an entry for north america united states virginia washington county but arctos recognizes north america united states virginia virginia washington county any idea what s going on there vanuatu appears to be missing an ocean requiring use of vanuatu rather than pacific ocean vanuatu is this correct btw will i have to restart the entire bulk upload if the missing geography terms are added or can i just fill in the added term in shouldbe and it will be okay thanks
0
319,589
27,387,196,494
IssuesEvent
2023-02-28 14:06:06
splendo/kaluga
https://api.github.com/repos/splendo/kaluga
closed
Unstable test testScanning in com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest (android integration test)
🤖android 1.0.0 ❄ flaky test
currently in https://github.com/splendo/kaluga/tree/feature/performance-and-tests only ``` java.lang.AssertionError: Expected Exactly(times=1) but got 0 times at com.splendo.kaluga.test.base.mock.MockMethodKt.fail(MockMethod.kt:35) at com.splendo.kaluga.test.base.mock.MockMethodKt.expect(MockMethod.kt:32) at com.splendo.kaluga.test.base.mock.MockMethodKt.access$expect(MockMethod.kt:1) at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:166) at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:158) at com.splendo.kaluga.test.base.mock.VerifyKt.verify(verify.kt:36) at com.splendo.kaluga.test.base.mock.VerifyKt.verify$default(verify.kt:18) at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invokeSuspend(ScanningStateRepoTest.kt:181) at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:8) at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:4) at com.splendo.kaluga.test.base.BaseFlowTest$mainAction$3.invokeSuspend(FlowTest.kt:264) at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) at android.os.Handler.handleCallback(Handler.java:883) at android.os.Handler.dispatchMessage(Handler.java:100) at android.os.Looper.loop(Looper.java:214) at android.app.ActivityThread.main(ActivityThread.java:7356) at java.lang.reflect.Method.invoke(Native Method) at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:492) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:930) ``` ``` 07-20 01:11:31.058: I/TestRunner(21080): started: testScanning(com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest) 07-20 01:11:31.096: V/LogKt$debug(21080): first test offered, starting collection 07-20 01:11:31.096: V/LogKt$debug(21080): launch flow scope... 07-20 01:11:31.097: V/LogKt$debug(21080): wait for main thread to be launched in StandaloneCoroutine{Active}@ef49f3e 07-20 01:11:31.100: V/LogKt$debug(21080): main scope launched, about to flow, test channel empty 07-20 01:11:31.100: V/LogKt$debug(21080): waited for main thread to be launched 07-20 01:11:31.100: V/LogKt$debug(21080): 1 in collection (including this one), offering 07-20 01:11:31.101: V/LogKt$debug(21080): start mainAction 07-20 01:11:31.101: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds 07-20 01:11:31.120: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@f17714a], test channel not empty " 07-20 01:11:31.122: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@abbf8bb) 07-20 01:11:31.126: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@abbf8bb) 07-20 01:11:31.134: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@abbf8bb) 07-20 01:11:31.135: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@f17714a] 07-20 01:11:31.135: V/LogKt$debug(21080): in main scope for mainAction 07-20 01:11:31.138: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@37411d8], test channel empty " 07-20 01:11:31.138: V/LogKt$debug(21080): did mainAction 07-20 01:11:31.139: V/LogKt$debug(21080): 1 in collection (including this one), offering 07-20 01:11:31.139: V/LogKt$debug(21080): start mainAction 07-20 01:11:31.140: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds 07-20 01:11:31.153: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@aeb9931) 07-20 01:11:31.157: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@aeb9931) 07-20 01:11:31.159: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@aeb9931) 07-20 01:11:31.159: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@37411d8] 07-20 01:11:31.159: V/LogKt$debug(21080): in main scope for mainAction 07-20 01:11:31.160: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@be35816], test channel empty " 07-20 01:11:31.160: V/LogKt$debug(21080): did mainAction 07-20 01:11:31.160: V/LogKt$debug(21080): 1 in collection (including this one), offering 07-20 01:11:31.161: V/LogKt$debug(21080): start mainAction 07-20 01:11:31.161: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds 07-20 01:11:31.163: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@ba76597) 07-20 01:11:31.167: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@ba76597) 07-20 01:11:31.169: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@ba76597) 07-20 01:11:31.170: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@be35816] 07-20 01:11:31.170: V/LogKt$debug(21080): in main scope for mainAction 07-20 01:11:31.172: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@1047384], test channel empty " 07-20 01:11:31.172: V/LogKt$debug(21080): did mainAction 07-20 01:11:31.173: V/LogKt$debug(21080): 1 in collection (including this one), offering 07-20 01:11:31.173: V/LogKt$debug(21080): start action 07-20 01:11:31.174: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds 07-20 01:11:31.176: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@62f976d) 07-20 01:11:31.178: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@62f976d) 07-20 01:11:31.179: V/LogKt$debug(21080): await all test blocks, but none found, skip waiting 07-20 01:11:31.179: V/LogKt$debug(21080): job: StandaloneCoroutine{Active}@ef49f3e 07-20 01:11:31.179: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@62f976d) 07-20 01:11:31.179: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@1047384] 07-20 01:11:31.181: V/LogKt$debug(21080): Ending flow, job canceled 07-20 01:11:31.181: V/LogKt$debug(21080): test channel closed 07-20 01:11:31.181: V/LogKt$debug(21080): did action 07-20 01:11:31.182: V/LogKt$debug(21080): start mainAction 07-20 01:11:31.182: V/LogKt$debug(21080): await all test blocks, but none found, skip waiting 07-20 01:11:31.183: V/LogKt$debug(21080): in main scope for mainAction 07-20 01:11:31.191: E/TestRunner(21080): failed: testScanning(com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest) 07-20 01:11:31.191: E/TestRunner(21080): ----- begin exception ----- 07-20 01:11:31.192: I/link_layer_controller(1331): Sending scan response 07-20 01:11:31.193: E/TestRunner(21080): java.lang.AssertionError: Expected Exactly(times=1) but got 0 times 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.MockMethodKt.fail(MockMethod.kt:35) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.MockMethodKt.expect(MockMethod.kt:32) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.MockMethodKt.access$expect(MockMethod.kt:1) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:166) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:158) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.VerifyKt.verify(verify.kt:36) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.VerifyKt.verify$default(verify.kt:18) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invokeSuspend(ScanningStateRepoTest.kt:182) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:8) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:4) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.BaseFlowTest$mainAction$3.invokeSuspend(FlowTest.kt:264) 07-20 01:11:31.193: E/TestRunner(21080): at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) 07-20 01:11:31.193: E/TestRunner(21080): at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) 07-20 01:11:31.193: E/TestRunner(21080): at android.os.Handler.handleCallback(Handler.java:883) 07-20 01:11:31.193: E/TestRunner(21080): at android.os.Handler.dispatchMessage(Handler.java:100) 07-20 01:11:31.193: E/TestRunner(21080): at android.os.Looper.loop(Looper.java:214) 07-20 01:11:31.193: E/TestRunner(21080): at android.app.ActivityThread.main(ActivityThread.java:7356) 07-20 01:11:31.193: E/TestRunner(21080): at java.lang.reflect.Method.invoke(Native Method) 07-20 01:11:31.193: E/TestRunner(21080): at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:492) 07-20 01:11:31.193: E/TestRunner(21080): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:930) 07-20 01:11:31.193: E/TestRunner(21080): ----- end exception ----- 07-20 01:11:31.195: I/TestRunner(21080): finished: testScanning(com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest) ```
1.0
Unstable test testScanning in com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest (android integration test) - currently in https://github.com/splendo/kaluga/tree/feature/performance-and-tests only ``` java.lang.AssertionError: Expected Exactly(times=1) but got 0 times at com.splendo.kaluga.test.base.mock.MockMethodKt.fail(MockMethod.kt:35) at com.splendo.kaluga.test.base.mock.MockMethodKt.expect(MockMethod.kt:32) at com.splendo.kaluga.test.base.mock.MockMethodKt.access$expect(MockMethod.kt:1) at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:166) at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:158) at com.splendo.kaluga.test.base.mock.VerifyKt.verify(verify.kt:36) at com.splendo.kaluga.test.base.mock.VerifyKt.verify$default(verify.kt:18) at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invokeSuspend(ScanningStateRepoTest.kt:181) at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:8) at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:4) at com.splendo.kaluga.test.base.BaseFlowTest$mainAction$3.invokeSuspend(FlowTest.kt:264) at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) at android.os.Handler.handleCallback(Handler.java:883) at android.os.Handler.dispatchMessage(Handler.java:100) at android.os.Looper.loop(Looper.java:214) at android.app.ActivityThread.main(ActivityThread.java:7356) at java.lang.reflect.Method.invoke(Native Method) at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:492) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:930) ``` ``` 07-20 01:11:31.058: I/TestRunner(21080): started: testScanning(com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest) 07-20 01:11:31.096: V/LogKt$debug(21080): first test offered, starting collection 07-20 01:11:31.096: V/LogKt$debug(21080): launch flow scope... 07-20 01:11:31.097: V/LogKt$debug(21080): wait for main thread to be launched in StandaloneCoroutine{Active}@ef49f3e 07-20 01:11:31.100: V/LogKt$debug(21080): main scope launched, about to flow, test channel empty 07-20 01:11:31.100: V/LogKt$debug(21080): waited for main thread to be launched 07-20 01:11:31.100: V/LogKt$debug(21080): 1 in collection (including this one), offering 07-20 01:11:31.101: V/LogKt$debug(21080): start mainAction 07-20 01:11:31.101: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds 07-20 01:11:31.120: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@f17714a], test channel not empty " 07-20 01:11:31.122: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@abbf8bb) 07-20 01:11:31.126: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@abbf8bb) 07-20 01:11:31.134: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@abbf8bb) 07-20 01:11:31.135: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@f17714a] 07-20 01:11:31.135: V/LogKt$debug(21080): in main scope for mainAction 07-20 01:11:31.138: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@37411d8], test channel empty " 07-20 01:11:31.138: V/LogKt$debug(21080): did mainAction 07-20 01:11:31.139: V/LogKt$debug(21080): 1 in collection (including this one), offering 07-20 01:11:31.139: V/LogKt$debug(21080): start mainAction 07-20 01:11:31.140: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds 07-20 01:11:31.153: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@aeb9931) 07-20 01:11:31.157: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@aeb9931) 07-20 01:11:31.159: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@aeb9931) 07-20 01:11:31.159: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@37411d8] 07-20 01:11:31.159: V/LogKt$debug(21080): in main scope for mainAction 07-20 01:11:31.160: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@be35816], test channel empty " 07-20 01:11:31.160: V/LogKt$debug(21080): did mainAction 07-20 01:11:31.160: V/LogKt$debug(21080): 1 in collection (including this one), offering 07-20 01:11:31.161: V/LogKt$debug(21080): start mainAction 07-20 01:11:31.161: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds 07-20 01:11:31.163: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@ba76597) 07-20 01:11:31.167: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@ba76597) 07-20 01:11:31.169: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@ba76597) 07-20 01:11:31.170: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Scanning@be35816] 07-20 01:11:31.170: V/LogKt$debug(21080): in main scope for mainAction 07-20 01:11:31.172: V/LogKt$debug(21080): in flow received [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@1047384], test channel empty " 07-20 01:11:31.172: V/LogKt$debug(21080): did mainAction 07-20 01:11:31.173: V/LogKt$debug(21080): 1 in collection (including this one), offering 07-20 01:11:31.173: V/LogKt$debug(21080): start action 07-20 01:11:31.174: V/LogKt$debug(21080): await all test blocks (1), give it 60000 milliseconds 07-20 01:11:31.176: V/LogKt$debug(21080): received test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@62f976d) 07-20 01:11:31.178: V/LogKt$debug(21080): ran test block (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Active}@62f976d) 07-20 01:11:31.179: V/LogKt$debug(21080): await all test blocks, but none found, skip waiting 07-20 01:11:31.179: V/LogKt$debug(21080): job: StandaloneCoroutine{Active}@ef49f3e 07-20 01:11:31.179: V/LogKt$debug(21080): completed (com.splendo.kaluga.bluetooth.BluetoothFlowTest.DeviceContext.(com.splendo.kaluga.bluetooth.scanner.ScanningState) -> kotlin.Unit, CompletableDeferredImpl{Completed}@62f976d) 07-20 01:11:31.179: V/LogKt$debug(21080): handeling value completed [com.splendo.kaluga.bluetooth.scanner.ScanningStateImpl$Enabled$Idle@1047384] 07-20 01:11:31.181: V/LogKt$debug(21080): Ending flow, job canceled 07-20 01:11:31.181: V/LogKt$debug(21080): test channel closed 07-20 01:11:31.181: V/LogKt$debug(21080): did action 07-20 01:11:31.182: V/LogKt$debug(21080): start mainAction 07-20 01:11:31.182: V/LogKt$debug(21080): await all test blocks, but none found, skip waiting 07-20 01:11:31.183: V/LogKt$debug(21080): in main scope for mainAction 07-20 01:11:31.191: E/TestRunner(21080): failed: testScanning(com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest) 07-20 01:11:31.191: E/TestRunner(21080): ----- begin exception ----- 07-20 01:11:31.192: I/link_layer_controller(1331): Sending scan response 07-20 01:11:31.193: E/TestRunner(21080): java.lang.AssertionError: Expected Exactly(times=1) but got 0 times 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.MockMethodKt.fail(MockMethod.kt:35) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.MockMethodKt.expect(MockMethod.kt:32) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.MockMethodKt.access$expect(MockMethod.kt:1) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:166) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.BaseMethodMock.verify$test_utils_base_debug(MockMethod.kt:158) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.VerifyKt.verify(verify.kt:36) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.mock.VerifyKt.verify$default(verify.kt:18) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invokeSuspend(ScanningStateRepoTest.kt:182) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:8) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest$testScanning$1$9.invoke(Unknown Source:4) 07-20 01:11:31.193: E/TestRunner(21080): at com.splendo.kaluga.test.base.BaseFlowTest$mainAction$3.invokeSuspend(FlowTest.kt:264) 07-20 01:11:31.193: E/TestRunner(21080): at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) 07-20 01:11:31.193: E/TestRunner(21080): at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) 07-20 01:11:31.193: E/TestRunner(21080): at android.os.Handler.handleCallback(Handler.java:883) 07-20 01:11:31.193: E/TestRunner(21080): at android.os.Handler.dispatchMessage(Handler.java:100) 07-20 01:11:31.193: E/TestRunner(21080): at android.os.Looper.loop(Looper.java:214) 07-20 01:11:31.193: E/TestRunner(21080): at android.app.ActivityThread.main(ActivityThread.java:7356) 07-20 01:11:31.193: E/TestRunner(21080): at java.lang.reflect.Method.invoke(Native Method) 07-20 01:11:31.193: E/TestRunner(21080): at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:492) 07-20 01:11:31.193: E/TestRunner(21080): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:930) 07-20 01:11:31.193: E/TestRunner(21080): ----- end exception ----- 07-20 01:11:31.195: I/TestRunner(21080): finished: testScanning(com.splendo.kaluga.bluetooth.scanner.ScanningStateRepoTest) ```
test
unstable test testscanning in com splendo kaluga bluetooth scanner scanningstaterepotest android integration test currently in only java lang assertionerror expected exactly times but got times at com splendo kaluga test base mock mockmethodkt fail mockmethod kt at com splendo kaluga test base mock mockmethodkt expect mockmethod kt at com splendo kaluga test base mock mockmethodkt access expect mockmethod kt at com splendo kaluga test base mock basemethodmock verify test utils base debug mockmethod kt at com splendo kaluga test base mock basemethodmock verify test utils base debug mockmethod kt at com splendo kaluga test base mock verifykt verify verify kt at com splendo kaluga test base mock verifykt verify default verify kt at com splendo kaluga bluetooth scanner scanningstaterepotest testscanning invokesuspend scanningstaterepotest kt at com splendo kaluga bluetooth scanner scanningstaterepotest testscanning invoke unknown source at com splendo kaluga bluetooth scanner scanningstaterepotest testscanning invoke unknown source at com splendo kaluga test base baseflowtest mainaction invokesuspend flowtest kt at kotlin coroutines jvm internal basecontinuationimpl resumewith continuationimpl kt at kotlinx coroutines dispatchedtask run dispatchedtask kt at android os handler handlecallback handler java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at com android internal os runtimeinit methodandargscaller run runtimeinit java at com android internal os zygoteinit main zygoteinit java i testrunner started testscanning com splendo kaluga bluetooth scanner scanningstaterepotest v logkt debug first test offered starting collection v logkt debug launch flow scope v logkt debug wait for main thread to be launched in standalonecoroutine active v logkt debug main scope launched about to flow test channel empty v logkt debug waited for main thread to be launched v logkt debug in collection including this one offering v logkt debug start mainaction v logkt debug await all test blocks give it milliseconds v logkt debug in flow received test channel not empty v logkt debug received test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug ran test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug completed com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl completed v logkt debug handeling value completed v logkt debug in main scope for mainaction v logkt debug in flow received test channel empty v logkt debug did mainaction v logkt debug in collection including this one offering v logkt debug start mainaction v logkt debug await all test blocks give it milliseconds v logkt debug received test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug ran test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug completed com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl completed v logkt debug handeling value completed v logkt debug in main scope for mainaction v logkt debug in flow received test channel empty v logkt debug did mainaction v logkt debug in collection including this one offering v logkt debug start mainaction v logkt debug await all test blocks give it milliseconds v logkt debug received test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug ran test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug completed com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl completed v logkt debug handeling value completed v logkt debug in main scope for mainaction v logkt debug in flow received test channel empty v logkt debug did mainaction v logkt debug in collection including this one offering v logkt debug start action v logkt debug await all test blocks give it milliseconds v logkt debug received test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug ran test block com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl active v logkt debug await all test blocks but none found skip waiting v logkt debug job standalonecoroutine active v logkt debug completed com splendo kaluga bluetooth bluetoothflowtest devicecontext com splendo kaluga bluetooth scanner scanningstate kotlin unit completabledeferredimpl completed v logkt debug handeling value completed v logkt debug ending flow job canceled v logkt debug test channel closed v logkt debug did action v logkt debug start mainaction v logkt debug await all test blocks but none found skip waiting v logkt debug in main scope for mainaction e testrunner failed testscanning com splendo kaluga bluetooth scanner scanningstaterepotest e testrunner begin exception i link layer controller sending scan response e testrunner java lang assertionerror expected exactly times but got times e testrunner at com splendo kaluga test base mock mockmethodkt fail mockmethod kt e testrunner at com splendo kaluga test base mock mockmethodkt expect mockmethod kt e testrunner at com splendo kaluga test base mock mockmethodkt access expect mockmethod kt e testrunner at com splendo kaluga test base mock basemethodmock verify test utils base debug mockmethod kt e testrunner at com splendo kaluga test base mock basemethodmock verify test utils base debug mockmethod kt e testrunner at com splendo kaluga test base mock verifykt verify verify kt e testrunner at com splendo kaluga test base mock verifykt verify default verify kt e testrunner at com splendo kaluga bluetooth scanner scanningstaterepotest testscanning invokesuspend scanningstaterepotest kt e testrunner at com splendo kaluga bluetooth scanner scanningstaterepotest testscanning invoke unknown source e testrunner at com splendo kaluga bluetooth scanner scanningstaterepotest testscanning invoke unknown source e testrunner at com splendo kaluga test base baseflowtest mainaction invokesuspend flowtest kt e testrunner at kotlin coroutines jvm internal basecontinuationimpl resumewith continuationimpl kt e testrunner at kotlinx coroutines dispatchedtask run dispatchedtask kt e testrunner at android os handler handlecallback handler java e testrunner at android os handler dispatchmessage handler java e testrunner at android os looper loop looper java e testrunner at android app activitythread main activitythread java e testrunner at java lang reflect method invoke native method e testrunner at com android internal os runtimeinit methodandargscaller run runtimeinit java e testrunner at com android internal os zygoteinit main zygoteinit java e testrunner end exception i testrunner finished testscanning com splendo kaluga bluetooth scanner scanningstaterepotest
1
265,676
23,188,710,710
IssuesEvent
2022-08-01 10:40:16
redpanda-data/redpanda
https://api.github.com/repos/redpanda-data/redpanda
closed
SchemaRegistryTest.test_protobuf
kind/bug area/schema_registry area/tests ci-failure
https://buildkite.com/redpanda/redpanda/builds/13300#018248a7-cb3a-4031-8a0e-8f4c7187645e ``` Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/ducktape/tests/runner_client.py", line 135, in run data = self.run_test() File "/usr/local/lib/python3.10/dist-packages/ducktape/tests/runner_client.py", line 227, in run_test return self.test_context.function(self.test) File "/root/tests/rptest/services/cluster.py", line 35, in wrapped r = f(self, *args, **kwargs) File "/root/tests/rptest/tests/schema_registry_test.py", line 999, in test_protobuf assert result_raw.status_code == requests.codes.unprocessable_entity AssertionError ```
1.0
SchemaRegistryTest.test_protobuf - https://buildkite.com/redpanda/redpanda/builds/13300#018248a7-cb3a-4031-8a0e-8f4c7187645e ``` Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/ducktape/tests/runner_client.py", line 135, in run data = self.run_test() File "/usr/local/lib/python3.10/dist-packages/ducktape/tests/runner_client.py", line 227, in run_test return self.test_context.function(self.test) File "/root/tests/rptest/services/cluster.py", line 35, in wrapped r = f(self, *args, **kwargs) File "/root/tests/rptest/tests/schema_registry_test.py", line 999, in test_protobuf assert result_raw.status_code == requests.codes.unprocessable_entity AssertionError ```
test
schemaregistrytest test protobuf traceback most recent call last file usr local lib dist packages ducktape tests runner client py line in run data self run test file usr local lib dist packages ducktape tests runner client py line in run test return self test context function self test file root tests rptest services cluster py line in wrapped r f self args kwargs file root tests rptest tests schema registry test py line in test protobuf assert result raw status code requests codes unprocessable entity assertionerror
1
352,082
32,044,242,484
IssuesEvent
2023-09-22 22:44:37
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
sql/tests: TestRandomSyntaxSchemaChangeColumn failed
C-test-failure O-robot branch-master T-sql-foundations
sql/tests.TestRandomSyntaxSchemaChangeColumn [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/11880882?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/11880882?buildTab=artifacts#/) on master @ [6a2097e4c2b6537be84f265bd55ac23d2cf83351](https://github.com/cockroachdb/cockroach/commits/6a2097e4c2b6537be84f265bd55ac23d2cf83351): Random syntax error: ``` rsg_test.go:889: Crash detected: server panic: statement exec timeout ``` Query: ``` ALTER TABLE ident.ident ADD CONSTRAINT IF NOT EXISTS TIES PRIMARY KEY ( DEFAULTS ident DESC NULLS LAST ); ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #111020 sql/tests: TestRandomSyntaxSchemaChangeColumn failed [C-test-failure O-robot T-sql-foundations branch-release-22.2] </p> </details> /cc @cockroachdb/sql-foundations <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRandomSyntaxSchemaChangeColumn.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-31759
1.0
sql/tests: TestRandomSyntaxSchemaChangeColumn failed - sql/tests.TestRandomSyntaxSchemaChangeColumn [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/11880882?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/11880882?buildTab=artifacts#/) on master @ [6a2097e4c2b6537be84f265bd55ac23d2cf83351](https://github.com/cockroachdb/cockroach/commits/6a2097e4c2b6537be84f265bd55ac23d2cf83351): Random syntax error: ``` rsg_test.go:889: Crash detected: server panic: statement exec timeout ``` Query: ``` ALTER TABLE ident.ident ADD CONSTRAINT IF NOT EXISTS TIES PRIMARY KEY ( DEFAULTS ident DESC NULLS LAST ); ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #111020 sql/tests: TestRandomSyntaxSchemaChangeColumn failed [C-test-failure O-robot T-sql-foundations branch-release-22.2] </p> </details> /cc @cockroachdb/sql-foundations <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRandomSyntaxSchemaChangeColumn.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-31759
test
sql tests testrandomsyntaxschemachangecolumn failed sql tests testrandomsyntaxschemachangecolumn with on master random syntax error rsg test go crash detected server panic statement exec timeout query alter table ident ident add constraint if not exists ties primary key defaults ident desc nulls last help see also same failure on other branches sql tests testrandomsyntaxschemachangecolumn failed cc cockroachdb sql foundations jira issue crdb
1
25,233
12,522,073,015
IssuesEvent
2020-06-03 18:30:34
OpenNeuroOrg/openneuro
https://api.github.com/repos/OpenNeuroOrg/openneuro
closed
Migrate existing git trees out of infrequent access tier storage
performance
The fastest way to do this would be a maintenance window to rsync them to another volume and back. This is a small subset of data, so the window could be fairly short. See #1458 for fixing this going forward.
True
Migrate existing git trees out of infrequent access tier storage - The fastest way to do this would be a maintenance window to rsync them to another volume and back. This is a small subset of data, so the window could be fairly short. See #1458 for fixing this going forward.
non_test
migrate existing git trees out of infrequent access tier storage the fastest way to do this would be a maintenance window to rsync them to another volume and back this is a small subset of data so the window could be fairly short see for fixing this going forward
0
802,641
29,042,869,449
IssuesEvent
2023-05-13 06:50:59
IdkwhatImD0ing/SlugLoop
https://api.github.com/repos/IdkwhatImD0ing/SlugLoop
closed
Implement Route Toggle Feature to Display Buses and Route Lines for Selected Routes
wontfix low-priority long-term
# Description: Building upon the previous issue of adding a JSON file for common bus routes, we need to enhance the web app by providing users with the ability to toggle between different routes. This feature should display only the buses corresponding to the selected route and show a line on the map representing the route path. We can utilize the JSON file provided in issue #41 to achieve this. # Issue Details: 1. No route toggle functionality: Users currently cannot switch between different bus routes and view the corresponding buses in the web app. 2. Missing route lines: The web app does not display a visual representation of the bus routes on the map. # Proposed Solution: 1. Implement route toggle feature: Add a user interface element, such as a dropdown menu or buttons, that allows users to select a specific bus route. 2. Filter displayed buses: When a route is selected, update the web app to show only the buses corresponding to the chosen route. 3. Display route lines: Upon selecting a route, display a line on the map that represents the route path, providing users with a visual guide of the bus journey. 4. Utilize JSON file from issue #41: Integrate the JSON file provided in issue #41 to store the route data, including the bus stops and route coordinates for displaying route lines. # Acceptance Criteria: 1. Users can toggle between different bus routes using a dedicated interface element in the web app. 2. When a route is selected, only the buses corresponding to that route are displayed. 3. A line representing the chosen bus route is shown on the map, providing users with a visual understanding of the route path. 4. The JSON file from issue #41 is effectively utilized for storing and managing route data.
1.0
Implement Route Toggle Feature to Display Buses and Route Lines for Selected Routes - # Description: Building upon the previous issue of adding a JSON file for common bus routes, we need to enhance the web app by providing users with the ability to toggle between different routes. This feature should display only the buses corresponding to the selected route and show a line on the map representing the route path. We can utilize the JSON file provided in issue #41 to achieve this. # Issue Details: 1. No route toggle functionality: Users currently cannot switch between different bus routes and view the corresponding buses in the web app. 2. Missing route lines: The web app does not display a visual representation of the bus routes on the map. # Proposed Solution: 1. Implement route toggle feature: Add a user interface element, such as a dropdown menu or buttons, that allows users to select a specific bus route. 2. Filter displayed buses: When a route is selected, update the web app to show only the buses corresponding to the chosen route. 3. Display route lines: Upon selecting a route, display a line on the map that represents the route path, providing users with a visual guide of the bus journey. 4. Utilize JSON file from issue #41: Integrate the JSON file provided in issue #41 to store the route data, including the bus stops and route coordinates for displaying route lines. # Acceptance Criteria: 1. Users can toggle between different bus routes using a dedicated interface element in the web app. 2. When a route is selected, only the buses corresponding to that route are displayed. 3. A line representing the chosen bus route is shown on the map, providing users with a visual understanding of the route path. 4. The JSON file from issue #41 is effectively utilized for storing and managing route data.
non_test
implement route toggle feature to display buses and route lines for selected routes description building upon the previous issue of adding a json file for common bus routes we need to enhance the web app by providing users with the ability to toggle between different routes this feature should display only the buses corresponding to the selected route and show a line on the map representing the route path we can utilize the json file provided in issue to achieve this issue details no route toggle functionality users currently cannot switch between different bus routes and view the corresponding buses in the web app missing route lines the web app does not display a visual representation of the bus routes on the map proposed solution implement route toggle feature add a user interface element such as a dropdown menu or buttons that allows users to select a specific bus route filter displayed buses when a route is selected update the web app to show only the buses corresponding to the chosen route display route lines upon selecting a route display a line on the map that represents the route path providing users with a visual guide of the bus journey utilize json file from issue integrate the json file provided in issue to store the route data including the bus stops and route coordinates for displaying route lines acceptance criteria users can toggle between different bus routes using a dedicated interface element in the web app when a route is selected only the buses corresponding to that route are displayed a line representing the chosen bus route is shown on the map providing users with a visual understanding of the route path the json file from issue is effectively utilized for storing and managing route data
0
151,106
13,390,775,279
IssuesEvent
2020-09-02 21:10:09
google/generic-webdriver-server
https://api.github.com/repos/google/generic-webdriver-server
opened
Document requirements for iframe embedding for Chromecast receiver
documentation
Sites can prevent iframe-embedding with the [X-Frame-Options](https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-Frame-Options) header. Though this should not be an issue for a test runner, one of the first things a person might try with the command-line tools is: ```sh chromecast-webdriver-cli --hostname=chromecast-hostname --url=https://www.google.com/ ``` However, google.com prevents iframe embedding, which means it cannot be shown in our Chromecast receiver. The user would see a blank screen instead. This should be clearly documented to avoid confusion.
1.0
Document requirements for iframe embedding for Chromecast receiver - Sites can prevent iframe-embedding with the [X-Frame-Options](https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-Frame-Options) header. Though this should not be an issue for a test runner, one of the first things a person might try with the command-line tools is: ```sh chromecast-webdriver-cli --hostname=chromecast-hostname --url=https://www.google.com/ ``` However, google.com prevents iframe embedding, which means it cannot be shown in our Chromecast receiver. The user would see a blank screen instead. This should be clearly documented to avoid confusion.
non_test
document requirements for iframe embedding for chromecast receiver sites can prevent iframe embedding with the header though this should not be an issue for a test runner one of the first things a person might try with the command line tools is sh chromecast webdriver cli hostname chromecast hostname url however google com prevents iframe embedding which means it cannot be shown in our chromecast receiver the user would see a blank screen instead this should be clearly documented to avoid confusion
0
307,189
26,518,717,896
IssuesEvent
2023-01-18 23:32:55
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_write_to_closures_in_inlining_dynamic_shapes (torch._dynamo.testing.make_test_cls_with_patches.<locals>.DummyTestClass)
module: flaky-tests skipped module: unknown
Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_write_to_closures_in_inlining_dynamic_shapes) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/10711572477). Over the past 72 hours, it has flakily failed in 4 workflow(s). **Debugging instructions (after clicking on the recent samples link):** To find relevant log snippets: 1. Click on the workflow logs linked above 2. Grep for `test_write_to_closures_in_inlining_dynamic_shapes` Error retrieving /opt/conda/lib/python3.10/site-packages/torch/_dynamo/testing.py: Error: Statuscode 301
1.0
DISABLED test_write_to_closures_in_inlining_dynamic_shapes (torch._dynamo.testing.make_test_cls_with_patches.<locals>.DummyTestClass) - Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_write_to_closures_in_inlining_dynamic_shapes) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/10711572477). Over the past 72 hours, it has flakily failed in 4 workflow(s). **Debugging instructions (after clicking on the recent samples link):** To find relevant log snippets: 1. Click on the workflow logs linked above 2. Grep for `test_write_to_closures_in_inlining_dynamic_shapes` Error retrieving /opt/conda/lib/python3.10/site-packages/torch/_dynamo/testing.py: Error: Statuscode 301
test
disabled test write to closures in inlining dynamic shapes torch dynamo testing make test cls with patches dummytestclass platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has flakily failed in workflow s debugging instructions after clicking on the recent samples link to find relevant log snippets click on the workflow logs linked above grep for test write to closures in inlining dynamic shapes error retrieving opt conda lib site packages torch dynamo testing py error statuscode
1
516,188
14,976,763,065
IssuesEvent
2021-01-28 08:33:41
magento/magento2
https://api.github.com/repos/magento/magento2
closed
Config data is re-encrypted when saved in the back office on Magento 2210
Component: Config Issue: Confirmed Issue: Format is valid Issue: Ready for Work Issue: needs update Priority: P2 Progress: dev in progress Reproduced on 2.4.x Severity: S2 Triage: Dev.Experience
### Preconditions 1. Magento 2.4-develop ### Steps to reproduce 1. Create a module with this a back office form, with such a field (system.xml): ``` [...] <field id="hmackey" translate="label comment" type="text" sortOrder="50" showInDefault="1" showInWebsite="1" showInStore="1"> <label>HMAC</label> <comment><![CDATA[...]]></comment> <validate>required-entry</validate> <config_path>foo/bar/hmackey</config_path> <backend_model>Magento\Config\Model\Config\Backend\Encrypted</backend_model> </field> [...] ``` 2. Go to the admin on the form page 3. The data is correctly shown, save the form ### Expected result 1. Data is correctly encrypted and saved in the database 2. De-crypted data is shown in the admin form ### Actual result 1. Data is correctly encrypted and saved in the database 2. EN-crypted data is shown in the admin form ### Notes I've tried the same configuration on previous versions and can't reproduce this issue, only on 2210. I can't see anything in the changelog between 2200 and 2210 that would explain such a behavior. Any ideas?
1.0
Config data is re-encrypted when saved in the back office on Magento 2210 - ### Preconditions 1. Magento 2.4-develop ### Steps to reproduce 1. Create a module with this a back office form, with such a field (system.xml): ``` [...] <field id="hmackey" translate="label comment" type="text" sortOrder="50" showInDefault="1" showInWebsite="1" showInStore="1"> <label>HMAC</label> <comment><![CDATA[...]]></comment> <validate>required-entry</validate> <config_path>foo/bar/hmackey</config_path> <backend_model>Magento\Config\Model\Config\Backend\Encrypted</backend_model> </field> [...] ``` 2. Go to the admin on the form page 3. The data is correctly shown, save the form ### Expected result 1. Data is correctly encrypted and saved in the database 2. De-crypted data is shown in the admin form ### Actual result 1. Data is correctly encrypted and saved in the database 2. EN-crypted data is shown in the admin form ### Notes I've tried the same configuration on previous versions and can't reproduce this issue, only on 2210. I can't see anything in the changelog between 2200 and 2210 that would explain such a behavior. Any ideas?
non_test
config data is re encrypted when saved in the back office on magento preconditions magento develop steps to reproduce create a module with this a back office form with such a field system xml field id hmackey translate label comment type text sortorder showindefault showinwebsite showinstore hmac required entry foo bar hmackey magento config model config backend encrypted go to the admin on the form page the data is correctly shown save the form expected result data is correctly encrypted and saved in the database de crypted data is shown in the admin form actual result data is correctly encrypted and saved in the database en crypted data is shown in the admin form notes i ve tried the same configuration on previous versions and can t reproduce this issue only on i can t see anything in the changelog between and that would explain such a behavior any ideas
0
188,101
22,046,151,458
IssuesEvent
2022-05-30 02:06:25
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
closed
CVE-2019-15291 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed
security vulnerability
## CVE-2019-15291 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/usb/b2c2/flexcop-usb.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/usb/b2c2/flexcop-usb.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel through 5.2.9. There is a NULL pointer dereference caused by a malicious USB device in the flexcop_usb_probe function in the drivers/media/usb/b2c2/flexcop-usb.c driver. <p>Publish Date: 2019-08-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15291>CVE-2019-15291</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2019-15291">https://www.linuxkernelcves.com/cves/CVE-2019-15291</a></p> <p>Release Date: 2019-09-06</p> <p>Fix Resolution: v5.5-rc1,v3.16.79,v4.14.157,v4.19.87,v4.4.204,v4.9.204,v5.3.14,v5.4.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-15291 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed - ## CVE-2019-15291 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/usb/b2c2/flexcop-usb.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/usb/b2c2/flexcop-usb.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel through 5.2.9. There is a NULL pointer dereference caused by a malicious USB device in the flexcop_usb_probe function in the drivers/media/usb/b2c2/flexcop-usb.c driver. <p>Publish Date: 2019-08-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15291>CVE-2019-15291</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2019-15291">https://www.linuxkernelcves.com/cves/CVE-2019-15291</a></p> <p>Release Date: 2019-09-06</p> <p>Fix Resolution: v5.5-rc1,v3.16.79,v4.14.157,v4.19.87,v4.4.204,v4.9.204,v5.3.14,v5.4.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers media usb flexcop usb c drivers media usb flexcop usb c vulnerability details an issue was discovered in the linux kernel through there is a null pointer dereference caused by a malicious usb device in the flexcop usb probe function in the drivers media usb flexcop usb c driver publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
527,309
15,339,629,750
IssuesEvent
2021-02-27 02:49:46
RubyMoney/money
https://api.github.com/repos/RubyMoney/money
closed
Explicit +/- change breaks existing API
priority:low
After https://github.com/RubyMoney/money/pull/852 has been merged, classes not inheriting Money cannot be handled anymore. I've got a codebase with the following class: ``` class Pricing class NullPrice < SimpleDelegator def initialize(item_id=nil) super(Money.zero) LogContext.add(null_priced_item_id: item_id) end def format(*_rules) "$N/A" end alias_method :to_s, :format alias_method :centless_format, :format def inspect "#<#{self.class.name}>" end end end ``` Which worked with the version `6.13.2`, but breaks in `6.13.3` when testing +/-: ``` Failure/Error: expect(Money.new(1_00) - subject).to be_eql Money.new(1_00) Unsupported argument type: Pricing::NullPrice ``` While this is possible to patch to work with the new update, I don't think a point release should break existing code. I think the new approach in general has an issue where the new object is constructed via `self.class.new` which gives inconsistent results: `Foo.new + Money.new -> Foo`, but `Money.new + Foo.new -> Money`
1.0
Explicit +/- change breaks existing API - After https://github.com/RubyMoney/money/pull/852 has been merged, classes not inheriting Money cannot be handled anymore. I've got a codebase with the following class: ``` class Pricing class NullPrice < SimpleDelegator def initialize(item_id=nil) super(Money.zero) LogContext.add(null_priced_item_id: item_id) end def format(*_rules) "$N/A" end alias_method :to_s, :format alias_method :centless_format, :format def inspect "#<#{self.class.name}>" end end end ``` Which worked with the version `6.13.2`, but breaks in `6.13.3` when testing +/-: ``` Failure/Error: expect(Money.new(1_00) - subject).to be_eql Money.new(1_00) Unsupported argument type: Pricing::NullPrice ``` While this is possible to patch to work with the new update, I don't think a point release should break existing code. I think the new approach in general has an issue where the new object is constructed via `self.class.new` which gives inconsistent results: `Foo.new + Money.new -> Foo`, but `Money.new + Foo.new -> Money`
non_test
explicit change breaks existing api after has been merged classes not inheriting money cannot be handled anymore i ve got a codebase with the following class class pricing class nullprice simpledelegator def initialize item id nil super money zero logcontext add null priced item id item id end def format rules n a end alias method to s format alias method centless format format def inspect end end end which worked with the version but breaks in when testing failure error expect money new subject to be eql money new unsupported argument type pricing nullprice while this is possible to patch to work with the new update i don t think a point release should break existing code i think the new approach in general has an issue where the new object is constructed via self class new which gives inconsistent results foo new money new foo but money new foo new money
0
432,034
30,264,328,470
IssuesEvent
2023-07-07 10:35:11
Eu4ng/TIL
https://api.github.com/repos/Eu4ng/TIL
opened
[UE5] 데이터 애셋을 할당하거나 인스턴스를 생성하는 예제
documentation
```cpp // Fill out your copyright notice in the Description page of Project Settings. #pragma once #include "CoreMinimal.h" #include "Engine/DataAsset.h" #include "DataAsset_CharacterAction.generated.h" USTRUCT(BlueprintType) struct FCharacterAction { GENERATED_BODY() UPROPERTY(EditDefaultsOnly) bool bUseInstance; UPROPERTY(EditDefaultsOnly, meta=(EditCondition="!bUseInstance", EditConditionHides)) UDataAsset_CharacterAction* DataAsset; UPROPERTY(EditDefaultsOnly, Instanced, meta=(EditCondition="bUseInstance", EditConditionHides)) UDataAsset_CharacterAction* Instance; }; /** * 무기 종류에 따른 애님 몽타주를 설정하는 데이터 애셋 */ UCLASS(EditInlineNew, DisplayName="CharacterAction") class RPGCORE_API UDataAsset_CharacterAction : public UDataAsset { GENERATED_BODY() UPROPERTY(EditDefaultsOnly, BlueprintGetter=GetSkeleton) USkeleton* Skeleton; UPROPERTY(EditDefaultsOnly, BlueprintGetter=GetAttackMontage) UAnimMontage* AttackMontage; public: UFUNCTION(BlueprintGetter) FORCEINLINE USkeleton* GetSkeleton() const { return Skeleton; } UFUNCTION(BlueprintGetter) FORCEINLINE UAnimMontage* GetAttackMontage() const { return AttackMontage; } UFUNCTION(BlueprintPure) static FORCEINLINE UDataAsset_CharacterAction* Get(const FCharacterAction& CharacterAction) { return CharacterAction.bUseInstance ? CharacterAction.Instance : CharacterAction.DataAsset; } }; ```
1.0
[UE5] 데이터 애셋을 할당하거나 인스턴스를 생성하는 예제 - ```cpp // Fill out your copyright notice in the Description page of Project Settings. #pragma once #include "CoreMinimal.h" #include "Engine/DataAsset.h" #include "DataAsset_CharacterAction.generated.h" USTRUCT(BlueprintType) struct FCharacterAction { GENERATED_BODY() UPROPERTY(EditDefaultsOnly) bool bUseInstance; UPROPERTY(EditDefaultsOnly, meta=(EditCondition="!bUseInstance", EditConditionHides)) UDataAsset_CharacterAction* DataAsset; UPROPERTY(EditDefaultsOnly, Instanced, meta=(EditCondition="bUseInstance", EditConditionHides)) UDataAsset_CharacterAction* Instance; }; /** * 무기 종류에 따른 애님 몽타주를 설정하는 데이터 애셋 */ UCLASS(EditInlineNew, DisplayName="CharacterAction") class RPGCORE_API UDataAsset_CharacterAction : public UDataAsset { GENERATED_BODY() UPROPERTY(EditDefaultsOnly, BlueprintGetter=GetSkeleton) USkeleton* Skeleton; UPROPERTY(EditDefaultsOnly, BlueprintGetter=GetAttackMontage) UAnimMontage* AttackMontage; public: UFUNCTION(BlueprintGetter) FORCEINLINE USkeleton* GetSkeleton() const { return Skeleton; } UFUNCTION(BlueprintGetter) FORCEINLINE UAnimMontage* GetAttackMontage() const { return AttackMontage; } UFUNCTION(BlueprintPure) static FORCEINLINE UDataAsset_CharacterAction* Get(const FCharacterAction& CharacterAction) { return CharacterAction.bUseInstance ? CharacterAction.Instance : CharacterAction.DataAsset; } }; ```
non_test
데이터 애셋을 할당하거나 인스턴스를 생성하는 예제 cpp fill out your copyright notice in the description page of project settings pragma once include coreminimal h include engine dataasset h include dataasset characteraction generated h ustruct blueprinttype struct fcharacteraction generated body uproperty editdefaultsonly bool buseinstance uproperty editdefaultsonly meta editcondition buseinstance editconditionhides udataasset characteraction dataasset uproperty editdefaultsonly instanced meta editcondition buseinstance editconditionhides udataasset characteraction instance 무기 종류에 따른 애님 몽타주를 설정하는 데이터 애셋 uclass editinlinenew displayname characteraction class rpgcore api udataasset characteraction public udataasset generated body uproperty editdefaultsonly blueprintgetter getskeleton uskeleton skeleton uproperty editdefaultsonly blueprintgetter getattackmontage uanimmontage attackmontage public ufunction blueprintgetter forceinline uskeleton getskeleton const return skeleton ufunction blueprintgetter forceinline uanimmontage getattackmontage const return attackmontage ufunction blueprintpure static forceinline udataasset characteraction get const fcharacteraction characteraction return characteraction buseinstance characteraction instance characteraction dataasset
0
34,455
4,927,210,621
IssuesEvent
2016-11-26 16:16:51
khartec/waltz
https://api.github.com/repos/khartec/waltz
closed
High Phys Flows: Lineage Edit page needlessly retrieves all phys flows for an app
fixed (test & close) performance
In order to render the candidate sources, all flows either sourcing from or to an application are retrieved. This is unnecessary as we only need the flows incoming into the application. The api calls used are: http://localhost:8443/api/physical-specification/application/APPLICATION/1746 http://localhost:8443/api/physical-flow/entity/APPLICATION/1746 (this one is expensive when lots of flows in system)
1.0
High Phys Flows: Lineage Edit page needlessly retrieves all phys flows for an app - In order to render the candidate sources, all flows either sourcing from or to an application are retrieved. This is unnecessary as we only need the flows incoming into the application. The api calls used are: http://localhost:8443/api/physical-specification/application/APPLICATION/1746 http://localhost:8443/api/physical-flow/entity/APPLICATION/1746 (this one is expensive when lots of flows in system)
test
high phys flows lineage edit page needlessly retrieves all phys flows for an app in order to render the candidate sources all flows either sourcing from or to an application are retrieved this is unnecessary as we only need the flows incoming into the application the api calls used are this one is expensive when lots of flows in system
1
87,739
8,120,627,907
IssuesEvent
2018-08-16 03:57:08
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
teamcity: failed tests on master: testrace/TestShowSessions, lint/TestLint
C-test-failure O-robot
The following tests appear to have failed: [#839592](https://teamcity.cockroachdb.com/viewLog.html?buildId=839592): ``` --- FAIL: testrace/TestShowSessions (2.450s) show_test.go:684: unexpected number of running sessions: 2, expected 1. Active sessions (results might have changed since the test checked): app: "$ internal-log-range-event", query: "INSERT INTO system.public.rangelog(\"timestamp\", \"rangeID\", \"storeID\", \"eventType\", \"otherRangeID\", info) VALUES ($1, $2, $3, $4, $5, $6)", last query: app: "", query: "SELECT active_queries, last_active_query, application_name FROM [SHOW CLUSTER SESSIONS]", last query: SELECT node_id, (now() - session_start)::FLOAT FROM [SHOW CLUSTER SESSIONS] WHERE application_name NOT LIKE 'internal-%' ------- Stdout: ------- W180815 20:40:34.423397 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180815 20:40:34.450676 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180815 20:40:34.451091 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:34.451159 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:34.455752 158099 server/config.go:496 [n?] 1 storage engine initialized I180815 20:40:34.456002 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180815 20:40:34.456043 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B I180815 20:40:34.514781 158099 server/node.go:373 [n?] **** cluster 6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 has been created I180815 20:40:34.515045 158099 server/server.go:1398 [n?] **** add additional nodes by specifying --join=127.0.0.1:41091 I180815 20:40:34.516516 158099 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41091" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365634515938360 I180815 20:40:34.534941 158099 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180815 20:40:34.537635 158099 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00} I180815 20:40:34.538394 158099 storage/stores.go:222 [n1] read 0 node addresses from persistent storage I180815 20:40:34.539206 158099 server/node.go:697 [n1] connecting to gossip network to verify cluster ID... I180815 20:40:34.539582 158099 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102" I180815 20:40:34.539943 158099 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180815 20:40:34.541685 158099 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:34.541909 158099 server/server.go:1833 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined I180815 20:40:34.543736 158325 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:34.545405 158099 server/server.go:1535 [n1] starting https server at 127.0.0.1:34275 (use: 127.0.0.1:34275) I180815 20:40:34.545680 158099 server/server.go:1537 [n1] starting grpc/postgres server at 127.0.0.1:41091 I180815 20:40:34.545753 158099 server/server.go:1538 [n1] advertising CockroachDB node at 127.0.0.1:41091 I180815 20:40:34.566927 158335 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180815 20:40:34.719270 158304 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] I180815 20:40:34.817666 158087 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] I180815 20:40:34.840768 158068 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root} I180815 20:40:34.933807 158160 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180815 20:40:35.038918 158386 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180815 20:40:35.116602 158341 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root} W180815 20:40:35.128647 158374 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=5ddfb628 key=/Local/Range/System/tsd/RangeDescriptor rw=true pri=0.02614628 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534365635.039580425,0 orig=1534365635.039580425,0 max=1534365635.039580425,0 wto=false rop=false seq=1 I180815 20:40:35.147031 158074 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180815 20:40:35.241602 158379 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] I180815 20:40:35.285924 158289 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root} I180815 20:40:35.328157 158390 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180815 20:40:35.401113 158408 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180815 20:40:35.470572 158398 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180815 20:40:35.530967 158452 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180815 20:40:35.548516 158385 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root} I180815 20:40:35.603367 158447 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180815 20:40:35.633708 158442 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established I180815 20:40:35.654186 158514 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root} I180815 20:40:35.677178 158463 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180815 20:40:35.720541 158458 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root} I180815 20:40:35.744993 158521 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180815 20:40:35.765024 158099 server/server.go:1620 [n1] done ensuring all necessary migrations have run I180815 20:40:35.765364 158099 server/server.go:1623 [n1] serving sql connections I180815 20:40:35.811858 158481 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41091} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365634515938360 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365634515938360 LastUp:1534365634515938360} I180815 20:40:35.813481 158479 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version I180815 20:40:35.820374 158564 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180815 20:40:35.889664 158497 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180815 20:40:35.951092 158556 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180815 20:40:36.009944 158598 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180815 20:40:36.059947 158619 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] I180815 20:40:36.119763 158606 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180815 20:40:36.175659 158630 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] W180815 20:40:36.287477 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180815 20:40:36.315831 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180815 20:40:36.316449 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:36.316518 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:36.321488 158099 server/config.go:496 [n?] 1 storage engine initialized I180815 20:40:36.321673 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180815 20:40:36.321701 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B W180815 20:40:36.321999 158099 gossip/gossip.go:1351 [n?] no incoming or outgoing connections I180815 20:40:36.322565 158099 server/server.go:1400 [n?] no stores bootstrapped and --join flag specified, awaiting init command. I180815 20:40:36.398063 158665 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:41091 I180815 20:40:36.399347 158763 gossip/server.go:219 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:32789} I180815 20:40:36.404521 158099 server/node.go:697 [n?] connecting to gossip network to verify cluster ID... I180815 20:40:36.405108 158099 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102" I180815 20:40:36.408670 158717 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:40:36.415122 158716 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:40:36.422331 158099 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:40:36.431430 158099 server/node.go:428 [n?] new node allocated ID 2 I180815 20:40:36.432094 158099 gossip/gossip.go:383 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:32789" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365636431659703 I180815 20:40:36.433009 158099 storage/stores.go:222 [n2] read 0 node addresses from persistent storage I180815 20:40:36.433961 158099 storage/stores.go:241 [n2] wrote 1 node addresses to persistent storage I180815 20:40:36.436697 158720 storage/stores.go:241 [n1] wrote 1 node addresses to persistent storage I180815 20:40:36.457872 158099 server/node.go:672 [n2] bootstrapped store [n2,s2] I180815 20:40:36.460045 158099 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes [] I180815 20:40:36.461355 158099 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:36.461544 158099 server/server.go:1833 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined I180815 20:40:36.465410 158099 server/server.go:1535 [n2] starting https server at 127.0.0.1:36335 (use: 127.0.0.1:36335) I180815 20:40:36.465633 158099 server/server.go:1537 [n2] starting grpc/postgres server at 127.0.0.1:32789 I180815 20:40:36.465700 158099 server/server.go:1538 [n2] advertising CockroachDB node at 127.0.0.1:32789 I180815 20:40:36.468619 158867 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:36.476134 158099 server/server.go:1620 [n2] done ensuring all necessary migrations have run I180815 20:40:36.476492 158099 server/server.go:1623 [n2] serving sql connections I180815 20:40:36.576176 158869 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version I180815 20:40:36.577336 158871 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:32789} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365636431659703 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365636431659703 LastUp:1534365636431659703} I180815 20:40:36.617219 158877 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established I180815 20:40:36.623403 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root} I180815 20:40:36.624884 158702 storage/replica_command.go:282 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23] I180815 20:40:36.685520 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_table", target: 53, info: {TableName:test.public.t Statement:CREATE TABLE test.public.t (num INT) User:root} I180815 20:40:36.689918 158686 storage/replica_command.go:282 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24] I180815 20:40:36.729831 158942 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] transport racer 1 [async] storage.split: processing replica 1 [async] kv.TxnCoordSender: heartbeat loop W180815 20:40:36.732234 158686 internal/client/txn.go:514 [split,n1,s1,r23/1:/{Table/52-Max}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer E180815 20:40:36.733096 158686 storage/queue.go:788 [split,n1,s1,r23/1:/{Table/52-Max}] unable to split [n1,s1,r23/1:/{Table/52-Max}] at key "/Table/53": split at key /Table/53 failed: node unavailable; try another peer I180815 20:40:36.733743 158942 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] transport racer I180815 20:40:36.789547 158280 kv/transport_race.go:67 transport race promotion: ran 39 iterations on up to 877 requests --- FAIL: lint/TestLint (146.790s) --- FAIL: lint/TestLint: TestLint/TestHelpURLs (41.520s) lint_test.go:945: https://www.cockroachlabs.com/docs/v2.1/set-vars.html : Head https://www.cockroachlabs.com/docs/v2.1/set-vars.html: EOF 324: https://www.cockroachlabs.com/docs/v2.1/set-vars.html 990: https://www.cockroachlabs.com/docs/v2.1/set-vars.html 1245:https://www.cockroachlabs.com/docs/v2.1/set-vars.html https://www.cockroachlabs.com/docs/v2.1/create-table-as.html : Head https://www.cockroachlabs.com/docs/v2.1/create-table-as.html: EOF 241:https://www.cockroachlabs.com/docs/v2.1/create-table-as.html 2 errors ------- Stdout: ------- === PAUSE TestLint/TestHelpURLs --- FAIL: lint/TestLint (146.790s) --- FAIL: lint/TestLint: TestLint/TestHelpURLs (41.520s) lint_test.go:945: https://www.cockroachlabs.com/docs/v2.1/set-vars.html : Head https://www.cockroachlabs.com/docs/v2.1/set-vars.html: EOF 324: https://www.cockroachlabs.com/docs/v2.1/set-vars.html 990: https://www.cockroachlabs.com/docs/v2.1/set-vars.html 1245:https://www.cockroachlabs.com/docs/v2.1/set-vars.html https://www.cockroachlabs.com/docs/v2.1/create-table-as.html : Head https://www.cockroachlabs.com/docs/v2.1/create-table-as.html: EOF 241:https://www.cockroachlabs.com/docs/v2.1/create-table-as.html 2 errors ------- Stdout: ------- === PAUSE TestLint/TestHelpURLs --- FAIL: testrace/TestShowSessions (3.280s) show_test.go:684: unexpected number of running sessions: 2, expected 1. Active sessions (results might have changed since the test checked): app: "$ internal-log-range-event", query: "INSERT INTO system.public.rangelog(\"timestamp\", \"rangeID\", \"storeID\", \"eventType\", \"otherRangeID\", info) VALUES ($1, $2, $3, $4, $5, $6)", last query: app: "", query: "SELECT active_queries, last_active_query, application_name FROM [SHOW CLUSTER SESSIONS]", last query: SELECT node_id, (now() - session_start)::FLOAT FROM [SHOW CLUSTER SESSIONS] WHERE application_name NOT LIKE 'internal-%' ------- Stdout: ------- W180815 20:40:34.423397 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180815 20:40:34.450676 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180815 20:40:34.451091 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:34.451159 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:34.455752 158099 server/config.go:496 [n?] 1 storage engine initialized I180815 20:40:34.456002 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180815 20:40:34.456043 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B I180815 20:40:34.514781 158099 server/node.go:373 [n?] **** cluster 6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 has been created I180815 20:40:34.515045 158099 server/server.go:1398 [n?] **** add additional nodes by specifying --join=127.0.0.1:41091 I180815 20:40:34.516516 158099 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41091" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365634515938360 I180815 20:40:34.534941 158099 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180815 20:40:34.537635 158099 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00} I180815 20:40:34.538394 158099 storage/stores.go:222 [n1] read 0 node addresses from persistent storage I180815 20:40:34.539206 158099 server/node.go:697 [n1] connecting to gossip network to verify cluster ID... I180815 20:40:34.539582 158099 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102" I180815 20:40:34.539943 158099 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180815 20:40:34.541685 158099 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:34.541909 158099 server/server.go:1833 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined I180815 20:40:34.543736 158325 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:34.545405 158099 server/server.go:1535 [n1] starting https server at 127.0.0.1:34275 (use: 127.0.0.1:34275) I180815 20:40:34.545680 158099 server/server.go:1537 [n1] starting grpc/postgres server at 127.0.0.1:41091 I180815 20:40:34.545753 158099 server/server.go:1538 [n1] advertising CockroachDB node at 127.0.0.1:41091 I180815 20:40:34.566927 158335 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180815 20:40:34.719270 158304 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] I180815 20:40:34.817666 158087 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] I180815 20:40:34.840768 158068 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root} I180815 20:40:34.933807 158160 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180815 20:40:35.038918 158386 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180815 20:40:35.116602 158341 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root} W180815 20:40:35.128647 158374 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=5ddfb628 key=/Local/Range/System/tsd/RangeDescriptor rw=true pri=0.02614628 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534365635.039580425,0 orig=1534365635.039580425,0 max=1534365635.039580425,0 wto=false rop=false seq=1 I180815 20:40:35.147031 158074 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180815 20:40:35.241602 158379 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] I180815 20:40:35.285924 158289 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root} I180815 20:40:35.328157 158390 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180815 20:40:35.401113 158408 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180815 20:40:35.470572 158398 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180815 20:40:35.530967 158452 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180815 20:40:35.548516 158385 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root} I180815 20:40:35.603367 158447 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180815 20:40:35.633708 158442 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established I180815 20:40:35.654186 158514 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root} I180815 20:40:35.677178 158463 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180815 20:40:35.720541 158458 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root} I180815 20:40:35.744993 158521 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180815 20:40:35.765024 158099 server/server.go:1620 [n1] done ensuring all necessary migrations have run I180815 20:40:35.765364 158099 server/server.go:1623 [n1] serving sql connections I180815 20:40:35.811858 158481 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41091} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365634515938360 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365634515938360 LastUp:1534365634515938360} I180815 20:40:35.813481 158479 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version I180815 20:40:35.820374 158564 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180815 20:40:35.889664 158497 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180815 20:40:35.951092 158556 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180815 20:40:36.009944 158598 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180815 20:40:36.059947 158619 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] I180815 20:40:36.119763 158606 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180815 20:40:36.175659 158630 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] W180815 20:40:36.287477 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180815 20:40:36.315831 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180815 20:40:36.316449 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:36.316518 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:36.321488 158099 server/config.go:496 [n?] 1 storage engine initialized I180815 20:40:36.321673 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180815 20:40:36.321701 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B W180815 20:40:36.321999 158099 gossip/gossip.go:1351 [n?] no incoming or outgoing connections I180815 20:40:36.322565 158099 server/server.go:1400 [n?] no stores bootstrapped and --join flag specified, awaiting init command. I180815 20:40:36.398063 158665 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:41091 I180815 20:40:36.399347 158763 gossip/server.go:219 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:32789} I180815 20:40:36.404521 158099 server/node.go:697 [n?] connecting to gossip network to verify cluster ID... I180815 20:40:36.405108 158099 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102" I180815 20:40:36.408670 158717 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:40:36.415122 158716 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:40:36.422331 158099 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:40:36.431430 158099 server/node.go:428 [n?] new node allocated ID 2 I180815 20:40:36.432094 158099 gossip/gossip.go:383 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:32789" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365636431659703 I180815 20:40:36.433009 158099 storage/stores.go:222 [n2] read 0 node addresses from persistent storage I180815 20:40:36.433961 158099 storage/stores.go:241 [n2] wrote 1 node addresses to persistent storage I180815 20:40:36.436697 158720 storage/stores.go:241 [n1] wrote 1 node addresses to persistent storage I180815 20:40:36.457872 158099 server/node.go:672 [n2] bootstrapped store [n2,s2] I180815 20:40:36.460045 158099 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes [] I180815 20:40:36.461355 158099 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:36.461544 158099 server/server.go:1833 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined I180815 20:40:36.465410 158099 server/server.go:1535 [n2] starting https server at 127.0.0.1:36335 (use: 127.0.0.1:36335) I180815 20:40:36.465633 158099 server/server.go:1537 [n2] starting grpc/postgres server at 127.0.0.1:32789 I180815 20:40:36.465700 158099 server/server.go:1538 [n2] advertising CockroachDB node at 127.0.0.1:32789 I180815 20:40:36.468619 158867 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:36.476134 158099 server/server.go:1620 [n2] done ensuring all necessary migrations have run I180815 20:40:36.476492 158099 server/server.go:1623 [n2] serving sql connections I180815 20:40:36.576176 158869 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version I180815 20:40:36.577336 158871 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:32789} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365636431659703 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365636431659703 LastUp:1534365636431659703} I180815 20:40:36.617219 158877 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established I180815 20:40:36.623403 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root} I180815 20:40:36.624884 158702 storage/replica_command.go:282 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23] I180815 20:40:36.685520 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_table", target: 53, info: {TableName:test.public.t Statement:CREATE TABLE test.public.t (num INT) User:root} I180815 20:40:36.689918 158686 storage/replica_command.go:282 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24] I180815 20:40:36.729831 158942 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] transport racer 1 [async] storage.split: processing replica 1 [async] kv.TxnCoordSender: heartbeat loop W180815 20:40:36.732234 158686 internal/client/txn.go:514 [split,n1,s1,r23/1:/{Table/52-Max}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer E180815 20:40:36.733096 158686 storage/queue.go:788 [split,n1,s1,r23/1:/{Table/52-Max}] unable to split [n1,s1,r23/1:/{Table/52-Max}] at key "/Table/53": split at key /Table/53 failed: node unavailable; try another peer I180815 20:40:36.733743 158942 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] transport racer I180815 20:40:36.789547 158280 kv/transport_race.go:67 transport race promotion: ran 39 iterations on up to 877 requests ------- Stdout: ------- W180815 20:20:04.735629 158449 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180815 20:20:04.756225 158449 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180815 20:20:04.756547 158449 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:20:04.756590 158449 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:20:04.770676 158449 server/config.go:496 [n?] 1 storage engine initialized I180815 20:20:04.770775 158449 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180815 20:20:04.770795 158449 server/config.go:499 [n?] store 0: in-memory, size 0 B I180815 20:20:04.774753 158449 server/node.go:373 [n?] **** cluster 897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a has been created I180815 20:20:04.774783 158449 server/server.go:1398 [n?] **** add additional nodes by specifying --join=127.0.0.1:35641 I180815 20:20:04.774934 158449 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:35641" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534364404774867605 I180815 20:20:04.788664 158449 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180815 20:20:04.788825 158449 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00} I180815 20:20:04.788933 158449 storage/stores.go:222 [n1] read 0 node addresses from persistent storage I180815 20:20:04.789122 158449 server/node.go:697 [n1] connecting to gossip network to verify cluster ID... I180815 20:20:04.792799 158449 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a" I180815 20:20:04.792912 158449 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180815 20:20:04.794388 158449 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:20:04.794466 158449 server/server.go:1833 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined I180815 20:20:04.794618 158449 server/server.go:1535 [n1] starting https server at 127.0.0.1:35159 (use: 127.0.0.1:35159) I180815 20:20:04.800514 158449 server/server.go:1537 [n1] starting grpc/postgres server at 127.0.0.1:35641 I180815 20:20:04.800600 158449 server/server.go:1538 [n1] advertising CockroachDB node at 127.0.0.1:35641 I180815 20:20:04.795116 158433 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180815 20:20:04.795346 158681 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory E180815 20:20:04.816788 158723 storage/queue.go:788 [replicate,n1,s1,r1/1:/{Min-System/}] range requires a replication change, but lacks a quorum of live replicas (0/1) I180815 20:20:04.817580 158712 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] I180815 20:20:04.995544 158709 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root} I180815 20:20:04.997521 158738 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] I180815 20:20:05.014104 158744 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180815 20:20:05.031468 158456 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root} I180815 20:20:05.039572 158756 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180815 20:20:05.057269 158729 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root} I180815 20:20:05.087390 158775 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root} I180815 20:20:05.090490 158462 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180815 20:20:05.100042 158794 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root} I180815 20:20:05.103041 158804 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] I180815 20:20:05.108352 158799 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root} I180815 20:20:05.110062 158381 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180815 20:20:05.114153 158449 server/server.go:1620 [n1] done ensuring all necessary migrations have run I180815 20:20:05.114189 158449 server/server.go:1623 [n1] serving sql connections I180815 20:20:05.117362 158850 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:35641} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534364404774867605 LocalityAddress:[]} ClusterID:897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a StartedAt:1534364404774867605 LastUp:1534364404774867605} I180815 20:20:05.117710 158736 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version I180815 20:20:05.120942 158749 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180815 20:20:05.126759 158782 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180815 20:20:05.133741 158869 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180815 20:20:05.148799 158890 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180815 20:20:05.155296 158901 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180815 20:20:05.177525 158811 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180815 20:20:05.184671 158931 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180815 20:20:05.193079 158916 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180815 20:20:05.213605 158926 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180815 20:20:05.220306 158986 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180815 20:20:05.226131 159001 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] W180815 20:20:05.232969 159014 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=530964b3 key=/Local/Range/Table/22/RangeDescriptor rw=true pri=0.02693795 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534364405.227916665,1 orig=1534364405.226151335,0 max=1534364405.226151335,0 wto=false rop=false seq=1 I180815 20:20:05.235499 158952 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180815 20:20:05.245022 159005 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] W180815 20:20:05.303250 158449 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180815 20:20:05.310656 158449 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180815 20:20:05.315402 158449 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:20:05.315480 158449 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:20:05.330854 158449 server/config.go:496 [n?] 1 storage engine initialized I180815 20:20:05.330964 158449 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180815 20:20:05.330983 158449 server/config.go:499 [n?] store 0: in-memory, size 0 B W180815 20:20:05.331113 158449 gossip/gossip.go:1351 [n?] no incoming or outgoing connections I180815 20:20:05.331177 158449 server/server.go:1400 [n?] no stores bootstrapped and --join flag specified, awaiting init command. I180815 20:20:05.344590 159092 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:35641 I180815 20:20:05.345720 159045 gossip/server.go:219 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:45155} I180815 20:20:05.360942 158449 server/node.go:697 [n?] connecting to gossip network to verify cluster ID... I180815 20:20:05.360981 158449 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a" I180815 20:20:05.361374 159100 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:20:05.365029 159099 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:20:05.366271 158449 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:20:05.367766 158449 server/node.go:428 [n?] new node allocated ID 2 I180815 20:20:05.367934 158449 gossip/gossip.go:383 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:45155" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534364405367814511 I180815 20:20:05.368073 158449 storage/stores.go:222 [n2] read 0 node addresses from persistent storage I180815 20:20:05.368143 158449 storage/stores.go:241 [n2] wrote 1 node addresses to persistent storage I180815 20:20:05.369673 158824 storage/stores.go:241 [n1] wrote 1 node addresses to persistent storage I180815 20:20:05.371353 158449 server/node.go:672 [n2] bootstrapped store [n2,s2] I180815 20:20:05.374076 158449 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes [] I180815 20:20:05.374378 158449 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:20:05.374427 158449 server/server.go:1833 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined I180815 20:20:05.374526 158449 server/server.go:1535 [n2] starting https server at 127.0.0.1:43861 (use: 127.0.0.1:43861) I180815 20:20:05.374557 158449 server/server.go:1537 [n2] starting grpc/postgres server at 127.0.0.1:45155 I180815 20:20:05.374579 158449 server/server.go:1538 [n2] advertising CockroachDB node at 127.0.0.1:45155 I180815 20:20:05.377097 159053 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:20:05.379673 158449 server/server.go:1620 [n2] done ensuring all necessary migrations have run I180815 20:20:05.397824 158449 server/server.go:1623 [n2] serving sql connections I180815 20:20:05.424198 159220 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established I180815 20:20:05.447142 159221 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version I180815 20:20:05.458760 159223 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:45155} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534364405367814511 LocalityAddress:[]} ClusterID:897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a StartedAt:1534364405367814511 LastUp:1534364405367814511} I180815 20:20:05.461728 159262 sql/event_log.go:126 [n1,client=127.0.0.1:39484,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root} I180815 20:20:05.466851 159264 storage/replica_command.go:282 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23] I180815 20:20:05.468550 159262 sql/event_log.go:126 [n1,client=127.0.0.1:39484,user=root] Event: "create_table", target: 53, info: {TableName:test.public.t Statement:CREATE TABLE test.public.t (num INT) User:root} I180815 20:20:05.495103 159292 storage/replica_command.go:282 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24] W180815 20:20:05.516400 159262 sql/crdb_internal.go:900 [n1,client=127.0.0.1:39484,user=root] failed to dial into node 2 (LIVE): initial connection heartbeat failed: rpc error: code = Unavailable desc = all SubConns are in TransientFailure, latest connection error: connection error: desc = "transport: Error while dialing dial tcp 127.0.0.1:45155: connect: connection refused" ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed tests on master: testrace/TestShowSessions, lint/TestLint - The following tests appear to have failed: [#839592](https://teamcity.cockroachdb.com/viewLog.html?buildId=839592): ``` --- FAIL: testrace/TestShowSessions (2.450s) show_test.go:684: unexpected number of running sessions: 2, expected 1. Active sessions (results might have changed since the test checked): app: "$ internal-log-range-event", query: "INSERT INTO system.public.rangelog(\"timestamp\", \"rangeID\", \"storeID\", \"eventType\", \"otherRangeID\", info) VALUES ($1, $2, $3, $4, $5, $6)", last query: app: "", query: "SELECT active_queries, last_active_query, application_name FROM [SHOW CLUSTER SESSIONS]", last query: SELECT node_id, (now() - session_start)::FLOAT FROM [SHOW CLUSTER SESSIONS] WHERE application_name NOT LIKE 'internal-%' ------- Stdout: ------- W180815 20:40:34.423397 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180815 20:40:34.450676 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180815 20:40:34.451091 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:34.451159 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:34.455752 158099 server/config.go:496 [n?] 1 storage engine initialized I180815 20:40:34.456002 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180815 20:40:34.456043 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B I180815 20:40:34.514781 158099 server/node.go:373 [n?] **** cluster 6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 has been created I180815 20:40:34.515045 158099 server/server.go:1398 [n?] **** add additional nodes by specifying --join=127.0.0.1:41091 I180815 20:40:34.516516 158099 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41091" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365634515938360 I180815 20:40:34.534941 158099 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180815 20:40:34.537635 158099 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00} I180815 20:40:34.538394 158099 storage/stores.go:222 [n1] read 0 node addresses from persistent storage I180815 20:40:34.539206 158099 server/node.go:697 [n1] connecting to gossip network to verify cluster ID... I180815 20:40:34.539582 158099 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102" I180815 20:40:34.539943 158099 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180815 20:40:34.541685 158099 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:34.541909 158099 server/server.go:1833 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined I180815 20:40:34.543736 158325 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:34.545405 158099 server/server.go:1535 [n1] starting https server at 127.0.0.1:34275 (use: 127.0.0.1:34275) I180815 20:40:34.545680 158099 server/server.go:1537 [n1] starting grpc/postgres server at 127.0.0.1:41091 I180815 20:40:34.545753 158099 server/server.go:1538 [n1] advertising CockroachDB node at 127.0.0.1:41091 I180815 20:40:34.566927 158335 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180815 20:40:34.719270 158304 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] I180815 20:40:34.817666 158087 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] I180815 20:40:34.840768 158068 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root} I180815 20:40:34.933807 158160 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180815 20:40:35.038918 158386 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180815 20:40:35.116602 158341 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root} W180815 20:40:35.128647 158374 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=5ddfb628 key=/Local/Range/System/tsd/RangeDescriptor rw=true pri=0.02614628 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534365635.039580425,0 orig=1534365635.039580425,0 max=1534365635.039580425,0 wto=false rop=false seq=1 I180815 20:40:35.147031 158074 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180815 20:40:35.241602 158379 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] I180815 20:40:35.285924 158289 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root} I180815 20:40:35.328157 158390 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180815 20:40:35.401113 158408 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180815 20:40:35.470572 158398 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180815 20:40:35.530967 158452 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180815 20:40:35.548516 158385 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root} I180815 20:40:35.603367 158447 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180815 20:40:35.633708 158442 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established I180815 20:40:35.654186 158514 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root} I180815 20:40:35.677178 158463 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180815 20:40:35.720541 158458 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root} I180815 20:40:35.744993 158521 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180815 20:40:35.765024 158099 server/server.go:1620 [n1] done ensuring all necessary migrations have run I180815 20:40:35.765364 158099 server/server.go:1623 [n1] serving sql connections I180815 20:40:35.811858 158481 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41091} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365634515938360 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365634515938360 LastUp:1534365634515938360} I180815 20:40:35.813481 158479 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version I180815 20:40:35.820374 158564 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180815 20:40:35.889664 158497 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180815 20:40:35.951092 158556 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180815 20:40:36.009944 158598 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180815 20:40:36.059947 158619 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] I180815 20:40:36.119763 158606 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180815 20:40:36.175659 158630 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] W180815 20:40:36.287477 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180815 20:40:36.315831 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180815 20:40:36.316449 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:36.316518 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:36.321488 158099 server/config.go:496 [n?] 1 storage engine initialized I180815 20:40:36.321673 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180815 20:40:36.321701 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B W180815 20:40:36.321999 158099 gossip/gossip.go:1351 [n?] no incoming or outgoing connections I180815 20:40:36.322565 158099 server/server.go:1400 [n?] no stores bootstrapped and --join flag specified, awaiting init command. I180815 20:40:36.398063 158665 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:41091 I180815 20:40:36.399347 158763 gossip/server.go:219 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:32789} I180815 20:40:36.404521 158099 server/node.go:697 [n?] connecting to gossip network to verify cluster ID... I180815 20:40:36.405108 158099 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102" I180815 20:40:36.408670 158717 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:40:36.415122 158716 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:40:36.422331 158099 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:40:36.431430 158099 server/node.go:428 [n?] new node allocated ID 2 I180815 20:40:36.432094 158099 gossip/gossip.go:383 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:32789" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365636431659703 I180815 20:40:36.433009 158099 storage/stores.go:222 [n2] read 0 node addresses from persistent storage I180815 20:40:36.433961 158099 storage/stores.go:241 [n2] wrote 1 node addresses to persistent storage I180815 20:40:36.436697 158720 storage/stores.go:241 [n1] wrote 1 node addresses to persistent storage I180815 20:40:36.457872 158099 server/node.go:672 [n2] bootstrapped store [n2,s2] I180815 20:40:36.460045 158099 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes [] I180815 20:40:36.461355 158099 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:36.461544 158099 server/server.go:1833 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined I180815 20:40:36.465410 158099 server/server.go:1535 [n2] starting https server at 127.0.0.1:36335 (use: 127.0.0.1:36335) I180815 20:40:36.465633 158099 server/server.go:1537 [n2] starting grpc/postgres server at 127.0.0.1:32789 I180815 20:40:36.465700 158099 server/server.go:1538 [n2] advertising CockroachDB node at 127.0.0.1:32789 I180815 20:40:36.468619 158867 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:36.476134 158099 server/server.go:1620 [n2] done ensuring all necessary migrations have run I180815 20:40:36.476492 158099 server/server.go:1623 [n2] serving sql connections I180815 20:40:36.576176 158869 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version I180815 20:40:36.577336 158871 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:32789} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365636431659703 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365636431659703 LastUp:1534365636431659703} I180815 20:40:36.617219 158877 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established I180815 20:40:36.623403 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root} I180815 20:40:36.624884 158702 storage/replica_command.go:282 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23] I180815 20:40:36.685520 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_table", target: 53, info: {TableName:test.public.t Statement:CREATE TABLE test.public.t (num INT) User:root} I180815 20:40:36.689918 158686 storage/replica_command.go:282 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24] I180815 20:40:36.729831 158942 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] transport racer 1 [async] storage.split: processing replica 1 [async] kv.TxnCoordSender: heartbeat loop W180815 20:40:36.732234 158686 internal/client/txn.go:514 [split,n1,s1,r23/1:/{Table/52-Max}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer E180815 20:40:36.733096 158686 storage/queue.go:788 [split,n1,s1,r23/1:/{Table/52-Max}] unable to split [n1,s1,r23/1:/{Table/52-Max}] at key "/Table/53": split at key /Table/53 failed: node unavailable; try another peer I180815 20:40:36.733743 158942 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] transport racer I180815 20:40:36.789547 158280 kv/transport_race.go:67 transport race promotion: ran 39 iterations on up to 877 requests --- FAIL: lint/TestLint (146.790s) --- FAIL: lint/TestLint: TestLint/TestHelpURLs (41.520s) lint_test.go:945: https://www.cockroachlabs.com/docs/v2.1/set-vars.html : Head https://www.cockroachlabs.com/docs/v2.1/set-vars.html: EOF 324: https://www.cockroachlabs.com/docs/v2.1/set-vars.html 990: https://www.cockroachlabs.com/docs/v2.1/set-vars.html 1245:https://www.cockroachlabs.com/docs/v2.1/set-vars.html https://www.cockroachlabs.com/docs/v2.1/create-table-as.html : Head https://www.cockroachlabs.com/docs/v2.1/create-table-as.html: EOF 241:https://www.cockroachlabs.com/docs/v2.1/create-table-as.html 2 errors ------- Stdout: ------- === PAUSE TestLint/TestHelpURLs --- FAIL: lint/TestLint (146.790s) --- FAIL: lint/TestLint: TestLint/TestHelpURLs (41.520s) lint_test.go:945: https://www.cockroachlabs.com/docs/v2.1/set-vars.html : Head https://www.cockroachlabs.com/docs/v2.1/set-vars.html: EOF 324: https://www.cockroachlabs.com/docs/v2.1/set-vars.html 990: https://www.cockroachlabs.com/docs/v2.1/set-vars.html 1245:https://www.cockroachlabs.com/docs/v2.1/set-vars.html https://www.cockroachlabs.com/docs/v2.1/create-table-as.html : Head https://www.cockroachlabs.com/docs/v2.1/create-table-as.html: EOF 241:https://www.cockroachlabs.com/docs/v2.1/create-table-as.html 2 errors ------- Stdout: ------- === PAUSE TestLint/TestHelpURLs --- FAIL: testrace/TestShowSessions (3.280s) show_test.go:684: unexpected number of running sessions: 2, expected 1. Active sessions (results might have changed since the test checked): app: "$ internal-log-range-event", query: "INSERT INTO system.public.rangelog(\"timestamp\", \"rangeID\", \"storeID\", \"eventType\", \"otherRangeID\", info) VALUES ($1, $2, $3, $4, $5, $6)", last query: app: "", query: "SELECT active_queries, last_active_query, application_name FROM [SHOW CLUSTER SESSIONS]", last query: SELECT node_id, (now() - session_start)::FLOAT FROM [SHOW CLUSTER SESSIONS] WHERE application_name NOT LIKE 'internal-%' ------- Stdout: ------- W180815 20:40:34.423397 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180815 20:40:34.450676 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180815 20:40:34.451091 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:34.451159 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:34.455752 158099 server/config.go:496 [n?] 1 storage engine initialized I180815 20:40:34.456002 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180815 20:40:34.456043 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B I180815 20:40:34.514781 158099 server/node.go:373 [n?] **** cluster 6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 has been created I180815 20:40:34.515045 158099 server/server.go:1398 [n?] **** add additional nodes by specifying --join=127.0.0.1:41091 I180815 20:40:34.516516 158099 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41091" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365634515938360 I180815 20:40:34.534941 158099 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180815 20:40:34.537635 158099 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00} I180815 20:40:34.538394 158099 storage/stores.go:222 [n1] read 0 node addresses from persistent storage I180815 20:40:34.539206 158099 server/node.go:697 [n1] connecting to gossip network to verify cluster ID... I180815 20:40:34.539582 158099 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102" I180815 20:40:34.539943 158099 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180815 20:40:34.541685 158099 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:34.541909 158099 server/server.go:1833 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined I180815 20:40:34.543736 158325 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:34.545405 158099 server/server.go:1535 [n1] starting https server at 127.0.0.1:34275 (use: 127.0.0.1:34275) I180815 20:40:34.545680 158099 server/server.go:1537 [n1] starting grpc/postgres server at 127.0.0.1:41091 I180815 20:40:34.545753 158099 server/server.go:1538 [n1] advertising CockroachDB node at 127.0.0.1:41091 I180815 20:40:34.566927 158335 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180815 20:40:34.719270 158304 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] I180815 20:40:34.817666 158087 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] I180815 20:40:34.840768 158068 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root} I180815 20:40:34.933807 158160 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180815 20:40:35.038918 158386 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180815 20:40:35.116602 158341 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root} W180815 20:40:35.128647 158374 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=5ddfb628 key=/Local/Range/System/tsd/RangeDescriptor rw=true pri=0.02614628 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534365635.039580425,0 orig=1534365635.039580425,0 max=1534365635.039580425,0 wto=false rop=false seq=1 I180815 20:40:35.147031 158074 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180815 20:40:35.241602 158379 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] I180815 20:40:35.285924 158289 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root} I180815 20:40:35.328157 158390 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180815 20:40:35.401113 158408 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180815 20:40:35.470572 158398 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180815 20:40:35.530967 158452 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180815 20:40:35.548516 158385 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root} I180815 20:40:35.603367 158447 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180815 20:40:35.633708 158442 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established I180815 20:40:35.654186 158514 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root} I180815 20:40:35.677178 158463 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180815 20:40:35.720541 158458 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root} I180815 20:40:35.744993 158521 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180815 20:40:35.765024 158099 server/server.go:1620 [n1] done ensuring all necessary migrations have run I180815 20:40:35.765364 158099 server/server.go:1623 [n1] serving sql connections I180815 20:40:35.811858 158481 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41091} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365634515938360 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365634515938360 LastUp:1534365634515938360} I180815 20:40:35.813481 158479 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version I180815 20:40:35.820374 158564 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180815 20:40:35.889664 158497 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180815 20:40:35.951092 158556 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180815 20:40:36.009944 158598 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180815 20:40:36.059947 158619 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] I180815 20:40:36.119763 158606 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180815 20:40:36.175659 158630 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] W180815 20:40:36.287477 158099 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180815 20:40:36.315831 158099 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180815 20:40:36.316449 158099 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:36.316518 158099 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:40:36.321488 158099 server/config.go:496 [n?] 1 storage engine initialized I180815 20:40:36.321673 158099 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180815 20:40:36.321701 158099 server/config.go:499 [n?] store 0: in-memory, size 0 B W180815 20:40:36.321999 158099 gossip/gossip.go:1351 [n?] no incoming or outgoing connections I180815 20:40:36.322565 158099 server/server.go:1400 [n?] no stores bootstrapped and --join flag specified, awaiting init command. I180815 20:40:36.398063 158665 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:41091 I180815 20:40:36.399347 158763 gossip/server.go:219 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:32789} I180815 20:40:36.404521 158099 server/node.go:697 [n?] connecting to gossip network to verify cluster ID... I180815 20:40:36.405108 158099 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "6c15fc2c-e9dc-4747-9d9c-f7f97a2da102" I180815 20:40:36.408670 158717 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:40:36.415122 158716 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:40:36.422331 158099 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:40:36.431430 158099 server/node.go:428 [n?] new node allocated ID 2 I180815 20:40:36.432094 158099 gossip/gossip.go:383 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:32789" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534365636431659703 I180815 20:40:36.433009 158099 storage/stores.go:222 [n2] read 0 node addresses from persistent storage I180815 20:40:36.433961 158099 storage/stores.go:241 [n2] wrote 1 node addresses to persistent storage I180815 20:40:36.436697 158720 storage/stores.go:241 [n1] wrote 1 node addresses to persistent storage I180815 20:40:36.457872 158099 server/node.go:672 [n2] bootstrapped store [n2,s2] I180815 20:40:36.460045 158099 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes [] I180815 20:40:36.461355 158099 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:36.461544 158099 server/server.go:1833 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined I180815 20:40:36.465410 158099 server/server.go:1535 [n2] starting https server at 127.0.0.1:36335 (use: 127.0.0.1:36335) I180815 20:40:36.465633 158099 server/server.go:1537 [n2] starting grpc/postgres server at 127.0.0.1:32789 I180815 20:40:36.465700 158099 server/server.go:1538 [n2] advertising CockroachDB node at 127.0.0.1:32789 I180815 20:40:36.468619 158867 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:40:36.476134 158099 server/server.go:1620 [n2] done ensuring all necessary migrations have run I180815 20:40:36.476492 158099 server/server.go:1623 [n2] serving sql connections I180815 20:40:36.576176 158869 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version I180815 20:40:36.577336 158871 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:32789} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534365636431659703 LocalityAddress:[]} ClusterID:6c15fc2c-e9dc-4747-9d9c-f7f97a2da102 StartedAt:1534365636431659703 LastUp:1534365636431659703} I180815 20:40:36.617219 158877 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established I180815 20:40:36.623403 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root} I180815 20:40:36.624884 158702 storage/replica_command.go:282 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23] I180815 20:40:36.685520 158934 sql/event_log.go:126 [n1,client=127.0.0.1:47914,user=root] Event: "create_table", target: 53, info: {TableName:test.public.t Statement:CREATE TABLE test.public.t (num INT) User:root} I180815 20:40:36.689918 158686 storage/replica_command.go:282 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24] I180815 20:40:36.729831 158942 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] transport racer 1 [async] storage.split: processing replica 1 [async] kv.TxnCoordSender: heartbeat loop W180815 20:40:36.732234 158686 internal/client/txn.go:514 [split,n1,s1,r23/1:/{Table/52-Max}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer E180815 20:40:36.733096 158686 storage/queue.go:788 [split,n1,s1,r23/1:/{Table/52-Max}] unable to split [n1,s1,r23/1:/{Table/52-Max}] at key "/Table/53": split at key /Table/53 failed: node unavailable; try another peer I180815 20:40:36.733743 158942 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] transport racer I180815 20:40:36.789547 158280 kv/transport_race.go:67 transport race promotion: ran 39 iterations on up to 877 requests ------- Stdout: ------- W180815 20:20:04.735629 158449 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180815 20:20:04.756225 158449 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180815 20:20:04.756547 158449 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:20:04.756590 158449 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:20:04.770676 158449 server/config.go:496 [n?] 1 storage engine initialized I180815 20:20:04.770775 158449 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180815 20:20:04.770795 158449 server/config.go:499 [n?] store 0: in-memory, size 0 B I180815 20:20:04.774753 158449 server/node.go:373 [n?] **** cluster 897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a has been created I180815 20:20:04.774783 158449 server/server.go:1398 [n?] **** add additional nodes by specifying --join=127.0.0.1:35641 I180815 20:20:04.774934 158449 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:35641" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534364404774867605 I180815 20:20:04.788664 158449 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180815 20:20:04.788825 158449 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00} I180815 20:20:04.788933 158449 storage/stores.go:222 [n1] read 0 node addresses from persistent storage I180815 20:20:04.789122 158449 server/node.go:697 [n1] connecting to gossip network to verify cluster ID... I180815 20:20:04.792799 158449 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a" I180815 20:20:04.792912 158449 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180815 20:20:04.794388 158449 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:20:04.794466 158449 server/server.go:1833 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined I180815 20:20:04.794618 158449 server/server.go:1535 [n1] starting https server at 127.0.0.1:35159 (use: 127.0.0.1:35159) I180815 20:20:04.800514 158449 server/server.go:1537 [n1] starting grpc/postgres server at 127.0.0.1:35641 I180815 20:20:04.800600 158449 server/server.go:1538 [n1] advertising CockroachDB node at 127.0.0.1:35641 I180815 20:20:04.795116 158433 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180815 20:20:04.795346 158681 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory E180815 20:20:04.816788 158723 storage/queue.go:788 [replicate,n1,s1,r1/1:/{Min-System/}] range requires a replication change, but lacks a quorum of live replicas (0/1) I180815 20:20:04.817580 158712 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] I180815 20:20:04.995544 158709 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root} I180815 20:20:04.997521 158738 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] I180815 20:20:05.014104 158744 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180815 20:20:05.031468 158456 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root} I180815 20:20:05.039572 158756 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180815 20:20:05.057269 158729 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root} I180815 20:20:05.087390 158775 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root} I180815 20:20:05.090490 158462 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180815 20:20:05.100042 158794 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root} I180815 20:20:05.103041 158804 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] I180815 20:20:05.108352 158799 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root} I180815 20:20:05.110062 158381 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180815 20:20:05.114153 158449 server/server.go:1620 [n1] done ensuring all necessary migrations have run I180815 20:20:05.114189 158449 server/server.go:1623 [n1] serving sql connections I180815 20:20:05.117362 158850 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:35641} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534364404774867605 LocalityAddress:[]} ClusterID:897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a StartedAt:1534364404774867605 LastUp:1534364404774867605} I180815 20:20:05.117710 158736 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version I180815 20:20:05.120942 158749 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180815 20:20:05.126759 158782 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180815 20:20:05.133741 158869 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180815 20:20:05.148799 158890 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180815 20:20:05.155296 158901 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180815 20:20:05.177525 158811 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180815 20:20:05.184671 158931 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180815 20:20:05.193079 158916 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180815 20:20:05.213605 158926 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180815 20:20:05.220306 158986 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180815 20:20:05.226131 159001 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] W180815 20:20:05.232969 159014 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=530964b3 key=/Local/Range/Table/22/RangeDescriptor rw=true pri=0.02693795 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534364405.227916665,1 orig=1534364405.226151335,0 max=1534364405.226151335,0 wto=false rop=false seq=1 I180815 20:20:05.235499 158952 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180815 20:20:05.245022 159005 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] W180815 20:20:05.303250 158449 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180815 20:20:05.310656 158449 server/server.go:827 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180815 20:20:05.315402 158449 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:20:05.315480 158449 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180815 20:20:05.330854 158449 server/config.go:496 [n?] 1 storage engine initialized I180815 20:20:05.330964 158449 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180815 20:20:05.330983 158449 server/config.go:499 [n?] store 0: in-memory, size 0 B W180815 20:20:05.331113 158449 gossip/gossip.go:1351 [n?] no incoming or outgoing connections I180815 20:20:05.331177 158449 server/server.go:1400 [n?] no stores bootstrapped and --join flag specified, awaiting init command. I180815 20:20:05.344590 159092 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:35641 I180815 20:20:05.345720 159045 gossip/server.go:219 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:45155} I180815 20:20:05.360942 158449 server/node.go:697 [n?] connecting to gossip network to verify cluster ID... I180815 20:20:05.360981 158449 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a" I180815 20:20:05.361374 159100 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:20:05.365029 159099 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:20:05.366271 158449 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180815 20:20:05.367766 158449 server/node.go:428 [n?] new node allocated ID 2 I180815 20:20:05.367934 158449 gossip/gossip.go:383 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:45155" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v2.1.0-alpha.20180702-1619-g2f15c44" started_at:1534364405367814511 I180815 20:20:05.368073 158449 storage/stores.go:222 [n2] read 0 node addresses from persistent storage I180815 20:20:05.368143 158449 storage/stores.go:241 [n2] wrote 1 node addresses to persistent storage I180815 20:20:05.369673 158824 storage/stores.go:241 [n1] wrote 1 node addresses to persistent storage I180815 20:20:05.371353 158449 server/node.go:672 [n2] bootstrapped store [n2,s2] I180815 20:20:05.374076 158449 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes [] I180815 20:20:05.374378 158449 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:20:05.374427 158449 server/server.go:1833 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined I180815 20:20:05.374526 158449 server/server.go:1535 [n2] starting https server at 127.0.0.1:43861 (use: 127.0.0.1:43861) I180815 20:20:05.374557 158449 server/server.go:1537 [n2] starting grpc/postgres server at 127.0.0.1:45155 I180815 20:20:05.374579 158449 server/server.go:1538 [n2] advertising CockroachDB node at 127.0.0.1:45155 I180815 20:20:05.377097 159053 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180815 20:20:05.379673 158449 server/server.go:1620 [n2] done ensuring all necessary migrations have run I180815 20:20:05.397824 158449 server/server.go:1623 [n2] serving sql connections I180815 20:20:05.424198 159220 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established I180815 20:20:05.447142 159221 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version I180815 20:20:05.458760 159223 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:45155} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v2.1.0-alpha.20180702-1619-g2f15c44 StartedAt:1534364405367814511 LocalityAddress:[]} ClusterID:897f3ffc-68b2-4d3e-b45e-2b1f6b60ae9a StartedAt:1534364405367814511 LastUp:1534364405367814511} I180815 20:20:05.461728 159262 sql/event_log.go:126 [n1,client=127.0.0.1:39484,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root} I180815 20:20:05.466851 159264 storage/replica_command.go:282 [split,n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23] I180815 20:20:05.468550 159262 sql/event_log.go:126 [n1,client=127.0.0.1:39484,user=root] Event: "create_table", target: 53, info: {TableName:test.public.t Statement:CREATE TABLE test.public.t (num INT) User:root} I180815 20:20:05.495103 159292 storage/replica_command.go:282 [split,n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24] W180815 20:20:05.516400 159262 sql/crdb_internal.go:900 [n1,client=127.0.0.1:39484,user=root] failed to dial into node 2 (LIVE): initial connection heartbeat failed: rpc error: code = Unavailable desc = all SubConns are in TransientFailure, latest connection error: connection error: desc = "transport: Error while dialing dial tcp 127.0.0.1:45155: connect: connection refused" ``` Please assign, take a look and update the issue accordingly.
test
teamcity failed tests on master testrace testshowsessions lint testlint the following tests appear to have failed fail testrace testshowsessions show test go unexpected number of running sessions expected active sessions results might have changed since the test checked app internal log range event query insert into system public rangelog timestamp rangeid storeid eventtype otherrangeid info values last query app query select active queries last active query application name from last query select node id now session start float from where application name not like internal stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases queries writes bytesperreplica pmax writesperreplica pmax storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at storage replica command go initiating a split of this range at key system storage replica command go initiating a split of this range at key system nodeliveness storage replica command go initiating a split of this range at key system nodelivenessmax sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user root storage replica command go initiating a split of this range at key system tsd storage replica command go initiating a split of this range at key system tse sql event log go event set cluster setting target info settingname version value user root storage intent resolver go failed to push during intent resolution failed to push split id key local range system tsd rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table systemconfigspan start storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname trace debug enable value false user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname cluster secret value gen random uuid string user root storage replica command go initiating a split of this range at key table rpc nodedialer nodedialer go connection to established sql event log go event create database target info databasename defaultdb statement create database if not exists defaultdb user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename postgres statement create database if not exists postgres user root storage replica command go initiating a split of this range at key table server server go done ensuring all necessary migrations have run server server go serving sql connections sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup server server update go no need to upgrade cluster already at the newest version storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b gossip gossip go no incoming or outgoing connections server server go no stores bootstrapped and join flag specified awaiting init command gossip client go started gossip client to gossip server go received initial cluster verification connection from tcp server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping server node go new node allocated id gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage stores go read node addresses from persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage server node go bootstrapped store server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go done ensuring all necessary migrations have run server server go serving sql connections server server update go no need to upgrade cluster already at the newest version sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup rpc nodedialer nodedialer go connection to established sql event log go event create database target info databasename test statement create database if not exists test user root storage replica command go initiating a split of this range at key table sql event log go event create table target info tablename test public t statement create table test public t num int user root storage replica command go initiating a split of this range at key table util stop stopper go quiescing tasks left transport racer storage split processing replica kv txncoordsender heartbeat loop internal client txn go failure aborting transaction node unavailable try another peer abort caused by node unavailable try another peer storage queue go unable to split at key table split at key table failed node unavailable try another peer util stop stopper go quiescing tasks left transport racer kv transport race go transport race promotion ran iterations on up to requests fail lint testlint fail lint testlint testlint testhelpurls lint test go head eof head eof errors stdout pause testlint testhelpurls fail lint testlint fail lint testlint testlint testhelpurls lint test go head eof head eof errors stdout pause testlint testhelpurls fail testrace testshowsessions show test go unexpected number of running sessions expected active sessions results might have changed since the test checked app internal log range event query insert into system public rangelog timestamp rangeid storeid eventtype otherrangeid info values last query app query select active queries last active query application name from last query select node id now session start float from where application name not like internal stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases queries writes bytesperreplica pmax writesperreplica pmax storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at storage replica command go initiating a split of this range at key system storage replica command go initiating a split of this range at key system nodeliveness storage replica command go initiating a split of this range at key system nodelivenessmax sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user root storage replica command go initiating a split of this range at key system tsd storage replica command go initiating a split of this range at key system tse sql event log go event set cluster setting target info settingname version value user root storage intent resolver go failed to push during intent resolution failed to push split id key local range system tsd rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table systemconfigspan start storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname trace debug enable value false user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname cluster secret value gen random uuid string user root storage replica command go initiating a split of this range at key table rpc nodedialer nodedialer go connection to established sql event log go event create database target info databasename defaultdb statement create database if not exists defaultdb user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename postgres statement create database if not exists postgres user root storage replica command go initiating a split of this range at key table server server go done ensuring all necessary migrations have run server server go serving sql connections sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup server server update go no need to upgrade cluster already at the newest version storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b gossip gossip go no incoming or outgoing connections server server go no stores bootstrapped and join flag specified awaiting init command gossip client go started gossip client to gossip server go received initial cluster verification connection from tcp server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping server node go new node allocated id gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage stores go read node addresses from persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage server node go bootstrapped store server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go done ensuring all necessary migrations have run server server go serving sql connections server server update go no need to upgrade cluster already at the newest version sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup rpc nodedialer nodedialer go connection to established sql event log go event create database target info databasename test statement create database if not exists test user root storage replica command go initiating a split of this range at key table sql event log go event create table target info tablename test public t statement create table test public t num int user root storage replica command go initiating a split of this range at key table util stop stopper go quiescing tasks left transport racer storage split processing replica kv txncoordsender heartbeat loop internal client txn go failure aborting transaction node unavailable try another peer abort caused by node unavailable try another peer storage queue go unable to split at key table split at key table failed node unavailable try another peer util stop stopper go quiescing tasks left transport racer kv transport race go transport race promotion ran iterations on up to requests stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases queries writes bytesperreplica pmax writesperreplica pmax storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at storage replica command go initiating a split of this range at key system server status recorder go available memory from cgroups eib exceeds system memory gib using system memory storage queue go range requires a replication change but lacks a quorum of live replicas storage replica command go initiating a split of this range at key system nodeliveness sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user root storage replica command go initiating a split of this range at key system nodelivenessmax storage replica command go initiating a split of this range at key system tsd sql event log go event set cluster setting target info settingname version value user root storage replica command go initiating a split of this range at key system tse sql event log go event set cluster setting target info settingname trace debug enable value false user root sql event log go event set cluster setting target info settingname cluster secret value gen random uuid string user root storage replica command go initiating a split of this range at key table systemconfigspan start sql event log go event create database target info databasename defaultdb statement create database if not exists defaultdb user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename postgres statement create database if not exists postgres user root storage replica command go initiating a split of this range at key table server server go done ensuring all necessary migrations have run server server go serving sql connections sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup server server update go no need to upgrade cluster already at the newest version storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b gossip gossip go no incoming or outgoing connections server server go no stores bootstrapped and join flag specified awaiting init command gossip client go started gossip client to gossip server go received initial cluster verification connection from tcp server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping server node go new node allocated id gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage stores go read node addresses from persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage server node go bootstrapped store server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go done ensuring all necessary migrations have run server server go serving sql connections rpc nodedialer nodedialer go connection to established server server update go no need to upgrade cluster already at the newest version sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup sql event log go event create database target info databasename test statement create database if not exists test user root storage replica command go initiating a split of this range at key table sql event log go event create table target info tablename test public t statement create table test public t num int user root storage replica command go initiating a split of this range at key table sql crdb internal go failed to dial into node live initial connection heartbeat failed rpc error code unavailable desc all subconns are in transientfailure latest connection error connection error desc transport error while dialing dial tcp connect connection refused please assign take a look and update the issue accordingly
1
304,090
26,251,701,804
IssuesEvent
2023-01-05 19:57:12
MetaMask/metamask-extension
https://api.github.com/repos/MetaMask/metamask-extension
closed
Write tests for Infura network client
area-testSuite type-refactor
As part of an effort to [merge the NetworkController that exists in this repo with the NetworkController that exists in the `controllers` repo](https://github.com/MetaMask/controllers/issues/753), we need to write tests for the JSON-RPC layer, and all of its custom behavior, that the NetworkController exposes. We have two "network clients" that handle JSON-RPC network requests, one for Infura and one for custom networks. The custom network client tests are covered under issue #16041. This issue is for testing the network client that is specific to Infura.
1.0
Write tests for Infura network client - As part of an effort to [merge the NetworkController that exists in this repo with the NetworkController that exists in the `controllers` repo](https://github.com/MetaMask/controllers/issues/753), we need to write tests for the JSON-RPC layer, and all of its custom behavior, that the NetworkController exposes. We have two "network clients" that handle JSON-RPC network requests, one for Infura and one for custom networks. The custom network client tests are covered under issue #16041. This issue is for testing the network client that is specific to Infura.
test
write tests for infura network client as part of an effort to we need to write tests for the json rpc layer and all of its custom behavior that the networkcontroller exposes we have two network clients that handle json rpc network requests one for infura and one for custom networks the custom network client tests are covered under issue this issue is for testing the network client that is specific to infura
1
273,022
23,722,526,481
IssuesEvent
2022-08-30 16:29:53
k3s-io/k3s
https://api.github.com/repos/k3s-io/k3s
opened
Add Integration Test for common/uncommon server/agent flag options
kind/task kind/test
Similar to https://github.com/k3s-io/k3s/issues/6056. This should add tests for different options that we see. We still need to define what is common vs. uncommon, but for example, a lot of users deploy k3s with traefik and svclb disabled, so that is one configuration to have in the tests here. Feel free to add configurations as comments to this issue for tracking and to make it easier to add tests for them! To start: ``` k3s server --disable=traefik,servicelb,metrics-server,local-storage,coredns --disable-network-policy \ --disable-agent --disable-cloud-controller --flannel-backend=none --disable-scheduler \ --kube-controller-manager-arg=controllers=*,-nodeipam,-nodelifecycle,-persistentvolume-binder,-attachdetach,-persistentvolume-expander,-cloud-node-lifecycle \ --kube-apiserver-arg=endpoint-reconciler-type=none ```
1.0
Add Integration Test for common/uncommon server/agent flag options - Similar to https://github.com/k3s-io/k3s/issues/6056. This should add tests for different options that we see. We still need to define what is common vs. uncommon, but for example, a lot of users deploy k3s with traefik and svclb disabled, so that is one configuration to have in the tests here. Feel free to add configurations as comments to this issue for tracking and to make it easier to add tests for them! To start: ``` k3s server --disable=traefik,servicelb,metrics-server,local-storage,coredns --disable-network-policy \ --disable-agent --disable-cloud-controller --flannel-backend=none --disable-scheduler \ --kube-controller-manager-arg=controllers=*,-nodeipam,-nodelifecycle,-persistentvolume-binder,-attachdetach,-persistentvolume-expander,-cloud-node-lifecycle \ --kube-apiserver-arg=endpoint-reconciler-type=none ```
test
add integration test for common uncommon server agent flag options similar to this should add tests for different options that we see we still need to define what is common vs uncommon but for example a lot of users deploy with traefik and svclb disabled so that is one configuration to have in the tests here feel free to add configurations as comments to this issue for tracking and to make it easier to add tests for them to start server disable traefik servicelb metrics server local storage coredns disable network policy disable agent disable cloud controller flannel backend none disable scheduler kube controller manager arg controllers nodeipam nodelifecycle persistentvolume binder attachdetach persistentvolume expander cloud node lifecycle kube apiserver arg endpoint reconciler type none
1
401,894
27,340,994,440
IssuesEvent
2023-02-26 19:50:48
chbackstrom/Assignment-5
https://api.github.com/repos/chbackstrom/Assignment-5
closed
Assignment 5: Completed for Review
documentation
My Assignment-5 repository includes a .Rmd file (in the "figures" folder, as instructed) including two exploratory iterations of a length vs. weight figure (by fish sex) and an expository figure enhancing the presentation according to concepts covered in lecture. Please let me know if you have any questions or issues accessing the file(s).
1.0
Assignment 5: Completed for Review - My Assignment-5 repository includes a .Rmd file (in the "figures" folder, as instructed) including two exploratory iterations of a length vs. weight figure (by fish sex) and an expository figure enhancing the presentation according to concepts covered in lecture. Please let me know if you have any questions or issues accessing the file(s).
non_test
assignment completed for review my assignment repository includes a rmd file in the figures folder as instructed including two exploratory iterations of a length vs weight figure by fish sex and an expository figure enhancing the presentation according to concepts covered in lecture please let me know if you have any questions or issues accessing the file s
0
324,519
9,904,974,012
IssuesEvent
2019-06-27 10:23:46
huridocs/uwazi
https://api.github.com/repos/huridocs/uwazi
closed
Menu translations get deleted
Bug Priority: Critical Status: Sprint
When editing settings > collection > custom CSS, Menu translations get deleted. 1. Add a menu to a blank Uwazi with at least 2 languages. 2. The translations are created in settings > translations > menu. 3. Go to settings > collection > custom css 4. Edit and save. 5. Go to settings > translations > menu. The translations have disappeared.
1.0
Menu translations get deleted - When editing settings > collection > custom CSS, Menu translations get deleted. 1. Add a menu to a blank Uwazi with at least 2 languages. 2. The translations are created in settings > translations > menu. 3. Go to settings > collection > custom css 4. Edit and save. 5. Go to settings > translations > menu. The translations have disappeared.
non_test
menu translations get deleted when editing settings collection custom css menu translations get deleted add a menu to a blank uwazi with at least languages the translations are created in settings translations menu go to settings collection custom css edit and save go to settings translations menu the translations have disappeared
0
38,845
5,201,415,356
IssuesEvent
2017-01-24 04:36:48
ElucidataInc/ElMaven
https://api.github.com/repos/ElucidataInc/ElMaven
closed
allow .txt extension for tab-delimited files
enhancement fixed medium tested
Maven always wants .tab for tab-delimited input files (like compound database) which is just mildly annoying since most people use Excel to edit the file, and Excel only saves as .txt . Let's just allow .txt extension and assume it's tab-delimited. (maybe at some point in the future we can do what spotfire and excel do and guess the format/delimiter, but this is extremely low priority).
1.0
allow .txt extension for tab-delimited files - Maven always wants .tab for tab-delimited input files (like compound database) which is just mildly annoying since most people use Excel to edit the file, and Excel only saves as .txt . Let's just allow .txt extension and assume it's tab-delimited. (maybe at some point in the future we can do what spotfire and excel do and guess the format/delimiter, but this is extremely low priority).
test
allow txt extension for tab delimited files maven always wants tab for tab delimited input files like compound database which is just mildly annoying since most people use excel to edit the file and excel only saves as txt let s just allow txt extension and assume it s tab delimited maybe at some point in the future we can do what spotfire and excel do and guess the format delimiter but this is extremely low priority
1
91,412
8,304,750,049
IssuesEvent
2018-09-21 22:39:40
excellaco/play-slick-postgres.g8
https://api.github.com/repos/excellaco/play-slick-postgres.g8
opened
Add Evolutions Tests
nice-to-have test-coverage
Though we can't see it in Statement Coverage report, it would be nice to include a test for applyEvolutions using H2. This can be helpful to warn of an evolution that will break when Play applies it (except of course in cases of merge conflicts).
1.0
Add Evolutions Tests - Though we can't see it in Statement Coverage report, it would be nice to include a test for applyEvolutions using H2. This can be helpful to warn of an evolution that will break when Play applies it (except of course in cases of merge conflicts).
test
add evolutions tests though we can t see it in statement coverage report it would be nice to include a test for applyevolutions using this can be helpful to warn of an evolution that will break when play applies it except of course in cases of merge conflicts
1
269,931
8,444,584,923
IssuesEvent
2018-10-18 18:52:27
ampproject/amphtml
https://api.github.com/repos/ampproject/amphtml
closed
amp-youtube autoplay does not work unless initially in viewport
Category: Audio&Video P1: High Priority Type: Bug
not a regression of our code but it used to work before a month ago or so. Go to https://ampbyexample.com/components/amp-youtube/ Scroll down See the autoplaying embed has started playing but is paused. something with youtube embed may have changed. /cc @kevinkassimo looks similar to the issue you found where messages where not received if user scrolls to amp-youtube.
1.0
amp-youtube autoplay does not work unless initially in viewport - not a regression of our code but it used to work before a month ago or so. Go to https://ampbyexample.com/components/amp-youtube/ Scroll down See the autoplaying embed has started playing but is paused. something with youtube embed may have changed. /cc @kevinkassimo looks similar to the issue you found where messages where not received if user scrolls to amp-youtube.
non_test
amp youtube autoplay does not work unless initially in viewport not a regression of our code but it used to work before a month ago or so go to scroll down see the autoplaying embed has started playing but is paused something with youtube embed may have changed cc kevinkassimo looks similar to the issue you found where messages where not received if user scrolls to amp youtube
0
471,387
13,565,846,606
IssuesEvent
2020-09-18 12:23:50
bc-ticketing/host-client
https://api.github.com/repos/bc-ticketing/host-client
opened
Export and Import Seating Plan as JSON
event feature frontend medium priority
A host should be able to export and import a json of a seating plan. To avoid a host creating the exact same seating plan every time a new event is planned, the host should be able to export the seating plan in a json. The next time an event is created, the host can just import this json file again, set the prices etc. for each type and then create the types without creating the whole plan from scratch.
1.0
Export and Import Seating Plan as JSON - A host should be able to export and import a json of a seating plan. To avoid a host creating the exact same seating plan every time a new event is planned, the host should be able to export the seating plan in a json. The next time an event is created, the host can just import this json file again, set the prices etc. for each type and then create the types without creating the whole plan from scratch.
non_test
export and import seating plan as json a host should be able to export and import a json of a seating plan to avoid a host creating the exact same seating plan every time a new event is planned the host should be able to export the seating plan in a json the next time an event is created the host can just import this json file again set the prices etc for each type and then create the types without creating the whole plan from scratch
0
146,192
13,173,470,756
IssuesEvent
2020-08-11 20:22:59
leomaurodesenv/data-science-api-framework
https://api.github.com/repos/leomaurodesenv/data-science-api-framework
closed
swagger integration
documentation enhancement
Automatize the [Swagger](https://swagger.io/) documentation to generate a simple documentation based on the flask. Additional documentation: - https://pypi.org/project/flask-swagger/ - https://flask-restplus.readthedocs.io/en/stable/swagger.html - http://michal.karzynski.pl/blog/2016/06/19/building-beautiful-restful-apis-using-flask-swagger-ui-flask-restplus/
1.0
swagger integration - Automatize the [Swagger](https://swagger.io/) documentation to generate a simple documentation based on the flask. Additional documentation: - https://pypi.org/project/flask-swagger/ - https://flask-restplus.readthedocs.io/en/stable/swagger.html - http://michal.karzynski.pl/blog/2016/06/19/building-beautiful-restful-apis-using-flask-swagger-ui-flask-restplus/
non_test
swagger integration automatize the documentation to generate a simple documentation based on the flask additional documentation
0
8,694
7,570,597,083
IssuesEvent
2018-04-23 09:29:33
matthiasbeyer/imag
https://api.github.com/repos/matthiasbeyer/imag
opened
Idea: libimagentrydatapoint
complexity/high kind/enhancement kind/infrastructure meta/importance/medium
One idea I had: A library for writing data points to entries. The library would not only handle unit conversions (whereas the unit is a thing which can be defined by the user of the library) but also aggregation, statistics and possibly even anomaly detection. --- This library can then be used for other tools, for example for a health tracker, a crypto currency price tracker etc etc --- Details have to be worked out of course.
1.0
Idea: libimagentrydatapoint - One idea I had: A library for writing data points to entries. The library would not only handle unit conversions (whereas the unit is a thing which can be defined by the user of the library) but also aggregation, statistics and possibly even anomaly detection. --- This library can then be used for other tools, for example for a health tracker, a crypto currency price tracker etc etc --- Details have to be worked out of course.
non_test
idea libimagentrydatapoint one idea i had a library for writing data points to entries the library would not only handle unit conversions whereas the unit is a thing which can be defined by the user of the library but also aggregation statistics and possibly even anomaly detection this library can then be used for other tools for example for a health tracker a crypto currency price tracker etc etc details have to be worked out of course
0
201,838
15,226,848,239
IssuesEvent
2021-02-18 09:26:12
WeiXian042901/fyp_repository
https://api.github.com/repos/WeiXian042901/fyp_repository
opened
FU_041_Quiz Play Page Two Options(Display Only Two Options)
Acceptance Test Quiz User
**Test Scenario** - User has chosen to attempt the Testing Title(Two Options) quiz **Test Case** - Check that the system displays only two available options for the users to choose from **Pre-Conditions** User has successfully entered the application User clicked on the “Quizzes” Option User selected the “Testing title(two options)” quiz option User clicked on the “Start Quiz” button. **Test-Steps** **Test Data** **Expected Results** - The system should only displays two available options for the users to chooses from **Actual Results** - The system only displays two available options for the users to chooses from **Pass/Fail** - Pass **Date Tested** - 10th February 2021 **Tested By** - Zachary Tan
1.0
FU_041_Quiz Play Page Two Options(Display Only Two Options) - **Test Scenario** - User has chosen to attempt the Testing Title(Two Options) quiz **Test Case** - Check that the system displays only two available options for the users to choose from **Pre-Conditions** User has successfully entered the application User clicked on the “Quizzes” Option User selected the “Testing title(two options)” quiz option User clicked on the “Start Quiz” button. **Test-Steps** **Test Data** **Expected Results** - The system should only displays two available options for the users to chooses from **Actual Results** - The system only displays two available options for the users to chooses from **Pass/Fail** - Pass **Date Tested** - 10th February 2021 **Tested By** - Zachary Tan
test
fu quiz play page two options display only two options test scenario user has chosen to attempt the testing title two options quiz test case check that the system displays only two available options for the users to choose from pre conditions user has successfully entered the application user clicked on the “quizzes” option user selected the “testing title two options ” quiz option user clicked on the “start quiz” button test steps test data expected results the system should only displays two available options for the users to chooses from actual results the system only displays two available options for the users to chooses from pass fail pass date tested february tested by zachary tan
1
284,223
24,584,201,248
IssuesEvent
2022-10-13 18:10:09
lightningnetwork/lnd
https://api.github.com/repos/lightningnetwork/lnd
closed
tests: race in TestBlockCacheMutexes tests
tests data-race
Saw a test run fail w/ a race, looks like it's in the mock backend itself and not the actual code: ``` ================== WARNING: DATA RACE Read at 0x00c0001ca2d8 by goroutine 108: github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:33 +0xda github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock-fm() <autogenerated>:1 +0x44 github.com/lightningnetwork/lnd/blockcache.(*BlockCache).GetBlock() /home/runner/work/lnd/lnd/blockcache/blockcache.go:52 +0x3b6 github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func1() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:1[77](https://github.com/lightningnetwork/lnd/actions/runs/3011169890/jobs/4838020176#step:6:78) +0xb0 github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func2() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:182 +0x47 Previous write at 0x00c0001ca2d8 by goroutine 51: github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:33 +0xf8 github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock-fm() <autogenerated>:1 +0x44 github.com/lightningnetwork/lnd/blockcache.(*BlockCache).GetBlock() /home/runner/work/lnd/lnd/blockcache/blockcache.go:52 +0x3b6 github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func1() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:174 +0x7b github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func2() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:1[82](https://github.com/lightningnetwork/lnd/actions/runs/3011169890/jobs/4838020176#step:6:83) +0x47 Goroutine 108 (running) created at: github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:172 +0x824 testing.tRunner() /opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:1446 +0x216 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:14[93](https://github.com/lightningnetwork/lnd/actions/runs/3011169890/jobs/4838020176#step:6:94) +0x47 Goroutine 51 (finished) created at: github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:172 +0x824 testing.tRunner() /opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:1446 +0x216 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:1493 +0x47 ================== --- FAIL: TestBlockCacheMutexes (0.01s) ```
1.0
tests: race in TestBlockCacheMutexes tests - Saw a test run fail w/ a race, looks like it's in the mock backend itself and not the actual code: ``` ================== WARNING: DATA RACE Read at 0x00c0001ca2d8 by goroutine 108: github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:33 +0xda github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock-fm() <autogenerated>:1 +0x44 github.com/lightningnetwork/lnd/blockcache.(*BlockCache).GetBlock() /home/runner/work/lnd/lnd/blockcache/blockcache.go:52 +0x3b6 github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func1() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:1[77](https://github.com/lightningnetwork/lnd/actions/runs/3011169890/jobs/4838020176#step:6:78) +0xb0 github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func2() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:182 +0x47 Previous write at 0x00c0001ca2d8 by goroutine 51: github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:33 +0xf8 github.com/lightningnetwork/lnd/blockcache.(*mockChainBackend).GetBlock-fm() <autogenerated>:1 +0x44 github.com/lightningnetwork/lnd/blockcache.(*BlockCache).GetBlock() /home/runner/work/lnd/lnd/blockcache/blockcache.go:52 +0x3b6 github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func1() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:174 +0x7b github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes.func2() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:1[82](https://github.com/lightningnetwork/lnd/actions/runs/3011169890/jobs/4838020176#step:6:83) +0x47 Goroutine 108 (running) created at: github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:172 +0x824 testing.tRunner() /opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:1446 +0x216 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:14[93](https://github.com/lightningnetwork/lnd/actions/runs/3011169890/jobs/4838020176#step:6:94) +0x47 Goroutine 51 (finished) created at: github.com/lightningnetwork/lnd/blockcache.TestBlockCacheMutexes() /home/runner/work/lnd/lnd/blockcache/blockcache_test.go:172 +0x824 testing.tRunner() /opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:1446 +0x216 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.19.0/x64/src/testing/testing.go:1493 +0x47 ================== --- FAIL: TestBlockCacheMutexes (0.01s) ```
test
tests race in testblockcachemutexes tests saw a test run fail w a race looks like it s in the mock backend itself and not the actual code warning data race read at by goroutine github com lightningnetwork lnd blockcache mockchainbackend getblock home runner work lnd lnd blockcache blockcache test go github com lightningnetwork lnd blockcache mockchainbackend getblock fm github com lightningnetwork lnd blockcache blockcache getblock home runner work lnd lnd blockcache blockcache go github com lightningnetwork lnd blockcache testblockcachemutexes home runner work lnd lnd blockcache blockcache test go github com lightningnetwork lnd blockcache testblockcachemutexes home runner work lnd lnd blockcache blockcache test go previous write at by goroutine github com lightningnetwork lnd blockcache mockchainbackend getblock home runner work lnd lnd blockcache blockcache test go github com lightningnetwork lnd blockcache mockchainbackend getblock fm github com lightningnetwork lnd blockcache blockcache getblock home runner work lnd lnd blockcache blockcache go github com lightningnetwork lnd blockcache testblockcachemutexes home runner work lnd lnd blockcache blockcache test go github com lightningnetwork lnd blockcache testblockcachemutexes home runner work lnd lnd blockcache blockcache test go goroutine running created at github com lightningnetwork lnd blockcache testblockcachemutexes home runner work lnd lnd blockcache blockcache test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go goroutine finished created at github com lightningnetwork lnd blockcache testblockcachemutexes home runner work lnd lnd blockcache blockcache test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go fail testblockcachemutexes
1
302,759
26,160,950,366
IssuesEvent
2022-12-31 14:16:55
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: tlp failed
C-test-failure O-robot O-roachtest release-blocker branch-release-22.2
roachtest.tlp [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=artifacts#/tlp) on release-22.2 @ [07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1](https://github.com/cockroachdb/cockroach/commits/07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1): ``` test artifacts and logs in: /artifacts/tlp/run_1 (test_impl.go:286).Fatal: pq: Use of partitions requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out. ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #94551 roachtest: tlp failed [C-test-failure O-roachtest O-robot T-sql-queries branch-master release-blocker] </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*tlp.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: tlp failed - roachtest.tlp [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=artifacts#/tlp) on release-22.2 @ [07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1](https://github.com/cockroachdb/cockroach/commits/07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1): ``` test artifacts and logs in: /artifacts/tlp/run_1 (test_impl.go:286).Fatal: pq: Use of partitions requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out. ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #94551 roachtest: tlp failed [C-test-failure O-roachtest O-robot T-sql-queries branch-master release-blocker] </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*tlp.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest tlp failed roachtest tlp with on release test artifacts and logs in artifacts tlp run test impl go fatal pq use of partitions requires an enterprise license your evaluation license expired on december if you re interested in getting a new license please contact subscriptions cockroachlabs com and we can help you out parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see same failure on other branches roachtest tlp failed cc cockroachdb sql queries
1
543,523
15,883,376,446
IssuesEvent
2021-04-09 17:18:06
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Ask password option: Provide explanation and image of management console config
Affected/5.4.0-Update1 Priority/Highest docs
In [1] the management console config section does not contain an image where the expected configs are displayed. Please provide it. Furthermore in management console under User on boarding section there are other configs as well. Please explain the usage of them and mention if they are not relevant to this particular case, [1] https://docs.wso2.com/display/IS540/Creating+Users+using+the+Ask+Password+Option
1.0
Ask password option: Provide explanation and image of management console config - In [1] the management console config section does not contain an image where the expected configs are displayed. Please provide it. Furthermore in management console under User on boarding section there are other configs as well. Please explain the usage of them and mention if they are not relevant to this particular case, [1] https://docs.wso2.com/display/IS540/Creating+Users+using+the+Ask+Password+Option
non_test
ask password option provide explanation and image of management console config in the management console config section does not contain an image where the expected configs are displayed please provide it furthermore in management console under user on boarding section there are other configs as well please explain the usage of them and mention if they are not relevant to this particular case
0
111,877
17,038,617,138
IssuesEvent
2021-07-05 10:23:45
tamirdahan/NodeTestBench
https://api.github.com/repos/tamirdahan/NodeTestBench
opened
CVE-2021-23343 (High) detected in path-parse-1.0.6.tgz
security vulnerability
## CVE-2021-23343 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>path-parse-1.0.6.tgz</b></p></summary> <p>Node.js path.parse() ponyfill</p> <p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz</a></p> <p>Path to dependency file: NodeTestBench/package.json</p> <p>Path to vulnerable library: NodeTestBench/node_modules/path-parse/package.json</p> <p> Dependency Hierarchy: - eslint-config-1.0.3.tgz (Root Library) - eslint-plugin-node-9.1.0.tgz - resolve-1.12.0.tgz - :x: **path-parse-1.0.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tamirdahan/NodeTestBench/commit/20c4644edbd7cd6f64e400c703d1594170fe90df">20c4644edbd7cd6f64e400c703d1594170fe90df</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-05-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jbgutierrez/path-parse/issues/8">https://github.com/jbgutierrez/path-parse/issues/8</a></p> <p>Release Date: 2021-05-04</p> <p>Fix Resolution: path-parse - 1.0.7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"path-parse","packageVersion":"1.0.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"@contrast/eslint-config:1.0.3;eslint-plugin-node:9.1.0;resolve:1.12.0;path-parse:1.0.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"path-parse - 1.0.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23343","vulnerabilityDetails":"All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-23343 (High) detected in path-parse-1.0.6.tgz - ## CVE-2021-23343 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>path-parse-1.0.6.tgz</b></p></summary> <p>Node.js path.parse() ponyfill</p> <p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz</a></p> <p>Path to dependency file: NodeTestBench/package.json</p> <p>Path to vulnerable library: NodeTestBench/node_modules/path-parse/package.json</p> <p> Dependency Hierarchy: - eslint-config-1.0.3.tgz (Root Library) - eslint-plugin-node-9.1.0.tgz - resolve-1.12.0.tgz - :x: **path-parse-1.0.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tamirdahan/NodeTestBench/commit/20c4644edbd7cd6f64e400c703d1594170fe90df">20c4644edbd7cd6f64e400c703d1594170fe90df</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-05-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jbgutierrez/path-parse/issues/8">https://github.com/jbgutierrez/path-parse/issues/8</a></p> <p>Release Date: 2021-05-04</p> <p>Fix Resolution: path-parse - 1.0.7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"path-parse","packageVersion":"1.0.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"@contrast/eslint-config:1.0.3;eslint-plugin-node:9.1.0;resolve:1.12.0;path-parse:1.0.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"path-parse - 1.0.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23343","vulnerabilityDetails":"All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_test
cve high detected in path parse tgz cve high severity vulnerability vulnerable library path parse tgz node js path parse ponyfill library home page a href path to dependency file nodetestbench package json path to vulnerable library nodetestbench node modules path parse package json dependency hierarchy eslint config tgz root library eslint plugin node tgz resolve tgz x path parse tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of package path parse are vulnerable to regular expression denial of service redos via splitdevicere splittailre and splitpathre regular expressions redos exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution path parse isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree contrast eslint config eslint plugin node resolve path parse isminimumfixversionavailable true minimumfixversion path parse basebranches vulnerabilityidentifier cve vulnerabilitydetails all versions of package path parse are vulnerable to regular expression denial of service redos via splitdevicere splittailre and splitpathre regular expressions redos exhibits polynomial worst case time complexity vulnerabilityurl
0
167,122
14,102,269,171
IssuesEvent
2020-11-06 08:28:45
lensapp/lens
https://api.github.com/repos/lensapp/lens
opened
Documentation for UI components
area/documentation area/ui
We need to document inner Lens UI components using one of the utilities: * Storybook (https://storybook.js.org/) * Docz (https://www.docz.site/) * Blueprint (https://blueprintjs.com/) Each documented component should have at least: * Brief description (if needed) * Its visual representation * List of props
1.0
Documentation for UI components - We need to document inner Lens UI components using one of the utilities: * Storybook (https://storybook.js.org/) * Docz (https://www.docz.site/) * Blueprint (https://blueprintjs.com/) Each documented component should have at least: * Brief description (if needed) * Its visual representation * List of props
non_test
documentation for ui components we need to document inner lens ui components using one of the utilities storybook docz blueprint each documented component should have at least brief description if needed its visual representation list of props
0
90,032
25,953,823,886
IssuesEvent
2022-12-18 00:14:31
QubesOS/updates-status
https://api.github.com/repos/QubesOS/updates-status
closed
desktop-linux-manager v4.2.0 (r4.2)
r4.2-vm-bookworm-building r4.2-host-cur-test r4.2-vm-bullseye-cur-test r4.2-vm-fc37-cur-test r4.2-vm-fc36-cur-test r4.2-vm-centos-stream8-cur-test
Update of desktop-linux-manager to v4.2.0 for Qubes r4.2, see comments below for details and build status. From commit: https://github.com/QubesOS/qubes-desktop-linux-manager/commit/f5f5a00a41217dd2c036843fd669968ee3e2248c [Changes since previous version](https://github.com/QubesOS/qubes-desktop-linux-manager/compare/v4.1.15...v4.2.0): QubesOS/qubes-desktop-linux-manager@f5f5a00 version 4.2.0 QubesOS/qubes-desktop-linux-manager@e116e35 Add detecting changes in raw policy text QubesOS/qubes-desktop-linux-manager@24b20bf Fix updating initial value in text dropdowns QubesOS/qubes-desktop-linux-manager@16badf1 Improve formatting QubesOS/qubes-desktop-linux-manager@68aa035 Remove a superfluous variable QubesOS/qubes-desktop-linux-manager@1bdc940 Refactor raw policy handler as a separate handler QubesOS/qubes-desktop-linux-manager@a6d431f Improve error handling in policy config QubesOS/qubes-desktop-linux-manager@e85099f Change Open File/URL policy to more reasonable version QubesOS/qubes-desktop-linux-manager@5c9bd66 Refactor ExceptionsList to a separate files QubesOS/qubes-desktop-linux-manager@8a4ec98 Improve AdminVM approach: remove dom0 from most policy configs QubesOS/qubes-desktop-linux-manager@e4dba08 Use --get-default-whitelist qubes-appmenus option for Create New Qube QubesOS/qubes-desktop-linux-manager@6f7ddab Add 'copy system info to global clipboard' button QubesOS/qubes-desktop-linux-manager@bde5175 Show errors on loading policy rules QubesOS/qubes-desktop-linux-manager@ea1a5e3 Add gitlab CI integration QubesOS/qubes-desktop-linux-manager@8abee61 New Config Tool and new New Qube tool Referenced issues: If you're release manager, you can issue GPG-inline signed command: * `Upload-component r4.2 desktop-linux-manager f5f5a00a41217dd2c036843fd669968ee3e2248c current all` (available 5 days from now) * `Upload-component r4.2 desktop-linux-manager f5f5a00a41217dd2c036843fd669968ee3e2248c security-testing` You can choose subset of distributions like: * `Upload-component r4.2 desktop-linux-manager f5f5a00a41217dd2c036843fd669968ee3e2248c current vm-bookworm,vm-fc37` (available 5 days from now) Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it). For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
1.0
desktop-linux-manager v4.2.0 (r4.2) - Update of desktop-linux-manager to v4.2.0 for Qubes r4.2, see comments below for details and build status. From commit: https://github.com/QubesOS/qubes-desktop-linux-manager/commit/f5f5a00a41217dd2c036843fd669968ee3e2248c [Changes since previous version](https://github.com/QubesOS/qubes-desktop-linux-manager/compare/v4.1.15...v4.2.0): QubesOS/qubes-desktop-linux-manager@f5f5a00 version 4.2.0 QubesOS/qubes-desktop-linux-manager@e116e35 Add detecting changes in raw policy text QubesOS/qubes-desktop-linux-manager@24b20bf Fix updating initial value in text dropdowns QubesOS/qubes-desktop-linux-manager@16badf1 Improve formatting QubesOS/qubes-desktop-linux-manager@68aa035 Remove a superfluous variable QubesOS/qubes-desktop-linux-manager@1bdc940 Refactor raw policy handler as a separate handler QubesOS/qubes-desktop-linux-manager@a6d431f Improve error handling in policy config QubesOS/qubes-desktop-linux-manager@e85099f Change Open File/URL policy to more reasonable version QubesOS/qubes-desktop-linux-manager@5c9bd66 Refactor ExceptionsList to a separate files QubesOS/qubes-desktop-linux-manager@8a4ec98 Improve AdminVM approach: remove dom0 from most policy configs QubesOS/qubes-desktop-linux-manager@e4dba08 Use --get-default-whitelist qubes-appmenus option for Create New Qube QubesOS/qubes-desktop-linux-manager@6f7ddab Add 'copy system info to global clipboard' button QubesOS/qubes-desktop-linux-manager@bde5175 Show errors on loading policy rules QubesOS/qubes-desktop-linux-manager@ea1a5e3 Add gitlab CI integration QubesOS/qubes-desktop-linux-manager@8abee61 New Config Tool and new New Qube tool Referenced issues: If you're release manager, you can issue GPG-inline signed command: * `Upload-component r4.2 desktop-linux-manager f5f5a00a41217dd2c036843fd669968ee3e2248c current all` (available 5 days from now) * `Upload-component r4.2 desktop-linux-manager f5f5a00a41217dd2c036843fd669968ee3e2248c security-testing` You can choose subset of distributions like: * `Upload-component r4.2 desktop-linux-manager f5f5a00a41217dd2c036843fd669968ee3e2248c current vm-bookworm,vm-fc37` (available 5 days from now) Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it). For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
non_test
desktop linux manager update of desktop linux manager to for qubes see comments below for details and build status from commit qubesos qubes desktop linux manager version qubesos qubes desktop linux manager add detecting changes in raw policy text qubesos qubes desktop linux manager fix updating initial value in text dropdowns qubesos qubes desktop linux manager improve formatting qubesos qubes desktop linux manager remove a superfluous variable qubesos qubes desktop linux manager refactor raw policy handler as a separate handler qubesos qubes desktop linux manager improve error handling in policy config qubesos qubes desktop linux manager change open file url policy to more reasonable version qubesos qubes desktop linux manager refactor exceptionslist to a separate files qubesos qubes desktop linux manager improve adminvm approach remove from most policy configs qubesos qubes desktop linux manager use get default whitelist qubes appmenus option for create new qube qubesos qubes desktop linux manager add copy system info to global clipboard button qubesos qubes desktop linux manager show errors on loading policy rules qubesos qubes desktop linux manager add gitlab ci integration qubesos qubes desktop linux manager new config tool and new new qube tool referenced issues if you re release manager you can issue gpg inline signed command upload component desktop linux manager current all available days from now upload component desktop linux manager security testing you can choose subset of distributions like upload component desktop linux manager current vm bookworm vm available days from now above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it for more information on how to test this update please take a look at
0
83,277
7,867,928,542
IssuesEvent
2018-06-23 15:06:27
jbeard4/SCION
https://api.github.com/repos/jbeard4/SCION
closed
fail test/scxml-test-framework/test/w3c-ecma/test187.txml.scxml
2.0.0 Node v0.10.24 Tests fail feature:invoke
[https://github.com/jbeard4/scxml-test-framework/tree/master/test/scxml-test-framework/test/w3c-ecma/test187.txml.scxml](https://github.com/jbeard4/scxml-test-framework/tree/master/test/scxml-test-framework/test/w3c-ecma/test187.txml.scxml) **Error** <code><pre>{ "name": "AssertionError", "actual": [ "subFinal" ], "expected": [ "pass" ], "operator": "deepEqual", "message": "[\"subFinal\"] deepEqual [\"pass\"]" }</code></pre> **Data**: <code><pre>{ "sessionToken": 551, "nextConfiguration": [ "subFinal" ] }</code></pre> **scxml**: ``` xml <?xml version="1.0" encoding="UTF-8"?> <!-- we test that delayed <send> is not sent if the sending session terminates. In this case, a subscript is invoked which sends the event childToParent delayed by 1 second, and then terminates. The parent session, should not receive childToParent. If it does, we fail. Otherwise the 10 sec timer expires and we pass --> <scxml xmlns="http://www.w3.org/2005/07/scxml" xmlns:conf="http://www.w3.org/2005/scxml-conformance" initial="s0" version="1.0" datamodel="ecmascript"> <state id="s0"> <onentry> <send event="timeout" delay="10s"/> </onentry> <invoke type="scxml"> <content> <!-- exit before the delayed send can execute --> <scxml initial="sub0" version="1.0" datamodel="ecmascript"> <state id="sub0"> <onentry> <send event="childToParent" target="#_parent" delay="1s"/> </onentry> <transition target="subFinal"/> </state> <final id="subFinal"/> </scxml> </content> </invoke> <transition event="childToParent" target="fail"/> <transition event="timeout" target="pass"/> </state> <final id="pass"><onentry><log label="Outcome" expr="'pass'"/></onentry></final> <final id="fail"><onentry><log label="Outcome" expr="'fail'"/></onentry></final> </scxml> ``` **JSON**: <code><pre>{ "initialConfiguration": [ "pass" ], "events": [] }</code></pre>
1.0
fail test/scxml-test-framework/test/w3c-ecma/test187.txml.scxml - [https://github.com/jbeard4/scxml-test-framework/tree/master/test/scxml-test-framework/test/w3c-ecma/test187.txml.scxml](https://github.com/jbeard4/scxml-test-framework/tree/master/test/scxml-test-framework/test/w3c-ecma/test187.txml.scxml) **Error** <code><pre>{ "name": "AssertionError", "actual": [ "subFinal" ], "expected": [ "pass" ], "operator": "deepEqual", "message": "[\"subFinal\"] deepEqual [\"pass\"]" }</code></pre> **Data**: <code><pre>{ "sessionToken": 551, "nextConfiguration": [ "subFinal" ] }</code></pre> **scxml**: ``` xml <?xml version="1.0" encoding="UTF-8"?> <!-- we test that delayed <send> is not sent if the sending session terminates. In this case, a subscript is invoked which sends the event childToParent delayed by 1 second, and then terminates. The parent session, should not receive childToParent. If it does, we fail. Otherwise the 10 sec timer expires and we pass --> <scxml xmlns="http://www.w3.org/2005/07/scxml" xmlns:conf="http://www.w3.org/2005/scxml-conformance" initial="s0" version="1.0" datamodel="ecmascript"> <state id="s0"> <onentry> <send event="timeout" delay="10s"/> </onentry> <invoke type="scxml"> <content> <!-- exit before the delayed send can execute --> <scxml initial="sub0" version="1.0" datamodel="ecmascript"> <state id="sub0"> <onentry> <send event="childToParent" target="#_parent" delay="1s"/> </onentry> <transition target="subFinal"/> </state> <final id="subFinal"/> </scxml> </content> </invoke> <transition event="childToParent" target="fail"/> <transition event="timeout" target="pass"/> </state> <final id="pass"><onentry><log label="Outcome" expr="'pass'"/></onentry></final> <final id="fail"><onentry><log label="Outcome" expr="'fail'"/></onentry></final> </scxml> ``` **JSON**: <code><pre>{ "initialConfiguration": [ "pass" ], "events": [] }</code></pre>
test
fail test scxml test framework test ecma txml scxml error name assertionerror actual subfinal expected pass operator deepequal message deepequal data sessiontoken nextconfiguration subfinal scxml xml is not sent if the sending session terminates in this case a subscript is invoked which sends the event childtoparent delayed by second and then terminates the parent session should not receive childtoparent if it does we fail otherwise the sec timer expires and we pass json initialconfiguration pass events
1
4,240
6,493,246,173
IssuesEvent
2017-08-21 16:14:00
Microsoft/UWPCommunityToolkit
https://api.github.com/repos/Microsoft/UWPCommunityToolkit
closed
Resolve service dependencies on obsolete HttpHelper class
help wanted PR in progress services
HttpHelper will become obsolete in v2.0 of the toolkit. We need to remove service dependencies from HttpHelper to HttpClient to prevent a large number of warnings upon building.
1.0
Resolve service dependencies on obsolete HttpHelper class - HttpHelper will become obsolete in v2.0 of the toolkit. We need to remove service dependencies from HttpHelper to HttpClient to prevent a large number of warnings upon building.
non_test
resolve service dependencies on obsolete httphelper class httphelper will become obsolete in of the toolkit we need to remove service dependencies from httphelper to httpclient to prevent a large number of warnings upon building
0
381,297
26,445,626,310
IssuesEvent
2023-01-16 06:58:22
GameCache/CodeTemplate
https://api.github.com/repos/GameCache/CodeTemplate
opened
Add Initial Structure
documentation enhancement
<!--- Please search issues before creating a new one. ---> <!--- This is just a template, feel free to customize if prudent. ---> ## Wanted Behavior <!--- Describe the wanted feature. ---> Ability to use this repository as a seed for others. ## Possible Solution <!--- Optional implementation suggestion. ---> Add a default implementation with the common structure to use for other projects. ## Additional Information <!--- For example, why do you want the feature. ---> Also provides a way to test new language enhancements for already created projects.
1.0
Add Initial Structure - <!--- Please search issues before creating a new one. ---> <!--- This is just a template, feel free to customize if prudent. ---> ## Wanted Behavior <!--- Describe the wanted feature. ---> Ability to use this repository as a seed for others. ## Possible Solution <!--- Optional implementation suggestion. ---> Add a default implementation with the common structure to use for other projects. ## Additional Information <!--- For example, why do you want the feature. ---> Also provides a way to test new language enhancements for already created projects.
non_test
add initial structure wanted behavior ability to use this repository as a seed for others possible solution add a default implementation with the common structure to use for other projects additional information also provides a way to test new language enhancements for already created projects
0
47,945
19,842,709,587
IssuesEvent
2022-01-21 00:17:19
Azure/azure-sdk-for-python
https://api.github.com/repos/Azure/azure-sdk-for-python
opened
cognitiveservices/azure-cognitiveservices-knowledge-qnamaker - cspell found spelling errors in public API surface
Cognitive Services Client
Spell check scanning of package at `sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker` detected spelling errors in the public API surface. This directory is opted out of PR spell checking in PR #22199 to keep PRs unblocked. ## What to do 1. Ensure Node.js is installed (https://nodejs.org/en/download/). 1. Delete the entry in `.vscode/cspell.json`'s `ignorePaths` field. It will look like: `sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/**`. You need to do this to enable checking the files. 1. From the root of the repo run spell check using `./eng/common/spelling/Invoke-Cspell.ps1 -ScanGlobs "sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/**"` 1. Fix detections according to http://aka.ms/azsdk/engsys/spellcheck use the "False positives" section to fix false positives 1. Check in changes (including the change to `.vscode/cspell.json` where the `ignorePaths` is updated to remove the entry for this service.). You may need to run `git add -f .vscode/cspell.json` to force adding the changes to the file in git. ## Spell checking output ``` ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/__init__.py:12:10 - Unknown word (amaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:29:14 - Unknown word (adto) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:29:40 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:30:14 - Unknown word (adto) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:30:32 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:33:14 - Unknown word (adocuments) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:54:14 - Unknown word (asearch) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:55:14 - Unknown word (asearch) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:56:14 - Unknown word (asearch) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:57:14 - Unknown word (asearch) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:79:14 - Unknown word (adto) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:79:36 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:80:14 - Unknown word (adto) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:80:28 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:83:14 - Unknown word (adocuments) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:112:10 - Unknown word (amaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/context_dto_py3.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/context_dto.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:22:37 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:25:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:25:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:30:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:62:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:22:37 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:25:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:25:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:30:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:62:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:21:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:21:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:30:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:21:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:21:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:30:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/endpoint_settings_dto_py3.py:20:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/endpoint_settings_dto.py:20:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_py3.py:28:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_py3.py:36:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_py3.py:40:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error_py3.py:27:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error_py3.py:35:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error_py3.py:39:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error.py:27:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error.py:35:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error.py:39:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_py3.py:22:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response.py:22:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error.py:28:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error.py:36:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error.py:40:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/feedback_records_dto_py3.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/feedback_records_dto.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/inner_error_model_py3.py:26:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/inner_error_model.py:26:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/knowledgebases_dto_py3.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/knowledgebases_dto.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation_py3.py:21:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation_py3.py:36:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation.py:21:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation.py:36:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:22:8 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:24:19 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:24:44 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:26:51 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:12:10 - Unknown word (adto) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:12:28 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:15:22 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:16:10 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:16:35 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:31:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:34:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:12:10 - Unknown word (adto) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:12:24 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:15:22 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:16:10 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:16:35 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:31:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:34:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:22:8 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:24:19 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:24:44 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:26:51 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:16:18 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:20:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:24:62 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:16:18 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:20:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:24:62 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context_py3.py:15:9 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context_py3.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context_py3.py:39:17 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context.py:15:9 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context.py:39:17 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:15:9 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:31:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:34:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:34:59 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:52:50 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:57:17 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:15:9 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:31:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:34:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:34:59 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:52:50 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:57:17 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_context_py3.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_context.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_list_py3.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_list.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_py3.py:30:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_py3.py:33:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_py3.py:37:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result.py:30:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result.py:33:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result.py:37:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:33:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:40:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:45:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:48:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:33:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:40:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:45:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:48:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:20:15 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:20:47 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:22:14 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:23:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:27:13 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:31:13 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:31:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:20:15 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:20:47 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:22:14 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:23:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:27:13 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:31:13 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:31:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_context_dto_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_context_dto.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_contents_dto_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_contents_dto.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:21:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:21:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:30:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:21:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:21:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:30:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_py3.py:20:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_py3.py:23:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_py3.py:26:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_update_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_update.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto.py:20:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto.py:23:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto.py:26:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto_py3.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_context_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_context.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata_py3.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_py3.py:27:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_py3.py:30:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_py3.py:33:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto.py:27:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto.py:30:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto.py:33:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/word_alterations_dto_py3.py:22:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/word_alterations_dto.py:22:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:47:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:50:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:93:50 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:102:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:47:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:50:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:100:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:103:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:47:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:50:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:93:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:102:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:47:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:50:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:100:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:103:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:231:29 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:236:19 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:236:41 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:239:18 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:240:68 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:251:45 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/operations.py:48:52 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/operations.py:51:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:29:6 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:43:32 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:47:64 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:60:67 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/CHANGELOG.md:14:76 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/CHANGELOG.md:18:87 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/CHANGELOG.md:18:135 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/sdk_packaging.toml:2:51 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/setup.py:15:51 - Unknown word (qnamaker) ```
1.0
cognitiveservices/azure-cognitiveservices-knowledge-qnamaker - cspell found spelling errors in public API surface - Spell check scanning of package at `sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker` detected spelling errors in the public API surface. This directory is opted out of PR spell checking in PR #22199 to keep PRs unblocked. ## What to do 1. Ensure Node.js is installed (https://nodejs.org/en/download/). 1. Delete the entry in `.vscode/cspell.json`'s `ignorePaths` field. It will look like: `sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/**`. You need to do this to enable checking the files. 1. From the root of the repo run spell check using `./eng/common/spelling/Invoke-Cspell.ps1 -ScanGlobs "sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/**"` 1. Fix detections according to http://aka.ms/azsdk/engsys/spellcheck use the "False positives" section to fix false positives 1. Check in changes (including the change to `.vscode/cspell.json` where the `ignorePaths` is updated to remove the entry for this service.). You may need to run `git add -f .vscode/cspell.json` to force adding the changes to the file in git. ## Spell checking output ``` ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/__init__.py:12:10 - Unknown word (amaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:29:14 - Unknown word (adto) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:29:40 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:30:14 - Unknown word (adto) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:30:32 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:33:14 - Unknown word (adocuments) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:54:14 - Unknown word (asearch) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:55:14 - Unknown word (asearch) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:56:14 - Unknown word (asearch) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:57:14 - Unknown word (asearch) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:79:14 - Unknown word (adto) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:79:36 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:80:14 - Unknown word (adto) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:80:28 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:83:14 - Unknown word (adocuments) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/__init__.py:112:10 - Unknown word (amaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/context_dto_py3.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/context_dto.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:22:37 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:25:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:25:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:30:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto_py3.py:62:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:22:37 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:25:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:25:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:30:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_dto.py:62:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:21:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:21:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto_py3.py:30:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:21:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:21:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/create_kb_input_dto.py:30:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/endpoint_settings_dto_py3.py:20:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/endpoint_settings_dto.py:20:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_py3.py:28:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_py3.py:36:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_py3.py:40:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error_py3.py:27:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error_py3.py:35:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error_py3.py:39:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error.py:27:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error.py:35:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_error.py:39:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response_py3.py:22:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error_response.py:22:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error.py:28:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error.py:36:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/error.py:40:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/feedback_records_dto_py3.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/feedback_records_dto.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/inner_error_model_py3.py:26:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/inner_error_model.py:26:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/knowledgebases_dto_py3.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/knowledgebases_dto.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation_py3.py:21:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation_py3.py:36:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation.py:21:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/operation.py:36:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:22:8 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:24:19 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:24:44 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_py3.py:26:51 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:12:10 - Unknown word (adto) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:12:28 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:15:22 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:16:10 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:16:35 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:31:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna_py3.py:34:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:12:10 - Unknown word (adto) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:12:24 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:15:22 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:16:10 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:16:35 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:31:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto_qna.py:34:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:22:8 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:24:19 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:24:44 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/prompt_dto.py:26:51 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:16:18 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:20:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto_py3.py:24:62 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:16:18 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:20:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adocuments_dto.py:24:62 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context_py3.py:15:9 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context_py3.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context_py3.py:39:17 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context.py:15:9 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_context.py:39:17 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:15:9 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:31:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:34:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:34:59 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:52:50 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto_py3.py:57:17 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:15:9 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:31:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:34:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:34:59 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:52:50 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_adto.py:57:17 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_context_py3.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_context.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_list_py3.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_list.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_py3.py:30:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_py3.py:33:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result_py3.py:37:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result.py:30:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result.py:33:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/qn_asearch_result.py:37:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:33:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:40:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:45:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto_py3.py:48:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:33:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:40:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:45:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/query_dto.py:48:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:20:15 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:20:47 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:22:14 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:23:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:27:13 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:31:13 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto_py3.py:31:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:20:15 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:20:47 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:22:14 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:23:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:27:13 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:31:13 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/replace_kb_dto.py:31:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_context_dto_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_context_dto.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_contents_dto_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_contents_dto.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:21:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:21:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add_py3.py:30:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:21:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:21:64 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:26:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_add.py:30:52 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_py3.py:20:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_py3.py:23:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_py3.py:26:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_update_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto_update.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto.py:20:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto.py:23:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_kb_operation_dto.py:26:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto_py3.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_metadata_dto.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_context_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_context.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata_py3.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata_py3.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata.py:20:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_metadata.py:23:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_py3.py:27:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_py3.py:30:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto_py3.py:33:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto.py:27:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto.py:30:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/update_qna_dto.py:33:41 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/word_alterations_dto_py3.py:22:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/models/word_alterations_dto.py:22:46 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:47:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:50:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:93:50 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/alterations_operations.py:102:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:47:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:50:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:100:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_keys_operations.py:103:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:47:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:50:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:93:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/endpoint_settings_operations.py:102:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:47:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:50:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:100:45 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:103:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:231:29 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:236:19 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:236:41 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:239:18 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:240:68 - Unknown word (ADTO) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/knowledgebase_operations.py:251:45 - Unknown word (alist) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/operations.py:48:52 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/operations/operations.py:51:75 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:29:6 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:43:32 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:47:64 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/azure/cognitiveservices/knowledge/qnamaker/qn_amaker_client.py:60:67 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/CHANGELOG.md:14:76 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/CHANGELOG.md:18:87 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/CHANGELOG.md:18:135 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/sdk_packaging.toml:2:51 - Unknown word (qnamaker) ./sdk/cognitiveservices/azure-cognitiveservices-knowledge-qnamaker/setup.py:15:51 - Unknown word (qnamaker) ```
non_test
cognitiveservices azure cognitiveservices knowledge qnamaker cspell found spelling errors in public api surface spell check scanning of package at sdk cognitiveservices azure cognitiveservices knowledge qnamaker detected spelling errors in the public api surface this directory is opted out of pr spell checking in pr to keep prs unblocked what to do ensure node js is installed delete the entry in vscode cspell json s ignorepaths field it will look like sdk cognitiveservices azure cognitiveservices knowledge qnamaker you need to do this to enable checking the files from the root of the repo run spell check using eng common spelling invoke cspell scanglobs sdk cognitiveservices azure cognitiveservices knowledge qnamaker fix detections according to use the false positives section to fix false positives check in changes including the change to vscode cspell json where the ignorepaths is updated to remove the entry for this service you may need to run git add f vscode cspell json to force adding the changes to the file in git spell checking output sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker init py unknown word amaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adocuments sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word asearch sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word asearch sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word asearch sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word asearch sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word adocuments sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models init py unknown word amaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models context dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models context dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models create kb input dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models endpoint settings dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models endpoint settings dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error response py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models error py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models feedback records dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models feedback records dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models inner error model py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models inner error model py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models knowledgebases dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models knowledgebases dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models operation py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models operation py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models operation py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models operation py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto qna py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models prompt dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adocuments dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto context py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto context py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto context py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto context py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto context py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto context py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn adto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result context py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result context py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result list py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result list py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models qn asearch result py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models query dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models replace kb dto py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update context dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update context dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb contents dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb contents dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto add py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto update py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto update py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update kb operation dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update metadata dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update metadata dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update metadata dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update metadata dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto context py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto context py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto metadata py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto metadata py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto metadata py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto metadata py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models update qna dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models word alterations dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker models word alterations dto py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations alterations operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations alterations operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations alterations operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations alterations operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint keys operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint keys operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint keys operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint keys operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint settings operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint settings operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint settings operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations endpoint settings operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word adto sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations knowledgebase operations py unknown word alist sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker operations operations py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker qn amaker client py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker qn amaker client py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker qn amaker client py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker azure cognitiveservices knowledge qnamaker qn amaker client py unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker changelog md unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker changelog md unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker changelog md unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker sdk packaging toml unknown word qnamaker sdk cognitiveservices azure cognitiveservices knowledge qnamaker setup py unknown word qnamaker
0
114,738
11,855,220,791
IssuesEvent
2020-03-25 03:30:09
kotest/kotest
https://api.github.com/repos/kotest/kotest
closed
Example of property-based testing does not compile in most recent version (3.4.2)
documentation
In the main readme, this section of code won't compile with the latest version of the library: ``` class PropertyExample: StringSpec({ "String size" { checkAll<String, String> { a, b -> (a + b) should haveLength(a.length + b.length) } } }) ``` But the version here does work (https://github.com/kotest/kotest/blob/master/doc/property_testing.md): ``` class PropertyExample: StringSpec() { init { "String size" { assertAll({ a: String, b: String -> (a + b).length shouldBe a.length + b.length }) } } } ```
1.0
Example of property-based testing does not compile in most recent version (3.4.2) - In the main readme, this section of code won't compile with the latest version of the library: ``` class PropertyExample: StringSpec({ "String size" { checkAll<String, String> { a, b -> (a + b) should haveLength(a.length + b.length) } } }) ``` But the version here does work (https://github.com/kotest/kotest/blob/master/doc/property_testing.md): ``` class PropertyExample: StringSpec() { init { "String size" { assertAll({ a: String, b: String -> (a + b).length shouldBe a.length + b.length }) } } } ```
non_test
example of property based testing does not compile in most recent version in the main readme this section of code won t compile with the latest version of the library class propertyexample stringspec string size checkall a b a b should havelength a length b length but the version here does work class propertyexample stringspec init string size assertall a string b string a b length shouldbe a length b length
0
120,887
4,796,091,952
IssuesEvent
2016-11-01 05:34:24
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
kubectl apply doesn't work with thirdparty resource
area/app-config-deployment component/kubectl priority/P2 team/CSI-API Machinery SIG team/ux
This is to keep track of known issues of thirdparty resource. cc @brendandburns
1.0
kubectl apply doesn't work with thirdparty resource - This is to keep track of known issues of thirdparty resource. cc @brendandburns
non_test
kubectl apply doesn t work with thirdparty resource this is to keep track of known issues of thirdparty resource cc brendandburns
0
799,757
28,313,502,353
IssuesEvent
2023-04-10 17:29:14
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
Python: 3.6 EOL. Drop 3.6 support?
lang/Python kind/internal cleanup priority/P2
Python 3.6 has been EOLed. Should we drop 3.6 support? Thanks!
1.0
Python: 3.6 EOL. Drop 3.6 support? - Python 3.6 has been EOLed. Should we drop 3.6 support? Thanks!
non_test
python eol drop support python has been eoled should we drop support thanks
0
78,629
7,656,275,666
IssuesEvent
2018-05-10 15:46:30
italia/spid
https://api.github.com/repos/italia/spid
closed
Richiesta aggiornamento metadati - Comune di Carpenedolo
aggiornamento md test metadata
Buongiorno, si richiede per conto del Comune di Carpenedolo l'aggiornamento dei metadati SPID deployati. I nuovi metadati sono disponibili all'endpoint https://sportellotelematico.comune.carpenedolo.bs.it/sites/default/files/spid-metadata/carpenedolo-signed.xml Grazie
1.0
Richiesta aggiornamento metadati - Comune di Carpenedolo - Buongiorno, si richiede per conto del Comune di Carpenedolo l'aggiornamento dei metadati SPID deployati. I nuovi metadati sono disponibili all'endpoint https://sportellotelematico.comune.carpenedolo.bs.it/sites/default/files/spid-metadata/carpenedolo-signed.xml Grazie
test
richiesta aggiornamento metadati comune di carpenedolo buongiorno si richiede per conto del comune di carpenedolo l aggiornamento dei metadati spid deployati i nuovi metadati sono disponibili all endpoint grazie
1
282,808
24,498,024,755
IssuesEvent
2022-10-10 10:21:33
wazuh/wazuh-qa
https://api.github.com/repos/wazuh/wazuh-qa
opened
Ubuntu Linux 22.04 SCA Policy - Update and rework - checks 5 to 5.2.9
team/qa type/qa-testing status/not-tracked
| Target version | Related issue | Related PR | | -------------- | ------------- | ----------------------------------------- | | 4.4.0 | #3391 | https://github.com/wazuh/wazuh/pull/10487 | | Check ID | Check Name | Implemented | Ready for review | QA review | | -------- | --------------------------------------------------------------------------- | ----------- | ---------------- | --------- | | 5 | Access, Authentication and Authorization | | | | | 5.1 | Configure time-based job schedulers | | | | | 5.1.1 | Ensure cron daemon is enabled and running (Automated) | | | | | 5.1.2 | Ensure permissions on /etc/crontab are configured (Automated) | | | | | 5.1.3 | Ensure permissions on /etc/cron.hourly are configured (Automated) | | | | | 5.1.4 | Ensure permissions on /etc/cron.daily are configured (Automated) | | | | | 5.1.5 | Ensure permissions on /etc/cron.weekly are configured (Automated) | | | | | 5.1.6 | Ensure permissions on /etc/cron.monthly are configured (Automated) | | | | | 5.1.7 | Ensure permissions on /etc/cron.d are configured (Automated) | | | | | 5.1.8 | Ensure cron is restricted to authorized users (Automated) | | | | | 5.1.9 | Ensure at is restricted to authorized users (Automated) | | | | | 5.2 | Configure SSH Server | | | | | 5.2.1 | Ensure permissions on /etc/ssh/sshd_config are configured (Automated) | | | | | 5.2.2 | Ensure permissions on SSH private host key files are configured (Automated) | | | | | 5.2.3 | Ensure permissions on SSH public host key files are configured (Automated) | | | | | 5.2.4 | Ensure SSH access is limited (Automated) | | | | | 5.2.5 | Ensure SSH LogLevel is appropriate (Automated) | | | | | 5.2.6 | Ensure SSH PAM is enabled (Automated) | | | | | 5.2.7 | Ensure SSH root login is disabled (Automated) | | | | | 5.2.8 | Ensure SSH HostbasedAuthentication is disabled (Automated) | | | | | 5.2.9 | Ensure SSH PermitEmptyPasswords is disabled (Automated) | | | |
1.0
Ubuntu Linux 22.04 SCA Policy - Update and rework - checks 5 to 5.2.9 - | Target version | Related issue | Related PR | | -------------- | ------------- | ----------------------------------------- | | 4.4.0 | #3391 | https://github.com/wazuh/wazuh/pull/10487 | | Check ID | Check Name | Implemented | Ready for review | QA review | | -------- | --------------------------------------------------------------------------- | ----------- | ---------------- | --------- | | 5 | Access, Authentication and Authorization | | | | | 5.1 | Configure time-based job schedulers | | | | | 5.1.1 | Ensure cron daemon is enabled and running (Automated) | | | | | 5.1.2 | Ensure permissions on /etc/crontab are configured (Automated) | | | | | 5.1.3 | Ensure permissions on /etc/cron.hourly are configured (Automated) | | | | | 5.1.4 | Ensure permissions on /etc/cron.daily are configured (Automated) | | | | | 5.1.5 | Ensure permissions on /etc/cron.weekly are configured (Automated) | | | | | 5.1.6 | Ensure permissions on /etc/cron.monthly are configured (Automated) | | | | | 5.1.7 | Ensure permissions on /etc/cron.d are configured (Automated) | | | | | 5.1.8 | Ensure cron is restricted to authorized users (Automated) | | | | | 5.1.9 | Ensure at is restricted to authorized users (Automated) | | | | | 5.2 | Configure SSH Server | | | | | 5.2.1 | Ensure permissions on /etc/ssh/sshd_config are configured (Automated) | | | | | 5.2.2 | Ensure permissions on SSH private host key files are configured (Automated) | | | | | 5.2.3 | Ensure permissions on SSH public host key files are configured (Automated) | | | | | 5.2.4 | Ensure SSH access is limited (Automated) | | | | | 5.2.5 | Ensure SSH LogLevel is appropriate (Automated) | | | | | 5.2.6 | Ensure SSH PAM is enabled (Automated) | | | | | 5.2.7 | Ensure SSH root login is disabled (Automated) | | | | | 5.2.8 | Ensure SSH HostbasedAuthentication is disabled (Automated) | | | | | 5.2.9 | Ensure SSH PermitEmptyPasswords is disabled (Automated) | | | |
test
ubuntu linux sca policy update and rework checks to target version related issue related pr check id check name implemented ready for review qa review access authentication and authorization configure time based job schedulers ensure cron daemon is enabled and running automated ensure permissions on etc crontab are configured automated ensure permissions on etc cron hourly are configured automated ensure permissions on etc cron daily are configured automated ensure permissions on etc cron weekly are configured automated ensure permissions on etc cron monthly are configured automated ensure permissions on etc cron d are configured automated ensure cron is restricted to authorized users automated ensure at is restricted to authorized users automated configure ssh server ensure permissions on etc ssh sshd config are configured automated ensure permissions on ssh private host key files are configured automated ensure permissions on ssh public host key files are configured automated ensure ssh access is limited automated ensure ssh loglevel is appropriate automated ensure ssh pam is enabled automated ensure ssh root login is disabled automated ensure ssh hostbasedauthentication is disabled automated ensure ssh permitemptypasswords is disabled automated
1
171,685
13,244,139,480
IssuesEvent
2020-08-19 12:36:51
MarcoConsiglio/shopping-list
https://api.github.com/repos/MarcoConsiglio/shopping-list
opened
Segnare i prodotti già presi
browser-test miglioria estetica
Quando un prodotto ha valorizzata la quantità nel carrello, allora segna il prodotto in modo tale da far capire che è stato già preso.
1.0
Segnare i prodotti già presi - Quando un prodotto ha valorizzata la quantità nel carrello, allora segna il prodotto in modo tale da far capire che è stato già preso.
test
segnare i prodotti già presi quando un prodotto ha valorizzata la quantità nel carrello allora segna il prodotto in modo tale da far capire che è stato già preso
1
302,369
26,140,936,855
IssuesEvent
2022-12-29 18:20:05
Hamlib/Hamlib
https://api.github.com/repos/Hamlib/Hamlib
closed
QRP-Labs QDX IF command parsing buggy
bug needs test
I'm using a QRP-Labs QDX which emulates TS-440. When requesting PTT state from rigctl (sending command 't') it returns 0 (RX) even when still transmitting. It appears that the 440 protocol returns accurate PTT state from 'P8' of the 'IF' command as per the spec. I'm going to suggest that hamlib reads PTT state from this, rather than return 0 as it does now. Discussion here: https://groups.io/g/QRPLabs/topic/95922381?p=Created,,,20,1,0,0::recentpostdate/sticky,,,20,2,0,95922381,previd=1672274690459582723,nextid=1672093401297648786 I might be wrong but this is as far as I've gotten in terms of debug.
1.0
QRP-Labs QDX IF command parsing buggy - I'm using a QRP-Labs QDX which emulates TS-440. When requesting PTT state from rigctl (sending command 't') it returns 0 (RX) even when still transmitting. It appears that the 440 protocol returns accurate PTT state from 'P8' of the 'IF' command as per the spec. I'm going to suggest that hamlib reads PTT state from this, rather than return 0 as it does now. Discussion here: https://groups.io/g/QRPLabs/topic/95922381?p=Created,,,20,1,0,0::recentpostdate/sticky,,,20,2,0,95922381,previd=1672274690459582723,nextid=1672093401297648786 I might be wrong but this is as far as I've gotten in terms of debug.
test
qrp labs qdx if command parsing buggy i m using a qrp labs qdx which emulates ts when requesting ptt state from rigctl sending command t it returns rx even when still transmitting it appears that the protocol returns accurate ptt state from of the if command as per the spec i m going to suggest that hamlib reads ptt state from this rather than return as it does now discussion here i might be wrong but this is as far as i ve gotten in terms of debug
1
60,385
6,689,743,960
IssuesEvent
2017-10-09 05:02:26
Kademi/kademi-dev
https://api.github.com/repos/Kademi/kademi-dev
closed
Improve UI of switch repo
Ready to Test QA
Instead of showing in dropdown list, we can show in combo with chosen plugin
1.0
Improve UI of switch repo - Instead of showing in dropdown list, we can show in combo with chosen plugin
test
improve ui of switch repo instead of showing in dropdown list we can show in combo with chosen plugin
1
99,122
16,430,784,870
IssuesEvent
2021-05-20 01:03:37
DavidSpek/pipelines
https://api.github.com/repos/DavidSpek/pipelines
opened
CVE-2021-29515 (Low) detected in tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl
security vulnerability
## CVE-2021-29515 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt</p> <p>Path to vulnerable library: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt,pipelines/samples/core/ai_platform/training</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. The implementation of `MatrixDiag*` operations(https://github.com/tensorflow/tensorflow/blob/4c4f420e68f1cfaf8f4b6e8e3eb857e9e4c3ff33/tensorflow/core/kernels/linalg/matrix_diag_op.cc#L195-L197) does not validate that the tensor arguments are non-empty. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range. <p>Publish Date: 2021-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29515>CVE-2021-29515</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-hc6c-75p4-hmq4">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-hc6c-75p4-hmq4</a></p> <p>Release Date: 2021-05-14</p> <p>Fix Resolution: tensorflow - 2.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-29515 (Low) detected in tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl - ## CVE-2021-29515 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt</p> <p>Path to vulnerable library: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt,pipelines/samples/core/ai_platform/training</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. The implementation of `MatrixDiag*` operations(https://github.com/tensorflow/tensorflow/blob/4c4f420e68f1cfaf8f4b6e8e3eb857e9e4c3ff33/tensorflow/core/kernels/linalg/matrix_diag_op.cc#L195-L197) does not validate that the tensor arguments are non-empty. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range. <p>Publish Date: 2021-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29515>CVE-2021-29515</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-hc6c-75p4-hmq4">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-hc6c-75p4-hmq4</a></p> <p>Release Date: 2021-05-14</p> <p>Fix Resolution: tensorflow - 2.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve low detected in tensorflow whl cve low severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file pipelines contrib components openvino ovms deployer containers requirements txt path to vulnerable library pipelines contrib components openvino ovms deployer containers requirements txt pipelines samples core ai platform training dependency hierarchy x tensorflow whl vulnerable library found in base branch master vulnerability details tensorflow is an end to end open source platform for machine learning the implementation of matrixdiag operations does not validate that the tensor arguments are non empty the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow step up your open source security game with whitesource
0
159,423
12,475,129,603
IssuesEvent
2020-05-29 10:57:53
tactilenews/100eyes
https://api.github.com/repos/tactilenews/100eyes
opened
Revise function "Hinweise anhängen"
Ergebnis Usertest
As an editor, the "attach notes"("Hinweise anhängen") function bothers me because it is very prominent, I don't understand it and it keeps me from asking a question.
1.0
Revise function "Hinweise anhängen" - As an editor, the "attach notes"("Hinweise anhängen") function bothers me because it is very prominent, I don't understand it and it keeps me from asking a question.
test
revise function hinweise anhängen as an editor the attach notes hinweise anhängen function bothers me because it is very prominent i don t understand it and it keeps me from asking a question
1
62,268
6,792,565,838
IssuesEvent
2017-11-01 01:17:16
googlefonts/fontbakery
https://api.github.com/repos/googlefonts/fontbakery
closed
Upstream UFO Check: Validate GLIF
testing
- [ ] Add an upstream UFO check that uses https://github.com/davelab6/glif-schema and http://lxml.de/validation.html to validate GLIF files
1.0
Upstream UFO Check: Validate GLIF - - [ ] Add an upstream UFO check that uses https://github.com/davelab6/glif-schema and http://lxml.de/validation.html to validate GLIF files
test
upstream ufo check validate glif add an upstream ufo check that uses and to validate glif files
1
52,440
12,965,507,827
IssuesEvent
2020-07-20 22:30:58
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
nixos.iso_minimal has a nondeterministic disk guid
0.kind: bug 6.topic: reproducible builds
**Describe the bug** When building `nixos.iso_minimal.x86_64-linux` twice, the GPT partition has a different GUID. **To Reproduce** Steps to reproduce the behavior: 1. `nix-build ./nixos/release-combined.nix -A nixos.iso_minimal.x86_64-linux` 2. `nix-build ./nixos/release-combined.nix -A nixos.iso_minimal.x86_64-linux --check --keep-failed` 3. use `gdisk` to see the GUID is not consistent (or `diffoscope` at `0x00000238`, https://en.wikipedia.org/wiki/GUID_Partition_Table) **Expected behavior** A deterministic GUID **Additional context** For additional background on why it's helpful for builds to be reproducible, see https://reproducible-builds.org/ I'd be happy to help hunt down and fix this problem further, but this is rather new territory for me, so I'd appreciate any help
1.0
nixos.iso_minimal has a nondeterministic disk guid - **Describe the bug** When building `nixos.iso_minimal.x86_64-linux` twice, the GPT partition has a different GUID. **To Reproduce** Steps to reproduce the behavior: 1. `nix-build ./nixos/release-combined.nix -A nixos.iso_minimal.x86_64-linux` 2. `nix-build ./nixos/release-combined.nix -A nixos.iso_minimal.x86_64-linux --check --keep-failed` 3. use `gdisk` to see the GUID is not consistent (or `diffoscope` at `0x00000238`, https://en.wikipedia.org/wiki/GUID_Partition_Table) **Expected behavior** A deterministic GUID **Additional context** For additional background on why it's helpful for builds to be reproducible, see https://reproducible-builds.org/ I'd be happy to help hunt down and fix this problem further, but this is rather new territory for me, so I'd appreciate any help
non_test
nixos iso minimal has a nondeterministic disk guid describe the bug when building nixos iso minimal linux twice the gpt partition has a different guid to reproduce steps to reproduce the behavior nix build nixos release combined nix a nixos iso minimal linux nix build nixos release combined nix a nixos iso minimal linux check keep failed use gdisk to see the guid is not consistent or diffoscope at expected behavior a deterministic guid additional context for additional background on why it s helpful for builds to be reproducible see i d be happy to help hunt down and fix this problem further but this is rather new territory for me so i d appreciate any help
0