Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
14,352
17,375,052,868
IssuesEvent
2021-07-30 19:37:16
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Reference to Obsolete Version of Node.js
Pri2 devops-cicd-process/tech devops/prod doc-enhancement
This section specifies Node.js version 6, which is obsolete: https://docs.microsoft.com/en-gb/azure/devops/pipelines/process/container-phases?view=azure-devops#bring-your-own-nodejs The text reads: "Node 6 is a safe choice. You can start from the node:6-alpine image." --- #### Document details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3339a2e0-be29-1363-f588-b231d4472c02 * Version Independent ID: 72dd11a3-704d-d0fd-6dfa-cf49f3352de3 * Content: [Container Jobs in Azure Pipelines and TFS - Azure Pipelines](https://docs.microsoft.com/en-gb/azure/devops/pipelines/process/container-phases?view=azure-devops) * Content Source: [docs/pipelines/process/container-phases.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/container-phases.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Reference to Obsolete Version of Node.js - This section specifies Node.js version 6, which is obsolete: https://docs.microsoft.com/en-gb/azure/devops/pipelines/process/container-phases?view=azure-devops#bring-your-own-nodejs The text reads: "Node 6 is a safe choice. You can start from the node:6-alpine image." --- #### Document details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3339a2e0-be29-1363-f588-b231d4472c02 * Version Independent ID: 72dd11a3-704d-d0fd-6dfa-cf49f3352de3 * Content: [Container Jobs in Azure Pipelines and TFS - Azure Pipelines](https://docs.microsoft.com/en-gb/azure/devops/pipelines/process/container-phases?view=azure-devops) * Content Source: [docs/pipelines/process/container-phases.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/container-phases.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
reference to obsolete version of node js this section specifies node js version which is obsolete the text reads node is a safe choice you can start from the node alpine image document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
21,789
30,297,622,262
IssuesEvent
2023-07-10 01:18:05
winter-telescope/mirar
https://api.github.com/repos/winter-telescope/mirar
opened
[BUG] Hardcoded assumption of split image
bug processors
**Describe the bug** Some processors have undeclared requirements that they only work if an image has been split, e.g: https://github.com/winter-telescope/mirar/blob/93ecacddd58f7814a30d045df4e53325aa7da570/mirar/references/ukirt.py#L655 This is super unintuitive behaviour, and breaks the golden rule that processors cannot expect a key unless it is listed in https://github.com/winter-telescope/mirar/blob/93ecacddd58f7814a30d045df4e53325aa7da570/mirar/paths.py#L291, or it is listed as a prerequisite requirement in the processors. Further, these processors do not really need the images to be split at all, the two things are independent. They just need to refer uniquely to a field. What should we do? I can think of: - rewrite the processors to be agnostic, - make the subcoordid required - Abolish subcoordid, and start modifying the required key `FIELDID` in split Alternative solutions welcome.
1.0
[BUG] Hardcoded assumption of split image - **Describe the bug** Some processors have undeclared requirements that they only work if an image has been split, e.g: https://github.com/winter-telescope/mirar/blob/93ecacddd58f7814a30d045df4e53325aa7da570/mirar/references/ukirt.py#L655 This is super unintuitive behaviour, and breaks the golden rule that processors cannot expect a key unless it is listed in https://github.com/winter-telescope/mirar/blob/93ecacddd58f7814a30d045df4e53325aa7da570/mirar/paths.py#L291, or it is listed as a prerequisite requirement in the processors. Further, these processors do not really need the images to be split at all, the two things are independent. They just need to refer uniquely to a field. What should we do? I can think of: - rewrite the processors to be agnostic, - make the subcoordid required - Abolish subcoordid, and start modifying the required key `FIELDID` in split Alternative solutions welcome.
process
hardcoded assumption of split image describe the bug some processors have undeclared requirements that they only work if an image has been split e g this is super unintuitive behaviour and breaks the golden rule that processors cannot expect a key unless it is listed in or it is listed as a prerequisite requirement in the processors further these processors do not really need the images to be split at all the two things are independent they just need to refer uniquely to a field what should we do i can think of rewrite the processors to be agnostic make the subcoordid required abolish subcoordid and start modifying the required key fieldid in split alternative solutions welcome
1
20,564
27,224,993,180
IssuesEvent
2023-02-21 09:05:25
redhat-developer/vscode-java
https://api.github.com/repos/redhat-developer/vscode-java
closed
Failed build when using records with mapstruct
bug upstream annotation-processing
##### Environment - Operating System: MacOS Ventura 13.1 - JDK version: openjdk-19 - Visual Studio Code version: 1.74.3 - Java extension version: 1.14.0 ##### Steps To Reproduce 1. Clone https://github.com/mapstruct/mapstruct-examples 2. Open mapstruct-record in Visual Studio Code 3. Rename source = "email" to source = "name" in CustomerMapper.java ##### Additional Informations I'm getting strange a error while using mapstruct with records: `No property named "name" exists in source parameter(s). Type "CustomerDto" has no properties.Java(0)` The error disappears after running "Java: rebuild project". This error is not a new one, I saw it with older versions of the extension too. When I try to run or debug the project a warning for "Build was failed" appears
1.0
Failed build when using records with mapstruct - ##### Environment - Operating System: MacOS Ventura 13.1 - JDK version: openjdk-19 - Visual Studio Code version: 1.74.3 - Java extension version: 1.14.0 ##### Steps To Reproduce 1. Clone https://github.com/mapstruct/mapstruct-examples 2. Open mapstruct-record in Visual Studio Code 3. Rename source = "email" to source = "name" in CustomerMapper.java ##### Additional Informations I'm getting strange a error while using mapstruct with records: `No property named "name" exists in source parameter(s). Type "CustomerDto" has no properties.Java(0)` The error disappears after running "Java: rebuild project". This error is not a new one, I saw it with older versions of the extension too. When I try to run or debug the project a warning for "Build was failed" appears
process
failed build when using records with mapstruct environment operating system macos ventura jdk version openjdk visual studio code version java extension version steps to reproduce clone open mapstruct record in visual studio code rename source email to source name in customermapper java additional informations i m getting strange a error while using mapstruct with records no property named name exists in source parameter s type customerdto has no properties java the error disappears after running java rebuild project this error is not a new one i saw it with older versions of the extension too when i try to run or debug the project a warning for build was failed appears
1
20,870
27,657,188,695
IssuesEvent
2023-03-12 04:21:41
pentium3/sys_reading
https://api.github.com/repos/pentium3/sys_reading
opened
CrystalPerf: Learning to Characterize the Performance of Dataflow Computation through Code Analysis
stream processing
https://www.usenix.org/system/files/atc21-tian.pdf
1.0
CrystalPerf: Learning to Characterize the Performance of Dataflow Computation through Code Analysis - https://www.usenix.org/system/files/atc21-tian.pdf
process
crystalperf learning to characterize the performance of dataflow computation through code analysis
1
10,517
13,299,248,606
IssuesEvent
2020-08-25 09:27:13
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
Fuzz test failed for tikv::coprocessor::codec::mysql::Duration::parse
sig/coprocessor type/bug
## Bug Report **What version of TiKV are you using?** 3.0.0-beta.1 **What operating system and CPU are you using?** macOS 10.14.4 **What did you do?** The fuzz test added by WIP PR #4608 failed. This is the test: ```rust pub fn fuzz_coprocessor_codec_duration_from_parse(data: &[u8]) -> Result<(), Error> { use std::io::Read; use tikv::coprocessor::codec::mysql::Duration; let mut cursor = Cursor::new(data); let fsp = cursor.read_as_i8()?; let mut buf: [u8; 32] = [b' '; 32]; cursor.read_exact(&mut buf)?; let d = Duration::parse(&buf, fsp)?; fuzz_duration(d, cursor) } ``` This is the error report: ``` thread '<unnamed>' panicked at 'attempt to multiply with overflow', src/coprocessor/codec/mysql/duration.rs:191:21 stack backtrace: 0: std::sys::unix::backtrace::tracing::imp::unwind_backtrace 1: std::sys_common::backtrace::_print 2: std::panicking::default_hook::{{closure}} 3: std::panicking::default_hook 4: std::panicking::rust_panic_with_hook 5: std::panicking::continue_panic_fmt 6: rust_begin_unwind 7: core::panicking::panic_fmt 8: core::panicking::panic 9: tikv::coprocessor::codec::mysql::duration::Duration::parse 10: fuzz_targets::fuzz_coprocessor_codec_duration_from_parse 11: rust_fuzzer_test_input 12: __rust_maybe_catch_panic 13: std::panicking::try 14: LLVMFuzzerTestOneInput 15: _ZN6fuzzer6Fuzzer15ExecuteCallbackEPKhm 16: _ZN6fuzzer6Fuzzer6RunOneEPKhmbPNS_9InputInfoEPb 17: _ZN6fuzzer6Fuzzer16MutateAndTestOneEv 18: _ZN6fuzzer6Fuzzer4LoopERKNSt3__16vectorINS1_12basic_stringIcNS1_11char_traitsIcEENS1_9allocatorIcEEEENS_16fuzzer_allocatorIS8_EEEE 19: _ZN6fuzzer12FuzzerDriverEPiPPPcPFiPKhmE 20: main ==56552== ERROR: libFuzzer: deadly signal #0 0x1166d1707 in __sanitizer_print_stack_trace (lib__rustc__clang_rt.asan_osx_dynamic.dylib:x86_64+0x68707) #1 0x110bd583b in fuzzer::Fuzzer::CrashCallback() (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b7383b) #2 0x110bd57ed in fuzzer::Fuzzer::StaticCrashSignalCallback() (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b737ed) #3 0x110c16437 in fuzzer::CrashHandler(int, __siginfo*, void*) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105bb4437) #4 0x7fff590abb5c in _sigtramp (libsystem_platform.dylib:x86_64+0x4b5c) #5 0x3500000bd0 (<unknown module>) #6 0x110c4a7ee in __rust_maybe_catch_panic (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105be87ee) #7 0x110bd3fde in std::panicking::try::hd4964c260fccd748 (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b71fde) #8 0x110bd3a72 in LLVMFuzzerTestOneInput (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b71a72) #9 0x110bd6f9a in fuzzer::Fuzzer::ExecuteCallback(unsigned char const*, unsigned long) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b74f9a) #10 0x110bd6999 in fuzzer::Fuzzer::RunOne(unsigned char const*, unsigned long, bool, fuzzer::InputInfo*, bool*) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b74999) #11 0x110bd85c1 in fuzzer::Fuzzer::MutateAndTestOne() (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b765c1) #12 0x110bd9871 in fuzzer::Fuzzer::Loop(std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, fuzzer::fuzzer_allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > > > const&) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b77871) #13 0x110c04ab8 in fuzzer::FuzzerDriver(int*, char***, int (*)(unsigned char const*, unsigned long)) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105ba2ab8) #14 0x110c27179 in main (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105bc5179) #15 0x7fff58ec63d4 in start (libdyld.dylib:x86_64+0x163d4) NOTE: libFuzzer has rudimentary signal handlers. Combine libFuzzer with AddressSanitizer or similar for better crash reports. SUMMARY: libFuzzer: deadly signal MS: 2 ChangeBit-CopyPart-; base unit: 4744b705bce806db21998c6b5c000865f87bf51a 0xff,0x35,0x35,0x35,0x35,0x35,0x35,0x35,0x35,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x20,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x32,0x33,0x35,0x35,0x35,0x0,0x1, \xff555555550000000 0000000000023555\x00\x01 artifact_prefix='./'; Test unit written to ./crash-eda52eca653e470cf2fed7ad070b4f09e122dab9 Base64: /zU1NTU1NTU1MDAwMDAwMCAwMDAwMDAwMDAwMDIzNTU1AAE= Running fuzzer failed: Libfuzzer exited with code Some(77) Error: 1 ```
1.0
Fuzz test failed for tikv::coprocessor::codec::mysql::Duration::parse - ## Bug Report **What version of TiKV are you using?** 3.0.0-beta.1 **What operating system and CPU are you using?** macOS 10.14.4 **What did you do?** The fuzz test added by WIP PR #4608 failed. This is the test: ```rust pub fn fuzz_coprocessor_codec_duration_from_parse(data: &[u8]) -> Result<(), Error> { use std::io::Read; use tikv::coprocessor::codec::mysql::Duration; let mut cursor = Cursor::new(data); let fsp = cursor.read_as_i8()?; let mut buf: [u8; 32] = [b' '; 32]; cursor.read_exact(&mut buf)?; let d = Duration::parse(&buf, fsp)?; fuzz_duration(d, cursor) } ``` This is the error report: ``` thread '<unnamed>' panicked at 'attempt to multiply with overflow', src/coprocessor/codec/mysql/duration.rs:191:21 stack backtrace: 0: std::sys::unix::backtrace::tracing::imp::unwind_backtrace 1: std::sys_common::backtrace::_print 2: std::panicking::default_hook::{{closure}} 3: std::panicking::default_hook 4: std::panicking::rust_panic_with_hook 5: std::panicking::continue_panic_fmt 6: rust_begin_unwind 7: core::panicking::panic_fmt 8: core::panicking::panic 9: tikv::coprocessor::codec::mysql::duration::Duration::parse 10: fuzz_targets::fuzz_coprocessor_codec_duration_from_parse 11: rust_fuzzer_test_input 12: __rust_maybe_catch_panic 13: std::panicking::try 14: LLVMFuzzerTestOneInput 15: _ZN6fuzzer6Fuzzer15ExecuteCallbackEPKhm 16: _ZN6fuzzer6Fuzzer6RunOneEPKhmbPNS_9InputInfoEPb 17: _ZN6fuzzer6Fuzzer16MutateAndTestOneEv 18: _ZN6fuzzer6Fuzzer4LoopERKNSt3__16vectorINS1_12basic_stringIcNS1_11char_traitsIcEENS1_9allocatorIcEEEENS_16fuzzer_allocatorIS8_EEEE 19: _ZN6fuzzer12FuzzerDriverEPiPPPcPFiPKhmE 20: main ==56552== ERROR: libFuzzer: deadly signal #0 0x1166d1707 in __sanitizer_print_stack_trace (lib__rustc__clang_rt.asan_osx_dynamic.dylib:x86_64+0x68707) #1 0x110bd583b in fuzzer::Fuzzer::CrashCallback() (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b7383b) #2 0x110bd57ed in fuzzer::Fuzzer::StaticCrashSignalCallback() (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b737ed) #3 0x110c16437 in fuzzer::CrashHandler(int, __siginfo*, void*) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105bb4437) #4 0x7fff590abb5c in _sigtramp (libsystem_platform.dylib:x86_64+0x4b5c) #5 0x3500000bd0 (<unknown module>) #6 0x110c4a7ee in __rust_maybe_catch_panic (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105be87ee) #7 0x110bd3fde in std::panicking::try::hd4964c260fccd748 (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b71fde) #8 0x110bd3a72 in LLVMFuzzerTestOneInput (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b71a72) #9 0x110bd6f9a in fuzzer::Fuzzer::ExecuteCallback(unsigned char const*, unsigned long) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b74f9a) #10 0x110bd6999 in fuzzer::Fuzzer::RunOne(unsigned char const*, unsigned long, bool, fuzzer::InputInfo*, bool*) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b74999) #11 0x110bd85c1 in fuzzer::Fuzzer::MutateAndTestOne() (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b765c1) #12 0x110bd9871 in fuzzer::Fuzzer::Loop(std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, fuzzer::fuzzer_allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > > > const&) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b77871) #13 0x110c04ab8 in fuzzer::FuzzerDriver(int*, char***, int (*)(unsigned char const*, unsigned long)) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105ba2ab8) #14 0x110c27179 in main (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105bc5179) #15 0x7fff58ec63d4 in start (libdyld.dylib:x86_64+0x163d4) NOTE: libFuzzer has rudimentary signal handlers. Combine libFuzzer with AddressSanitizer or similar for better crash reports. SUMMARY: libFuzzer: deadly signal MS: 2 ChangeBit-CopyPart-; base unit: 4744b705bce806db21998c6b5c000865f87bf51a 0xff,0x35,0x35,0x35,0x35,0x35,0x35,0x35,0x35,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x20,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x32,0x33,0x35,0x35,0x35,0x0,0x1, \xff555555550000000 0000000000023555\x00\x01 artifact_prefix='./'; Test unit written to ./crash-eda52eca653e470cf2fed7ad070b4f09e122dab9 Base64: /zU1NTU1NTU1MDAwMDAwMCAwMDAwMDAwMDAwMDIzNTU1AAE= Running fuzzer failed: Libfuzzer exited with code Some(77) Error: 1 ```
process
fuzz test failed for tikv coprocessor codec mysql duration parse bug report what version of tikv are you using beta what operating system and cpu are you using macos what did you do the fuzz test added by wip pr failed this is the test rust pub fn fuzz coprocessor codec duration from parse data result use std io read use tikv coprocessor codec mysql duration let mut cursor cursor new data let fsp cursor read as let mut buf cursor read exact mut buf let d duration parse buf fsp fuzz duration d cursor this is the error report thread panicked at attempt to multiply with overflow src coprocessor codec mysql duration rs stack backtrace std sys unix backtrace tracing imp unwind backtrace std sys common backtrace print std panicking default hook closure std panicking default hook std panicking rust panic with hook std panicking continue panic fmt rust begin unwind core panicking panic fmt core panicking panic tikv coprocessor codec mysql duration duration parse fuzz targets fuzz coprocessor codec duration from parse rust fuzzer test input rust maybe catch panic std panicking try llvmfuzzertestoneinput eeee main error libfuzzer deadly signal in sanitizer print stack trace lib rustc clang rt asan osx dynamic dylib in fuzzer fuzzer crashcallback fuzz coprocessor codec duration from parse in fuzzer fuzzer staticcrashsignalcallback fuzz coprocessor codec duration from parse in fuzzer crashhandler int siginfo void fuzz coprocessor codec duration from parse in sigtramp libsystem platform dylib in rust maybe catch panic fuzz coprocessor codec duration from parse in std panicking try fuzz coprocessor codec duration from parse in llvmfuzzertestoneinput fuzz coprocessor codec duration from parse in fuzzer fuzzer executecallback unsigned char const unsigned long fuzz coprocessor codec duration from parse in fuzzer fuzzer runone unsigned char const unsigned long bool fuzzer inputinfo bool fuzz coprocessor codec duration from parse in fuzzer fuzzer mutateandtestone fuzz coprocessor codec duration from parse in fuzzer fuzzer loop std vector std allocator fuzzer fuzzer allocator std allocator const fuzz coprocessor codec duration from parse in fuzzer fuzzerdriver int char int unsigned char const unsigned long fuzz coprocessor codec duration from parse in main fuzz coprocessor codec duration from parse in start libdyld dylib note libfuzzer has rudimentary signal handlers combine libfuzzer with addresssanitizer or similar for better crash reports summary libfuzzer deadly signal ms changebit copypart base unit artifact prefix test unit written to crash running fuzzer failed libfuzzer exited with code some error
1
36,251
6,521,838,234
IssuesEvent
2017-08-28 22:22:43
NAVADMC/ADSM
https://api.github.com/repos/NAVADMC/ADSM
closed
Fix Unhelpful help text
Documentation
While working on #245 I've noticed places with unhelpful help text: - Destruction capacity - The relational function used to define the daily destruction capacity. - X and Y axis labels would be more useful - Units detected before triggering vaccination - The number of clinical Units which must be detected before the initiation of a vaccination program. - I think this prevents any vaccinations from happening before the "vaccination program". Does it start up instantly? - Vaccination capacity - Relational function used to define the daily vaccination capacity. - What variable/events does this actually affect in the simulation? - Test Delay - Function that describes the delay in obtaining test results. - This is a PDF... so I may be able to deduce the Y axis is a percentage and that this is a distribution of possible values for "delay". But I have to pause and really think about that because the help text says nothing about the "distribution of delays" - Zone direct movement - Function the describes direct movement rate. (and indirect) - It's not immediately obvious that this is our "Quarantine" option as "Quarantine" does not appear anywhere in the program. Some thing like "Models imperfect Quarantine as a percentage of normal Direct Movement in days since the Zone was first established." Is there a "Documentation Person" I can assign these to?
1.0
Fix Unhelpful help text - While working on #245 I've noticed places with unhelpful help text: - Destruction capacity - The relational function used to define the daily destruction capacity. - X and Y axis labels would be more useful - Units detected before triggering vaccination - The number of clinical Units which must be detected before the initiation of a vaccination program. - I think this prevents any vaccinations from happening before the "vaccination program". Does it start up instantly? - Vaccination capacity - Relational function used to define the daily vaccination capacity. - What variable/events does this actually affect in the simulation? - Test Delay - Function that describes the delay in obtaining test results. - This is a PDF... so I may be able to deduce the Y axis is a percentage and that this is a distribution of possible values for "delay". But I have to pause and really think about that because the help text says nothing about the "distribution of delays" - Zone direct movement - Function the describes direct movement rate. (and indirect) - It's not immediately obvious that this is our "Quarantine" option as "Quarantine" does not appear anywhere in the program. Some thing like "Models imperfect Quarantine as a percentage of normal Direct Movement in days since the Zone was first established." Is there a "Documentation Person" I can assign these to?
non_process
fix unhelpful help text while working on i ve noticed places with unhelpful help text destruction capacity the relational function used to define the daily destruction capacity x and y axis labels would be more useful units detected before triggering vaccination the number of clinical units which must be detected before the initiation of a vaccination program i think this prevents any vaccinations from happening before the vaccination program does it start up instantly vaccination capacity relational function used to define the daily vaccination capacity what variable events does this actually affect in the simulation test delay function that describes the delay in obtaining test results this is a pdf so i may be able to deduce the y axis is a percentage and that this is a distribution of possible values for delay but i have to pause and really think about that because the help text says nothing about the distribution of delays zone direct movement function the describes direct movement rate and indirect it s not immediately obvious that this is our quarantine option as quarantine does not appear anywhere in the program some thing like models imperfect quarantine as a percentage of normal direct movement in days since the zone was first established is there a documentation person i can assign these to
0
67,672
13,012,210,798
IssuesEvent
2020-07-25 04:00:41
jcodwell/MentalHealthRPG
https://api.github.com/repos/jcodwell/MentalHealthRPG
opened
Health Bar Functionalities Implementation
New Feature - Code/Unity
Code and develop a script that can be used as a player character or enemy health bar. This code will be complete if the health bar can - [ ] Move the graphics bar up and down based on value. - [ ] Be able to attach to a game object (such as a player/NPC or enemy) - [ ] Be able to interact with the hitbox by going down when collision is detected.
1.0
Health Bar Functionalities Implementation - Code and develop a script that can be used as a player character or enemy health bar. This code will be complete if the health bar can - [ ] Move the graphics bar up and down based on value. - [ ] Be able to attach to a game object (such as a player/NPC or enemy) - [ ] Be able to interact with the hitbox by going down when collision is detected.
non_process
health bar functionalities implementation code and develop a script that can be used as a player character or enemy health bar this code will be complete if the health bar can move the graphics bar up and down based on value be able to attach to a game object such as a player npc or enemy be able to interact with the hitbox by going down when collision is detected
0
368,252
25,784,145,913
IssuesEvent
2022-12-09 18:39:14
vector-im/matrix-bot-sdk
https://api.github.com/repos/vector-im/matrix-bot-sdk
opened
Build & host the web documentation
documentation
Instead of linking to the documentation pages for the upstream project at https://turt2live.github.io/matrix-bot-sdk/index.html, this fork should have its own documentation hosted on a different domain & with any downstream changes included.
1.0
Build & host the web documentation - Instead of linking to the documentation pages for the upstream project at https://turt2live.github.io/matrix-bot-sdk/index.html, this fork should have its own documentation hosted on a different domain & with any downstream changes included.
non_process
build host the web documentation instead of linking to the documentation pages for the upstream project at this fork should have its own documentation hosted on a different domain with any downstream changes included
0
661,871
22,093,062,074
IssuesEvent
2022-06-01 07:48:53
eclipse/dirigible
https://api.github.com/repos/eclipse/dirigible
opened
[API] Copy API should return useful data.
enhancement API web-ide priority-medium efforts-medium
**Describe the bug** Currently, the copy API, does not return any useful data. In order to properly update the file tree in the front-end, without reloading the entire structure, we need the new workspace path and git status in the response. The contentType would also be nice to have. If the copied object is a project, the response should also include the git boolean and gitName fields. Optionally, it would be good if the workspace name is not included in the file path but is included as another key in the response. **Desktop:** - OS: macOS 12.4 - Browser: Firefox 100 - Version: Dirigible 6.2.29
1.0
[API] Copy API should return useful data. - **Describe the bug** Currently, the copy API, does not return any useful data. In order to properly update the file tree in the front-end, without reloading the entire structure, we need the new workspace path and git status in the response. The contentType would also be nice to have. If the copied object is a project, the response should also include the git boolean and gitName fields. Optionally, it would be good if the workspace name is not included in the file path but is included as another key in the response. **Desktop:** - OS: macOS 12.4 - Browser: Firefox 100 - Version: Dirigible 6.2.29
non_process
copy api should return useful data describe the bug currently the copy api does not return any useful data in order to properly update the file tree in the front end without reloading the entire structure we need the new workspace path and git status in the response the contenttype would also be nice to have if the copied object is a project the response should also include the git boolean and gitname fields optionally it would be good if the workspace name is not included in the file path but is included as another key in the response desktop os macos browser firefox version dirigible
0
10,988
13,785,603,290
IssuesEvent
2020-10-08 23:19:55
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
TIN creation crash
Bug Crash/Data Corruption Feedback Processing
Crash ID: 185de4a9a11eb29f05672d8b176e731c0d36ee2f ``` Stack Trace MathUtils::lineIntersection : DualEdgeTriangulation::insertForcedSegment : DualEdgeTriangulation::addLine : QgsTinInterpolator::insertData : QgsTinInterpolator::initialize : QgsTinInterpolator::interpolatePoint : PyInit__analysis : QgsGridFileWriter::writeFile : PyInit__analysis : PyMethodDef_RawFastCallKeywords : PyMethodDef_RawFastCallKeywords : PyEval_EvalFrameDefault : PyFunction_FastCallDict : PyMethodDef_RawFastCallDict : PyObject_Call : PyInit_sip : CPLStringList::empty : PyInit__core : QgsProcessingAlgorithm::runPrepared : QgsProcessingAlgRunnerTask::run : PyInit__core : QgsTask::start : QThreadPoolPrivate::reset : QThread::start : BaseThreadInitThunk : RtlUserThreadStart : ``` QGIS Info QGIS Version: 3.8.0-Zanzibar QGIS code revision: 11aff65f10 Compiled against Qt: 5.11.2 Running against Qt: 5.11.2 Compiled against GDAL: 2.4.1 Running against GDAL: 2.4.1 System Info CPU Type: x86_64 Kernel Type: winnt Kernel Version: 10.0.17763
1.0
TIN creation crash - Crash ID: 185de4a9a11eb29f05672d8b176e731c0d36ee2f ``` Stack Trace MathUtils::lineIntersection : DualEdgeTriangulation::insertForcedSegment : DualEdgeTriangulation::addLine : QgsTinInterpolator::insertData : QgsTinInterpolator::initialize : QgsTinInterpolator::interpolatePoint : PyInit__analysis : QgsGridFileWriter::writeFile : PyInit__analysis : PyMethodDef_RawFastCallKeywords : PyMethodDef_RawFastCallKeywords : PyEval_EvalFrameDefault : PyFunction_FastCallDict : PyMethodDef_RawFastCallDict : PyObject_Call : PyInit_sip : CPLStringList::empty : PyInit__core : QgsProcessingAlgorithm::runPrepared : QgsProcessingAlgRunnerTask::run : PyInit__core : QgsTask::start : QThreadPoolPrivate::reset : QThread::start : BaseThreadInitThunk : RtlUserThreadStart : ``` QGIS Info QGIS Version: 3.8.0-Zanzibar QGIS code revision: 11aff65f10 Compiled against Qt: 5.11.2 Running against Qt: 5.11.2 Compiled against GDAL: 2.4.1 Running against GDAL: 2.4.1 System Info CPU Type: x86_64 Kernel Type: winnt Kernel Version: 10.0.17763
process
tin creation crash crash id stack trace mathutils lineintersection dualedgetriangulation insertforcedsegment dualedgetriangulation addline qgstininterpolator insertdata qgstininterpolator initialize qgstininterpolator interpolatepoint pyinit analysis qgsgridfilewriter writefile pyinit analysis pymethoddef rawfastcallkeywords pymethoddef rawfastcallkeywords pyeval evalframedefault pyfunction fastcalldict pymethoddef rawfastcalldict pyobject call pyinit sip cplstringlist empty pyinit core qgsprocessingalgorithm runprepared qgsprocessingalgrunnertask run pyinit core qgstask start qthreadpoolprivate reset qthread start basethreadinitthunk rtluserthreadstart qgis info qgis version zanzibar qgis code revision compiled against qt running against qt compiled against gdal running against gdal system info cpu type kernel type winnt kernel version
1
32,308
8,826,423,247
IssuesEvent
2019-01-03 02:00:08
biojs/biojs-component-builder
https://api.github.com/repos/biojs/biojs-component-builder
closed
[F] Create API endpoint to request bundled components
registry/component-builder
Should return 200 and the bundled file if it is available or 302 (Found) with a Retry-After header if the component still needs to be build.
1.0
[F] Create API endpoint to request bundled components - Should return 200 and the bundled file if it is available or 302 (Found) with a Retry-After header if the component still needs to be build.
non_process
create api endpoint to request bundled components should return and the bundled file if it is available or found with a retry after header if the component still needs to be build
0
430,187
30,120,390,455
IssuesEvent
2023-06-30 14:43:46
ant-media/Ant-Media-Server
https://api.github.com/repos/ant-media/Ant-Media-Server
opened
React Native SDK - Documentation and Blogpost Update
documentation/blog-post
It is difficult to compile and run the SDK following the existing guide, therefore need to review the document section-wise and make it easy to compile and run the samples for beginners. #5201
1.0
React Native SDK - Documentation and Blogpost Update - It is difficult to compile and run the SDK following the existing guide, therefore need to review the document section-wise and make it easy to compile and run the samples for beginners. #5201
non_process
react native sdk documentation and blogpost update it is difficult to compile and run the sdk following the existing guide therefore need to review the document section wise and make it easy to compile and run the samples for beginners
0
8,047
11,220,662,924
IssuesEvent
2020-01-07 16:14:12
code4romania/expert-consultation-api
https://api.github.com/repos/code4romania/expert-consultation-api
closed
[Documents] Implement document breakdown logic
document processing documents java spring
As a user of the Legal Consultation platform, after loading a new document to the platform, I want to be able to see the breakdown of the document. The document needs to be split into logical units of data, based on the structure of the proposition. ![Doc breakdown edit](https://user-images.githubusercontent.com/15039873/58744287-92861e00-83f5-11e9-99c7-438d1b4f94e3.png)
1.0
[Documents] Implement document breakdown logic - As a user of the Legal Consultation platform, after loading a new document to the platform, I want to be able to see the breakdown of the document. The document needs to be split into logical units of data, based on the structure of the proposition. ![Doc breakdown edit](https://user-images.githubusercontent.com/15039873/58744287-92861e00-83f5-11e9-99c7-438d1b4f94e3.png)
process
implement document breakdown logic as a user of the legal consultation platform after loading a new document to the platform i want to be able to see the breakdown of the document the document needs to be split into logical units of data based on the structure of the proposition
1
44,072
2,899,105,672
IssuesEvent
2015-06-17 09:14:51
greenlion/PHP-SQL-Parser
https://api.github.com/repos/greenlion/PHP-SQL-Parser
closed
fake aliases for table names
bug imported Priority-Medium
_From [pho...@gmx.de](https://code.google.com/u/109317404671582518013/) on January 20, 2012 09:12:19_ Hi, I try to use your parser within a small project. I need the table names and their aliases within a SELECT statement to check, which columns are used within the ORDER BY clause. The first test produces a problem: SELECT * FROM answers ORDER BY qid The parser generates an alias ANSWERS, but the table doesn't have an alias. So it is not possible to find column names within the ORDER clause with the alias (the alias can or cannot be a valid alias). Please remove the alias from the resulting array, if the table doesn't have an alias within the SQL statement. Thanks a lot Andre _Original issue: http://code.google.com/p/php-sql-parser/issues/detail?id=18_
1.0
fake aliases for table names - _From [pho...@gmx.de](https://code.google.com/u/109317404671582518013/) on January 20, 2012 09:12:19_ Hi, I try to use your parser within a small project. I need the table names and their aliases within a SELECT statement to check, which columns are used within the ORDER BY clause. The first test produces a problem: SELECT * FROM answers ORDER BY qid The parser generates an alias ANSWERS, but the table doesn't have an alias. So it is not possible to find column names within the ORDER clause with the alias (the alias can or cannot be a valid alias). Please remove the alias from the resulting array, if the table doesn't have an alias within the SQL statement. Thanks a lot Andre _Original issue: http://code.google.com/p/php-sql-parser/issues/detail?id=18_
non_process
fake aliases for table names from on january hi i try to use your parser within a small project i need the table names and their aliases within a select statement to check which columns are used within the order by clause the first test produces a problem select from answers order by qid the parser generates an alias answers but the table doesn t have an alias so it is not possible to find column names within the order clause with the alias the alias can or cannot be a valid alias please remove the alias from the resulting array if the table doesn t have an alias within the sql statement thanks a lot andre original issue
0
15,971
20,187,935,196
IssuesEvent
2022-02-11 00:55:17
quark-engine/quark-engine
https://api.github.com/repos/quark-engine/quark-engine
closed
Quark label report does not work any longer UnboundLocalError
bug work-in-progress issue-processing-state-04
quark with option `-l detailed` or `-l max` crashes with `UnboundLocalError`. This used to work, so probably a _regression_? ``` # quark -a smsthief.apk -l detailed ... \__> \/ \/ v22.1.1 An Obfuscation-Neglect Android Malware Scoring System Requested API level 30 is larger than maximum we have, returning API level 28 instead. 100%|██████████████████████████████████████████████████████████████████████████████████████████| 181/181 [00:43<00:00, 4.18it/s] [*] Total Label found: 34 [*] Rules with label which max confidence >= 80%: 12 Traceback (most recent call last): File "/usr/local/bin/quark", line 8, in <module> sys.exit(entry_point()) File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 1128, in __call__ return self.main(*args, **kwargs) File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 1053, in main rv = self.invoke(ctx) File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 1395, in invoke return ctx.invoke(self.callback, **ctx.params) File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 754, in invoke return __callback(*args, **kwargs) File "/usr/local/lib/python3.8/dist-packages/quark/cli.py", line 304, in entry_point data.show_label_report(rule, all_labels, label) File "/usr/local/lib/python3.8/dist-packages/quark/core/quark.py", line 650, in show_label_report self.quark_analysis.label_report_table.field_names = [ File "/usr/local/lib/python3.8/dist-packages/prettytable.py", line 416, in _set_field_names if self._align and old_names: UnboundLocalError: local variable 'old_names' referenced before assignment ``` **To Reproduce** - Please get an APK. I tested with 2 totally different ones, and got the crash in both case, so I believe this does not depend on the APK. - Run `quark -a your.apk -l detailed` or `quark -a your.apk -l max` **Desktop (please complete the following information):** - OS: Docker container from *Ubuntu 20.04* (https://github.com/cryptax/androidre/blob/master/Dockerfile.re) - Version: `v22.1.1` - Python `3.8.10`
1.0
Quark label report does not work any longer UnboundLocalError - quark with option `-l detailed` or `-l max` crashes with `UnboundLocalError`. This used to work, so probably a _regression_? ``` # quark -a smsthief.apk -l detailed ... \__> \/ \/ v22.1.1 An Obfuscation-Neglect Android Malware Scoring System Requested API level 30 is larger than maximum we have, returning API level 28 instead. 100%|██████████████████████████████████████████████████████████████████████████████████████████| 181/181 [00:43<00:00, 4.18it/s] [*] Total Label found: 34 [*] Rules with label which max confidence >= 80%: 12 Traceback (most recent call last): File "/usr/local/bin/quark", line 8, in <module> sys.exit(entry_point()) File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 1128, in __call__ return self.main(*args, **kwargs) File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 1053, in main rv = self.invoke(ctx) File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 1395, in invoke return ctx.invoke(self.callback, **ctx.params) File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 754, in invoke return __callback(*args, **kwargs) File "/usr/local/lib/python3.8/dist-packages/quark/cli.py", line 304, in entry_point data.show_label_report(rule, all_labels, label) File "/usr/local/lib/python3.8/dist-packages/quark/core/quark.py", line 650, in show_label_report self.quark_analysis.label_report_table.field_names = [ File "/usr/local/lib/python3.8/dist-packages/prettytable.py", line 416, in _set_field_names if self._align and old_names: UnboundLocalError: local variable 'old_names' referenced before assignment ``` **To Reproduce** - Please get an APK. I tested with 2 totally different ones, and got the crash in both case, so I believe this does not depend on the APK. - Run `quark -a your.apk -l detailed` or `quark -a your.apk -l max` **Desktop (please complete the following information):** - OS: Docker container from *Ubuntu 20.04* (https://github.com/cryptax/androidre/blob/master/Dockerfile.re) - Version: `v22.1.1` - Python `3.8.10`
process
quark label report does not work any longer unboundlocalerror quark with option l detailed or l max crashes with unboundlocalerror this used to work so probably a regression quark a smsthief apk l detailed an obfuscation neglect android malware scoring system requested api level is larger than maximum we have returning api level instead ██████████████████████████████████████████████████████████████████████████████████████████ total label found rules with label which max confidence traceback most recent call last file usr local bin quark line in sys exit entry point file usr local lib dist packages click core py line in call return self main args kwargs file usr local lib dist packages click core py line in main rv self invoke ctx file usr local lib dist packages click core py line in invoke return ctx invoke self callback ctx params file usr local lib dist packages click core py line in invoke return callback args kwargs file usr local lib dist packages quark cli py line in entry point data show label report rule all labels label file usr local lib dist packages quark core quark py line in show label report self quark analysis label report table field names file usr local lib dist packages prettytable py line in set field names if self align and old names unboundlocalerror local variable old names referenced before assignment to reproduce please get an apk i tested with totally different ones and got the crash in both case so i believe this does not depend on the apk run quark a your apk l detailed or quark a your apk l max desktop please complete the following information os docker container from ubuntu version python
1
140,639
21,179,246,556
IssuesEvent
2022-04-08 05:55:13
HSLdevcom/jore4
https://api.github.com/repos/HSLdevcom/jore4
closed
As a public transport planner I want to upload timetables that does not affect compensations to the operator
design Timetables compensations
In some rare cases the compensation has been agreed in a different way between HSL and the operator, not in the normal basis. In these rare cases, compensation is calculated outside Jore4.
1.0
As a public transport planner I want to upload timetables that does not affect compensations to the operator - In some rare cases the compensation has been agreed in a different way between HSL and the operator, not in the normal basis. In these rare cases, compensation is calculated outside Jore4.
non_process
as a public transport planner i want to upload timetables that does not affect compensations to the operator in some rare cases the compensation has been agreed in a different way between hsl and the operator not in the normal basis in these rare cases compensation is calculated outside
0
9,106
12,619,873,591
IssuesEvent
2020-06-13 03:07:48
w-a-r-m-inventory-system/Food-Pantry-Inventory
https://api.github.com/repos/w-a-r-m-inventory-system/Food-Pantry-Inventory
closed
Implement permission levels though out the application
document requirement
- Only login page does not require the LoginMixIn - Each page will verify if the user is allowed to access that page - Throws a polite error message and allow return to previous page if not - So far, only the Add user page and the change user permission page will have conditional logic based on the permission of the user - Perhaps this can be isolated to one class or function that contains the necessary logic
1.0
Implement permission levels though out the application - - Only login page does not require the LoginMixIn - Each page will verify if the user is allowed to access that page - Throws a polite error message and allow return to previous page if not - So far, only the Add user page and the change user permission page will have conditional logic based on the permission of the user - Perhaps this can be isolated to one class or function that contains the necessary logic
non_process
implement permission levels though out the application only login page does not require the loginmixin each page will verify if the user is allowed to access that page throws a polite error message and allow return to previous page if not so far only the add user page and the change user permission page will have conditional logic based on the permission of the user perhaps this can be isolated to one class or function that contains the necessary logic
0
10,668
13,456,825,250
IssuesEvent
2020-09-09 08:23:32
jgraley/inferno-cpp2v
https://api.github.com/repos/jgraley/inferno-cpp2v
opened
Partial constraint testing
Constraint Processing General Optimisation
Add eg `Constraint::PartialTest`. Allow values to be provided for any subset of the constraint's variables. See about the feasibility of curtailing the conjecture walk when no values remain to be checked (optimisation).
1.0
Partial constraint testing - Add eg `Constraint::PartialTest`. Allow values to be provided for any subset of the constraint's variables. See about the feasibility of curtailing the conjecture walk when no values remain to be checked (optimisation).
process
partial constraint testing add eg constraint partialtest allow values to be provided for any subset of the constraint s variables see about the feasibility of curtailing the conjecture walk when no values remain to be checked optimisation
1
19,169
25,270,038,222
IssuesEvent
2022-11-16 08:39:30
tesseract-ocr/tesseract
https://api.github.com/repos/tesseract-ocr/tesseract
closed
Tesseract hangs
process hangs
### Environment * **Tesseract Version**: tesseract 5.2.0 leptonica-1.82.0 libgif 5.2.1 : libjpeg 8d (libjpeg-turbo 2.1.4) : libpng 1.6.37+apng : libtiff 4.4.0 : zlib 1.2.12 : libwebp 1.2.4 Found OpenMP 201811 Found libarchive 3.6.1 zlib/1.2.12 liblzma/5.2.5 bz2lib/1.0.8 liblz4/1.9.4 libzstd/1.5.2 Found libcurl/7.85.0 OpenSSL/1.1.1o zlib/1.2.12 libpsl/0.21.1 (+libidn2/2.3.3) libssh2/1.10.0 nghttp2/1.48.0 * **Platform**: <!-- either `uname -a` output, or if Windows, version and 32-bit or 64-bit --> FreeBSD hostname.com 13.1-RELEASE-p3 FreeBSD 13.1-RELEASE-p3 GENERIC amd64 ### Current Behavior: Tesseract hangs without any result. Hanged process even can not be killed. Commands like ` tesseract anyimage.png stdout` or ` tesseract -l eng+rus anyimage.png stdout` ### Expected Behavior: Tesseract works (produces result and exits)
1.0
Tesseract hangs - ### Environment * **Tesseract Version**: tesseract 5.2.0 leptonica-1.82.0 libgif 5.2.1 : libjpeg 8d (libjpeg-turbo 2.1.4) : libpng 1.6.37+apng : libtiff 4.4.0 : zlib 1.2.12 : libwebp 1.2.4 Found OpenMP 201811 Found libarchive 3.6.1 zlib/1.2.12 liblzma/5.2.5 bz2lib/1.0.8 liblz4/1.9.4 libzstd/1.5.2 Found libcurl/7.85.0 OpenSSL/1.1.1o zlib/1.2.12 libpsl/0.21.1 (+libidn2/2.3.3) libssh2/1.10.0 nghttp2/1.48.0 * **Platform**: <!-- either `uname -a` output, or if Windows, version and 32-bit or 64-bit --> FreeBSD hostname.com 13.1-RELEASE-p3 FreeBSD 13.1-RELEASE-p3 GENERIC amd64 ### Current Behavior: Tesseract hangs without any result. Hanged process even can not be killed. Commands like ` tesseract anyimage.png stdout` or ` tesseract -l eng+rus anyimage.png stdout` ### Expected Behavior: Tesseract works (produces result and exits)
process
tesseract hangs environment tesseract version tesseract leptonica libgif libjpeg libjpeg turbo libpng apng libtiff zlib libwebp found openmp found libarchive zlib liblzma libzstd found libcurl openssl zlib libpsl platform freebsd hostname com release freebsd release generic current behavior tesseract hangs without any result hanged process even can not be killed commands like tesseract anyimage png stdout or tesseract l eng rus anyimage png stdout expected behavior tesseract works produces result and exits
1
521,416
15,109,303,081
IssuesEvent
2021-02-08 17:40:18
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
Minimal libc malloc() is unprotected
Regression bug priority: high
This is a reminder that PR #31898 must be merged before the 2.5 release.
1.0
Minimal libc malloc() is unprotected - This is a reminder that PR #31898 must be merged before the 2.5 release.
non_process
minimal libc malloc is unprotected this is a reminder that pr must be merged before the release
0
15,717
19,849,206,015
IssuesEvent
2022-01-21 10:21:19
ooi-data/RS01SLBS-LJ01A-10-ADCPTE101-streamed-adcp_pd0_beam_parsed
https://api.github.com/repos/ooi-data/RS01SLBS-LJ01A-10-ADCPTE101-streamed-adcp_pd0_beam_parsed
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T10:21:18.388301. ## Details Flow name: `RS01SLBS-LJ01A-10-ADCPTE101-streamed-adcp_pd0_beam_parsed` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T10:21:18.388301. ## Details Flow name: `RS01SLBS-LJ01A-10-ADCPTE101-streamed-adcp_pd0_beam_parsed` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name streamed adcp beam parsed task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
328,944
10,007,297,738
IssuesEvent
2019-07-14 09:29:39
gama-platform/gama
https://api.github.com/repos/gama-platform/gama
closed
Exception in an aspect with two displays can freeze GAMA when opening the displays
> Bug Affects Usability Concerns Interface Display All Priority High Version Git
**Describe the bug** Let's consider an experiment with 2 displays and that the first display displays a species with an Exception thrown in the aspect. Let's now consider that the Preference allowing to show the Exception thrown in displays is activated. Then when I run such a model, GAMA freezes and no Exception appears in GAMA. Some exceptions are displayed in the Eclipse console. Notes that : - when there is only one display with the Exception inside The exception is displayed in GAMA. - when the Preference is set to not show the exceptions in the displays then the freeze does not occur **To Reproduce** Run the following model: ``` model NewModel global { init { create toot number: 10; write "maître"; } } species toot { list<int> l <- [1,2]; aspect c { draw circle(10); } aspect default { loop i from: 0 to: length(l) { draw circle(l[i]); } } } experiment name type: gui { output { display d2 { species toot; } display d { species toot aspect: c; } } } ``` **Desktop (please complete the following information):** - OS: macOSX - GAMA version: git
1.0
Exception in an aspect with two displays can freeze GAMA when opening the displays - **Describe the bug** Let's consider an experiment with 2 displays and that the first display displays a species with an Exception thrown in the aspect. Let's now consider that the Preference allowing to show the Exception thrown in displays is activated. Then when I run such a model, GAMA freezes and no Exception appears in GAMA. Some exceptions are displayed in the Eclipse console. Notes that : - when there is only one display with the Exception inside The exception is displayed in GAMA. - when the Preference is set to not show the exceptions in the displays then the freeze does not occur **To Reproduce** Run the following model: ``` model NewModel global { init { create toot number: 10; write "maître"; } } species toot { list<int> l <- [1,2]; aspect c { draw circle(10); } aspect default { loop i from: 0 to: length(l) { draw circle(l[i]); } } } experiment name type: gui { output { display d2 { species toot; } display d { species toot aspect: c; } } } ``` **Desktop (please complete the following information):** - OS: macOSX - GAMA version: git
non_process
exception in an aspect with two displays can freeze gama when opening the displays describe the bug let s consider an experiment with displays and that the first display displays a species with an exception thrown in the aspect let s now consider that the preference allowing to show the exception thrown in displays is activated then when i run such a model gama freezes and no exception appears in gama some exceptions are displayed in the eclipse console notes that when there is only one display with the exception inside the exception is displayed in gama when the preference is set to not show the exceptions in the displays then the freeze does not occur to reproduce run the following model model newmodel global init create toot number write maître species toot list l aspect c draw circle aspect default loop i from to length l draw circle l experiment name type gui output display species toot display d species toot aspect c desktop please complete the following information os macosx gama version git
0
15,632
19,783,392,957
IssuesEvent
2022-01-18 01:39:05
linuxdeepin/developer-center
https://api.github.com/repos/linuxdeepin/developer-center
closed
Validity Sensors VFS491 without controllers
other | delay processing
the VFS491 validation sensor, used for fingerprint scanning, has no drivers in linux deepin. for this sensor only proprietary drivers are available https://ftp.hp.com/pub/softpaq/sp84501-85000/sp84530.tar https://github.com/rindeal/libfprint-vfs_proprietary-driver these drivers work for various models of fingerprint readers used in computers.
1.0
Validity Sensors VFS491 without controllers - the VFS491 validation sensor, used for fingerprint scanning, has no drivers in linux deepin. for this sensor only proprietary drivers are available https://ftp.hp.com/pub/softpaq/sp84501-85000/sp84530.tar https://github.com/rindeal/libfprint-vfs_proprietary-driver these drivers work for various models of fingerprint readers used in computers.
process
validity sensors without controllers the validation sensor used for fingerprint scanning has no drivers in linux deepin for this sensor only proprietary drivers are available these drivers work for various models of fingerprint readers used in computers
1
437,867
12,603,622,331
IssuesEvent
2020-06-11 13:46:27
cilium/cilium
https://api.github.com/repos/cilium/cilium
closed
CI: provide agent logs when cilium pods are in crash loopback
area/CI priority/high project/ci-force
When agent's pod goes into crash loopback and therefore CI results into failure, it would be helpful to see why that is happening. Right now the output is not useful, e.g. the zip file doesn't collect the agent log. [Example dump](https://github.com/cilium/cilium/files/4576627/f3ee8175_K8sChaosTest_Connectivity_demo_application_Endpoint_can_still_connect_while_Cilium_is_not_running.1.zip) from such CI run where the pod went into this state when init.sh was failing due to compilation error. The issue does not happen when the CI is run locally.
1.0
CI: provide agent logs when cilium pods are in crash loopback - When agent's pod goes into crash loopback and therefore CI results into failure, it would be helpful to see why that is happening. Right now the output is not useful, e.g. the zip file doesn't collect the agent log. [Example dump](https://github.com/cilium/cilium/files/4576627/f3ee8175_K8sChaosTest_Connectivity_demo_application_Endpoint_can_still_connect_while_Cilium_is_not_running.1.zip) from such CI run where the pod went into this state when init.sh was failing due to compilation error. The issue does not happen when the CI is run locally.
non_process
ci provide agent logs when cilium pods are in crash loopback when agent s pod goes into crash loopback and therefore ci results into failure it would be helpful to see why that is happening right now the output is not useful e g the zip file doesn t collect the agent log from such ci run where the pod went into this state when init sh was failing due to compilation error the issue does not happen when the ci is run locally
0
10,795
6,922,324,037
IssuesEvent
2017-11-30 02:30:26
dryuen/ist5313-final-project-template
https://api.github.com/repos/dryuen/ist5313-final-project-template
closed
Group 1- Module 3- Hotspot
developer enhancement f17-group1 other usability
Hotspot should include example to help the user identify the steps.
True
Group 1- Module 3- Hotspot - Hotspot should include example to help the user identify the steps.
non_process
group module hotspot hotspot should include example to help the user identify the steps
0
12,994
15,358,513,454
IssuesEvent
2021-03-01 14:53:14
edwardsmarc/CASFRI
https://api.github.com/repos/edwardsmarc/CASFRI
opened
TT_ProduceInvGeoHistory('SK01') fails on one row
bug high post-translation process
ERROR: TT_PolygonGeoHistory() ERROR: TT_SafeDifference() failed on SK01-xxxxxxxxxxxxUTM-xxxxxxxxxx-1348605004-0343222...
1.0
TT_ProduceInvGeoHistory('SK01') fails on one row - ERROR: TT_PolygonGeoHistory() ERROR: TT_SafeDifference() failed on SK01-xxxxxxxxxxxxUTM-xxxxxxxxxx-1348605004-0343222...
process
tt produceinvgeohistory fails on one row error tt polygongeohistory error tt safedifference failed on xxxxxxxxxxxxutm xxxxxxxxxx
1
70,171
15,053,124,142
IssuesEvent
2021-02-03 15:58:06
istio/istio
https://api.github.com/repos/istio/istio
closed
Log mTLS permissive warning only once
area/networking/envoy area/security kind/enhancement
``` 2020-11-03T18:17:23.189271Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/ 2020-11-03T18:17:23.190018Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/ 2020-11-03T18:17:23.192851Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/ 2020-11-03T18:17:23.193503Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/ ``` log is very spammy. Should log it once for the lifetime of a pod
True
Log mTLS permissive warning only once - ``` 2020-11-03T18:17:23.189271Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/ 2020-11-03T18:17:23.190018Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/ 2020-11-03T18:17:23.192851Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/ 2020-11-03T18:17:23.193503Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/ ``` log is very spammy. Should log it once for the lifetime of a pod
non_process
log mtls permissive warning only once warning envoy filter mtls permissive mode is used connection can be either plaintext or tls and client cert can be omitted please consider to upgrade to mtls strict mode for more secure configuration that only allows tls connection with client cert see warning envoy filter mtls permissive mode is used connection can be either plaintext or tls and client cert can be omitted please consider to upgrade to mtls strict mode for more secure configuration that only allows tls connection with client cert see warning envoy filter mtls permissive mode is used connection can be either plaintext or tls and client cert can be omitted please consider to upgrade to mtls strict mode for more secure configuration that only allows tls connection with client cert see warning envoy filter mtls permissive mode is used connection can be either plaintext or tls and client cert can be omitted please consider to upgrade to mtls strict mode for more secure configuration that only allows tls connection with client cert see log is very spammy should log it once for the lifetime of a pod
0
160,629
25,198,998,867
IssuesEvent
2022-11-12 22:14:16
authelia/authelia
https://api.github.com/repos/authelia/authelia
opened
Design: Support for multiple WebAuthn devices for each user
priority/4/normal type/feature status/needs-design
### Description Desired user scenarios: * Users can add more than one WebAuthn device to their account ([listed on Authelia's WebAuthn roadmap](https://www.authelia.com/roadmap/active/webauthn/#multi-device-registration)) * Registered device name can be provided by the user during device registration * Users can manage their existing WebAuthn devices within their account ([listed on Authelia's Dashboard / Control Panel roadmap](https://www.authelia.com/roadmap/active/dashboard-control-panel/#initial-implementation)) * View information about registered device(s) * Rename existing device(s) * Add additional device(s) * Delete existing device(s) ### Use Case It's common and recommended for users with hardware security tokens (ex. Yubikeys) to have more than one. As an example, I have a backup Yubikey in case my main key gets lost. Each of my Yubikeys are configured in my accounts where I use them for 2-factor authentication. Hardware security tokens are usable via the WebAuthn protocol, which Authelia already uses and supports. Sites which support hardware security tokens for 2-factor authentication generally support this use case. Examples include Google, Dropbox, and GitHub. This feature was requested in https://github.com/authelia/authelia/issues/275. ### Details For reference, this is the current WebAuthn device registration flow: 1. On the login frame (at either the `OneFactor` or `TwoFactor` authentication levels), the user clicks the "Lost your device?" link ![registration-flow-link](https://user-images.githubusercontent.com/2831985/201496556-947cbf09-dbc6-4794-af70-5c3656315735.png) 2. Authelia generates a token, encoded as JWT, and sends the user an email containing a link with the JWT token 3. User clicks the link in their email 4. Authelia retrieves the token from the URI, decodes and verifies the JWT, and [performs the WebAuthn attestation ceremony](https://github.com/authelia/authelia/blob/c7a668c897cbda49d858a19e40705dca88e1ca3e/web/src/views/DeviceRegistration/RegisterWebauthn.tsx#L35). 5. On successful WebAuthn device attestation, the new device info is stored in the Authelia database. [Currently the device registration is hardcoded with a description of `Primary`](https://github.com/authelia/authelia/blob/c7a668c897cbda49d858a19e40705dca88e1ca3e/internal/handlers/handler_register_webauthn.go#L137). If an existing device was registered, it is overwritten. 6. On WebAuthn device registration success, the user is redirected back to the login page (which then re-prompts for 2FA authentication with the new device). #### Proposed changes: Users can add more than one WebAuthn device to their account * As mentioned on Authelia's [WebAuthn roadmap](https://www.authelia.com/roadmap/active/webauthn/#multi-device-registration), the storage backend has existing support for multiple devices. Devices are overwritten on re-registration simply due to the hardcoded name (referred to in the backend as "description") `Primary` which is used for all devices. The hardcoded device name should be removed so it can be provided to the backend via the API as in https://github.com/authelia/authelia/pull/4363. * Support entry of a user-provided device name as part of the WebAuthn device registration flow. This can then be provided to the backend on registration via the previously mentioned API change. * The user should be prevented from reusing the name of existing device within their account by both the UI and backend #### Proposed changes: Users can manage their existing WebAuthn devices within their account * New UI is needed for device management. From the roadmap and discussions on Discord, the desired feature is an extensible user account settings UI. Although the UI may only support WebAuthn device management initially, this UI would later provide options for management of other 2FA options and other unrelated settings (such as Authelia service configuration options, perhaps). * Some existing features work around a settings UI is present in the [`feat-settings-ui` branch](https://github.com/authelia/authelia/tree/feat-settings-ui). The UI in this branch currently looks like this: ![feat-settings-ui](https://user-images.githubusercontent.com/2831985/201496602-a4a765f9-6991-4642-99e8-8fc5bcc25305.png) * WebAuthn settings UI should display information about registered devices: * User-entered device name ("description" in the backend) * Additional details such as registered date, and last used date. These values are already stored in the database. * No API endpoint currently exists for retrieving configured WebAuthn device information for a user. A new API endpoint is needed. (A new endpoint is present in the [`feat-settings-ui` branch](https://github.com/authelia/authelia/tree/feat-settings-ui).) * WebAuthn settings UI should allow registration of multiple devices. Some options: 1. Reuse existing registration flow. Without additional changes, re-launching the flow will send the user another email with a link+token. Opening this page will start the WebAuthn device attestation/registration flow. Pros: Smallest code change. Cons: Cumbersome for the user, as it requires an email round-trip for each new added device. 2. Reuse existing registration flow, without requiring email verification when already logged in at the `TwoFactor` level. Pros: Still a smaller code change. Less cumbersome for the user. No expected security impact as the user already had to verify their email to register/log in with 2FA in the first place. Cons: Redirection to existing flow may be jarring for users. 3. Reuse existing registration flow in a popup modal over the new settings UI, without requiring email verification when already logged in at the `TwoFactor` level. Pros: Most intuitive flow for the user. Modal popup makes it more obvious the registration function operates within the settings UI space. No expected security impact as the user already had to verify their email to register/log in with 2FA in the first place. Cons: Largest code change. * **Note**: Skipping email verification when logged in at the `TwoFactor` level could be accomplished by simply succeeding the [identity verification middleware](https://github.com/authelia/authelia/blob/c7a668c897cbda49d858a19e40705dca88e1ca3e/internal/middlewares/identity_verification.go) if the user's auth level is `TwoFactor`. I sketched this out here: https://github.com/smkent/authelia/pull/1 * WebAuthn settings UI should allow rename of multiple devices. * A new API endpoint is needed for modifying a WebAuthn device. * WebAuthn settings UI should allow deletion of multiple devices. * A new API endpoint is needed for deleting a WebAuthn device. * When all WebAuthn devices are deleted, my preferred option is to: 1. Do nothing. The user will remain logged in at the `TwoFactor` level within their current session. Logging in again will automatically restart the 2FA device registration prompt if no other 2FA methods are enabled. Pros: Less disruptive flow for the user. User can delete their last device and (re-)add a device in the settings UI without their session ending. #### Additional considerations * Email-based identity verification may be reworked via https://github.com/authelia/authelia/issues/3801, which would replace the JWT token with a randomly generated one time password. This may affect some of the proposed feature implementation above. ### Documentation _No response_
1.0
Design: Support for multiple WebAuthn devices for each user - ### Description Desired user scenarios: * Users can add more than one WebAuthn device to their account ([listed on Authelia's WebAuthn roadmap](https://www.authelia.com/roadmap/active/webauthn/#multi-device-registration)) * Registered device name can be provided by the user during device registration * Users can manage their existing WebAuthn devices within their account ([listed on Authelia's Dashboard / Control Panel roadmap](https://www.authelia.com/roadmap/active/dashboard-control-panel/#initial-implementation)) * View information about registered device(s) * Rename existing device(s) * Add additional device(s) * Delete existing device(s) ### Use Case It's common and recommended for users with hardware security tokens (ex. Yubikeys) to have more than one. As an example, I have a backup Yubikey in case my main key gets lost. Each of my Yubikeys are configured in my accounts where I use them for 2-factor authentication. Hardware security tokens are usable via the WebAuthn protocol, which Authelia already uses and supports. Sites which support hardware security tokens for 2-factor authentication generally support this use case. Examples include Google, Dropbox, and GitHub. This feature was requested in https://github.com/authelia/authelia/issues/275. ### Details For reference, this is the current WebAuthn device registration flow: 1. On the login frame (at either the `OneFactor` or `TwoFactor` authentication levels), the user clicks the "Lost your device?" link ![registration-flow-link](https://user-images.githubusercontent.com/2831985/201496556-947cbf09-dbc6-4794-af70-5c3656315735.png) 2. Authelia generates a token, encoded as JWT, and sends the user an email containing a link with the JWT token 3. User clicks the link in their email 4. Authelia retrieves the token from the URI, decodes and verifies the JWT, and [performs the WebAuthn attestation ceremony](https://github.com/authelia/authelia/blob/c7a668c897cbda49d858a19e40705dca88e1ca3e/web/src/views/DeviceRegistration/RegisterWebauthn.tsx#L35). 5. On successful WebAuthn device attestation, the new device info is stored in the Authelia database. [Currently the device registration is hardcoded with a description of `Primary`](https://github.com/authelia/authelia/blob/c7a668c897cbda49d858a19e40705dca88e1ca3e/internal/handlers/handler_register_webauthn.go#L137). If an existing device was registered, it is overwritten. 6. On WebAuthn device registration success, the user is redirected back to the login page (which then re-prompts for 2FA authentication with the new device). #### Proposed changes: Users can add more than one WebAuthn device to their account * As mentioned on Authelia's [WebAuthn roadmap](https://www.authelia.com/roadmap/active/webauthn/#multi-device-registration), the storage backend has existing support for multiple devices. Devices are overwritten on re-registration simply due to the hardcoded name (referred to in the backend as "description") `Primary` which is used for all devices. The hardcoded device name should be removed so it can be provided to the backend via the API as in https://github.com/authelia/authelia/pull/4363. * Support entry of a user-provided device name as part of the WebAuthn device registration flow. This can then be provided to the backend on registration via the previously mentioned API change. * The user should be prevented from reusing the name of existing device within their account by both the UI and backend #### Proposed changes: Users can manage their existing WebAuthn devices within their account * New UI is needed for device management. From the roadmap and discussions on Discord, the desired feature is an extensible user account settings UI. Although the UI may only support WebAuthn device management initially, this UI would later provide options for management of other 2FA options and other unrelated settings (such as Authelia service configuration options, perhaps). * Some existing features work around a settings UI is present in the [`feat-settings-ui` branch](https://github.com/authelia/authelia/tree/feat-settings-ui). The UI in this branch currently looks like this: ![feat-settings-ui](https://user-images.githubusercontent.com/2831985/201496602-a4a765f9-6991-4642-99e8-8fc5bcc25305.png) * WebAuthn settings UI should display information about registered devices: * User-entered device name ("description" in the backend) * Additional details such as registered date, and last used date. These values are already stored in the database. * No API endpoint currently exists for retrieving configured WebAuthn device information for a user. A new API endpoint is needed. (A new endpoint is present in the [`feat-settings-ui` branch](https://github.com/authelia/authelia/tree/feat-settings-ui).) * WebAuthn settings UI should allow registration of multiple devices. Some options: 1. Reuse existing registration flow. Without additional changes, re-launching the flow will send the user another email with a link+token. Opening this page will start the WebAuthn device attestation/registration flow. Pros: Smallest code change. Cons: Cumbersome for the user, as it requires an email round-trip for each new added device. 2. Reuse existing registration flow, without requiring email verification when already logged in at the `TwoFactor` level. Pros: Still a smaller code change. Less cumbersome for the user. No expected security impact as the user already had to verify their email to register/log in with 2FA in the first place. Cons: Redirection to existing flow may be jarring for users. 3. Reuse existing registration flow in a popup modal over the new settings UI, without requiring email verification when already logged in at the `TwoFactor` level. Pros: Most intuitive flow for the user. Modal popup makes it more obvious the registration function operates within the settings UI space. No expected security impact as the user already had to verify their email to register/log in with 2FA in the first place. Cons: Largest code change. * **Note**: Skipping email verification when logged in at the `TwoFactor` level could be accomplished by simply succeeding the [identity verification middleware](https://github.com/authelia/authelia/blob/c7a668c897cbda49d858a19e40705dca88e1ca3e/internal/middlewares/identity_verification.go) if the user's auth level is `TwoFactor`. I sketched this out here: https://github.com/smkent/authelia/pull/1 * WebAuthn settings UI should allow rename of multiple devices. * A new API endpoint is needed for modifying a WebAuthn device. * WebAuthn settings UI should allow deletion of multiple devices. * A new API endpoint is needed for deleting a WebAuthn device. * When all WebAuthn devices are deleted, my preferred option is to: 1. Do nothing. The user will remain logged in at the `TwoFactor` level within their current session. Logging in again will automatically restart the 2FA device registration prompt if no other 2FA methods are enabled. Pros: Less disruptive flow for the user. User can delete their last device and (re-)add a device in the settings UI without their session ending. #### Additional considerations * Email-based identity verification may be reworked via https://github.com/authelia/authelia/issues/3801, which would replace the JWT token with a randomly generated one time password. This may affect some of the proposed feature implementation above. ### Documentation _No response_
non_process
design support for multiple webauthn devices for each user description desired user scenarios users can add more than one webauthn device to their account registered device name can be provided by the user during device registration users can manage their existing webauthn devices within their account view information about registered device s rename existing device s add additional device s delete existing device s use case it s common and recommended for users with hardware security tokens ex yubikeys to have more than one as an example i have a backup yubikey in case my main key gets lost each of my yubikeys are configured in my accounts where i use them for factor authentication hardware security tokens are usable via the webauthn protocol which authelia already uses and supports sites which support hardware security tokens for factor authentication generally support this use case examples include google dropbox and github this feature was requested in details for reference this is the current webauthn device registration flow on the login frame at either the onefactor or twofactor authentication levels the user clicks the lost your device link authelia generates a token encoded as jwt and sends the user an email containing a link with the jwt token user clicks the link in their email authelia retrieves the token from the uri decodes and verifies the jwt and on successful webauthn device attestation the new device info is stored in the authelia database if an existing device was registered it is overwritten on webauthn device registration success the user is redirected back to the login page which then re prompts for authentication with the new device proposed changes users can add more than one webauthn device to their account as mentioned on authelia s the storage backend has existing support for multiple devices devices are overwritten on re registration simply due to the hardcoded name referred to in the backend as description primary which is used for all devices the hardcoded device name should be removed so it can be provided to the backend via the api as in support entry of a user provided device name as part of the webauthn device registration flow this can then be provided to the backend on registration via the previously mentioned api change the user should be prevented from reusing the name of existing device within their account by both the ui and backend proposed changes users can manage their existing webauthn devices within their account new ui is needed for device management from the roadmap and discussions on discord the desired feature is an extensible user account settings ui although the ui may only support webauthn device management initially this ui would later provide options for management of other options and other unrelated settings such as authelia service configuration options perhaps some existing features work around a settings ui is present in the the ui in this branch currently looks like this webauthn settings ui should display information about registered devices user entered device name description in the backend additional details such as registered date and last used date these values are already stored in the database no api endpoint currently exists for retrieving configured webauthn device information for a user a new api endpoint is needed a new endpoint is present in the webauthn settings ui should allow registration of multiple devices some options reuse existing registration flow without additional changes re launching the flow will send the user another email with a link token opening this page will start the webauthn device attestation registration flow pros smallest code change cons cumbersome for the user as it requires an email round trip for each new added device reuse existing registration flow without requiring email verification when already logged in at the twofactor level pros still a smaller code change less cumbersome for the user no expected security impact as the user already had to verify their email to register log in with in the first place cons redirection to existing flow may be jarring for users reuse existing registration flow in a popup modal over the new settings ui without requiring email verification when already logged in at the twofactor level pros most intuitive flow for the user modal popup makes it more obvious the registration function operates within the settings ui space no expected security impact as the user already had to verify their email to register log in with in the first place cons largest code change note skipping email verification when logged in at the twofactor level could be accomplished by simply succeeding the if the user s auth level is twofactor i sketched this out here webauthn settings ui should allow rename of multiple devices a new api endpoint is needed for modifying a webauthn device webauthn settings ui should allow deletion of multiple devices a new api endpoint is needed for deleting a webauthn device when all webauthn devices are deleted my preferred option is to do nothing the user will remain logged in at the twofactor level within their current session logging in again will automatically restart the device registration prompt if no other methods are enabled pros less disruptive flow for the user user can delete their last device and re add a device in the settings ui without their session ending additional considerations email based identity verification may be reworked via which would replace the jwt token with a randomly generated one time password this may affect some of the proposed feature implementation above documentation no response
0
755,211
26,421,286,243
IssuesEvent
2023-01-13 20:47:01
DSpace/DSpace
https://api.github.com/repos/DSpace/DSpace
opened
Request a Copy's "helpdesk" strategy requires authentication to respond
bug authorization medium priority help wanted
**Describe the bug** Discovered while testing #8319 In DSpace 6, selecting the "helpdesk" strategy for Request a Copy allowed anyone receiving the helpdesk email to respond to the request _without authenticating into DSpace._ This is currently not possible in DSpace 7. **To Reproduce** Steps to reproduce the behavior: 1. Activate request a copy with Helpdesk Strategy: https://wiki.lyrasis.org/display/DSDOC7x/Request+a+Copy#RequestaCopy-(Optional)RequestingacopywithHelpDeskworkflow 2. As anonymous user, look at an embargoed item, click on the file to display the request a copy form. 3. Complete and send the request a copy form 4. As helpdesk staff, look at the email inbox 5. Open the Request a copy email and click on the link to respond to it. 6. **You are now required to login to DSpace** in order to respond. Previously this was available anonymously. 7. After logging in, you can approve or decline the request. **Expected behavior** You should be able to approve or deny the request without authenticating. The link sent to the helpdesk staff includes a unique token which is all that is necessary to respond to the request. **Related work** Related to #8319 and originally reported in #8304
1.0
Request a Copy's "helpdesk" strategy requires authentication to respond - **Describe the bug** Discovered while testing #8319 In DSpace 6, selecting the "helpdesk" strategy for Request a Copy allowed anyone receiving the helpdesk email to respond to the request _without authenticating into DSpace._ This is currently not possible in DSpace 7. **To Reproduce** Steps to reproduce the behavior: 1. Activate request a copy with Helpdesk Strategy: https://wiki.lyrasis.org/display/DSDOC7x/Request+a+Copy#RequestaCopy-(Optional)RequestingacopywithHelpDeskworkflow 2. As anonymous user, look at an embargoed item, click on the file to display the request a copy form. 3. Complete and send the request a copy form 4. As helpdesk staff, look at the email inbox 5. Open the Request a copy email and click on the link to respond to it. 6. **You are now required to login to DSpace** in order to respond. Previously this was available anonymously. 7. After logging in, you can approve or decline the request. **Expected behavior** You should be able to approve or deny the request without authenticating. The link sent to the helpdesk staff includes a unique token which is all that is necessary to respond to the request. **Related work** Related to #8319 and originally reported in #8304
non_process
request a copy s helpdesk strategy requires authentication to respond describe the bug discovered while testing in dspace selecting the helpdesk strategy for request a copy allowed anyone receiving the helpdesk email to respond to the request without authenticating into dspace this is currently not possible in dspace to reproduce steps to reproduce the behavior activate request a copy with helpdesk strategy as anonymous user look at an embargoed item click on the file to display the request a copy form complete and send the request a copy form as helpdesk staff look at the email inbox open the request a copy email and click on the link to respond to it you are now required to login to dspace in order to respond previously this was available anonymously after logging in you can approve or decline the request expected behavior you should be able to approve or deny the request without authenticating the link sent to the helpdesk staff includes a unique token which is all that is necessary to respond to the request related work related to and originally reported in
0
322,680
9,821,524,880
IssuesEvent
2019-06-14 07:28:50
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
start.parimatch.com - see bug description
browser-firefox engine-gecko priority-normal
<!-- @browser: Firefox 68.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://start.parimatch.com/az/fight/0/5/?btag=a_16060b_1047c_375&siteid=16060 **Browser / Version**: Firefox 68.0 **Operating System**: Windows 7 **Tested Another Browser**: Unknown **Problem type**: Something else **Description**: site contains virus **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/6/a6a0aa41-a2d3-4ac4-94e0-5864ceffa5ac-thumb.jpeg)](https://webcompat.com/uploads/2019/6/a6a0aa41-a2d3-4ac4-94e0-5864ceffa5ac.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190610153228</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
start.parimatch.com - see bug description - <!-- @browser: Firefox 68.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://start.parimatch.com/az/fight/0/5/?btag=a_16060b_1047c_375&siteid=16060 **Browser / Version**: Firefox 68.0 **Operating System**: Windows 7 **Tested Another Browser**: Unknown **Problem type**: Something else **Description**: site contains virus **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/6/a6a0aa41-a2d3-4ac4-94e0-5864ceffa5ac-thumb.jpeg)](https://webcompat.com/uploads/2019/6/a6a0aa41-a2d3-4ac4-94e0-5864ceffa5ac.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190610153228</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
start parimatch com see bug description url browser version firefox operating system windows tested another browser unknown problem type something else description site contains virus steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta from with ❤️
0
5,664
8,531,628,714
IssuesEvent
2018-11-04 14:14:40
aiidateam/aiida_core
https://api.github.com/repos/aiidateam/aiida_core
opened
Add optional TransporTask for JobProcess to copy over files
priority/nice to have requires discussion topic/Engine topic/JobCalculationAndProcess type/feature request
A common use case, is that after a `JobProcess` has finished, one might want to copy some files from the work directory, which is often on a scratch space, to a more durable location and attach them through a `RemoteFolder` to the `JobCalculation` node. This is different from the files that are already retrieved for parsing, as often these additional files are big, and one wants to keep them on the remote computer. A typical example are big files that are necessary to restart a calculation from, that are too big to retrieve but also too costly to lose and have to recompute. There is a simple and naive solution of just running an inline calculation that will open transport to the remote computer and use `copy_from_remote_to_remote` to copy the files and create a `RemoteFolder` object to attach to the computer, but this can be dangerous. If an exception occurs, the entire process, most likely the `WorkChain` that launched the `JobCalculation` will except as well. Of course one can catch the exceptions but then the files will not be copied. The `JobProcess` infrastructure has a rich architecture to deal with tasks that require transport and can fail. The best option would be to leverage this functionality for this particular purpose by creating an optional `TransportTask` to copy files after the job has finished, but before going over to the retrieval and parse step.
1.0
Add optional TransporTask for JobProcess to copy over files - A common use case, is that after a `JobProcess` has finished, one might want to copy some files from the work directory, which is often on a scratch space, to a more durable location and attach them through a `RemoteFolder` to the `JobCalculation` node. This is different from the files that are already retrieved for parsing, as often these additional files are big, and one wants to keep them on the remote computer. A typical example are big files that are necessary to restart a calculation from, that are too big to retrieve but also too costly to lose and have to recompute. There is a simple and naive solution of just running an inline calculation that will open transport to the remote computer and use `copy_from_remote_to_remote` to copy the files and create a `RemoteFolder` object to attach to the computer, but this can be dangerous. If an exception occurs, the entire process, most likely the `WorkChain` that launched the `JobCalculation` will except as well. Of course one can catch the exceptions but then the files will not be copied. The `JobProcess` infrastructure has a rich architecture to deal with tasks that require transport and can fail. The best option would be to leverage this functionality for this particular purpose by creating an optional `TransportTask` to copy files after the job has finished, but before going over to the retrieval and parse step.
process
add optional transportask for jobprocess to copy over files a common use case is that after a jobprocess has finished one might want to copy some files from the work directory which is often on a scratch space to a more durable location and attach them through a remotefolder to the jobcalculation node this is different from the files that are already retrieved for parsing as often these additional files are big and one wants to keep them on the remote computer a typical example are big files that are necessary to restart a calculation from that are too big to retrieve but also too costly to lose and have to recompute there is a simple and naive solution of just running an inline calculation that will open transport to the remote computer and use copy from remote to remote to copy the files and create a remotefolder object to attach to the computer but this can be dangerous if an exception occurs the entire process most likely the workchain that launched the jobcalculation will except as well of course one can catch the exceptions but then the files will not be copied the jobprocess infrastructure has a rich architecture to deal with tasks that require transport and can fail the best option would be to leverage this functionality for this particular purpose by creating an optional transporttask to copy files after the job has finished but before going over to the retrieval and parse step
1
6,086
8,948,425,063
IssuesEvent
2019-01-25 02:19:44
bisq-network/bisq
https://api.github.com/repos/bisq-network/bisq
closed
Add new P2P message in trade process to report errors to the peer
in:trade-process was:dropped
Related to https://github.com/bisq-network/exchange/issues/1044 it would help to find out the cause of errors if the peer sends a P2P message to the trade peer in the trade process so he get informed why the peer stopped responding. Currently if an error occurs the peers stops the protocol but do not send any message to the peer.
1.0
Add new P2P message in trade process to report errors to the peer - Related to https://github.com/bisq-network/exchange/issues/1044 it would help to find out the cause of errors if the peer sends a P2P message to the trade peer in the trade process so he get informed why the peer stopped responding. Currently if an error occurs the peers stops the protocol but do not send any message to the peer.
process
add new message in trade process to report errors to the peer related to it would help to find out the cause of errors if the peer sends a message to the trade peer in the trade process so he get informed why the peer stopped responding currently if an error occurs the peers stops the protocol but do not send any message to the peer
1
77,383
21,783,728,326
IssuesEvent
2022-05-13 22:29:19
dotnet/arcade
https://api.github.com/repos/dotnet/arcade
closed
Build failed: dotnet-arcade-validation-official/main #20220511.6
Build Failed
Build [#20220511.6](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=1766460) partiallySucceeded ## :warning: : internal / dotnet-arcade-validation-official partiallySucceeded ### Summary **Finished** - Thu, 12 May 2022 17:08:41 GMT **Duration** - 3 minutes **Requested for** - DotNet Bot **Reason** - batchedCI ### Details #### Promote Arcade to '.NET Eng - Latest' channel - :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/1766460/logs/410) - Artifact Promote_Arcade_To_Latest_Promote_Arcade_to_'.NET_Eng_-_Latest'_channel_SBOM already exists for build 1766460. ### Changes - [ac86b36b](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/ac86b36bdb46a65e1e5ec35b40468d81b4616b47) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220511.16 (#3111)
1.0
Build failed: dotnet-arcade-validation-official/main #20220511.6 - Build [#20220511.6](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=1766460) partiallySucceeded ## :warning: : internal / dotnet-arcade-validation-official partiallySucceeded ### Summary **Finished** - Thu, 12 May 2022 17:08:41 GMT **Duration** - 3 minutes **Requested for** - DotNet Bot **Reason** - batchedCI ### Details #### Promote Arcade to '.NET Eng - Latest' channel - :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/1766460/logs/410) - Artifact Promote_Arcade_To_Latest_Promote_Arcade_to_'.NET_Eng_-_Latest'_channel_SBOM already exists for build 1766460. ### Changes - [ac86b36b](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/ac86b36bdb46a65e1e5ec35b40468d81b4616b47) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220511.16 (#3111)
non_process
build failed dotnet arcade validation official main build partiallysucceeded warning internal dotnet arcade validation official partiallysucceeded summary finished thu may gmt duration minutes requested for dotnet bot reason batchedci details promote arcade to net eng latest channel x artifact promote arcade to latest promote arcade to net eng latest channel sbom already exists for build changes dotnet maestro update dependencies from build
0
220,350
24,564,929,014
IssuesEvent
2022-10-13 01:26:35
RG4421/skyux-sdk-builder
https://api.github.com/repos/RG4421/skyux-sdk-builder
opened
CVE-2022-37599 (Medium) detected in multiple libraries
security vulnerability
## CVE-2022-37599 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>loader-utils-2.0.0.tgz</b>, <b>loader-utils-1.4.0.tgz</b>, <b>loader-utils-1.2.3.tgz</b></p></summary> <p> <details><summary><b>loader-utils-2.0.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - :x: **loader-utils-2.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>loader-utils-1.4.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/sass-loader/node_modules/loader-utils/package.json,/node_modules/webpack/node_modules/loader-utils/package.json,/node_modules/istanbul-instrumenter-loader/node_modules/loader-utils/package.json,/node_modules/karma-webpack/node_modules/loader-utils/package.json,/node_modules/html-webpack-plugin/node_modules/loader-utils/package.json,/node_modules/raw-loader/node_modules/loader-utils/package.json,/node_modules/babel-loader/node_modules/loader-utils/package.json,/node_modules/awesome-typescript-loader/node_modules/loader-utils/package.json,/node_modules/css-loader/node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - babel-loader-8.1.0.tgz (Root Library) - :x: **loader-utils-1.4.0.tgz** (Vulnerable Library) </details> <details><summary><b>loader-utils-1.2.3.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.2.3.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.2.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/webpack-cli/node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - webpack-cli-3.3.11.tgz (Root Library) - :x: **loader-utils-1.2.3.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the resourcePath variable in interpolateName.js. <p>Publish Date: 2022-10-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599>CVE-2022-37599</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
True
CVE-2022-37599 (Medium) detected in multiple libraries - ## CVE-2022-37599 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>loader-utils-2.0.0.tgz</b>, <b>loader-utils-1.4.0.tgz</b>, <b>loader-utils-1.2.3.tgz</b></p></summary> <p> <details><summary><b>loader-utils-2.0.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - :x: **loader-utils-2.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>loader-utils-1.4.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/sass-loader/node_modules/loader-utils/package.json,/node_modules/webpack/node_modules/loader-utils/package.json,/node_modules/istanbul-instrumenter-loader/node_modules/loader-utils/package.json,/node_modules/karma-webpack/node_modules/loader-utils/package.json,/node_modules/html-webpack-plugin/node_modules/loader-utils/package.json,/node_modules/raw-loader/node_modules/loader-utils/package.json,/node_modules/babel-loader/node_modules/loader-utils/package.json,/node_modules/awesome-typescript-loader/node_modules/loader-utils/package.json,/node_modules/css-loader/node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - babel-loader-8.1.0.tgz (Root Library) - :x: **loader-utils-1.4.0.tgz** (Vulnerable Library) </details> <details><summary><b>loader-utils-1.2.3.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.2.3.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.2.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/webpack-cli/node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - webpack-cli-3.3.11.tgz (Root Library) - :x: **loader-utils-1.2.3.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the resourcePath variable in interpolateName.js. <p>Publish Date: 2022-10-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599>CVE-2022-37599</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
non_process
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries loader utils tgz loader utils tgz loader utils tgz loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules loader utils package json dependency hierarchy x loader utils tgz vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules sass loader node modules loader utils package json node modules webpack node modules loader utils package json node modules istanbul instrumenter loader node modules loader utils package json node modules karma webpack node modules loader utils package json node modules html webpack plugin node modules loader utils package json node modules raw loader node modules loader utils package json node modules babel loader node modules loader utils package json node modules awesome typescript loader node modules loader utils package json node modules css loader node modules loader utils package json dependency hierarchy babel loader tgz root library x loader utils tgz vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules webpack cli node modules loader utils package json dependency hierarchy webpack cli tgz root library x loader utils tgz vulnerable library found in base branch master vulnerability details a regular expression denial of service redos flaw was found in function interpolatename in interpolatename js in webpack loader utils via the resourcepath variable in interpolatename js publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href
0
10,727
13,530,333,370
IssuesEvent
2020-09-15 19:44:43
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Duplicate join generated when (manually) joining a table with a remapped field
.Query Language (MBQL) Priority:P2 Querying/Processor Type:Bug
We join twice when a remapping is defined. To repo: 1) Define a remapping Orders/prodcut_id -> Products/Category 2) create a question in QB joining Orders and Products Resulting MBQL has two joins to Orders instead of one.
1.0
Duplicate join generated when (manually) joining a table with a remapped field - We join twice when a remapping is defined. To repo: 1) Define a remapping Orders/prodcut_id -> Products/Category 2) create a question in QB joining Orders and Products Resulting MBQL has two joins to Orders instead of one.
process
duplicate join generated when manually joining a table with a remapped field we join twice when a remapping is defined to repo define a remapping orders prodcut id products category create a question in qb joining orders and products resulting mbql has two joins to orders instead of one
1
16,566
21,578,365,259
IssuesEvent
2022-05-02 15:57:12
cypress-io/cypress-documentation
https://api.github.com/repos/cypress-io/cypress-documentation
opened
Investigate Spell Checker
process: internal docs
### Subject Docs Internal ### Description Spell checking could be part of a lint process to cut down on manual reviewing and editing of basic spelling errors. @elylucas
1.0
Investigate Spell Checker - ### Subject Docs Internal ### Description Spell checking could be part of a lint process to cut down on manual reviewing and editing of basic spelling errors. @elylucas
process
investigate spell checker subject docs internal description spell checking could be part of a lint process to cut down on manual reviewing and editing of basic spelling errors elylucas
1
16,435
21,316,757,301
IssuesEvent
2022-04-16 12:16:22
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Common index preprocess
feature priority/medium preprocess index stale
## Description Add preprocessing module that collects index terms and groups them. The grouping should be extendable so plugins can contribute grouping code to handle e.g. additional languages.
1.0
Common index preprocess - ## Description Add preprocessing module that collects index terms and groups them. The grouping should be extendable so plugins can contribute grouping code to handle e.g. additional languages.
process
common index preprocess description add preprocessing module that collects index terms and groups them the grouping should be extendable so plugins can contribute grouping code to handle e g additional languages
1
30,569
6,176,278,319
IssuesEvent
2017-07-01 12:04:07
larcenists/larceny
https://api.github.com/repos/larcenists/larceny
closed
case sensitivity modes when loading file via command line may surprise users
C: ERR5RS P: trivial R: fixed T: defect
**Reported by: pnkfelix on Mon Jun 30 16:39:04 2008** While running the heap building script for Larceny, I ran into some surprising undefined global variable errors when some of the FFI files had been updated (and thus were being loaded from source) while others were still being loaded from fasl files. The errors looked like this: ``` % ./larceny.bin -stopcopy -- src/Build/iasn-larceny-heap.sch ... Warning: loading source in favor of stale fasl file: /Users/pnkfelix/larcenytest/larceny-gcbench/larceny_src/lib/Ffi/ffi-i386.sch Error: Undefined global variable "ffi/i386-C-callout-cdecl". Entering debugger; type "?" for help. debug> ``` The heap building script is run on top of the bootstrap heap, usually via the invocation `larceny.bin -stopcopy` followed by loading the heap build script via a command line parameter. There were a couple of things that surprised me: - loading the heap building script via the REPL worked, while loading the script via the command line broke (as illustrated at the top of the ticket). - my initial attempts to make a small test case to expose the problem failed, because I tried to replicate the problem using the full fledged larceny.heap, rather than the bootstrap heap. At this point, I believe I understand where the behavior is coming from (I think its from the fact that [source:trunk/larceny_src/src/Lib/Common/main.sch main.sch] adjusts the case sensitivity '''after''' invoking `failsafe-process-arguments`). Here is a transcript of interactions with a larceny executable that illustrate the weirdness. (The fourth load of `foo.sch` is the one that I am worried about users being confused by.) ``` % cat foo.sch (define Hello-World 'hELLO-wORLD!) (display Hello-World) (newline) (exit) % ./larceny Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified) larceny.heap, built on Mon Jun 30 16:36:29 EDT 2008 > (load "foo.sch") hELLO-wORLD! % ./larceny -- foo.sch Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified) hELLO-wORLD! % ./larceny.bin LARCENY_ROOT not set; using current directory Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified) > (load "foo.sch") hELLO-wORLD! % ./larceny.bin -- foo.sch LARCENY_ROOT not set; using current directory Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified) hello-world! % ``` This is not a huge issue, since I do not expect the average user to be doing anything starting from the bootstrap heap; if the system starts from the larceny.heap, then it starts with that heap's case sensitivity mode when loading files via the command line arguments. But it is an issue for the heap building scripts like [source:trunk/larceny_src/src/Build/iasn-larceny-heap.sch iasn-larceny-heap.sch] or [source:trunk/larceny_src/src/Build/sparc-larceny-heap.sch sparc-larceny-heap.sch]. (After all, what sparked the problem for me initially were issues with the FFI .sch files being loaded in case-insensitive mode while the other FFI .sch files had been compiled in case-sensitive mode.) I am currently planning to work around this problem by adding `no-fold-case` directives to the FFI source code, so that those files will continue to define their identifiers in a case-sensitive manner even when loaded from source. (It would also guard against the files being mis-compiled by a compiler that happens to be running in case-insensitive mode.) So that will be a band-aid for the problem with respect to the FFI source code. (Another fix would be to remove all case-sensitive identifiers from our libraries, but that is a questionable strategy for fixing this sort of thing.) But we (that is, Will) may want to consider moving the invocation of `adjust-case-sensitivity!` above the loads of the init files and command line argument processing. (Additionally, the way things are structured now, a user's attempt to change the default case-sensitivity in a personal init file will be foiled by the subsequent invocation of `adjust-case-sensitivity!`)
1.0
case sensitivity modes when loading file via command line may surprise users - **Reported by: pnkfelix on Mon Jun 30 16:39:04 2008** While running the heap building script for Larceny, I ran into some surprising undefined global variable errors when some of the FFI files had been updated (and thus were being loaded from source) while others were still being loaded from fasl files. The errors looked like this: ``` % ./larceny.bin -stopcopy -- src/Build/iasn-larceny-heap.sch ... Warning: loading source in favor of stale fasl file: /Users/pnkfelix/larcenytest/larceny-gcbench/larceny_src/lib/Ffi/ffi-i386.sch Error: Undefined global variable "ffi/i386-C-callout-cdecl". Entering debugger; type "?" for help. debug> ``` The heap building script is run on top of the bootstrap heap, usually via the invocation `larceny.bin -stopcopy` followed by loading the heap build script via a command line parameter. There were a couple of things that surprised me: - loading the heap building script via the REPL worked, while loading the script via the command line broke (as illustrated at the top of the ticket). - my initial attempts to make a small test case to expose the problem failed, because I tried to replicate the problem using the full fledged larceny.heap, rather than the bootstrap heap. At this point, I believe I understand where the behavior is coming from (I think its from the fact that [source:trunk/larceny_src/src/Lib/Common/main.sch main.sch] adjusts the case sensitivity '''after''' invoking `failsafe-process-arguments`). Here is a transcript of interactions with a larceny executable that illustrate the weirdness. (The fourth load of `foo.sch` is the one that I am worried about users being confused by.) ``` % cat foo.sch (define Hello-World 'hELLO-wORLD!) (display Hello-World) (newline) (exit) % ./larceny Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified) larceny.heap, built on Mon Jun 30 16:36:29 EDT 2008 > (load "foo.sch") hELLO-wORLD! % ./larceny -- foo.sch Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified) hELLO-wORLD! % ./larceny.bin LARCENY_ROOT not set; using current directory Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified) > (load "foo.sch") hELLO-wORLD! % ./larceny.bin -- foo.sch LARCENY_ROOT not set; using current directory Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified) hello-world! % ``` This is not a huge issue, since I do not expect the average user to be doing anything starting from the bootstrap heap; if the system starts from the larceny.heap, then it starts with that heap's case sensitivity mode when loading files via the command line arguments. But it is an issue for the heap building scripts like [source:trunk/larceny_src/src/Build/iasn-larceny-heap.sch iasn-larceny-heap.sch] or [source:trunk/larceny_src/src/Build/sparc-larceny-heap.sch sparc-larceny-heap.sch]. (After all, what sparked the problem for me initially were issues with the FFI .sch files being loaded in case-insensitive mode while the other FFI .sch files had been compiled in case-sensitive mode.) I am currently planning to work around this problem by adding `no-fold-case` directives to the FFI source code, so that those files will continue to define their identifiers in a case-sensitive manner even when loaded from source. (It would also guard against the files being mis-compiled by a compiler that happens to be running in case-insensitive mode.) So that will be a band-aid for the problem with respect to the FFI source code. (Another fix would be to remove all case-sensitive identifiers from our libraries, but that is a questionable strategy for fixing this sort of thing.) But we (that is, Will) may want to consider moving the invocation of `adjust-case-sensitivity!` above the loads of the init files and command line argument processing. (Additionally, the way things are structured now, a user's attempt to change the default case-sensitivity in a personal init file will be foiled by the subsequent invocation of `adjust-case-sensitivity!`)
non_process
case sensitivity modes when loading file via command line may surprise users reported by pnkfelix on mon jun while running the heap building script for larceny i ran into some surprising undefined global variable errors when some of the ffi files had been updated and thus were being loaded from source while others were still being loaded from fasl files the errors looked like this larceny bin stopcopy src build iasn larceny heap sch warning loading source in favor of stale fasl file users pnkfelix larcenytest larceny gcbench larceny src lib ffi ffi sch error undefined global variable ffi c callout cdecl entering debugger type for help debug the heap building script is run on top of the bootstrap heap usually via the invocation larceny bin stopcopy followed by loading the heap build script via a command line parameter there were a couple of things that surprised me loading the heap building script via the repl worked while loading the script via the command line broke as illustrated at the top of the ticket my initial attempts to make a small test case to expose the problem failed because i tried to replicate the problem using the full fledged larceny heap rather than the bootstrap heap at this point i believe i understand where the behavior is coming from i think its from the fact that adjusts the case sensitivity after invoking failsafe process arguments here is a transcript of interactions with a larceny executable that illustrate the weirdness the fourth load of foo sch is the one that i am worried about users being confused by cat foo sch define hello world hello world display hello world newline exit larceny larceny fluoridation jun precise posix unix unified larceny heap built on mon jun edt load foo sch hello world larceny foo sch larceny fluoridation jun precise posix unix unified hello world larceny bin larceny root not set using current directory larceny fluoridation jun precise posix unix unified load foo sch hello world larceny bin foo sch larceny root not set using current directory larceny fluoridation jun precise posix unix unified hello world this is not a huge issue since i do not expect the average user to be doing anything starting from the bootstrap heap if the system starts from the larceny heap then it starts with that heap s case sensitivity mode when loading files via the command line arguments but it is an issue for the heap building scripts like or after all what sparked the problem for me initially were issues with the ffi sch files being loaded in case insensitive mode while the other ffi sch files had been compiled in case sensitive mode i am currently planning to work around this problem by adding no fold case directives to the ffi source code so that those files will continue to define their identifiers in a case sensitive manner even when loaded from source it would also guard against the files being mis compiled by a compiler that happens to be running in case insensitive mode so that will be a band aid for the problem with respect to the ffi source code another fix would be to remove all case sensitive identifiers from our libraries but that is a questionable strategy for fixing this sort of thing but we that is will may want to consider moving the invocation of adjust case sensitivity above the loads of the init files and command line argument processing additionally the way things are structured now a user s attempt to change the default case sensitivity in a personal init file will be foiled by the subsequent invocation of adjust case sensitivity
0
180,047
13,916,711,303
IssuesEvent
2020-10-21 04:06:21
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Incorrect SELinux labels for kubelet artifacts for node tests on Fedora CoreOS
kind/failing-test sig/node
<!-- Please only use this template for submitting reports about continuously failing tests or jobs in Kubernetes CI --> **Which jobs are failing**: pull-kubernetes-node-crio-e2e **Which test(s) are failing**: N/A **Since when has it been failing**: N/A **Testgrid link**: https://storage.googleapis.com/kubernetes-jenkins/pr-logs/pull/93486/pull-kubernetes-node-crio-e2e/1308737597702410240/artifacts/tmp-node-e2e-2f579b16-fedora-coreos-32-20200907-3-0-gcp-x86-64-system.log **Reason for failure**: ``` Sep 23 12:20:21 localhost systemd[1]: Started /tmp/node-e2e-20200923T121804/kubelet --kubeconfig /tmp/node-e2e-20200923T121804/kubeconfig --root-dir /var/lib/kubelet --v 4 --logtostderr --dynamic-config-dir /tmp/node-e2e-20200923T121804/dynamic-kubelet-config --network-plugin=kubenet --cni-bin-dir /tmp/node-e2e-20200923T121804/cni/bin --cni-conf-dir /tmp/node-e2e-20200923T121804/cni/net.d --cni-cache-dir /tmp/node-e2e-20200923T121804/cni/cache --hostname-override tmp-node-e2e-2f579b16-fedora-coreos-32-20200907-3-0-gcp-x86-64 --container-runtime remote --container-runtime-endpoint unix:///run/crio/crio.sock --config /tmp/node-e2e-20200923T121804/kubelet-config --cgroups-per-qos=true --cgroup-root=/. Sep 23 12:20:21 localhost audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=kubelet-20200923T121804 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 23 12:20:21 localhost systemd[4212]: kubelet-20200923T121804.service: Failed to execute command: Permission denied Sep 23 12:20:21 localhost systemd[4212]: kubelet-20200923T121804.service: Failed at step EXEC spawning /tmp/node-e2e-20200923T121804/kubelet: Permission denied Sep 23 12:20:21 localhost audit[4212]: AVC avc: denied { execute } for pid=4212 comm="(kubelet)" name="kubelet" dev="tmpfs" ino=62327 scontext=system_u:system_r:init_t:s0 tcontext=unconfined_u:object_r:user_tmp_t:s0 tclass=file permissive=0 Sep 23 12:20:21 localhost systemd[1]: kubelet-20200923T121804.service: Main process exited, code=exited, status=203/EXEC Sep 23 12:20:21 localhost systemd[1]: kubelet-20200923T121804.service: Failed with result 'exit-code'. Sep 23 12:20:21 localhost audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=kubelet-20200923T121804 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' ``` **Anything else we need to know**:
1.0
Incorrect SELinux labels for kubelet artifacts for node tests on Fedora CoreOS - <!-- Please only use this template for submitting reports about continuously failing tests or jobs in Kubernetes CI --> **Which jobs are failing**: pull-kubernetes-node-crio-e2e **Which test(s) are failing**: N/A **Since when has it been failing**: N/A **Testgrid link**: https://storage.googleapis.com/kubernetes-jenkins/pr-logs/pull/93486/pull-kubernetes-node-crio-e2e/1308737597702410240/artifacts/tmp-node-e2e-2f579b16-fedora-coreos-32-20200907-3-0-gcp-x86-64-system.log **Reason for failure**: ``` Sep 23 12:20:21 localhost systemd[1]: Started /tmp/node-e2e-20200923T121804/kubelet --kubeconfig /tmp/node-e2e-20200923T121804/kubeconfig --root-dir /var/lib/kubelet --v 4 --logtostderr --dynamic-config-dir /tmp/node-e2e-20200923T121804/dynamic-kubelet-config --network-plugin=kubenet --cni-bin-dir /tmp/node-e2e-20200923T121804/cni/bin --cni-conf-dir /tmp/node-e2e-20200923T121804/cni/net.d --cni-cache-dir /tmp/node-e2e-20200923T121804/cni/cache --hostname-override tmp-node-e2e-2f579b16-fedora-coreos-32-20200907-3-0-gcp-x86-64 --container-runtime remote --container-runtime-endpoint unix:///run/crio/crio.sock --config /tmp/node-e2e-20200923T121804/kubelet-config --cgroups-per-qos=true --cgroup-root=/. Sep 23 12:20:21 localhost audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=kubelet-20200923T121804 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 23 12:20:21 localhost systemd[4212]: kubelet-20200923T121804.service: Failed to execute command: Permission denied Sep 23 12:20:21 localhost systemd[4212]: kubelet-20200923T121804.service: Failed at step EXEC spawning /tmp/node-e2e-20200923T121804/kubelet: Permission denied Sep 23 12:20:21 localhost audit[4212]: AVC avc: denied { execute } for pid=4212 comm="(kubelet)" name="kubelet" dev="tmpfs" ino=62327 scontext=system_u:system_r:init_t:s0 tcontext=unconfined_u:object_r:user_tmp_t:s0 tclass=file permissive=0 Sep 23 12:20:21 localhost systemd[1]: kubelet-20200923T121804.service: Main process exited, code=exited, status=203/EXEC Sep 23 12:20:21 localhost systemd[1]: kubelet-20200923T121804.service: Failed with result 'exit-code'. Sep 23 12:20:21 localhost audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=kubelet-20200923T121804 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' ``` **Anything else we need to know**:
non_process
incorrect selinux labels for kubelet artifacts for node tests on fedora coreos which jobs are failing pull kubernetes node crio which test s are failing n a since when has it been failing n a testgrid link reason for failure sep localhost systemd started tmp node kubelet kubeconfig tmp node kubeconfig root dir var lib kubelet v logtostderr dynamic config dir tmp node dynamic kubelet config network plugin kubenet cni bin dir tmp node cni bin cni conf dir tmp node cni net d cni cache dir tmp node cni cache hostname override tmp node fedora coreos gcp container runtime remote container runtime endpoint unix run crio crio sock config tmp node kubelet config cgroups per qos true cgroup root sep localhost audit service start pid uid auid ses subj system u system r init t msg unit kubelet comm systemd exe usr lib systemd systemd hostname addr terminal res success sep localhost systemd kubelet service failed to execute command permission denied sep localhost systemd kubelet service failed at step exec spawning tmp node kubelet permission denied sep localhost audit avc avc denied execute for pid comm kubelet name kubelet dev tmpfs ino scontext system u system r init t tcontext unconfined u object r user tmp t tclass file permissive sep localhost systemd kubelet service main process exited code exited status exec sep localhost systemd kubelet service failed with result exit code sep localhost audit service stop pid uid auid ses subj system u system r init t msg unit kubelet comm systemd exe usr lib systemd systemd hostname addr terminal res failed anything else we need to know
0
4,764
7,631,987,777
IssuesEvent
2018-05-05 09:42:48
AmpersandTarski/Ampersand
https://api.github.com/repos/AmpersandTarski/Ampersand
closed
Cleanup copied templates folder in prototype output dir
component:prototype generator priority:low software process
After generating the prototype, this folder is not used anymore. This issue requests to delete the folder as final step in the prototype generator (e.g. after calling `composer ---install`) Prevents users wasting time in case they don't know the purpose of this folder.
1.0
Cleanup copied templates folder in prototype output dir - After generating the prototype, this folder is not used anymore. This issue requests to delete the folder as final step in the prototype generator (e.g. after calling `composer ---install`) Prevents users wasting time in case they don't know the purpose of this folder.
process
cleanup copied templates folder in prototype output dir after generating the prototype this folder is not used anymore this issue requests to delete the folder as final step in the prototype generator e g after calling composer install prevents users wasting time in case they don t know the purpose of this folder
1
250,754
7,987,224,593
IssuesEvent
2018-07-19 06:57:23
architecture-building-systems/CityEnergyAnalyst
https://api.github.com/repos/architecture-building-systems/CityEnergyAnalyst
opened
Bug in ARU air-conditioning model calculations
Priority 1 bug
@martin-mosteiro reported strange behavior of the air conditioning model for the case when there is dehumidification and the interior air temperature is below the cooling supply temperature. this happens in spring in a very poorly insulated building outside the heating season.
1.0
Bug in ARU air-conditioning model calculations - @martin-mosteiro reported strange behavior of the air conditioning model for the case when there is dehumidification and the interior air temperature is below the cooling supply temperature. this happens in spring in a very poorly insulated building outside the heating season.
non_process
bug in aru air conditioning model calculations martin mosteiro reported strange behavior of the air conditioning model for the case when there is dehumidification and the interior air temperature is below the cooling supply temperature this happens in spring in a very poorly insulated building outside the heating season
0
19,375
25,501,988,434
IssuesEvent
2022-11-28 05:33:48
prusa3d/Prusa-Firmware
https://api.github.com/repos/prusa3d/Prusa-Firmware
closed
Put version info in release filenames
enhancement processing
Minor nit, but makes life a bit easier. Instead of the files being named like `FW_MK3-EINSy10a.hex`, add the version in so we get something like `FW_3.7.0_MK3-EINSy10a.hex`. That should give a nice sort order even if you have multiple printers of different models.
1.0
Put version info in release filenames - Minor nit, but makes life a bit easier. Instead of the files being named like `FW_MK3-EINSy10a.hex`, add the version in so we get something like `FW_3.7.0_MK3-EINSy10a.hex`. That should give a nice sort order even if you have multiple printers of different models.
process
put version info in release filenames minor nit but makes life a bit easier instead of the files being named like fw hex add the version in so we get something like fw hex that should give a nice sort order even if you have multiple printers of different models
1
11,224
14,004,032,107
IssuesEvent
2020-10-28 16:35:41
retaildevcrews/ngsa
https://api.github.com/repos/retaildevcrews/ngsa
reopened
Checkpoint Review - November
CPR Process
This checklist is for verifying the necessary details for a checkpoint review. ## [Engagement Name] - [Month] Checkpoint Review - [Milestone/Sprint], [Month] - [Date] ### Planned vs. Done - List of high-level delivery plan from last month vs done log. _This typically would be described at the EPIC/Workstream level and provide a high-level view of progress vs plan. Note: Whenever possible link to artifacts from sources such as ADO vs pasting content from other documents_ | Epic / Workstream | Plan for the month | Status | Comments | | ------------- | ------------- | ------------- | ------------- | | Workstream1 | Y/N | Y/N | Y/N | Y/N | | Workstream2| Y/N | Y/N | Y/N | Y/N | | Workstream3| Y/N | Y/N | Y/N | Y/N | | Workstream4 | Y/N | Y/N | Y/N | Y/N | Question to add into talk track: - What was planned to be done this month? - What was actually done this month? - Changes in plan versus last - Key design decisions and their outcomes ### Work-Item Level _This is typically linked directly from the Agile Planning Tool (GITHUB, ADO, Jira). Note: Whenever possible link to artifacts from sources such as ADO vs pasting content from other documents_ | ID | Title | Work Item Type | Assigned To | State | Tags| | ------------- | ------------- | ------------- | ------------- | ------------- | ------------- | | | Workitem 1 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N | | | Workitem 2 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N | | | Workitem 3 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N | | | Workitem 4 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N | ### ML/AI Status | Topic | Status | Migration | | ------------- | ------------- | ------------- | | Responsible AI (RAIL compliance at current project phase) | Y/N | Y/N | | Evaluation Framework Readiness (metrics, test, data, tooling) | Y/N | Y/N | | Performance Evaulation (baseline, iterations | Y/N | Y/N | ### Key Changes / Decisions / Outcomes _This is a list of any decisions or architecture changes that have been made since the last checkpoint and the action plan for remediation. Note: Whenever possible link to artifacts from sources such as ADO vs pasting content. You can keep a decision log in your project repo or wiki to preserve this information for the customer._ | Key Decision / Architectural change | Remediation Plan | | ------------- | ------------- | | Key Decision 1 | Y/N | | Key Decision 2 | Y/N | | Key Decision 3 | Y/N | ## Engineering Fundamentals _The latest [Engineering Fundamentals Assessment Report](https://aka.ms/EngineeringFundamentals) for the engagement_ ### Retrospective Summary _Provide details from the retrospectives over the last month_ - What went well: - What didn't go well: - Action Items: ### Risks, Issues, Additional Asks _These items are copied from the initial Gameplan and are updated each month. NOTE: Whenever possible link to the artifacts from sources such as ADO vs pasting content from other documents, See Risk Tag example below_ | Risk / Issue | Mitigation | Status| | ------------- | ------------- | ------------- | | Risk / Issue 1 | Y/N | Y/N | | Risk / Issue 2 | Y/N | Y/N | | Risk / Issue 3 | Y/N | Y/N | _Include links to CSE Feedback and Feedback submitted to product groups as outlined in [CSE Engineering Feedback Playbook](https://cwcwiki.com/wiki/CSE_Engineering_Feedback_Playbook)_ ### Final Checkpoint _If this is the final checkpoint for an engagement, please also include a link to the [engagement wrap-up document ](https://aka.ms/wrapuptemplate)_ ### Tasks - [ ] Upload checkpoint review to Artifact Hub 24hrs in advance - [ ] Tech Lead to complete Engineering Assessment via Snowball
1.0
Checkpoint Review - November - This checklist is for verifying the necessary details for a checkpoint review. ## [Engagement Name] - [Month] Checkpoint Review - [Milestone/Sprint], [Month] - [Date] ### Planned vs. Done - List of high-level delivery plan from last month vs done log. _This typically would be described at the EPIC/Workstream level and provide a high-level view of progress vs plan. Note: Whenever possible link to artifacts from sources such as ADO vs pasting content from other documents_ | Epic / Workstream | Plan for the month | Status | Comments | | ------------- | ------------- | ------------- | ------------- | | Workstream1 | Y/N | Y/N | Y/N | Y/N | | Workstream2| Y/N | Y/N | Y/N | Y/N | | Workstream3| Y/N | Y/N | Y/N | Y/N | | Workstream4 | Y/N | Y/N | Y/N | Y/N | Question to add into talk track: - What was planned to be done this month? - What was actually done this month? - Changes in plan versus last - Key design decisions and their outcomes ### Work-Item Level _This is typically linked directly from the Agile Planning Tool (GITHUB, ADO, Jira). Note: Whenever possible link to artifacts from sources such as ADO vs pasting content from other documents_ | ID | Title | Work Item Type | Assigned To | State | Tags| | ------------- | ------------- | ------------- | ------------- | ------------- | ------------- | | | Workitem 1 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N | | | Workitem 2 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N | | | Workitem 3 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N | | | Workitem 4 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N | ### ML/AI Status | Topic | Status | Migration | | ------------- | ------------- | ------------- | | Responsible AI (RAIL compliance at current project phase) | Y/N | Y/N | | Evaluation Framework Readiness (metrics, test, data, tooling) | Y/N | Y/N | | Performance Evaulation (baseline, iterations | Y/N | Y/N | ### Key Changes / Decisions / Outcomes _This is a list of any decisions or architecture changes that have been made since the last checkpoint and the action plan for remediation. Note: Whenever possible link to artifacts from sources such as ADO vs pasting content. You can keep a decision log in your project repo or wiki to preserve this information for the customer._ | Key Decision / Architectural change | Remediation Plan | | ------------- | ------------- | | Key Decision 1 | Y/N | | Key Decision 2 | Y/N | | Key Decision 3 | Y/N | ## Engineering Fundamentals _The latest [Engineering Fundamentals Assessment Report](https://aka.ms/EngineeringFundamentals) for the engagement_ ### Retrospective Summary _Provide details from the retrospectives over the last month_ - What went well: - What didn't go well: - Action Items: ### Risks, Issues, Additional Asks _These items are copied from the initial Gameplan and are updated each month. NOTE: Whenever possible link to the artifacts from sources such as ADO vs pasting content from other documents, See Risk Tag example below_ | Risk / Issue | Mitigation | Status| | ------------- | ------------- | ------------- | | Risk / Issue 1 | Y/N | Y/N | | Risk / Issue 2 | Y/N | Y/N | | Risk / Issue 3 | Y/N | Y/N | _Include links to CSE Feedback and Feedback submitted to product groups as outlined in [CSE Engineering Feedback Playbook](https://cwcwiki.com/wiki/CSE_Engineering_Feedback_Playbook)_ ### Final Checkpoint _If this is the final checkpoint for an engagement, please also include a link to the [engagement wrap-up document ](https://aka.ms/wrapuptemplate)_ ### Tasks - [ ] Upload checkpoint review to Artifact Hub 24hrs in advance - [ ] Tech Lead to complete Engineering Assessment via Snowball
process
checkpoint review november this checklist is for verifying the necessary details for a checkpoint review checkpoint review planned vs done list of high level delivery plan from last month vs done log this typically would be described at the epic workstream level and provide a high level view of progress vs plan note whenever possible link to artifacts from sources such as ado vs pasting content from other documents epic workstream plan for the month status comments y n y n y n y n y n y n y n y n y n y n y n y n y n y n y n y n question to add into talk track what was planned to be done this month what was actually done this month changes in plan versus last key design decisions and their outcomes work item level this is typically linked directly from the agile planning tool github ado jira note whenever possible link to artifacts from sources such as ado vs pasting content from other documents id title work item type assigned to state tags workitem y n y n y n y n y n y n workitem y n y n y n y n y n y n workitem y n y n y n y n y n y n workitem y n y n y n y n y n y n ml ai status topic status migration responsible ai rail compliance at current project phase y n y n evaluation framework readiness metrics test data tooling y n y n performance evaulation baseline iterations y n y n key changes decisions outcomes this is a list of any decisions or architecture changes that have been made since the last checkpoint and the action plan for remediation note whenever possible link to artifacts from sources such as ado vs pasting content you can keep a decision log in your project repo or wiki to preserve this information for the customer key decision architectural change remediation plan key decision y n key decision y n key decision y n engineering fundamentals the latest for the engagement retrospective summary provide details from the retrospectives over the last month what went well what didn t go well action items risks issues additional asks these items are copied from the initial gameplan and are updated each month note whenever possible link to the artifacts from sources such as ado vs pasting content from other documents see risk tag example below risk issue mitigation status risk issue y n y n risk issue y n y n risk issue y n y n include links to cse feedback and feedback submitted to product groups as outlined in final checkpoint if this is the final checkpoint for an engagement please also include a link to the tasks upload checkpoint review to artifact hub in advance tech lead to complete engineering assessment via snowball
1
1,989
4,816,895,086
IssuesEvent
2016-11-04 11:45:14
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Request for unifying shell command form when using exec and spawn
child_process
* **Version / Platform**: v4.5.0 / MacOS <!-- Enter your issue details below this comment. --> Based on this site, there is a difference when calling an external script using `exec` and `spawn` methods. https://nodejs.org/api/child_process.html#child_process_child_process_exec_command_options_callback One takes a string with space separated arguments and the second one takes a command in the first parameter and an array containing arguments in the second parameter. This is a real pain when we want to call multiple chained commands (for example: `clean dist && test --all && publish`) and stream their output.
1.0
Request for unifying shell command form when using exec and spawn - * **Version / Platform**: v4.5.0 / MacOS <!-- Enter your issue details below this comment. --> Based on this site, there is a difference when calling an external script using `exec` and `spawn` methods. https://nodejs.org/api/child_process.html#child_process_child_process_exec_command_options_callback One takes a string with space separated arguments and the second one takes a command in the first parameter and an array containing arguments in the second parameter. This is a real pain when we want to call multiple chained commands (for example: `clean dist && test --all && publish`) and stream their output.
process
request for unifying shell command form when using exec and spawn version platform macos based on this site there is a difference when calling an external script using exec and spawn methods one takes a string with space separated arguments and the second one takes a command in the first parameter and an array containing arguments in the second parameter this is a real pain when we want to call multiple chained commands for example clean dist test all publish and stream their output
1
24
2,496,270,287
IssuesEvent
2015-01-06 18:18:04
vivo-isf/vivo-isf-ontology
https://api.github.com/repos/vivo-isf/vivo-isf-ontology
closed
Courtship
biological_process imported
_From [fcold...@eagle-i.org](https://code.google.com/u/113677139039624182507/) on May 21, 2013 07:50:44_ \<b>**** Use the form below to request a new term ****</b> \<b>**** Scroll down to see a term request example ****</b> &#13; \<b>Please indicate the label for the proposed term:</b> Courtship&#13; &#13; \<b>Please provide a textual definition (with source):</b> "Courtship can be defined as behavioral interactions between males and females, the evolutionary objective of which is copulation and the ultimate perpetuation of the species."&#13; &#13; \<a href="http://cshprotocols.cshlp.org/content/2012/11/pdb.prot071647.long" rel="nofollow">http://cshprotocols.cshlp.org/content/2012/11/pdb.prot071647.long</a>&#13; &#13; \<b>Please add an example of usage for proposed term:</b> Term is to be used as a "Biological Process Studied" for a model organism&#13; &#13; \<b>Please provide any additional optional information below. (e.g. desired</b> \<b>asserted SuperClass in ERO hierarchy or Reference Branch)</b> &#13; \<b>[ ] Instrument</b> [X] Biological process&#13; \<b>[ ] Disease</b> \<b>[ ] Human studies</b> \<b>[ ] Instrument</b> \<b>[ ] Organism</b> \<b>[ ] Reagent</b> \<b>[ ] Software</b> \<b>[ ] Technique</b> \<b>[ ] Organization</b> &#13; \<b>Additional info:</b> &#13; &#13; &#13; &#13; \<b>*** Term request example ****</b> &#13; \<b>Please indicate the label for the proposed term: four-terminal resistance</b> \<b>sensor</b> &#13; &#13; Please provide a textual definition (with source): "Four-terminal&#13; \<b>resistance sensors are electrical impedance measuring instruments that use</b> \<b>separate pairs of current-carrying and voltage-sensing electrodes to make</b> \<b>accurate measurements that can be used to compute a material's electrical</b> resistance." \<a href="http://en.wikipedia.org/wiki/Four-terminal_sensing" rel="nofollow">http://en.wikipedia.org/wiki/Four-terminal_sensing</a>&#13; &#13; &#13; \<b>Please add an example of usage for proposed term: Measuring the inherent</b> \<b>(per square) resistance of doped silicon.</b> &#13; &#13; \<b>Please provide any additional optional information below. (e.g. desired</b> \<b>asserted SuperClass in ERO hierarchy or Reference Branch)</b> &#13; \<b>[X] Instrument</b> \<b>[ ] Biological process</b> \<b>[ ] Disease</b> \<b>[ ] Human studies</b> \<b>[ ] Instrument</b> \<b>[ ] Organism</b> \<b>[ ] Reagent</b> \<b>[ ] Software</b> \<b>[ ] Technique</b> \<b>[ ] Organization</b> &#13; \<b>Additional info: AKA - 4T sensors, 4-wire sensor, or 4-point probe</b> _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=217_
1.0
Courtship - _From [fcold...@eagle-i.org](https://code.google.com/u/113677139039624182507/) on May 21, 2013 07:50:44_ \<b>**** Use the form below to request a new term ****</b> \<b>**** Scroll down to see a term request example ****</b> &#13; \<b>Please indicate the label for the proposed term:</b> Courtship&#13; &#13; \<b>Please provide a textual definition (with source):</b> "Courtship can be defined as behavioral interactions between males and females, the evolutionary objective of which is copulation and the ultimate perpetuation of the species."&#13; &#13; \<a href="http://cshprotocols.cshlp.org/content/2012/11/pdb.prot071647.long" rel="nofollow">http://cshprotocols.cshlp.org/content/2012/11/pdb.prot071647.long</a>&#13; &#13; \<b>Please add an example of usage for proposed term:</b> Term is to be used as a "Biological Process Studied" for a model organism&#13; &#13; \<b>Please provide any additional optional information below. (e.g. desired</b> \<b>asserted SuperClass in ERO hierarchy or Reference Branch)</b> &#13; \<b>[ ] Instrument</b> [X] Biological process&#13; \<b>[ ] Disease</b> \<b>[ ] Human studies</b> \<b>[ ] Instrument</b> \<b>[ ] Organism</b> \<b>[ ] Reagent</b> \<b>[ ] Software</b> \<b>[ ] Technique</b> \<b>[ ] Organization</b> &#13; \<b>Additional info:</b> &#13; &#13; &#13; &#13; \<b>*** Term request example ****</b> &#13; \<b>Please indicate the label for the proposed term: four-terminal resistance</b> \<b>sensor</b> &#13; &#13; Please provide a textual definition (with source): "Four-terminal&#13; \<b>resistance sensors are electrical impedance measuring instruments that use</b> \<b>separate pairs of current-carrying and voltage-sensing electrodes to make</b> \<b>accurate measurements that can be used to compute a material's electrical</b> resistance." \<a href="http://en.wikipedia.org/wiki/Four-terminal_sensing" rel="nofollow">http://en.wikipedia.org/wiki/Four-terminal_sensing</a>&#13; &#13; &#13; \<b>Please add an example of usage for proposed term: Measuring the inherent</b> \<b>(per square) resistance of doped silicon.</b> &#13; &#13; \<b>Please provide any additional optional information below. (e.g. desired</b> \<b>asserted SuperClass in ERO hierarchy or Reference Branch)</b> &#13; \<b>[X] Instrument</b> \<b>[ ] Biological process</b> \<b>[ ] Disease</b> \<b>[ ] Human studies</b> \<b>[ ] Instrument</b> \<b>[ ] Organism</b> \<b>[ ] Reagent</b> \<b>[ ] Software</b> \<b>[ ] Technique</b> \<b>[ ] Organization</b> &#13; \<b>Additional info: AKA - 4T sensors, 4-wire sensor, or 4-point probe</b> _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=217_
process
courtship from on may use the form below to request a new term scroll down to see a term request example please indicate the label for the proposed term courtship please provide a textual definition with source courtship can be defined as behavioral interactions between males and females the evolutionary objective of which is copulation and the ultimate perpetuation of the species please add an example of usage for proposed term term is to be used as a biological process studied for a model organism please provide any additional optional information below e g desired asserted superclass in ero hierarchy or reference branch instrument biological process disease human studies instrument organism reagent software technique organization additional info term request example please indicate the label for the proposed term four terminal resistance sensor please provide a textual definition with source four terminal resistance sensors are electrical impedance measuring instruments that use separate pairs of current carrying and voltage sensing electrodes to make accurate measurements that can be used to compute a material s electrical resistance please add an example of usage for proposed term measuring the inherent per square resistance of doped silicon please provide any additional optional information below e g desired asserted superclass in ero hierarchy or reference branch instrument biological process disease human studies instrument organism reagent software technique organization additional info aka sensors wire sensor or point probe original issue
1
4,863
7,746,917,344
IssuesEvent
2018-05-30 00:03:57
AppFolioOnboarding/image-sharer-ChaoHuangAtAppfolio
https://api.github.com/repos/AppFolioOnboarding/image-sharer-ChaoHuangAtAppfolio
closed
Image Seeds
in process
#### As a developer I want at least 20 images to be created when I setup the database. __Story__: Working on this application has been a lot of fun, right? You are fast becoming a git expert, and an expert at working with a deployment pipeline. Speaking of development and deployment pipelines, is it not pretty annoying that everytime a new deployment is created it starts out with an empty database? It seems like everytime a review application is created we need to manually add a handful of images in order to manually test the application. Would it not be great if each deployment was created with a handful of existing images? __Acceptance criteria__: - [ ] After running `rake db:setup` at least 20 unique images are shown in the homepage. - [ ] When a new app is deployed on heroku (e.g., every review app), that app automatically has at least 20 unique images on its homepage. __Dependencies__: - Image Index
1.0
Image Seeds - #### As a developer I want at least 20 images to be created when I setup the database. __Story__: Working on this application has been a lot of fun, right? You are fast becoming a git expert, and an expert at working with a deployment pipeline. Speaking of development and deployment pipelines, is it not pretty annoying that everytime a new deployment is created it starts out with an empty database? It seems like everytime a review application is created we need to manually add a handful of images in order to manually test the application. Would it not be great if each deployment was created with a handful of existing images? __Acceptance criteria__: - [ ] After running `rake db:setup` at least 20 unique images are shown in the homepage. - [ ] When a new app is deployed on heroku (e.g., every review app), that app automatically has at least 20 unique images on its homepage. __Dependencies__: - Image Index
process
image seeds as a developer i want at least images to be created when i setup the database story working on this application has been a lot of fun right you are fast becoming a git expert and an expert at working with a deployment pipeline speaking of development and deployment pipelines is it not pretty annoying that everytime a new deployment is created it starts out with an empty database it seems like everytime a review application is created we need to manually add a handful of images in order to manually test the application would it not be great if each deployment was created with a handful of existing images acceptance criteria after running rake db setup at least unique images are shown in the homepage when a new app is deployed on heroku e g every review app that app automatically has at least unique images on its homepage dependencies image index
1
16,541
21,566,803,155
IssuesEvent
2022-05-02 00:12:35
jmacost5/CPP-528-Project
https://api.github.com/repos/jmacost5/CPP-528-Project
closed
Commits have useful names and clear descriptions
TEAM PROCESS
Good example: 'Create .rds file that stores original and final predictive model' Bad example: 'Updated files' In general, good commits start with a present tense verb that summarizes your work in 50 character or less. If I can’t tell exactly what change/edit was made, neither will you six months from now.
1.0
Commits have useful names and clear descriptions - Good example: 'Create .rds file that stores original and final predictive model' Bad example: 'Updated files' In general, good commits start with a present tense verb that summarizes your work in 50 character or less. If I can’t tell exactly what change/edit was made, neither will you six months from now.
process
commits have useful names and clear descriptions good example create rds file that stores original and final predictive model bad example updated files in general good commits start with a present tense verb that summarizes your work in character or less if i can’t tell exactly what change edit was made neither will you six months from now
1
3,826
6,802,323,816
IssuesEvent
2017-11-02 19:47:35
WikiWatershed/model-my-watershed
https://api.github.com/repos/WikiWatershed/model-my-watershed
closed
Geoprocessing API: Validate Shape Does Not Self Intersect
Geoprocessing API tested/verified WPF
We already check for self-intersections on the frontend via `leaflet-draw` and `turf-kink`, but we now have to guard against bad input from API users. Add a check on AOIs submitted to the analyze endpoints.
1.0
Geoprocessing API: Validate Shape Does Not Self Intersect - We already check for self-intersections on the frontend via `leaflet-draw` and `turf-kink`, but we now have to guard against bad input from API users. Add a check on AOIs submitted to the analyze endpoints.
process
geoprocessing api validate shape does not self intersect we already check for self intersections on the frontend via leaflet draw and turf kink but we now have to guard against bad input from api users add a check on aois submitted to the analyze endpoints
1
258,715
27,578,011,620
IssuesEvent
2023-03-08 14:23:27
DEV-REPO-URIEL/TEST_AINAT_CHANGE
https://api.github.com/repos/DEV-REPO-URIEL/TEST_AINAT_CHANGE
opened
chart.js-2.9.3.tgz: 1 vulnerabilities (highest severity is: 7.5)
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>chart.js-2.9.3.tgz</b></p></summary> <p>Simple HTML5 charts using the canvas element.</p> <p>Library home page: <a href="https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz">https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/chart.js/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/DEV-REPO-URIEL/TEST_AINAT_CHANGE/commit/1855a834302dd3f0182c78b12ce583c7af3a2921">1855a834302dd3f0182c78b12ce583c7af3a2921</a></p></details> #### <img src='https://whitesource-resources.whitesourcesoftware.com/suggestedVersion.png' width=19 height=20> Mend has checked all newer package trees, and you are on the least vulnerable package! #### Please note: There might be a version that explicitly solves one or more of the vulnerabilities listed below, but we do not recommend it. For more info about the optional fixes, check the section “Details” below. ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (chart.js version) | Fix PR available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2020-7746](https://www.mend.io/vulnerability-database/CVE-2020-7746) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | chart.js-2.9.3.tgz | Direct | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-7746</summary> ### Vulnerable Library - <b>chart.js-2.9.3.tgz</b></p> <p>Simple HTML5 charts using the canvas element.</p> <p>Library home page: <a href="https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz">https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/chart.js/package.json</p> <p> Dependency Hierarchy: - :x: **chart.js-2.9.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/DEV-REPO-URIEL/TEST_AINAT_CHANGE/commit/1855a834302dd3f0182c78b12ce583c7af3a2921">1855a834302dd3f0182c78b12ce583c7af3a2921</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> This affects the package chart.js before 2.9.4. The options parameter is not properly sanitized when it is processed. When the options are processed, the existing options (or the defaults options) are deeply merged with provided options. However, during this operation, the keys of the object being set are not checked, leading to a prototype pollution. <p>Publish Date: 2020-10-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-7746>CVE-2020-7746</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7746">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7746</a></p> <p>Release Date: 2020-10-29</p> <p>Fix Resolution: chart.js - 2.9.4</p> </p> <p></p> <p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p> </details>
True
chart.js-2.9.3.tgz: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>chart.js-2.9.3.tgz</b></p></summary> <p>Simple HTML5 charts using the canvas element.</p> <p>Library home page: <a href="https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz">https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/chart.js/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/DEV-REPO-URIEL/TEST_AINAT_CHANGE/commit/1855a834302dd3f0182c78b12ce583c7af3a2921">1855a834302dd3f0182c78b12ce583c7af3a2921</a></p></details> #### <img src='https://whitesource-resources.whitesourcesoftware.com/suggestedVersion.png' width=19 height=20> Mend has checked all newer package trees, and you are on the least vulnerable package! #### Please note: There might be a version that explicitly solves one or more of the vulnerabilities listed below, but we do not recommend it. For more info about the optional fixes, check the section “Details” below. ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (chart.js version) | Fix PR available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2020-7746](https://www.mend.io/vulnerability-database/CVE-2020-7746) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | chart.js-2.9.3.tgz | Direct | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-7746</summary> ### Vulnerable Library - <b>chart.js-2.9.3.tgz</b></p> <p>Simple HTML5 charts using the canvas element.</p> <p>Library home page: <a href="https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz">https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/chart.js/package.json</p> <p> Dependency Hierarchy: - :x: **chart.js-2.9.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/DEV-REPO-URIEL/TEST_AINAT_CHANGE/commit/1855a834302dd3f0182c78b12ce583c7af3a2921">1855a834302dd3f0182c78b12ce583c7af3a2921</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> This affects the package chart.js before 2.9.4. The options parameter is not properly sanitized when it is processed. When the options are processed, the existing options (or the defaults options) are deeply merged with provided options. However, during this operation, the keys of the object being set are not checked, leading to a prototype pollution. <p>Publish Date: 2020-10-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-7746>CVE-2020-7746</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7746">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7746</a></p> <p>Release Date: 2020-10-29</p> <p>Fix Resolution: chart.js - 2.9.4</p> </p> <p></p> <p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p> </details>
non_process
chart js tgz vulnerabilities highest severity is vulnerable library chart js tgz simple charts using the canvas element library home page a href path to dependency file package json path to vulnerable library node modules chart js package json found in head commit a href mend has checked all newer package trees and you are on the least vulnerable package please note there might be a version that explicitly solves one or more of the vulnerabilities listed below but we do not recommend it for more info about the optional fixes check the section “details” below vulnerabilities cve severity cvss dependency type fixed in chart js version fix pr available high chart js tgz direct n a details cve vulnerable library chart js tgz simple charts using the canvas element library home page a href path to dependency file package json path to vulnerable library node modules chart js package json dependency hierarchy x chart js tgz vulnerable library found in head commit a href found in base branch main vulnerability details this affects the package chart js before the options parameter is not properly sanitized when it is processed when the options are processed the existing options or the defaults options are deeply merged with provided options however during this operation the keys of the object being set are not checked leading to a prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution chart js in order to enable automatic remediation please create
0
222,574
7,433,973,611
IssuesEvent
2018-03-26 09:27:09
CS2103JAN2018-W09-B2/main
https://api.github.com/repos/CS2103JAN2018-W09-B2/main
opened
Create new UI for drafting email
priority.high type.enhancement type.task
#31 Need UI for user to input information like subject title, content and attachments before sending email.
1.0
Create new UI for drafting email - #31 Need UI for user to input information like subject title, content and attachments before sending email.
non_process
create new ui for drafting email need ui for user to input information like subject title content and attachments before sending email
0
7,177
10,318,919,399
IssuesEvent
2019-08-30 16:03:53
googleapis/nodejs-logging-winston
https://api.github.com/repos/googleapis/nodejs-logging-winston
closed
Cut 2.0.0 release
type: process
While the [2.0.0 release PR](https://github.com/googleapis/nodejs-logging-winston/pull/365) is open, it is not quite optimal from the release notes point of view. For semver majors, it is important that the release notes communicate clearly what is breaking. I pushed a commit with [a better changelog](https://github.com/googleapis/nodejs-logging-winston/commit/443ee2310e7eaaa1c14e21d3f7e4d13e0bb27261), but that got overwritten by the release automation. How do we do a release with proper release notes?
1.0
Cut 2.0.0 release - While the [2.0.0 release PR](https://github.com/googleapis/nodejs-logging-winston/pull/365) is open, it is not quite optimal from the release notes point of view. For semver majors, it is important that the release notes communicate clearly what is breaking. I pushed a commit with [a better changelog](https://github.com/googleapis/nodejs-logging-winston/commit/443ee2310e7eaaa1c14e21d3f7e4d13e0bb27261), but that got overwritten by the release automation. How do we do a release with proper release notes?
process
cut release while the is open it is not quite optimal from the release notes point of view for semver majors it is important that the release notes communicate clearly what is breaking i pushed a commit with but that got overwritten by the release automation how do we do a release with proper release notes
1
318,806
27,321,992,334
IssuesEvent
2023-02-24 20:46:51
brave/brave-ios
https://api.github.com/repos/brave/brave-ios
closed
Manual test run for `1.48` on `iPhone` running `iOS 16`
QA/Yes release-notes/exclude QA Pass - iPhone tests iPhone iOS 16
## Installer - [x] Check that installer is close to the size of the last release - [x] Check the Brave version in About and make sure it is EXACTLY as expected ## Data - [x] Verify that data from the previous build appears in the updated build as expected (bookmarks, history, etc.) - [x] Verify that cookies from the previous build are preserved after upgrade - [x] Verify saved passwords are retained after upgrade - [x] Verify stats are retained after upgrade - [x] Verify sync chain created in the previous version is still retained on upgrade - [x] Verify per-site settings are preserved after upgrade ## Bookmarks - [x] Verify that creating a bookmark works - [x] Verify that tapping a bookmark from bookmark manager loads the bookmark - [x] Verify that deleting a bookmark works - [x] Verify that creating a bookmark folder works - [x] Verify that creating a bookmark inside the created folder works - [x] Verify that you can add a bookmark directly inside a bookmark folder - [x] Verify that you can delete a bookmark in edit mode - [x] Verify that you can delete a bookmark folder with bookmarks inside - [x] Verify adding a bookmark domain, subpaths is retained and you are successfully able to visit the domain subpath in a new tab ## Favourites - [x] Verify editing favourites and changing their URLs updates the favicons accordingly - [x] Verify that you can remove favourites - [x] Verify that you can add new favourites from the `Share with...` menu ## Context menus - [x] Make sure context menu items in the URL bar work - [x] Make sure context menu items on content work with no selected text - [x] Make sure context menu items on content work with selected text - [x] Make sure context menu items on content work inside an editable control (input, textarea, or contenteditable) - [x] Verify you can open links in a new tab or private tab. Ensure you can open links in background tabs. ## Downloads - [x] Verify that you can save an image from a site - [x] Verify that you are able to save a GIF image ## Reader Mode - [x] Visit `http://theverge.com`, open any article, verify the reader mode icon is shown in the URL bar - [x] Verify tapping on the reader mode icon opens the article in reader mode - [x] Edit reader mode settings and open different pages in reader mode and verify if the setting is retained across each article ## Gestures - [x] Verify pinch to zoom in/out gestures work on `https://www.homedepot.com` - [x] Verify that navigating to a different origin resets the zoom - [x] Swipe back and forward to navigate, verify this works as expected ## Password Managers - [x] Verify tapping on 1Password on the slide-out keyboard launches 1Password app and you're able to select the stored credentials - [x] Verify tapping on Bitwarden password manager in the autofill field launches the app and auto-fills the stored data ## Shields Settings - [x] Enable all switches in `Settings` and visit a site and disable `Block Scripts`. Kill and relaunch the app and verify if the site Shields settings are retained ## Browser Lock - [x] Verify browser lock enables device passcode to lock/unlock the browser - [x] Verify swipe up/swipe down with browser in focus doesn't ask for device passcode - [x] Verify cancel passcode shows unlock message - [x] Remove the app from memory and relaunch, enter a wrong passcode, the browser should not be unlocked - [x] Verify cancel fingerprint confirmation/face unlock shows enter passcode window when fingerprint/face unlock is set up on the device ## Brave Rewards/Ads - [x] Verify wallet is auto-created after enabling Rewards - [x] Verify when you click on the Brave Rewards panel while on a site, the panel displays if the site is verified or not - [x] Verify ads are only shown while the app is being used - [x] Verify tapping on an ad notification shows the landing page - [x] Verify inline-content ads show on Brave News and are correctly redeemed from the server - [x] Verify `view`,`clicked`, `landed`, and `dismiss` states are logged based on the action ## Sync - [x] Verify you are able to join Sync chain by scanning the QR code - [x] Verify you are able to join Sync chain using code words - [x] Verify you are able to create a Sync chain on the device and add other devices to the chain via QR code/code words - [x] Verify that bookmarks from other devices on the chain show up on the mobile device after sync completes - [x] Verify newly created bookmarks get synced to all devices on the Sync chain - [x] Verify existing bookmarks before joining Sync chain also gets sync'd to all devices on the Sync chain - [x] Verify history sync is disabled by default. Enabling it starts sync'ing history items as well (limited to 200 entries). - [x] Verify sync works on an upgraded profile and new bookmarks added post-upgrade sync across devices on the chain - [x] Verify you can create a standalone Sync chain with one device ## Playlist & Autoplay - [x] Verify Playlist is enabled by default on a clean install - [x] Verify when a site with audio/video is detected, `Playlist` button is shown in URL bar and a notification badge on the menu - [x] Verify clicking the `Playlist` button adds audio/video to playlist and initiates offline play - [x] Verify able to add audio/video to playlist via share menu - [x] Verify able to add audio/video to playlist via long press context menu - [x] Verify Autoplay is disabled by default - [x] Verify with Autoplay enabled, able to play audio/video automatically - [x] Verify with background audio enabled, able to play audio/video with device locked or app in background ## Brave News - [x] Verify Brave News opt-in is shown on a clean install - [x] Verify once opted-in Brave News starts showing articles - [x] Verify inline-content ads are shown on Brave News - [x] Verify able to add custom RSS feed to the list of sources - [x] Verify able to disable content from a particular publisher by long-press on the card ## Bravery settings - [x] Check that block ad and unblock ad works on `http://slashdot.org` - [x] Check that toggling to blocking and allow ads works as expected - [x] Verify that tapping through a cert error in `https://badssl.com` works - [x] Enable `Block Scripts` globally from `Settings`. Visit `https://twitter.com`, nothing should load. Tap on `Shields` and disable `Block Scripts`; the page should load properly. - [x] Verify that default Brave Shields settings take effect on pages with no site settings ### Fingerprint Tests - [x] Verify that turning on fingerprinting protection in preferences shows `3 fingerprints blocked` at `https://browserleaks.com/canvas`. Verify that turning it off in the Bravery menu shows `0 fingerprints blocked`. - [x] Verify that the audio fingerprint is blocked at `https://audiofingerprint.openwpm.com` when fingerprinting protection is on - [x] Verify that `https://diafygi.github.io/webrtc-ips/` doesn't leak IP address when `Block all fingerprinting protection` is on ## Content tests - [x] Go to `https://feedly.com` and make sure that the password can be saved. Make sure the saved password is auto-populated when you visit the site again. - [x] Open an email on `https://mail.google.com` and tap on a link. Make sure it works. - [x] Verify that PDF is loaded over `HTTPS` at `https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf` - [x] Verify that PDF is loaded over `HTTP` at `http://www.pdf995.com/samples/pdf.pdf` - [x] Verify that `https://mixed-script.badssl.com` shows up as grey, not red (no mixed-content scripts are run) - [x] Verify that search results from `https://startpage.com` open in a new tab (due to target being _blank_) ## WebAuthn - [x] Verify browser prompts for security key when trying to log in to accounts - [x] Verify you are able to successfully authenticate an account using security key when prompted - [x] Verify quickly connecting and disconnecting doesn't cause issues authenticating using security keys - [x] Verify you are able to use security keys on both normal and private tabs - [x] Verify you are able to use security keys to authenticate using NFC ## App linker - [x] Long-press on a link in the Twitter app to get the share picker, choose Brave. Verify Brave doesn't crash after opening the link. ## Session storage - [x] Verify that tabs restore when closed, including active tab
1.0
Manual test run for `1.48` on `iPhone` running `iOS 16` - ## Installer - [x] Check that installer is close to the size of the last release - [x] Check the Brave version in About and make sure it is EXACTLY as expected ## Data - [x] Verify that data from the previous build appears in the updated build as expected (bookmarks, history, etc.) - [x] Verify that cookies from the previous build are preserved after upgrade - [x] Verify saved passwords are retained after upgrade - [x] Verify stats are retained after upgrade - [x] Verify sync chain created in the previous version is still retained on upgrade - [x] Verify per-site settings are preserved after upgrade ## Bookmarks - [x] Verify that creating a bookmark works - [x] Verify that tapping a bookmark from bookmark manager loads the bookmark - [x] Verify that deleting a bookmark works - [x] Verify that creating a bookmark folder works - [x] Verify that creating a bookmark inside the created folder works - [x] Verify that you can add a bookmark directly inside a bookmark folder - [x] Verify that you can delete a bookmark in edit mode - [x] Verify that you can delete a bookmark folder with bookmarks inside - [x] Verify adding a bookmark domain, subpaths is retained and you are successfully able to visit the domain subpath in a new tab ## Favourites - [x] Verify editing favourites and changing their URLs updates the favicons accordingly - [x] Verify that you can remove favourites - [x] Verify that you can add new favourites from the `Share with...` menu ## Context menus - [x] Make sure context menu items in the URL bar work - [x] Make sure context menu items on content work with no selected text - [x] Make sure context menu items on content work with selected text - [x] Make sure context menu items on content work inside an editable control (input, textarea, or contenteditable) - [x] Verify you can open links in a new tab or private tab. Ensure you can open links in background tabs. ## Downloads - [x] Verify that you can save an image from a site - [x] Verify that you are able to save a GIF image ## Reader Mode - [x] Visit `http://theverge.com`, open any article, verify the reader mode icon is shown in the URL bar - [x] Verify tapping on the reader mode icon opens the article in reader mode - [x] Edit reader mode settings and open different pages in reader mode and verify if the setting is retained across each article ## Gestures - [x] Verify pinch to zoom in/out gestures work on `https://www.homedepot.com` - [x] Verify that navigating to a different origin resets the zoom - [x] Swipe back and forward to navigate, verify this works as expected ## Password Managers - [x] Verify tapping on 1Password on the slide-out keyboard launches 1Password app and you're able to select the stored credentials - [x] Verify tapping on Bitwarden password manager in the autofill field launches the app and auto-fills the stored data ## Shields Settings - [x] Enable all switches in `Settings` and visit a site and disable `Block Scripts`. Kill and relaunch the app and verify if the site Shields settings are retained ## Browser Lock - [x] Verify browser lock enables device passcode to lock/unlock the browser - [x] Verify swipe up/swipe down with browser in focus doesn't ask for device passcode - [x] Verify cancel passcode shows unlock message - [x] Remove the app from memory and relaunch, enter a wrong passcode, the browser should not be unlocked - [x] Verify cancel fingerprint confirmation/face unlock shows enter passcode window when fingerprint/face unlock is set up on the device ## Brave Rewards/Ads - [x] Verify wallet is auto-created after enabling Rewards - [x] Verify when you click on the Brave Rewards panel while on a site, the panel displays if the site is verified or not - [x] Verify ads are only shown while the app is being used - [x] Verify tapping on an ad notification shows the landing page - [x] Verify inline-content ads show on Brave News and are correctly redeemed from the server - [x] Verify `view`,`clicked`, `landed`, and `dismiss` states are logged based on the action ## Sync - [x] Verify you are able to join Sync chain by scanning the QR code - [x] Verify you are able to join Sync chain using code words - [x] Verify you are able to create a Sync chain on the device and add other devices to the chain via QR code/code words - [x] Verify that bookmarks from other devices on the chain show up on the mobile device after sync completes - [x] Verify newly created bookmarks get synced to all devices on the Sync chain - [x] Verify existing bookmarks before joining Sync chain also gets sync'd to all devices on the Sync chain - [x] Verify history sync is disabled by default. Enabling it starts sync'ing history items as well (limited to 200 entries). - [x] Verify sync works on an upgraded profile and new bookmarks added post-upgrade sync across devices on the chain - [x] Verify you can create a standalone Sync chain with one device ## Playlist & Autoplay - [x] Verify Playlist is enabled by default on a clean install - [x] Verify when a site with audio/video is detected, `Playlist` button is shown in URL bar and a notification badge on the menu - [x] Verify clicking the `Playlist` button adds audio/video to playlist and initiates offline play - [x] Verify able to add audio/video to playlist via share menu - [x] Verify able to add audio/video to playlist via long press context menu - [x] Verify Autoplay is disabled by default - [x] Verify with Autoplay enabled, able to play audio/video automatically - [x] Verify with background audio enabled, able to play audio/video with device locked or app in background ## Brave News - [x] Verify Brave News opt-in is shown on a clean install - [x] Verify once opted-in Brave News starts showing articles - [x] Verify inline-content ads are shown on Brave News - [x] Verify able to add custom RSS feed to the list of sources - [x] Verify able to disable content from a particular publisher by long-press on the card ## Bravery settings - [x] Check that block ad and unblock ad works on `http://slashdot.org` - [x] Check that toggling to blocking and allow ads works as expected - [x] Verify that tapping through a cert error in `https://badssl.com` works - [x] Enable `Block Scripts` globally from `Settings`. Visit `https://twitter.com`, nothing should load. Tap on `Shields` and disable `Block Scripts`; the page should load properly. - [x] Verify that default Brave Shields settings take effect on pages with no site settings ### Fingerprint Tests - [x] Verify that turning on fingerprinting protection in preferences shows `3 fingerprints blocked` at `https://browserleaks.com/canvas`. Verify that turning it off in the Bravery menu shows `0 fingerprints blocked`. - [x] Verify that the audio fingerprint is blocked at `https://audiofingerprint.openwpm.com` when fingerprinting protection is on - [x] Verify that `https://diafygi.github.io/webrtc-ips/` doesn't leak IP address when `Block all fingerprinting protection` is on ## Content tests - [x] Go to `https://feedly.com` and make sure that the password can be saved. Make sure the saved password is auto-populated when you visit the site again. - [x] Open an email on `https://mail.google.com` and tap on a link. Make sure it works. - [x] Verify that PDF is loaded over `HTTPS` at `https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf` - [x] Verify that PDF is loaded over `HTTP` at `http://www.pdf995.com/samples/pdf.pdf` - [x] Verify that `https://mixed-script.badssl.com` shows up as grey, not red (no mixed-content scripts are run) - [x] Verify that search results from `https://startpage.com` open in a new tab (due to target being _blank_) ## WebAuthn - [x] Verify browser prompts for security key when trying to log in to accounts - [x] Verify you are able to successfully authenticate an account using security key when prompted - [x] Verify quickly connecting and disconnecting doesn't cause issues authenticating using security keys - [x] Verify you are able to use security keys on both normal and private tabs - [x] Verify you are able to use security keys to authenticate using NFC ## App linker - [x] Long-press on a link in the Twitter app to get the share picker, choose Brave. Verify Brave doesn't crash after opening the link. ## Session storage - [x] Verify that tabs restore when closed, including active tab
non_process
manual test run for on iphone running ios installer check that installer is close to the size of the last release check the brave version in about and make sure it is exactly as expected data verify that data from the previous build appears in the updated build as expected bookmarks history etc verify that cookies from the previous build are preserved after upgrade verify saved passwords are retained after upgrade verify stats are retained after upgrade verify sync chain created in the previous version is still retained on upgrade verify per site settings are preserved after upgrade bookmarks verify that creating a bookmark works verify that tapping a bookmark from bookmark manager loads the bookmark verify that deleting a bookmark works verify that creating a bookmark folder works verify that creating a bookmark inside the created folder works verify that you can add a bookmark directly inside a bookmark folder verify that you can delete a bookmark in edit mode verify that you can delete a bookmark folder with bookmarks inside verify adding a bookmark domain subpaths is retained and you are successfully able to visit the domain subpath in a new tab favourites verify editing favourites and changing their urls updates the favicons accordingly verify that you can remove favourites verify that you can add new favourites from the share with menu context menus make sure context menu items in the url bar work make sure context menu items on content work with no selected text make sure context menu items on content work with selected text make sure context menu items on content work inside an editable control input textarea or contenteditable verify you can open links in a new tab or private tab ensure you can open links in background tabs downloads verify that you can save an image from a site verify that you are able to save a gif image reader mode visit open any article verify the reader mode icon is shown in the url bar verify tapping on the reader mode icon opens the article in reader mode edit reader mode settings and open different pages in reader mode and verify if the setting is retained across each article gestures verify pinch to zoom in out gestures work on verify that navigating to a different origin resets the zoom swipe back and forward to navigate verify this works as expected password managers verify tapping on on the slide out keyboard launches app and you re able to select the stored credentials verify tapping on bitwarden password manager in the autofill field launches the app and auto fills the stored data shields settings enable all switches in settings and visit a site and disable block scripts kill and relaunch the app and verify if the site shields settings are retained browser lock verify browser lock enables device passcode to lock unlock the browser verify swipe up swipe down with browser in focus doesn t ask for device passcode verify cancel passcode shows unlock message remove the app from memory and relaunch enter a wrong passcode the browser should not be unlocked verify cancel fingerprint confirmation face unlock shows enter passcode window when fingerprint face unlock is set up on the device brave rewards ads verify wallet is auto created after enabling rewards verify when you click on the brave rewards panel while on a site the panel displays if the site is verified or not verify ads are only shown while the app is being used verify tapping on an ad notification shows the landing page verify inline content ads show on brave news and are correctly redeemed from the server verify view clicked landed and dismiss states are logged based on the action sync verify you are able to join sync chain by scanning the qr code verify you are able to join sync chain using code words verify you are able to create a sync chain on the device and add other devices to the chain via qr code code words verify that bookmarks from other devices on the chain show up on the mobile device after sync completes verify newly created bookmarks get synced to all devices on the sync chain verify existing bookmarks before joining sync chain also gets sync d to all devices on the sync chain verify history sync is disabled by default enabling it starts sync ing history items as well limited to entries verify sync works on an upgraded profile and new bookmarks added post upgrade sync across devices on the chain verify you can create a standalone sync chain with one device playlist autoplay verify playlist is enabled by default on a clean install verify when a site with audio video is detected playlist button is shown in url bar and a notification badge on the menu verify clicking the playlist button adds audio video to playlist and initiates offline play verify able to add audio video to playlist via share menu verify able to add audio video to playlist via long press context menu verify autoplay is disabled by default verify with autoplay enabled able to play audio video automatically verify with background audio enabled able to play audio video with device locked or app in background brave news verify brave news opt in is shown on a clean install verify once opted in brave news starts showing articles verify inline content ads are shown on brave news verify able to add custom rss feed to the list of sources verify able to disable content from a particular publisher by long press on the card bravery settings check that block ad and unblock ad works on check that toggling to blocking and allow ads works as expected verify that tapping through a cert error in works enable block scripts globally from settings visit nothing should load tap on shields and disable block scripts the page should load properly verify that default brave shields settings take effect on pages with no site settings fingerprint tests verify that turning on fingerprinting protection in preferences shows fingerprints blocked at verify that turning it off in the bravery menu shows fingerprints blocked verify that the audio fingerprint is blocked at when fingerprinting protection is on verify that doesn t leak ip address when block all fingerprinting protection is on content tests go to and make sure that the password can be saved make sure the saved password is auto populated when you visit the site again open an email on and tap on a link make sure it works verify that pdf is loaded over https at verify that pdf is loaded over http at verify that shows up as grey not red no mixed content scripts are run verify that search results from open in a new tab due to target being blank webauthn verify browser prompts for security key when trying to log in to accounts verify you are able to successfully authenticate an account using security key when prompted verify quickly connecting and disconnecting doesn t cause issues authenticating using security keys verify you are able to use security keys on both normal and private tabs verify you are able to use security keys to authenticate using nfc app linker long press on a link in the twitter app to get the share picker choose brave verify brave doesn t crash after opening the link session storage verify that tabs restore when closed including active tab
0
13,697
16,455,144,759
IssuesEvent
2021-05-21 11:29:31
metabase/metabase
https://api.github.com/repos/metabase/metabase
opened
Empty statement results causes reducer error, which causes the entire query to fail
.Regression Database/SQLServer Priority:P2 Querying/Processor Type:Bug
**Describe the bug** While multi-statements are officially not supported #4050, 0.39.0 introduced an upgraded driver #15287, which now causes empty statement results to produce an error like `Select statement did not produce a ResultSet for native query`, which prevent the entire query from working. This is has been reported for SQL Server, but might be a more general problem. It might be possible to rewrite some queries to [CTEs](https://www.metabase.com/learn/building-analytics/sql-templates/sql-cte.html), but for some queries this might not be feasible or require a lot of work. **To Reproduce** 1. Native query > SQL Server: ``` drop table if exists #myTemporaryTable; select 1; ``` 2. Errors with `Select statement did not produce a ResultSet for native query` ![image](https://user-images.githubusercontent.com/1447303/119130209-2fee1880-ba38-11eb-8603-b8178a8c2800.png) <details><summary>Full stacktrace</summary> ``` 2021-05-21 13:26:33,775 ERROR middleware.catch-exceptions :: Error processing query: null {:database_id 20, :started_at #t "2021-05-21T13:26:31.148017+02:00[Europe/Copenhagen]", :via [{:status :failed, :class clojure.lang.ExceptionInfo, :error "Error executing query", :stacktrace ["--> driver.sql_jdbc.execute$execute_reducible_query$fn__79533.invoke(execute.clj:470)" "driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:467)" "driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)" "driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:462)" "driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)" "driver.sql_jdbc$fn__80985.invokeStatic(sql_jdbc.clj:54)" "driver.sql_jdbc$fn__80985.invoke(sql_jdbc.clj:52)" "query_processor.context$executef.invokeStatic(context.clj:59)" "query_processor.context$executef.invoke(context.clj:48)" "query_processor.context.default$default_runf.invokeStatic(default.clj:68)" "query_processor.context.default$default_runf.invoke(default.clj:66)" "query_processor.context$runf.invokeStatic(context.clj:45)" "query_processor.context$runf.invoke(context.clj:39)" "query_processor.reducible$pivot.invokeStatic(reducible.clj:34)" "query_processor.reducible$pivot.invoke(reducible.clj:31)" "query_processor.middleware.mbql_to_native$mbql__GT_native$fn__47305.invoke(mbql_to_native.clj:25)" "query_processor.middleware.check_features$check_features$fn__46547.invoke(check_features.clj:39)" "query_processor.middleware.limit$limit$fn__47291.invoke(limit.clj:37)" "query_processor.middleware.cache$maybe_return_cached_results$fn__45999.invoke(cache.clj:211)" "query_processor.middleware.optimize_temporal_filters$optimize_temporal_filters$fn__47551.invoke(optimize_temporal_filters.clj:204)" "query_processor.middleware.validate_temporal_bucketing$validate_temporal_bucketing$fn__49481.invoke(validate_temporal_bucketing.clj:50)" "query_processor.middleware.auto_parse_filter_values$auto_parse_filter_values$fn__45118.invoke(auto_parse_filter_values.clj:43)" "query_processor.middleware.wrap_value_literals$wrap_value_literals$fn__41435.invoke(wrap_value_literals.clj:161)" "query_processor.middleware.annotate$add_column_info$fn__41310.invoke(annotate.clj:605)" "query_processor.middleware.permissions$check_query_permissions$fn__46419.invoke(permissions.clj:81)" "query_processor.middleware.pre_alias_aggregations$pre_alias_aggregations$fn__48409.invoke(pre_alias_aggregations.clj:40)" "query_processor.middleware.cumulative_aggregations$handle_cumulative_aggregations$fn__46620.invoke(cumulative_aggregations.clj:60)" "query_processor.middleware.resolve_joined_fields$resolve_joined_fields$fn__48708.invoke(resolve_joined_fields.clj:102)" "query_processor.middleware.resolve_joins$resolve_joins$fn__49021.invoke(resolve_joins.clj:171)" "query_processor.middleware.add_implicit_joins$add_implicit_joins$fn__44694.invoke(add_implicit_joins.clj:190)" "query_processor.middleware.large_int_id$convert_id_to_string$fn__47255.invoke(large_int_id.clj:59)" "query_processor.middleware.format_rows$format_rows$fn__47236.invoke(format_rows.clj:74)" "query_processor.middleware.add_default_temporal_unit$add_default_temporal_unit$fn__43988.invoke(add_default_temporal_unit.clj:23)" "query_processor.middleware.desugar$desugar$fn__46686.invoke(desugar.clj:21)" "query_processor.middleware.binning$update_binning_strategy$fn__45505.invoke(binning.clj:227)" "query_processor.middleware.resolve_fields$resolve_fields$fn__46222.invoke(resolve_fields.clj:34)" "query_processor.middleware.add_dimension_projections$add_remapping$fn__44343.invoke(add_dimension_projections.clj:312)" "query_processor.middleware.add_implicit_clauses$add_implicit_clauses$fn__44572.invoke(add_implicit_clauses.clj:147)" "query_processor.middleware.upgrade_field_literals$upgrade_field_literals$fn__49430.invoke(upgrade_field_literals.clj:40)" "query_processor.middleware.add_source_metadata$add_source_metadata_for_source_queries$fn__44857.invoke(add_source_metadata.clj:123)" "query_processor.middleware.reconcile_breakout_and_order_by_bucketing$reconcile_breakout_and_order_by_bucketing$fn__48583.invoke(reconcile_breakout_and_order_by_bucketing.clj:100)" "query_processor.middleware.auto_bucket_datetimes$auto_bucket_datetimes$fn__45065.invoke(auto_bucket_datetimes.clj:147)" "query_processor.middleware.resolve_source_table$resolve_source_tables$fn__46269.invoke(resolve_source_table.clj:45)" "query_processor.middleware.parameters$substitute_parameters$fn__48391.invoke(parameters.clj:111)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__46321.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__46942.invoke(expand_macros.clj:155)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__44866.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__49383.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594$fn__48598.invoke(resolve_database_and_driver.clj:31)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594.invoke(resolve_database_and_driver.clj:25)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__47182.invoke(fetch_source_query.clj:274)" "query_processor.middleware.store$initialize_store$fn__49392$fn__49393.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:44)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__49392.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__49437.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__47318.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__44712.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49368.invoke(results_metadata.clj:147)" "query_processor.middleware.constraints$add_default_userland_constraints$fn__46563.invoke(constraints.clj:42)" "query_processor.middleware.process_userland_query$process_userland_query$fn__48480.invoke(process_userland_query.clj:135)" "query_processor.middleware.catch_exceptions$catch_exceptions$fn__46503.invoke(catch_exceptions.clj:173)" "query_processor.reducible$async_qp$qp_STAR___37989$thunk__37990.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___37989.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___37998$fn__38001.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___37998.invoke(reducible.clj:134)" "query_processor$process_userland_query.invokeStatic(query_processor.clj:241)" "query_processor$process_userland_query.doInvoke(query_processor.clj:237)" "query_processor$fn__49527$process_query_and_save_execution_BANG___49536$fn__49539.invoke(query_processor.clj:253)" "query_processor$fn__49527$process_query_and_save_execution_BANG___49536.invoke(query_processor.clj:245)" "query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580$fn__49583.invoke(query_processor.clj:265)" "query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580.invoke(query_processor.clj:258)" "api.dataset$run_query_async$fn__55784.invoke(dataset.clj:56)" "query_processor.streaming$streaming_response_STAR_$fn__55763$fn__55764.invoke(streaming.clj:72)" "query_processor.streaming$streaming_response_STAR_$fn__55763.invoke(streaming.clj:71)" "async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)" "async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)" "async.streaming_response$do_f_async$fn__16071.invoke(streaming_response.clj:84)"], :error_type :invalid-query, :ex-data {:sql "-- Metabase:: userID: 1 queryType: native queryHash: 1b671306840b04531e6a4a4fcc56cd0d6dbfdb671c65eb3c14850b139b358a26\ndrop table if exists #myTemporaryTable;\nselect 1;", :params nil, :type :invalid-query}}], :error_type :invalid-query, :json_query {:type "native", :native {:query "drop table if exists #myTemporaryTable;\nselect 1;", :template-tags {}}, :database 20, :parameters [], :middleware {:js-int-to-string? true, :add-default-userland-constraints? true}}, :status :failed, :class clojure.lang.ExceptionInfo, :stacktrace ["--> driver.sql_jdbc.execute$fn__79453.invokeStatic(execute.clj:336)" "driver.sql_jdbc.execute$fn__79453.invoke(execute.clj:332)" "driver.sql_jdbc.execute$execute_statement_or_prepared_statement_BANG_.invokeStatic(execute.clj:342)" "driver.sql_jdbc.execute$execute_statement_or_prepared_statement_BANG_.invoke(execute.clj:339)" "driver.sql_jdbc.execute$execute_reducible_query$fn__79533.invoke(execute.clj:468)" "driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:467)" "driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)" "driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:462)" "driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)" "driver.sql_jdbc$fn__80985.invokeStatic(sql_jdbc.clj:54)" "driver.sql_jdbc$fn__80985.invoke(sql_jdbc.clj:52)" "query_processor.context$executef.invokeStatic(context.clj:59)" "query_processor.context$executef.invoke(context.clj:48)" "query_processor.context.default$default_runf.invokeStatic(default.clj:68)" "query_processor.context.default$default_runf.invoke(default.clj:66)" "query_processor.context$runf.invokeStatic(context.clj:45)" "query_processor.context$runf.invoke(context.clj:39)" "query_processor.reducible$pivot.invokeStatic(reducible.clj:34)" "query_processor.reducible$pivot.invoke(reducible.clj:31)" "query_processor.middleware.mbql_to_native$mbql__GT_native$fn__47305.invoke(mbql_to_native.clj:25)" "query_processor.middleware.check_features$check_features$fn__46547.invoke(check_features.clj:39)" "query_processor.middleware.limit$limit$fn__47291.invoke(limit.clj:37)" "query_processor.middleware.cache$maybe_return_cached_results$fn__45999.invoke(cache.clj:211)" "query_processor.middleware.optimize_temporal_filters$optimize_temporal_filters$fn__47551.invoke(optimize_temporal_filters.clj:204)" "query_processor.middleware.validate_temporal_bucketing$validate_temporal_bucketing$fn__49481.invoke(validate_temporal_bucketing.clj:50)" "query_processor.middleware.auto_parse_filter_values$auto_parse_filter_values$fn__45118.invoke(auto_parse_filter_values.clj:43)" "query_processor.middleware.wrap_value_literals$wrap_value_literals$fn__41435.invoke(wrap_value_literals.clj:161)" "query_processor.middleware.annotate$add_column_info$fn__41310.invoke(annotate.clj:605)" "query_processor.middleware.permissions$check_query_permissions$fn__46419.invoke(permissions.clj:81)" "query_processor.middleware.pre_alias_aggregations$pre_alias_aggregations$fn__48409.invoke(pre_alias_aggregations.clj:40)" "query_processor.middleware.cumulative_aggregations$handle_cumulative_aggregations$fn__46620.invoke(cumulative_aggregations.clj:60)" "query_processor.middleware.resolve_joined_fields$resolve_joined_fields$fn__48708.invoke(resolve_joined_fields.clj:102)" "query_processor.middleware.resolve_joins$resolve_joins$fn__49021.invoke(resolve_joins.clj:171)" "query_processor.middleware.add_implicit_joins$add_implicit_joins$fn__44694.invoke(add_implicit_joins.clj:190)" "query_processor.middleware.large_int_id$convert_id_to_string$fn__47255.invoke(large_int_id.clj:59)" "query_processor.middleware.format_rows$format_rows$fn__47236.invoke(format_rows.clj:74)" "query_processor.middleware.add_default_temporal_unit$add_default_temporal_unit$fn__43988.invoke(add_default_temporal_unit.clj:23)" "query_processor.middleware.desugar$desugar$fn__46686.invoke(desugar.clj:21)" "query_processor.middleware.binning$update_binning_strategy$fn__45505.invoke(binning.clj:227)" "query_processor.middleware.resolve_fields$resolve_fields$fn__46222.invoke(resolve_fields.clj:34)" "query_processor.middleware.add_dimension_projections$add_remapping$fn__44343.invoke(add_dimension_projections.clj:312)" "query_processor.middleware.add_implicit_clauses$add_implicit_clauses$fn__44572.invoke(add_implicit_clauses.clj:147)" "query_processor.middleware.upgrade_field_literals$upgrade_field_literals$fn__49430.invoke(upgrade_field_literals.clj:40)" "query_processor.middleware.add_source_metadata$add_source_metadata_for_source_queries$fn__44857.invoke(add_source_metadata.clj:123)" "query_processor.middleware.reconcile_breakout_and_order_by_bucketing$reconcile_breakout_and_order_by_bucketing$fn__48583.invoke(reconcile_breakout_and_order_by_bucketing.clj:100)" "query_processor.middleware.auto_bucket_datetimes$auto_bucket_datetimes$fn__45065.invoke(auto_bucket_datetimes.clj:147)" "query_processor.middleware.resolve_source_table$resolve_source_tables$fn__46269.invoke(resolve_source_table.clj:45)" "query_processor.middleware.parameters$substitute_parameters$fn__48391.invoke(parameters.clj:111)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__46321.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__46942.invoke(expand_macros.clj:155)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__44866.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__49383.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594$fn__48598.invoke(resolve_database_and_driver.clj:31)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594.invoke(resolve_database_and_driver.clj:25)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__47182.invoke(fetch_source_query.clj:274)" "query_processor.middleware.store$initialize_store$fn__49392$fn__49393.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:44)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__49392.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__49437.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__47318.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__44712.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49368.invoke(results_metadata.clj:147)" "query_processor.middleware.constraints$add_default_userland_constraints$fn__46563.invoke(constraints.clj:42)" "query_processor.middleware.process_userland_query$process_userland_query$fn__48480.invoke(process_userland_query.clj:135)" "query_processor.middleware.catch_exceptions$catch_exceptions$fn__46503.invoke(catch_exceptions.clj:173)" "query_processor.reducible$async_qp$qp_STAR___37989$thunk__37990.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___37989.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___37998$fn__38001.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___37998.invoke(reducible.clj:134)" "query_processor$process_userland_query.invokeStatic(query_processor.clj:241)" "query_processor$process_userland_query.doInvoke(query_processor.clj:237)" "query_processor$fn__49527$process_query_and_save_execution_BANG___49536$fn__49539.invoke(query_processor.clj:253)" "query_processor$fn__49527$process_query_and_save_execution_BANG___49536.invoke(query_processor.clj:245)" "query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580$fn__49583.invoke(query_processor.clj:265)" "query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580.invoke(query_processor.clj:258)" "api.dataset$run_query_async$fn__55784.invoke(dataset.clj:56)" "query_processor.streaming$streaming_response_STAR_$fn__55763$fn__55764.invoke(streaming.clj:72)" "query_processor.streaming$streaming_response_STAR_$fn__55763.invoke(streaming.clj:71)" "async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)" "async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)" "async.streaming_response$do_f_async$fn__16071.invoke(streaming_response.clj:84)"], :context :ad-hoc, :error "Select statement did not produce a ResultSet for native query", :row_count 0, :running_time 0, :ex-data {:sql "-- Metabase:: userID: 1 queryType: native queryHash: 1b671306840b04531e6a4a4fcc56cd0d6dbfdb671c65eb3c14850b139b358a26\ndrop table if exists #myTemporaryTable;\nselect 1;", :driver :sqlserver}, :data {:rows [], :cols []}} 2021-05-21 13:26:33,823 DEBUG middleware.log :: POST /api/dataset 202 [ASYNC: completed] 2.7 s (6 DB calls) App DB connections: 0/7 Jetty threads: 3/50 (4 idle, 0 queued) (50 total active threads) Queries in flight: 1 (0 queued) ``` </details> **Expected behavior** The first statement to be executed and empty resultset to be ignored, and then run second statement and show result from that. This is 0.38.4: ![image](https://user-images.githubusercontent.com/1447303/119129980-e56c9c00-ba37-11eb-9b34-97400025e1f9.png) **Information about your Metabase Installation:** Regression since 0.39.0 - tested 0.38.4 thru 0.39.2 **Additional context** https://discourse.metabase.com/t/sql-queries-broken-in-0-39-0-1/14468
1.0
Empty statement results causes reducer error, which causes the entire query to fail - **Describe the bug** While multi-statements are officially not supported #4050, 0.39.0 introduced an upgraded driver #15287, which now causes empty statement results to produce an error like `Select statement did not produce a ResultSet for native query`, which prevent the entire query from working. This is has been reported for SQL Server, but might be a more general problem. It might be possible to rewrite some queries to [CTEs](https://www.metabase.com/learn/building-analytics/sql-templates/sql-cte.html), but for some queries this might not be feasible or require a lot of work. **To Reproduce** 1. Native query > SQL Server: ``` drop table if exists #myTemporaryTable; select 1; ``` 2. Errors with `Select statement did not produce a ResultSet for native query` ![image](https://user-images.githubusercontent.com/1447303/119130209-2fee1880-ba38-11eb-8603-b8178a8c2800.png) <details><summary>Full stacktrace</summary> ``` 2021-05-21 13:26:33,775 ERROR middleware.catch-exceptions :: Error processing query: null {:database_id 20, :started_at #t "2021-05-21T13:26:31.148017+02:00[Europe/Copenhagen]", :via [{:status :failed, :class clojure.lang.ExceptionInfo, :error "Error executing query", :stacktrace ["--> driver.sql_jdbc.execute$execute_reducible_query$fn__79533.invoke(execute.clj:470)" "driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:467)" "driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)" "driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:462)" "driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)" "driver.sql_jdbc$fn__80985.invokeStatic(sql_jdbc.clj:54)" "driver.sql_jdbc$fn__80985.invoke(sql_jdbc.clj:52)" "query_processor.context$executef.invokeStatic(context.clj:59)" "query_processor.context$executef.invoke(context.clj:48)" "query_processor.context.default$default_runf.invokeStatic(default.clj:68)" "query_processor.context.default$default_runf.invoke(default.clj:66)" "query_processor.context$runf.invokeStatic(context.clj:45)" "query_processor.context$runf.invoke(context.clj:39)" "query_processor.reducible$pivot.invokeStatic(reducible.clj:34)" "query_processor.reducible$pivot.invoke(reducible.clj:31)" "query_processor.middleware.mbql_to_native$mbql__GT_native$fn__47305.invoke(mbql_to_native.clj:25)" "query_processor.middleware.check_features$check_features$fn__46547.invoke(check_features.clj:39)" "query_processor.middleware.limit$limit$fn__47291.invoke(limit.clj:37)" "query_processor.middleware.cache$maybe_return_cached_results$fn__45999.invoke(cache.clj:211)" "query_processor.middleware.optimize_temporal_filters$optimize_temporal_filters$fn__47551.invoke(optimize_temporal_filters.clj:204)" "query_processor.middleware.validate_temporal_bucketing$validate_temporal_bucketing$fn__49481.invoke(validate_temporal_bucketing.clj:50)" "query_processor.middleware.auto_parse_filter_values$auto_parse_filter_values$fn__45118.invoke(auto_parse_filter_values.clj:43)" "query_processor.middleware.wrap_value_literals$wrap_value_literals$fn__41435.invoke(wrap_value_literals.clj:161)" "query_processor.middleware.annotate$add_column_info$fn__41310.invoke(annotate.clj:605)" "query_processor.middleware.permissions$check_query_permissions$fn__46419.invoke(permissions.clj:81)" "query_processor.middleware.pre_alias_aggregations$pre_alias_aggregations$fn__48409.invoke(pre_alias_aggregations.clj:40)" "query_processor.middleware.cumulative_aggregations$handle_cumulative_aggregations$fn__46620.invoke(cumulative_aggregations.clj:60)" "query_processor.middleware.resolve_joined_fields$resolve_joined_fields$fn__48708.invoke(resolve_joined_fields.clj:102)" "query_processor.middleware.resolve_joins$resolve_joins$fn__49021.invoke(resolve_joins.clj:171)" "query_processor.middleware.add_implicit_joins$add_implicit_joins$fn__44694.invoke(add_implicit_joins.clj:190)" "query_processor.middleware.large_int_id$convert_id_to_string$fn__47255.invoke(large_int_id.clj:59)" "query_processor.middleware.format_rows$format_rows$fn__47236.invoke(format_rows.clj:74)" "query_processor.middleware.add_default_temporal_unit$add_default_temporal_unit$fn__43988.invoke(add_default_temporal_unit.clj:23)" "query_processor.middleware.desugar$desugar$fn__46686.invoke(desugar.clj:21)" "query_processor.middleware.binning$update_binning_strategy$fn__45505.invoke(binning.clj:227)" "query_processor.middleware.resolve_fields$resolve_fields$fn__46222.invoke(resolve_fields.clj:34)" "query_processor.middleware.add_dimension_projections$add_remapping$fn__44343.invoke(add_dimension_projections.clj:312)" "query_processor.middleware.add_implicit_clauses$add_implicit_clauses$fn__44572.invoke(add_implicit_clauses.clj:147)" "query_processor.middleware.upgrade_field_literals$upgrade_field_literals$fn__49430.invoke(upgrade_field_literals.clj:40)" "query_processor.middleware.add_source_metadata$add_source_metadata_for_source_queries$fn__44857.invoke(add_source_metadata.clj:123)" "query_processor.middleware.reconcile_breakout_and_order_by_bucketing$reconcile_breakout_and_order_by_bucketing$fn__48583.invoke(reconcile_breakout_and_order_by_bucketing.clj:100)" "query_processor.middleware.auto_bucket_datetimes$auto_bucket_datetimes$fn__45065.invoke(auto_bucket_datetimes.clj:147)" "query_processor.middleware.resolve_source_table$resolve_source_tables$fn__46269.invoke(resolve_source_table.clj:45)" "query_processor.middleware.parameters$substitute_parameters$fn__48391.invoke(parameters.clj:111)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__46321.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__46942.invoke(expand_macros.clj:155)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__44866.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__49383.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594$fn__48598.invoke(resolve_database_and_driver.clj:31)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594.invoke(resolve_database_and_driver.clj:25)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__47182.invoke(fetch_source_query.clj:274)" "query_processor.middleware.store$initialize_store$fn__49392$fn__49393.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:44)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__49392.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__49437.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__47318.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__44712.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49368.invoke(results_metadata.clj:147)" "query_processor.middleware.constraints$add_default_userland_constraints$fn__46563.invoke(constraints.clj:42)" "query_processor.middleware.process_userland_query$process_userland_query$fn__48480.invoke(process_userland_query.clj:135)" "query_processor.middleware.catch_exceptions$catch_exceptions$fn__46503.invoke(catch_exceptions.clj:173)" "query_processor.reducible$async_qp$qp_STAR___37989$thunk__37990.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___37989.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___37998$fn__38001.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___37998.invoke(reducible.clj:134)" "query_processor$process_userland_query.invokeStatic(query_processor.clj:241)" "query_processor$process_userland_query.doInvoke(query_processor.clj:237)" "query_processor$fn__49527$process_query_and_save_execution_BANG___49536$fn__49539.invoke(query_processor.clj:253)" "query_processor$fn__49527$process_query_and_save_execution_BANG___49536.invoke(query_processor.clj:245)" "query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580$fn__49583.invoke(query_processor.clj:265)" "query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580.invoke(query_processor.clj:258)" "api.dataset$run_query_async$fn__55784.invoke(dataset.clj:56)" "query_processor.streaming$streaming_response_STAR_$fn__55763$fn__55764.invoke(streaming.clj:72)" "query_processor.streaming$streaming_response_STAR_$fn__55763.invoke(streaming.clj:71)" "async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)" "async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)" "async.streaming_response$do_f_async$fn__16071.invoke(streaming_response.clj:84)"], :error_type :invalid-query, :ex-data {:sql "-- Metabase:: userID: 1 queryType: native queryHash: 1b671306840b04531e6a4a4fcc56cd0d6dbfdb671c65eb3c14850b139b358a26\ndrop table if exists #myTemporaryTable;\nselect 1;", :params nil, :type :invalid-query}}], :error_type :invalid-query, :json_query {:type "native", :native {:query "drop table if exists #myTemporaryTable;\nselect 1;", :template-tags {}}, :database 20, :parameters [], :middleware {:js-int-to-string? true, :add-default-userland-constraints? true}}, :status :failed, :class clojure.lang.ExceptionInfo, :stacktrace ["--> driver.sql_jdbc.execute$fn__79453.invokeStatic(execute.clj:336)" "driver.sql_jdbc.execute$fn__79453.invoke(execute.clj:332)" "driver.sql_jdbc.execute$execute_statement_or_prepared_statement_BANG_.invokeStatic(execute.clj:342)" "driver.sql_jdbc.execute$execute_statement_or_prepared_statement_BANG_.invoke(execute.clj:339)" "driver.sql_jdbc.execute$execute_reducible_query$fn__79533.invoke(execute.clj:468)" "driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:467)" "driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)" "driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:462)" "driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)" "driver.sql_jdbc$fn__80985.invokeStatic(sql_jdbc.clj:54)" "driver.sql_jdbc$fn__80985.invoke(sql_jdbc.clj:52)" "query_processor.context$executef.invokeStatic(context.clj:59)" "query_processor.context$executef.invoke(context.clj:48)" "query_processor.context.default$default_runf.invokeStatic(default.clj:68)" "query_processor.context.default$default_runf.invoke(default.clj:66)" "query_processor.context$runf.invokeStatic(context.clj:45)" "query_processor.context$runf.invoke(context.clj:39)" "query_processor.reducible$pivot.invokeStatic(reducible.clj:34)" "query_processor.reducible$pivot.invoke(reducible.clj:31)" "query_processor.middleware.mbql_to_native$mbql__GT_native$fn__47305.invoke(mbql_to_native.clj:25)" "query_processor.middleware.check_features$check_features$fn__46547.invoke(check_features.clj:39)" "query_processor.middleware.limit$limit$fn__47291.invoke(limit.clj:37)" "query_processor.middleware.cache$maybe_return_cached_results$fn__45999.invoke(cache.clj:211)" "query_processor.middleware.optimize_temporal_filters$optimize_temporal_filters$fn__47551.invoke(optimize_temporal_filters.clj:204)" "query_processor.middleware.validate_temporal_bucketing$validate_temporal_bucketing$fn__49481.invoke(validate_temporal_bucketing.clj:50)" "query_processor.middleware.auto_parse_filter_values$auto_parse_filter_values$fn__45118.invoke(auto_parse_filter_values.clj:43)" "query_processor.middleware.wrap_value_literals$wrap_value_literals$fn__41435.invoke(wrap_value_literals.clj:161)" "query_processor.middleware.annotate$add_column_info$fn__41310.invoke(annotate.clj:605)" "query_processor.middleware.permissions$check_query_permissions$fn__46419.invoke(permissions.clj:81)" "query_processor.middleware.pre_alias_aggregations$pre_alias_aggregations$fn__48409.invoke(pre_alias_aggregations.clj:40)" "query_processor.middleware.cumulative_aggregations$handle_cumulative_aggregations$fn__46620.invoke(cumulative_aggregations.clj:60)" "query_processor.middleware.resolve_joined_fields$resolve_joined_fields$fn__48708.invoke(resolve_joined_fields.clj:102)" "query_processor.middleware.resolve_joins$resolve_joins$fn__49021.invoke(resolve_joins.clj:171)" "query_processor.middleware.add_implicit_joins$add_implicit_joins$fn__44694.invoke(add_implicit_joins.clj:190)" "query_processor.middleware.large_int_id$convert_id_to_string$fn__47255.invoke(large_int_id.clj:59)" "query_processor.middleware.format_rows$format_rows$fn__47236.invoke(format_rows.clj:74)" "query_processor.middleware.add_default_temporal_unit$add_default_temporal_unit$fn__43988.invoke(add_default_temporal_unit.clj:23)" "query_processor.middleware.desugar$desugar$fn__46686.invoke(desugar.clj:21)" "query_processor.middleware.binning$update_binning_strategy$fn__45505.invoke(binning.clj:227)" "query_processor.middleware.resolve_fields$resolve_fields$fn__46222.invoke(resolve_fields.clj:34)" "query_processor.middleware.add_dimension_projections$add_remapping$fn__44343.invoke(add_dimension_projections.clj:312)" "query_processor.middleware.add_implicit_clauses$add_implicit_clauses$fn__44572.invoke(add_implicit_clauses.clj:147)" "query_processor.middleware.upgrade_field_literals$upgrade_field_literals$fn__49430.invoke(upgrade_field_literals.clj:40)" "query_processor.middleware.add_source_metadata$add_source_metadata_for_source_queries$fn__44857.invoke(add_source_metadata.clj:123)" "query_processor.middleware.reconcile_breakout_and_order_by_bucketing$reconcile_breakout_and_order_by_bucketing$fn__48583.invoke(reconcile_breakout_and_order_by_bucketing.clj:100)" "query_processor.middleware.auto_bucket_datetimes$auto_bucket_datetimes$fn__45065.invoke(auto_bucket_datetimes.clj:147)" "query_processor.middleware.resolve_source_table$resolve_source_tables$fn__46269.invoke(resolve_source_table.clj:45)" "query_processor.middleware.parameters$substitute_parameters$fn__48391.invoke(parameters.clj:111)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__46321.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__46942.invoke(expand_macros.clj:155)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__44866.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__49383.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594$fn__48598.invoke(resolve_database_and_driver.clj:31)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594.invoke(resolve_database_and_driver.clj:25)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__47182.invoke(fetch_source_query.clj:274)" "query_processor.middleware.store$initialize_store$fn__49392$fn__49393.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:44)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__49392.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__49437.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__47318.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__44712.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49368.invoke(results_metadata.clj:147)" "query_processor.middleware.constraints$add_default_userland_constraints$fn__46563.invoke(constraints.clj:42)" "query_processor.middleware.process_userland_query$process_userland_query$fn__48480.invoke(process_userland_query.clj:135)" "query_processor.middleware.catch_exceptions$catch_exceptions$fn__46503.invoke(catch_exceptions.clj:173)" "query_processor.reducible$async_qp$qp_STAR___37989$thunk__37990.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___37989.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___37998$fn__38001.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___37998.invoke(reducible.clj:134)" "query_processor$process_userland_query.invokeStatic(query_processor.clj:241)" "query_processor$process_userland_query.doInvoke(query_processor.clj:237)" "query_processor$fn__49527$process_query_and_save_execution_BANG___49536$fn__49539.invoke(query_processor.clj:253)" "query_processor$fn__49527$process_query_and_save_execution_BANG___49536.invoke(query_processor.clj:245)" "query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580$fn__49583.invoke(query_processor.clj:265)" "query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580.invoke(query_processor.clj:258)" "api.dataset$run_query_async$fn__55784.invoke(dataset.clj:56)" "query_processor.streaming$streaming_response_STAR_$fn__55763$fn__55764.invoke(streaming.clj:72)" "query_processor.streaming$streaming_response_STAR_$fn__55763.invoke(streaming.clj:71)" "async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)" "async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)" "async.streaming_response$do_f_async$fn__16071.invoke(streaming_response.clj:84)"], :context :ad-hoc, :error "Select statement did not produce a ResultSet for native query", :row_count 0, :running_time 0, :ex-data {:sql "-- Metabase:: userID: 1 queryType: native queryHash: 1b671306840b04531e6a4a4fcc56cd0d6dbfdb671c65eb3c14850b139b358a26\ndrop table if exists #myTemporaryTable;\nselect 1;", :driver :sqlserver}, :data {:rows [], :cols []}} 2021-05-21 13:26:33,823 DEBUG middleware.log :: POST /api/dataset 202 [ASYNC: completed] 2.7 s (6 DB calls) App DB connections: 0/7 Jetty threads: 3/50 (4 idle, 0 queued) (50 total active threads) Queries in flight: 1 (0 queued) ``` </details> **Expected behavior** The first statement to be executed and empty resultset to be ignored, and then run second statement and show result from that. This is 0.38.4: ![image](https://user-images.githubusercontent.com/1447303/119129980-e56c9c00-ba37-11eb-9b34-97400025e1f9.png) **Information about your Metabase Installation:** Regression since 0.39.0 - tested 0.38.4 thru 0.39.2 **Additional context** https://discourse.metabase.com/t/sql-queries-broken-in-0-39-0-1/14468
process
empty statement results causes reducer error which causes the entire query to fail describe the bug while multi statements are officially not supported introduced an upgraded driver which now causes empty statement results to produce an error like select statement did not produce a resultset for native query which prevent the entire query from working this is has been reported for sql server but might be a more general problem it might be possible to rewrite some queries to but for some queries this might not be feasible or require a lot of work to reproduce native query sql server drop table if exists mytemporarytable select errors with select statement did not produce a resultset for native query full stacktrace error middleware catch exceptions error processing query null database id started at t via status failed class clojure lang exceptioninfo error error executing query stacktrace driver sql jdbc execute execute reducible query fn invoke execute clj driver sql jdbc execute execute reducible query invokestatic execute clj driver sql jdbc execute execute reducible query invoke execute clj driver sql jdbc execute execute reducible query invokestatic execute clj driver sql jdbc execute execute reducible query invoke execute clj driver sql jdbc fn invokestatic sql jdbc clj driver sql jdbc fn invoke sql jdbc clj query processor context executef invokestatic context clj query processor context executef invoke context clj query processor context default default runf invokestatic default clj query processor context default default runf invoke default clj query processor context runf invokestatic context clj query processor context runf invoke context clj query processor reducible pivot invokestatic reducible clj query processor reducible pivot invoke reducible clj query processor middleware mbql to native mbql gt native fn invoke mbql to native clj query processor middleware check features check features fn invoke check features clj query processor middleware limit limit fn invoke limit clj query processor middleware cache maybe return cached results fn invoke cache clj query processor middleware optimize temporal filters optimize temporal filters fn invoke optimize temporal filters clj query processor middleware validate temporal bucketing validate temporal bucketing fn invoke validate temporal bucketing clj query processor middleware auto parse filter values auto parse filter values fn invoke auto parse filter values clj query processor middleware wrap value literals wrap value literals fn invoke wrap value literals clj query processor middleware annotate add column info fn invoke annotate clj query processor middleware permissions check query permissions fn invoke permissions clj query processor middleware pre alias aggregations pre alias aggregations fn invoke pre alias aggregations clj query processor middleware cumulative aggregations handle cumulative aggregations fn invoke cumulative aggregations clj query processor middleware resolve joined fields resolve joined fields fn invoke resolve joined fields clj query processor middleware resolve joins resolve joins fn invoke resolve joins clj query processor middleware add implicit joins add implicit joins fn invoke add implicit joins clj query processor middleware large int id convert id to string fn invoke large int id clj query processor middleware format rows format rows fn invoke format rows clj query processor middleware add default temporal unit add default temporal unit fn invoke add default temporal unit clj query processor middleware desugar desugar fn invoke desugar clj query processor middleware binning update binning strategy fn invoke binning clj query processor middleware resolve fields resolve fields fn invoke resolve fields clj query processor middleware add dimension projections add remapping fn invoke add dimension projections clj query processor middleware add implicit clauses add implicit clauses fn invoke add implicit clauses clj query processor middleware upgrade field literals upgrade field literals fn invoke upgrade field literals clj query processor middleware add source metadata add source metadata for source queries fn invoke add source metadata clj query processor middleware reconcile breakout and order by bucketing reconcile breakout and order by bucketing fn invoke reconcile breakout and order by bucketing clj query processor middleware auto bucket datetimes auto bucket datetimes fn invoke auto bucket datetimes clj query processor middleware resolve source table resolve source tables fn invoke resolve source table clj query processor middleware parameters substitute parameters fn invoke parameters clj query processor middleware resolve referenced resolve referenced card resources fn invoke resolve referenced clj query processor middleware expand macros expand macros fn invoke expand macros clj query processor middleware add timezone info add timezone info fn invoke add timezone info clj query processor middleware splice params in response splice params in response fn invoke splice params in response clj query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj driver do with driver invokestatic driver clj driver do with driver invoke driver clj query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware validate validate query fn invoke validate clj query processor middleware normalize query normalize fn invoke normalize query clj query processor middleware add rows truncated add rows truncated fn invoke add rows truncated clj query processor middleware results metadata record and return metadata bang fn invoke results metadata clj query processor middleware constraints add default userland constraints fn invoke constraints clj query processor middleware process userland query process userland query fn invoke process userland query clj query processor middleware catch exceptions catch exceptions fn invoke catch exceptions clj query processor reducible async qp qp star thunk invoke reducible clj query processor reducible async qp qp star invoke reducible clj query processor reducible sync qp qp star fn invoke reducible clj query processor reducible sync qp qp star invoke reducible clj query processor process userland query invokestatic query processor clj query processor process userland query doinvoke query processor clj query processor fn process query and save execution bang fn invoke query processor clj query processor fn process query and save execution bang invoke query processor clj query processor fn process query and save with max results constraints bang fn invoke query processor clj query processor fn process query and save with max results constraints bang invoke query processor clj api dataset run query async fn invoke dataset clj query processor streaming streaming response star fn fn invoke streaming clj query processor streaming streaming response star fn invoke streaming clj async streaming response do f star invokestatic streaming response clj async streaming response do f star invoke streaming response clj async streaming response do f async fn invoke streaming response clj error type invalid query ex data sql metabase userid querytype native queryhash ndrop table if exists mytemporarytable nselect params nil type invalid query error type invalid query json query type native native query drop table if exists mytemporarytable nselect template tags database parameters middleware js int to string true add default userland constraints true status failed class clojure lang exceptioninfo stacktrace driver sql jdbc execute fn invokestatic execute clj driver sql jdbc execute fn invoke execute clj driver sql jdbc execute execute statement or prepared statement bang invokestatic execute clj driver sql jdbc execute execute statement or prepared statement bang invoke execute clj driver sql jdbc execute execute reducible query fn invoke execute clj driver sql jdbc execute execute reducible query invokestatic execute clj driver sql jdbc execute execute reducible query invoke execute clj driver sql jdbc execute execute reducible query invokestatic execute clj driver sql jdbc execute execute reducible query invoke execute clj driver sql jdbc fn invokestatic sql jdbc clj driver sql jdbc fn invoke sql jdbc clj query processor context executef invokestatic context clj query processor context executef invoke context clj query processor context default default runf invokestatic default clj query processor context default default runf invoke default clj query processor context runf invokestatic context clj query processor context runf invoke context clj query processor reducible pivot invokestatic reducible clj query processor reducible pivot invoke reducible clj query processor middleware mbql to native mbql gt native fn invoke mbql to native clj query processor middleware check features check features fn invoke check features clj query processor middleware limit limit fn invoke limit clj query processor middleware cache maybe return cached results fn invoke cache clj query processor middleware optimize temporal filters optimize temporal filters fn invoke optimize temporal filters clj query processor middleware validate temporal bucketing validate temporal bucketing fn invoke validate temporal bucketing clj query processor middleware auto parse filter values auto parse filter values fn invoke auto parse filter values clj query processor middleware wrap value literals wrap value literals fn invoke wrap value literals clj query processor middleware annotate add column info fn invoke annotate clj query processor middleware permissions check query permissions fn invoke permissions clj query processor middleware pre alias aggregations pre alias aggregations fn invoke pre alias aggregations clj query processor middleware cumulative aggregations handle cumulative aggregations fn invoke cumulative aggregations clj query processor middleware resolve joined fields resolve joined fields fn invoke resolve joined fields clj query processor middleware resolve joins resolve joins fn invoke resolve joins clj query processor middleware add implicit joins add implicit joins fn invoke add implicit joins clj query processor middleware large int id convert id to string fn invoke large int id clj query processor middleware format rows format rows fn invoke format rows clj query processor middleware add default temporal unit add default temporal unit fn invoke add default temporal unit clj query processor middleware desugar desugar fn invoke desugar clj query processor middleware binning update binning strategy fn invoke binning clj query processor middleware resolve fields resolve fields fn invoke resolve fields clj query processor middleware add dimension projections add remapping fn invoke add dimension projections clj query processor middleware add implicit clauses add implicit clauses fn invoke add implicit clauses clj query processor middleware upgrade field literals upgrade field literals fn invoke upgrade field literals clj query processor middleware add source metadata add source metadata for source queries fn invoke add source metadata clj query processor middleware reconcile breakout and order by bucketing reconcile breakout and order by bucketing fn invoke reconcile breakout and order by bucketing clj query processor middleware auto bucket datetimes auto bucket datetimes fn invoke auto bucket datetimes clj query processor middleware resolve source table resolve source tables fn invoke resolve source table clj query processor middleware parameters substitute parameters fn invoke parameters clj query processor middleware resolve referenced resolve referenced card resources fn invoke resolve referenced clj query processor middleware expand macros expand macros fn invoke expand macros clj query processor middleware add timezone info add timezone info fn invoke add timezone info clj query processor middleware splice params in response splice params in response fn invoke splice params in response clj query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj driver do with driver invokestatic driver clj driver do with driver invoke driver clj query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware validate validate query fn invoke validate clj query processor middleware normalize query normalize fn invoke normalize query clj query processor middleware add rows truncated add rows truncated fn invoke add rows truncated clj query processor middleware results metadata record and return metadata bang fn invoke results metadata clj query processor middleware constraints add default userland constraints fn invoke constraints clj query processor middleware process userland query process userland query fn invoke process userland query clj query processor middleware catch exceptions catch exceptions fn invoke catch exceptions clj query processor reducible async qp qp star thunk invoke reducible clj query processor reducible async qp qp star invoke reducible clj query processor reducible sync qp qp star fn invoke reducible clj query processor reducible sync qp qp star invoke reducible clj query processor process userland query invokestatic query processor clj query processor process userland query doinvoke query processor clj query processor fn process query and save execution bang fn invoke query processor clj query processor fn process query and save execution bang invoke query processor clj query processor fn process query and save with max results constraints bang fn invoke query processor clj query processor fn process query and save with max results constraints bang invoke query processor clj api dataset run query async fn invoke dataset clj query processor streaming streaming response star fn fn invoke streaming clj query processor streaming streaming response star fn invoke streaming clj async streaming response do f star invokestatic streaming response clj async streaming response do f star invoke streaming response clj async streaming response do f async fn invoke streaming response clj context ad hoc error select statement did not produce a resultset for native query row count running time ex data sql metabase userid querytype native queryhash ndrop table if exists mytemporarytable nselect driver sqlserver data rows cols debug middleware log post api dataset s db calls app db connections jetty threads idle queued total active threads queries in flight queued expected behavior the first statement to be executed and empty resultset to be ignored and then run second statement and show result from that this is information about your metabase installation regression since tested thru additional context
1
18,827
24,729,720,012
IssuesEvent
2022-10-20 16:31:38
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Attributes processor is not identified when regex does not match
processor/attributes
**Describe the bug** When a regular expression doesn't match a log line, it fails, but is not identified in the error message, making it hard for the developer to understand which processor entry is failing. **Steps to reproduce** Define a pipeline with two different processors such as: ``` attributes/quorum_log_file: actions: - key: "file_name" pattern: ^geth-(?P<host>.*)\.log action: extract attributes/tessera_log_file: actions: - key: "file_name" pattern: ^tessera-(?P<host>.*)\.log action: extract ``` Add those 2 processors to the same pipeline. Run the pipeline. **What did you expect to see?** I expect to see no errors, or a warning that one of the regular expression has not matched, at most. **What did you see instead?** I see this error: ``` 2021-05-25T17:47:17.107Z ERROR Failed to process entry {"kind": "receiver", "name": "filelog/quorum", "operator_id": "$.regex_parser", "operator_type": "regex_parser", "error": "regex pattern does not match", "action": "send", "entry": {"timestamp":"2021-05-25T17:47:17.107868766Z","severity":0,"attributes":{"file_name":"geth-5b637c3a018d.log"},"body":"DEBUG[05-25|17:47:16.943] Recalculated downloader QoS values rtt=20s confidence=1.000 ttl=1m0s"}} ``` Neither `attributes/quorum_log_file` nor `attributes/tessera_log_file` are named. **What version did you use?** Latest **Environment** OS: Ubuntu 20.04 **Additional context** Additionally, I don't think this should error out at all :)
1.0
Attributes processor is not identified when regex does not match - **Describe the bug** When a regular expression doesn't match a log line, it fails, but is not identified in the error message, making it hard for the developer to understand which processor entry is failing. **Steps to reproduce** Define a pipeline with two different processors such as: ``` attributes/quorum_log_file: actions: - key: "file_name" pattern: ^geth-(?P<host>.*)\.log action: extract attributes/tessera_log_file: actions: - key: "file_name" pattern: ^tessera-(?P<host>.*)\.log action: extract ``` Add those 2 processors to the same pipeline. Run the pipeline. **What did you expect to see?** I expect to see no errors, or a warning that one of the regular expression has not matched, at most. **What did you see instead?** I see this error: ``` 2021-05-25T17:47:17.107Z ERROR Failed to process entry {"kind": "receiver", "name": "filelog/quorum", "operator_id": "$.regex_parser", "operator_type": "regex_parser", "error": "regex pattern does not match", "action": "send", "entry": {"timestamp":"2021-05-25T17:47:17.107868766Z","severity":0,"attributes":{"file_name":"geth-5b637c3a018d.log"},"body":"DEBUG[05-25|17:47:16.943] Recalculated downloader QoS values rtt=20s confidence=1.000 ttl=1m0s"}} ``` Neither `attributes/quorum_log_file` nor `attributes/tessera_log_file` are named. **What version did you use?** Latest **Environment** OS: Ubuntu 20.04 **Additional context** Additionally, I don't think this should error out at all :)
process
attributes processor is not identified when regex does not match describe the bug when a regular expression doesn t match a log line it fails but is not identified in the error message making it hard for the developer to understand which processor entry is failing steps to reproduce define a pipeline with two different processors such as attributes quorum log file actions key file name pattern geth p log action extract attributes tessera log file actions key file name pattern tessera p log action extract add those processors to the same pipeline run the pipeline what did you expect to see i expect to see no errors or a warning that one of the regular expression has not matched at most what did you see instead i see this error error failed to process entry kind receiver name filelog quorum operator id regex parser operator type regex parser error regex pattern does not match action send entry timestamp severity attributes file name geth log body debug recalculated downloader qos values rtt confidence ttl neither attributes quorum log file nor attributes tessera log file are named what version did you use latest environment os ubuntu additional context additionally i don t think this should error out at all
1
989
3,453,993,088
IssuesEvent
2015-12-17 14:09:20
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
opened
Track hostgroup in stats_mysql_query_digest
ADMIN CONNECTION POOL enhancement MYSQL QUERY PROCESSOR ROUTING
That is an enhancement. Currently table stats_mysql_query_digest tracks several metrics, but not hostgroup. Tracking also hostgroup will allow to better understand how efficient is query routing.
1.0
Track hostgroup in stats_mysql_query_digest - That is an enhancement. Currently table stats_mysql_query_digest tracks several metrics, but not hostgroup. Tracking also hostgroup will allow to better understand how efficient is query routing.
process
track hostgroup in stats mysql query digest that is an enhancement currently table stats mysql query digest tracks several metrics but not hostgroup tracking also hostgroup will allow to better understand how efficient is query routing
1
770,756
27,055,136,198
IssuesEvent
2023-02-13 15:41:56
ufosc/Club_Website_2
https://api.github.com/repos/ufosc/Club_Website_2
closed
About button should link to #title_animation
bug/error good first issue help wanted high priority difficulty: easy
**Problem:** the About button in the navigation bar redirects to the `#about` anchor, which no longer exists because the about section was removed. **TODO:** Navigate through the site and ensure that all About buttons link to the `#title_animation` anchor. The first button that comes to mind is in `views/components/navbar.ejs` - but there might be more.
1.0
About button should link to #title_animation - **Problem:** the About button in the navigation bar redirects to the `#about` anchor, which no longer exists because the about section was removed. **TODO:** Navigate through the site and ensure that all About buttons link to the `#title_animation` anchor. The first button that comes to mind is in `views/components/navbar.ejs` - but there might be more.
non_process
about button should link to title animation problem the about button in the navigation bar redirects to the about anchor which no longer exists because the about section was removed todo navigate through the site and ensure that all about buttons link to the title animation anchor the first button that comes to mind is in views components navbar ejs but there might be more
0
790,362
27,823,868,555
IssuesEvent
2023-03-19 14:45:32
bounswe/bounswe2023group7
https://api.github.com/repos/bounswe/bounswe2023group7
closed
Research about Mock-Up
type : research priority : high status : need_review effort : low
### Issue Description I'm going to research what is Mock-Up, why is it useful, why should we have it. Then I'm going to document it. ### Tasks _No response_ ### Extra Comments _No response_ ### Final Step _No response_ ### Deadline of the Issue 19.03.2023 - 23.59 ### Reviewer @melihgezerr ### Deadline for the Review 20.03.2023 - 23.59
1.0
Research about Mock-Up - ### Issue Description I'm going to research what is Mock-Up, why is it useful, why should we have it. Then I'm going to document it. ### Tasks _No response_ ### Extra Comments _No response_ ### Final Step _No response_ ### Deadline of the Issue 19.03.2023 - 23.59 ### Reviewer @melihgezerr ### Deadline for the Review 20.03.2023 - 23.59
non_process
research about mock up issue description i m going to research what is mock up why is it useful why should we have it then i m going to document it tasks no response extra comments no response final step no response deadline of the issue reviewer melihgezerr deadline for the review
0
18,673
24,591,535,134
IssuesEvent
2022-10-14 03:01:56
Tencent/tdesign-miniprogram
https://api.github.com/repos/Tencent/tdesign-miniprogram
closed
[t-tabs] 结合t-navbar时使用sticky属性导致标签无法以t-navbar下边缘为吸顶位置
good first issue in process
### tdesign-miniprogram 版本 0.22.0-beta.0 ### 重现链接 _No response_ ### 重现步骤 ``` html <t-navbar   title="标题"   leftIcon="chevron-left"   homeIcon="home"   bind:go-home="onGoHome"   t-class-left-icon="left-icon-back"   t-class-home-icon="home-icon"   t-class-title="nav-title" /> <t-tabs defaultValue="{{0}}" sticky="{{true}}" bind:change="onTabsChange" bind:click="onTabsClick" bind:scroll="onStickyScroll" t-class="custom-tabs" > <t-tab-panel label="全部" value="0"> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>bb</view> </t-tab-panel> <t-tab-panel label="未完成" value="1"> <view>未完成</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>未完成</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>未完成</view> </t-tab-panel> <t-tab-panel label="已完成" value="2">已完成</t-tab-panel> </t-tabs> ``` ### 期望结果 选项卡上边界以t-navbar下边界自动吸顶 ### 实际结果 选项卡没有吸顶效果。 ### 框架版本 0.22.0-beta.0 ### 浏览器版本 _No response_ ### 系统版本 _No response_ ### Node版本 _No response_ ### 补充说明 _No response_
1.0
[t-tabs] 结合t-navbar时使用sticky属性导致标签无法以t-navbar下边缘为吸顶位置 - ### tdesign-miniprogram 版本 0.22.0-beta.0 ### 重现链接 _No response_ ### 重现步骤 ``` html <t-navbar   title="标题"   leftIcon="chevron-left"   homeIcon="home"   bind:go-home="onGoHome"   t-class-left-icon="left-icon-back"   t-class-home-icon="home-icon"   t-class-title="nav-title" /> <t-tabs defaultValue="{{0}}" sticky="{{true}}" bind:change="onTabsChange" bind:click="onTabsClick" bind:scroll="onStickyScroll" t-class="custom-tabs" > <t-tab-panel label="全部" value="0"> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>bb</view> </t-tab-panel> <t-tab-panel label="未完成" value="1"> <view>未完成</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view> <view>未完成</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>未完成</view> </t-tab-panel> <t-tab-panel label="已完成" value="2">已完成</t-tab-panel> </t-tabs> ``` ### 期望结果 选项卡上边界以t-navbar下边界自动吸顶 ### 实际结果 选项卡没有吸顶效果。 ### 框架版本 0.22.0-beta.0 ### 浏览器版本 _No response_ ### 系统版本 _No response_ ### Node版本 _No response_ ### 补充说明 _No response_
process
结合t navbar时使用sticky属性导致标签无法以t navbar下边缘为吸顶位置 tdesign miniprogram 版本 beta 重现链接 no response 重现步骤 html t navbar   title 标题   lefticon chevron left   homeicon home   bind go home ongohome   t class left icon left icon back   t class home icon home icon   t class title nav title t tabs defaultvalue sticky true bind change ontabschange bind click ontabsclick bind scroll onstickyscroll t class custom tabs aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa bb 未完成 aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa 未完成 aaaa aaaa aaaa 未完成 已完成 期望结果 选项卡上边界以t navbar下边界自动吸顶 实际结果 选项卡没有吸顶效果。 框架版本 beta 浏览器版本 no response 系统版本 no response node版本 no response 补充说明 no response
1
13,445
3,332,927,738
IssuesEvent
2015-11-11 22:21:57
t2davis/Squad
https://api.github.com/repos/t2davis/Squad
opened
buddybuild feedback from Mariajfdavis@gmail.com on build #12: The keyboard stays on the screen when you go to the map if u…
1.0 BetaTester
Feedback from Mariajfdavis@gmail.com : The keyboard stays on the screen when you go to the map if u had it open in messages <img src="https://s3-us-west-2.amazonaws.com/buddybuild-screenshots/5642b68257de0f0100c2f1c9/5643a14fbbd0c90100b82c5f/c543d717-8402-4b0b-9f6e-a90062b6e181.jpg" width="33%" height="33%" />
1.0
buddybuild feedback from Mariajfdavis@gmail.com on build #12: The keyboard stays on the screen when you go to the map if u… - Feedback from Mariajfdavis@gmail.com : The keyboard stays on the screen when you go to the map if u had it open in messages <img src="https://s3-us-west-2.amazonaws.com/buddybuild-screenshots/5642b68257de0f0100c2f1c9/5643a14fbbd0c90100b82c5f/c543d717-8402-4b0b-9f6e-a90062b6e181.jpg" width="33%" height="33%" />
non_process
buddybuild feedback from mariajfdavis gmail com on build the keyboard stays on the screen when you go to the map if u… feedback from mariajfdavis gmail com the keyboard stays on the screen when you go to the map if u had it open in messages
0
555,743
16,468,898,141
IssuesEvent
2021-05-23 01:56:50
googleapis/google-api-go-client
https://api.github.com/repos/googleapis/google-api-go-client
closed
Google Youtube API V3 PlaylistItemSnippet Position 0 Is Ignored
priority: p2 type: bug
#### Environment details - Programming language: Golang - OS: WSL2 Windows - Language runtime version: go version go1.16 linux/amd64 - Package version: 0.47.0 #### Steps to reproduce 1. Create a playlist 2. Add video to a playlist 3. Add video to a playlist at position 0 4. Notice that it is added to the end of the playlist rather than the first position. ``` &youtube.PlaylistItem{ Snippet: &youtube.PlaylistItemSnippet{ Position: 0, }, } ``` #### Note PlaylistItemSnippet Position is int64 rather than *int64 like Position in ChannelSectionSnippet. As such it is ignored with by `json:"omitempty"`.
1.0
Google Youtube API V3 PlaylistItemSnippet Position 0 Is Ignored - #### Environment details - Programming language: Golang - OS: WSL2 Windows - Language runtime version: go version go1.16 linux/amd64 - Package version: 0.47.0 #### Steps to reproduce 1. Create a playlist 2. Add video to a playlist 3. Add video to a playlist at position 0 4. Notice that it is added to the end of the playlist rather than the first position. ``` &youtube.PlaylistItem{ Snippet: &youtube.PlaylistItemSnippet{ Position: 0, }, } ``` #### Note PlaylistItemSnippet Position is int64 rather than *int64 like Position in ChannelSectionSnippet. As such it is ignored with by `json:"omitempty"`.
non_process
google youtube api playlistitemsnippet position is ignored environment details programming language golang os windows language runtime version go version linux package version steps to reproduce create a playlist add video to a playlist add video to a playlist at position notice that it is added to the end of the playlist rather than the first position youtube playlistitem snippet youtube playlistitemsnippet position note playlistitemsnippet position is rather than like position in channelsectionsnippet as such it is ignored with by json omitempty
0
19,596
3,776,504,971
IssuesEvent
2016-03-17 16:52:27
leeensminger/DelDOT-NPDES-Field-Tool
https://api.github.com/repos/leeensminger/DelDOT-NPDES-Field-Tool
closed
Can buffer for drawing barrel conveyances be larger?
fixed will be ready for testing in next release.
When there are more than two barrels next to each other, the buffer used to select the upstream/downstream structures is not large enough to encompass the structures. Could the buffer be larger for just barrel conveyances? ![image](https://cloud.githubusercontent.com/assets/16919958/12978974/362b8848-d0a2-11e5-9625-b9548fe706d6.png)
1.0
Can buffer for drawing barrel conveyances be larger? - When there are more than two barrels next to each other, the buffer used to select the upstream/downstream structures is not large enough to encompass the structures. Could the buffer be larger for just barrel conveyances? ![image](https://cloud.githubusercontent.com/assets/16919958/12978974/362b8848-d0a2-11e5-9625-b9548fe706d6.png)
non_process
can buffer for drawing barrel conveyances be larger when there are more than two barrels next to each other the buffer used to select the upstream downstream structures is not large enough to encompass the structures could the buffer be larger for just barrel conveyances
0
4,516
7,360,185,455
IssuesEvent
2018-03-10 16:00:20
ODiogoSilva/assemblerflow
https://api.github.com/repos/ODiogoSilva/assemblerflow
opened
Add container attribute to Process
enhancement process
Add the `container` attribute to the Process base class. This will be used to determine the container directive in the nextflow configuration file.
1.0
Add container attribute to Process - Add the `container` attribute to the Process base class. This will be used to determine the container directive in the nextflow configuration file.
process
add container attribute to process add the container attribute to the process base class this will be used to determine the container directive in the nextflow configuration file
1
20,960
27,817,510,295
IssuesEvent
2023-03-18 21:19:10
cse442-at-ub/project_s23-iweatherify
https://api.github.com/repos/cse442-at-ub/project_s23-iweatherify
closed
Save the units and temperature settings to the database
Processing Task Sprint 2
**Task Tests** *Test 1* 1. Go to the following URL: https://github.com/cse442-at-ub/project_s23-iweatherify/tree/dev 2. Click on the green `<> Code` button and download the ZIP file. ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/75c741f7-5ca3-4f64-879f-df960ad51a8b) 3. Unzip the downloaded file to a folder on your computer. 4. Open a terminal and navigate to the git repository folder using the `cd` command. 5. Run the `npm install` command in the terminal to install the necessary dependencies. 6. Run the `npm start` command in the terminal to start the application. 7. Check the output from the npm start command for the URL to access the application. The URL should be a localhost address (e.g., http://localhost:8080). 8. Navigate to http://localhost:8080/#/login 9. Ensure you have logged in to our app to see the page use UserID: `UB442` and Password:`Myub442@!` to login 10. Go to URL: http://localhost:8080/#/unitsSettings 11. Verify that the units page is displayed ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/3f9cbb63-e582-43f0-91bf-16970daf57eb) 12. Change the temperature unit to Celsius (°C) 13. Change the wind unit to km/h 14. Change the pressure unit to mm 15. Change the distance unit to km 16. Open the browser inspector tool and select console 17. Click the save button 18. You should see the message: `Units saved successfully.` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/ee5e91dc-9854-4474-b931-329a77f89996) 19. You should see the message: `{message: 'User settings saved successfully.'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/f6c50295-ae68-4e31-a536-93bc29bea727) 18. Open a different tab and go to: https://www-student.cse.buffalo.edu/tools/db/phpmyadmin/index.php 19. Input username: `jpan26` and password: `50314999` 20. Make sure the server choice is `oceanus.cse.buffalo.edu:3306` 21. Click go and you should see this page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/443bf59a-f5fd-4dbe-88a8-645493eaa713) 22. Click `cse442_2023_spring_team_a_db` first and then `saved_units` on the left side of the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/9cb87c0a-e93f-4327-9640-8f6c0478c3a2) 23. Verify you see a row with the exact same information as shown by the picture ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/fefc62d6-7c3f-4da5-948a-49b0de95a56d) *Test 2* 1. Repeat steps 1 to 9 from `Test 1` 2. Go to URL: http://localhost:8080/#/tempSettings 3. Verify that the temperature setting page is displayed ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/1feae072-4846-46ae-8639-9958248e9158) 4. Open the browser inspector tool and select console 5. Change the hot temperature to 80, you can either use the slider or input box and click save 6. You should see the message: `{result: 'success'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/689287f3-8376-4ed4-abd6-ff73816eb604) 7. You should see the message: `Temperatures Saved Successfully` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/30bea4ec-67c2-4410-855a-6ee558b521f7) 8. Change the warm temperature to 65, you can either use the slider or input box and click save 9. You should see the message: `{result: 'success'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/689287f3-8376-4ed4-abd6-ff73816eb604) 10. You should see the message: `Temperatures Saved Successfully` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/30bea4ec-67c2-4410-855a-6ee558b521f7) 11. Change the ideal temperature to 50, you can either use the slider or input box and click save 12. You should see the message: `{result: 'success'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/689287f3-8376-4ed4-abd6-ff73816eb604) 13. You should see the message: `Temperatures Saved Successfully` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/30bea4ec-67c2-4410-855a-6ee558b521f7) 14. Change the chilly temperature to 0, you can either use the slider or input box and click save 15. You should see the message: `{result: 'success'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/689287f3-8376-4ed4-abd6-ff73816eb604) 16. You should see the message: `Temperatures Saved Successfully` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/30bea4ec-67c2-4410-855a-6ee558b521f7) 17. Change the cold temperature to -65, you can either use the slider or input box and click save 18. You should see the message: `{result: 'success'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/689287f3-8376-4ed4-abd6-ff73816eb604) 19. You should see the message: `Temperatures Saved Successfully` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/30bea4ec-67c2-4410-855a-6ee558b521f7) 20. Change the freezing temperature to -80, you can either use the slider or input box and click save 21. You should see the message: `{result: 'success'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/689287f3-8376-4ed4-abd6-ff73816eb604) 22. You should see the message: `Temperatures Saved Successfully` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/30bea4ec-67c2-4410-855a-6ee558b521f7) 23. Repeat steps 18 to 21 from `Test 1` 24. Click `cse442_2023_spring_team_a_db` first and then `saved_temperatures` on the left side of the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/ae9e6ed8-1291-4690-89ab-ce8cb4223843) 25. Verify you see a row with the exact same information as shown by the picture ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/9331cadd-35f2-4ac4-b366-44722b56430e)
1.0
Save the units and temperature settings to the database - **Task Tests** *Test 1* 1. Go to the following URL: https://github.com/cse442-at-ub/project_s23-iweatherify/tree/dev 2. Click on the green `<> Code` button and download the ZIP file. ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/75c741f7-5ca3-4f64-879f-df960ad51a8b) 3. Unzip the downloaded file to a folder on your computer. 4. Open a terminal and navigate to the git repository folder using the `cd` command. 5. Run the `npm install` command in the terminal to install the necessary dependencies. 6. Run the `npm start` command in the terminal to start the application. 7. Check the output from the npm start command for the URL to access the application. The URL should be a localhost address (e.g., http://localhost:8080). 8. Navigate to http://localhost:8080/#/login 9. Ensure you have logged in to our app to see the page use UserID: `UB442` and Password:`Myub442@!` to login 10. Go to URL: http://localhost:8080/#/unitsSettings 11. Verify that the units page is displayed ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/3f9cbb63-e582-43f0-91bf-16970daf57eb) 12. Change the temperature unit to Celsius (°C) 13. Change the wind unit to km/h 14. Change the pressure unit to mm 15. Change the distance unit to km 16. Open the browser inspector tool and select console 17. Click the save button 18. You should see the message: `Units saved successfully.` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/ee5e91dc-9854-4474-b931-329a77f89996) 19. You should see the message: `{message: 'User settings saved successfully.'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/f6c50295-ae68-4e31-a536-93bc29bea727) 18. Open a different tab and go to: https://www-student.cse.buffalo.edu/tools/db/phpmyadmin/index.php 19. Input username: `jpan26` and password: `50314999` 20. Make sure the server choice is `oceanus.cse.buffalo.edu:3306` 21. Click go and you should see this page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/443bf59a-f5fd-4dbe-88a8-645493eaa713) 22. Click `cse442_2023_spring_team_a_db` first and then `saved_units` on the left side of the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/9cb87c0a-e93f-4327-9640-8f6c0478c3a2) 23. Verify you see a row with the exact same information as shown by the picture ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/fefc62d6-7c3f-4da5-948a-49b0de95a56d) *Test 2* 1. Repeat steps 1 to 9 from `Test 1` 2. Go to URL: http://localhost:8080/#/tempSettings 3. Verify that the temperature setting page is displayed ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/1feae072-4846-46ae-8639-9958248e9158) 4. Open the browser inspector tool and select console 5. Change the hot temperature to 80, you can either use the slider or input box and click save 6. You should see the message: `{result: 'success'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/689287f3-8376-4ed4-abd6-ff73816eb604) 7. You should see the message: `Temperatures Saved Successfully` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/30bea4ec-67c2-4410-855a-6ee558b521f7) 8. Change the warm temperature to 65, you can either use the slider or input box and click save 9. You should see the message: `{result: 'success'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/689287f3-8376-4ed4-abd6-ff73816eb604) 10. You should see the message: `Temperatures Saved Successfully` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/30bea4ec-67c2-4410-855a-6ee558b521f7) 11. Change the ideal temperature to 50, you can either use the slider or input box and click save 12. You should see the message: `{result: 'success'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/689287f3-8376-4ed4-abd6-ff73816eb604) 13. You should see the message: `Temperatures Saved Successfully` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/30bea4ec-67c2-4410-855a-6ee558b521f7) 14. Change the chilly temperature to 0, you can either use the slider or input box and click save 15. You should see the message: `{result: 'success'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/689287f3-8376-4ed4-abd6-ff73816eb604) 16. You should see the message: `Temperatures Saved Successfully` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/30bea4ec-67c2-4410-855a-6ee558b521f7) 17. Change the cold temperature to -65, you can either use the slider or input box and click save 18. You should see the message: `{result: 'success'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/689287f3-8376-4ed4-abd6-ff73816eb604) 19. You should see the message: `Temperatures Saved Successfully` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/30bea4ec-67c2-4410-855a-6ee558b521f7) 20. Change the freezing temperature to -80, you can either use the slider or input box and click save 21. You should see the message: `{result: 'success'}` in the console ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/689287f3-8376-4ed4-abd6-ff73816eb604) 22. You should see the message: `Temperatures Saved Successfully` on the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/30bea4ec-67c2-4410-855a-6ee558b521f7) 23. Repeat steps 18 to 21 from `Test 1` 24. Click `cse442_2023_spring_team_a_db` first and then `saved_temperatures` on the left side of the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/ae9e6ed8-1291-4690-89ab-ce8cb4223843) 25. Verify you see a row with the exact same information as shown by the picture ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/9331cadd-35f2-4ac4-b366-44722b56430e)
process
save the units and temperature settings to the database task tests test go to the following url click on the green code button and download the zip file unzip the downloaded file to a folder on your computer open a terminal and navigate to the git repository folder using the cd command run the npm install command in the terminal to install the necessary dependencies run the npm start command in the terminal to start the application check the output from the npm start command for the url to access the application the url should be a localhost address e g navigate to ensure you have logged in to our app to see the page use userid and password to login go to url verify that the units page is displayed change the temperature unit to celsius °c change the wind unit to km h change the pressure unit to mm change the distance unit to km open the browser inspector tool and select console click the save button you should see the message units saved successfully on the page you should see the message message user settings saved successfully in the console open a different tab and go to input username and password make sure the server choice is oceanus cse buffalo edu click go and you should see this page click spring team a db first and then saved units on the left side of the page verify you see a row with the exact same information as shown by the picture test repeat steps to from test go to url verify that the temperature setting page is displayed open the browser inspector tool and select console change the hot temperature to you can either use the slider or input box and click save you should see the message result success in the console you should see the message temperatures saved successfully on the page change the warm temperature to you can either use the slider or input box and click save you should see the message result success in the console you should see the message temperatures saved successfully on the page change the ideal temperature to you can either use the slider or input box and click save you should see the message result success in the console you should see the message temperatures saved successfully on the page change the chilly temperature to you can either use the slider or input box and click save you should see the message result success in the console you should see the message temperatures saved successfully on the page change the cold temperature to you can either use the slider or input box and click save you should see the message result success in the console you should see the message temperatures saved successfully on the page change the freezing temperature to you can either use the slider or input box and click save you should see the message result success in the console you should see the message temperatures saved successfully on the page repeat steps to from test click spring team a db first and then saved temperatures on the left side of the page verify you see a row with the exact same information as shown by the picture
1
67,601
27,960,815,404
IssuesEvent
2023-03-24 15:31:39
LiskHQ/lisk-service
https://api.github.com/repos/LiskHQ/lisk-service
opened
Add functional tests for app-registry
type: test service/blockchain-app-registry
### Description Add missing functional tests for app-registry. ### Acceptance Criteria - Functional tests are added for the following methods in `services/blockchain-app-registry/shared/metadataIndex.js` - indexAppMeta - indexMetadataFromFile - deleteAppMeta - deleteTokensMeta - deleteIndexedMetadataFromFile - indexAllBlockchainAppsMeta
1.0
Add functional tests for app-registry - ### Description Add missing functional tests for app-registry. ### Acceptance Criteria - Functional tests are added for the following methods in `services/blockchain-app-registry/shared/metadataIndex.js` - indexAppMeta - indexMetadataFromFile - deleteAppMeta - deleteTokensMeta - deleteIndexedMetadataFromFile - indexAllBlockchainAppsMeta
non_process
add functional tests for app registry description add missing functional tests for app registry acceptance criteria functional tests are added for the following methods in services blockchain app registry shared metadataindex js indexappmeta indexmetadatafromfile deleteappmeta deletetokensmeta deleteindexedmetadatafromfile indexallblockchainappsmeta
0
32,167
13,770,403,330
IssuesEvent
2020-10-07 20:10:45
Azure/azure-sdk-for-python
https://api.github.com/repos/Azure/azure-sdk-for-python
opened
[Servicebus] Docstring review and bolstering pre-GA
Client Service Bus
pre-GA fit-and-finish. Review all of our docstrings. Ensure text is clear, accurate, up-to-date and comprehensive. Things to focus on: - are all params covered? - rval and returns? (both defined type and description) - custom errors being thrown? - typedefs in all the places? (both # type: and docstring types)
1.0
[Servicebus] Docstring review and bolstering pre-GA - pre-GA fit-and-finish. Review all of our docstrings. Ensure text is clear, accurate, up-to-date and comprehensive. Things to focus on: - are all params covered? - rval and returns? (both defined type and description) - custom errors being thrown? - typedefs in all the places? (both # type: and docstring types)
non_process
docstring review and bolstering pre ga pre ga fit and finish review all of our docstrings ensure text is clear accurate up to date and comprehensive things to focus on are all params covered rval and returns both defined type and description custom errors being thrown typedefs in all the places both type and docstring types
0
35,643
14,789,886,763
IssuesEvent
2021-01-12 11:12:27
terraform-providers/terraform-provider-azurerm
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
closed
RoleAssignmentNotFound with azurerm_role_assignment
bug service/authorization
### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform (and AzureRM Provider) Version Terraform v0.13.3 + provider registry.terraform.io/hashicorp/azurerm v2.32.0 + provider registry.terraform.io/hashicorp/external v2.0.0 ### Affected Resource(s) * `azurerm_role_assigment` ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "azurerm_user_assigned_identity" "tc" { name = "tc" resource_group_name = azurerm_resource_group.rg.name location = azurerm_resource_group.rg.location tags = local.default_tags } resource "azurerm_role_assignment" "tc-acr" { scope = data.azurerm_resources.acr.resources[0].id role_definition_name = "AcrPull" principal_id = azurerm_user_assigned_identity.tc.principal_id skip_service_principal_aad_check = true } resource "azurerm_role_assignment" "tc-privatedns" { scope = azurerm_resource_group.rg.id role_definition_name = "Private DNS Zone Contributor" principal_id = azurerm_user_assigned_identity.tc.principal_id skip_service_principal_aad_check = true } data "azurerm_resource_group" "dns" { name = "dns" } resource "azurerm_role_assignment" "tc-dns" { scope = data.azurerm_resource_group.dns.id role_definition_name = "DNS Zone Contributor" principal_id = azurerm_user_assigned_identity.tc.principal_id skip_service_principal_aad_check = true } resource "azurerm_role_assignment" "tc-vm" { scope = azurerm_resource_group.rg.id role_definition_name = "Virtual Machine Contributor" principal_id = azurerm_user_assigned_identity.tc.principal_id skip_service_principal_aad_check = true } resource "azurerm_role_assignment" "tc-monitoring" { scope = azurerm_resource_group.rg.id role_definition_name = "Monitoring Reader" principal_id = azurerm_user_assigned_identity.traffic-control.principal_id skip_service_principal_aad_check = true } resource "azurerm_role_assignment" "tc-aks" { scope = azurerm_resource_group.rg.id role_definition_name = "Azure Kubernetes Service Cluster User Role" principal_id = azurerm_user_assigned_identity.traffic-control.principal_id skip_service_principal_aad_check = true } resource "azurerm_linux_virtual_machine_scale_set" "tc" { name = "tc" identity { type = "UserAssigned" identity_ids = [azurerm_user_assigned_identity.tc.id] } ... depends_on = [ azurerm_role_assignment.tc-acr, azurerm_role_assignment.tc-privatedns, azurerm_role_assignment.tc-dns, azurerm_role_assignment.tc-vm, azurerm_role_assignment.tc-monitoring, azurerm_role_assignment.tc-aks ] } resource "azurerm_linux_virtual_machine_scale_set" "te" { for_each = var.availability_zones name = "te-${each.value}" identity { type = "UserAssigned" identity_ids = [azurerm_user_assigned_identity.te.id] } depends_on = [ azurerm_role_assignment.te-acr ] } ``` ### Debug Output ``` Still creating... [2m0s elapsed] azurerm_linux_virtual_machine_scale_set.te["3"]: Still creating... [2m10s elapsed] azurerm_linux_virtual_machine_scale_set.traffic-envoy["3"]: Creation complete after 2m15s [id=/subscriptions/xxxx/resourceGroups/yyyy/providers/Microsoft.Compute/virtualMachineScaleSets/te-3] Error: authorization.RoleAssignmentsClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code="RoleAssignmentNotFound" Message="The role assignment '920ffd05-df3c-308b-2e01-c8c58481998e' is not found." on tc.tf line 8, in resource "azurerm_role_assignment" "tc-acr": 8: resource "azurerm_role_assignment" "tc-acr" { Error: authorization.RoleAssignmentsClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code="RoleAssignmentNotFound" Message="The role assignment '1fda1763-4e9f-2c9f-b0a3-a2581b68e457' is not found." on tc.tf line 26, in resource "azurerm_role_assignment" "tc-dns": 26: resource \\\"azurerm_role_assignment\\\" \\\"tc-dns\\\" { Error: authorization.RoleAssignmentsClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code=\\\"RoleAssignmentNotFound\\\" Message=\\\"The role assignment '8f820931-964c-84bf-ecc7-f3c2385fb6a9' is not found." on tc.tf line 47, in resource "azurerm_role_assignment" "tc-aks": 47: resource \\\"azurerm_role_assignment\\\" "tc-aks" ``` ### Expected Behaviour Success ### Actual Behaviour Failure ### Important Factoids Before I placed the dependency between the VMSS and role assignments, the failure between tc role assignment would occur on the tc VMSS. Once the dependency was added it shifted to the te VMSS. Not all role assignments fail, its usually 2 or 3, guess its timing related. * #0000
1.0
RoleAssignmentNotFound with azurerm_role_assignment - ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform (and AzureRM Provider) Version Terraform v0.13.3 + provider registry.terraform.io/hashicorp/azurerm v2.32.0 + provider registry.terraform.io/hashicorp/external v2.0.0 ### Affected Resource(s) * `azurerm_role_assigment` ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "azurerm_user_assigned_identity" "tc" { name = "tc" resource_group_name = azurerm_resource_group.rg.name location = azurerm_resource_group.rg.location tags = local.default_tags } resource "azurerm_role_assignment" "tc-acr" { scope = data.azurerm_resources.acr.resources[0].id role_definition_name = "AcrPull" principal_id = azurerm_user_assigned_identity.tc.principal_id skip_service_principal_aad_check = true } resource "azurerm_role_assignment" "tc-privatedns" { scope = azurerm_resource_group.rg.id role_definition_name = "Private DNS Zone Contributor" principal_id = azurerm_user_assigned_identity.tc.principal_id skip_service_principal_aad_check = true } data "azurerm_resource_group" "dns" { name = "dns" } resource "azurerm_role_assignment" "tc-dns" { scope = data.azurerm_resource_group.dns.id role_definition_name = "DNS Zone Contributor" principal_id = azurerm_user_assigned_identity.tc.principal_id skip_service_principal_aad_check = true } resource "azurerm_role_assignment" "tc-vm" { scope = azurerm_resource_group.rg.id role_definition_name = "Virtual Machine Contributor" principal_id = azurerm_user_assigned_identity.tc.principal_id skip_service_principal_aad_check = true } resource "azurerm_role_assignment" "tc-monitoring" { scope = azurerm_resource_group.rg.id role_definition_name = "Monitoring Reader" principal_id = azurerm_user_assigned_identity.traffic-control.principal_id skip_service_principal_aad_check = true } resource "azurerm_role_assignment" "tc-aks" { scope = azurerm_resource_group.rg.id role_definition_name = "Azure Kubernetes Service Cluster User Role" principal_id = azurerm_user_assigned_identity.traffic-control.principal_id skip_service_principal_aad_check = true } resource "azurerm_linux_virtual_machine_scale_set" "tc" { name = "tc" identity { type = "UserAssigned" identity_ids = [azurerm_user_assigned_identity.tc.id] } ... depends_on = [ azurerm_role_assignment.tc-acr, azurerm_role_assignment.tc-privatedns, azurerm_role_assignment.tc-dns, azurerm_role_assignment.tc-vm, azurerm_role_assignment.tc-monitoring, azurerm_role_assignment.tc-aks ] } resource "azurerm_linux_virtual_machine_scale_set" "te" { for_each = var.availability_zones name = "te-${each.value}" identity { type = "UserAssigned" identity_ids = [azurerm_user_assigned_identity.te.id] } depends_on = [ azurerm_role_assignment.te-acr ] } ``` ### Debug Output ``` Still creating... [2m0s elapsed] azurerm_linux_virtual_machine_scale_set.te["3"]: Still creating... [2m10s elapsed] azurerm_linux_virtual_machine_scale_set.traffic-envoy["3"]: Creation complete after 2m15s [id=/subscriptions/xxxx/resourceGroups/yyyy/providers/Microsoft.Compute/virtualMachineScaleSets/te-3] Error: authorization.RoleAssignmentsClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code="RoleAssignmentNotFound" Message="The role assignment '920ffd05-df3c-308b-2e01-c8c58481998e' is not found." on tc.tf line 8, in resource "azurerm_role_assignment" "tc-acr": 8: resource "azurerm_role_assignment" "tc-acr" { Error: authorization.RoleAssignmentsClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code="RoleAssignmentNotFound" Message="The role assignment '1fda1763-4e9f-2c9f-b0a3-a2581b68e457' is not found." on tc.tf line 26, in resource "azurerm_role_assignment" "tc-dns": 26: resource \\\"azurerm_role_assignment\\\" \\\"tc-dns\\\" { Error: authorization.RoleAssignmentsClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code=\\\"RoleAssignmentNotFound\\\" Message=\\\"The role assignment '8f820931-964c-84bf-ecc7-f3c2385fb6a9' is not found." on tc.tf line 47, in resource "azurerm_role_assignment" "tc-aks": 47: resource \\\"azurerm_role_assignment\\\" "tc-aks" ``` ### Expected Behaviour Success ### Actual Behaviour Failure ### Important Factoids Before I placed the dependency between the VMSS and role assignments, the failure between tc role assignment would occur on the tc VMSS. Once the dependency was added it shifted to the te VMSS. Not all role assignments fail, its usually 2 or 3, guess its timing related. * #0000
non_process
roleassignmentnotfound with azurerm role assignment community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform and azurerm provider version terraform provider registry terraform io hashicorp azurerm provider registry terraform io hashicorp external affected resource s azurerm role assigment terraform configuration files hcl resource azurerm user assigned identity tc name tc resource group name azurerm resource group rg name location azurerm resource group rg location tags local default tags resource azurerm role assignment tc acr scope data azurerm resources acr resources id role definition name acrpull principal id azurerm user assigned identity tc principal id skip service principal aad check true resource azurerm role assignment tc privatedns scope azurerm resource group rg id role definition name private dns zone contributor principal id azurerm user assigned identity tc principal id skip service principal aad check true data azurerm resource group dns name dns resource azurerm role assignment tc dns scope data azurerm resource group dns id role definition name dns zone contributor principal id azurerm user assigned identity tc principal id skip service principal aad check true resource azurerm role assignment tc vm scope azurerm resource group rg id role definition name virtual machine contributor principal id azurerm user assigned identity tc principal id skip service principal aad check true resource azurerm role assignment tc monitoring scope azurerm resource group rg id role definition name monitoring reader principal id azurerm user assigned identity traffic control principal id skip service principal aad check true resource azurerm role assignment tc aks scope azurerm resource group rg id role definition name azure kubernetes service cluster user role principal id azurerm user assigned identity traffic control principal id skip service principal aad check true resource azurerm linux virtual machine scale set tc name tc identity type userassigned identity ids depends on azurerm role assignment tc acr azurerm role assignment tc privatedns azurerm role assignment tc dns azurerm role assignment tc vm azurerm role assignment tc monitoring azurerm role assignment tc aks resource azurerm linux virtual machine scale set te for each var availability zones name te each value identity type userassigned identity ids depends on azurerm role assignment te acr debug output still creating azurerm linux virtual machine scale set te still creating azurerm linux virtual machine scale set traffic envoy creation complete after error authorization roleassignmentsclient get failure responding to request statuscode original error autorest azure service returned an error status code roleassignmentnotfound message the role assignment is not found on tc tf line in resource azurerm role assignment tc acr resource azurerm role assignment tc acr error authorization roleassignmentsclient get failure responding to request statuscode original error autorest azure service returned an error status code roleassignmentnotfound message the role assignment is not found on tc tf line in resource azurerm role assignment tc dns resource azurerm role assignment tc dns error authorization roleassignmentsclient get failure responding to request statuscode original error autorest azure service returned an error status code roleassignmentnotfound message the role assignment is not found on tc tf line in resource azurerm role assignment tc aks resource azurerm role assignment tc aks expected behaviour success actual behaviour failure important factoids before i placed the dependency between the vmss and role assignments the failure between tc role assignment would occur on the tc vmss once the dependency was added it shifted to the te vmss not all role assignments fail its usually or guess its timing related
0
4,737
7,594,724,809
IssuesEvent
2018-04-27 00:49:35
agroportal/agroportal_web_ui
https://api.github.com/repos/agroportal/agroportal_web_ui
closed
Biorefinery & Transmat failed to parse
ontology processing problem
Error from parsing log file (Biorefinery): Illegal rdf:nodeID value '_:genid259' there is an equivalent error for Transmat. This error seems to have been identified in the NCBO BioPortal: see - [https://github.com/ncbo/bioportal-project/issues/32#event-1226205997](https://github.com/ncbo/bioportal-project/issues/32#event-1226205997) - [https://github.com/ncbo/bioportal-project/issues/9](https://github.com/ncbo/bioportal-project/issues/9) - https://sourceforge.net/p/owlapi/mailman/owlapi-developer/thread/CAD2jOMO3yMeQ7Ojk7TtNdMoshqQCQ87ke37aQWJDNeqZzvCAew%40mail.gmail.com/#msg35964428 @jvendetti Did you solve this problem?
1.0
Biorefinery & Transmat failed to parse - Error from parsing log file (Biorefinery): Illegal rdf:nodeID value '_:genid259' there is an equivalent error for Transmat. This error seems to have been identified in the NCBO BioPortal: see - [https://github.com/ncbo/bioportal-project/issues/32#event-1226205997](https://github.com/ncbo/bioportal-project/issues/32#event-1226205997) - [https://github.com/ncbo/bioportal-project/issues/9](https://github.com/ncbo/bioportal-project/issues/9) - https://sourceforge.net/p/owlapi/mailman/owlapi-developer/thread/CAD2jOMO3yMeQ7Ojk7TtNdMoshqQCQ87ke37aQWJDNeqZzvCAew%40mail.gmail.com/#msg35964428 @jvendetti Did you solve this problem?
process
biorefinery transmat failed to parse error from parsing log file biorefinery illegal rdf nodeid value there is an equivalent error for transmat this error seems to have been identified in the ncbo bioportal see jvendetti did you solve this problem
1
5,575
8,410,084,216
IssuesEvent
2018-10-12 09:28:16
threefoldtech/jumpscale_lib
https://api.github.com/repos/threefoldtech/jumpscale_lib
closed
Use the base token value instead of hastings for the txfee in the j.clients.rivine.wallet creation
process_wontfix
And rename it to transactionfee. might be that the docs are wrong, they seem to contradict: ```python from JumpscaleLib.clients.blockchain.rivine.RivineWallet import RivineWallet wallet = RivineWallet(seed=seed, bc_network='https://explorer.testnet.threefoldtoken.com/', bc_network_password='test123', nr_keys_per_seed=5, minerfee=100000000) # where seed is the seed you have or generated # bc_network: is the url to the blockchain network explorer node # bc_network_password: is the password to use while communicating with the chain explorer node # nr_keys_per_seed: is how many keys to generate per seed # minerfee: How many hastings should be added as minerfee ``` Or alternatively you can configure the wallet instance using the following code: ```python client_data = {'bc_address': 'https://explorer.testnet.threefoldtoken.com/', 'password_': 'test123', 'minerfee': 10, 'nr_keys_per_seed': 5, 'seed_': seed} rivine_client = j.clients.rivine.get('mytestwallet', data=client_data) rivine_client.config.save() wallet = rivine_client.wallet ``` while it should be 0.1
1.0
Use the base token value instead of hastings for the txfee in the j.clients.rivine.wallet creation - And rename it to transactionfee. might be that the docs are wrong, they seem to contradict: ```python from JumpscaleLib.clients.blockchain.rivine.RivineWallet import RivineWallet wallet = RivineWallet(seed=seed, bc_network='https://explorer.testnet.threefoldtoken.com/', bc_network_password='test123', nr_keys_per_seed=5, minerfee=100000000) # where seed is the seed you have or generated # bc_network: is the url to the blockchain network explorer node # bc_network_password: is the password to use while communicating with the chain explorer node # nr_keys_per_seed: is how many keys to generate per seed # minerfee: How many hastings should be added as minerfee ``` Or alternatively you can configure the wallet instance using the following code: ```python client_data = {'bc_address': 'https://explorer.testnet.threefoldtoken.com/', 'password_': 'test123', 'minerfee': 10, 'nr_keys_per_seed': 5, 'seed_': seed} rivine_client = j.clients.rivine.get('mytestwallet', data=client_data) rivine_client.config.save() wallet = rivine_client.wallet ``` while it should be 0.1
process
use the base token value instead of hastings for the txfee in the j clients rivine wallet creation and rename it to transactionfee might be that the docs are wrong they seem to contradict python from jumpscalelib clients blockchain rivine rivinewallet import rivinewallet wallet rivinewallet seed seed bc network bc network password nr keys per seed minerfee where seed is the seed you have or generated bc network is the url to the blockchain network explorer node bc network password is the password to use while communicating with the chain explorer node nr keys per seed is how many keys to generate per seed minerfee how many hastings should be added as minerfee or alternatively you can configure the wallet instance using the following code python client data bc address password minerfee nr keys per seed seed seed rivine client j clients rivine get mytestwallet data client data rivine client config save wallet rivine client wallet while it should be
1
6,164
9,049,614,521
IssuesEvent
2019-02-12 05:34:13
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
NullPointerException when Publishing a Learning Group Map
bug needs reproduction preprocess/chunking priority/high
I'm using DITA OT 2.4.6 and when I attempt to publish a Learning Group Map to PDF or HTML5, I'm getting this error below. I tested with DITA OT 2.0 and it is working fine. [LearningMap.zip](https://github.com/dita-ot/dita-ot/files/1019705/LearningMap.zip) chunk: [chunk] Processing file:A$L000022.ditamap BUILD FAILED The following error occurred while executing this line: lib\dita-ot\plugins\org.dita.base\build_preprocess.xml:308: java.lang.NullPointerException at org.dita.dost.writer.AbstractChunkTopicParser.generateOutputFile(AbstractChunkTopicParser.java:340) at org.dita.dost.writer.ChunkTopicParser.processChunk(ChunkTopicParser.java:149) at org.dita.dost.writer.ChunkTopicParser.write(ChunkTopicParser.java:60) at org.dita.dost.reader.ChunkMapReader.processCombineChunk(ChunkMapReader.java:536) at org.dita.dost.reader.ChunkMapReader.processTopicref(ChunkMapReader.java:341) at org.dita.dost.reader.ChunkMapReader.processChildTopicref(ChunkMapReader.java:513) at org.dita.dost.reader.ChunkMapReader.processTopicref(ChunkMapReader.java:366) at org.dita.dost.reader.ChunkMapReader.process(ChunkMapReader.java:142) at org.dita.dost.writer.AbstractDomFilter.read(AbstractDomFilter.java:55) at org.dita.dost.reader.ChunkMapReader.read(ChunkMapReader.java:118) at org.dita.dost.module.ChunkModule.execute(ChunkModule.java:80) at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:70) at org.dita.dost.invoker.ExtensibleAntInvoker.execute(ExtensibleAntInvoker.java:222) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:293) at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) at org.apache.tools.ant.Task.perform(Task.java:348) at org.apache.tools.ant.Target.execute(Target.java:435) at org.apache.tools.ant.Target.performTasks(Target.java:456) at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1405) at org.apache.tools.ant.helper.SingleCheckExecutor.executeTargets(SingleCheckExecutor.java:38) at org.apache.tools.ant.Project.executeTargets(Project.java:1260) at org.apache.tools.ant.taskdefs.Ant.execute(Ant.java:441) at org.apache.tools.ant.taskdefs.CallTarget.execute(CallTarget.java:105) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:293) at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) at org.apache.tools.ant.Task.perform(Task.java:348) at org.apache.tools.ant.Target.execute(Target.java:435) at org.apache.tools.ant.Target.performTasks(Target.java:456) at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1405) at org.apache.tools.ant.Project.executeTarget(Project.java:1376) at org.apache.tools.ant.helper.DefaultExecutor.executeTargets(DefaultExecutor.java:41) at org.apache.tools.ant.Project.executeTargets(Project.java:1260) at org.apache.tools.ant.Main.runBuild(Main.java:854) at org.apache.tools.ant.Main.startAnt(Main.java:236) at org.apache.tools.ant.launch.Launcher.run(Launcher.java:285) at org.apache.tools.ant.launch.Launcher.main(Launcher.java:112)
1.0
NullPointerException when Publishing a Learning Group Map - I'm using DITA OT 2.4.6 and when I attempt to publish a Learning Group Map to PDF or HTML5, I'm getting this error below. I tested with DITA OT 2.0 and it is working fine. [LearningMap.zip](https://github.com/dita-ot/dita-ot/files/1019705/LearningMap.zip) chunk: [chunk] Processing file:A$L000022.ditamap BUILD FAILED The following error occurred while executing this line: lib\dita-ot\plugins\org.dita.base\build_preprocess.xml:308: java.lang.NullPointerException at org.dita.dost.writer.AbstractChunkTopicParser.generateOutputFile(AbstractChunkTopicParser.java:340) at org.dita.dost.writer.ChunkTopicParser.processChunk(ChunkTopicParser.java:149) at org.dita.dost.writer.ChunkTopicParser.write(ChunkTopicParser.java:60) at org.dita.dost.reader.ChunkMapReader.processCombineChunk(ChunkMapReader.java:536) at org.dita.dost.reader.ChunkMapReader.processTopicref(ChunkMapReader.java:341) at org.dita.dost.reader.ChunkMapReader.processChildTopicref(ChunkMapReader.java:513) at org.dita.dost.reader.ChunkMapReader.processTopicref(ChunkMapReader.java:366) at org.dita.dost.reader.ChunkMapReader.process(ChunkMapReader.java:142) at org.dita.dost.writer.AbstractDomFilter.read(AbstractDomFilter.java:55) at org.dita.dost.reader.ChunkMapReader.read(ChunkMapReader.java:118) at org.dita.dost.module.ChunkModule.execute(ChunkModule.java:80) at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:70) at org.dita.dost.invoker.ExtensibleAntInvoker.execute(ExtensibleAntInvoker.java:222) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:293) at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) at org.apache.tools.ant.Task.perform(Task.java:348) at org.apache.tools.ant.Target.execute(Target.java:435) at org.apache.tools.ant.Target.performTasks(Target.java:456) at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1405) at org.apache.tools.ant.helper.SingleCheckExecutor.executeTargets(SingleCheckExecutor.java:38) at org.apache.tools.ant.Project.executeTargets(Project.java:1260) at org.apache.tools.ant.taskdefs.Ant.execute(Ant.java:441) at org.apache.tools.ant.taskdefs.CallTarget.execute(CallTarget.java:105) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:293) at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) at org.apache.tools.ant.Task.perform(Task.java:348) at org.apache.tools.ant.Target.execute(Target.java:435) at org.apache.tools.ant.Target.performTasks(Target.java:456) at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1405) at org.apache.tools.ant.Project.executeTarget(Project.java:1376) at org.apache.tools.ant.helper.DefaultExecutor.executeTargets(DefaultExecutor.java:41) at org.apache.tools.ant.Project.executeTargets(Project.java:1260) at org.apache.tools.ant.Main.runBuild(Main.java:854) at org.apache.tools.ant.Main.startAnt(Main.java:236) at org.apache.tools.ant.launch.Launcher.run(Launcher.java:285) at org.apache.tools.ant.launch.Launcher.main(Launcher.java:112)
process
nullpointerexception when publishing a learning group map i m using dita ot and when i attempt to publish a learning group map to pdf or i m getting this error below i tested with dita ot and it is working fine chunk processing file a ditamap build failed the following error occurred while executing this line lib dita ot plugins org dita base build preprocess xml java lang nullpointerexception at org dita dost writer abstractchunktopicparser generateoutputfile abstractchunktopicparser java at org dita dost writer chunktopicparser processchunk chunktopicparser java at org dita dost writer chunktopicparser write chunktopicparser java at org dita dost reader chunkmapreader processcombinechunk chunkmapreader java at org dita dost reader chunkmapreader processtopicref chunkmapreader java at org dita dost reader chunkmapreader processchildtopicref chunkmapreader java at org dita dost reader chunkmapreader processtopicref chunkmapreader java at org dita dost reader chunkmapreader process chunkmapreader java at org dita dost writer abstractdomfilter read abstractdomfilter java at org dita dost reader chunkmapreader read chunkmapreader java at org dita dost module chunkmodule execute chunkmodule java at org dita dost pipeline pipelinefacade execute pipelinefacade java at org dita dost invoker extensibleantinvoker execute extensibleantinvoker java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache tools ant dispatch dispatchutils execute dispatchutils java at org apache tools ant task perform task java at org apache tools ant target execute target java at org apache tools ant target performtasks target java at org apache tools ant project executesortedtargets project java at org apache tools ant helper singlecheckexecutor executetargets singlecheckexecutor java at org apache tools ant project executetargets project java at org apache tools ant taskdefs ant execute ant java at org apache tools ant taskdefs calltarget execute calltarget java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache tools ant dispatch dispatchutils execute dispatchutils java at org apache tools ant task perform task java at org apache tools ant target execute target java at org apache tools ant target performtasks target java at org apache tools ant project executesortedtargets project java at org apache tools ant project executetarget project java at org apache tools ant helper defaultexecutor executetargets defaultexecutor java at org apache tools ant project executetargets project java at org apache tools ant main runbuild main java at org apache tools ant main startant main java at org apache tools ant launch launcher run launcher java at org apache tools ant launch launcher main launcher java
1
1,951
4,773,979,274
IssuesEvent
2016-10-27 03:49:37
nodejs/node
https://api.github.com/repos/nodejs/node
opened
doc: need better link to signals list in process documentation
doc good first contribution process
* **Version**: all * **Platform**: n/a * **Subsystem**: doc In the 'Signal Events' section of the `process` documentation, there is a link to a man page that supposedly lists the names of the various signals that can be listened for. However, the page that is currently linked ([sigaction(2)](http://man7.org/linux/man-pages/man2/sigaction.2.html)) doesn't present all of the valid signals and the ones that it does show are not formatted very well. I think a better link would be to [signal(7)](http://man7.org/linux/man-pages/man7/signal.7.html), which *does* list all of the signal names and in an easy to spot/read format.
1.0
doc: need better link to signals list in process documentation - * **Version**: all * **Platform**: n/a * **Subsystem**: doc In the 'Signal Events' section of the `process` documentation, there is a link to a man page that supposedly lists the names of the various signals that can be listened for. However, the page that is currently linked ([sigaction(2)](http://man7.org/linux/man-pages/man2/sigaction.2.html)) doesn't present all of the valid signals and the ones that it does show are not formatted very well. I think a better link would be to [signal(7)](http://man7.org/linux/man-pages/man7/signal.7.html), which *does* list all of the signal names and in an easy to spot/read format.
process
doc need better link to signals list in process documentation version all platform n a subsystem doc in the signal events section of the process documentation there is a link to a man page that supposedly lists the names of the various signals that can be listened for however the page that is currently linked doesn t present all of the valid signals and the ones that it does show are not formatted very well i think a better link would be to which does list all of the signal names and in an easy to spot read format
1
16,455
21,334,029,560
IssuesEvent
2022-04-18 12:26:06
km4ack/patmenu2
https://api.github.com/repos/km4ack/patmenu2
closed
Backup/Save current config file
enhancement in process
There should be a way to save/backup the currently loaded config file.
1.0
Backup/Save current config file - There should be a way to save/backup the currently loaded config file.
process
backup save current config file there should be a way to save backup the currently loaded config file
1
18,886
3,091,421,457
IssuesEvent
2015-08-26 13:07:35
kattunga/uib
https://api.github.com/repos/kattunga/uib
closed
Exceptions text better readability
auto-migrated Priority-Medium Type-Defect
``` UibLib.pas Line 1528 was: Exception.Message := Exception.Message + 'Error Code: ' + IntToStr(Exception.FErrorCode); better: Exception.Message := Exception.Message + '. Error Code: ' + IntToStr(Exception.FErrorCode); Error Code is to be separated from previous GDS Code. ``` Original issue reported on code.google.com by `Arioch...@gmail.com` on 31 Oct 2011 at 7:56
1.0
Exceptions text better readability - ``` UibLib.pas Line 1528 was: Exception.Message := Exception.Message + 'Error Code: ' + IntToStr(Exception.FErrorCode); better: Exception.Message := Exception.Message + '. Error Code: ' + IntToStr(Exception.FErrorCode); Error Code is to be separated from previous GDS Code. ``` Original issue reported on code.google.com by `Arioch...@gmail.com` on 31 Oct 2011 at 7:56
non_process
exceptions text better readability uiblib pas line was exception message exception message error code inttostr exception ferrorcode better exception message exception message error code inttostr exception ferrorcode error code is to be separated from previous gds code original issue reported on code google com by arioch gmail com on oct at
0
16,353
9,784,102,245
IssuesEvent
2019-06-08 16:15:50
alexcorvi/apexo
https://api.github.com/repos/alexcorvi/apexo
closed
CVE-2019-6283 (Medium) detected in node-sass-v4.12.0
security vulnerability
## CVE-2019-6283 - Medium Severity Vulnerability <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.12.0</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> </p> </details> </p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Library Source Files (125)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /apexo/node_modules/node-sass/src/libsass/src/expand.hpp - /apexo/node_modules/node-sass/src/libsass/src/color_maps.cpp - /apexo/node_modules/node-sass/src/libsass/src/sass_util.hpp - /apexo/node_modules/node-sass/src/libsass/src/utf8/unchecked.h - /apexo/node_modules/node-sass/src/libsass/src/output.hpp - /apexo/node_modules/node-sass/src/libsass/src/sass_values.hpp - /apexo/node_modules/node-sass/src/libsass/src/util.hpp - /apexo/node_modules/node-sass/src/libsass/src/emitter.hpp - /apexo/node_modules/node-sass/src/libsass/src/lexer.cpp - /apexo/node_modules/node-sass/src/libsass/test/test_node.cpp - /apexo/node_modules/node-sass/src/libsass/src/plugins.cpp - /apexo/node_modules/node-sass/src/libsass/include/sass/base.h - /apexo/node_modules/node-sass/src/libsass/src/position.hpp - /apexo/node_modules/node-sass/src/libsass/src/subset_map.hpp - /apexo/node_modules/node-sass/src/libsass/src/operation.hpp - /apexo/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp - /apexo/node_modules/node-sass/src/libsass/src/error_handling.hpp - /apexo/node_modules/node-sass/src/custom_importer_bridge.cpp - /apexo/node_modules/node-sass/src/libsass/contrib/plugin.cpp - /apexo/node_modules/node-sass/src/libsass/src/functions.hpp - /apexo/node_modules/node-sass/src/libsass/test/test_superselector.cpp - /apexo/node_modules/node-sass/src/libsass/src/eval.hpp - /apexo/node_modules/node-sass/src/libsass/src/utf8_string.hpp - /apexo/node_modules/node-sass/src/sass_context_wrapper.h - /apexo/node_modules/node-sass/src/libsass/src/error_handling.cpp - /apexo/node_modules/node-sass/src/libsass/src/node.cpp - /apexo/node_modules/node-sass/src/libsass/src/parser.cpp - /apexo/node_modules/node-sass/src/libsass/src/subset_map.cpp - /apexo/node_modules/node-sass/src/libsass/src/emitter.cpp - /apexo/node_modules/node-sass/src/libsass/src/listize.cpp - /apexo/node_modules/node-sass/src/libsass/src/ast.hpp - /apexo/node_modules/node-sass/src/libsass/src/sass_functions.hpp - /apexo/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp - /apexo/node_modules/node-sass/src/libsass/src/output.cpp - /apexo/node_modules/node-sass/src/libsass/src/check_nesting.cpp - /apexo/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp - /apexo/node_modules/node-sass/src/libsass/src/functions.cpp - /apexo/node_modules/node-sass/src/libsass/src/cssize.hpp - /apexo/node_modules/node-sass/src/libsass/src/prelexer.cpp - /apexo/node_modules/node-sass/src/libsass/src/paths.hpp - /apexo/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp - /apexo/node_modules/node-sass/src/libsass/src/inspect.hpp - /apexo/node_modules/node-sass/src/sass_types/color.cpp - /apexo/node_modules/node-sass/src/libsass/test/test_unification.cpp - /apexo/node_modules/node-sass/src/libsass/src/values.cpp - /apexo/node_modules/node-sass/src/libsass/src/sass_util.cpp - /apexo/node_modules/node-sass/src/libsass/src/source_map.hpp - /apexo/node_modules/node-sass/src/sass_types/list.h - /apexo/node_modules/node-sass/src/libsass/src/check_nesting.hpp - /apexo/node_modules/node-sass/src/libsass/src/json.cpp - /apexo/node_modules/node-sass/src/libsass/src/units.cpp - /apexo/node_modules/node-sass/src/libsass/src/units.hpp - /apexo/node_modules/node-sass/src/libsass/src/context.cpp - /apexo/node_modules/node-sass/src/libsass/src/utf8/checked.h - /apexo/node_modules/node-sass/src/libsass/src/listize.hpp - /apexo/node_modules/node-sass/src/sass_types/string.cpp - /apexo/node_modules/node-sass/src/libsass/src/prelexer.hpp - /apexo/node_modules/node-sass/src/libsass/src/context.hpp - /apexo/node_modules/node-sass/src/sass_types/boolean.h - /apexo/node_modules/node-sass/src/libsass/include/sass2scss.h - /apexo/node_modules/node-sass/src/libsass/src/eval.cpp - /apexo/node_modules/node-sass/src/libsass/src/expand.cpp - /apexo/node_modules/node-sass/src/sass_types/factory.cpp - /apexo/node_modules/node-sass/src/libsass/src/operators.cpp - /apexo/node_modules/node-sass/src/sass_types/boolean.cpp - /apexo/node_modules/node-sass/src/libsass/src/source_map.cpp - /apexo/node_modules/node-sass/src/sass_types/value.h - /apexo/node_modules/node-sass/src/libsass/src/utf8_string.cpp - /apexo/node_modules/node-sass/src/callback_bridge.h - /apexo/node_modules/node-sass/src/libsass/src/file.cpp - /apexo/node_modules/node-sass/src/libsass/src/sass.cpp - /apexo/node_modules/node-sass/src/libsass/src/node.hpp - /apexo/node_modules/node-sass/src/libsass/src/environment.cpp - /apexo/node_modules/node-sass/src/libsass/src/extend.hpp - /apexo/node_modules/node-sass/src/libsass/src/sass_context.hpp - /apexo/node_modules/node-sass/src/libsass/src/operators.hpp - /apexo/node_modules/node-sass/src/libsass/src/constants.hpp - /apexo/node_modules/node-sass/src/libsass/src/sass.hpp - /apexo/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp - /apexo/node_modules/node-sass/src/libsass/src/parser.hpp - /apexo/node_modules/node-sass/src/libsass/src/constants.cpp - /apexo/node_modules/node-sass/src/sass_types/list.cpp - /apexo/node_modules/node-sass/src/libsass/src/cssize.cpp - /apexo/node_modules/node-sass/src/libsass/include/sass/functions.h - /apexo/node_modules/node-sass/src/libsass/src/util.cpp - /apexo/node_modules/node-sass/src/custom_function_bridge.cpp - /apexo/node_modules/node-sass/src/custom_importer_bridge.h - /apexo/node_modules/node-sass/src/libsass/src/bind.cpp - /apexo/node_modules/node-sass/src/libsass/src/inspect.cpp - /apexo/node_modules/node-sass/src/libsass/src/sass_functions.cpp - /apexo/node_modules/node-sass/src/libsass/src/backtrace.cpp - /apexo/node_modules/node-sass/src/libsass/src/extend.cpp - /apexo/node_modules/node-sass/src/sass_types/sass_value_wrapper.h - /apexo/node_modules/node-sass/src/libsass/src/debugger.hpp - /apexo/node_modules/node-sass/src/libsass/src/cencode.c - /apexo/node_modules/node-sass/src/libsass/src/base64vlq.cpp - /apexo/node_modules/node-sass/src/sass_types/number.cpp - /apexo/node_modules/node-sass/src/sass_types/color.h - /apexo/node_modules/node-sass/src/libsass/src/c99func.c - /apexo/node_modules/node-sass/src/libsass/src/position.cpp - /apexo/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp - /apexo/node_modules/node-sass/src/libsass/src/sass_values.cpp - /apexo/node_modules/node-sass/src/libsass/include/sass/values.h - /apexo/node_modules/node-sass/src/libsass/test/test_subset_map.cpp - /apexo/node_modules/node-sass/src/libsass/src/sass2scss.cpp - /apexo/node_modules/node-sass/src/sass_types/null.cpp - /apexo/node_modules/node-sass/src/libsass/src/ast.cpp - /apexo/node_modules/node-sass/src/libsass/include/sass/context.h - /apexo/node_modules/node-sass/src/libsass/src/to_c.cpp - /apexo/node_modules/node-sass/src/libsass/src/to_value.hpp - /apexo/node_modules/node-sass/src/libsass/src/color_maps.hpp - /apexo/node_modules/node-sass/src/sass_context_wrapper.cpp - /apexo/node_modules/node-sass/src/libsass/script/test-leaks.pl - /apexo/node_modules/node-sass/src/libsass/src/lexer.hpp - /apexo/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp - /apexo/node_modules/node-sass/src/libsass/src/to_c.hpp - /apexo/node_modules/node-sass/src/sass_types/map.cpp - /apexo/node_modules/node-sass/src/libsass/src/to_value.cpp - /apexo/node_modules/node-sass/src/libsass/src/b64/encode.h - /apexo/node_modules/node-sass/src/libsass/src/file.hpp - /apexo/node_modules/node-sass/src/libsass/src/environment.hpp - /apexo/node_modules/node-sass/src/libsass/src/plugins.hpp - /apexo/node_modules/node-sass/src/binding.cpp - /apexo/node_modules/node-sass/src/libsass/src/sass_context.cpp - /apexo/node_modules/node-sass/src/libsass/src/debug.hpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::parenthese_scope in prelexer.hpp. <p>Publish Date: 2019-01-14 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6283>CVE-2019-6283</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-6283 (Medium) detected in node-sass-v4.12.0 - ## CVE-2019-6283 - Medium Severity Vulnerability <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.12.0</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> </p> </details> </p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Library Source Files (125)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /apexo/node_modules/node-sass/src/libsass/src/expand.hpp - /apexo/node_modules/node-sass/src/libsass/src/color_maps.cpp - /apexo/node_modules/node-sass/src/libsass/src/sass_util.hpp - /apexo/node_modules/node-sass/src/libsass/src/utf8/unchecked.h - /apexo/node_modules/node-sass/src/libsass/src/output.hpp - /apexo/node_modules/node-sass/src/libsass/src/sass_values.hpp - /apexo/node_modules/node-sass/src/libsass/src/util.hpp - /apexo/node_modules/node-sass/src/libsass/src/emitter.hpp - /apexo/node_modules/node-sass/src/libsass/src/lexer.cpp - /apexo/node_modules/node-sass/src/libsass/test/test_node.cpp - /apexo/node_modules/node-sass/src/libsass/src/plugins.cpp - /apexo/node_modules/node-sass/src/libsass/include/sass/base.h - /apexo/node_modules/node-sass/src/libsass/src/position.hpp - /apexo/node_modules/node-sass/src/libsass/src/subset_map.hpp - /apexo/node_modules/node-sass/src/libsass/src/operation.hpp - /apexo/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp - /apexo/node_modules/node-sass/src/libsass/src/error_handling.hpp - /apexo/node_modules/node-sass/src/custom_importer_bridge.cpp - /apexo/node_modules/node-sass/src/libsass/contrib/plugin.cpp - /apexo/node_modules/node-sass/src/libsass/src/functions.hpp - /apexo/node_modules/node-sass/src/libsass/test/test_superselector.cpp - /apexo/node_modules/node-sass/src/libsass/src/eval.hpp - /apexo/node_modules/node-sass/src/libsass/src/utf8_string.hpp - /apexo/node_modules/node-sass/src/sass_context_wrapper.h - /apexo/node_modules/node-sass/src/libsass/src/error_handling.cpp - /apexo/node_modules/node-sass/src/libsass/src/node.cpp - /apexo/node_modules/node-sass/src/libsass/src/parser.cpp - /apexo/node_modules/node-sass/src/libsass/src/subset_map.cpp - /apexo/node_modules/node-sass/src/libsass/src/emitter.cpp - /apexo/node_modules/node-sass/src/libsass/src/listize.cpp - /apexo/node_modules/node-sass/src/libsass/src/ast.hpp - /apexo/node_modules/node-sass/src/libsass/src/sass_functions.hpp - /apexo/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp - /apexo/node_modules/node-sass/src/libsass/src/output.cpp - /apexo/node_modules/node-sass/src/libsass/src/check_nesting.cpp - /apexo/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp - /apexo/node_modules/node-sass/src/libsass/src/functions.cpp - /apexo/node_modules/node-sass/src/libsass/src/cssize.hpp - /apexo/node_modules/node-sass/src/libsass/src/prelexer.cpp - /apexo/node_modules/node-sass/src/libsass/src/paths.hpp - /apexo/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp - /apexo/node_modules/node-sass/src/libsass/src/inspect.hpp - /apexo/node_modules/node-sass/src/sass_types/color.cpp - /apexo/node_modules/node-sass/src/libsass/test/test_unification.cpp - /apexo/node_modules/node-sass/src/libsass/src/values.cpp - /apexo/node_modules/node-sass/src/libsass/src/sass_util.cpp - /apexo/node_modules/node-sass/src/libsass/src/source_map.hpp - /apexo/node_modules/node-sass/src/sass_types/list.h - /apexo/node_modules/node-sass/src/libsass/src/check_nesting.hpp - /apexo/node_modules/node-sass/src/libsass/src/json.cpp - /apexo/node_modules/node-sass/src/libsass/src/units.cpp - /apexo/node_modules/node-sass/src/libsass/src/units.hpp - /apexo/node_modules/node-sass/src/libsass/src/context.cpp - /apexo/node_modules/node-sass/src/libsass/src/utf8/checked.h - /apexo/node_modules/node-sass/src/libsass/src/listize.hpp - /apexo/node_modules/node-sass/src/sass_types/string.cpp - /apexo/node_modules/node-sass/src/libsass/src/prelexer.hpp - /apexo/node_modules/node-sass/src/libsass/src/context.hpp - /apexo/node_modules/node-sass/src/sass_types/boolean.h - /apexo/node_modules/node-sass/src/libsass/include/sass2scss.h - /apexo/node_modules/node-sass/src/libsass/src/eval.cpp - /apexo/node_modules/node-sass/src/libsass/src/expand.cpp - /apexo/node_modules/node-sass/src/sass_types/factory.cpp - /apexo/node_modules/node-sass/src/libsass/src/operators.cpp - /apexo/node_modules/node-sass/src/sass_types/boolean.cpp - /apexo/node_modules/node-sass/src/libsass/src/source_map.cpp - /apexo/node_modules/node-sass/src/sass_types/value.h - /apexo/node_modules/node-sass/src/libsass/src/utf8_string.cpp - /apexo/node_modules/node-sass/src/callback_bridge.h - /apexo/node_modules/node-sass/src/libsass/src/file.cpp - /apexo/node_modules/node-sass/src/libsass/src/sass.cpp - /apexo/node_modules/node-sass/src/libsass/src/node.hpp - /apexo/node_modules/node-sass/src/libsass/src/environment.cpp - /apexo/node_modules/node-sass/src/libsass/src/extend.hpp - /apexo/node_modules/node-sass/src/libsass/src/sass_context.hpp - /apexo/node_modules/node-sass/src/libsass/src/operators.hpp - /apexo/node_modules/node-sass/src/libsass/src/constants.hpp - /apexo/node_modules/node-sass/src/libsass/src/sass.hpp - /apexo/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp - /apexo/node_modules/node-sass/src/libsass/src/parser.hpp - /apexo/node_modules/node-sass/src/libsass/src/constants.cpp - /apexo/node_modules/node-sass/src/sass_types/list.cpp - /apexo/node_modules/node-sass/src/libsass/src/cssize.cpp - /apexo/node_modules/node-sass/src/libsass/include/sass/functions.h - /apexo/node_modules/node-sass/src/libsass/src/util.cpp - /apexo/node_modules/node-sass/src/custom_function_bridge.cpp - /apexo/node_modules/node-sass/src/custom_importer_bridge.h - /apexo/node_modules/node-sass/src/libsass/src/bind.cpp - /apexo/node_modules/node-sass/src/libsass/src/inspect.cpp - /apexo/node_modules/node-sass/src/libsass/src/sass_functions.cpp - /apexo/node_modules/node-sass/src/libsass/src/backtrace.cpp - /apexo/node_modules/node-sass/src/libsass/src/extend.cpp - /apexo/node_modules/node-sass/src/sass_types/sass_value_wrapper.h - /apexo/node_modules/node-sass/src/libsass/src/debugger.hpp - /apexo/node_modules/node-sass/src/libsass/src/cencode.c - /apexo/node_modules/node-sass/src/libsass/src/base64vlq.cpp - /apexo/node_modules/node-sass/src/sass_types/number.cpp - /apexo/node_modules/node-sass/src/sass_types/color.h - /apexo/node_modules/node-sass/src/libsass/src/c99func.c - /apexo/node_modules/node-sass/src/libsass/src/position.cpp - /apexo/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp - /apexo/node_modules/node-sass/src/libsass/src/sass_values.cpp - /apexo/node_modules/node-sass/src/libsass/include/sass/values.h - /apexo/node_modules/node-sass/src/libsass/test/test_subset_map.cpp - /apexo/node_modules/node-sass/src/libsass/src/sass2scss.cpp - /apexo/node_modules/node-sass/src/sass_types/null.cpp - /apexo/node_modules/node-sass/src/libsass/src/ast.cpp - /apexo/node_modules/node-sass/src/libsass/include/sass/context.h - /apexo/node_modules/node-sass/src/libsass/src/to_c.cpp - /apexo/node_modules/node-sass/src/libsass/src/to_value.hpp - /apexo/node_modules/node-sass/src/libsass/src/color_maps.hpp - /apexo/node_modules/node-sass/src/sass_context_wrapper.cpp - /apexo/node_modules/node-sass/src/libsass/script/test-leaks.pl - /apexo/node_modules/node-sass/src/libsass/src/lexer.hpp - /apexo/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp - /apexo/node_modules/node-sass/src/libsass/src/to_c.hpp - /apexo/node_modules/node-sass/src/sass_types/map.cpp - /apexo/node_modules/node-sass/src/libsass/src/to_value.cpp - /apexo/node_modules/node-sass/src/libsass/src/b64/encode.h - /apexo/node_modules/node-sass/src/libsass/src/file.hpp - /apexo/node_modules/node-sass/src/libsass/src/environment.hpp - /apexo/node_modules/node-sass/src/libsass/src/plugins.hpp - /apexo/node_modules/node-sass/src/binding.cpp - /apexo/node_modules/node-sass/src/libsass/src/sass_context.cpp - /apexo/node_modules/node-sass/src/libsass/src/debug.hpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::parenthese_scope in prelexer.hpp. <p>Publish Date: 2019-01-14 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6283>CVE-2019-6283</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in node sass cve medium severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries apexo node modules node sass src libsass src expand hpp apexo node modules node sass src libsass src color maps cpp apexo node modules node sass src libsass src sass util hpp apexo node modules node sass src libsass src unchecked h apexo node modules node sass src libsass src output hpp apexo node modules node sass src libsass src sass values hpp apexo node modules node sass src libsass src util hpp apexo node modules node sass src libsass src emitter hpp apexo node modules node sass src libsass src lexer cpp apexo node modules node sass src libsass test test node cpp apexo node modules node sass src libsass src plugins cpp apexo node modules node sass src libsass include sass base h apexo node modules node sass src libsass src position hpp apexo node modules node sass src libsass src subset map hpp apexo node modules node sass src libsass src operation hpp apexo node modules node sass src libsass src remove placeholders cpp apexo node modules node sass src libsass src error handling hpp apexo node modules node sass src custom importer bridge cpp apexo node modules node sass src libsass contrib plugin cpp apexo node modules node sass src libsass src functions hpp apexo node modules node sass src libsass test test superselector cpp apexo node modules node sass src libsass src eval hpp apexo node modules node sass src libsass src string hpp apexo node modules node sass src sass context wrapper h apexo node modules node sass src libsass src error handling cpp apexo node modules node sass src libsass src node cpp apexo node modules node sass src libsass src parser cpp apexo node modules node sass src libsass src subset map cpp apexo node modules node sass src libsass src emitter cpp apexo node modules node sass src libsass src listize cpp apexo node modules node sass src libsass src ast hpp apexo node modules node sass src libsass src sass functions hpp apexo node modules node sass src libsass src memory sharedptr cpp apexo node modules node sass src libsass src output cpp apexo node modules node sass src libsass src check nesting cpp apexo node modules node sass src libsass src ast def macros hpp apexo node modules node sass src libsass src functions cpp apexo node modules node sass src libsass src cssize hpp apexo node modules node sass src libsass src prelexer cpp apexo node modules node sass src libsass src paths hpp apexo node modules node sass src libsass src ast fwd decl hpp apexo node modules node sass src libsass src inspect hpp apexo node modules node sass src sass types color cpp apexo node modules node sass src libsass test test unification cpp apexo node modules node sass src libsass src values cpp apexo node modules node sass src libsass src sass util cpp apexo node modules node sass src libsass src source map hpp apexo node modules node sass src sass types list h apexo node modules node sass src libsass src check nesting hpp apexo node modules node sass src libsass src json cpp apexo node modules node sass src libsass src units cpp apexo node modules node sass src libsass src units hpp apexo node modules node sass src libsass src context cpp apexo node modules node sass src libsass src checked h apexo node modules node sass src libsass src listize hpp apexo node modules node sass src sass types string cpp apexo node modules node sass src libsass src prelexer hpp apexo node modules node sass src libsass src context hpp apexo node modules node sass src sass types boolean h apexo node modules node sass src libsass include h apexo node modules node sass src libsass src eval cpp apexo node modules node sass src libsass src expand cpp apexo node modules node sass src sass types factory cpp apexo node modules node sass src libsass src operators cpp apexo node modules node sass src sass types boolean cpp apexo node modules node sass src libsass src source map cpp apexo node modules node sass src sass types value h apexo node modules node sass src libsass src string cpp apexo node modules node sass src callback bridge h apexo node modules node sass src libsass src file cpp apexo node modules node sass src libsass src sass cpp apexo node modules node sass src libsass src node hpp apexo node modules node sass src libsass src environment cpp apexo node modules node sass src libsass src extend hpp apexo node modules node sass src libsass src sass context hpp apexo node modules node sass src libsass src operators hpp apexo node modules node sass src libsass src constants hpp apexo node modules node sass src libsass src sass hpp apexo node modules node sass src libsass src ast fwd decl cpp apexo node modules node sass src libsass src parser hpp apexo node modules node sass src libsass src constants cpp apexo node modules node sass src sass types list cpp apexo node modules node sass src libsass src cssize cpp apexo node modules node sass src libsass include sass functions h apexo node modules node sass src libsass src util cpp apexo node modules node sass src custom function bridge cpp apexo node modules node sass src custom importer bridge h apexo node modules node sass src libsass src bind cpp apexo node modules node sass src libsass src inspect cpp apexo node modules node sass src libsass src sass functions cpp apexo node modules node sass src libsass src backtrace cpp apexo node modules node sass src libsass src extend cpp apexo node modules node sass src sass types sass value wrapper h apexo node modules node sass src libsass src debugger hpp apexo node modules node sass src libsass src cencode c apexo node modules node sass src libsass src cpp apexo node modules node sass src sass types number cpp apexo node modules node sass src sass types color h apexo node modules node sass src libsass src c apexo node modules node sass src libsass src position cpp apexo node modules node sass src libsass src remove placeholders hpp apexo node modules node sass src libsass src sass values cpp apexo node modules node sass src libsass include sass values h apexo node modules node sass src libsass test test subset map cpp apexo node modules node sass src libsass src cpp apexo node modules node sass src sass types null cpp apexo node modules node sass src libsass src ast cpp apexo node modules node sass src libsass include sass context h apexo node modules node sass src libsass src to c cpp apexo node modules node sass src libsass src to value hpp apexo node modules node sass src libsass src color maps hpp apexo node modules node sass src sass context wrapper cpp apexo node modules node sass src libsass script test leaks pl apexo node modules node sass src libsass src lexer hpp apexo node modules node sass src libsass src memory sharedptr hpp apexo node modules node sass src libsass src to c hpp apexo node modules node sass src sass types map cpp apexo node modules node sass src libsass src to value cpp apexo node modules node sass src libsass src encode h apexo node modules node sass src libsass src file hpp apexo node modules node sass src libsass src environment hpp apexo node modules node sass src libsass src plugins hpp apexo node modules node sass src binding cpp apexo node modules node sass src libsass src sass context cpp apexo node modules node sass src libsass src debug hpp vulnerability details in libsass a heap based buffer over read exists in sass prelexer parenthese scope in prelexer hpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource
0
252,087
21,556,148,023
IssuesEvent
2022-04-30 13:00:42
poikilos/EnlivenMinetest
https://api.github.com/repos/poikilos/EnlivenMinetest
closed
airitems (air tools) textures are blurry
fixed - resolved in next Final Minetest release patch available externally
They were scaled up using an edge-detecting resize method then scaled back down. airitems 16x16 originals are at https://git.minetest.org/zander999/air_items mods/coderbuild/coderblocks/textures/ - [x] air_pick.png - [x] air_pick_upgraded.png - [x] air_sword.png Patchable via (set BUCKET_GAME then): ``` cd /tmp git clone https://git.minetest.org/zander999/air_items.git /tmp/air_items cd /tmp/air_items cp textures/air_pick.png $BUCKET_GAME/mods/coderbuild/coderblocks/textures/ cp textures/air_pick_upgraded.png $BUCKET_GAME/mods/coderbuild/coderblocks/textures/ cp textures/air_sword.png $BUCKET_GAME/mods/coderbuild/coderblocks/textures/ ```
1.0
airitems (air tools) textures are blurry - They were scaled up using an edge-detecting resize method then scaled back down. airitems 16x16 originals are at https://git.minetest.org/zander999/air_items mods/coderbuild/coderblocks/textures/ - [x] air_pick.png - [x] air_pick_upgraded.png - [x] air_sword.png Patchable via (set BUCKET_GAME then): ``` cd /tmp git clone https://git.minetest.org/zander999/air_items.git /tmp/air_items cd /tmp/air_items cp textures/air_pick.png $BUCKET_GAME/mods/coderbuild/coderblocks/textures/ cp textures/air_pick_upgraded.png $BUCKET_GAME/mods/coderbuild/coderblocks/textures/ cp textures/air_sword.png $BUCKET_GAME/mods/coderbuild/coderblocks/textures/ ```
non_process
airitems air tools textures are blurry they were scaled up using an edge detecting resize method then scaled back down airitems originals are at mods coderbuild coderblocks textures air pick png air pick upgraded png air sword png patchable via set bucket game then cd tmp git clone tmp air items cd tmp air items cp textures air pick png bucket game mods coderbuild coderblocks textures cp textures air pick upgraded png bucket game mods coderbuild coderblocks textures cp textures air sword png bucket game mods coderbuild coderblocks textures
0
36,677
2,810,587,032
IssuesEvent
2015-05-17 01:11:24
yearography/web
https://api.github.com/repos/yearography/web
closed
Invalid character found in landing page
category: frontend priority: now type: bug
It would appear as though some jackass left an 'e' in the header of the landing page. Please fix asap.
1.0
Invalid character found in landing page - It would appear as though some jackass left an 'e' in the header of the landing page. Please fix asap.
non_process
invalid character found in landing page it would appear as though some jackass left an e in the header of the landing page please fix asap
0
197,638
15,687,108,527
IssuesEvent
2021-03-25 13:19:19
hdl/awesome
https://api.github.com/repos/hdl/awesome
opened
How to handle BibTeX references?
documentation enhancement help wanted theme
In the context of #184, @rodrigomelo9 asked https://github.com/verilog-to-routing/vtr-verilog-to-routing/issues/1691. There, several academic/bibliographic references were provided. We don't have a mechanism for showing that content in the site. However, since many of the open source EDA projects are related to academia and research publications, it would be very desirable to support BibTeX. There are some solutions: - https://github.com/wowchemy/hugo-academic-cli - https://gohugo.io/content-management/formats/#external-helper-asciidoctor Yet, either of them requires some effort. They don't seem to be built-in. Meanwhile, we might have a `*.bib` file and suggest users to browse it with [JabRef](https://www.jabref.org/), [Zotero](https://www.zotero.org/), etc.
1.0
How to handle BibTeX references? - In the context of #184, @rodrigomelo9 asked https://github.com/verilog-to-routing/vtr-verilog-to-routing/issues/1691. There, several academic/bibliographic references were provided. We don't have a mechanism for showing that content in the site. However, since many of the open source EDA projects are related to academia and research publications, it would be very desirable to support BibTeX. There are some solutions: - https://github.com/wowchemy/hugo-academic-cli - https://gohugo.io/content-management/formats/#external-helper-asciidoctor Yet, either of them requires some effort. They don't seem to be built-in. Meanwhile, we might have a `*.bib` file and suggest users to browse it with [JabRef](https://www.jabref.org/), [Zotero](https://www.zotero.org/), etc.
non_process
how to handle bibtex references in the context of asked there several academic bibliographic references were provided we don t have a mechanism for showing that content in the site however since many of the open source eda projects are related to academia and research publications it would be very desirable to support bibtex there are some solutions yet either of them requires some effort they don t seem to be built in meanwhile we might have a bib file and suggest users to browse it with etc
0
57,672
24,187,760,967
IssuesEvent
2022-09-23 14:40:39
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
closed
Add Arterial Management & Austin Police Department to the Coordination Partners list
Impact: 2-Major Service: Dev Need: 1-Must Have Workgroup: VZ Type: Enhancement
Staff would like to add Arterial Management and Austin Police Department as additional options for selections from the Coordination Partners drop down in the Recommendations section. Please add them to the `atd__coordination_partners_lkp` table. ![Screen Shot 2022-09-19 at 8.04.52 AM.png](https://images.zenhubusercontent.com/5d68810b05d7792664be403f/9dc11df8-eead-4227-a8f8-36a4502bbd1b)
1.0
Add Arterial Management & Austin Police Department to the Coordination Partners list - Staff would like to add Arterial Management and Austin Police Department as additional options for selections from the Coordination Partners drop down in the Recommendations section. Please add them to the `atd__coordination_partners_lkp` table. ![Screen Shot 2022-09-19 at 8.04.52 AM.png](https://images.zenhubusercontent.com/5d68810b05d7792664be403f/9dc11df8-eead-4227-a8f8-36a4502bbd1b)
non_process
add arterial management austin police department to the coordination partners list staff would like to add arterial management and austin police department as additional options for selections from the coordination partners drop down in the recommendations section please add them to the atd coordination partners lkp table
0
2,899
5,887,623,343
IssuesEvent
2017-05-17 08:01:29
g8os/core0
https://api.github.com/repos/g8os/core0
closed
Add a way to stream the output of a command
process_wontfix type_feature
It would be interesting to have a command which stream the output of a command run in a container. When starting a command, there is no way to know the status (stdout, stderr) of that given command right now, until it complete.
1.0
Add a way to stream the output of a command - It would be interesting to have a command which stream the output of a command run in a container. When starting a command, there is no way to know the status (stdout, stderr) of that given command right now, until it complete.
process
add a way to stream the output of a command it would be interesting to have a command which stream the output of a command run in a container when starting a command there is no way to know the status stdout stderr of that given command right now until it complete
1
10,530
13,309,453,683
IssuesEvent
2020-08-26 04:04:08
Zjmainstay/blog-comment
https://api.github.com/repos/Zjmainstay/blog-comment
opened
Shell+Crontab实现服务进程监控重启 - Zjmainstay学习笔记 | 是非明于学习,境界升于内省
Gitalk _shell-checkprocess
http://www.zjmainstay.cn/shell-checkprocess 本文提供一个shell脚本,结合crontab的使用可以实现对服务器服务进程的运行情况进行监控,及时重启服务器关键服务进程,避免关键服务进程长期终止情况发生。
1.0
Shell+Crontab实现服务进程监控重启 - Zjmainstay学习笔记 | 是非明于学习,境界升于内省 - http://www.zjmainstay.cn/shell-checkprocess 本文提供一个shell脚本,结合crontab的使用可以实现对服务器服务进程的运行情况进行监控,及时重启服务器关键服务进程,避免关键服务进程长期终止情况发生。
process
shell crontab实现服务进程监控重启 zjmainstay学习笔记 是非明于学习,境界升于内省 本文提供一个shell脚本,结合crontab的使用可以实现对服务器服务进程的运行情况进行监控,及时重启服务器关键服务进程,避免关键服务进程长期终止情况发生。
1
455
2,894,791,486
IssuesEvent
2015-06-16 03:13:01
hammerlab/pileup.js
https://api.github.com/repos/hammerlab/pileup.js
opened
Use async/await
process
Flow 0.12 added support for the async & await keywords. The Babel transpiler supports them. Using these features of ES6/7 would greatly simplify the async code in pileup.js. https://github.com/facebook/flow/releases/tag/v0.12.0
1.0
Use async/await - Flow 0.12 added support for the async & await keywords. The Babel transpiler supports them. Using these features of ES6/7 would greatly simplify the async code in pileup.js. https://github.com/facebook/flow/releases/tag/v0.12.0
process
use async await flow added support for the async await keywords the babel transpiler supports them using these features of would greatly simplify the async code in pileup js
1
10,371
6,689,312,033
IssuesEvent
2017-10-09 00:43:31
uqbar-project/wollok
https://api.github.com/repos/uqbar-project/wollok
closed
REPL continues to run after program has been modified
component: repl in progress usability
If you have a program running in the REPL and you modify it, the REPL continues running the old program. I would like to give a try to update the program in execution, maybe it is easier than it appears at first sight. But still, if that is not possible, we should take some action when the running program is modified. Otherwise, programmers tend to get confused because they can't see any behavior change after program modifications. Two simple actions, we should either a. Stop de REPL, or b. If we allow to continue running, at least add some kind of visual clue about running an obsolete program: change a color or icon, display a warning, etc.
True
REPL continues to run after program has been modified - If you have a program running in the REPL and you modify it, the REPL continues running the old program. I would like to give a try to update the program in execution, maybe it is easier than it appears at first sight. But still, if that is not possible, we should take some action when the running program is modified. Otherwise, programmers tend to get confused because they can't see any behavior change after program modifications. Two simple actions, we should either a. Stop de REPL, or b. If we allow to continue running, at least add some kind of visual clue about running an obsolete program: change a color or icon, display a warning, etc.
non_process
repl continues to run after program has been modified if you have a program running in the repl and you modify it the repl continues running the old program i would like to give a try to update the program in execution maybe it is easier than it appears at first sight but still if that is not possible we should take some action when the running program is modified otherwise programmers tend to get confused because they can t see any behavior change after program modifications two simple actions we should either a stop de repl or b if we allow to continue running at least add some kind of visual clue about running an obsolete program change a color or icon display a warning etc
0
14,628
17,767,680,194
IssuesEvent
2021-08-30 09:37:53
googleapis/nodejs-dns
https://api.github.com/repos/googleapis/nodejs-dns
closed
Dependency Dashboard
type: process api: dns
This issue contains a list of Renovate updates and their statuses. ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2 ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/460) (`mocha`, `@types/mocha`) - [ ] <!-- recreate-branch=renovate/arrify-3.x -->[fix(deps): update dependency arrify to v3](../pull/448) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2 ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/460) (`mocha`, `@types/mocha`) - [ ] <!-- recreate-branch=renovate/arrify-3.x -->[fix(deps): update dependency arrify to v3](../pull/448) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses awaiting schedule these updates are awaiting their schedule click on a checkbox to get an update now chore deps update actions setup node action to ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull mocha types mocha pull check this box to trigger a request for renovate to run again on this repository
1
380,132
11,254,160,680
IssuesEvent
2020-01-11 21:25:04
Kaktushose/levelbot2
https://api.github.com/repos/Kaktushose/levelbot2
opened
update Command
enhancement low priority member team
**Steht dein Feature-Wunsch im Zusammenhang mit einem Problem? Bitte beschreibe es.** Falls es passiert, dass der Bot einen User nicht erfasst, z.B. durch Ausfälle, fehlt dieser in der DB. **Beschreibe die Lösung, die du gerne hättest** Command `!update`, der die Datenbank auf Fehler prüft
1.0
update Command - **Steht dein Feature-Wunsch im Zusammenhang mit einem Problem? Bitte beschreibe es.** Falls es passiert, dass der Bot einen User nicht erfasst, z.B. durch Ausfälle, fehlt dieser in der DB. **Beschreibe die Lösung, die du gerne hättest** Command `!update`, der die Datenbank auf Fehler prüft
non_process
update command steht dein feature wunsch im zusammenhang mit einem problem bitte beschreibe es falls es passiert dass der bot einen user nicht erfasst z b durch ausfälle fehlt dieser in der db beschreibe die lösung die du gerne hättest command update der die datenbank auf fehler prüft
0
34,642
12,294,674,151
IssuesEvent
2020-05-11 01:05:09
finos/secref-data
https://api.github.com/repos/finos/secref-data
opened
WS-2020-0070 (High) detected in lodash-4.17.15.tgz
security vulnerability
## WS-2020-0070 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.15.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/secref-data/website/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/secref-data/website/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - docusaurus-1.14.4.tgz (Root Library) - :x: **lodash-4.17.15.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> a prototype pollution vulnerability in lodash. It allows an attacker to inject properties on Object.prototype <p>Publish Date: 2020-04-28 <p>URL: <a href=https://hackerone.com/reports/712065>WS-2020-0070</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.15","isTransitiveDependency":true,"dependencyTree":"docusaurus:1.14.4;lodash:4.17.15","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"WS-2020-0070","vulnerabilityDetails":"a prototype pollution vulnerability in lodash. It allows an attacker to inject properties on Object.prototype","vulnerabilityUrl":"https://hackerone.com/reports/712065","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
WS-2020-0070 (High) detected in lodash-4.17.15.tgz - ## WS-2020-0070 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.15.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/secref-data/website/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/secref-data/website/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - docusaurus-1.14.4.tgz (Root Library) - :x: **lodash-4.17.15.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> a prototype pollution vulnerability in lodash. It allows an attacker to inject properties on Object.prototype <p>Publish Date: 2020-04-28 <p>URL: <a href=https://hackerone.com/reports/712065>WS-2020-0070</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.15","isTransitiveDependency":true,"dependencyTree":"docusaurus:1.14.4;lodash:4.17.15","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"WS-2020-0070","vulnerabilityDetails":"a prototype pollution vulnerability in lodash. It allows an attacker to inject properties on Object.prototype","vulnerabilityUrl":"https://hackerone.com/reports/712065","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
ws high detected in lodash tgz ws high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file tmp ws scm secref data website package json path to vulnerable library tmp ws scm secref data website node modules lodash package json dependency hierarchy docusaurus tgz root library x lodash tgz vulnerable library vulnerability details a prototype pollution vulnerability in lodash it allows an attacker to inject properties on object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails a prototype pollution vulnerability in lodash it allows an attacker to inject properties on object prototype vulnerabilityurl
0
16,193
20,674,211,081
IssuesEvent
2022-03-10 07:28:10
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
CockroachDB: Change the integer native types to the idiomatic ones
process/candidate topic: schema topic: native database types topic: schema validation engines/data model parser team/migrations topic: cockroachdb team/psl-wg
CockroachDB has two sets of names for integer column types, as described [in their docs](https://www.cockroachlabs.com/docs/v21.2/int). We started by forking the postgresql connector to support crdb, so we currently work with the aliases. To recap this in a table: <table> <tr> <td>Integer size <td>Postgres <td>CockroachDB <tr> <td>16 bits <td>SMALLINT <td>SMALLINT <tr> <td>32 bits <td>INTEGER <td>? (depends on session/server settings, may or may not be available, is not expressible by default) <tr> <td>64 bits <td>BIGINT <td>BIGINT / depends (INTEGER could be 64 bits, depending on settings) </table> There is however a solution to avoid ambiguities around integers: use the cockroachdb integer types `INT2`, `INT4` and `INT8`. We already use `INT4` in the migration engine to avoid ambiguities. This issue is here to propose that we replace the ambiguous/complicated `@db.SmallInt`, `@db.Integer` and `@db.BigInt` native types with the unambiguous/clear `@db.Int2`, `@db.Int4` and `@db.Int8` within the cockroachdb provider.
1.0
CockroachDB: Change the integer native types to the idiomatic ones - CockroachDB has two sets of names for integer column types, as described [in their docs](https://www.cockroachlabs.com/docs/v21.2/int). We started by forking the postgresql connector to support crdb, so we currently work with the aliases. To recap this in a table: <table> <tr> <td>Integer size <td>Postgres <td>CockroachDB <tr> <td>16 bits <td>SMALLINT <td>SMALLINT <tr> <td>32 bits <td>INTEGER <td>? (depends on session/server settings, may or may not be available, is not expressible by default) <tr> <td>64 bits <td>BIGINT <td>BIGINT / depends (INTEGER could be 64 bits, depending on settings) </table> There is however a solution to avoid ambiguities around integers: use the cockroachdb integer types `INT2`, `INT4` and `INT8`. We already use `INT4` in the migration engine to avoid ambiguities. This issue is here to propose that we replace the ambiguous/complicated `@db.SmallInt`, `@db.Integer` and `@db.BigInt` native types with the unambiguous/clear `@db.Int2`, `@db.Int4` and `@db.Int8` within the cockroachdb provider.
process
cockroachdb change the integer native types to the idiomatic ones cockroachdb has two sets of names for integer column types as described we started by forking the postgresql connector to support crdb so we currently work with the aliases to recap this in a table integer size postgres cockroachdb bits smallint smallint bits integer depends on session server settings may or may not be available is not expressible by default bits bigint bigint depends integer could be bits depending on settings there is however a solution to avoid ambiguities around integers use the cockroachdb integer types and we already use in the migration engine to avoid ambiguities this issue is here to propose that we replace the ambiguous complicated db smallint db integer and db bigint native types with the unambiguous clear db db and db within the cockroachdb provider
1
11,763
14,594,656,277
IssuesEvent
2020-12-20 07:11:56
rdoddanavar/hpr-sim
https://api.github.com/repos/rdoddanavar/hpr-sim
closed
util_unit.py: Temperature conversion
pre-processing
Implement lambda functions in `config_unit.yml` w/ corresponding routine in `util_unit.py` to handle more complex unit conversions
1.0
util_unit.py: Temperature conversion - Implement lambda functions in `config_unit.yml` w/ corresponding routine in `util_unit.py` to handle more complex unit conversions
process
util unit py temperature conversion implement lambda functions in config unit yml w corresponding routine in util unit py to handle more complex unit conversions
1
20,112
26,652,476,344
IssuesEvent
2023-01-25 14:38:16
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
How to understand decision_wait of tail_sampling
bug question processor/tailsampling
### Component(s) processor/tailsampling ### What happened? ## Description the decision_wait describe by: ``` decision_wait (default = 30s): Wait time since the first span of a trace before making a sampling decision ```` When I make two policies to get long-request-time traces and the error traces like this: ``` tail_sampling: decision_wait: 60s num_traces: 5000 expected_new_traces_per_sec: 10 policies: [ { name: errors-policy, type: numeric_attribute, numeric_attribute: {key: http.status_code, min_value: 499, max_value: 600} }, { name: latency-5s, type: latency, latency: {threshold_ms: 5000} }, ] ``` So if the trace is error already, the processor will delivery the trace in 60s later? In the real world, Microservice cross call make more than 60s long-request. So If I want to catch the threshold_ms>60000, the decision_wait must more than 60s, better about decision_wait=80s. And all the trace will store to the Jaeger at least delay decision_wait=80s later? Correct me if I'm wrong, Thank you. ## Steps to Reproduce ## Expected Result ## Actual Result ### Collector version 0.68.0 ### Environment information ## Environment OS: (e.g., "Ubuntu 20.04") Compiler(if manually compiled): (e.g., "go 14.2") ### OpenTelemetry Collector configuration _No response_ ### Log output _No response_ ### Additional context _No response_
1.0
How to understand decision_wait of tail_sampling - ### Component(s) processor/tailsampling ### What happened? ## Description the decision_wait describe by: ``` decision_wait (default = 30s): Wait time since the first span of a trace before making a sampling decision ```` When I make two policies to get long-request-time traces and the error traces like this: ``` tail_sampling: decision_wait: 60s num_traces: 5000 expected_new_traces_per_sec: 10 policies: [ { name: errors-policy, type: numeric_attribute, numeric_attribute: {key: http.status_code, min_value: 499, max_value: 600} }, { name: latency-5s, type: latency, latency: {threshold_ms: 5000} }, ] ``` So if the trace is error already, the processor will delivery the trace in 60s later? In the real world, Microservice cross call make more than 60s long-request. So If I want to catch the threshold_ms>60000, the decision_wait must more than 60s, better about decision_wait=80s. And all the trace will store to the Jaeger at least delay decision_wait=80s later? Correct me if I'm wrong, Thank you. ## Steps to Reproduce ## Expected Result ## Actual Result ### Collector version 0.68.0 ### Environment information ## Environment OS: (e.g., "Ubuntu 20.04") Compiler(if manually compiled): (e.g., "go 14.2") ### OpenTelemetry Collector configuration _No response_ ### Log output _No response_ ### Additional context _No response_
process
how to understand decision wait of tail sampling component s processor tailsampling what happened description the decision wait describe by decision wait default wait time since the first span of a trace before making a sampling decision when i make two policies to get long request time traces and the error traces like this tail sampling decision wait num traces expected new traces per sec policies name errors policy type numeric attribute numeric attribute key http status code min value max value name latency type latency latency threshold ms so if the trace is error already the processor will delivery the trace in later in the real world microservice cross call make more than long request so if i want to catch the threshold ms the decision wait must more than better about decision wait and all the trace will store to the jaeger at least delay decision wait later correct me if i m wrong thank you steps to reproduce expected result actual result collector version environment information environment os e g ubuntu compiler if manually compiled e g go opentelemetry collector configuration no response log output no response additional context no response
1
14,153
17,064,861,025
IssuesEvent
2021-07-07 05:38:00
ECP-WarpX/WarpX
https://api.github.com/repos/ECP-WarpX/WarpX
closed
yt: nbody species appearing
component: post-processing component: tests
I've been having the issue when using a newer version of yt, that the extra species nbody appears when reading in plot files. For regular post processing, this is not a problem. However, when dealing with regression tests, it does cause problems. Since that species is not present in the json files the tests fail. The other way around happens when I create a regression test - the nbody species is added to the locally created json file and then the test fails when run on azure since there is no nbody species there. Have others seen this? I have a hack to the `Regression/Checksum/checksum.py` file that fixes this, adding the line `species_list.discard('nbody')` after line 54 in the file.
1.0
yt: nbody species appearing - I've been having the issue when using a newer version of yt, that the extra species nbody appears when reading in plot files. For regular post processing, this is not a problem. However, when dealing with regression tests, it does cause problems. Since that species is not present in the json files the tests fail. The other way around happens when I create a regression test - the nbody species is added to the locally created json file and then the test fails when run on azure since there is no nbody species there. Have others seen this? I have a hack to the `Regression/Checksum/checksum.py` file that fixes this, adding the line `species_list.discard('nbody')` after line 54 in the file.
process
yt nbody species appearing i ve been having the issue when using a newer version of yt that the extra species nbody appears when reading in plot files for regular post processing this is not a problem however when dealing with regression tests it does cause problems since that species is not present in the json files the tests fail the other way around happens when i create a regression test the nbody species is added to the locally created json file and then the test fails when run on azure since there is no nbody species there have others seen this i have a hack to the regression checksum checksum py file that fixes this adding the line species list discard nbody after line in the file
1
123,154
16,449,477,271
IssuesEvent
2021-05-21 02:04:01
mobigen/IRIS-BigData-Platform
https://api.github.com/repos/mobigen/IRIS-BigData-Platform
closed
보고서 - 콤보박스 안에 스크롤이 가능한 체크박스 형태
#Design #IBP Priority: P2 Status: Backlog
## 기능 요청 ## 문제가 무엇인지에 대한 명확하고 간결한 설명 부탁드립니다. 보고서 - 콤보박스 안에 스크롤이 가능한 체크박스 형태 콤보박스처럼 접었다가 펼수 있는 체크박스 형태와 스크롤이 가능한 체크박스 형태를 원합니다. ## 원하는 솔루션 설명 ## { 원하는 기능에 대한 명확하고 간결한 설명 부탁드립니다 } psi 시연 과제를 하다보니 한정된 대시보드 화면 사이즈에서 현재 아이리스 체크박스의 형태로는 체크박스의 내용이 많아질 경우(200개 넘어가는경우 등) 화면상의 제약이 있어 화면의 대부분을 체크박스가 차지하는 경우가 생겼습니다. 현재 태블로에서는 체크박스가 콤보박스내부에 체크박스가 존재하고, 스크롤이 가능한 형태로 되어 있는데, 기능 개발이 되었으면 좋겠습니다. ## 고려한 다른 대안 ## ## 기타 ## 기능 요청에 대한 다른 의견 또는 스크린샷을 여기에 모두 추가 부탁드립니다. <img width="237" alt="스크린샷 2020-03-06 오후 1 34 28" src="https://user-images.githubusercontent.com/36151180/76053151-9a3fa800-5faf-11ea-9bc0-e7469fa927ed.png"> { 고려한 대체 솔루션이나 기능에 대한 명확하고 간결한 설명 부탁드립니다. } <img width="245" alt="스크린샷 2020-03-06 오후 1 34 23" src="https://user-images.githubusercontent.com/36151180/76053145-97dd4e00-5faf-11ea-95b9-7ee060df86a2.png"> ![KakaoTalk_Photo_2020-03-06-13-42-36](https://user-images.githubusercontent.com/36151180/76053385-6022d600-5fb0-11ea-9222-48a28aa62342.png)
1.0
보고서 - 콤보박스 안에 스크롤이 가능한 체크박스 형태 - ## 기능 요청 ## 문제가 무엇인지에 대한 명확하고 간결한 설명 부탁드립니다. 보고서 - 콤보박스 안에 스크롤이 가능한 체크박스 형태 콤보박스처럼 접었다가 펼수 있는 체크박스 형태와 스크롤이 가능한 체크박스 형태를 원합니다. ## 원하는 솔루션 설명 ## { 원하는 기능에 대한 명확하고 간결한 설명 부탁드립니다 } psi 시연 과제를 하다보니 한정된 대시보드 화면 사이즈에서 현재 아이리스 체크박스의 형태로는 체크박스의 내용이 많아질 경우(200개 넘어가는경우 등) 화면상의 제약이 있어 화면의 대부분을 체크박스가 차지하는 경우가 생겼습니다. 현재 태블로에서는 체크박스가 콤보박스내부에 체크박스가 존재하고, 스크롤이 가능한 형태로 되어 있는데, 기능 개발이 되었으면 좋겠습니다. ## 고려한 다른 대안 ## ## 기타 ## 기능 요청에 대한 다른 의견 또는 스크린샷을 여기에 모두 추가 부탁드립니다. <img width="237" alt="스크린샷 2020-03-06 오후 1 34 28" src="https://user-images.githubusercontent.com/36151180/76053151-9a3fa800-5faf-11ea-9bc0-e7469fa927ed.png"> { 고려한 대체 솔루션이나 기능에 대한 명확하고 간결한 설명 부탁드립니다. } <img width="245" alt="스크린샷 2020-03-06 오후 1 34 23" src="https://user-images.githubusercontent.com/36151180/76053145-97dd4e00-5faf-11ea-95b9-7ee060df86a2.png"> ![KakaoTalk_Photo_2020-03-06-13-42-36](https://user-images.githubusercontent.com/36151180/76053385-6022d600-5fb0-11ea-9222-48a28aa62342.png)
non_process
보고서 콤보박스 안에 스크롤이 가능한 체크박스 형태 기능 요청 문제가 무엇인지에 대한 명확하고 간결한 설명 부탁드립니다 보고서 콤보박스 안에 스크롤이 가능한 체크박스 형태 콤보박스처럼 접었다가 펼수 있는 체크박스 형태와 스크롤이 가능한 체크박스 형태를 원합니다 원하는 솔루션 설명 원하는 기능에 대한 명확하고 간결한 설명 부탁드립니다 psi 시연 과제를 하다보니 한정된 대시보드 화면 사이즈에서 현재 아이리스 체크박스의 형태로는 체크박스의 내용이 많아질 경우 넘어가는경우 등 화면상의 제약이 있어 화면의 대부분을 체크박스가 차지하는 경우가 생겼습니다 현재 태블로에서는 체크박스가 콤보박스내부에 체크박스가 존재하고 스크롤이 가능한 형태로 되어 있는데 기능 개발이 되었으면 좋겠습니다 고려한 다른 대안 기타 기능 요청에 대한 다른 의견 또는 스크린샷을 여기에 모두 추가 부탁드립니다 img width alt 스크린샷 오후 src 고려한 대체 솔루션이나 기능에 대한 명확하고 간결한 설명 부탁드립니다 img width alt 스크린샷 오후 src
0
10,922
3,441,380,871
IssuesEvent
2015-12-14 18:13:57
dotnet/wcf
https://api.github.com/repos/dotnet/wcf
closed
Improve legend for what's supported table and update all cells to reflect RC1
documentation
2 parts to this task: 1. Improve the granularity of the emoji's to distinguish better between "what works completely" to "what works for the paths we've tested" 2. Go through and choose the right emoji's for each table cell
1.0
Improve legend for what's supported table and update all cells to reflect RC1 - 2 parts to this task: 1. Improve the granularity of the emoji's to distinguish better between "what works completely" to "what works for the paths we've tested" 2. Go through and choose the right emoji's for each table cell
non_process
improve legend for what s supported table and update all cells to reflect parts to this task improve the granularity of the emoji s to distinguish better between what works completely to what works for the paths we ve tested go through and choose the right emoji s for each table cell
0
648,712
21,192,587,242
IssuesEvent
2022-04-08 19:17:21
status-im/status-desktop
https://api.github.com/repos/status-im/status-desktop
opened
image missing in get your keys screen
bug onboarding priority 2: medium
# Bug Report ## Description ![Screen Shot 2022-04-08 at 3 12 54 PM](https://user-images.githubusercontent.com/176720/162509143-0be0d15a-e3f0-47e2-841f-2936db30b1f2.png) ## Steps to reproduce 1. use a new data folder 2. start app and choose "I am new to Status" #### Expected behavior ![Screen Shot 2022-04-08 at 3 16 36 PM](https://user-images.githubusercontent.com/176720/162509622-0d400f5c-9b3d-413a-8f37-c9e4947dde46.png)
1.0
image missing in get your keys screen - # Bug Report ## Description ![Screen Shot 2022-04-08 at 3 12 54 PM](https://user-images.githubusercontent.com/176720/162509143-0be0d15a-e3f0-47e2-841f-2936db30b1f2.png) ## Steps to reproduce 1. use a new data folder 2. start app and choose "I am new to Status" #### Expected behavior ![Screen Shot 2022-04-08 at 3 16 36 PM](https://user-images.githubusercontent.com/176720/162509622-0d400f5c-9b3d-413a-8f37-c9e4947dde46.png)
non_process
image missing in get your keys screen bug report description steps to reproduce use a new data folder start app and choose i am new to status expected behavior
0
707
9,979,386,976
IssuesEvent
2019-07-09 22:40:57
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
opened
Completion crashes (fault) if more than instance of Roslyn completion service (different languages) are attached to the single completion session
Area-IDE Tenet-Reliability
from activity.log ``` Exception from ActivityLog.xml: System.ArgumentException: An entry with the same key already exists. at System.Collections.Specialized.ListDictionary.Add(Object key, Object value) at System.Collections.Specialized.HybridDictionary.Add(Object key, Object value) at Microsoft.VisualStudio.Utilities.PropertyCollection.AddProperty(Object key, Object property) at Microsoft.CodeAnalysis.Editor.Implementation.IntelliSense.AsyncCompletion.CompletionSource.&lt;GetCompletionContextAsync&gt;d__23.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at Microsoft.VisualStudio.Text.Utilities.GuardedOperations.&lt;CallExtensionPointAsync&gt;d__34`1.MoveNext() --- End of stack trace from previous location where exception was thrown --- at Microsoft.VisualStudio.Telemetry.WindowsErrorReporting.WatsonReport.GetClrWatsonExceptionInfo(Exception exceptionObject) ```
True
Completion crashes (fault) if more than instance of Roslyn completion service (different languages) are attached to the single completion session - from activity.log ``` Exception from ActivityLog.xml: System.ArgumentException: An entry with the same key already exists. at System.Collections.Specialized.ListDictionary.Add(Object key, Object value) at System.Collections.Specialized.HybridDictionary.Add(Object key, Object value) at Microsoft.VisualStudio.Utilities.PropertyCollection.AddProperty(Object key, Object property) at Microsoft.CodeAnalysis.Editor.Implementation.IntelliSense.AsyncCompletion.CompletionSource.&lt;GetCompletionContextAsync&gt;d__23.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at Microsoft.VisualStudio.Text.Utilities.GuardedOperations.&lt;CallExtensionPointAsync&gt;d__34`1.MoveNext() --- End of stack trace from previous location where exception was thrown --- at Microsoft.VisualStudio.Telemetry.WindowsErrorReporting.WatsonReport.GetClrWatsonExceptionInfo(Exception exceptionObject) ```
non_process
completion crashes fault if more than instance of roslyn completion service different languages are attached to the single completion session from activity log exception from activitylog xml system argumentexception an entry with the same key already exists at system collections specialized listdictionary add object key object value at system collections specialized hybriddictionary add object key object value at microsoft visualstudio utilities propertycollection addproperty object key object property at microsoft codeanalysis editor implementation intellisense asynccompletion completionsource lt getcompletioncontextasync gt d movenext end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at microsoft visualstudio text utilities guardedoperations lt callextensionpointasync gt d movenext end of stack trace from previous location where exception was thrown at microsoft visualstudio telemetry windowserrorreporting watsonreport getclrwatsonexceptioninfo exception exceptionobject
0
21,992
30,485,991,134
IssuesEvent
2023-07-18 02:16:17
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
Windows: Reduce raciness of `spawn()` vs. `SetCurrentDirectory`
O-windows C-bug T-libs A-process
I was inspecting `resolve_exe()` specifically looking for racy coding patterns. The code currently contains this: ```rust if !path::is_file_name(exe_path) { if has_exe_suffix { // The application name is a path to a `.exe` file. // Let `CreateProcessW` figure out if it exists or not. return Ok(exe_path.into()); } let mut path = PathBuf::from(exe_path); // Append `.exe` if not already there. path = path::append_suffix(path, EXE_SUFFIX.as_ref()); if path.try_exists().unwrap_or(false) { return Ok(path); } else { // It's ok to use `set_extension` here because the intent is to // remove the extension that was just added. path.set_extension(""); return Ok(path); } ``` Consider what happens when we pass a relative path containing a path separator, e.g. "target/something". This code will check if the file `target/something.exe` exists, relative to the current working directory and if so, return `"target/something.exe"`. This function is used by the implementation of `spawn()`. Roughly, spawn looks like this: ```rust let program = resolve_exe(&self.program, || env::var_os("PATH"), child_paths)?; ... let (dirp, _data) = make_dirp(self.cwd.as_ref())?; ... let program = to_u16s(&program)?; unsafe { cvt(c::CreateProcessW( program.as_ptr(), ... dirp, // lpCurrentDirectory &mut pi, )) }?; ``` In theory the current working directory could change at any time. To minimize the effect that would have on `spawn()`, it seems like it could instead be written like this: ```rust let cwd = self.cwd.as_ref(); let program = resolve_exe(cwd, &self.program, || env::var_os("PATH"), child_paths)?; ... let (dirp, _data) = make_dirp(cwd)?; ... let program = to_u16s(&program)?; unsafe { cvt(c::CreateProcessW( program.as_ptr(), ... dirp, // lpCurrentDirectory &mut pi, )) }?; ``` And then change `resolve_exe` so that it always returns an absolute path, where it always resolves any relative paths relative to the value of the `cwd` parameter passed in. `spawn()` and `SetCurrentDirectory` would still be racy, but at least all the path resolution within `spawn` would be internally consistent with respect to the current directory at a point in time.
1.0
Windows: Reduce raciness of `spawn()` vs. `SetCurrentDirectory` - I was inspecting `resolve_exe()` specifically looking for racy coding patterns. The code currently contains this: ```rust if !path::is_file_name(exe_path) { if has_exe_suffix { // The application name is a path to a `.exe` file. // Let `CreateProcessW` figure out if it exists or not. return Ok(exe_path.into()); } let mut path = PathBuf::from(exe_path); // Append `.exe` if not already there. path = path::append_suffix(path, EXE_SUFFIX.as_ref()); if path.try_exists().unwrap_or(false) { return Ok(path); } else { // It's ok to use `set_extension` here because the intent is to // remove the extension that was just added. path.set_extension(""); return Ok(path); } ``` Consider what happens when we pass a relative path containing a path separator, e.g. "target/something". This code will check if the file `target/something.exe` exists, relative to the current working directory and if so, return `"target/something.exe"`. This function is used by the implementation of `spawn()`. Roughly, spawn looks like this: ```rust let program = resolve_exe(&self.program, || env::var_os("PATH"), child_paths)?; ... let (dirp, _data) = make_dirp(self.cwd.as_ref())?; ... let program = to_u16s(&program)?; unsafe { cvt(c::CreateProcessW( program.as_ptr(), ... dirp, // lpCurrentDirectory &mut pi, )) }?; ``` In theory the current working directory could change at any time. To minimize the effect that would have on `spawn()`, it seems like it could instead be written like this: ```rust let cwd = self.cwd.as_ref(); let program = resolve_exe(cwd, &self.program, || env::var_os("PATH"), child_paths)?; ... let (dirp, _data) = make_dirp(cwd)?; ... let program = to_u16s(&program)?; unsafe { cvt(c::CreateProcessW( program.as_ptr(), ... dirp, // lpCurrentDirectory &mut pi, )) }?; ``` And then change `resolve_exe` so that it always returns an absolute path, where it always resolves any relative paths relative to the value of the `cwd` parameter passed in. `spawn()` and `SetCurrentDirectory` would still be racy, but at least all the path resolution within `spawn` would be internally consistent with respect to the current directory at a point in time.
process
windows reduce raciness of spawn vs setcurrentdirectory i was inspecting resolve exe specifically looking for racy coding patterns the code currently contains this rust if path is file name exe path if has exe suffix the application name is a path to a exe file let createprocessw figure out if it exists or not return ok exe path into let mut path pathbuf from exe path append exe if not already there path path append suffix path exe suffix as ref if path try exists unwrap or false return ok path else it s ok to use set extension here because the intent is to remove the extension that was just added path set extension return ok path consider what happens when we pass a relative path containing a path separator e g target something this code will check if the file target something exe exists relative to the current working directory and if so return target something exe this function is used by the implementation of spawn roughly spawn looks like this rust let program resolve exe self program env var os path child paths let dirp data make dirp self cwd as ref let program to program unsafe cvt c createprocessw program as ptr dirp lpcurrentdirectory mut pi in theory the current working directory could change at any time to minimize the effect that would have on spawn it seems like it could instead be written like this rust let cwd self cwd as ref let program resolve exe cwd self program env var os path child paths let dirp data make dirp cwd let program to program unsafe cvt c createprocessw program as ptr dirp lpcurrentdirectory mut pi and then change resolve exe so that it always returns an absolute path where it always resolves any relative paths relative to the value of the cwd parameter passed in spawn and setcurrentdirectory would still be racy but at least all the path resolution within spawn would be internally consistent with respect to the current directory at a point in time
1
88,031
25,283,521,921
IssuesEvent
2022-11-16 17:24:58
getll/pokemon-web-services
https://api.github.com/repos/getll/pokemon-web-services
closed
TASK 2.b: Populate Games, Generations, Gyms, Locations Tables
Build #1 Sub Task
Creating the Games, Generations, Gyms, and Locations tables. ### Requirements - XAMPP - phpMyAdmin ### Dependencies - None
1.0
TASK 2.b: Populate Games, Generations, Gyms, Locations Tables - Creating the Games, Generations, Gyms, and Locations tables. ### Requirements - XAMPP - phpMyAdmin ### Dependencies - None
non_process
task b populate games generations gyms locations tables creating the games generations gyms and locations tables requirements xampp phpmyadmin dependencies none
0
413,226
12,061,556,032
IssuesEvent
2020-04-16 00:10:24
eclipse-ee4j/glassfish
https://api.github.com/repos/eclipse-ee4j/glassfish
closed
gf4, web console athentication failed with 64 bit jdk7.
Component: admin ERR: Assignee Priority: Major Stale Type: Bug
The administration web console denies login if glassfish 4.0 ml is started with a JAVA_HOME that poits to a 64 bit jdk. This happens even if there is no passord for the admin user. The asadmin login command always works instead. The same glassfish installation if run with the corresponding x86 version of the jdk (jdk1.7.0_55 x86) works as advertised and does not exhibit the problem. #### Environment win7 x64, jdk1.7.0_55 x64. #### Affected Versions [4.0]
1.0
gf4, web console athentication failed with 64 bit jdk7. - The administration web console denies login if glassfish 4.0 ml is started with a JAVA_HOME that poits to a 64 bit jdk. This happens even if there is no passord for the admin user. The asadmin login command always works instead. The same glassfish installation if run with the corresponding x86 version of the jdk (jdk1.7.0_55 x86) works as advertised and does not exhibit the problem. #### Environment win7 x64, jdk1.7.0_55 x64. #### Affected Versions [4.0]
non_process
web console athentication failed with bit the administration web console denies login if glassfish ml is started with a java home that poits to a bit jdk this happens even if there is no passord for the admin user the asadmin login command always works instead the same glassfish installation if run with the corresponding version of the jdk works as advertised and does not exhibit the problem environment affected versions
0
5,329
8,142,963,715
IssuesEvent
2018-08-21 09:19:15
aiidateam/aiida_core
https://api.github.com/repos/aiidateam/aiida_core
closed
JobCalculation states not properly set when JobProcess is terminated
priority/critical-blocking topic/JobCalculationAndProcess topic/Workflows type/bug
The `state` attribute of a `JobCalculation` should be set accordingly when the `JobProcess` reaches a terminal state. When it finishes normally, nothing should be done most likely, however, when it reaches `Killed` or `Excepted` it is crucial that the calculation state be updated accordingly
1.0
JobCalculation states not properly set when JobProcess is terminated - The `state` attribute of a `JobCalculation` should be set accordingly when the `JobProcess` reaches a terminal state. When it finishes normally, nothing should be done most likely, however, when it reaches `Killed` or `Excepted` it is crucial that the calculation state be updated accordingly
process
jobcalculation states not properly set when jobprocess is terminated the state attribute of a jobcalculation should be set accordingly when the jobprocess reaches a terminal state when it finishes normally nothing should be done most likely however when it reaches killed or excepted it is crucial that the calculation state be updated accordingly
1
133,989
5,218,120,275
IssuesEvent
2017-01-26 15:44:11
michaeljcalkins/rangersteve-ideas
https://api.github.com/repos/michaeljcalkins/rangersteve-ideas
opened
Optimal entity interpolation
Priority: Critical
I don't know if tweening itself is performance heavy but we need some sort of interpolation.
1.0
Optimal entity interpolation - I don't know if tweening itself is performance heavy but we need some sort of interpolation.
non_process
optimal entity interpolation i don t know if tweening itself is performance heavy but we need some sort of interpolation
0
13,238
15,706,655,033
IssuesEvent
2021-03-26 17:43:44
netlify/build
https://api.github.com/repos/netlify/build
closed
Improve lingering processes message color
proj/lingering-processes type: chore
Background at https://github.com/netlify/build/issues/2371 The lingering processes message is a warning, not an error. Its color should be yellow, not red. We should use this new warning color for other warnings in `@netlify/build` and `@netlify/config` as well.
1.0
Improve lingering processes message color - Background at https://github.com/netlify/build/issues/2371 The lingering processes message is a warning, not an error. Its color should be yellow, not red. We should use this new warning color for other warnings in `@netlify/build` and `@netlify/config` as well.
process
improve lingering processes message color background at the lingering processes message is a warning not an error its color should be yellow not red we should use this new warning color for other warnings in netlify build and netlify config as well
1
2,385
5,187,641,399
IssuesEvent
2017-01-20 17:24:41
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
closed
No processes are displayed when starting process within task app
browser: all bug comp: activiti-processList
1. Go to task app 2. Go to processes 3. Start a process **Expected results** List of avaliable processes are displayed **Actual results** No processes are displayed N.b. processes display within individual apps <img width="576" alt="screen shot 2017-01-09 at 11 50 07" src="https://cloud.githubusercontent.com/assets/13200338/21765758/bb52885c-d662-11e6-865d-0ee686d29adc.png">
1.0
No processes are displayed when starting process within task app - 1. Go to task app 2. Go to processes 3. Start a process **Expected results** List of avaliable processes are displayed **Actual results** No processes are displayed N.b. processes display within individual apps <img width="576" alt="screen shot 2017-01-09 at 11 50 07" src="https://cloud.githubusercontent.com/assets/13200338/21765758/bb52885c-d662-11e6-865d-0ee686d29adc.png">
process
no processes are displayed when starting process within task app go to task app go to processes start a process expected results list of avaliable processes are displayed actual results no processes are displayed n b processes display within individual apps img width alt screen shot at src
1
18,850
24,763,987,847
IssuesEvent
2022-10-22 09:06:48
home-climate-control/dz
https://api.github.com/repos/home-climate-control/dz
closed
SingleStageUnitController demand output is wrong
annoyance process control
### Expected Behavior `HvacCommand` emitted by `SingleStageUnitController` reflects the actual demand value. ### Actual Behavior The value is always 1 when the HVAC is requested to be on. This throws off the metrics.
1.0
SingleStageUnitController demand output is wrong - ### Expected Behavior `HvacCommand` emitted by `SingleStageUnitController` reflects the actual demand value. ### Actual Behavior The value is always 1 when the HVAC is requested to be on. This throws off the metrics.
process
singlestageunitcontroller demand output is wrong expected behavior hvaccommand emitted by singlestageunitcontroller reflects the actual demand value actual behavior the value is always when the hvac is requested to be on this throws off the metrics
1