Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
14,352
| 17,375,052,868
|
IssuesEvent
|
2021-07-30 19:37:16
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Reference to Obsolete Version of Node.js
|
Pri2 devops-cicd-process/tech devops/prod doc-enhancement
|
This section specifies Node.js version 6, which is obsolete:
https://docs.microsoft.com/en-gb/azure/devops/pipelines/process/container-phases?view=azure-devops#bring-your-own-nodejs
The text reads:
"Node 6 is a safe choice. You can start from the node:6-alpine image."
---
#### Document details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 3339a2e0-be29-1363-f588-b231d4472c02
* Version Independent ID: 72dd11a3-704d-d0fd-6dfa-cf49f3352de3
* Content: [Container Jobs in Azure Pipelines and TFS - Azure Pipelines](https://docs.microsoft.com/en-gb/azure/devops/pipelines/process/container-phases?view=azure-devops)
* Content Source: [docs/pipelines/process/container-phases.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/container-phases.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Reference to Obsolete Version of Node.js - This section specifies Node.js version 6, which is obsolete:
https://docs.microsoft.com/en-gb/azure/devops/pipelines/process/container-phases?view=azure-devops#bring-your-own-nodejs
The text reads:
"Node 6 is a safe choice. You can start from the node:6-alpine image."
---
#### Document details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 3339a2e0-be29-1363-f588-b231d4472c02
* Version Independent ID: 72dd11a3-704d-d0fd-6dfa-cf49f3352de3
* Content: [Container Jobs in Azure Pipelines and TFS - Azure Pipelines](https://docs.microsoft.com/en-gb/azure/devops/pipelines/process/container-phases?view=azure-devops)
* Content Source: [docs/pipelines/process/container-phases.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/container-phases.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
reference to obsolete version of node js this section specifies node js version which is obsolete the text reads node is a safe choice you can start from the node alpine image document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
21,789
| 30,297,622,262
|
IssuesEvent
|
2023-07-10 01:18:05
|
winter-telescope/mirar
|
https://api.github.com/repos/winter-telescope/mirar
|
opened
|
[BUG] Hardcoded assumption of split image
|
bug processors
|
**Describe the bug**
Some processors have undeclared requirements that they only work if an image has been split, e.g:
https://github.com/winter-telescope/mirar/blob/93ecacddd58f7814a30d045df4e53325aa7da570/mirar/references/ukirt.py#L655
This is super unintuitive behaviour, and breaks the golden rule that processors cannot expect a key unless it is listed in https://github.com/winter-telescope/mirar/blob/93ecacddd58f7814a30d045df4e53325aa7da570/mirar/paths.py#L291, or it is listed as a prerequisite requirement in the processors. Further, these processors do not really need the images to be split at all, the two things are independent. They just need to refer uniquely to a field.
What should we do?
I can think of:
- rewrite the processors to be agnostic,
- make the subcoordid required
- Abolish subcoordid, and start modifying the required key `FIELDID` in split
Alternative solutions welcome.
|
1.0
|
[BUG] Hardcoded assumption of split image - **Describe the bug**
Some processors have undeclared requirements that they only work if an image has been split, e.g:
https://github.com/winter-telescope/mirar/blob/93ecacddd58f7814a30d045df4e53325aa7da570/mirar/references/ukirt.py#L655
This is super unintuitive behaviour, and breaks the golden rule that processors cannot expect a key unless it is listed in https://github.com/winter-telescope/mirar/blob/93ecacddd58f7814a30d045df4e53325aa7da570/mirar/paths.py#L291, or it is listed as a prerequisite requirement in the processors. Further, these processors do not really need the images to be split at all, the two things are independent. They just need to refer uniquely to a field.
What should we do?
I can think of:
- rewrite the processors to be agnostic,
- make the subcoordid required
- Abolish subcoordid, and start modifying the required key `FIELDID` in split
Alternative solutions welcome.
|
process
|
hardcoded assumption of split image describe the bug some processors have undeclared requirements that they only work if an image has been split e g this is super unintuitive behaviour and breaks the golden rule that processors cannot expect a key unless it is listed in or it is listed as a prerequisite requirement in the processors further these processors do not really need the images to be split at all the two things are independent they just need to refer uniquely to a field what should we do i can think of rewrite the processors to be agnostic make the subcoordid required abolish subcoordid and start modifying the required key fieldid in split alternative solutions welcome
| 1
|
20,564
| 27,224,993,180
|
IssuesEvent
|
2023-02-21 09:05:25
|
redhat-developer/vscode-java
|
https://api.github.com/repos/redhat-developer/vscode-java
|
closed
|
Failed build when using records with mapstruct
|
bug upstream annotation-processing
|
##### Environment
- Operating System: MacOS Ventura 13.1
- JDK version: openjdk-19
- Visual Studio Code version: 1.74.3
- Java extension version: 1.14.0
##### Steps To Reproduce
1. Clone https://github.com/mapstruct/mapstruct-examples
2. Open mapstruct-record in Visual Studio Code
3. Rename source = "email" to source = "name" in CustomerMapper.java
##### Additional Informations
I'm getting strange a error while using mapstruct with records:
`No property named "name" exists in source parameter(s). Type "CustomerDto" has no properties.Java(0)`
The error disappears after running "Java: rebuild project".
This error is not a new one, I saw it with older versions of the extension too. When I try to run or debug the project a warning for "Build was failed" appears
|
1.0
|
Failed build when using records with mapstruct - ##### Environment
- Operating System: MacOS Ventura 13.1
- JDK version: openjdk-19
- Visual Studio Code version: 1.74.3
- Java extension version: 1.14.0
##### Steps To Reproduce
1. Clone https://github.com/mapstruct/mapstruct-examples
2. Open mapstruct-record in Visual Studio Code
3. Rename source = "email" to source = "name" in CustomerMapper.java
##### Additional Informations
I'm getting strange a error while using mapstruct with records:
`No property named "name" exists in source parameter(s). Type "CustomerDto" has no properties.Java(0)`
The error disappears after running "Java: rebuild project".
This error is not a new one, I saw it with older versions of the extension too. When I try to run or debug the project a warning for "Build was failed" appears
|
process
|
failed build when using records with mapstruct environment operating system macos ventura jdk version openjdk visual studio code version java extension version steps to reproduce clone open mapstruct record in visual studio code rename source email to source name in customermapper java additional informations i m getting strange a error while using mapstruct with records no property named name exists in source parameter s type customerdto has no properties java the error disappears after running java rebuild project this error is not a new one i saw it with older versions of the extension too when i try to run or debug the project a warning for build was failed appears
| 1
|
20,870
| 27,657,188,695
|
IssuesEvent
|
2023-03-12 04:21:41
|
pentium3/sys_reading
|
https://api.github.com/repos/pentium3/sys_reading
|
opened
|
CrystalPerf: Learning to Characterize the Performance of Dataflow Computation through Code Analysis
|
stream processing
|
https://www.usenix.org/system/files/atc21-tian.pdf
|
1.0
|
CrystalPerf: Learning to Characterize the Performance of Dataflow Computation through Code Analysis - https://www.usenix.org/system/files/atc21-tian.pdf
|
process
|
crystalperf learning to characterize the performance of dataflow computation through code analysis
| 1
|
10,517
| 13,299,248,606
|
IssuesEvent
|
2020-08-25 09:27:13
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
Fuzz test failed for tikv::coprocessor::codec::mysql::Duration::parse
|
sig/coprocessor type/bug
|
## Bug Report
**What version of TiKV are you using?**
3.0.0-beta.1
**What operating system and CPU are you using?**
macOS 10.14.4
**What did you do?**
The fuzz test added by WIP PR #4608 failed. This is the test:
```rust
pub fn fuzz_coprocessor_codec_duration_from_parse(data: &[u8]) -> Result<(), Error> {
use std::io::Read;
use tikv::coprocessor::codec::mysql::Duration;
let mut cursor = Cursor::new(data);
let fsp = cursor.read_as_i8()?;
let mut buf: [u8; 32] = [b' '; 32];
cursor.read_exact(&mut buf)?;
let d = Duration::parse(&buf, fsp)?;
fuzz_duration(d, cursor)
}
```
This is the error report:
```
thread '<unnamed>' panicked at 'attempt to multiply with overflow', src/coprocessor/codec/mysql/duration.rs:191:21
stack backtrace:
0: std::sys::unix::backtrace::tracing::imp::unwind_backtrace
1: std::sys_common::backtrace::_print
2: std::panicking::default_hook::{{closure}}
3: std::panicking::default_hook
4: std::panicking::rust_panic_with_hook
5: std::panicking::continue_panic_fmt
6: rust_begin_unwind
7: core::panicking::panic_fmt
8: core::panicking::panic
9: tikv::coprocessor::codec::mysql::duration::Duration::parse
10: fuzz_targets::fuzz_coprocessor_codec_duration_from_parse
11: rust_fuzzer_test_input
12: __rust_maybe_catch_panic
13: std::panicking::try
14: LLVMFuzzerTestOneInput
15: _ZN6fuzzer6Fuzzer15ExecuteCallbackEPKhm
16: _ZN6fuzzer6Fuzzer6RunOneEPKhmbPNS_9InputInfoEPb
17: _ZN6fuzzer6Fuzzer16MutateAndTestOneEv
18: _ZN6fuzzer6Fuzzer4LoopERKNSt3__16vectorINS1_12basic_stringIcNS1_11char_traitsIcEENS1_9allocatorIcEEEENS_16fuzzer_allocatorIS8_EEEE
19: _ZN6fuzzer12FuzzerDriverEPiPPPcPFiPKhmE
20: main
==56552== ERROR: libFuzzer: deadly signal
#0 0x1166d1707 in __sanitizer_print_stack_trace (lib__rustc__clang_rt.asan_osx_dynamic.dylib:x86_64+0x68707)
#1 0x110bd583b in fuzzer::Fuzzer::CrashCallback() (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b7383b)
#2 0x110bd57ed in fuzzer::Fuzzer::StaticCrashSignalCallback() (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b737ed)
#3 0x110c16437 in fuzzer::CrashHandler(int, __siginfo*, void*) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105bb4437)
#4 0x7fff590abb5c in _sigtramp (libsystem_platform.dylib:x86_64+0x4b5c)
#5 0x3500000bd0 (<unknown module>)
#6 0x110c4a7ee in __rust_maybe_catch_panic (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105be87ee)
#7 0x110bd3fde in std::panicking::try::hd4964c260fccd748 (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b71fde)
#8 0x110bd3a72 in LLVMFuzzerTestOneInput (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b71a72)
#9 0x110bd6f9a in fuzzer::Fuzzer::ExecuteCallback(unsigned char const*, unsigned long) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b74f9a)
#10 0x110bd6999 in fuzzer::Fuzzer::RunOne(unsigned char const*, unsigned long, bool, fuzzer::InputInfo*, bool*) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b74999)
#11 0x110bd85c1 in fuzzer::Fuzzer::MutateAndTestOne() (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b765c1)
#12 0x110bd9871 in fuzzer::Fuzzer::Loop(std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, fuzzer::fuzzer_allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > > > const&) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b77871)
#13 0x110c04ab8 in fuzzer::FuzzerDriver(int*, char***, int (*)(unsigned char const*, unsigned long)) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105ba2ab8)
#14 0x110c27179 in main (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105bc5179)
#15 0x7fff58ec63d4 in start (libdyld.dylib:x86_64+0x163d4)
NOTE: libFuzzer has rudimentary signal handlers.
Combine libFuzzer with AddressSanitizer or similar for better crash reports.
SUMMARY: libFuzzer: deadly signal
MS: 2 ChangeBit-CopyPart-; base unit: 4744b705bce806db21998c6b5c000865f87bf51a
0xff,0x35,0x35,0x35,0x35,0x35,0x35,0x35,0x35,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x20,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x32,0x33,0x35,0x35,0x35,0x0,0x1,
\xff555555550000000 0000000000023555\x00\x01
artifact_prefix='./'; Test unit written to ./crash-eda52eca653e470cf2fed7ad070b4f09e122dab9
Base64: /zU1NTU1NTU1MDAwMDAwMCAwMDAwMDAwMDAwMDIzNTU1AAE=
Running fuzzer failed: Libfuzzer exited with code Some(77)
Error: 1
```
|
1.0
|
Fuzz test failed for tikv::coprocessor::codec::mysql::Duration::parse - ## Bug Report
**What version of TiKV are you using?**
3.0.0-beta.1
**What operating system and CPU are you using?**
macOS 10.14.4
**What did you do?**
The fuzz test added by WIP PR #4608 failed. This is the test:
```rust
pub fn fuzz_coprocessor_codec_duration_from_parse(data: &[u8]) -> Result<(), Error> {
use std::io::Read;
use tikv::coprocessor::codec::mysql::Duration;
let mut cursor = Cursor::new(data);
let fsp = cursor.read_as_i8()?;
let mut buf: [u8; 32] = [b' '; 32];
cursor.read_exact(&mut buf)?;
let d = Duration::parse(&buf, fsp)?;
fuzz_duration(d, cursor)
}
```
This is the error report:
```
thread '<unnamed>' panicked at 'attempt to multiply with overflow', src/coprocessor/codec/mysql/duration.rs:191:21
stack backtrace:
0: std::sys::unix::backtrace::tracing::imp::unwind_backtrace
1: std::sys_common::backtrace::_print
2: std::panicking::default_hook::{{closure}}
3: std::panicking::default_hook
4: std::panicking::rust_panic_with_hook
5: std::panicking::continue_panic_fmt
6: rust_begin_unwind
7: core::panicking::panic_fmt
8: core::panicking::panic
9: tikv::coprocessor::codec::mysql::duration::Duration::parse
10: fuzz_targets::fuzz_coprocessor_codec_duration_from_parse
11: rust_fuzzer_test_input
12: __rust_maybe_catch_panic
13: std::panicking::try
14: LLVMFuzzerTestOneInput
15: _ZN6fuzzer6Fuzzer15ExecuteCallbackEPKhm
16: _ZN6fuzzer6Fuzzer6RunOneEPKhmbPNS_9InputInfoEPb
17: _ZN6fuzzer6Fuzzer16MutateAndTestOneEv
18: _ZN6fuzzer6Fuzzer4LoopERKNSt3__16vectorINS1_12basic_stringIcNS1_11char_traitsIcEENS1_9allocatorIcEEEENS_16fuzzer_allocatorIS8_EEEE
19: _ZN6fuzzer12FuzzerDriverEPiPPPcPFiPKhmE
20: main
==56552== ERROR: libFuzzer: deadly signal
#0 0x1166d1707 in __sanitizer_print_stack_trace (lib__rustc__clang_rt.asan_osx_dynamic.dylib:x86_64+0x68707)
#1 0x110bd583b in fuzzer::Fuzzer::CrashCallback() (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b7383b)
#2 0x110bd57ed in fuzzer::Fuzzer::StaticCrashSignalCallback() (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b737ed)
#3 0x110c16437 in fuzzer::CrashHandler(int, __siginfo*, void*) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105bb4437)
#4 0x7fff590abb5c in _sigtramp (libsystem_platform.dylib:x86_64+0x4b5c)
#5 0x3500000bd0 (<unknown module>)
#6 0x110c4a7ee in __rust_maybe_catch_panic (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105be87ee)
#7 0x110bd3fde in std::panicking::try::hd4964c260fccd748 (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b71fde)
#8 0x110bd3a72 in LLVMFuzzerTestOneInput (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b71a72)
#9 0x110bd6f9a in fuzzer::Fuzzer::ExecuteCallback(unsigned char const*, unsigned long) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b74f9a)
#10 0x110bd6999 in fuzzer::Fuzzer::RunOne(unsigned char const*, unsigned long, bool, fuzzer::InputInfo*, bool*) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b74999)
#11 0x110bd85c1 in fuzzer::Fuzzer::MutateAndTestOne() (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b765c1)
#12 0x110bd9871 in fuzzer::Fuzzer::Loop(std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, fuzzer::fuzzer_allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > > > const&) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105b77871)
#13 0x110c04ab8 in fuzzer::FuzzerDriver(int*, char***, int (*)(unsigned char const*, unsigned long)) (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105ba2ab8)
#14 0x110c27179 in main (fuzz_coprocessor_codec_duration_from_parse:x86_64+0x105bc5179)
#15 0x7fff58ec63d4 in start (libdyld.dylib:x86_64+0x163d4)
NOTE: libFuzzer has rudimentary signal handlers.
Combine libFuzzer with AddressSanitizer or similar for better crash reports.
SUMMARY: libFuzzer: deadly signal
MS: 2 ChangeBit-CopyPart-; base unit: 4744b705bce806db21998c6b5c000865f87bf51a
0xff,0x35,0x35,0x35,0x35,0x35,0x35,0x35,0x35,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x20,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x30,0x32,0x33,0x35,0x35,0x35,0x0,0x1,
\xff555555550000000 0000000000023555\x00\x01
artifact_prefix='./'; Test unit written to ./crash-eda52eca653e470cf2fed7ad070b4f09e122dab9
Base64: /zU1NTU1NTU1MDAwMDAwMCAwMDAwMDAwMDAwMDIzNTU1AAE=
Running fuzzer failed: Libfuzzer exited with code Some(77)
Error: 1
```
|
process
|
fuzz test failed for tikv coprocessor codec mysql duration parse bug report what version of tikv are you using beta what operating system and cpu are you using macos what did you do the fuzz test added by wip pr failed this is the test rust pub fn fuzz coprocessor codec duration from parse data result use std io read use tikv coprocessor codec mysql duration let mut cursor cursor new data let fsp cursor read as let mut buf cursor read exact mut buf let d duration parse buf fsp fuzz duration d cursor this is the error report thread panicked at attempt to multiply with overflow src coprocessor codec mysql duration rs stack backtrace std sys unix backtrace tracing imp unwind backtrace std sys common backtrace print std panicking default hook closure std panicking default hook std panicking rust panic with hook std panicking continue panic fmt rust begin unwind core panicking panic fmt core panicking panic tikv coprocessor codec mysql duration duration parse fuzz targets fuzz coprocessor codec duration from parse rust fuzzer test input rust maybe catch panic std panicking try llvmfuzzertestoneinput eeee main error libfuzzer deadly signal in sanitizer print stack trace lib rustc clang rt asan osx dynamic dylib in fuzzer fuzzer crashcallback fuzz coprocessor codec duration from parse in fuzzer fuzzer staticcrashsignalcallback fuzz coprocessor codec duration from parse in fuzzer crashhandler int siginfo void fuzz coprocessor codec duration from parse in sigtramp libsystem platform dylib in rust maybe catch panic fuzz coprocessor codec duration from parse in std panicking try fuzz coprocessor codec duration from parse in llvmfuzzertestoneinput fuzz coprocessor codec duration from parse in fuzzer fuzzer executecallback unsigned char const unsigned long fuzz coprocessor codec duration from parse in fuzzer fuzzer runone unsigned char const unsigned long bool fuzzer inputinfo bool fuzz coprocessor codec duration from parse in fuzzer fuzzer mutateandtestone fuzz coprocessor codec duration from parse in fuzzer fuzzer loop std vector std allocator fuzzer fuzzer allocator std allocator const fuzz coprocessor codec duration from parse in fuzzer fuzzerdriver int char int unsigned char const unsigned long fuzz coprocessor codec duration from parse in main fuzz coprocessor codec duration from parse in start libdyld dylib note libfuzzer has rudimentary signal handlers combine libfuzzer with addresssanitizer or similar for better crash reports summary libfuzzer deadly signal ms changebit copypart base unit artifact prefix test unit written to crash running fuzzer failed libfuzzer exited with code some error
| 1
|
36,251
| 6,521,838,234
|
IssuesEvent
|
2017-08-28 22:22:43
|
NAVADMC/ADSM
|
https://api.github.com/repos/NAVADMC/ADSM
|
closed
|
Fix Unhelpful help text
|
Documentation
|
While working on #245 I've noticed places with unhelpful help text:
- Destruction capacity - The relational function used to define the daily destruction capacity.
- X and Y axis labels would be more useful
- Units detected before triggering vaccination - The number of clinical Units which must be detected before the initiation of a vaccination program.
- I think this prevents any vaccinations from happening before the "vaccination program". Does it start up instantly?
- Vaccination capacity - Relational function used to define the daily vaccination capacity.
- What variable/events does this actually affect in the simulation?
- Test Delay - Function that describes the delay in obtaining test results.
- This is a PDF... so I may be able to deduce the Y axis is a percentage and that this is a distribution of possible values for "delay". But I have to pause and really think about that because the help text says nothing about the "distribution of delays"
- Zone direct movement - Function the describes direct movement rate. (and indirect)
- It's not immediately obvious that this is our "Quarantine" option as "Quarantine" does not appear anywhere in the program. Some thing like "Models imperfect Quarantine as a percentage of normal Direct Movement in days since the Zone was first established."
Is there a "Documentation Person" I can assign these to?
|
1.0
|
Fix Unhelpful help text - While working on #245 I've noticed places with unhelpful help text:
- Destruction capacity - The relational function used to define the daily destruction capacity.
- X and Y axis labels would be more useful
- Units detected before triggering vaccination - The number of clinical Units which must be detected before the initiation of a vaccination program.
- I think this prevents any vaccinations from happening before the "vaccination program". Does it start up instantly?
- Vaccination capacity - Relational function used to define the daily vaccination capacity.
- What variable/events does this actually affect in the simulation?
- Test Delay - Function that describes the delay in obtaining test results.
- This is a PDF... so I may be able to deduce the Y axis is a percentage and that this is a distribution of possible values for "delay". But I have to pause and really think about that because the help text says nothing about the "distribution of delays"
- Zone direct movement - Function the describes direct movement rate. (and indirect)
- It's not immediately obvious that this is our "Quarantine" option as "Quarantine" does not appear anywhere in the program. Some thing like "Models imperfect Quarantine as a percentage of normal Direct Movement in days since the Zone was first established."
Is there a "Documentation Person" I can assign these to?
|
non_process
|
fix unhelpful help text while working on i ve noticed places with unhelpful help text destruction capacity the relational function used to define the daily destruction capacity x and y axis labels would be more useful units detected before triggering vaccination the number of clinical units which must be detected before the initiation of a vaccination program i think this prevents any vaccinations from happening before the vaccination program does it start up instantly vaccination capacity relational function used to define the daily vaccination capacity what variable events does this actually affect in the simulation test delay function that describes the delay in obtaining test results this is a pdf so i may be able to deduce the y axis is a percentage and that this is a distribution of possible values for delay but i have to pause and really think about that because the help text says nothing about the distribution of delays zone direct movement function the describes direct movement rate and indirect it s not immediately obvious that this is our quarantine option as quarantine does not appear anywhere in the program some thing like models imperfect quarantine as a percentage of normal direct movement in days since the zone was first established is there a documentation person i can assign these to
| 0
|
67,672
| 13,012,210,798
|
IssuesEvent
|
2020-07-25 04:00:41
|
jcodwell/MentalHealthRPG
|
https://api.github.com/repos/jcodwell/MentalHealthRPG
|
opened
|
Health Bar Functionalities Implementation
|
New Feature - Code/Unity
|
Code and develop a script that can be used as a player character or enemy health bar.
This code will be complete if the health bar can
- [ ] Move the graphics bar up and down based on value.
- [ ] Be able to attach to a game object (such as a player/NPC or enemy)
- [ ] Be able to interact with the hitbox by going down when collision is detected.
|
1.0
|
Health Bar Functionalities Implementation - Code and develop a script that can be used as a player character or enemy health bar.
This code will be complete if the health bar can
- [ ] Move the graphics bar up and down based on value.
- [ ] Be able to attach to a game object (such as a player/NPC or enemy)
- [ ] Be able to interact with the hitbox by going down when collision is detected.
|
non_process
|
health bar functionalities implementation code and develop a script that can be used as a player character or enemy health bar this code will be complete if the health bar can move the graphics bar up and down based on value be able to attach to a game object such as a player npc or enemy be able to interact with the hitbox by going down when collision is detected
| 0
|
368,252
| 25,784,145,913
|
IssuesEvent
|
2022-12-09 18:39:14
|
vector-im/matrix-bot-sdk
|
https://api.github.com/repos/vector-im/matrix-bot-sdk
|
opened
|
Build & host the web documentation
|
documentation
|
Instead of linking to the documentation pages for the upstream project at https://turt2live.github.io/matrix-bot-sdk/index.html, this fork should have its own documentation hosted on a different domain & with any downstream changes included.
|
1.0
|
Build & host the web documentation - Instead of linking to the documentation pages for the upstream project at https://turt2live.github.io/matrix-bot-sdk/index.html, this fork should have its own documentation hosted on a different domain & with any downstream changes included.
|
non_process
|
build host the web documentation instead of linking to the documentation pages for the upstream project at this fork should have its own documentation hosted on a different domain with any downstream changes included
| 0
|
661,871
| 22,093,062,074
|
IssuesEvent
|
2022-06-01 07:48:53
|
eclipse/dirigible
|
https://api.github.com/repos/eclipse/dirigible
|
opened
|
[API] Copy API should return useful data.
|
enhancement API web-ide priority-medium efforts-medium
|
**Describe the bug**
Currently, the copy API, does not return any useful data. In order to properly update the file tree in the front-end, without reloading the entire structure, we need the new workspace path and git status in the response. The contentType would also be nice to have.
If the copied object is a project, the response should also include the git boolean and gitName fields.
Optionally, it would be good if the workspace name is not included in the file path but is included as another key in the response.
**Desktop:**
- OS: macOS 12.4
- Browser: Firefox 100
- Version: Dirigible 6.2.29
|
1.0
|
[API] Copy API should return useful data. - **Describe the bug**
Currently, the copy API, does not return any useful data. In order to properly update the file tree in the front-end, without reloading the entire structure, we need the new workspace path and git status in the response. The contentType would also be nice to have.
If the copied object is a project, the response should also include the git boolean and gitName fields.
Optionally, it would be good if the workspace name is not included in the file path but is included as another key in the response.
**Desktop:**
- OS: macOS 12.4
- Browser: Firefox 100
- Version: Dirigible 6.2.29
|
non_process
|
copy api should return useful data describe the bug currently the copy api does not return any useful data in order to properly update the file tree in the front end without reloading the entire structure we need the new workspace path and git status in the response the contenttype would also be nice to have if the copied object is a project the response should also include the git boolean and gitname fields optionally it would be good if the workspace name is not included in the file path but is included as another key in the response desktop os macos browser firefox version dirigible
| 0
|
10,988
| 13,785,603,290
|
IssuesEvent
|
2020-10-08 23:19:55
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
TIN creation crash
|
Bug Crash/Data Corruption Feedback Processing
|
Crash ID: 185de4a9a11eb29f05672d8b176e731c0d36ee2f
```
Stack Trace
MathUtils::lineIntersection :
DualEdgeTriangulation::insertForcedSegment :
DualEdgeTriangulation::addLine :
QgsTinInterpolator::insertData :
QgsTinInterpolator::initialize :
QgsTinInterpolator::interpolatePoint :
PyInit__analysis :
QgsGridFileWriter::writeFile :
PyInit__analysis :
PyMethodDef_RawFastCallKeywords :
PyMethodDef_RawFastCallKeywords :
PyEval_EvalFrameDefault :
PyFunction_FastCallDict :
PyMethodDef_RawFastCallDict :
PyObject_Call :
PyInit_sip :
CPLStringList::empty :
PyInit__core :
QgsProcessingAlgorithm::runPrepared :
QgsProcessingAlgRunnerTask::run :
PyInit__core :
QgsTask::start :
QThreadPoolPrivate::reset :
QThread::start :
BaseThreadInitThunk :
RtlUserThreadStart :
```
QGIS Info
QGIS Version: 3.8.0-Zanzibar
QGIS code revision: 11aff65f10
Compiled against Qt: 5.11.2
Running against Qt: 5.11.2
Compiled against GDAL: 2.4.1
Running against GDAL: 2.4.1
System Info
CPU Type: x86_64
Kernel Type: winnt
Kernel Version: 10.0.17763
|
1.0
|
TIN creation crash - Crash ID: 185de4a9a11eb29f05672d8b176e731c0d36ee2f
```
Stack Trace
MathUtils::lineIntersection :
DualEdgeTriangulation::insertForcedSegment :
DualEdgeTriangulation::addLine :
QgsTinInterpolator::insertData :
QgsTinInterpolator::initialize :
QgsTinInterpolator::interpolatePoint :
PyInit__analysis :
QgsGridFileWriter::writeFile :
PyInit__analysis :
PyMethodDef_RawFastCallKeywords :
PyMethodDef_RawFastCallKeywords :
PyEval_EvalFrameDefault :
PyFunction_FastCallDict :
PyMethodDef_RawFastCallDict :
PyObject_Call :
PyInit_sip :
CPLStringList::empty :
PyInit__core :
QgsProcessingAlgorithm::runPrepared :
QgsProcessingAlgRunnerTask::run :
PyInit__core :
QgsTask::start :
QThreadPoolPrivate::reset :
QThread::start :
BaseThreadInitThunk :
RtlUserThreadStart :
```
QGIS Info
QGIS Version: 3.8.0-Zanzibar
QGIS code revision: 11aff65f10
Compiled against Qt: 5.11.2
Running against Qt: 5.11.2
Compiled against GDAL: 2.4.1
Running against GDAL: 2.4.1
System Info
CPU Type: x86_64
Kernel Type: winnt
Kernel Version: 10.0.17763
|
process
|
tin creation crash crash id stack trace mathutils lineintersection dualedgetriangulation insertforcedsegment dualedgetriangulation addline qgstininterpolator insertdata qgstininterpolator initialize qgstininterpolator interpolatepoint pyinit analysis qgsgridfilewriter writefile pyinit analysis pymethoddef rawfastcallkeywords pymethoddef rawfastcallkeywords pyeval evalframedefault pyfunction fastcalldict pymethoddef rawfastcalldict pyobject call pyinit sip cplstringlist empty pyinit core qgsprocessingalgorithm runprepared qgsprocessingalgrunnertask run pyinit core qgstask start qthreadpoolprivate reset qthread start basethreadinitthunk rtluserthreadstart qgis info qgis version zanzibar qgis code revision compiled against qt running against qt compiled against gdal running against gdal system info cpu type kernel type winnt kernel version
| 1
|
32,308
| 8,826,423,247
|
IssuesEvent
|
2019-01-03 02:00:08
|
biojs/biojs-component-builder
|
https://api.github.com/repos/biojs/biojs-component-builder
|
closed
|
[F] Create API endpoint to request bundled components
|
registry/component-builder
|
Should return 200 and the bundled file if it is available or 302 (Found) with a Retry-After header if the component still needs to be build.
|
1.0
|
[F] Create API endpoint to request bundled components - Should return 200 and the bundled file if it is available or 302 (Found) with a Retry-After header if the component still needs to be build.
|
non_process
|
create api endpoint to request bundled components should return and the bundled file if it is available or found with a retry after header if the component still needs to be build
| 0
|
430,187
| 30,120,390,455
|
IssuesEvent
|
2023-06-30 14:43:46
|
ant-media/Ant-Media-Server
|
https://api.github.com/repos/ant-media/Ant-Media-Server
|
opened
|
React Native SDK - Documentation and Blogpost Update
|
documentation/blog-post
|
It is difficult to compile and run the SDK following the existing guide, therefore need to review the document section-wise and make it easy to compile and run the samples for beginners.
#5201
|
1.0
|
React Native SDK - Documentation and Blogpost Update - It is difficult to compile and run the SDK following the existing guide, therefore need to review the document section-wise and make it easy to compile and run the samples for beginners.
#5201
|
non_process
|
react native sdk documentation and blogpost update it is difficult to compile and run the sdk following the existing guide therefore need to review the document section wise and make it easy to compile and run the samples for beginners
| 0
|
8,047
| 11,220,662,924
|
IssuesEvent
|
2020-01-07 16:14:12
|
code4romania/expert-consultation-api
|
https://api.github.com/repos/code4romania/expert-consultation-api
|
closed
|
[Documents] Implement document breakdown logic
|
document processing documents java spring
|
As a user of the Legal Consultation platform, after loading a new document to the platform, I want to be able to see the breakdown of the document. The document needs to be split into logical units of data, based on the structure of the proposition.

|
1.0
|
[Documents] Implement document breakdown logic - As a user of the Legal Consultation platform, after loading a new document to the platform, I want to be able to see the breakdown of the document. The document needs to be split into logical units of data, based on the structure of the proposition.

|
process
|
implement document breakdown logic as a user of the legal consultation platform after loading a new document to the platform i want to be able to see the breakdown of the document the document needs to be split into logical units of data based on the structure of the proposition
| 1
|
44,072
| 2,899,105,672
|
IssuesEvent
|
2015-06-17 09:14:51
|
greenlion/PHP-SQL-Parser
|
https://api.github.com/repos/greenlion/PHP-SQL-Parser
|
closed
|
fake aliases for table names
|
bug imported Priority-Medium
|
_From [pho...@gmx.de](https://code.google.com/u/109317404671582518013/) on January 20, 2012 09:12:19_
Hi,
I try to use your parser within a small project. I need the table names and their aliases within a SELECT statement to check, which columns are used within the ORDER BY clause. The first test produces a problem:
SELECT * FROM answers ORDER BY qid
The parser generates an alias ANSWERS, but the table doesn't have an alias. So it is not possible to find column names within the ORDER clause with the alias (the alias can or cannot be a valid alias).
Please remove the alias from the resulting array, if the table doesn't have an alias within the SQL statement.
Thanks a lot
Andre
_Original issue: http://code.google.com/p/php-sql-parser/issues/detail?id=18_
|
1.0
|
fake aliases for table names - _From [pho...@gmx.de](https://code.google.com/u/109317404671582518013/) on January 20, 2012 09:12:19_
Hi,
I try to use your parser within a small project. I need the table names and their aliases within a SELECT statement to check, which columns are used within the ORDER BY clause. The first test produces a problem:
SELECT * FROM answers ORDER BY qid
The parser generates an alias ANSWERS, but the table doesn't have an alias. So it is not possible to find column names within the ORDER clause with the alias (the alias can or cannot be a valid alias).
Please remove the alias from the resulting array, if the table doesn't have an alias within the SQL statement.
Thanks a lot
Andre
_Original issue: http://code.google.com/p/php-sql-parser/issues/detail?id=18_
|
non_process
|
fake aliases for table names from on january hi i try to use your parser within a small project i need the table names and their aliases within a select statement to check which columns are used within the order by clause the first test produces a problem select from answers order by qid the parser generates an alias answers but the table doesn t have an alias so it is not possible to find column names within the order clause with the alias the alias can or cannot be a valid alias please remove the alias from the resulting array if the table doesn t have an alias within the sql statement thanks a lot andre original issue
| 0
|
15,971
| 20,187,935,196
|
IssuesEvent
|
2022-02-11 00:55:17
|
quark-engine/quark-engine
|
https://api.github.com/repos/quark-engine/quark-engine
|
closed
|
Quark label report does not work any longer UnboundLocalError
|
bug work-in-progress issue-processing-state-04
|
quark with option `-l detailed` or `-l max` crashes with `UnboundLocalError`. This used to work, so probably a _regression_?
```
# quark -a smsthief.apk -l detailed
...
\__> \/ \/ v22.1.1
An Obfuscation-Neglect Android Malware Scoring System
Requested API level 30 is larger than maximum we have, returning API level 28 instead.
100%|██████████████████████████████████████████████████████████████████████████████████████████| 181/181 [00:43<00:00, 4.18it/s]
[*] Total Label found: 34
[*] Rules with label which max confidence >= 80%: 12
Traceback (most recent call last):
File "/usr/local/bin/quark", line 8, in <module>
sys.exit(entry_point())
File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 1128, in __call__
return self.main(*args, **kwargs)
File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 1053, in main
rv = self.invoke(ctx)
File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 1395, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 754, in invoke
return __callback(*args, **kwargs)
File "/usr/local/lib/python3.8/dist-packages/quark/cli.py", line 304, in entry_point
data.show_label_report(rule, all_labels, label)
File "/usr/local/lib/python3.8/dist-packages/quark/core/quark.py", line 650, in show_label_report
self.quark_analysis.label_report_table.field_names = [
File "/usr/local/lib/python3.8/dist-packages/prettytable.py", line 416, in _set_field_names
if self._align and old_names:
UnboundLocalError: local variable 'old_names' referenced before assignment
```
**To Reproduce**
- Please get an APK. I tested with 2 totally different ones, and got the crash in both case, so I believe this does not depend on the APK.
- Run `quark -a your.apk -l detailed` or `quark -a your.apk -l max`
**Desktop (please complete the following information):**
- OS: Docker container from *Ubuntu 20.04* (https://github.com/cryptax/androidre/blob/master/Dockerfile.re)
- Version: `v22.1.1`
- Python `3.8.10`
|
1.0
|
Quark label report does not work any longer UnboundLocalError - quark with option `-l detailed` or `-l max` crashes with `UnboundLocalError`. This used to work, so probably a _regression_?
```
# quark -a smsthief.apk -l detailed
...
\__> \/ \/ v22.1.1
An Obfuscation-Neglect Android Malware Scoring System
Requested API level 30 is larger than maximum we have, returning API level 28 instead.
100%|██████████████████████████████████████████████████████████████████████████████████████████| 181/181 [00:43<00:00, 4.18it/s]
[*] Total Label found: 34
[*] Rules with label which max confidence >= 80%: 12
Traceback (most recent call last):
File "/usr/local/bin/quark", line 8, in <module>
sys.exit(entry_point())
File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 1128, in __call__
return self.main(*args, **kwargs)
File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 1053, in main
rv = self.invoke(ctx)
File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 1395, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/local/lib/python3.8/dist-packages/click/core.py", line 754, in invoke
return __callback(*args, **kwargs)
File "/usr/local/lib/python3.8/dist-packages/quark/cli.py", line 304, in entry_point
data.show_label_report(rule, all_labels, label)
File "/usr/local/lib/python3.8/dist-packages/quark/core/quark.py", line 650, in show_label_report
self.quark_analysis.label_report_table.field_names = [
File "/usr/local/lib/python3.8/dist-packages/prettytable.py", line 416, in _set_field_names
if self._align and old_names:
UnboundLocalError: local variable 'old_names' referenced before assignment
```
**To Reproduce**
- Please get an APK. I tested with 2 totally different ones, and got the crash in both case, so I believe this does not depend on the APK.
- Run `quark -a your.apk -l detailed` or `quark -a your.apk -l max`
**Desktop (please complete the following information):**
- OS: Docker container from *Ubuntu 20.04* (https://github.com/cryptax/androidre/blob/master/Dockerfile.re)
- Version: `v22.1.1`
- Python `3.8.10`
|
process
|
quark label report does not work any longer unboundlocalerror quark with option l detailed or l max crashes with unboundlocalerror this used to work so probably a regression quark a smsthief apk l detailed an obfuscation neglect android malware scoring system requested api level is larger than maximum we have returning api level instead ██████████████████████████████████████████████████████████████████████████████████████████ total label found rules with label which max confidence traceback most recent call last file usr local bin quark line in sys exit entry point file usr local lib dist packages click core py line in call return self main args kwargs file usr local lib dist packages click core py line in main rv self invoke ctx file usr local lib dist packages click core py line in invoke return ctx invoke self callback ctx params file usr local lib dist packages click core py line in invoke return callback args kwargs file usr local lib dist packages quark cli py line in entry point data show label report rule all labels label file usr local lib dist packages quark core quark py line in show label report self quark analysis label report table field names file usr local lib dist packages prettytable py line in set field names if self align and old names unboundlocalerror local variable old names referenced before assignment to reproduce please get an apk i tested with totally different ones and got the crash in both case so i believe this does not depend on the apk run quark a your apk l detailed or quark a your apk l max desktop please complete the following information os docker container from ubuntu version python
| 1
|
140,639
| 21,179,246,556
|
IssuesEvent
|
2022-04-08 05:55:13
|
HSLdevcom/jore4
|
https://api.github.com/repos/HSLdevcom/jore4
|
closed
|
As a public transport planner I want to upload timetables that does not affect compensations to the operator
|
design Timetables compensations
|
In some rare cases the compensation has been agreed in a different way between HSL and the operator, not in the normal basis. In these rare cases, compensation is calculated outside Jore4.
|
1.0
|
As a public transport planner I want to upload timetables that does not affect compensations to the operator - In some rare cases the compensation has been agreed in a different way between HSL and the operator, not in the normal basis. In these rare cases, compensation is calculated outside Jore4.
|
non_process
|
as a public transport planner i want to upload timetables that does not affect compensations to the operator in some rare cases the compensation has been agreed in a different way between hsl and the operator not in the normal basis in these rare cases compensation is calculated outside
| 0
|
9,106
| 12,619,873,591
|
IssuesEvent
|
2020-06-13 03:07:48
|
w-a-r-m-inventory-system/Food-Pantry-Inventory
|
https://api.github.com/repos/w-a-r-m-inventory-system/Food-Pantry-Inventory
|
closed
|
Implement permission levels though out the application
|
document requirement
|
- Only login page does not require the LoginMixIn
- Each page will verify if the user is allowed to access that page
- Throws a polite error message and allow return to previous page if not
- So far, only the Add user page and the change user permission page will have conditional logic based on the permission of the user
- Perhaps this can be isolated to one class or function that contains the necessary logic
|
1.0
|
Implement permission levels though out the application - - Only login page does not require the LoginMixIn
- Each page will verify if the user is allowed to access that page
- Throws a polite error message and allow return to previous page if not
- So far, only the Add user page and the change user permission page will have conditional logic based on the permission of the user
- Perhaps this can be isolated to one class or function that contains the necessary logic
|
non_process
|
implement permission levels though out the application only login page does not require the loginmixin each page will verify if the user is allowed to access that page throws a polite error message and allow return to previous page if not so far only the add user page and the change user permission page will have conditional logic based on the permission of the user perhaps this can be isolated to one class or function that contains the necessary logic
| 0
|
10,668
| 13,456,825,250
|
IssuesEvent
|
2020-09-09 08:23:32
|
jgraley/inferno-cpp2v
|
https://api.github.com/repos/jgraley/inferno-cpp2v
|
opened
|
Partial constraint testing
|
Constraint Processing General Optimisation
|
Add eg `Constraint::PartialTest`. Allow values to be provided for any subset of the constraint's variables.
See about the feasibility of curtailing the conjecture walk when no values remain to be checked (optimisation).
|
1.0
|
Partial constraint testing - Add eg `Constraint::PartialTest`. Allow values to be provided for any subset of the constraint's variables.
See about the feasibility of curtailing the conjecture walk when no values remain to be checked (optimisation).
|
process
|
partial constraint testing add eg constraint partialtest allow values to be provided for any subset of the constraint s variables see about the feasibility of curtailing the conjecture walk when no values remain to be checked optimisation
| 1
|
19,169
| 25,270,038,222
|
IssuesEvent
|
2022-11-16 08:39:30
|
tesseract-ocr/tesseract
|
https://api.github.com/repos/tesseract-ocr/tesseract
|
closed
|
Tesseract hangs
|
process hangs
|
### Environment
* **Tesseract Version**:
tesseract 5.2.0
leptonica-1.82.0
libgif 5.2.1 : libjpeg 8d (libjpeg-turbo 2.1.4) : libpng 1.6.37+apng : libtiff 4.4.0 : zlib 1.2.12 : libwebp 1.2.4
Found OpenMP 201811
Found libarchive 3.6.1 zlib/1.2.12 liblzma/5.2.5 bz2lib/1.0.8 liblz4/1.9.4 libzstd/1.5.2
Found libcurl/7.85.0 OpenSSL/1.1.1o zlib/1.2.12 libpsl/0.21.1 (+libidn2/2.3.3) libssh2/1.10.0 nghttp2/1.48.0
* **Platform**: <!-- either `uname -a` output, or if Windows, version and 32-bit or 64-bit -->
FreeBSD hostname.com 13.1-RELEASE-p3 FreeBSD 13.1-RELEASE-p3 GENERIC amd64
### Current Behavior:
Tesseract hangs without any result. Hanged process even can not be killed.
Commands like ` tesseract anyimage.png stdout` or ` tesseract -l eng+rus anyimage.png stdout`
### Expected Behavior:
Tesseract works (produces result and exits)
|
1.0
|
Tesseract hangs - ### Environment
* **Tesseract Version**:
tesseract 5.2.0
leptonica-1.82.0
libgif 5.2.1 : libjpeg 8d (libjpeg-turbo 2.1.4) : libpng 1.6.37+apng : libtiff 4.4.0 : zlib 1.2.12 : libwebp 1.2.4
Found OpenMP 201811
Found libarchive 3.6.1 zlib/1.2.12 liblzma/5.2.5 bz2lib/1.0.8 liblz4/1.9.4 libzstd/1.5.2
Found libcurl/7.85.0 OpenSSL/1.1.1o zlib/1.2.12 libpsl/0.21.1 (+libidn2/2.3.3) libssh2/1.10.0 nghttp2/1.48.0
* **Platform**: <!-- either `uname -a` output, or if Windows, version and 32-bit or 64-bit -->
FreeBSD hostname.com 13.1-RELEASE-p3 FreeBSD 13.1-RELEASE-p3 GENERIC amd64
### Current Behavior:
Tesseract hangs without any result. Hanged process even can not be killed.
Commands like ` tesseract anyimage.png stdout` or ` tesseract -l eng+rus anyimage.png stdout`
### Expected Behavior:
Tesseract works (produces result and exits)
|
process
|
tesseract hangs environment tesseract version tesseract leptonica libgif libjpeg libjpeg turbo libpng apng libtiff zlib libwebp found openmp found libarchive zlib liblzma libzstd found libcurl openssl zlib libpsl platform freebsd hostname com release freebsd release generic current behavior tesseract hangs without any result hanged process even can not be killed commands like tesseract anyimage png stdout or tesseract l eng rus anyimage png stdout expected behavior tesseract works produces result and exits
| 1
|
521,416
| 15,109,303,081
|
IssuesEvent
|
2021-02-08 17:40:18
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
Minimal libc malloc() is unprotected
|
Regression bug priority: high
|
This is a reminder that PR #31898 must be merged before the 2.5 release.
|
1.0
|
Minimal libc malloc() is unprotected - This is a reminder that PR #31898 must be merged before the 2.5 release.
|
non_process
|
minimal libc malloc is unprotected this is a reminder that pr must be merged before the release
| 0
|
15,717
| 19,849,206,015
|
IssuesEvent
|
2022-01-21 10:21:19
|
ooi-data/RS01SLBS-LJ01A-10-ADCPTE101-streamed-adcp_pd0_beam_parsed
|
https://api.github.com/repos/ooi-data/RS01SLBS-LJ01A-10-ADCPTE101-streamed-adcp_pd0_beam_parsed
|
opened
|
🛑 Processing failed: ValueError
|
process
|
## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T10:21:18.388301.
## Details
Flow name: `RS01SLBS-LJ01A-10-ADCPTE101-streamed-adcp_pd0_beam_parsed`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
1.0
|
🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T10:21:18.388301.
## Details
Flow name: `RS01SLBS-LJ01A-10-ADCPTE101-streamed-adcp_pd0_beam_parsed`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
process
|
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name streamed adcp beam parsed task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
| 1
|
328,944
| 10,007,297,738
|
IssuesEvent
|
2019-07-14 09:29:39
|
gama-platform/gama
|
https://api.github.com/repos/gama-platform/gama
|
closed
|
Exception in an aspect with two displays can freeze GAMA when opening the displays
|
> Bug Affects Usability Concerns Interface Display All Priority High Version Git
|
**Describe the bug**
Let's consider an experiment with 2 displays and that the first display displays a species with an Exception thrown in the aspect.
Let's now consider that the Preference allowing to show the Exception thrown in displays is activated.
Then when I run such a model, GAMA freezes and no Exception appears in GAMA. Some exceptions are displayed in the Eclipse console.
Notes that :
- when there is only one display with the Exception inside
The exception is displayed in GAMA.
- when the Preference is set to not show the exceptions in the displays
then the freeze does not occur
**To Reproduce**
Run the following model:
```
model NewModel
global {
init {
create toot number: 10;
write "maître";
}
}
species toot {
list<int> l <- [1,2];
aspect c {
draw circle(10);
}
aspect default {
loop i from: 0 to: length(l) {
draw circle(l[i]);
}
}
}
experiment name type: gui {
output {
display d2 {
species toot;
}
display d {
species toot aspect: c;
}
}
}
```
**Desktop (please complete the following information):**
- OS: macOSX
- GAMA version: git
|
1.0
|
Exception in an aspect with two displays can freeze GAMA when opening the displays - **Describe the bug**
Let's consider an experiment with 2 displays and that the first display displays a species with an Exception thrown in the aspect.
Let's now consider that the Preference allowing to show the Exception thrown in displays is activated.
Then when I run such a model, GAMA freezes and no Exception appears in GAMA. Some exceptions are displayed in the Eclipse console.
Notes that :
- when there is only one display with the Exception inside
The exception is displayed in GAMA.
- when the Preference is set to not show the exceptions in the displays
then the freeze does not occur
**To Reproduce**
Run the following model:
```
model NewModel
global {
init {
create toot number: 10;
write "maître";
}
}
species toot {
list<int> l <- [1,2];
aspect c {
draw circle(10);
}
aspect default {
loop i from: 0 to: length(l) {
draw circle(l[i]);
}
}
}
experiment name type: gui {
output {
display d2 {
species toot;
}
display d {
species toot aspect: c;
}
}
}
```
**Desktop (please complete the following information):**
- OS: macOSX
- GAMA version: git
|
non_process
|
exception in an aspect with two displays can freeze gama when opening the displays describe the bug let s consider an experiment with displays and that the first display displays a species with an exception thrown in the aspect let s now consider that the preference allowing to show the exception thrown in displays is activated then when i run such a model gama freezes and no exception appears in gama some exceptions are displayed in the eclipse console notes that when there is only one display with the exception inside the exception is displayed in gama when the preference is set to not show the exceptions in the displays then the freeze does not occur to reproduce run the following model model newmodel global init create toot number write maître species toot list l aspect c draw circle aspect default loop i from to length l draw circle l experiment name type gui output display species toot display d species toot aspect c desktop please complete the following information os macosx gama version git
| 0
|
15,632
| 19,783,392,957
|
IssuesEvent
|
2022-01-18 01:39:05
|
linuxdeepin/developer-center
|
https://api.github.com/repos/linuxdeepin/developer-center
|
closed
|
Validity Sensors VFS491 without controllers
|
other | delay processing
|
the VFS491 validation sensor, used for fingerprint scanning, has no drivers in linux deepin.
for this sensor only proprietary drivers are available https://ftp.hp.com/pub/softpaq/sp84501-85000/sp84530.tar https://github.com/rindeal/libfprint-vfs_proprietary-driver
these drivers work for various models of fingerprint readers used in computers.
|
1.0
|
Validity Sensors VFS491 without controllers - the VFS491 validation sensor, used for fingerprint scanning, has no drivers in linux deepin.
for this sensor only proprietary drivers are available https://ftp.hp.com/pub/softpaq/sp84501-85000/sp84530.tar https://github.com/rindeal/libfprint-vfs_proprietary-driver
these drivers work for various models of fingerprint readers used in computers.
|
process
|
validity sensors without controllers the validation sensor used for fingerprint scanning has no drivers in linux deepin for this sensor only proprietary drivers are available these drivers work for various models of fingerprint readers used in computers
| 1
|
437,867
| 12,603,622,331
|
IssuesEvent
|
2020-06-11 13:46:27
|
cilium/cilium
|
https://api.github.com/repos/cilium/cilium
|
closed
|
CI: provide agent logs when cilium pods are in crash loopback
|
area/CI priority/high project/ci-force
|
When agent's pod goes into crash loopback and therefore CI results into failure, it would be helpful to see why that is happening. Right now the output is not useful, e.g. the zip file doesn't collect the agent log. [Example dump](https://github.com/cilium/cilium/files/4576627/f3ee8175_K8sChaosTest_Connectivity_demo_application_Endpoint_can_still_connect_while_Cilium_is_not_running.1.zip) from such CI run where the pod went into this state when init.sh was failing due to compilation error. The issue does not happen when the CI is run locally.
|
1.0
|
CI: provide agent logs when cilium pods are in crash loopback - When agent's pod goes into crash loopback and therefore CI results into failure, it would be helpful to see why that is happening. Right now the output is not useful, e.g. the zip file doesn't collect the agent log. [Example dump](https://github.com/cilium/cilium/files/4576627/f3ee8175_K8sChaosTest_Connectivity_demo_application_Endpoint_can_still_connect_while_Cilium_is_not_running.1.zip) from such CI run where the pod went into this state when init.sh was failing due to compilation error. The issue does not happen when the CI is run locally.
|
non_process
|
ci provide agent logs when cilium pods are in crash loopback when agent s pod goes into crash loopback and therefore ci results into failure it would be helpful to see why that is happening right now the output is not useful e g the zip file doesn t collect the agent log from such ci run where the pod went into this state when init sh was failing due to compilation error the issue does not happen when the ci is run locally
| 0
|
10,795
| 6,922,324,037
|
IssuesEvent
|
2017-11-30 02:30:26
|
dryuen/ist5313-final-project-template
|
https://api.github.com/repos/dryuen/ist5313-final-project-template
|
closed
|
Group 1- Module 3- Hotspot
|
developer enhancement f17-group1 other usability
|
Hotspot should include example to help the user identify the steps.
|
True
|
Group 1- Module 3- Hotspot - Hotspot should include example to help the user identify the steps.
|
non_process
|
group module hotspot hotspot should include example to help the user identify the steps
| 0
|
12,994
| 15,358,513,454
|
IssuesEvent
|
2021-03-01 14:53:14
|
edwardsmarc/CASFRI
|
https://api.github.com/repos/edwardsmarc/CASFRI
|
opened
|
TT_ProduceInvGeoHistory('SK01') fails on one row
|
bug high post-translation process
|
ERROR: TT_PolygonGeoHistory() ERROR: TT_SafeDifference() failed on SK01-xxxxxxxxxxxxUTM-xxxxxxxxxx-1348605004-0343222...
|
1.0
|
TT_ProduceInvGeoHistory('SK01') fails on one row - ERROR: TT_PolygonGeoHistory() ERROR: TT_SafeDifference() failed on SK01-xxxxxxxxxxxxUTM-xxxxxxxxxx-1348605004-0343222...
|
process
|
tt produceinvgeohistory fails on one row error tt polygongeohistory error tt safedifference failed on xxxxxxxxxxxxutm xxxxxxxxxx
| 1
|
70,171
| 15,053,124,142
|
IssuesEvent
|
2021-02-03 15:58:06
|
istio/istio
|
https://api.github.com/repos/istio/istio
|
closed
|
Log mTLS permissive warning only once
|
area/networking/envoy area/security kind/enhancement
|
```
2020-11-03T18:17:23.189271Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/
2020-11-03T18:17:23.190018Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/
2020-11-03T18:17:23.192851Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/
2020-11-03T18:17:23.193503Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/
```
log is very spammy. Should log it once for the lifetime of a pod
|
True
|
Log mTLS permissive warning only once - ```
2020-11-03T18:17:23.189271Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/
2020-11-03T18:17:23.190018Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/
2020-11-03T18:17:23.192851Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/
2020-11-03T18:17:23.193503Z warning envoy filter mTLS PERMISSIVE mode is used, connection can be either plaintext or TLS, and client cert can be omitted. Please consider to upgrade to mTLS STRICT mode for more secure configuration that only allows TLS connection with client cert. See https://istio.io/docs/tasks/security/mtls-migration/
```
log is very spammy. Should log it once for the lifetime of a pod
|
non_process
|
log mtls permissive warning only once warning envoy filter mtls permissive mode is used connection can be either plaintext or tls and client cert can be omitted please consider to upgrade to mtls strict mode for more secure configuration that only allows tls connection with client cert see warning envoy filter mtls permissive mode is used connection can be either plaintext or tls and client cert can be omitted please consider to upgrade to mtls strict mode for more secure configuration that only allows tls connection with client cert see warning envoy filter mtls permissive mode is used connection can be either plaintext or tls and client cert can be omitted please consider to upgrade to mtls strict mode for more secure configuration that only allows tls connection with client cert see warning envoy filter mtls permissive mode is used connection can be either plaintext or tls and client cert can be omitted please consider to upgrade to mtls strict mode for more secure configuration that only allows tls connection with client cert see log is very spammy should log it once for the lifetime of a pod
| 0
|
160,629
| 25,198,998,867
|
IssuesEvent
|
2022-11-12 22:14:16
|
authelia/authelia
|
https://api.github.com/repos/authelia/authelia
|
opened
|
Design: Support for multiple WebAuthn devices for each user
|
priority/4/normal type/feature status/needs-design
|
### Description
Desired user scenarios:
* Users can add more than one WebAuthn device to their account ([listed on Authelia's WebAuthn roadmap](https://www.authelia.com/roadmap/active/webauthn/#multi-device-registration))
* Registered device name can be provided by the user during device registration
* Users can manage their existing WebAuthn devices within their account ([listed on Authelia's Dashboard / Control Panel roadmap](https://www.authelia.com/roadmap/active/dashboard-control-panel/#initial-implementation))
* View information about registered device(s)
* Rename existing device(s)
* Add additional device(s)
* Delete existing device(s)
### Use Case
It's common and recommended for users with hardware security tokens (ex. Yubikeys) to have more than one. As an example, I have a backup Yubikey in case my main key gets lost. Each of my Yubikeys are configured in my accounts where I use them for 2-factor authentication. Hardware security tokens are usable via the WebAuthn protocol, which Authelia already uses and supports.
Sites which support hardware security tokens for 2-factor authentication generally support this use case. Examples include Google, Dropbox, and GitHub.
This feature was requested in https://github.com/authelia/authelia/issues/275.
### Details
For reference, this is the current WebAuthn device registration flow:
1. On the login frame (at either the `OneFactor` or `TwoFactor` authentication levels), the user clicks the "Lost your device?" link

2. Authelia generates a token, encoded as JWT, and sends the user an email containing a link with the JWT token
3. User clicks the link in their email
4. Authelia retrieves the token from the URI, decodes and verifies the JWT, and [performs the WebAuthn attestation ceremony](https://github.com/authelia/authelia/blob/c7a668c897cbda49d858a19e40705dca88e1ca3e/web/src/views/DeviceRegistration/RegisterWebauthn.tsx#L35).
5. On successful WebAuthn device attestation, the new device info is stored in the Authelia database. [Currently the device registration is hardcoded with a description of `Primary`](https://github.com/authelia/authelia/blob/c7a668c897cbda49d858a19e40705dca88e1ca3e/internal/handlers/handler_register_webauthn.go#L137). If an existing device was registered, it is overwritten.
6. On WebAuthn device registration success, the user is redirected back to the login page (which then re-prompts for 2FA authentication with the new device).
#### Proposed changes: Users can add more than one WebAuthn device to their account
* As mentioned on Authelia's [WebAuthn roadmap](https://www.authelia.com/roadmap/active/webauthn/#multi-device-registration), the storage backend has existing support for multiple devices. Devices are overwritten on re-registration simply due to the hardcoded name (referred to in the backend as "description") `Primary` which is used for all devices. The hardcoded device name should be removed so it can be provided to the backend via the API as in https://github.com/authelia/authelia/pull/4363.
* Support entry of a user-provided device name as part of the WebAuthn device registration flow. This can then be provided to the backend on registration via the previously mentioned API change.
* The user should be prevented from reusing the name of existing device within their account by both the UI and backend
#### Proposed changes: Users can manage their existing WebAuthn devices within their account
* New UI is needed for device management. From the roadmap and discussions on Discord, the desired feature is an extensible user account settings UI. Although the UI may only support WebAuthn device management initially, this UI would later provide options for management of other 2FA options and other unrelated settings (such as Authelia service configuration options, perhaps).
* Some existing features work around a settings UI is present in the [`feat-settings-ui` branch](https://github.com/authelia/authelia/tree/feat-settings-ui). The UI in this branch currently looks like this:

* WebAuthn settings UI should display information about registered devices:
* User-entered device name ("description" in the backend)
* Additional details such as registered date, and last used date. These values are already stored in the database.
* No API endpoint currently exists for retrieving configured WebAuthn device information for a user. A new API endpoint is needed. (A new endpoint is present in the [`feat-settings-ui` branch](https://github.com/authelia/authelia/tree/feat-settings-ui).)
* WebAuthn settings UI should allow registration of multiple devices. Some options:
1. Reuse existing registration flow. Without additional changes, re-launching the flow will send the user another email with a link+token. Opening this page will start the WebAuthn device attestation/registration flow. Pros: Smallest code change. Cons: Cumbersome for the user, as it requires an email round-trip for each new added device.
2. Reuse existing registration flow, without requiring email verification when already logged in at the `TwoFactor` level. Pros: Still a smaller code change. Less cumbersome for the user. No expected security impact as the user already had to verify their email to register/log in with 2FA in the first place. Cons: Redirection to existing flow may be jarring for users.
3. Reuse existing registration flow in a popup modal over the new settings UI, without requiring email verification when already logged in at the `TwoFactor` level. Pros: Most intuitive flow for the user. Modal popup makes it more obvious the registration function operates within the settings UI space. No expected security impact as the user already had to verify their email to register/log in with 2FA in the first place. Cons: Largest code change.
* **Note**: Skipping email verification when logged in at the `TwoFactor` level could be accomplished by simply succeeding the [identity verification middleware](https://github.com/authelia/authelia/blob/c7a668c897cbda49d858a19e40705dca88e1ca3e/internal/middlewares/identity_verification.go) if the user's auth level is `TwoFactor`. I sketched this out here: https://github.com/smkent/authelia/pull/1
* WebAuthn settings UI should allow rename of multiple devices.
* A new API endpoint is needed for modifying a WebAuthn device.
* WebAuthn settings UI should allow deletion of multiple devices.
* A new API endpoint is needed for deleting a WebAuthn device.
* When all WebAuthn devices are deleted, my preferred option is to:
1. Do nothing. The user will remain logged in at the `TwoFactor` level within their current session. Logging in again will automatically restart the 2FA device registration prompt if no other 2FA methods are enabled. Pros: Less disruptive flow for the user. User can delete their last device and (re-)add a device in the settings UI without their session ending.
#### Additional considerations
* Email-based identity verification may be reworked via https://github.com/authelia/authelia/issues/3801, which would replace the JWT token with a randomly generated one time password. This may affect some of the proposed feature implementation above.
### Documentation
_No response_
|
1.0
|
Design: Support for multiple WebAuthn devices for each user - ### Description
Desired user scenarios:
* Users can add more than one WebAuthn device to their account ([listed on Authelia's WebAuthn roadmap](https://www.authelia.com/roadmap/active/webauthn/#multi-device-registration))
* Registered device name can be provided by the user during device registration
* Users can manage their existing WebAuthn devices within their account ([listed on Authelia's Dashboard / Control Panel roadmap](https://www.authelia.com/roadmap/active/dashboard-control-panel/#initial-implementation))
* View information about registered device(s)
* Rename existing device(s)
* Add additional device(s)
* Delete existing device(s)
### Use Case
It's common and recommended for users with hardware security tokens (ex. Yubikeys) to have more than one. As an example, I have a backup Yubikey in case my main key gets lost. Each of my Yubikeys are configured in my accounts where I use them for 2-factor authentication. Hardware security tokens are usable via the WebAuthn protocol, which Authelia already uses and supports.
Sites which support hardware security tokens for 2-factor authentication generally support this use case. Examples include Google, Dropbox, and GitHub.
This feature was requested in https://github.com/authelia/authelia/issues/275.
### Details
For reference, this is the current WebAuthn device registration flow:
1. On the login frame (at either the `OneFactor` or `TwoFactor` authentication levels), the user clicks the "Lost your device?" link

2. Authelia generates a token, encoded as JWT, and sends the user an email containing a link with the JWT token
3. User clicks the link in their email
4. Authelia retrieves the token from the URI, decodes and verifies the JWT, and [performs the WebAuthn attestation ceremony](https://github.com/authelia/authelia/blob/c7a668c897cbda49d858a19e40705dca88e1ca3e/web/src/views/DeviceRegistration/RegisterWebauthn.tsx#L35).
5. On successful WebAuthn device attestation, the new device info is stored in the Authelia database. [Currently the device registration is hardcoded with a description of `Primary`](https://github.com/authelia/authelia/blob/c7a668c897cbda49d858a19e40705dca88e1ca3e/internal/handlers/handler_register_webauthn.go#L137). If an existing device was registered, it is overwritten.
6. On WebAuthn device registration success, the user is redirected back to the login page (which then re-prompts for 2FA authentication with the new device).
#### Proposed changes: Users can add more than one WebAuthn device to their account
* As mentioned on Authelia's [WebAuthn roadmap](https://www.authelia.com/roadmap/active/webauthn/#multi-device-registration), the storage backend has existing support for multiple devices. Devices are overwritten on re-registration simply due to the hardcoded name (referred to in the backend as "description") `Primary` which is used for all devices. The hardcoded device name should be removed so it can be provided to the backend via the API as in https://github.com/authelia/authelia/pull/4363.
* Support entry of a user-provided device name as part of the WebAuthn device registration flow. This can then be provided to the backend on registration via the previously mentioned API change.
* The user should be prevented from reusing the name of existing device within their account by both the UI and backend
#### Proposed changes: Users can manage their existing WebAuthn devices within their account
* New UI is needed for device management. From the roadmap and discussions on Discord, the desired feature is an extensible user account settings UI. Although the UI may only support WebAuthn device management initially, this UI would later provide options for management of other 2FA options and other unrelated settings (such as Authelia service configuration options, perhaps).
* Some existing features work around a settings UI is present in the [`feat-settings-ui` branch](https://github.com/authelia/authelia/tree/feat-settings-ui). The UI in this branch currently looks like this:

* WebAuthn settings UI should display information about registered devices:
* User-entered device name ("description" in the backend)
* Additional details such as registered date, and last used date. These values are already stored in the database.
* No API endpoint currently exists for retrieving configured WebAuthn device information for a user. A new API endpoint is needed. (A new endpoint is present in the [`feat-settings-ui` branch](https://github.com/authelia/authelia/tree/feat-settings-ui).)
* WebAuthn settings UI should allow registration of multiple devices. Some options:
1. Reuse existing registration flow. Without additional changes, re-launching the flow will send the user another email with a link+token. Opening this page will start the WebAuthn device attestation/registration flow. Pros: Smallest code change. Cons: Cumbersome for the user, as it requires an email round-trip for each new added device.
2. Reuse existing registration flow, without requiring email verification when already logged in at the `TwoFactor` level. Pros: Still a smaller code change. Less cumbersome for the user. No expected security impact as the user already had to verify their email to register/log in with 2FA in the first place. Cons: Redirection to existing flow may be jarring for users.
3. Reuse existing registration flow in a popup modal over the new settings UI, without requiring email verification when already logged in at the `TwoFactor` level. Pros: Most intuitive flow for the user. Modal popup makes it more obvious the registration function operates within the settings UI space. No expected security impact as the user already had to verify their email to register/log in with 2FA in the first place. Cons: Largest code change.
* **Note**: Skipping email verification when logged in at the `TwoFactor` level could be accomplished by simply succeeding the [identity verification middleware](https://github.com/authelia/authelia/blob/c7a668c897cbda49d858a19e40705dca88e1ca3e/internal/middlewares/identity_verification.go) if the user's auth level is `TwoFactor`. I sketched this out here: https://github.com/smkent/authelia/pull/1
* WebAuthn settings UI should allow rename of multiple devices.
* A new API endpoint is needed for modifying a WebAuthn device.
* WebAuthn settings UI should allow deletion of multiple devices.
* A new API endpoint is needed for deleting a WebAuthn device.
* When all WebAuthn devices are deleted, my preferred option is to:
1. Do nothing. The user will remain logged in at the `TwoFactor` level within their current session. Logging in again will automatically restart the 2FA device registration prompt if no other 2FA methods are enabled. Pros: Less disruptive flow for the user. User can delete their last device and (re-)add a device in the settings UI without their session ending.
#### Additional considerations
* Email-based identity verification may be reworked via https://github.com/authelia/authelia/issues/3801, which would replace the JWT token with a randomly generated one time password. This may affect some of the proposed feature implementation above.
### Documentation
_No response_
|
non_process
|
design support for multiple webauthn devices for each user description desired user scenarios users can add more than one webauthn device to their account registered device name can be provided by the user during device registration users can manage their existing webauthn devices within their account view information about registered device s rename existing device s add additional device s delete existing device s use case it s common and recommended for users with hardware security tokens ex yubikeys to have more than one as an example i have a backup yubikey in case my main key gets lost each of my yubikeys are configured in my accounts where i use them for factor authentication hardware security tokens are usable via the webauthn protocol which authelia already uses and supports sites which support hardware security tokens for factor authentication generally support this use case examples include google dropbox and github this feature was requested in details for reference this is the current webauthn device registration flow on the login frame at either the onefactor or twofactor authentication levels the user clicks the lost your device link authelia generates a token encoded as jwt and sends the user an email containing a link with the jwt token user clicks the link in their email authelia retrieves the token from the uri decodes and verifies the jwt and on successful webauthn device attestation the new device info is stored in the authelia database if an existing device was registered it is overwritten on webauthn device registration success the user is redirected back to the login page which then re prompts for authentication with the new device proposed changes users can add more than one webauthn device to their account as mentioned on authelia s the storage backend has existing support for multiple devices devices are overwritten on re registration simply due to the hardcoded name referred to in the backend as description primary which is used for all devices the hardcoded device name should be removed so it can be provided to the backend via the api as in support entry of a user provided device name as part of the webauthn device registration flow this can then be provided to the backend on registration via the previously mentioned api change the user should be prevented from reusing the name of existing device within their account by both the ui and backend proposed changes users can manage their existing webauthn devices within their account new ui is needed for device management from the roadmap and discussions on discord the desired feature is an extensible user account settings ui although the ui may only support webauthn device management initially this ui would later provide options for management of other options and other unrelated settings such as authelia service configuration options perhaps some existing features work around a settings ui is present in the the ui in this branch currently looks like this webauthn settings ui should display information about registered devices user entered device name description in the backend additional details such as registered date and last used date these values are already stored in the database no api endpoint currently exists for retrieving configured webauthn device information for a user a new api endpoint is needed a new endpoint is present in the webauthn settings ui should allow registration of multiple devices some options reuse existing registration flow without additional changes re launching the flow will send the user another email with a link token opening this page will start the webauthn device attestation registration flow pros smallest code change cons cumbersome for the user as it requires an email round trip for each new added device reuse existing registration flow without requiring email verification when already logged in at the twofactor level pros still a smaller code change less cumbersome for the user no expected security impact as the user already had to verify their email to register log in with in the first place cons redirection to existing flow may be jarring for users reuse existing registration flow in a popup modal over the new settings ui without requiring email verification when already logged in at the twofactor level pros most intuitive flow for the user modal popup makes it more obvious the registration function operates within the settings ui space no expected security impact as the user already had to verify their email to register log in with in the first place cons largest code change note skipping email verification when logged in at the twofactor level could be accomplished by simply succeeding the if the user s auth level is twofactor i sketched this out here webauthn settings ui should allow rename of multiple devices a new api endpoint is needed for modifying a webauthn device webauthn settings ui should allow deletion of multiple devices a new api endpoint is needed for deleting a webauthn device when all webauthn devices are deleted my preferred option is to do nothing the user will remain logged in at the twofactor level within their current session logging in again will automatically restart the device registration prompt if no other methods are enabled pros less disruptive flow for the user user can delete their last device and re add a device in the settings ui without their session ending additional considerations email based identity verification may be reworked via which would replace the jwt token with a randomly generated one time password this may affect some of the proposed feature implementation above documentation no response
| 0
|
755,211
| 26,421,286,243
|
IssuesEvent
|
2023-01-13 20:47:01
|
DSpace/DSpace
|
https://api.github.com/repos/DSpace/DSpace
|
opened
|
Request a Copy's "helpdesk" strategy requires authentication to respond
|
bug authorization medium priority help wanted
|
**Describe the bug**
Discovered while testing #8319
In DSpace 6, selecting the "helpdesk" strategy for Request a Copy allowed anyone receiving the helpdesk email to respond to the request _without authenticating into DSpace._ This is currently not possible in DSpace 7.
**To Reproduce**
Steps to reproduce the behavior:
1. Activate request a copy with Helpdesk Strategy: https://wiki.lyrasis.org/display/DSDOC7x/Request+a+Copy#RequestaCopy-(Optional)RequestingacopywithHelpDeskworkflow
2. As anonymous user, look at an embargoed item, click on the file to display the request a copy form.
3. Complete and send the request a copy form
4. As helpdesk staff, look at the email inbox
5. Open the Request a copy email and click on the link to respond to it.
6. **You are now required to login to DSpace** in order to respond. Previously this was available anonymously.
7. After logging in, you can approve or decline the request.
**Expected behavior**
You should be able to approve or deny the request without authenticating. The link sent to the helpdesk staff includes a unique token which is all that is necessary to respond to the request.
**Related work**
Related to #8319 and originally reported in #8304
|
1.0
|
Request a Copy's "helpdesk" strategy requires authentication to respond - **Describe the bug**
Discovered while testing #8319
In DSpace 6, selecting the "helpdesk" strategy for Request a Copy allowed anyone receiving the helpdesk email to respond to the request _without authenticating into DSpace._ This is currently not possible in DSpace 7.
**To Reproduce**
Steps to reproduce the behavior:
1. Activate request a copy with Helpdesk Strategy: https://wiki.lyrasis.org/display/DSDOC7x/Request+a+Copy#RequestaCopy-(Optional)RequestingacopywithHelpDeskworkflow
2. As anonymous user, look at an embargoed item, click on the file to display the request a copy form.
3. Complete and send the request a copy form
4. As helpdesk staff, look at the email inbox
5. Open the Request a copy email and click on the link to respond to it.
6. **You are now required to login to DSpace** in order to respond. Previously this was available anonymously.
7. After logging in, you can approve or decline the request.
**Expected behavior**
You should be able to approve or deny the request without authenticating. The link sent to the helpdesk staff includes a unique token which is all that is necessary to respond to the request.
**Related work**
Related to #8319 and originally reported in #8304
|
non_process
|
request a copy s helpdesk strategy requires authentication to respond describe the bug discovered while testing in dspace selecting the helpdesk strategy for request a copy allowed anyone receiving the helpdesk email to respond to the request without authenticating into dspace this is currently not possible in dspace to reproduce steps to reproduce the behavior activate request a copy with helpdesk strategy as anonymous user look at an embargoed item click on the file to display the request a copy form complete and send the request a copy form as helpdesk staff look at the email inbox open the request a copy email and click on the link to respond to it you are now required to login to dspace in order to respond previously this was available anonymously after logging in you can approve or decline the request expected behavior you should be able to approve or deny the request without authenticating the link sent to the helpdesk staff includes a unique token which is all that is necessary to respond to the request related work related to and originally reported in
| 0
|
322,680
| 9,821,524,880
|
IssuesEvent
|
2019-06-14 07:28:50
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
start.parimatch.com - see bug description
|
browser-firefox engine-gecko priority-normal
|
<!-- @browser: Firefox 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:68.0) Gecko/20100101 Firefox/68.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://start.parimatch.com/az/fight/0/5/?btag=a_16060b_1047c_375&siteid=16060
**Browser / Version**: Firefox 68.0
**Operating System**: Windows 7
**Tested Another Browser**: Unknown
**Problem type**: Something else
**Description**: site contains virus
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/6/a6a0aa41-a2d3-4ac4-94e0-5864ceffa5ac.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190610153228</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
start.parimatch.com - see bug description - <!-- @browser: Firefox 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:68.0) Gecko/20100101 Firefox/68.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://start.parimatch.com/az/fight/0/5/?btag=a_16060b_1047c_375&siteid=16060
**Browser / Version**: Firefox 68.0
**Operating System**: Windows 7
**Tested Another Browser**: Unknown
**Problem type**: Something else
**Description**: site contains virus
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/6/a6a0aa41-a2d3-4ac4-94e0-5864ceffa5ac.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190610153228</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
start parimatch com see bug description url browser version firefox operating system windows tested another browser unknown problem type something else description site contains virus steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta from with ❤️
| 0
|
5,664
| 8,531,628,714
|
IssuesEvent
|
2018-11-04 14:14:40
|
aiidateam/aiida_core
|
https://api.github.com/repos/aiidateam/aiida_core
|
opened
|
Add optional TransporTask for JobProcess to copy over files
|
priority/nice to have requires discussion topic/Engine topic/JobCalculationAndProcess type/feature request
|
A common use case, is that after a `JobProcess` has finished, one might want to copy some files from the work directory, which is often on a scratch space, to a more durable location and attach them through a `RemoteFolder` to the `JobCalculation` node. This is different from the files that are already retrieved for parsing, as often these additional files are big, and one wants to keep them on the remote computer. A typical example are big files that are necessary to restart a calculation from, that are too big to retrieve but also too costly to lose and have to recompute. There is a simple and naive solution of just running an inline calculation that will open transport to the remote computer and use `copy_from_remote_to_remote` to copy the files and create a `RemoteFolder` object to attach to the computer, but this can be dangerous. If an exception occurs, the entire process, most likely the `WorkChain` that launched the `JobCalculation` will except as well. Of course one can catch the exceptions but then the files will not be copied. The `JobProcess` infrastructure has a rich architecture to deal with tasks that require transport and can fail. The best option would be to leverage this functionality for this particular purpose by creating an optional `TransportTask` to copy files after the job has finished, but before going over to the retrieval and parse step.
|
1.0
|
Add optional TransporTask for JobProcess to copy over files - A common use case, is that after a `JobProcess` has finished, one might want to copy some files from the work directory, which is often on a scratch space, to a more durable location and attach them through a `RemoteFolder` to the `JobCalculation` node. This is different from the files that are already retrieved for parsing, as often these additional files are big, and one wants to keep them on the remote computer. A typical example are big files that are necessary to restart a calculation from, that are too big to retrieve but also too costly to lose and have to recompute. There is a simple and naive solution of just running an inline calculation that will open transport to the remote computer and use `copy_from_remote_to_remote` to copy the files and create a `RemoteFolder` object to attach to the computer, but this can be dangerous. If an exception occurs, the entire process, most likely the `WorkChain` that launched the `JobCalculation` will except as well. Of course one can catch the exceptions but then the files will not be copied. The `JobProcess` infrastructure has a rich architecture to deal with tasks that require transport and can fail. The best option would be to leverage this functionality for this particular purpose by creating an optional `TransportTask` to copy files after the job has finished, but before going over to the retrieval and parse step.
|
process
|
add optional transportask for jobprocess to copy over files a common use case is that after a jobprocess has finished one might want to copy some files from the work directory which is often on a scratch space to a more durable location and attach them through a remotefolder to the jobcalculation node this is different from the files that are already retrieved for parsing as often these additional files are big and one wants to keep them on the remote computer a typical example are big files that are necessary to restart a calculation from that are too big to retrieve but also too costly to lose and have to recompute there is a simple and naive solution of just running an inline calculation that will open transport to the remote computer and use copy from remote to remote to copy the files and create a remotefolder object to attach to the computer but this can be dangerous if an exception occurs the entire process most likely the workchain that launched the jobcalculation will except as well of course one can catch the exceptions but then the files will not be copied the jobprocess infrastructure has a rich architecture to deal with tasks that require transport and can fail the best option would be to leverage this functionality for this particular purpose by creating an optional transporttask to copy files after the job has finished but before going over to the retrieval and parse step
| 1
|
6,086
| 8,948,425,063
|
IssuesEvent
|
2019-01-25 02:19:44
|
bisq-network/bisq
|
https://api.github.com/repos/bisq-network/bisq
|
closed
|
Add new P2P message in trade process to report errors to the peer
|
in:trade-process was:dropped
|
Related to https://github.com/bisq-network/exchange/issues/1044 it would help to find out the cause of errors if the peer sends a P2P message to the trade peer in the trade process so he get informed why the peer stopped responding. Currently if an error occurs the peers stops the protocol but do not send any message to the peer.
|
1.0
|
Add new P2P message in trade process to report errors to the peer - Related to https://github.com/bisq-network/exchange/issues/1044 it would help to find out the cause of errors if the peer sends a P2P message to the trade peer in the trade process so he get informed why the peer stopped responding. Currently if an error occurs the peers stops the protocol but do not send any message to the peer.
|
process
|
add new message in trade process to report errors to the peer related to it would help to find out the cause of errors if the peer sends a message to the trade peer in the trade process so he get informed why the peer stopped responding currently if an error occurs the peers stops the protocol but do not send any message to the peer
| 1
|
77,383
| 21,783,728,326
|
IssuesEvent
|
2022-05-13 22:29:19
|
dotnet/arcade
|
https://api.github.com/repos/dotnet/arcade
|
closed
|
Build failed: dotnet-arcade-validation-official/main #20220511.6
|
Build Failed
|
Build [#20220511.6](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=1766460) partiallySucceeded
## :warning: : internal / dotnet-arcade-validation-official partiallySucceeded
### Summary
**Finished** - Thu, 12 May 2022 17:08:41 GMT
**Duration** - 3 minutes
**Requested for** - DotNet Bot
**Reason** - batchedCI
### Details
#### Promote Arcade to '.NET Eng - Latest' channel
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/1766460/logs/410) - Artifact Promote_Arcade_To_Latest_Promote_Arcade_to_'.NET_Eng_-_Latest'_channel_SBOM already exists for build 1766460.
### Changes
- [ac86b36b](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/ac86b36bdb46a65e1e5ec35b40468d81b4616b47) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220511.16 (#3111)
|
1.0
|
Build failed: dotnet-arcade-validation-official/main #20220511.6 - Build [#20220511.6](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=1766460) partiallySucceeded
## :warning: : internal / dotnet-arcade-validation-official partiallySucceeded
### Summary
**Finished** - Thu, 12 May 2022 17:08:41 GMT
**Duration** - 3 minutes
**Requested for** - DotNet Bot
**Reason** - batchedCI
### Details
#### Promote Arcade to '.NET Eng - Latest' channel
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/1766460/logs/410) - Artifact Promote_Arcade_To_Latest_Promote_Arcade_to_'.NET_Eng_-_Latest'_channel_SBOM already exists for build 1766460.
### Changes
- [ac86b36b](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/ac86b36bdb46a65e1e5ec35b40468d81b4616b47) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20220511.16 (#3111)
|
non_process
|
build failed dotnet arcade validation official main build partiallysucceeded warning internal dotnet arcade validation official partiallysucceeded summary finished thu may gmt duration minutes requested for dotnet bot reason batchedci details promote arcade to net eng latest channel x artifact promote arcade to latest promote arcade to net eng latest channel sbom already exists for build changes dotnet maestro update dependencies from build
| 0
|
220,350
| 24,564,929,014
|
IssuesEvent
|
2022-10-13 01:26:35
|
RG4421/skyux-sdk-builder
|
https://api.github.com/repos/RG4421/skyux-sdk-builder
|
opened
|
CVE-2022-37599 (Medium) detected in multiple libraries
|
security vulnerability
|
## CVE-2022-37599 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>loader-utils-2.0.0.tgz</b>, <b>loader-utils-1.4.0.tgz</b>, <b>loader-utils-1.2.3.tgz</b></p></summary>
<p>
<details><summary><b>loader-utils-2.0.0.tgz</b></p></summary>
<p>utils for webpack loaders</p>
<p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/loader-utils/package.json</p>
<p>
Dependency Hierarchy:
- :x: **loader-utils-2.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>loader-utils-1.4.0.tgz</b></p></summary>
<p>utils for webpack loaders</p>
<p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/sass-loader/node_modules/loader-utils/package.json,/node_modules/webpack/node_modules/loader-utils/package.json,/node_modules/istanbul-instrumenter-loader/node_modules/loader-utils/package.json,/node_modules/karma-webpack/node_modules/loader-utils/package.json,/node_modules/html-webpack-plugin/node_modules/loader-utils/package.json,/node_modules/raw-loader/node_modules/loader-utils/package.json,/node_modules/babel-loader/node_modules/loader-utils/package.json,/node_modules/awesome-typescript-loader/node_modules/loader-utils/package.json,/node_modules/css-loader/node_modules/loader-utils/package.json</p>
<p>
Dependency Hierarchy:
- babel-loader-8.1.0.tgz (Root Library)
- :x: **loader-utils-1.4.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>loader-utils-1.2.3.tgz</b></p></summary>
<p>utils for webpack loaders</p>
<p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.2.3.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.2.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/webpack-cli/node_modules/loader-utils/package.json</p>
<p>
Dependency Hierarchy:
- webpack-cli-3.3.11.tgz (Root Library)
- :x: **loader-utils-1.2.3.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the resourcePath variable in interpolateName.js.
<p>Publish Date: 2022-10-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599>CVE-2022-37599</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
|
True
|
CVE-2022-37599 (Medium) detected in multiple libraries - ## CVE-2022-37599 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>loader-utils-2.0.0.tgz</b>, <b>loader-utils-1.4.0.tgz</b>, <b>loader-utils-1.2.3.tgz</b></p></summary>
<p>
<details><summary><b>loader-utils-2.0.0.tgz</b></p></summary>
<p>utils for webpack loaders</p>
<p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/loader-utils/package.json</p>
<p>
Dependency Hierarchy:
- :x: **loader-utils-2.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>loader-utils-1.4.0.tgz</b></p></summary>
<p>utils for webpack loaders</p>
<p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/sass-loader/node_modules/loader-utils/package.json,/node_modules/webpack/node_modules/loader-utils/package.json,/node_modules/istanbul-instrumenter-loader/node_modules/loader-utils/package.json,/node_modules/karma-webpack/node_modules/loader-utils/package.json,/node_modules/html-webpack-plugin/node_modules/loader-utils/package.json,/node_modules/raw-loader/node_modules/loader-utils/package.json,/node_modules/babel-loader/node_modules/loader-utils/package.json,/node_modules/awesome-typescript-loader/node_modules/loader-utils/package.json,/node_modules/css-loader/node_modules/loader-utils/package.json</p>
<p>
Dependency Hierarchy:
- babel-loader-8.1.0.tgz (Root Library)
- :x: **loader-utils-1.4.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>loader-utils-1.2.3.tgz</b></p></summary>
<p>utils for webpack loaders</p>
<p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.2.3.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.2.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/webpack-cli/node_modules/loader-utils/package.json</p>
<p>
Dependency Hierarchy:
- webpack-cli-3.3.11.tgz (Root Library)
- :x: **loader-utils-1.2.3.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the resourcePath variable in interpolateName.js.
<p>Publish Date: 2022-10-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599>CVE-2022-37599</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
|
non_process
|
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries loader utils tgz loader utils tgz loader utils tgz loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules loader utils package json dependency hierarchy x loader utils tgz vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules sass loader node modules loader utils package json node modules webpack node modules loader utils package json node modules istanbul instrumenter loader node modules loader utils package json node modules karma webpack node modules loader utils package json node modules html webpack plugin node modules loader utils package json node modules raw loader node modules loader utils package json node modules babel loader node modules loader utils package json node modules awesome typescript loader node modules loader utils package json node modules css loader node modules loader utils package json dependency hierarchy babel loader tgz root library x loader utils tgz vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules webpack cli node modules loader utils package json dependency hierarchy webpack cli tgz root library x loader utils tgz vulnerable library found in base branch master vulnerability details a regular expression denial of service redos flaw was found in function interpolatename in interpolatename js in webpack loader utils via the resourcepath variable in interpolatename js publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href
| 0
|
10,727
| 13,530,333,370
|
IssuesEvent
|
2020-09-15 19:44:43
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Duplicate join generated when (manually) joining a table with a remapped field
|
.Query Language (MBQL) Priority:P2 Querying/Processor Type:Bug
|
We join twice when a remapping is defined.
To repo:
1) Define a remapping Orders/prodcut_id -> Products/Category
2) create a question in QB joining Orders and Products
Resulting MBQL has two joins to Orders instead of one.
|
1.0
|
Duplicate join generated when (manually) joining a table with a remapped field - We join twice when a remapping is defined.
To repo:
1) Define a remapping Orders/prodcut_id -> Products/Category
2) create a question in QB joining Orders and Products
Resulting MBQL has two joins to Orders instead of one.
|
process
|
duplicate join generated when manually joining a table with a remapped field we join twice when a remapping is defined to repo define a remapping orders prodcut id products category create a question in qb joining orders and products resulting mbql has two joins to orders instead of one
| 1
|
16,566
| 21,578,365,259
|
IssuesEvent
|
2022-05-02 15:57:12
|
cypress-io/cypress-documentation
|
https://api.github.com/repos/cypress-io/cypress-documentation
|
opened
|
Investigate Spell Checker
|
process: internal docs
|
### Subject
Docs Internal
### Description
Spell checking could be part of a lint process to cut down on manual reviewing and editing of basic spelling errors.
@elylucas
|
1.0
|
Investigate Spell Checker - ### Subject
Docs Internal
### Description
Spell checking could be part of a lint process to cut down on manual reviewing and editing of basic spelling errors.
@elylucas
|
process
|
investigate spell checker subject docs internal description spell checking could be part of a lint process to cut down on manual reviewing and editing of basic spelling errors elylucas
| 1
|
16,435
| 21,316,757,301
|
IssuesEvent
|
2022-04-16 12:16:22
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Common index preprocess
|
feature priority/medium preprocess index stale
|
## Description
Add preprocessing module that collects index terms and groups them. The grouping should be extendable so plugins can contribute grouping code to handle e.g. additional languages.
|
1.0
|
Common index preprocess - ## Description
Add preprocessing module that collects index terms and groups them. The grouping should be extendable so plugins can contribute grouping code to handle e.g. additional languages.
|
process
|
common index preprocess description add preprocessing module that collects index terms and groups them the grouping should be extendable so plugins can contribute grouping code to handle e g additional languages
| 1
|
30,569
| 6,176,278,319
|
IssuesEvent
|
2017-07-01 12:04:07
|
larcenists/larceny
|
https://api.github.com/repos/larcenists/larceny
|
closed
|
case sensitivity modes when loading file via command line may surprise users
|
C: ERR5RS P: trivial R: fixed T: defect
|
**Reported by: pnkfelix on Mon Jun 30 16:39:04 2008**
While running the heap building script for Larceny, I ran into some surprising undefined global variable errors when some of the FFI files had been updated (and thus were being loaded from source) while others were still being loaded from fasl files.
The errors looked like this:
```
% ./larceny.bin -stopcopy -- src/Build/iasn-larceny-heap.sch
...
Warning: loading source in favor of stale fasl file: /Users/pnkfelix/larcenytest/larceny-gcbench/larceny_src/lib/Ffi/ffi-i386.sch
Error: Undefined global variable "ffi/i386-C-callout-cdecl".
Entering debugger; type "?" for help.
debug>
```
The heap building script is run on top of the bootstrap heap, usually via the invocation `larceny.bin -stopcopy` followed by loading the heap build script via a command line parameter.
There were a couple of things that surprised me:
- loading the heap building script via the REPL worked, while loading the script via the command line broke (as illustrated at the top of the ticket).
- my initial attempts to make a small test case to expose the problem failed, because I tried to replicate the problem using the full fledged larceny.heap, rather than the bootstrap heap.
At this point, I believe I understand where the behavior is coming from (I think its from the fact that [source:trunk/larceny_src/src/Lib/Common/main.sch main.sch] adjusts the case sensitivity '''after''' invoking `failsafe-process-arguments`).
Here is a transcript of interactions with a larceny executable that illustrate the weirdness. (The fourth load of `foo.sch` is the one that I am worried about users being confused by.)
```
% cat foo.sch
(define Hello-World 'hELLO-wORLD!)
(display Hello-World)
(newline)
(exit)
% ./larceny
Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified)
larceny.heap, built on Mon Jun 30 16:36:29 EDT 2008
> (load "foo.sch")
hELLO-wORLD!
% ./larceny -- foo.sch
Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified)
hELLO-wORLD!
% ./larceny.bin
LARCENY_ROOT not set; using current directory
Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified)
> (load "foo.sch")
hELLO-wORLD!
% ./larceny.bin -- foo.sch
LARCENY_ROOT not set; using current directory
Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified)
hello-world!
%
```
This is not a huge issue, since I do not expect the average user to be doing anything starting from the bootstrap heap; if the system starts from the larceny.heap, then it starts with that heap's case sensitivity mode when loading files via the command line arguments.
But it is an issue for the heap building scripts like [source:trunk/larceny_src/src/Build/iasn-larceny-heap.sch iasn-larceny-heap.sch] or [source:trunk/larceny_src/src/Build/sparc-larceny-heap.sch sparc-larceny-heap.sch].
(After all, what sparked the problem for me initially were issues with the FFI .sch files being loaded in case-insensitive mode while the other FFI .sch files had been compiled in case-sensitive mode.)
I am currently planning to work around this problem by adding `no-fold-case` directives to the FFI source code, so that those files will continue to define their identifiers in a case-sensitive manner even when loaded from source. (It would also guard against the files being mis-compiled by a compiler that happens to be running in case-insensitive mode.) So that will be a band-aid for the problem with respect to the FFI source code. (Another fix would be to remove all case-sensitive identifiers from our libraries, but that is a questionable strategy for fixing this sort of thing.)
But we (that is, Will) may want to consider moving the invocation of `adjust-case-sensitivity!` above the loads of the init files and command line argument processing. (Additionally, the way things are structured now, a user's attempt to change the default case-sensitivity in a personal init file will be foiled by the subsequent invocation of `adjust-case-sensitivity!`)
|
1.0
|
case sensitivity modes when loading file via command line may surprise users - **Reported by: pnkfelix on Mon Jun 30 16:39:04 2008**
While running the heap building script for Larceny, I ran into some surprising undefined global variable errors when some of the FFI files had been updated (and thus were being loaded from source) while others were still being loaded from fasl files.
The errors looked like this:
```
% ./larceny.bin -stopcopy -- src/Build/iasn-larceny-heap.sch
...
Warning: loading source in favor of stale fasl file: /Users/pnkfelix/larcenytest/larceny-gcbench/larceny_src/lib/Ffi/ffi-i386.sch
Error: Undefined global variable "ffi/i386-C-callout-cdecl".
Entering debugger; type "?" for help.
debug>
```
The heap building script is run on top of the bootstrap heap, usually via the invocation `larceny.bin -stopcopy` followed by loading the heap build script via a command line parameter.
There were a couple of things that surprised me:
- loading the heap building script via the REPL worked, while loading the script via the command line broke (as illustrated at the top of the ticket).
- my initial attempts to make a small test case to expose the problem failed, because I tried to replicate the problem using the full fledged larceny.heap, rather than the bootstrap heap.
At this point, I believe I understand where the behavior is coming from (I think its from the fact that [source:trunk/larceny_src/src/Lib/Common/main.sch main.sch] adjusts the case sensitivity '''after''' invoking `failsafe-process-arguments`).
Here is a transcript of interactions with a larceny executable that illustrate the weirdness. (The fourth load of `foo.sch` is the one that I am worried about users being confused by.)
```
% cat foo.sch
(define Hello-World 'hELLO-wORLD!)
(display Hello-World)
(newline)
(exit)
% ./larceny
Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified)
larceny.heap, built on Mon Jun 30 16:36:29 EDT 2008
> (load "foo.sch")
hELLO-wORLD!
% ./larceny -- foo.sch
Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified)
hELLO-wORLD!
% ./larceny.bin
LARCENY_ROOT not set; using current directory
Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified)
> (load "foo.sch")
hELLO-wORLD!
% ./larceny.bin -- foo.sch
LARCENY_ROOT not set; using current directory
Larceny v0.961 "Fluoridation" (Jun 30 2008 15:52:46, precise:Posix Unix:unified)
hello-world!
%
```
This is not a huge issue, since I do not expect the average user to be doing anything starting from the bootstrap heap; if the system starts from the larceny.heap, then it starts with that heap's case sensitivity mode when loading files via the command line arguments.
But it is an issue for the heap building scripts like [source:trunk/larceny_src/src/Build/iasn-larceny-heap.sch iasn-larceny-heap.sch] or [source:trunk/larceny_src/src/Build/sparc-larceny-heap.sch sparc-larceny-heap.sch].
(After all, what sparked the problem for me initially were issues with the FFI .sch files being loaded in case-insensitive mode while the other FFI .sch files had been compiled in case-sensitive mode.)
I am currently planning to work around this problem by adding `no-fold-case` directives to the FFI source code, so that those files will continue to define their identifiers in a case-sensitive manner even when loaded from source. (It would also guard against the files being mis-compiled by a compiler that happens to be running in case-insensitive mode.) So that will be a band-aid for the problem with respect to the FFI source code. (Another fix would be to remove all case-sensitive identifiers from our libraries, but that is a questionable strategy for fixing this sort of thing.)
But we (that is, Will) may want to consider moving the invocation of `adjust-case-sensitivity!` above the loads of the init files and command line argument processing. (Additionally, the way things are structured now, a user's attempt to change the default case-sensitivity in a personal init file will be foiled by the subsequent invocation of `adjust-case-sensitivity!`)
|
non_process
|
case sensitivity modes when loading file via command line may surprise users reported by pnkfelix on mon jun while running the heap building script for larceny i ran into some surprising undefined global variable errors when some of the ffi files had been updated and thus were being loaded from source while others were still being loaded from fasl files the errors looked like this larceny bin stopcopy src build iasn larceny heap sch warning loading source in favor of stale fasl file users pnkfelix larcenytest larceny gcbench larceny src lib ffi ffi sch error undefined global variable ffi c callout cdecl entering debugger type for help debug the heap building script is run on top of the bootstrap heap usually via the invocation larceny bin stopcopy followed by loading the heap build script via a command line parameter there were a couple of things that surprised me loading the heap building script via the repl worked while loading the script via the command line broke as illustrated at the top of the ticket my initial attempts to make a small test case to expose the problem failed because i tried to replicate the problem using the full fledged larceny heap rather than the bootstrap heap at this point i believe i understand where the behavior is coming from i think its from the fact that adjusts the case sensitivity after invoking failsafe process arguments here is a transcript of interactions with a larceny executable that illustrate the weirdness the fourth load of foo sch is the one that i am worried about users being confused by cat foo sch define hello world hello world display hello world newline exit larceny larceny fluoridation jun precise posix unix unified larceny heap built on mon jun edt load foo sch hello world larceny foo sch larceny fluoridation jun precise posix unix unified hello world larceny bin larceny root not set using current directory larceny fluoridation jun precise posix unix unified load foo sch hello world larceny bin foo sch larceny root not set using current directory larceny fluoridation jun precise posix unix unified hello world this is not a huge issue since i do not expect the average user to be doing anything starting from the bootstrap heap if the system starts from the larceny heap then it starts with that heap s case sensitivity mode when loading files via the command line arguments but it is an issue for the heap building scripts like or after all what sparked the problem for me initially were issues with the ffi sch files being loaded in case insensitive mode while the other ffi sch files had been compiled in case sensitive mode i am currently planning to work around this problem by adding no fold case directives to the ffi source code so that those files will continue to define their identifiers in a case sensitive manner even when loaded from source it would also guard against the files being mis compiled by a compiler that happens to be running in case insensitive mode so that will be a band aid for the problem with respect to the ffi source code another fix would be to remove all case sensitive identifiers from our libraries but that is a questionable strategy for fixing this sort of thing but we that is will may want to consider moving the invocation of adjust case sensitivity above the loads of the init files and command line argument processing additionally the way things are structured now a user s attempt to change the default case sensitivity in a personal init file will be foiled by the subsequent invocation of adjust case sensitivity
| 0
|
180,047
| 13,916,711,303
|
IssuesEvent
|
2020-10-21 04:06:21
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
Incorrect SELinux labels for kubelet artifacts for node tests on Fedora CoreOS
|
kind/failing-test sig/node
|
<!-- Please only use this template for submitting reports about continuously failing tests or jobs in Kubernetes CI -->
**Which jobs are failing**:
pull-kubernetes-node-crio-e2e
**Which test(s) are failing**:
N/A
**Since when has it been failing**:
N/A
**Testgrid link**:
https://storage.googleapis.com/kubernetes-jenkins/pr-logs/pull/93486/pull-kubernetes-node-crio-e2e/1308737597702410240/artifacts/tmp-node-e2e-2f579b16-fedora-coreos-32-20200907-3-0-gcp-x86-64-system.log
**Reason for failure**:
```
Sep 23 12:20:21 localhost systemd[1]: Started /tmp/node-e2e-20200923T121804/kubelet --kubeconfig /tmp/node-e2e-20200923T121804/kubeconfig --root-dir /var/lib/kubelet --v 4 --logtostderr --dynamic-config-dir /tmp/node-e2e-20200923T121804/dynamic-kubelet-config --network-plugin=kubenet --cni-bin-dir /tmp/node-e2e-20200923T121804/cni/bin --cni-conf-dir /tmp/node-e2e-20200923T121804/cni/net.d --cni-cache-dir /tmp/node-e2e-20200923T121804/cni/cache --hostname-override tmp-node-e2e-2f579b16-fedora-coreos-32-20200907-3-0-gcp-x86-64 --container-runtime remote --container-runtime-endpoint unix:///run/crio/crio.sock --config /tmp/node-e2e-20200923T121804/kubelet-config --cgroups-per-qos=true --cgroup-root=/.
Sep 23 12:20:21 localhost audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=kubelet-20200923T121804 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Sep 23 12:20:21 localhost systemd[4212]: kubelet-20200923T121804.service: Failed to execute command: Permission denied
Sep 23 12:20:21 localhost systemd[4212]: kubelet-20200923T121804.service: Failed at step EXEC spawning /tmp/node-e2e-20200923T121804/kubelet: Permission denied
Sep 23 12:20:21 localhost audit[4212]: AVC avc: denied { execute } for pid=4212 comm="(kubelet)" name="kubelet" dev="tmpfs" ino=62327 scontext=system_u:system_r:init_t:s0 tcontext=unconfined_u:object_r:user_tmp_t:s0 tclass=file permissive=0
Sep 23 12:20:21 localhost systemd[1]: kubelet-20200923T121804.service: Main process exited, code=exited, status=203/EXEC
Sep 23 12:20:21 localhost systemd[1]: kubelet-20200923T121804.service: Failed with result 'exit-code'.
Sep 23 12:20:21 localhost audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=kubelet-20200923T121804 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed'
```
**Anything else we need to know**:
|
1.0
|
Incorrect SELinux labels for kubelet artifacts for node tests on Fedora CoreOS - <!-- Please only use this template for submitting reports about continuously failing tests or jobs in Kubernetes CI -->
**Which jobs are failing**:
pull-kubernetes-node-crio-e2e
**Which test(s) are failing**:
N/A
**Since when has it been failing**:
N/A
**Testgrid link**:
https://storage.googleapis.com/kubernetes-jenkins/pr-logs/pull/93486/pull-kubernetes-node-crio-e2e/1308737597702410240/artifacts/tmp-node-e2e-2f579b16-fedora-coreos-32-20200907-3-0-gcp-x86-64-system.log
**Reason for failure**:
```
Sep 23 12:20:21 localhost systemd[1]: Started /tmp/node-e2e-20200923T121804/kubelet --kubeconfig /tmp/node-e2e-20200923T121804/kubeconfig --root-dir /var/lib/kubelet --v 4 --logtostderr --dynamic-config-dir /tmp/node-e2e-20200923T121804/dynamic-kubelet-config --network-plugin=kubenet --cni-bin-dir /tmp/node-e2e-20200923T121804/cni/bin --cni-conf-dir /tmp/node-e2e-20200923T121804/cni/net.d --cni-cache-dir /tmp/node-e2e-20200923T121804/cni/cache --hostname-override tmp-node-e2e-2f579b16-fedora-coreos-32-20200907-3-0-gcp-x86-64 --container-runtime remote --container-runtime-endpoint unix:///run/crio/crio.sock --config /tmp/node-e2e-20200923T121804/kubelet-config --cgroups-per-qos=true --cgroup-root=/.
Sep 23 12:20:21 localhost audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=kubelet-20200923T121804 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Sep 23 12:20:21 localhost systemd[4212]: kubelet-20200923T121804.service: Failed to execute command: Permission denied
Sep 23 12:20:21 localhost systemd[4212]: kubelet-20200923T121804.service: Failed at step EXEC spawning /tmp/node-e2e-20200923T121804/kubelet: Permission denied
Sep 23 12:20:21 localhost audit[4212]: AVC avc: denied { execute } for pid=4212 comm="(kubelet)" name="kubelet" dev="tmpfs" ino=62327 scontext=system_u:system_r:init_t:s0 tcontext=unconfined_u:object_r:user_tmp_t:s0 tclass=file permissive=0
Sep 23 12:20:21 localhost systemd[1]: kubelet-20200923T121804.service: Main process exited, code=exited, status=203/EXEC
Sep 23 12:20:21 localhost systemd[1]: kubelet-20200923T121804.service: Failed with result 'exit-code'.
Sep 23 12:20:21 localhost audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:init_t:s0 msg='unit=kubelet-20200923T121804 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed'
```
**Anything else we need to know**:
|
non_process
|
incorrect selinux labels for kubelet artifacts for node tests on fedora coreos which jobs are failing pull kubernetes node crio which test s are failing n a since when has it been failing n a testgrid link reason for failure sep localhost systemd started tmp node kubelet kubeconfig tmp node kubeconfig root dir var lib kubelet v logtostderr dynamic config dir tmp node dynamic kubelet config network plugin kubenet cni bin dir tmp node cni bin cni conf dir tmp node cni net d cni cache dir tmp node cni cache hostname override tmp node fedora coreos gcp container runtime remote container runtime endpoint unix run crio crio sock config tmp node kubelet config cgroups per qos true cgroup root sep localhost audit service start pid uid auid ses subj system u system r init t msg unit kubelet comm systemd exe usr lib systemd systemd hostname addr terminal res success sep localhost systemd kubelet service failed to execute command permission denied sep localhost systemd kubelet service failed at step exec spawning tmp node kubelet permission denied sep localhost audit avc avc denied execute for pid comm kubelet name kubelet dev tmpfs ino scontext system u system r init t tcontext unconfined u object r user tmp t tclass file permissive sep localhost systemd kubelet service main process exited code exited status exec sep localhost systemd kubelet service failed with result exit code sep localhost audit service stop pid uid auid ses subj system u system r init t msg unit kubelet comm systemd exe usr lib systemd systemd hostname addr terminal res failed anything else we need to know
| 0
|
4,764
| 7,631,987,777
|
IssuesEvent
|
2018-05-05 09:42:48
|
AmpersandTarski/Ampersand
|
https://api.github.com/repos/AmpersandTarski/Ampersand
|
closed
|
Cleanup copied templates folder in prototype output dir
|
component:prototype generator priority:low software process
|
After generating the prototype, this folder is not used anymore.
This issue requests to delete the folder as final step in the prototype generator (e.g. after calling `composer ---install`)
Prevents users wasting time in case they don't know the purpose of this folder.
|
1.0
|
Cleanup copied templates folder in prototype output dir - After generating the prototype, this folder is not used anymore.
This issue requests to delete the folder as final step in the prototype generator (e.g. after calling `composer ---install`)
Prevents users wasting time in case they don't know the purpose of this folder.
|
process
|
cleanup copied templates folder in prototype output dir after generating the prototype this folder is not used anymore this issue requests to delete the folder as final step in the prototype generator e g after calling composer install prevents users wasting time in case they don t know the purpose of this folder
| 1
|
250,754
| 7,987,224,593
|
IssuesEvent
|
2018-07-19 06:57:23
|
architecture-building-systems/CityEnergyAnalyst
|
https://api.github.com/repos/architecture-building-systems/CityEnergyAnalyst
|
opened
|
Bug in ARU air-conditioning model calculations
|
Priority 1 bug
|
@martin-mosteiro reported strange behavior of the air conditioning model for the case when there is dehumidification and the interior air temperature is below the cooling supply temperature. this happens in spring in a very poorly insulated building outside the heating season.
|
1.0
|
Bug in ARU air-conditioning model calculations - @martin-mosteiro reported strange behavior of the air conditioning model for the case when there is dehumidification and the interior air temperature is below the cooling supply temperature. this happens in spring in a very poorly insulated building outside the heating season.
|
non_process
|
bug in aru air conditioning model calculations martin mosteiro reported strange behavior of the air conditioning model for the case when there is dehumidification and the interior air temperature is below the cooling supply temperature this happens in spring in a very poorly insulated building outside the heating season
| 0
|
19,375
| 25,501,988,434
|
IssuesEvent
|
2022-11-28 05:33:48
|
prusa3d/Prusa-Firmware
|
https://api.github.com/repos/prusa3d/Prusa-Firmware
|
closed
|
Put version info in release filenames
|
enhancement processing
|
Minor nit, but makes life a bit easier. Instead of the files being named like `FW_MK3-EINSy10a.hex`, add the version in so we get something like `FW_3.7.0_MK3-EINSy10a.hex`. That should give a nice sort order even if you have multiple printers of different models.
|
1.0
|
Put version info in release filenames - Minor nit, but makes life a bit easier. Instead of the files being named like `FW_MK3-EINSy10a.hex`, add the version in so we get something like `FW_3.7.0_MK3-EINSy10a.hex`. That should give a nice sort order even if you have multiple printers of different models.
|
process
|
put version info in release filenames minor nit but makes life a bit easier instead of the files being named like fw hex add the version in so we get something like fw hex that should give a nice sort order even if you have multiple printers of different models
| 1
|
11,224
| 14,004,032,107
|
IssuesEvent
|
2020-10-28 16:35:41
|
retaildevcrews/ngsa
|
https://api.github.com/repos/retaildevcrews/ngsa
|
reopened
|
Checkpoint Review - November
|
CPR Process
|
This checklist is for verifying the necessary details for a checkpoint review.
## [Engagement Name] - [Month] Checkpoint Review
- [Milestone/Sprint], [Month]
- [Date]
### Planned vs. Done
- List of high-level delivery plan from last month vs done log.
_This typically would be described at the EPIC/Workstream level and provide a high-level view of progress vs plan.
Note: Whenever possible link to artifacts from sources such as ADO vs pasting content from other documents_
| Epic / Workstream | Plan for the month | Status | Comments |
| ------------- | ------------- | ------------- | ------------- |
| Workstream1 | Y/N | Y/N | Y/N | Y/N |
| Workstream2| Y/N | Y/N | Y/N | Y/N |
| Workstream3| Y/N | Y/N | Y/N | Y/N |
| Workstream4 | Y/N | Y/N | Y/N | Y/N |
Question to add into talk track:
- What was planned to be done this month?
- What was actually done this month?
- Changes in plan versus last
- Key design decisions and their outcomes
### Work-Item Level
_This is typically linked directly from the Agile Planning Tool (GITHUB, ADO, Jira).
Note: Whenever possible link to artifacts from sources such as ADO vs pasting content from other documents_
| ID | Title | Work Item Type | Assigned To | State | Tags|
| ------------- | ------------- | ------------- | ------------- | ------------- | ------------- |
| | Workitem 1 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N |
| | Workitem 2 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N |
| | Workitem 3 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N |
| | Workitem 4 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N |
### ML/AI Status
| Topic | Status | Migration |
| ------------- | ------------- | ------------- |
| Responsible AI (RAIL compliance at current project phase) | Y/N | Y/N |
| Evaluation Framework Readiness (metrics, test, data, tooling) | Y/N | Y/N |
| Performance Evaulation (baseline, iterations | Y/N | Y/N |
### Key Changes / Decisions / Outcomes
_This is a list of any decisions or architecture changes that have been made since the last checkpoint and the action plan for remediation.
Note: Whenever possible link to artifacts from sources such as ADO vs pasting content. You can keep a decision log in your project repo or wiki to preserve this information for the customer._
| Key Decision / Architectural change | Remediation Plan |
| ------------- | ------------- |
| Key Decision 1 | Y/N |
| Key Decision 2 | Y/N |
| Key Decision 3 | Y/N |
## Engineering Fundamentals
_The latest [Engineering Fundamentals Assessment Report](https://aka.ms/EngineeringFundamentals) for the engagement_
### Retrospective Summary
_Provide details from the retrospectives over the last month_
- What went well:
- What didn't go well:
- Action Items:
### Risks, Issues, Additional Asks
_These items are copied from the initial Gameplan and are updated each month.
NOTE: Whenever possible link to the artifacts from sources such as ADO vs pasting content from other documents, See Risk Tag example below_
| Risk / Issue | Mitigation | Status|
| ------------- | ------------- | ------------- |
| Risk / Issue 1 | Y/N | Y/N |
| Risk / Issue 2 | Y/N | Y/N |
| Risk / Issue 3 | Y/N | Y/N |
_Include links to CSE Feedback and Feedback submitted to product groups as outlined in [CSE Engineering Feedback Playbook](https://cwcwiki.com/wiki/CSE_Engineering_Feedback_Playbook)_
### Final Checkpoint
_If this is the final checkpoint for an engagement, please also include a link to the [engagement wrap-up document ](https://aka.ms/wrapuptemplate)_
### Tasks
- [ ] Upload checkpoint review to Artifact Hub 24hrs in advance
- [ ] Tech Lead to complete Engineering Assessment via Snowball
|
1.0
|
Checkpoint Review - November - This checklist is for verifying the necessary details for a checkpoint review.
## [Engagement Name] - [Month] Checkpoint Review
- [Milestone/Sprint], [Month]
- [Date]
### Planned vs. Done
- List of high-level delivery plan from last month vs done log.
_This typically would be described at the EPIC/Workstream level and provide a high-level view of progress vs plan.
Note: Whenever possible link to artifacts from sources such as ADO vs pasting content from other documents_
| Epic / Workstream | Plan for the month | Status | Comments |
| ------------- | ------------- | ------------- | ------------- |
| Workstream1 | Y/N | Y/N | Y/N | Y/N |
| Workstream2| Y/N | Y/N | Y/N | Y/N |
| Workstream3| Y/N | Y/N | Y/N | Y/N |
| Workstream4 | Y/N | Y/N | Y/N | Y/N |
Question to add into talk track:
- What was planned to be done this month?
- What was actually done this month?
- Changes in plan versus last
- Key design decisions and their outcomes
### Work-Item Level
_This is typically linked directly from the Agile Planning Tool (GITHUB, ADO, Jira).
Note: Whenever possible link to artifacts from sources such as ADO vs pasting content from other documents_
| ID | Title | Work Item Type | Assigned To | State | Tags|
| ------------- | ------------- | ------------- | ------------- | ------------- | ------------- |
| | Workitem 1 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N |
| | Workitem 2 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N |
| | Workitem 3 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N |
| | Workitem 4 | Y/N | Y/N | Y/N | Y/N | Y/N | Y/N |
### ML/AI Status
| Topic | Status | Migration |
| ------------- | ------------- | ------------- |
| Responsible AI (RAIL compliance at current project phase) | Y/N | Y/N |
| Evaluation Framework Readiness (metrics, test, data, tooling) | Y/N | Y/N |
| Performance Evaulation (baseline, iterations | Y/N | Y/N |
### Key Changes / Decisions / Outcomes
_This is a list of any decisions or architecture changes that have been made since the last checkpoint and the action plan for remediation.
Note: Whenever possible link to artifacts from sources such as ADO vs pasting content. You can keep a decision log in your project repo or wiki to preserve this information for the customer._
| Key Decision / Architectural change | Remediation Plan |
| ------------- | ------------- |
| Key Decision 1 | Y/N |
| Key Decision 2 | Y/N |
| Key Decision 3 | Y/N |
## Engineering Fundamentals
_The latest [Engineering Fundamentals Assessment Report](https://aka.ms/EngineeringFundamentals) for the engagement_
### Retrospective Summary
_Provide details from the retrospectives over the last month_
- What went well:
- What didn't go well:
- Action Items:
### Risks, Issues, Additional Asks
_These items are copied from the initial Gameplan and are updated each month.
NOTE: Whenever possible link to the artifacts from sources such as ADO vs pasting content from other documents, See Risk Tag example below_
| Risk / Issue | Mitigation | Status|
| ------------- | ------------- | ------------- |
| Risk / Issue 1 | Y/N | Y/N |
| Risk / Issue 2 | Y/N | Y/N |
| Risk / Issue 3 | Y/N | Y/N |
_Include links to CSE Feedback and Feedback submitted to product groups as outlined in [CSE Engineering Feedback Playbook](https://cwcwiki.com/wiki/CSE_Engineering_Feedback_Playbook)_
### Final Checkpoint
_If this is the final checkpoint for an engagement, please also include a link to the [engagement wrap-up document ](https://aka.ms/wrapuptemplate)_
### Tasks
- [ ] Upload checkpoint review to Artifact Hub 24hrs in advance
- [ ] Tech Lead to complete Engineering Assessment via Snowball
|
process
|
checkpoint review november this checklist is for verifying the necessary details for a checkpoint review checkpoint review planned vs done list of high level delivery plan from last month vs done log this typically would be described at the epic workstream level and provide a high level view of progress vs plan note whenever possible link to artifacts from sources such as ado vs pasting content from other documents epic workstream plan for the month status comments y n y n y n y n y n y n y n y n y n y n y n y n y n y n y n y n question to add into talk track what was planned to be done this month what was actually done this month changes in plan versus last key design decisions and their outcomes work item level this is typically linked directly from the agile planning tool github ado jira note whenever possible link to artifacts from sources such as ado vs pasting content from other documents id title work item type assigned to state tags workitem y n y n y n y n y n y n workitem y n y n y n y n y n y n workitem y n y n y n y n y n y n workitem y n y n y n y n y n y n ml ai status topic status migration responsible ai rail compliance at current project phase y n y n evaluation framework readiness metrics test data tooling y n y n performance evaulation baseline iterations y n y n key changes decisions outcomes this is a list of any decisions or architecture changes that have been made since the last checkpoint and the action plan for remediation note whenever possible link to artifacts from sources such as ado vs pasting content you can keep a decision log in your project repo or wiki to preserve this information for the customer key decision architectural change remediation plan key decision y n key decision y n key decision y n engineering fundamentals the latest for the engagement retrospective summary provide details from the retrospectives over the last month what went well what didn t go well action items risks issues additional asks these items are copied from the initial gameplan and are updated each month note whenever possible link to the artifacts from sources such as ado vs pasting content from other documents see risk tag example below risk issue mitigation status risk issue y n y n risk issue y n y n risk issue y n y n include links to cse feedback and feedback submitted to product groups as outlined in final checkpoint if this is the final checkpoint for an engagement please also include a link to the tasks upload checkpoint review to artifact hub in advance tech lead to complete engineering assessment via snowball
| 1
|
1,989
| 4,816,895,086
|
IssuesEvent
|
2016-11-04 11:45:14
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Request for unifying shell command form when using exec and spawn
|
child_process
|
* **Version / Platform**:
v4.5.0 / MacOS
<!-- Enter your issue details below this comment. -->
Based on this site, there is a difference when calling an external script using `exec` and `spawn` methods. https://nodejs.org/api/child_process.html#child_process_child_process_exec_command_options_callback
One takes a string with space separated arguments and the second one takes a command in the first parameter and an array containing arguments in the second parameter. This is a real pain when we want to call multiple chained commands (for example: `clean dist && test --all && publish`) and stream their output.
|
1.0
|
Request for unifying shell command form when using exec and spawn - * **Version / Platform**:
v4.5.0 / MacOS
<!-- Enter your issue details below this comment. -->
Based on this site, there is a difference when calling an external script using `exec` and `spawn` methods. https://nodejs.org/api/child_process.html#child_process_child_process_exec_command_options_callback
One takes a string with space separated arguments and the second one takes a command in the first parameter and an array containing arguments in the second parameter. This is a real pain when we want to call multiple chained commands (for example: `clean dist && test --all && publish`) and stream their output.
|
process
|
request for unifying shell command form when using exec and spawn version platform macos based on this site there is a difference when calling an external script using exec and spawn methods one takes a string with space separated arguments and the second one takes a command in the first parameter and an array containing arguments in the second parameter this is a real pain when we want to call multiple chained commands for example clean dist test all publish and stream their output
| 1
|
24
| 2,496,270,287
|
IssuesEvent
|
2015-01-06 18:18:04
|
vivo-isf/vivo-isf-ontology
|
https://api.github.com/repos/vivo-isf/vivo-isf-ontology
|
closed
|
Courtship
|
biological_process imported
|
_From [fcold...@eagle-i.org](https://code.google.com/u/113677139039624182507/) on May 21, 2013 07:50:44_
\<b>**** Use the form below to request a new term ****</b>
\<b>**** Scroll down to see a term request example ****</b>
\<b>Please indicate the label for the proposed term:</b>
Courtship
\<b>Please provide a textual definition (with source):</b>
"Courtship can be defined as behavioral interactions between males and females, the evolutionary objective of which is copulation and the ultimate perpetuation of the species."
\<a href="http://cshprotocols.cshlp.org/content/2012/11/pdb.prot071647.long" rel="nofollow">http://cshprotocols.cshlp.org/content/2012/11/pdb.prot071647.long</a>
\<b>Please add an example of usage for proposed term:</b>
Term is to be used as a "Biological Process Studied" for a model organism
\<b>Please provide any additional optional information below. (e.g. desired</b>
\<b>asserted SuperClass in ERO hierarchy or Reference Branch)</b>
\<b>[ ] Instrument</b>
[X] Biological process
\<b>[ ] Disease</b>
\<b>[ ] Human studies</b>
\<b>[ ] Instrument</b>
\<b>[ ] Organism</b>
\<b>[ ] Reagent</b>
\<b>[ ] Software</b>
\<b>[ ] Technique</b>
\<b>[ ] Organization</b>
\<b>Additional info:</b>
\<b>*** Term request example ****</b>
\<b>Please indicate the label for the proposed term: four-terminal resistance</b>
\<b>sensor</b>
Please provide a textual definition (with source): "Four-terminal
\<b>resistance sensors are electrical impedance measuring instruments that use</b>
\<b>separate pairs of current-carrying and voltage-sensing electrodes to make</b>
\<b>accurate measurements that can be used to compute a material's electrical</b>
resistance." \<a href="http://en.wikipedia.org/wiki/Four-terminal_sensing" rel="nofollow">http://en.wikipedia.org/wiki/Four-terminal_sensing</a>
\<b>Please add an example of usage for proposed term: Measuring the inherent</b>
\<b>(per square) resistance of doped silicon.</b>
\<b>Please provide any additional optional information below. (e.g. desired</b>
\<b>asserted SuperClass in ERO hierarchy or Reference Branch)</b>
\<b>[X] Instrument</b>
\<b>[ ] Biological process</b>
\<b>[ ] Disease</b>
\<b>[ ] Human studies</b>
\<b>[ ] Instrument</b>
\<b>[ ] Organism</b>
\<b>[ ] Reagent</b>
\<b>[ ] Software</b>
\<b>[ ] Technique</b>
\<b>[ ] Organization</b>
\<b>Additional info: AKA - 4T sensors, 4-wire sensor, or 4-point probe</b>
_Original issue: http://code.google.com/p/eagle-i/issues/detail?id=217_
|
1.0
|
Courtship - _From [fcold...@eagle-i.org](https://code.google.com/u/113677139039624182507/) on May 21, 2013 07:50:44_
\<b>**** Use the form below to request a new term ****</b>
\<b>**** Scroll down to see a term request example ****</b>
\<b>Please indicate the label for the proposed term:</b>
Courtship
\<b>Please provide a textual definition (with source):</b>
"Courtship can be defined as behavioral interactions between males and females, the evolutionary objective of which is copulation and the ultimate perpetuation of the species."
\<a href="http://cshprotocols.cshlp.org/content/2012/11/pdb.prot071647.long" rel="nofollow">http://cshprotocols.cshlp.org/content/2012/11/pdb.prot071647.long</a>
\<b>Please add an example of usage for proposed term:</b>
Term is to be used as a "Biological Process Studied" for a model organism
\<b>Please provide any additional optional information below. (e.g. desired</b>
\<b>asserted SuperClass in ERO hierarchy or Reference Branch)</b>
\<b>[ ] Instrument</b>
[X] Biological process
\<b>[ ] Disease</b>
\<b>[ ] Human studies</b>
\<b>[ ] Instrument</b>
\<b>[ ] Organism</b>
\<b>[ ] Reagent</b>
\<b>[ ] Software</b>
\<b>[ ] Technique</b>
\<b>[ ] Organization</b>
\<b>Additional info:</b>
\<b>*** Term request example ****</b>
\<b>Please indicate the label for the proposed term: four-terminal resistance</b>
\<b>sensor</b>
Please provide a textual definition (with source): "Four-terminal
\<b>resistance sensors are electrical impedance measuring instruments that use</b>
\<b>separate pairs of current-carrying and voltage-sensing electrodes to make</b>
\<b>accurate measurements that can be used to compute a material's electrical</b>
resistance." \<a href="http://en.wikipedia.org/wiki/Four-terminal_sensing" rel="nofollow">http://en.wikipedia.org/wiki/Four-terminal_sensing</a>
\<b>Please add an example of usage for proposed term: Measuring the inherent</b>
\<b>(per square) resistance of doped silicon.</b>
\<b>Please provide any additional optional information below. (e.g. desired</b>
\<b>asserted SuperClass in ERO hierarchy or Reference Branch)</b>
\<b>[X] Instrument</b>
\<b>[ ] Biological process</b>
\<b>[ ] Disease</b>
\<b>[ ] Human studies</b>
\<b>[ ] Instrument</b>
\<b>[ ] Organism</b>
\<b>[ ] Reagent</b>
\<b>[ ] Software</b>
\<b>[ ] Technique</b>
\<b>[ ] Organization</b>
\<b>Additional info: AKA - 4T sensors, 4-wire sensor, or 4-point probe</b>
_Original issue: http://code.google.com/p/eagle-i/issues/detail?id=217_
|
process
|
courtship from on may use the form below to request a new term scroll down to see a term request example please indicate the label for the proposed term courtship please provide a textual definition with source courtship can be defined as behavioral interactions between males and females the evolutionary objective of which is copulation and the ultimate perpetuation of the species please add an example of usage for proposed term term is to be used as a biological process studied for a model organism please provide any additional optional information below e g desired asserted superclass in ero hierarchy or reference branch instrument biological process disease human studies instrument organism reagent software technique organization additional info term request example please indicate the label for the proposed term four terminal resistance sensor please provide a textual definition with source four terminal resistance sensors are electrical impedance measuring instruments that use separate pairs of current carrying and voltage sensing electrodes to make accurate measurements that can be used to compute a material s electrical resistance please add an example of usage for proposed term measuring the inherent per square resistance of doped silicon please provide any additional optional information below e g desired asserted superclass in ero hierarchy or reference branch instrument biological process disease human studies instrument organism reagent software technique organization additional info aka sensors wire sensor or point probe original issue
| 1
|
4,863
| 7,746,917,344
|
IssuesEvent
|
2018-05-30 00:03:57
|
AppFolioOnboarding/image-sharer-ChaoHuangAtAppfolio
|
https://api.github.com/repos/AppFolioOnboarding/image-sharer-ChaoHuangAtAppfolio
|
closed
|
Image Seeds
|
in process
|
#### As a developer I want at least 20 images to be created when I setup the database.
__Story__:
Working on this application has been a lot of fun, right? You are fast
becoming a git expert, and an expert at working with a deployment
pipeline. Speaking of development and deployment pipelines, is it not pretty
annoying that everytime a new deployment is created it starts out with an empty
database? It seems like everytime a review application is created we need to
manually add a handful of images in order to manually test the
application. Would it not be great if each deployment was created with a
handful of existing images?
__Acceptance criteria__:
- [ ] After running `rake db:setup` at least 20 unique images are shown in the
homepage.
- [ ] When a new app is deployed on heroku (e.g., every review app), that app
automatically has at least 20 unique images on its homepage.
__Dependencies__:
- Image Index
|
1.0
|
Image Seeds - #### As a developer I want at least 20 images to be created when I setup the database.
__Story__:
Working on this application has been a lot of fun, right? You are fast
becoming a git expert, and an expert at working with a deployment
pipeline. Speaking of development and deployment pipelines, is it not pretty
annoying that everytime a new deployment is created it starts out with an empty
database? It seems like everytime a review application is created we need to
manually add a handful of images in order to manually test the
application. Would it not be great if each deployment was created with a
handful of existing images?
__Acceptance criteria__:
- [ ] After running `rake db:setup` at least 20 unique images are shown in the
homepage.
- [ ] When a new app is deployed on heroku (e.g., every review app), that app
automatically has at least 20 unique images on its homepage.
__Dependencies__:
- Image Index
|
process
|
image seeds as a developer i want at least images to be created when i setup the database story working on this application has been a lot of fun right you are fast becoming a git expert and an expert at working with a deployment pipeline speaking of development and deployment pipelines is it not pretty annoying that everytime a new deployment is created it starts out with an empty database it seems like everytime a review application is created we need to manually add a handful of images in order to manually test the application would it not be great if each deployment was created with a handful of existing images acceptance criteria after running rake db setup at least unique images are shown in the homepage when a new app is deployed on heroku e g every review app that app automatically has at least unique images on its homepage dependencies image index
| 1
|
16,541
| 21,566,803,155
|
IssuesEvent
|
2022-05-02 00:12:35
|
jmacost5/CPP-528-Project
|
https://api.github.com/repos/jmacost5/CPP-528-Project
|
closed
|
Commits have useful names and clear descriptions
|
TEAM PROCESS
|
Good example: 'Create .rds file that stores original and final predictive model'
Bad example: 'Updated files'
In general, good commits start with a present tense verb that summarizes your work in 50 character or less. If I can’t tell exactly what change/edit was made, neither will you six months from now.
|
1.0
|
Commits have useful names and clear descriptions - Good example: 'Create .rds file that stores original and final predictive model'
Bad example: 'Updated files'
In general, good commits start with a present tense verb that summarizes your work in 50 character or less. If I can’t tell exactly what change/edit was made, neither will you six months from now.
|
process
|
commits have useful names and clear descriptions good example create rds file that stores original and final predictive model bad example updated files in general good commits start with a present tense verb that summarizes your work in character or less if i can’t tell exactly what change edit was made neither will you six months from now
| 1
|
3,826
| 6,802,323,816
|
IssuesEvent
|
2017-11-02 19:47:35
|
WikiWatershed/model-my-watershed
|
https://api.github.com/repos/WikiWatershed/model-my-watershed
|
closed
|
Geoprocessing API: Validate Shape Does Not Self Intersect
|
Geoprocessing API tested/verified WPF
|
We already check for self-intersections on the frontend via `leaflet-draw` and `turf-kink`, but we now have to guard against bad input from API users. Add a check on AOIs submitted to the analyze endpoints.
|
1.0
|
Geoprocessing API: Validate Shape Does Not Self Intersect - We already check for self-intersections on the frontend via `leaflet-draw` and `turf-kink`, but we now have to guard against bad input from API users. Add a check on AOIs submitted to the analyze endpoints.
|
process
|
geoprocessing api validate shape does not self intersect we already check for self intersections on the frontend via leaflet draw and turf kink but we now have to guard against bad input from api users add a check on aois submitted to the analyze endpoints
| 1
|
258,715
| 27,578,011,620
|
IssuesEvent
|
2023-03-08 14:23:27
|
DEV-REPO-URIEL/TEST_AINAT_CHANGE
|
https://api.github.com/repos/DEV-REPO-URIEL/TEST_AINAT_CHANGE
|
opened
|
chart.js-2.9.3.tgz: 1 vulnerabilities (highest severity is: 7.5)
|
Mend: dependency security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>chart.js-2.9.3.tgz</b></p></summary>
<p>Simple HTML5 charts using the canvas element.</p>
<p>Library home page: <a href="https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz">https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/chart.js/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/DEV-REPO-URIEL/TEST_AINAT_CHANGE/commit/1855a834302dd3f0182c78b12ce583c7af3a2921">1855a834302dd3f0182c78b12ce583c7af3a2921</a></p></details>
#### <img src='https://whitesource-resources.whitesourcesoftware.com/suggestedVersion.png' width=19 height=20> Mend has checked all newer package trees, and you are on the least vulnerable package!
#### Please note: There might be a version that explicitly solves one or more of the vulnerabilities listed below, but we do not recommend it. For more info about the optional fixes, check the section “Details” below.
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (chart.js version) | Fix PR available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2020-7746](https://www.mend.io/vulnerability-database/CVE-2020-7746) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | chart.js-2.9.3.tgz | Direct | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-7746</summary>
### Vulnerable Library - <b>chart.js-2.9.3.tgz</b></p>
<p>Simple HTML5 charts using the canvas element.</p>
<p>Library home page: <a href="https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz">https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/chart.js/package.json</p>
<p>
Dependency Hierarchy:
- :x: **chart.js-2.9.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/DEV-REPO-URIEL/TEST_AINAT_CHANGE/commit/1855a834302dd3f0182c78b12ce583c7af3a2921">1855a834302dd3f0182c78b12ce583c7af3a2921</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
This affects the package chart.js before 2.9.4. The options parameter is not properly sanitized when it is processed. When the options are processed, the existing options (or the defaults options) are deeply merged with provided options. However, during this operation, the keys of the object being set are not checked, leading to a prototype pollution.
<p>Publish Date: 2020-10-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-7746>CVE-2020-7746</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7746">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7746</a></p>
<p>Release Date: 2020-10-29</p>
<p>Fix Resolution: chart.js - 2.9.4</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details>
|
True
|
chart.js-2.9.3.tgz: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>chart.js-2.9.3.tgz</b></p></summary>
<p>Simple HTML5 charts using the canvas element.</p>
<p>Library home page: <a href="https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz">https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/chart.js/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/DEV-REPO-URIEL/TEST_AINAT_CHANGE/commit/1855a834302dd3f0182c78b12ce583c7af3a2921">1855a834302dd3f0182c78b12ce583c7af3a2921</a></p></details>
#### <img src='https://whitesource-resources.whitesourcesoftware.com/suggestedVersion.png' width=19 height=20> Mend has checked all newer package trees, and you are on the least vulnerable package!
#### Please note: There might be a version that explicitly solves one or more of the vulnerabilities listed below, but we do not recommend it. For more info about the optional fixes, check the section “Details” below.
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (chart.js version) | Fix PR available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2020-7746](https://www.mend.io/vulnerability-database/CVE-2020-7746) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | chart.js-2.9.3.tgz | Direct | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-7746</summary>
### Vulnerable Library - <b>chart.js-2.9.3.tgz</b></p>
<p>Simple HTML5 charts using the canvas element.</p>
<p>Library home page: <a href="https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz">https://registry.npmjs.org/chart.js/-/chart.js-2.9.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/chart.js/package.json</p>
<p>
Dependency Hierarchy:
- :x: **chart.js-2.9.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/DEV-REPO-URIEL/TEST_AINAT_CHANGE/commit/1855a834302dd3f0182c78b12ce583c7af3a2921">1855a834302dd3f0182c78b12ce583c7af3a2921</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
This affects the package chart.js before 2.9.4. The options parameter is not properly sanitized when it is processed. When the options are processed, the existing options (or the defaults options) are deeply merged with provided options. However, during this operation, the keys of the object being set are not checked, leading to a prototype pollution.
<p>Publish Date: 2020-10-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-7746>CVE-2020-7746</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7746">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7746</a></p>
<p>Release Date: 2020-10-29</p>
<p>Fix Resolution: chart.js - 2.9.4</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details>
|
non_process
|
chart js tgz vulnerabilities highest severity is vulnerable library chart js tgz simple charts using the canvas element library home page a href path to dependency file package json path to vulnerable library node modules chart js package json found in head commit a href mend has checked all newer package trees and you are on the least vulnerable package please note there might be a version that explicitly solves one or more of the vulnerabilities listed below but we do not recommend it for more info about the optional fixes check the section “details” below vulnerabilities cve severity cvss dependency type fixed in chart js version fix pr available high chart js tgz direct n a details cve vulnerable library chart js tgz simple charts using the canvas element library home page a href path to dependency file package json path to vulnerable library node modules chart js package json dependency hierarchy x chart js tgz vulnerable library found in head commit a href found in base branch main vulnerability details this affects the package chart js before the options parameter is not properly sanitized when it is processed when the options are processed the existing options or the defaults options are deeply merged with provided options however during this operation the keys of the object being set are not checked leading to a prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution chart js in order to enable automatic remediation please create
| 0
|
222,574
| 7,433,973,611
|
IssuesEvent
|
2018-03-26 09:27:09
|
CS2103JAN2018-W09-B2/main
|
https://api.github.com/repos/CS2103JAN2018-W09-B2/main
|
opened
|
Create new UI for drafting email
|
priority.high type.enhancement type.task
|
#31
Need UI for user to input information like subject title, content and attachments before sending email.
|
1.0
|
Create new UI for drafting email - #31
Need UI for user to input information like subject title, content and attachments before sending email.
|
non_process
|
create new ui for drafting email need ui for user to input information like subject title content and attachments before sending email
| 0
|
7,177
| 10,318,919,399
|
IssuesEvent
|
2019-08-30 16:03:53
|
googleapis/nodejs-logging-winston
|
https://api.github.com/repos/googleapis/nodejs-logging-winston
|
closed
|
Cut 2.0.0 release
|
type: process
|
While the [2.0.0 release PR](https://github.com/googleapis/nodejs-logging-winston/pull/365) is open, it is not quite optimal from the release notes point of view. For semver majors, it is important that the release notes communicate clearly what is breaking. I pushed a commit with [a better changelog](https://github.com/googleapis/nodejs-logging-winston/commit/443ee2310e7eaaa1c14e21d3f7e4d13e0bb27261), but that got overwritten by the release automation.
How do we do a release with proper release notes?
|
1.0
|
Cut 2.0.0 release - While the [2.0.0 release PR](https://github.com/googleapis/nodejs-logging-winston/pull/365) is open, it is not quite optimal from the release notes point of view. For semver majors, it is important that the release notes communicate clearly what is breaking. I pushed a commit with [a better changelog](https://github.com/googleapis/nodejs-logging-winston/commit/443ee2310e7eaaa1c14e21d3f7e4d13e0bb27261), but that got overwritten by the release automation.
How do we do a release with proper release notes?
|
process
|
cut release while the is open it is not quite optimal from the release notes point of view for semver majors it is important that the release notes communicate clearly what is breaking i pushed a commit with but that got overwritten by the release automation how do we do a release with proper release notes
| 1
|
318,806
| 27,321,992,334
|
IssuesEvent
|
2023-02-24 20:46:51
|
brave/brave-ios
|
https://api.github.com/repos/brave/brave-ios
|
closed
|
Manual test run for `1.48` on `iPhone` running `iOS 16`
|
QA/Yes release-notes/exclude QA Pass - iPhone tests iPhone iOS 16
|
## Installer
- [x] Check that installer is close to the size of the last release
- [x] Check the Brave version in About and make sure it is EXACTLY as expected
## Data
- [x] Verify that data from the previous build appears in the updated build as expected (bookmarks, history, etc.)
- [x] Verify that cookies from the previous build are preserved after upgrade
- [x] Verify saved passwords are retained after upgrade
- [x] Verify stats are retained after upgrade
- [x] Verify sync chain created in the previous version is still retained on upgrade
- [x] Verify per-site settings are preserved after upgrade
## Bookmarks
- [x] Verify that creating a bookmark works
- [x] Verify that tapping a bookmark from bookmark manager loads the bookmark
- [x] Verify that deleting a bookmark works
- [x] Verify that creating a bookmark folder works
- [x] Verify that creating a bookmark inside the created folder works
- [x] Verify that you can add a bookmark directly inside a bookmark folder
- [x] Verify that you can delete a bookmark in edit mode
- [x] Verify that you can delete a bookmark folder with bookmarks inside
- [x] Verify adding a bookmark domain, subpaths is retained and you are successfully able to visit the domain subpath in a new tab
## Favourites
- [x] Verify editing favourites and changing their URLs updates the favicons accordingly
- [x] Verify that you can remove favourites
- [x] Verify that you can add new favourites from the `Share with...` menu
## Context menus
- [x] Make sure context menu items in the URL bar work
- [x] Make sure context menu items on content work with no selected text
- [x] Make sure context menu items on content work with selected text
- [x] Make sure context menu items on content work inside an editable control (input, textarea, or contenteditable)
- [x] Verify you can open links in a new tab or private tab. Ensure you can open links in background tabs.
## Downloads
- [x] Verify that you can save an image from a site
- [x] Verify that you are able to save a GIF image
## Reader Mode
- [x] Visit `http://theverge.com`, open any article, verify the reader mode icon is shown in the URL bar
- [x] Verify tapping on the reader mode icon opens the article in reader mode
- [x] Edit reader mode settings and open different pages in reader mode and verify if the setting is retained across each article
## Gestures
- [x] Verify pinch to zoom in/out gestures work on `https://www.homedepot.com`
- [x] Verify that navigating to a different origin resets the zoom
- [x] Swipe back and forward to navigate, verify this works as expected
## Password Managers
- [x] Verify tapping on 1Password on the slide-out keyboard launches 1Password app and you're able to select the stored credentials
- [x] Verify tapping on Bitwarden password manager in the autofill field launches the app and auto-fills the stored data
## Shields Settings
- [x] Enable all switches in `Settings` and visit a site and disable `Block Scripts`. Kill and relaunch the app and verify if the site Shields settings are retained
## Browser Lock
- [x] Verify browser lock enables device passcode to lock/unlock the browser
- [x] Verify swipe up/swipe down with browser in focus doesn't ask for device passcode
- [x] Verify cancel passcode shows unlock message
- [x] Remove the app from memory and relaunch, enter a wrong passcode, the browser should not be unlocked
- [x] Verify cancel fingerprint confirmation/face unlock shows enter passcode window when fingerprint/face unlock is set up on the device
## Brave Rewards/Ads
- [x] Verify wallet is auto-created after enabling Rewards
- [x] Verify when you click on the Brave Rewards panel while on a site, the panel displays if the site is verified or not
- [x] Verify ads are only shown while the app is being used
- [x] Verify tapping on an ad notification shows the landing page
- [x] Verify inline-content ads show on Brave News and are correctly redeemed from the server
- [x] Verify `view`,`clicked`, `landed`, and `dismiss` states are logged based on the action
## Sync
- [x] Verify you are able to join Sync chain by scanning the QR code
- [x] Verify you are able to join Sync chain using code words
- [x] Verify you are able to create a Sync chain on the device and add other devices to the chain via QR code/code words
- [x] Verify that bookmarks from other devices on the chain show up on the mobile device after sync completes
- [x] Verify newly created bookmarks get synced to all devices on the Sync chain
- [x] Verify existing bookmarks before joining Sync chain also gets sync'd to all devices on the Sync chain
- [x] Verify history sync is disabled by default. Enabling it starts sync'ing history items as well (limited to 200 entries).
- [x] Verify sync works on an upgraded profile and new bookmarks added post-upgrade sync across devices on the chain
- [x] Verify you can create a standalone Sync chain with one device
## Playlist & Autoplay
- [x] Verify Playlist is enabled by default on a clean install
- [x] Verify when a site with audio/video is detected, `Playlist` button is shown in URL bar and a notification badge on the menu
- [x] Verify clicking the `Playlist` button adds audio/video to playlist and initiates offline play
- [x] Verify able to add audio/video to playlist via share menu
- [x] Verify able to add audio/video to playlist via long press context menu
- [x] Verify Autoplay is disabled by default
- [x] Verify with Autoplay enabled, able to play audio/video automatically
- [x] Verify with background audio enabled, able to play audio/video with device locked or app in background
## Brave News
- [x] Verify Brave News opt-in is shown on a clean install
- [x] Verify once opted-in Brave News starts showing articles
- [x] Verify inline-content ads are shown on Brave News
- [x] Verify able to add custom RSS feed to the list of sources
- [x] Verify able to disable content from a particular publisher by long-press on the card
## Bravery settings
- [x] Check that block ad and unblock ad works on `http://slashdot.org`
- [x] Check that toggling to blocking and allow ads works as expected
- [x] Verify that tapping through a cert error in `https://badssl.com` works
- [x] Enable `Block Scripts` globally from `Settings`. Visit `https://twitter.com`, nothing should load. Tap on `Shields` and disable `Block Scripts`; the page should load properly.
- [x] Verify that default Brave Shields settings take effect on pages with no site settings
### Fingerprint Tests
- [x] Verify that turning on fingerprinting protection in preferences shows `3 fingerprints blocked` at `https://browserleaks.com/canvas`. Verify that turning it off in the Bravery menu shows `0 fingerprints blocked`.
- [x] Verify that the audio fingerprint is blocked at `https://audiofingerprint.openwpm.com` when fingerprinting protection is on
- [x] Verify that `https://diafygi.github.io/webrtc-ips/` doesn't leak IP address when `Block all fingerprinting protection` is on
## Content tests
- [x] Go to `https://feedly.com` and make sure that the password can be saved. Make sure the saved password is auto-populated when you visit the site again.
- [x] Open an email on `https://mail.google.com` and tap on a link. Make sure it works.
- [x] Verify that PDF is loaded over `HTTPS` at `https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf`
- [x] Verify that PDF is loaded over `HTTP` at `http://www.pdf995.com/samples/pdf.pdf`
- [x] Verify that `https://mixed-script.badssl.com` shows up as grey, not red (no mixed-content scripts are run)
- [x] Verify that search results from `https://startpage.com` open in a new tab (due to target being _blank_)
## WebAuthn
- [x] Verify browser prompts for security key when trying to log in to accounts
- [x] Verify you are able to successfully authenticate an account using security key when prompted
- [x] Verify quickly connecting and disconnecting doesn't cause issues authenticating using security keys
- [x] Verify you are able to use security keys on both normal and private tabs
- [x] Verify you are able to use security keys to authenticate using NFC
## App linker
- [x] Long-press on a link in the Twitter app to get the share picker, choose Brave. Verify Brave doesn't crash after opening the link.
## Session storage
- [x] Verify that tabs restore when closed, including active tab
|
1.0
|
Manual test run for `1.48` on `iPhone` running `iOS 16` - ## Installer
- [x] Check that installer is close to the size of the last release
- [x] Check the Brave version in About and make sure it is EXACTLY as expected
## Data
- [x] Verify that data from the previous build appears in the updated build as expected (bookmarks, history, etc.)
- [x] Verify that cookies from the previous build are preserved after upgrade
- [x] Verify saved passwords are retained after upgrade
- [x] Verify stats are retained after upgrade
- [x] Verify sync chain created in the previous version is still retained on upgrade
- [x] Verify per-site settings are preserved after upgrade
## Bookmarks
- [x] Verify that creating a bookmark works
- [x] Verify that tapping a bookmark from bookmark manager loads the bookmark
- [x] Verify that deleting a bookmark works
- [x] Verify that creating a bookmark folder works
- [x] Verify that creating a bookmark inside the created folder works
- [x] Verify that you can add a bookmark directly inside a bookmark folder
- [x] Verify that you can delete a bookmark in edit mode
- [x] Verify that you can delete a bookmark folder with bookmarks inside
- [x] Verify adding a bookmark domain, subpaths is retained and you are successfully able to visit the domain subpath in a new tab
## Favourites
- [x] Verify editing favourites and changing their URLs updates the favicons accordingly
- [x] Verify that you can remove favourites
- [x] Verify that you can add new favourites from the `Share with...` menu
## Context menus
- [x] Make sure context menu items in the URL bar work
- [x] Make sure context menu items on content work with no selected text
- [x] Make sure context menu items on content work with selected text
- [x] Make sure context menu items on content work inside an editable control (input, textarea, or contenteditable)
- [x] Verify you can open links in a new tab or private tab. Ensure you can open links in background tabs.
## Downloads
- [x] Verify that you can save an image from a site
- [x] Verify that you are able to save a GIF image
## Reader Mode
- [x] Visit `http://theverge.com`, open any article, verify the reader mode icon is shown in the URL bar
- [x] Verify tapping on the reader mode icon opens the article in reader mode
- [x] Edit reader mode settings and open different pages in reader mode and verify if the setting is retained across each article
## Gestures
- [x] Verify pinch to zoom in/out gestures work on `https://www.homedepot.com`
- [x] Verify that navigating to a different origin resets the zoom
- [x] Swipe back and forward to navigate, verify this works as expected
## Password Managers
- [x] Verify tapping on 1Password on the slide-out keyboard launches 1Password app and you're able to select the stored credentials
- [x] Verify tapping on Bitwarden password manager in the autofill field launches the app and auto-fills the stored data
## Shields Settings
- [x] Enable all switches in `Settings` and visit a site and disable `Block Scripts`. Kill and relaunch the app and verify if the site Shields settings are retained
## Browser Lock
- [x] Verify browser lock enables device passcode to lock/unlock the browser
- [x] Verify swipe up/swipe down with browser in focus doesn't ask for device passcode
- [x] Verify cancel passcode shows unlock message
- [x] Remove the app from memory and relaunch, enter a wrong passcode, the browser should not be unlocked
- [x] Verify cancel fingerprint confirmation/face unlock shows enter passcode window when fingerprint/face unlock is set up on the device
## Brave Rewards/Ads
- [x] Verify wallet is auto-created after enabling Rewards
- [x] Verify when you click on the Brave Rewards panel while on a site, the panel displays if the site is verified or not
- [x] Verify ads are only shown while the app is being used
- [x] Verify tapping on an ad notification shows the landing page
- [x] Verify inline-content ads show on Brave News and are correctly redeemed from the server
- [x] Verify `view`,`clicked`, `landed`, and `dismiss` states are logged based on the action
## Sync
- [x] Verify you are able to join Sync chain by scanning the QR code
- [x] Verify you are able to join Sync chain using code words
- [x] Verify you are able to create a Sync chain on the device and add other devices to the chain via QR code/code words
- [x] Verify that bookmarks from other devices on the chain show up on the mobile device after sync completes
- [x] Verify newly created bookmarks get synced to all devices on the Sync chain
- [x] Verify existing bookmarks before joining Sync chain also gets sync'd to all devices on the Sync chain
- [x] Verify history sync is disabled by default. Enabling it starts sync'ing history items as well (limited to 200 entries).
- [x] Verify sync works on an upgraded profile and new bookmarks added post-upgrade sync across devices on the chain
- [x] Verify you can create a standalone Sync chain with one device
## Playlist & Autoplay
- [x] Verify Playlist is enabled by default on a clean install
- [x] Verify when a site with audio/video is detected, `Playlist` button is shown in URL bar and a notification badge on the menu
- [x] Verify clicking the `Playlist` button adds audio/video to playlist and initiates offline play
- [x] Verify able to add audio/video to playlist via share menu
- [x] Verify able to add audio/video to playlist via long press context menu
- [x] Verify Autoplay is disabled by default
- [x] Verify with Autoplay enabled, able to play audio/video automatically
- [x] Verify with background audio enabled, able to play audio/video with device locked or app in background
## Brave News
- [x] Verify Brave News opt-in is shown on a clean install
- [x] Verify once opted-in Brave News starts showing articles
- [x] Verify inline-content ads are shown on Brave News
- [x] Verify able to add custom RSS feed to the list of sources
- [x] Verify able to disable content from a particular publisher by long-press on the card
## Bravery settings
- [x] Check that block ad and unblock ad works on `http://slashdot.org`
- [x] Check that toggling to blocking and allow ads works as expected
- [x] Verify that tapping through a cert error in `https://badssl.com` works
- [x] Enable `Block Scripts` globally from `Settings`. Visit `https://twitter.com`, nothing should load. Tap on `Shields` and disable `Block Scripts`; the page should load properly.
- [x] Verify that default Brave Shields settings take effect on pages with no site settings
### Fingerprint Tests
- [x] Verify that turning on fingerprinting protection in preferences shows `3 fingerprints blocked` at `https://browserleaks.com/canvas`. Verify that turning it off in the Bravery menu shows `0 fingerprints blocked`.
- [x] Verify that the audio fingerprint is blocked at `https://audiofingerprint.openwpm.com` when fingerprinting protection is on
- [x] Verify that `https://diafygi.github.io/webrtc-ips/` doesn't leak IP address when `Block all fingerprinting protection` is on
## Content tests
- [x] Go to `https://feedly.com` and make sure that the password can be saved. Make sure the saved password is auto-populated when you visit the site again.
- [x] Open an email on `https://mail.google.com` and tap on a link. Make sure it works.
- [x] Verify that PDF is loaded over `HTTPS` at `https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf`
- [x] Verify that PDF is loaded over `HTTP` at `http://www.pdf995.com/samples/pdf.pdf`
- [x] Verify that `https://mixed-script.badssl.com` shows up as grey, not red (no mixed-content scripts are run)
- [x] Verify that search results from `https://startpage.com` open in a new tab (due to target being _blank_)
## WebAuthn
- [x] Verify browser prompts for security key when trying to log in to accounts
- [x] Verify you are able to successfully authenticate an account using security key when prompted
- [x] Verify quickly connecting and disconnecting doesn't cause issues authenticating using security keys
- [x] Verify you are able to use security keys on both normal and private tabs
- [x] Verify you are able to use security keys to authenticate using NFC
## App linker
- [x] Long-press on a link in the Twitter app to get the share picker, choose Brave. Verify Brave doesn't crash after opening the link.
## Session storage
- [x] Verify that tabs restore when closed, including active tab
|
non_process
|
manual test run for on iphone running ios installer check that installer is close to the size of the last release check the brave version in about and make sure it is exactly as expected data verify that data from the previous build appears in the updated build as expected bookmarks history etc verify that cookies from the previous build are preserved after upgrade verify saved passwords are retained after upgrade verify stats are retained after upgrade verify sync chain created in the previous version is still retained on upgrade verify per site settings are preserved after upgrade bookmarks verify that creating a bookmark works verify that tapping a bookmark from bookmark manager loads the bookmark verify that deleting a bookmark works verify that creating a bookmark folder works verify that creating a bookmark inside the created folder works verify that you can add a bookmark directly inside a bookmark folder verify that you can delete a bookmark in edit mode verify that you can delete a bookmark folder with bookmarks inside verify adding a bookmark domain subpaths is retained and you are successfully able to visit the domain subpath in a new tab favourites verify editing favourites and changing their urls updates the favicons accordingly verify that you can remove favourites verify that you can add new favourites from the share with menu context menus make sure context menu items in the url bar work make sure context menu items on content work with no selected text make sure context menu items on content work with selected text make sure context menu items on content work inside an editable control input textarea or contenteditable verify you can open links in a new tab or private tab ensure you can open links in background tabs downloads verify that you can save an image from a site verify that you are able to save a gif image reader mode visit open any article verify the reader mode icon is shown in the url bar verify tapping on the reader mode icon opens the article in reader mode edit reader mode settings and open different pages in reader mode and verify if the setting is retained across each article gestures verify pinch to zoom in out gestures work on verify that navigating to a different origin resets the zoom swipe back and forward to navigate verify this works as expected password managers verify tapping on on the slide out keyboard launches app and you re able to select the stored credentials verify tapping on bitwarden password manager in the autofill field launches the app and auto fills the stored data shields settings enable all switches in settings and visit a site and disable block scripts kill and relaunch the app and verify if the site shields settings are retained browser lock verify browser lock enables device passcode to lock unlock the browser verify swipe up swipe down with browser in focus doesn t ask for device passcode verify cancel passcode shows unlock message remove the app from memory and relaunch enter a wrong passcode the browser should not be unlocked verify cancel fingerprint confirmation face unlock shows enter passcode window when fingerprint face unlock is set up on the device brave rewards ads verify wallet is auto created after enabling rewards verify when you click on the brave rewards panel while on a site the panel displays if the site is verified or not verify ads are only shown while the app is being used verify tapping on an ad notification shows the landing page verify inline content ads show on brave news and are correctly redeemed from the server verify view clicked landed and dismiss states are logged based on the action sync verify you are able to join sync chain by scanning the qr code verify you are able to join sync chain using code words verify you are able to create a sync chain on the device and add other devices to the chain via qr code code words verify that bookmarks from other devices on the chain show up on the mobile device after sync completes verify newly created bookmarks get synced to all devices on the sync chain verify existing bookmarks before joining sync chain also gets sync d to all devices on the sync chain verify history sync is disabled by default enabling it starts sync ing history items as well limited to entries verify sync works on an upgraded profile and new bookmarks added post upgrade sync across devices on the chain verify you can create a standalone sync chain with one device playlist autoplay verify playlist is enabled by default on a clean install verify when a site with audio video is detected playlist button is shown in url bar and a notification badge on the menu verify clicking the playlist button adds audio video to playlist and initiates offline play verify able to add audio video to playlist via share menu verify able to add audio video to playlist via long press context menu verify autoplay is disabled by default verify with autoplay enabled able to play audio video automatically verify with background audio enabled able to play audio video with device locked or app in background brave news verify brave news opt in is shown on a clean install verify once opted in brave news starts showing articles verify inline content ads are shown on brave news verify able to add custom rss feed to the list of sources verify able to disable content from a particular publisher by long press on the card bravery settings check that block ad and unblock ad works on check that toggling to blocking and allow ads works as expected verify that tapping through a cert error in works enable block scripts globally from settings visit nothing should load tap on shields and disable block scripts the page should load properly verify that default brave shields settings take effect on pages with no site settings fingerprint tests verify that turning on fingerprinting protection in preferences shows fingerprints blocked at verify that turning it off in the bravery menu shows fingerprints blocked verify that the audio fingerprint is blocked at when fingerprinting protection is on verify that doesn t leak ip address when block all fingerprinting protection is on content tests go to and make sure that the password can be saved make sure the saved password is auto populated when you visit the site again open an email on and tap on a link make sure it works verify that pdf is loaded over https at verify that pdf is loaded over http at verify that shows up as grey not red no mixed content scripts are run verify that search results from open in a new tab due to target being blank webauthn verify browser prompts for security key when trying to log in to accounts verify you are able to successfully authenticate an account using security key when prompted verify quickly connecting and disconnecting doesn t cause issues authenticating using security keys verify you are able to use security keys on both normal and private tabs verify you are able to use security keys to authenticate using nfc app linker long press on a link in the twitter app to get the share picker choose brave verify brave doesn t crash after opening the link session storage verify that tabs restore when closed including active tab
| 0
|
13,697
| 16,455,144,759
|
IssuesEvent
|
2021-05-21 11:29:31
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
opened
|
Empty statement results causes reducer error, which causes the entire query to fail
|
.Regression Database/SQLServer Priority:P2 Querying/Processor Type:Bug
|
**Describe the bug**
While multi-statements are officially not supported #4050, 0.39.0 introduced an upgraded driver #15287, which now causes empty statement results to produce an error like `Select statement did not produce a ResultSet for native query`, which prevent the entire query from working.
This is has been reported for SQL Server, but might be a more general problem.
It might be possible to rewrite some queries to [CTEs](https://www.metabase.com/learn/building-analytics/sql-templates/sql-cte.html), but for some queries this might not be feasible or require a lot of work.
**To Reproduce**
1. Native query > SQL Server:
```
drop table if exists #myTemporaryTable;
select 1;
```
2. Errors with `Select statement did not produce a ResultSet for native query`

<details><summary>Full stacktrace</summary>
```
2021-05-21 13:26:33,775 ERROR middleware.catch-exceptions :: Error processing query: null
{:database_id 20,
:started_at #t "2021-05-21T13:26:31.148017+02:00[Europe/Copenhagen]",
:via
[{:status :failed,
:class clojure.lang.ExceptionInfo,
:error "Error executing query",
:stacktrace
["--> driver.sql_jdbc.execute$execute_reducible_query$fn__79533.invoke(execute.clj:470)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:467)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:462)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)"
"driver.sql_jdbc$fn__80985.invokeStatic(sql_jdbc.clj:54)"
"driver.sql_jdbc$fn__80985.invoke(sql_jdbc.clj:52)"
"query_processor.context$executef.invokeStatic(context.clj:59)"
"query_processor.context$executef.invoke(context.clj:48)"
"query_processor.context.default$default_runf.invokeStatic(default.clj:68)"
"query_processor.context.default$default_runf.invoke(default.clj:66)"
"query_processor.context$runf.invokeStatic(context.clj:45)"
"query_processor.context$runf.invoke(context.clj:39)"
"query_processor.reducible$pivot.invokeStatic(reducible.clj:34)"
"query_processor.reducible$pivot.invoke(reducible.clj:31)"
"query_processor.middleware.mbql_to_native$mbql__GT_native$fn__47305.invoke(mbql_to_native.clj:25)"
"query_processor.middleware.check_features$check_features$fn__46547.invoke(check_features.clj:39)"
"query_processor.middleware.limit$limit$fn__47291.invoke(limit.clj:37)"
"query_processor.middleware.cache$maybe_return_cached_results$fn__45999.invoke(cache.clj:211)"
"query_processor.middleware.optimize_temporal_filters$optimize_temporal_filters$fn__47551.invoke(optimize_temporal_filters.clj:204)"
"query_processor.middleware.validate_temporal_bucketing$validate_temporal_bucketing$fn__49481.invoke(validate_temporal_bucketing.clj:50)"
"query_processor.middleware.auto_parse_filter_values$auto_parse_filter_values$fn__45118.invoke(auto_parse_filter_values.clj:43)"
"query_processor.middleware.wrap_value_literals$wrap_value_literals$fn__41435.invoke(wrap_value_literals.clj:161)"
"query_processor.middleware.annotate$add_column_info$fn__41310.invoke(annotate.clj:605)"
"query_processor.middleware.permissions$check_query_permissions$fn__46419.invoke(permissions.clj:81)"
"query_processor.middleware.pre_alias_aggregations$pre_alias_aggregations$fn__48409.invoke(pre_alias_aggregations.clj:40)"
"query_processor.middleware.cumulative_aggregations$handle_cumulative_aggregations$fn__46620.invoke(cumulative_aggregations.clj:60)"
"query_processor.middleware.resolve_joined_fields$resolve_joined_fields$fn__48708.invoke(resolve_joined_fields.clj:102)"
"query_processor.middleware.resolve_joins$resolve_joins$fn__49021.invoke(resolve_joins.clj:171)"
"query_processor.middleware.add_implicit_joins$add_implicit_joins$fn__44694.invoke(add_implicit_joins.clj:190)"
"query_processor.middleware.large_int_id$convert_id_to_string$fn__47255.invoke(large_int_id.clj:59)"
"query_processor.middleware.format_rows$format_rows$fn__47236.invoke(format_rows.clj:74)"
"query_processor.middleware.add_default_temporal_unit$add_default_temporal_unit$fn__43988.invoke(add_default_temporal_unit.clj:23)"
"query_processor.middleware.desugar$desugar$fn__46686.invoke(desugar.clj:21)"
"query_processor.middleware.binning$update_binning_strategy$fn__45505.invoke(binning.clj:227)"
"query_processor.middleware.resolve_fields$resolve_fields$fn__46222.invoke(resolve_fields.clj:34)"
"query_processor.middleware.add_dimension_projections$add_remapping$fn__44343.invoke(add_dimension_projections.clj:312)"
"query_processor.middleware.add_implicit_clauses$add_implicit_clauses$fn__44572.invoke(add_implicit_clauses.clj:147)"
"query_processor.middleware.upgrade_field_literals$upgrade_field_literals$fn__49430.invoke(upgrade_field_literals.clj:40)"
"query_processor.middleware.add_source_metadata$add_source_metadata_for_source_queries$fn__44857.invoke(add_source_metadata.clj:123)"
"query_processor.middleware.reconcile_breakout_and_order_by_bucketing$reconcile_breakout_and_order_by_bucketing$fn__48583.invoke(reconcile_breakout_and_order_by_bucketing.clj:100)"
"query_processor.middleware.auto_bucket_datetimes$auto_bucket_datetimes$fn__45065.invoke(auto_bucket_datetimes.clj:147)"
"query_processor.middleware.resolve_source_table$resolve_source_tables$fn__46269.invoke(resolve_source_table.clj:45)"
"query_processor.middleware.parameters$substitute_parameters$fn__48391.invoke(parameters.clj:111)"
"query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__46321.invoke(resolve_referenced.clj:79)"
"query_processor.middleware.expand_macros$expand_macros$fn__46942.invoke(expand_macros.clj:155)"
"query_processor.middleware.add_timezone_info$add_timezone_info$fn__44866.invoke(add_timezone_info.clj:15)"
"query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__49383.invoke(splice_params_in_response.clj:32)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594$fn__48598.invoke(resolve_database_and_driver.clj:31)"
"driver$do_with_driver.invokeStatic(driver.clj:60)"
"driver$do_with_driver.invoke(driver.clj:56)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594.invoke(resolve_database_and_driver.clj:25)"
"query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__47182.invoke(fetch_source_query.clj:274)"
"query_processor.middleware.store$initialize_store$fn__49392$fn__49393.invoke(store.clj:11)"
"query_processor.store$do_with_store.invokeStatic(store.clj:44)"
"query_processor.store$do_with_store.invoke(store.clj:38)"
"query_processor.middleware.store$initialize_store$fn__49392.invoke(store.clj:10)"
"query_processor.middleware.validate$validate_query$fn__49437.invoke(validate.clj:10)"
"query_processor.middleware.normalize_query$normalize$fn__47318.invoke(normalize_query.clj:22)"
"query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__44712.invoke(add_rows_truncated.clj:35)"
"query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49368.invoke(results_metadata.clj:147)"
"query_processor.middleware.constraints$add_default_userland_constraints$fn__46563.invoke(constraints.clj:42)"
"query_processor.middleware.process_userland_query$process_userland_query$fn__48480.invoke(process_userland_query.clj:135)"
"query_processor.middleware.catch_exceptions$catch_exceptions$fn__46503.invoke(catch_exceptions.clj:173)"
"query_processor.reducible$async_qp$qp_STAR___37989$thunk__37990.invoke(reducible.clj:103)"
"query_processor.reducible$async_qp$qp_STAR___37989.invoke(reducible.clj:109)"
"query_processor.reducible$sync_qp$qp_STAR___37998$fn__38001.invoke(reducible.clj:135)"
"query_processor.reducible$sync_qp$qp_STAR___37998.invoke(reducible.clj:134)"
"query_processor$process_userland_query.invokeStatic(query_processor.clj:241)"
"query_processor$process_userland_query.doInvoke(query_processor.clj:237)"
"query_processor$fn__49527$process_query_and_save_execution_BANG___49536$fn__49539.invoke(query_processor.clj:253)"
"query_processor$fn__49527$process_query_and_save_execution_BANG___49536.invoke(query_processor.clj:245)"
"query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580$fn__49583.invoke(query_processor.clj:265)"
"query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580.invoke(query_processor.clj:258)"
"api.dataset$run_query_async$fn__55784.invoke(dataset.clj:56)"
"query_processor.streaming$streaming_response_STAR_$fn__55763$fn__55764.invoke(streaming.clj:72)"
"query_processor.streaming$streaming_response_STAR_$fn__55763.invoke(streaming.clj:71)"
"async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)"
"async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)"
"async.streaming_response$do_f_async$fn__16071.invoke(streaming_response.clj:84)"],
:error_type :invalid-query,
:ex-data
{:sql
"-- Metabase:: userID: 1 queryType: native queryHash: 1b671306840b04531e6a4a4fcc56cd0d6dbfdb671c65eb3c14850b139b358a26\ndrop table if exists #myTemporaryTable;\nselect 1;",
:params nil,
:type :invalid-query}}],
:error_type :invalid-query,
:json_query
{:type "native",
:native {:query "drop table if exists #myTemporaryTable;\nselect 1;", :template-tags {}},
:database 20,
:parameters [],
:middleware {:js-int-to-string? true, :add-default-userland-constraints? true}},
:status :failed,
:class clojure.lang.ExceptionInfo,
:stacktrace
["--> driver.sql_jdbc.execute$fn__79453.invokeStatic(execute.clj:336)"
"driver.sql_jdbc.execute$fn__79453.invoke(execute.clj:332)"
"driver.sql_jdbc.execute$execute_statement_or_prepared_statement_BANG_.invokeStatic(execute.clj:342)"
"driver.sql_jdbc.execute$execute_statement_or_prepared_statement_BANG_.invoke(execute.clj:339)"
"driver.sql_jdbc.execute$execute_reducible_query$fn__79533.invoke(execute.clj:468)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:467)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:462)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)"
"driver.sql_jdbc$fn__80985.invokeStatic(sql_jdbc.clj:54)"
"driver.sql_jdbc$fn__80985.invoke(sql_jdbc.clj:52)"
"query_processor.context$executef.invokeStatic(context.clj:59)"
"query_processor.context$executef.invoke(context.clj:48)"
"query_processor.context.default$default_runf.invokeStatic(default.clj:68)"
"query_processor.context.default$default_runf.invoke(default.clj:66)"
"query_processor.context$runf.invokeStatic(context.clj:45)"
"query_processor.context$runf.invoke(context.clj:39)"
"query_processor.reducible$pivot.invokeStatic(reducible.clj:34)"
"query_processor.reducible$pivot.invoke(reducible.clj:31)"
"query_processor.middleware.mbql_to_native$mbql__GT_native$fn__47305.invoke(mbql_to_native.clj:25)"
"query_processor.middleware.check_features$check_features$fn__46547.invoke(check_features.clj:39)"
"query_processor.middleware.limit$limit$fn__47291.invoke(limit.clj:37)"
"query_processor.middleware.cache$maybe_return_cached_results$fn__45999.invoke(cache.clj:211)"
"query_processor.middleware.optimize_temporal_filters$optimize_temporal_filters$fn__47551.invoke(optimize_temporal_filters.clj:204)"
"query_processor.middleware.validate_temporal_bucketing$validate_temporal_bucketing$fn__49481.invoke(validate_temporal_bucketing.clj:50)"
"query_processor.middleware.auto_parse_filter_values$auto_parse_filter_values$fn__45118.invoke(auto_parse_filter_values.clj:43)"
"query_processor.middleware.wrap_value_literals$wrap_value_literals$fn__41435.invoke(wrap_value_literals.clj:161)"
"query_processor.middleware.annotate$add_column_info$fn__41310.invoke(annotate.clj:605)"
"query_processor.middleware.permissions$check_query_permissions$fn__46419.invoke(permissions.clj:81)"
"query_processor.middleware.pre_alias_aggregations$pre_alias_aggregations$fn__48409.invoke(pre_alias_aggregations.clj:40)"
"query_processor.middleware.cumulative_aggregations$handle_cumulative_aggregations$fn__46620.invoke(cumulative_aggregations.clj:60)"
"query_processor.middleware.resolve_joined_fields$resolve_joined_fields$fn__48708.invoke(resolve_joined_fields.clj:102)"
"query_processor.middleware.resolve_joins$resolve_joins$fn__49021.invoke(resolve_joins.clj:171)"
"query_processor.middleware.add_implicit_joins$add_implicit_joins$fn__44694.invoke(add_implicit_joins.clj:190)"
"query_processor.middleware.large_int_id$convert_id_to_string$fn__47255.invoke(large_int_id.clj:59)"
"query_processor.middleware.format_rows$format_rows$fn__47236.invoke(format_rows.clj:74)"
"query_processor.middleware.add_default_temporal_unit$add_default_temporal_unit$fn__43988.invoke(add_default_temporal_unit.clj:23)"
"query_processor.middleware.desugar$desugar$fn__46686.invoke(desugar.clj:21)"
"query_processor.middleware.binning$update_binning_strategy$fn__45505.invoke(binning.clj:227)"
"query_processor.middleware.resolve_fields$resolve_fields$fn__46222.invoke(resolve_fields.clj:34)"
"query_processor.middleware.add_dimension_projections$add_remapping$fn__44343.invoke(add_dimension_projections.clj:312)"
"query_processor.middleware.add_implicit_clauses$add_implicit_clauses$fn__44572.invoke(add_implicit_clauses.clj:147)"
"query_processor.middleware.upgrade_field_literals$upgrade_field_literals$fn__49430.invoke(upgrade_field_literals.clj:40)"
"query_processor.middleware.add_source_metadata$add_source_metadata_for_source_queries$fn__44857.invoke(add_source_metadata.clj:123)"
"query_processor.middleware.reconcile_breakout_and_order_by_bucketing$reconcile_breakout_and_order_by_bucketing$fn__48583.invoke(reconcile_breakout_and_order_by_bucketing.clj:100)"
"query_processor.middleware.auto_bucket_datetimes$auto_bucket_datetimes$fn__45065.invoke(auto_bucket_datetimes.clj:147)"
"query_processor.middleware.resolve_source_table$resolve_source_tables$fn__46269.invoke(resolve_source_table.clj:45)"
"query_processor.middleware.parameters$substitute_parameters$fn__48391.invoke(parameters.clj:111)"
"query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__46321.invoke(resolve_referenced.clj:79)"
"query_processor.middleware.expand_macros$expand_macros$fn__46942.invoke(expand_macros.clj:155)"
"query_processor.middleware.add_timezone_info$add_timezone_info$fn__44866.invoke(add_timezone_info.clj:15)"
"query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__49383.invoke(splice_params_in_response.clj:32)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594$fn__48598.invoke(resolve_database_and_driver.clj:31)"
"driver$do_with_driver.invokeStatic(driver.clj:60)"
"driver$do_with_driver.invoke(driver.clj:56)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594.invoke(resolve_database_and_driver.clj:25)"
"query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__47182.invoke(fetch_source_query.clj:274)"
"query_processor.middleware.store$initialize_store$fn__49392$fn__49393.invoke(store.clj:11)"
"query_processor.store$do_with_store.invokeStatic(store.clj:44)"
"query_processor.store$do_with_store.invoke(store.clj:38)"
"query_processor.middleware.store$initialize_store$fn__49392.invoke(store.clj:10)"
"query_processor.middleware.validate$validate_query$fn__49437.invoke(validate.clj:10)"
"query_processor.middleware.normalize_query$normalize$fn__47318.invoke(normalize_query.clj:22)"
"query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__44712.invoke(add_rows_truncated.clj:35)"
"query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49368.invoke(results_metadata.clj:147)"
"query_processor.middleware.constraints$add_default_userland_constraints$fn__46563.invoke(constraints.clj:42)"
"query_processor.middleware.process_userland_query$process_userland_query$fn__48480.invoke(process_userland_query.clj:135)"
"query_processor.middleware.catch_exceptions$catch_exceptions$fn__46503.invoke(catch_exceptions.clj:173)"
"query_processor.reducible$async_qp$qp_STAR___37989$thunk__37990.invoke(reducible.clj:103)"
"query_processor.reducible$async_qp$qp_STAR___37989.invoke(reducible.clj:109)"
"query_processor.reducible$sync_qp$qp_STAR___37998$fn__38001.invoke(reducible.clj:135)"
"query_processor.reducible$sync_qp$qp_STAR___37998.invoke(reducible.clj:134)"
"query_processor$process_userland_query.invokeStatic(query_processor.clj:241)"
"query_processor$process_userland_query.doInvoke(query_processor.clj:237)"
"query_processor$fn__49527$process_query_and_save_execution_BANG___49536$fn__49539.invoke(query_processor.clj:253)"
"query_processor$fn__49527$process_query_and_save_execution_BANG___49536.invoke(query_processor.clj:245)"
"query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580$fn__49583.invoke(query_processor.clj:265)"
"query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580.invoke(query_processor.clj:258)"
"api.dataset$run_query_async$fn__55784.invoke(dataset.clj:56)"
"query_processor.streaming$streaming_response_STAR_$fn__55763$fn__55764.invoke(streaming.clj:72)"
"query_processor.streaming$streaming_response_STAR_$fn__55763.invoke(streaming.clj:71)"
"async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)"
"async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)"
"async.streaming_response$do_f_async$fn__16071.invoke(streaming_response.clj:84)"],
:context :ad-hoc,
:error "Select statement did not produce a ResultSet for native query",
:row_count 0,
:running_time 0,
:ex-data
{:sql
"-- Metabase:: userID: 1 queryType: native queryHash: 1b671306840b04531e6a4a4fcc56cd0d6dbfdb671c65eb3c14850b139b358a26\ndrop table if exists #myTemporaryTable;\nselect 1;",
:driver :sqlserver},
:data {:rows [], :cols []}}
2021-05-21 13:26:33,823 DEBUG middleware.log :: POST /api/dataset 202 [ASYNC: completed] 2.7 s (6 DB calls) App DB connections: 0/7 Jetty threads: 3/50 (4 idle, 0 queued) (50 total active threads) Queries in flight: 1 (0 queued)
```
</details>
**Expected behavior**
The first statement to be executed and empty resultset to be ignored, and then run second statement and show result from that.
This is 0.38.4:

**Information about your Metabase Installation:**
Regression since 0.39.0 - tested 0.38.4 thru 0.39.2
**Additional context**
https://discourse.metabase.com/t/sql-queries-broken-in-0-39-0-1/14468
|
1.0
|
Empty statement results causes reducer error, which causes the entire query to fail - **Describe the bug**
While multi-statements are officially not supported #4050, 0.39.0 introduced an upgraded driver #15287, which now causes empty statement results to produce an error like `Select statement did not produce a ResultSet for native query`, which prevent the entire query from working.
This is has been reported for SQL Server, but might be a more general problem.
It might be possible to rewrite some queries to [CTEs](https://www.metabase.com/learn/building-analytics/sql-templates/sql-cte.html), but for some queries this might not be feasible or require a lot of work.
**To Reproduce**
1. Native query > SQL Server:
```
drop table if exists #myTemporaryTable;
select 1;
```
2. Errors with `Select statement did not produce a ResultSet for native query`

<details><summary>Full stacktrace</summary>
```
2021-05-21 13:26:33,775 ERROR middleware.catch-exceptions :: Error processing query: null
{:database_id 20,
:started_at #t "2021-05-21T13:26:31.148017+02:00[Europe/Copenhagen]",
:via
[{:status :failed,
:class clojure.lang.ExceptionInfo,
:error "Error executing query",
:stacktrace
["--> driver.sql_jdbc.execute$execute_reducible_query$fn__79533.invoke(execute.clj:470)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:467)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:462)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)"
"driver.sql_jdbc$fn__80985.invokeStatic(sql_jdbc.clj:54)"
"driver.sql_jdbc$fn__80985.invoke(sql_jdbc.clj:52)"
"query_processor.context$executef.invokeStatic(context.clj:59)"
"query_processor.context$executef.invoke(context.clj:48)"
"query_processor.context.default$default_runf.invokeStatic(default.clj:68)"
"query_processor.context.default$default_runf.invoke(default.clj:66)"
"query_processor.context$runf.invokeStatic(context.clj:45)"
"query_processor.context$runf.invoke(context.clj:39)"
"query_processor.reducible$pivot.invokeStatic(reducible.clj:34)"
"query_processor.reducible$pivot.invoke(reducible.clj:31)"
"query_processor.middleware.mbql_to_native$mbql__GT_native$fn__47305.invoke(mbql_to_native.clj:25)"
"query_processor.middleware.check_features$check_features$fn__46547.invoke(check_features.clj:39)"
"query_processor.middleware.limit$limit$fn__47291.invoke(limit.clj:37)"
"query_processor.middleware.cache$maybe_return_cached_results$fn__45999.invoke(cache.clj:211)"
"query_processor.middleware.optimize_temporal_filters$optimize_temporal_filters$fn__47551.invoke(optimize_temporal_filters.clj:204)"
"query_processor.middleware.validate_temporal_bucketing$validate_temporal_bucketing$fn__49481.invoke(validate_temporal_bucketing.clj:50)"
"query_processor.middleware.auto_parse_filter_values$auto_parse_filter_values$fn__45118.invoke(auto_parse_filter_values.clj:43)"
"query_processor.middleware.wrap_value_literals$wrap_value_literals$fn__41435.invoke(wrap_value_literals.clj:161)"
"query_processor.middleware.annotate$add_column_info$fn__41310.invoke(annotate.clj:605)"
"query_processor.middleware.permissions$check_query_permissions$fn__46419.invoke(permissions.clj:81)"
"query_processor.middleware.pre_alias_aggregations$pre_alias_aggregations$fn__48409.invoke(pre_alias_aggregations.clj:40)"
"query_processor.middleware.cumulative_aggregations$handle_cumulative_aggregations$fn__46620.invoke(cumulative_aggregations.clj:60)"
"query_processor.middleware.resolve_joined_fields$resolve_joined_fields$fn__48708.invoke(resolve_joined_fields.clj:102)"
"query_processor.middleware.resolve_joins$resolve_joins$fn__49021.invoke(resolve_joins.clj:171)"
"query_processor.middleware.add_implicit_joins$add_implicit_joins$fn__44694.invoke(add_implicit_joins.clj:190)"
"query_processor.middleware.large_int_id$convert_id_to_string$fn__47255.invoke(large_int_id.clj:59)"
"query_processor.middleware.format_rows$format_rows$fn__47236.invoke(format_rows.clj:74)"
"query_processor.middleware.add_default_temporal_unit$add_default_temporal_unit$fn__43988.invoke(add_default_temporal_unit.clj:23)"
"query_processor.middleware.desugar$desugar$fn__46686.invoke(desugar.clj:21)"
"query_processor.middleware.binning$update_binning_strategy$fn__45505.invoke(binning.clj:227)"
"query_processor.middleware.resolve_fields$resolve_fields$fn__46222.invoke(resolve_fields.clj:34)"
"query_processor.middleware.add_dimension_projections$add_remapping$fn__44343.invoke(add_dimension_projections.clj:312)"
"query_processor.middleware.add_implicit_clauses$add_implicit_clauses$fn__44572.invoke(add_implicit_clauses.clj:147)"
"query_processor.middleware.upgrade_field_literals$upgrade_field_literals$fn__49430.invoke(upgrade_field_literals.clj:40)"
"query_processor.middleware.add_source_metadata$add_source_metadata_for_source_queries$fn__44857.invoke(add_source_metadata.clj:123)"
"query_processor.middleware.reconcile_breakout_and_order_by_bucketing$reconcile_breakout_and_order_by_bucketing$fn__48583.invoke(reconcile_breakout_and_order_by_bucketing.clj:100)"
"query_processor.middleware.auto_bucket_datetimes$auto_bucket_datetimes$fn__45065.invoke(auto_bucket_datetimes.clj:147)"
"query_processor.middleware.resolve_source_table$resolve_source_tables$fn__46269.invoke(resolve_source_table.clj:45)"
"query_processor.middleware.parameters$substitute_parameters$fn__48391.invoke(parameters.clj:111)"
"query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__46321.invoke(resolve_referenced.clj:79)"
"query_processor.middleware.expand_macros$expand_macros$fn__46942.invoke(expand_macros.clj:155)"
"query_processor.middleware.add_timezone_info$add_timezone_info$fn__44866.invoke(add_timezone_info.clj:15)"
"query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__49383.invoke(splice_params_in_response.clj:32)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594$fn__48598.invoke(resolve_database_and_driver.clj:31)"
"driver$do_with_driver.invokeStatic(driver.clj:60)"
"driver$do_with_driver.invoke(driver.clj:56)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594.invoke(resolve_database_and_driver.clj:25)"
"query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__47182.invoke(fetch_source_query.clj:274)"
"query_processor.middleware.store$initialize_store$fn__49392$fn__49393.invoke(store.clj:11)"
"query_processor.store$do_with_store.invokeStatic(store.clj:44)"
"query_processor.store$do_with_store.invoke(store.clj:38)"
"query_processor.middleware.store$initialize_store$fn__49392.invoke(store.clj:10)"
"query_processor.middleware.validate$validate_query$fn__49437.invoke(validate.clj:10)"
"query_processor.middleware.normalize_query$normalize$fn__47318.invoke(normalize_query.clj:22)"
"query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__44712.invoke(add_rows_truncated.clj:35)"
"query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49368.invoke(results_metadata.clj:147)"
"query_processor.middleware.constraints$add_default_userland_constraints$fn__46563.invoke(constraints.clj:42)"
"query_processor.middleware.process_userland_query$process_userland_query$fn__48480.invoke(process_userland_query.clj:135)"
"query_processor.middleware.catch_exceptions$catch_exceptions$fn__46503.invoke(catch_exceptions.clj:173)"
"query_processor.reducible$async_qp$qp_STAR___37989$thunk__37990.invoke(reducible.clj:103)"
"query_processor.reducible$async_qp$qp_STAR___37989.invoke(reducible.clj:109)"
"query_processor.reducible$sync_qp$qp_STAR___37998$fn__38001.invoke(reducible.clj:135)"
"query_processor.reducible$sync_qp$qp_STAR___37998.invoke(reducible.clj:134)"
"query_processor$process_userland_query.invokeStatic(query_processor.clj:241)"
"query_processor$process_userland_query.doInvoke(query_processor.clj:237)"
"query_processor$fn__49527$process_query_and_save_execution_BANG___49536$fn__49539.invoke(query_processor.clj:253)"
"query_processor$fn__49527$process_query_and_save_execution_BANG___49536.invoke(query_processor.clj:245)"
"query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580$fn__49583.invoke(query_processor.clj:265)"
"query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580.invoke(query_processor.clj:258)"
"api.dataset$run_query_async$fn__55784.invoke(dataset.clj:56)"
"query_processor.streaming$streaming_response_STAR_$fn__55763$fn__55764.invoke(streaming.clj:72)"
"query_processor.streaming$streaming_response_STAR_$fn__55763.invoke(streaming.clj:71)"
"async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)"
"async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)"
"async.streaming_response$do_f_async$fn__16071.invoke(streaming_response.clj:84)"],
:error_type :invalid-query,
:ex-data
{:sql
"-- Metabase:: userID: 1 queryType: native queryHash: 1b671306840b04531e6a4a4fcc56cd0d6dbfdb671c65eb3c14850b139b358a26\ndrop table if exists #myTemporaryTable;\nselect 1;",
:params nil,
:type :invalid-query}}],
:error_type :invalid-query,
:json_query
{:type "native",
:native {:query "drop table if exists #myTemporaryTable;\nselect 1;", :template-tags {}},
:database 20,
:parameters [],
:middleware {:js-int-to-string? true, :add-default-userland-constraints? true}},
:status :failed,
:class clojure.lang.ExceptionInfo,
:stacktrace
["--> driver.sql_jdbc.execute$fn__79453.invokeStatic(execute.clj:336)"
"driver.sql_jdbc.execute$fn__79453.invoke(execute.clj:332)"
"driver.sql_jdbc.execute$execute_statement_or_prepared_statement_BANG_.invokeStatic(execute.clj:342)"
"driver.sql_jdbc.execute$execute_statement_or_prepared_statement_BANG_.invoke(execute.clj:339)"
"driver.sql_jdbc.execute$execute_reducible_query$fn__79533.invoke(execute.clj:468)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:467)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)"
"driver.sql_jdbc.execute$execute_reducible_query.invokeStatic(execute.clj:462)"
"driver.sql_jdbc.execute$execute_reducible_query.invoke(execute.clj:453)"
"driver.sql_jdbc$fn__80985.invokeStatic(sql_jdbc.clj:54)"
"driver.sql_jdbc$fn__80985.invoke(sql_jdbc.clj:52)"
"query_processor.context$executef.invokeStatic(context.clj:59)"
"query_processor.context$executef.invoke(context.clj:48)"
"query_processor.context.default$default_runf.invokeStatic(default.clj:68)"
"query_processor.context.default$default_runf.invoke(default.clj:66)"
"query_processor.context$runf.invokeStatic(context.clj:45)"
"query_processor.context$runf.invoke(context.clj:39)"
"query_processor.reducible$pivot.invokeStatic(reducible.clj:34)"
"query_processor.reducible$pivot.invoke(reducible.clj:31)"
"query_processor.middleware.mbql_to_native$mbql__GT_native$fn__47305.invoke(mbql_to_native.clj:25)"
"query_processor.middleware.check_features$check_features$fn__46547.invoke(check_features.clj:39)"
"query_processor.middleware.limit$limit$fn__47291.invoke(limit.clj:37)"
"query_processor.middleware.cache$maybe_return_cached_results$fn__45999.invoke(cache.clj:211)"
"query_processor.middleware.optimize_temporal_filters$optimize_temporal_filters$fn__47551.invoke(optimize_temporal_filters.clj:204)"
"query_processor.middleware.validate_temporal_bucketing$validate_temporal_bucketing$fn__49481.invoke(validate_temporal_bucketing.clj:50)"
"query_processor.middleware.auto_parse_filter_values$auto_parse_filter_values$fn__45118.invoke(auto_parse_filter_values.clj:43)"
"query_processor.middleware.wrap_value_literals$wrap_value_literals$fn__41435.invoke(wrap_value_literals.clj:161)"
"query_processor.middleware.annotate$add_column_info$fn__41310.invoke(annotate.clj:605)"
"query_processor.middleware.permissions$check_query_permissions$fn__46419.invoke(permissions.clj:81)"
"query_processor.middleware.pre_alias_aggregations$pre_alias_aggregations$fn__48409.invoke(pre_alias_aggregations.clj:40)"
"query_processor.middleware.cumulative_aggregations$handle_cumulative_aggregations$fn__46620.invoke(cumulative_aggregations.clj:60)"
"query_processor.middleware.resolve_joined_fields$resolve_joined_fields$fn__48708.invoke(resolve_joined_fields.clj:102)"
"query_processor.middleware.resolve_joins$resolve_joins$fn__49021.invoke(resolve_joins.clj:171)"
"query_processor.middleware.add_implicit_joins$add_implicit_joins$fn__44694.invoke(add_implicit_joins.clj:190)"
"query_processor.middleware.large_int_id$convert_id_to_string$fn__47255.invoke(large_int_id.clj:59)"
"query_processor.middleware.format_rows$format_rows$fn__47236.invoke(format_rows.clj:74)"
"query_processor.middleware.add_default_temporal_unit$add_default_temporal_unit$fn__43988.invoke(add_default_temporal_unit.clj:23)"
"query_processor.middleware.desugar$desugar$fn__46686.invoke(desugar.clj:21)"
"query_processor.middleware.binning$update_binning_strategy$fn__45505.invoke(binning.clj:227)"
"query_processor.middleware.resolve_fields$resolve_fields$fn__46222.invoke(resolve_fields.clj:34)"
"query_processor.middleware.add_dimension_projections$add_remapping$fn__44343.invoke(add_dimension_projections.clj:312)"
"query_processor.middleware.add_implicit_clauses$add_implicit_clauses$fn__44572.invoke(add_implicit_clauses.clj:147)"
"query_processor.middleware.upgrade_field_literals$upgrade_field_literals$fn__49430.invoke(upgrade_field_literals.clj:40)"
"query_processor.middleware.add_source_metadata$add_source_metadata_for_source_queries$fn__44857.invoke(add_source_metadata.clj:123)"
"query_processor.middleware.reconcile_breakout_and_order_by_bucketing$reconcile_breakout_and_order_by_bucketing$fn__48583.invoke(reconcile_breakout_and_order_by_bucketing.clj:100)"
"query_processor.middleware.auto_bucket_datetimes$auto_bucket_datetimes$fn__45065.invoke(auto_bucket_datetimes.clj:147)"
"query_processor.middleware.resolve_source_table$resolve_source_tables$fn__46269.invoke(resolve_source_table.clj:45)"
"query_processor.middleware.parameters$substitute_parameters$fn__48391.invoke(parameters.clj:111)"
"query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__46321.invoke(resolve_referenced.clj:79)"
"query_processor.middleware.expand_macros$expand_macros$fn__46942.invoke(expand_macros.clj:155)"
"query_processor.middleware.add_timezone_info$add_timezone_info$fn__44866.invoke(add_timezone_info.clj:15)"
"query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__49383.invoke(splice_params_in_response.clj:32)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594$fn__48598.invoke(resolve_database_and_driver.clj:31)"
"driver$do_with_driver.invokeStatic(driver.clj:60)"
"driver$do_with_driver.invoke(driver.clj:56)"
"query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__48594.invoke(resolve_database_and_driver.clj:25)"
"query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__47182.invoke(fetch_source_query.clj:274)"
"query_processor.middleware.store$initialize_store$fn__49392$fn__49393.invoke(store.clj:11)"
"query_processor.store$do_with_store.invokeStatic(store.clj:44)"
"query_processor.store$do_with_store.invoke(store.clj:38)"
"query_processor.middleware.store$initialize_store$fn__49392.invoke(store.clj:10)"
"query_processor.middleware.validate$validate_query$fn__49437.invoke(validate.clj:10)"
"query_processor.middleware.normalize_query$normalize$fn__47318.invoke(normalize_query.clj:22)"
"query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__44712.invoke(add_rows_truncated.clj:35)"
"query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49368.invoke(results_metadata.clj:147)"
"query_processor.middleware.constraints$add_default_userland_constraints$fn__46563.invoke(constraints.clj:42)"
"query_processor.middleware.process_userland_query$process_userland_query$fn__48480.invoke(process_userland_query.clj:135)"
"query_processor.middleware.catch_exceptions$catch_exceptions$fn__46503.invoke(catch_exceptions.clj:173)"
"query_processor.reducible$async_qp$qp_STAR___37989$thunk__37990.invoke(reducible.clj:103)"
"query_processor.reducible$async_qp$qp_STAR___37989.invoke(reducible.clj:109)"
"query_processor.reducible$sync_qp$qp_STAR___37998$fn__38001.invoke(reducible.clj:135)"
"query_processor.reducible$sync_qp$qp_STAR___37998.invoke(reducible.clj:134)"
"query_processor$process_userland_query.invokeStatic(query_processor.clj:241)"
"query_processor$process_userland_query.doInvoke(query_processor.clj:237)"
"query_processor$fn__49527$process_query_and_save_execution_BANG___49536$fn__49539.invoke(query_processor.clj:253)"
"query_processor$fn__49527$process_query_and_save_execution_BANG___49536.invoke(query_processor.clj:245)"
"query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580$fn__49583.invoke(query_processor.clj:265)"
"query_processor$fn__49571$process_query_and_save_with_max_results_constraints_BANG___49580.invoke(query_processor.clj:258)"
"api.dataset$run_query_async$fn__55784.invoke(dataset.clj:56)"
"query_processor.streaming$streaming_response_STAR_$fn__55763$fn__55764.invoke(streaming.clj:72)"
"query_processor.streaming$streaming_response_STAR_$fn__55763.invoke(streaming.clj:71)"
"async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)"
"async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)"
"async.streaming_response$do_f_async$fn__16071.invoke(streaming_response.clj:84)"],
:context :ad-hoc,
:error "Select statement did not produce a ResultSet for native query",
:row_count 0,
:running_time 0,
:ex-data
{:sql
"-- Metabase:: userID: 1 queryType: native queryHash: 1b671306840b04531e6a4a4fcc56cd0d6dbfdb671c65eb3c14850b139b358a26\ndrop table if exists #myTemporaryTable;\nselect 1;",
:driver :sqlserver},
:data {:rows [], :cols []}}
2021-05-21 13:26:33,823 DEBUG middleware.log :: POST /api/dataset 202 [ASYNC: completed] 2.7 s (6 DB calls) App DB connections: 0/7 Jetty threads: 3/50 (4 idle, 0 queued) (50 total active threads) Queries in flight: 1 (0 queued)
```
</details>
**Expected behavior**
The first statement to be executed and empty resultset to be ignored, and then run second statement and show result from that.
This is 0.38.4:

**Information about your Metabase Installation:**
Regression since 0.39.0 - tested 0.38.4 thru 0.39.2
**Additional context**
https://discourse.metabase.com/t/sql-queries-broken-in-0-39-0-1/14468
|
process
|
empty statement results causes reducer error which causes the entire query to fail describe the bug while multi statements are officially not supported introduced an upgraded driver which now causes empty statement results to produce an error like select statement did not produce a resultset for native query which prevent the entire query from working this is has been reported for sql server but might be a more general problem it might be possible to rewrite some queries to but for some queries this might not be feasible or require a lot of work to reproduce native query sql server drop table if exists mytemporarytable select errors with select statement did not produce a resultset for native query full stacktrace error middleware catch exceptions error processing query null database id started at t via status failed class clojure lang exceptioninfo error error executing query stacktrace driver sql jdbc execute execute reducible query fn invoke execute clj driver sql jdbc execute execute reducible query invokestatic execute clj driver sql jdbc execute execute reducible query invoke execute clj driver sql jdbc execute execute reducible query invokestatic execute clj driver sql jdbc execute execute reducible query invoke execute clj driver sql jdbc fn invokestatic sql jdbc clj driver sql jdbc fn invoke sql jdbc clj query processor context executef invokestatic context clj query processor context executef invoke context clj query processor context default default runf invokestatic default clj query processor context default default runf invoke default clj query processor context runf invokestatic context clj query processor context runf invoke context clj query processor reducible pivot invokestatic reducible clj query processor reducible pivot invoke reducible clj query processor middleware mbql to native mbql gt native fn invoke mbql to native clj query processor middleware check features check features fn invoke check features clj query processor middleware limit limit fn invoke limit clj query processor middleware cache maybe return cached results fn invoke cache clj query processor middleware optimize temporal filters optimize temporal filters fn invoke optimize temporal filters clj query processor middleware validate temporal bucketing validate temporal bucketing fn invoke validate temporal bucketing clj query processor middleware auto parse filter values auto parse filter values fn invoke auto parse filter values clj query processor middleware wrap value literals wrap value literals fn invoke wrap value literals clj query processor middleware annotate add column info fn invoke annotate clj query processor middleware permissions check query permissions fn invoke permissions clj query processor middleware pre alias aggregations pre alias aggregations fn invoke pre alias aggregations clj query processor middleware cumulative aggregations handle cumulative aggregations fn invoke cumulative aggregations clj query processor middleware resolve joined fields resolve joined fields fn invoke resolve joined fields clj query processor middleware resolve joins resolve joins fn invoke resolve joins clj query processor middleware add implicit joins add implicit joins fn invoke add implicit joins clj query processor middleware large int id convert id to string fn invoke large int id clj query processor middleware format rows format rows fn invoke format rows clj query processor middleware add default temporal unit add default temporal unit fn invoke add default temporal unit clj query processor middleware desugar desugar fn invoke desugar clj query processor middleware binning update binning strategy fn invoke binning clj query processor middleware resolve fields resolve fields fn invoke resolve fields clj query processor middleware add dimension projections add remapping fn invoke add dimension projections clj query processor middleware add implicit clauses add implicit clauses fn invoke add implicit clauses clj query processor middleware upgrade field literals upgrade field literals fn invoke upgrade field literals clj query processor middleware add source metadata add source metadata for source queries fn invoke add source metadata clj query processor middleware reconcile breakout and order by bucketing reconcile breakout and order by bucketing fn invoke reconcile breakout and order by bucketing clj query processor middleware auto bucket datetimes auto bucket datetimes fn invoke auto bucket datetimes clj query processor middleware resolve source table resolve source tables fn invoke resolve source table clj query processor middleware parameters substitute parameters fn invoke parameters clj query processor middleware resolve referenced resolve referenced card resources fn invoke resolve referenced clj query processor middleware expand macros expand macros fn invoke expand macros clj query processor middleware add timezone info add timezone info fn invoke add timezone info clj query processor middleware splice params in response splice params in response fn invoke splice params in response clj query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj driver do with driver invokestatic driver clj driver do with driver invoke driver clj query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware validate validate query fn invoke validate clj query processor middleware normalize query normalize fn invoke normalize query clj query processor middleware add rows truncated add rows truncated fn invoke add rows truncated clj query processor middleware results metadata record and return metadata bang fn invoke results metadata clj query processor middleware constraints add default userland constraints fn invoke constraints clj query processor middleware process userland query process userland query fn invoke process userland query clj query processor middleware catch exceptions catch exceptions fn invoke catch exceptions clj query processor reducible async qp qp star thunk invoke reducible clj query processor reducible async qp qp star invoke reducible clj query processor reducible sync qp qp star fn invoke reducible clj query processor reducible sync qp qp star invoke reducible clj query processor process userland query invokestatic query processor clj query processor process userland query doinvoke query processor clj query processor fn process query and save execution bang fn invoke query processor clj query processor fn process query and save execution bang invoke query processor clj query processor fn process query and save with max results constraints bang fn invoke query processor clj query processor fn process query and save with max results constraints bang invoke query processor clj api dataset run query async fn invoke dataset clj query processor streaming streaming response star fn fn invoke streaming clj query processor streaming streaming response star fn invoke streaming clj async streaming response do f star invokestatic streaming response clj async streaming response do f star invoke streaming response clj async streaming response do f async fn invoke streaming response clj error type invalid query ex data sql metabase userid querytype native queryhash ndrop table if exists mytemporarytable nselect params nil type invalid query error type invalid query json query type native native query drop table if exists mytemporarytable nselect template tags database parameters middleware js int to string true add default userland constraints true status failed class clojure lang exceptioninfo stacktrace driver sql jdbc execute fn invokestatic execute clj driver sql jdbc execute fn invoke execute clj driver sql jdbc execute execute statement or prepared statement bang invokestatic execute clj driver sql jdbc execute execute statement or prepared statement bang invoke execute clj driver sql jdbc execute execute reducible query fn invoke execute clj driver sql jdbc execute execute reducible query invokestatic execute clj driver sql jdbc execute execute reducible query invoke execute clj driver sql jdbc execute execute reducible query invokestatic execute clj driver sql jdbc execute execute reducible query invoke execute clj driver sql jdbc fn invokestatic sql jdbc clj driver sql jdbc fn invoke sql jdbc clj query processor context executef invokestatic context clj query processor context executef invoke context clj query processor context default default runf invokestatic default clj query processor context default default runf invoke default clj query processor context runf invokestatic context clj query processor context runf invoke context clj query processor reducible pivot invokestatic reducible clj query processor reducible pivot invoke reducible clj query processor middleware mbql to native mbql gt native fn invoke mbql to native clj query processor middleware check features check features fn invoke check features clj query processor middleware limit limit fn invoke limit clj query processor middleware cache maybe return cached results fn invoke cache clj query processor middleware optimize temporal filters optimize temporal filters fn invoke optimize temporal filters clj query processor middleware validate temporal bucketing validate temporal bucketing fn invoke validate temporal bucketing clj query processor middleware auto parse filter values auto parse filter values fn invoke auto parse filter values clj query processor middleware wrap value literals wrap value literals fn invoke wrap value literals clj query processor middleware annotate add column info fn invoke annotate clj query processor middleware permissions check query permissions fn invoke permissions clj query processor middleware pre alias aggregations pre alias aggregations fn invoke pre alias aggregations clj query processor middleware cumulative aggregations handle cumulative aggregations fn invoke cumulative aggregations clj query processor middleware resolve joined fields resolve joined fields fn invoke resolve joined fields clj query processor middleware resolve joins resolve joins fn invoke resolve joins clj query processor middleware add implicit joins add implicit joins fn invoke add implicit joins clj query processor middleware large int id convert id to string fn invoke large int id clj query processor middleware format rows format rows fn invoke format rows clj query processor middleware add default temporal unit add default temporal unit fn invoke add default temporal unit clj query processor middleware desugar desugar fn invoke desugar clj query processor middleware binning update binning strategy fn invoke binning clj query processor middleware resolve fields resolve fields fn invoke resolve fields clj query processor middleware add dimension projections add remapping fn invoke add dimension projections clj query processor middleware add implicit clauses add implicit clauses fn invoke add implicit clauses clj query processor middleware upgrade field literals upgrade field literals fn invoke upgrade field literals clj query processor middleware add source metadata add source metadata for source queries fn invoke add source metadata clj query processor middleware reconcile breakout and order by bucketing reconcile breakout and order by bucketing fn invoke reconcile breakout and order by bucketing clj query processor middleware auto bucket datetimes auto bucket datetimes fn invoke auto bucket datetimes clj query processor middleware resolve source table resolve source tables fn invoke resolve source table clj query processor middleware parameters substitute parameters fn invoke parameters clj query processor middleware resolve referenced resolve referenced card resources fn invoke resolve referenced clj query processor middleware expand macros expand macros fn invoke expand macros clj query processor middleware add timezone info add timezone info fn invoke add timezone info clj query processor middleware splice params in response splice params in response fn invoke splice params in response clj query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj driver do with driver invokestatic driver clj driver do with driver invoke driver clj query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware validate validate query fn invoke validate clj query processor middleware normalize query normalize fn invoke normalize query clj query processor middleware add rows truncated add rows truncated fn invoke add rows truncated clj query processor middleware results metadata record and return metadata bang fn invoke results metadata clj query processor middleware constraints add default userland constraints fn invoke constraints clj query processor middleware process userland query process userland query fn invoke process userland query clj query processor middleware catch exceptions catch exceptions fn invoke catch exceptions clj query processor reducible async qp qp star thunk invoke reducible clj query processor reducible async qp qp star invoke reducible clj query processor reducible sync qp qp star fn invoke reducible clj query processor reducible sync qp qp star invoke reducible clj query processor process userland query invokestatic query processor clj query processor process userland query doinvoke query processor clj query processor fn process query and save execution bang fn invoke query processor clj query processor fn process query and save execution bang invoke query processor clj query processor fn process query and save with max results constraints bang fn invoke query processor clj query processor fn process query and save with max results constraints bang invoke query processor clj api dataset run query async fn invoke dataset clj query processor streaming streaming response star fn fn invoke streaming clj query processor streaming streaming response star fn invoke streaming clj async streaming response do f star invokestatic streaming response clj async streaming response do f star invoke streaming response clj async streaming response do f async fn invoke streaming response clj context ad hoc error select statement did not produce a resultset for native query row count running time ex data sql metabase userid querytype native queryhash ndrop table if exists mytemporarytable nselect driver sqlserver data rows cols debug middleware log post api dataset s db calls app db connections jetty threads idle queued total active threads queries in flight queued expected behavior the first statement to be executed and empty resultset to be ignored and then run second statement and show result from that this is information about your metabase installation regression since tested thru additional context
| 1
|
18,827
| 24,729,720,012
|
IssuesEvent
|
2022-10-20 16:31:38
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Attributes processor is not identified when regex does not match
|
processor/attributes
|
**Describe the bug**
When a regular expression doesn't match a log line, it fails, but is not identified in the error message, making it hard for the developer to understand which processor entry is failing.
**Steps to reproduce**
Define a pipeline with two different processors such as:
```
attributes/quorum_log_file:
actions:
- key: "file_name"
pattern: ^geth-(?P<host>.*)\.log
action: extract
attributes/tessera_log_file:
actions:
- key: "file_name"
pattern: ^tessera-(?P<host>.*)\.log
action: extract
```
Add those 2 processors to the same pipeline.
Run the pipeline.
**What did you expect to see?**
I expect to see no errors, or a warning that one of the regular expression has not matched, at most.
**What did you see instead?**
I see this error:
```
2021-05-25T17:47:17.107Z ERROR Failed to process entry {"kind": "receiver", "name": "filelog/quorum", "operator_id": "$.regex_parser", "operator_type": "regex_parser", "error": "regex pattern does not match", "action": "send", "entry": {"timestamp":"2021-05-25T17:47:17.107868766Z","severity":0,"attributes":{"file_name":"geth-5b637c3a018d.log"},"body":"DEBUG[05-25|17:47:16.943] Recalculated downloader QoS values rtt=20s confidence=1.000 ttl=1m0s"}}
```
Neither `attributes/quorum_log_file` nor `attributes/tessera_log_file` are named.
**What version did you use?**
Latest
**Environment**
OS: Ubuntu 20.04
**Additional context**
Additionally, I don't think this should error out at all :)
|
1.0
|
Attributes processor is not identified when regex does not match - **Describe the bug**
When a regular expression doesn't match a log line, it fails, but is not identified in the error message, making it hard for the developer to understand which processor entry is failing.
**Steps to reproduce**
Define a pipeline with two different processors such as:
```
attributes/quorum_log_file:
actions:
- key: "file_name"
pattern: ^geth-(?P<host>.*)\.log
action: extract
attributes/tessera_log_file:
actions:
- key: "file_name"
pattern: ^tessera-(?P<host>.*)\.log
action: extract
```
Add those 2 processors to the same pipeline.
Run the pipeline.
**What did you expect to see?**
I expect to see no errors, or a warning that one of the regular expression has not matched, at most.
**What did you see instead?**
I see this error:
```
2021-05-25T17:47:17.107Z ERROR Failed to process entry {"kind": "receiver", "name": "filelog/quorum", "operator_id": "$.regex_parser", "operator_type": "regex_parser", "error": "regex pattern does not match", "action": "send", "entry": {"timestamp":"2021-05-25T17:47:17.107868766Z","severity":0,"attributes":{"file_name":"geth-5b637c3a018d.log"},"body":"DEBUG[05-25|17:47:16.943] Recalculated downloader QoS values rtt=20s confidence=1.000 ttl=1m0s"}}
```
Neither `attributes/quorum_log_file` nor `attributes/tessera_log_file` are named.
**What version did you use?**
Latest
**Environment**
OS: Ubuntu 20.04
**Additional context**
Additionally, I don't think this should error out at all :)
|
process
|
attributes processor is not identified when regex does not match describe the bug when a regular expression doesn t match a log line it fails but is not identified in the error message making it hard for the developer to understand which processor entry is failing steps to reproduce define a pipeline with two different processors such as attributes quorum log file actions key file name pattern geth p log action extract attributes tessera log file actions key file name pattern tessera p log action extract add those processors to the same pipeline run the pipeline what did you expect to see i expect to see no errors or a warning that one of the regular expression has not matched at most what did you see instead i see this error error failed to process entry kind receiver name filelog quorum operator id regex parser operator type regex parser error regex pattern does not match action send entry timestamp severity attributes file name geth log body debug recalculated downloader qos values rtt confidence ttl neither attributes quorum log file nor attributes tessera log file are named what version did you use latest environment os ubuntu additional context additionally i don t think this should error out at all
| 1
|
989
| 3,453,993,088
|
IssuesEvent
|
2015-12-17 14:09:20
|
sysown/proxysql
|
https://api.github.com/repos/sysown/proxysql
|
opened
|
Track hostgroup in stats_mysql_query_digest
|
ADMIN CONNECTION POOL enhancement MYSQL QUERY PROCESSOR ROUTING
|
That is an enhancement.
Currently table stats_mysql_query_digest tracks several metrics, but not hostgroup.
Tracking also hostgroup will allow to better understand how efficient is query routing.
|
1.0
|
Track hostgroup in stats_mysql_query_digest - That is an enhancement.
Currently table stats_mysql_query_digest tracks several metrics, but not hostgroup.
Tracking also hostgroup will allow to better understand how efficient is query routing.
|
process
|
track hostgroup in stats mysql query digest that is an enhancement currently table stats mysql query digest tracks several metrics but not hostgroup tracking also hostgroup will allow to better understand how efficient is query routing
| 1
|
770,756
| 27,055,136,198
|
IssuesEvent
|
2023-02-13 15:41:56
|
ufosc/Club_Website_2
|
https://api.github.com/repos/ufosc/Club_Website_2
|
closed
|
About button should link to #title_animation
|
bug/error good first issue help wanted high priority difficulty: easy
|
**Problem:**
the About button in the navigation bar redirects to the `#about` anchor, which no longer exists because the about section was removed.
**TODO:**
Navigate through the site and ensure that all About buttons link to the `#title_animation` anchor. The first button that comes to mind is in `views/components/navbar.ejs` - but there might be more.
|
1.0
|
About button should link to #title_animation - **Problem:**
the About button in the navigation bar redirects to the `#about` anchor, which no longer exists because the about section was removed.
**TODO:**
Navigate through the site and ensure that all About buttons link to the `#title_animation` anchor. The first button that comes to mind is in `views/components/navbar.ejs` - but there might be more.
|
non_process
|
about button should link to title animation problem the about button in the navigation bar redirects to the about anchor which no longer exists because the about section was removed todo navigate through the site and ensure that all about buttons link to the title animation anchor the first button that comes to mind is in views components navbar ejs but there might be more
| 0
|
790,362
| 27,823,868,555
|
IssuesEvent
|
2023-03-19 14:45:32
|
bounswe/bounswe2023group7
|
https://api.github.com/repos/bounswe/bounswe2023group7
|
closed
|
Research about Mock-Up
|
type : research priority : high status : need_review effort : low
|
### Issue Description
I'm going to research what is Mock-Up, why is it useful, why should we have it. Then I'm going to document it.
### Tasks
_No response_
### Extra Comments
_No response_
### Final Step
_No response_
### Deadline of the Issue
19.03.2023 - 23.59
### Reviewer
@melihgezerr
### Deadline for the Review
20.03.2023 - 23.59
|
1.0
|
Research about Mock-Up - ### Issue Description
I'm going to research what is Mock-Up, why is it useful, why should we have it. Then I'm going to document it.
### Tasks
_No response_
### Extra Comments
_No response_
### Final Step
_No response_
### Deadline of the Issue
19.03.2023 - 23.59
### Reviewer
@melihgezerr
### Deadline for the Review
20.03.2023 - 23.59
|
non_process
|
research about mock up issue description i m going to research what is mock up why is it useful why should we have it then i m going to document it tasks no response extra comments no response final step no response deadline of the issue reviewer melihgezerr deadline for the review
| 0
|
18,673
| 24,591,535,134
|
IssuesEvent
|
2022-10-14 03:01:56
|
Tencent/tdesign-miniprogram
|
https://api.github.com/repos/Tencent/tdesign-miniprogram
|
closed
|
[t-tabs] 结合t-navbar时使用sticky属性导致标签无法以t-navbar下边缘为吸顶位置
|
good first issue in process
|
### tdesign-miniprogram 版本
0.22.0-beta.0
### 重现链接
_No response_
### 重现步骤
``` html
<t-navbar
title="标题"
leftIcon="chevron-left"
homeIcon="home"
bind:go-home="onGoHome"
t-class-left-icon="left-icon-back"
t-class-home-icon="home-icon"
t-class-title="nav-title"
/>
<t-tabs
defaultValue="{{0}}"
sticky="{{true}}"
bind:change="onTabsChange"
bind:click="onTabsClick"
bind:scroll="onStickyScroll"
t-class="custom-tabs"
>
<t-tab-panel label="全部" value="0">
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>bb</view>
</t-tab-panel>
<t-tab-panel label="未完成" value="1">
<view>未完成</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>未完成</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>未完成</view>
</t-tab-panel>
<t-tab-panel label="已完成" value="2">已完成</t-tab-panel>
</t-tabs>
```
### 期望结果
选项卡上边界以t-navbar下边界自动吸顶
### 实际结果
选项卡没有吸顶效果。
### 框架版本
0.22.0-beta.0
### 浏览器版本
_No response_
### 系统版本
_No response_
### Node版本
_No response_
### 补充说明
_No response_
|
1.0
|
[t-tabs] 结合t-navbar时使用sticky属性导致标签无法以t-navbar下边缘为吸顶位置 - ### tdesign-miniprogram 版本
0.22.0-beta.0
### 重现链接
_No response_
### 重现步骤
``` html
<t-navbar
title="标题"
leftIcon="chevron-left"
homeIcon="home"
bind:go-home="onGoHome"
t-class-left-icon="left-icon-back"
t-class-home-icon="home-icon"
t-class-title="nav-title"
/>
<t-tabs
defaultValue="{{0}}"
sticky="{{true}}"
bind:change="onTabsChange"
bind:click="onTabsClick"
bind:scroll="onStickyScroll"
t-class="custom-tabs"
>
<t-tab-panel label="全部" value="0">
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>bb</view>
</t-tab-panel>
<t-tab-panel label="未完成" value="1">
<view>未完成</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>aaaa</view>
<view>未完成</view><view>aaaa</view><view>aaaa</view><view>aaaa</view><view>未完成</view>
</t-tab-panel>
<t-tab-panel label="已完成" value="2">已完成</t-tab-panel>
</t-tabs>
```
### 期望结果
选项卡上边界以t-navbar下边界自动吸顶
### 实际结果
选项卡没有吸顶效果。
### 框架版本
0.22.0-beta.0
### 浏览器版本
_No response_
### 系统版本
_No response_
### Node版本
_No response_
### 补充说明
_No response_
|
process
|
结合t navbar时使用sticky属性导致标签无法以t navbar下边缘为吸顶位置 tdesign miniprogram 版本 beta 重现链接 no response 重现步骤 html t navbar title 标题 lefticon chevron left homeicon home bind go home ongohome t class left icon left icon back t class home icon home icon t class title nav title t tabs defaultvalue sticky true bind change ontabschange bind click ontabsclick bind scroll onstickyscroll t class custom tabs aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa bb 未完成 aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa aaaa 未完成 aaaa aaaa aaaa 未完成 已完成 期望结果 选项卡上边界以t navbar下边界自动吸顶 实际结果 选项卡没有吸顶效果。 框架版本 beta 浏览器版本 no response 系统版本 no response node版本 no response 补充说明 no response
| 1
|
13,445
| 3,332,927,738
|
IssuesEvent
|
2015-11-11 22:21:57
|
t2davis/Squad
|
https://api.github.com/repos/t2davis/Squad
|
opened
|
buddybuild feedback from Mariajfdavis@gmail.com on build #12: The keyboard stays on the screen when you go to the map if u…
|
1.0 BetaTester
|
Feedback from Mariajfdavis@gmail.com : The keyboard stays on the screen when you go to the map if u had it open in messages
<img src="https://s3-us-west-2.amazonaws.com/buddybuild-screenshots/5642b68257de0f0100c2f1c9/5643a14fbbd0c90100b82c5f/c543d717-8402-4b0b-9f6e-a90062b6e181.jpg" width="33%" height="33%" />
|
1.0
|
buddybuild feedback from Mariajfdavis@gmail.com on build #12: The keyboard stays on the screen when you go to the map if u… - Feedback from Mariajfdavis@gmail.com : The keyboard stays on the screen when you go to the map if u had it open in messages
<img src="https://s3-us-west-2.amazonaws.com/buddybuild-screenshots/5642b68257de0f0100c2f1c9/5643a14fbbd0c90100b82c5f/c543d717-8402-4b0b-9f6e-a90062b6e181.jpg" width="33%" height="33%" />
|
non_process
|
buddybuild feedback from mariajfdavis gmail com on build the keyboard stays on the screen when you go to the map if u… feedback from mariajfdavis gmail com the keyboard stays on the screen when you go to the map if u had it open in messages
| 0
|
555,743
| 16,468,898,141
|
IssuesEvent
|
2021-05-23 01:56:50
|
googleapis/google-api-go-client
|
https://api.github.com/repos/googleapis/google-api-go-client
|
closed
|
Google Youtube API V3 PlaylistItemSnippet Position 0 Is Ignored
|
priority: p2 type: bug
|
#### Environment details
- Programming language: Golang
- OS: WSL2 Windows
- Language runtime version: go version go1.16 linux/amd64
- Package version: 0.47.0
#### Steps to reproduce
1. Create a playlist
2. Add video to a playlist
3. Add video to a playlist at position 0
4. Notice that it is added to the end of the playlist rather than the first position.
```
&youtube.PlaylistItem{
Snippet: &youtube.PlaylistItemSnippet{
Position: 0,
},
}
```
#### Note
PlaylistItemSnippet Position is int64 rather than *int64 like Position in ChannelSectionSnippet.
As such it is ignored with by `json:"omitempty"`.
|
1.0
|
Google Youtube API V3 PlaylistItemSnippet Position 0 Is Ignored - #### Environment details
- Programming language: Golang
- OS: WSL2 Windows
- Language runtime version: go version go1.16 linux/amd64
- Package version: 0.47.0
#### Steps to reproduce
1. Create a playlist
2. Add video to a playlist
3. Add video to a playlist at position 0
4. Notice that it is added to the end of the playlist rather than the first position.
```
&youtube.PlaylistItem{
Snippet: &youtube.PlaylistItemSnippet{
Position: 0,
},
}
```
#### Note
PlaylistItemSnippet Position is int64 rather than *int64 like Position in ChannelSectionSnippet.
As such it is ignored with by `json:"omitempty"`.
|
non_process
|
google youtube api playlistitemsnippet position is ignored environment details programming language golang os windows language runtime version go version linux package version steps to reproduce create a playlist add video to a playlist add video to a playlist at position notice that it is added to the end of the playlist rather than the first position youtube playlistitem snippet youtube playlistitemsnippet position note playlistitemsnippet position is rather than like position in channelsectionsnippet as such it is ignored with by json omitempty
| 0
|
19,596
| 3,776,504,971
|
IssuesEvent
|
2016-03-17 16:52:27
|
leeensminger/DelDOT-NPDES-Field-Tool
|
https://api.github.com/repos/leeensminger/DelDOT-NPDES-Field-Tool
|
closed
|
Can buffer for drawing barrel conveyances be larger?
|
fixed will be ready for testing in next release.
|
When there are more than two barrels next to each other, the buffer used to select the upstream/downstream structures is not large enough to encompass the structures. Could the buffer be larger for just barrel conveyances?

|
1.0
|
Can buffer for drawing barrel conveyances be larger? - When there are more than two barrels next to each other, the buffer used to select the upstream/downstream structures is not large enough to encompass the structures. Could the buffer be larger for just barrel conveyances?

|
non_process
|
can buffer for drawing barrel conveyances be larger when there are more than two barrels next to each other the buffer used to select the upstream downstream structures is not large enough to encompass the structures could the buffer be larger for just barrel conveyances
| 0
|
4,516
| 7,360,185,455
|
IssuesEvent
|
2018-03-10 16:00:20
|
ODiogoSilva/assemblerflow
|
https://api.github.com/repos/ODiogoSilva/assemblerflow
|
opened
|
Add container attribute to Process
|
enhancement process
|
Add the `container` attribute to the Process base class. This will be used to determine the container directive in the nextflow configuration file.
|
1.0
|
Add container attribute to Process - Add the `container` attribute to the Process base class. This will be used to determine the container directive in the nextflow configuration file.
|
process
|
add container attribute to process add the container attribute to the process base class this will be used to determine the container directive in the nextflow configuration file
| 1
|
20,960
| 27,817,510,295
|
IssuesEvent
|
2023-03-18 21:19:10
|
cse442-at-ub/project_s23-iweatherify
|
https://api.github.com/repos/cse442-at-ub/project_s23-iweatherify
|
closed
|
Save the units and temperature settings to the database
|
Processing Task Sprint 2
|
**Task Tests**
*Test 1*
1. Go to the following URL: https://github.com/cse442-at-ub/project_s23-iweatherify/tree/dev
2. Click on the green `<> Code` button and download the ZIP file.

3. Unzip the downloaded file to a folder on your computer.
4. Open a terminal and navigate to the git repository folder using the `cd` command.
5. Run the `npm install` command in the terminal to install the necessary dependencies.
6. Run the `npm start` command in the terminal to start the application.
7. Check the output from the npm start command for the URL to access the application. The URL should be a localhost address (e.g., http://localhost:8080).
8. Navigate to http://localhost:8080/#/login
9. Ensure you have logged in to our app to see the page use UserID: `UB442` and Password:`Myub442@!` to login
10. Go to URL: http://localhost:8080/#/unitsSettings
11. Verify that the units page is displayed

12. Change the temperature unit to Celsius (°C)
13. Change the wind unit to km/h
14. Change the pressure unit to mm
15. Change the distance unit to km
16. Open the browser inspector tool and select console
17. Click the save button
18. You should see the message: `Units saved successfully.` on the page

19. You should see the message: `{message: 'User settings saved successfully.'}` in the console

18. Open a different tab and go to: https://www-student.cse.buffalo.edu/tools/db/phpmyadmin/index.php
19. Input username: `jpan26` and password: `50314999`
20. Make sure the server choice is `oceanus.cse.buffalo.edu:3306`
21. Click go and you should see this page

22. Click `cse442_2023_spring_team_a_db` first and then `saved_units` on the left side of the page

23. Verify you see a row with the exact same information as shown by the picture

*Test 2*
1. Repeat steps 1 to 9 from `Test 1`
2. Go to URL: http://localhost:8080/#/tempSettings
3. Verify that the temperature setting page is displayed

4. Open the browser inspector tool and select console
5. Change the hot temperature to 80, you can either use the slider or input box and click save
6. You should see the message: `{result: 'success'}` in the console

7. You should see the message: `Temperatures Saved Successfully` on the page

8. Change the warm temperature to 65, you can either use the slider or input box and click save
9. You should see the message: `{result: 'success'}` in the console

10. You should see the message: `Temperatures Saved Successfully` on the page

11. Change the ideal temperature to 50, you can either use the slider or input box and click save
12. You should see the message: `{result: 'success'}` in the console

13. You should see the message: `Temperatures Saved Successfully` on the page

14. Change the chilly temperature to 0, you can either use the slider or input box and click save
15. You should see the message: `{result: 'success'}` in the console

16. You should see the message: `Temperatures Saved Successfully` on the page

17. Change the cold temperature to -65, you can either use the slider or input box and click save
18. You should see the message: `{result: 'success'}` in the console

19. You should see the message: `Temperatures Saved Successfully` on the page

20. Change the freezing temperature to -80, you can either use the slider or input box and click save
21. You should see the message: `{result: 'success'}` in the console

22. You should see the message: `Temperatures Saved Successfully` on the page

23. Repeat steps 18 to 21 from `Test 1`
24. Click `cse442_2023_spring_team_a_db` first and then `saved_temperatures` on the left side of the page

25. Verify you see a row with the exact same information as shown by the picture

|
1.0
|
Save the units and temperature settings to the database - **Task Tests**
*Test 1*
1. Go to the following URL: https://github.com/cse442-at-ub/project_s23-iweatherify/tree/dev
2. Click on the green `<> Code` button and download the ZIP file.

3. Unzip the downloaded file to a folder on your computer.
4. Open a terminal and navigate to the git repository folder using the `cd` command.
5. Run the `npm install` command in the terminal to install the necessary dependencies.
6. Run the `npm start` command in the terminal to start the application.
7. Check the output from the npm start command for the URL to access the application. The URL should be a localhost address (e.g., http://localhost:8080).
8. Navigate to http://localhost:8080/#/login
9. Ensure you have logged in to our app to see the page use UserID: `UB442` and Password:`Myub442@!` to login
10. Go to URL: http://localhost:8080/#/unitsSettings
11. Verify that the units page is displayed

12. Change the temperature unit to Celsius (°C)
13. Change the wind unit to km/h
14. Change the pressure unit to mm
15. Change the distance unit to km
16. Open the browser inspector tool and select console
17. Click the save button
18. You should see the message: `Units saved successfully.` on the page

19. You should see the message: `{message: 'User settings saved successfully.'}` in the console

18. Open a different tab and go to: https://www-student.cse.buffalo.edu/tools/db/phpmyadmin/index.php
19. Input username: `jpan26` and password: `50314999`
20. Make sure the server choice is `oceanus.cse.buffalo.edu:3306`
21. Click go and you should see this page

22. Click `cse442_2023_spring_team_a_db` first and then `saved_units` on the left side of the page

23. Verify you see a row with the exact same information as shown by the picture

*Test 2*
1. Repeat steps 1 to 9 from `Test 1`
2. Go to URL: http://localhost:8080/#/tempSettings
3. Verify that the temperature setting page is displayed

4. Open the browser inspector tool and select console
5. Change the hot temperature to 80, you can either use the slider or input box and click save
6. You should see the message: `{result: 'success'}` in the console

7. You should see the message: `Temperatures Saved Successfully` on the page

8. Change the warm temperature to 65, you can either use the slider or input box and click save
9. You should see the message: `{result: 'success'}` in the console

10. You should see the message: `Temperatures Saved Successfully` on the page

11. Change the ideal temperature to 50, you can either use the slider or input box and click save
12. You should see the message: `{result: 'success'}` in the console

13. You should see the message: `Temperatures Saved Successfully` on the page

14. Change the chilly temperature to 0, you can either use the slider or input box and click save
15. You should see the message: `{result: 'success'}` in the console

16. You should see the message: `Temperatures Saved Successfully` on the page

17. Change the cold temperature to -65, you can either use the slider or input box and click save
18. You should see the message: `{result: 'success'}` in the console

19. You should see the message: `Temperatures Saved Successfully` on the page

20. Change the freezing temperature to -80, you can either use the slider or input box and click save
21. You should see the message: `{result: 'success'}` in the console

22. You should see the message: `Temperatures Saved Successfully` on the page

23. Repeat steps 18 to 21 from `Test 1`
24. Click `cse442_2023_spring_team_a_db` first and then `saved_temperatures` on the left side of the page

25. Verify you see a row with the exact same information as shown by the picture

|
process
|
save the units and temperature settings to the database task tests test go to the following url click on the green code button and download the zip file unzip the downloaded file to a folder on your computer open a terminal and navigate to the git repository folder using the cd command run the npm install command in the terminal to install the necessary dependencies run the npm start command in the terminal to start the application check the output from the npm start command for the url to access the application the url should be a localhost address e g navigate to ensure you have logged in to our app to see the page use userid and password to login go to url verify that the units page is displayed change the temperature unit to celsius °c change the wind unit to km h change the pressure unit to mm change the distance unit to km open the browser inspector tool and select console click the save button you should see the message units saved successfully on the page you should see the message message user settings saved successfully in the console open a different tab and go to input username and password make sure the server choice is oceanus cse buffalo edu click go and you should see this page click spring team a db first and then saved units on the left side of the page verify you see a row with the exact same information as shown by the picture test repeat steps to from test go to url verify that the temperature setting page is displayed open the browser inspector tool and select console change the hot temperature to you can either use the slider or input box and click save you should see the message result success in the console you should see the message temperatures saved successfully on the page change the warm temperature to you can either use the slider or input box and click save you should see the message result success in the console you should see the message temperatures saved successfully on the page change the ideal temperature to you can either use the slider or input box and click save you should see the message result success in the console you should see the message temperatures saved successfully on the page change the chilly temperature to you can either use the slider or input box and click save you should see the message result success in the console you should see the message temperatures saved successfully on the page change the cold temperature to you can either use the slider or input box and click save you should see the message result success in the console you should see the message temperatures saved successfully on the page change the freezing temperature to you can either use the slider or input box and click save you should see the message result success in the console you should see the message temperatures saved successfully on the page repeat steps to from test click spring team a db first and then saved temperatures on the left side of the page verify you see a row with the exact same information as shown by the picture
| 1
|
67,601
| 27,960,815,404
|
IssuesEvent
|
2023-03-24 15:31:39
|
LiskHQ/lisk-service
|
https://api.github.com/repos/LiskHQ/lisk-service
|
opened
|
Add functional tests for app-registry
|
type: test service/blockchain-app-registry
|
### Description
Add missing functional tests for app-registry.
### Acceptance Criteria
- Functional tests are added for the following methods in `services/blockchain-app-registry/shared/metadataIndex.js`
- indexAppMeta
- indexMetadataFromFile
- deleteAppMeta
- deleteTokensMeta
- deleteIndexedMetadataFromFile
- indexAllBlockchainAppsMeta
|
1.0
|
Add functional tests for app-registry - ### Description
Add missing functional tests for app-registry.
### Acceptance Criteria
- Functional tests are added for the following methods in `services/blockchain-app-registry/shared/metadataIndex.js`
- indexAppMeta
- indexMetadataFromFile
- deleteAppMeta
- deleteTokensMeta
- deleteIndexedMetadataFromFile
- indexAllBlockchainAppsMeta
|
non_process
|
add functional tests for app registry description add missing functional tests for app registry acceptance criteria functional tests are added for the following methods in services blockchain app registry shared metadataindex js indexappmeta indexmetadatafromfile deleteappmeta deletetokensmeta deleteindexedmetadatafromfile indexallblockchainappsmeta
| 0
|
32,167
| 13,770,403,330
|
IssuesEvent
|
2020-10-07 20:10:45
|
Azure/azure-sdk-for-python
|
https://api.github.com/repos/Azure/azure-sdk-for-python
|
opened
|
[Servicebus] Docstring review and bolstering pre-GA
|
Client Service Bus
|
pre-GA fit-and-finish.
Review all of our docstrings. Ensure text is clear, accurate, up-to-date and comprehensive.
Things to focus on:
- are all params covered?
- rval and returns? (both defined type and description)
- custom errors being thrown?
- typedefs in all the places? (both # type: and docstring types)
|
1.0
|
[Servicebus] Docstring review and bolstering pre-GA - pre-GA fit-and-finish.
Review all of our docstrings. Ensure text is clear, accurate, up-to-date and comprehensive.
Things to focus on:
- are all params covered?
- rval and returns? (both defined type and description)
- custom errors being thrown?
- typedefs in all the places? (both # type: and docstring types)
|
non_process
|
docstring review and bolstering pre ga pre ga fit and finish review all of our docstrings ensure text is clear accurate up to date and comprehensive things to focus on are all params covered rval and returns both defined type and description custom errors being thrown typedefs in all the places both type and docstring types
| 0
|
35,643
| 14,789,886,763
|
IssuesEvent
|
2021-01-12 11:12:27
|
terraform-providers/terraform-provider-azurerm
|
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
|
closed
|
RoleAssignmentNotFound with azurerm_role_assignment
|
bug service/authorization
|
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform (and AzureRM Provider) Version
Terraform v0.13.3
+ provider registry.terraform.io/hashicorp/azurerm v2.32.0
+ provider registry.terraform.io/hashicorp/external v2.0.0
### Affected Resource(s)
* `azurerm_role_assigment`
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "azurerm_user_assigned_identity" "tc" {
name = "tc"
resource_group_name = azurerm_resource_group.rg.name
location = azurerm_resource_group.rg.location
tags = local.default_tags
}
resource "azurerm_role_assignment" "tc-acr" {
scope = data.azurerm_resources.acr.resources[0].id
role_definition_name = "AcrPull"
principal_id = azurerm_user_assigned_identity.tc.principal_id
skip_service_principal_aad_check = true
}
resource "azurerm_role_assignment" "tc-privatedns" {
scope = azurerm_resource_group.rg.id
role_definition_name = "Private DNS Zone Contributor"
principal_id = azurerm_user_assigned_identity.tc.principal_id
skip_service_principal_aad_check = true
}
data "azurerm_resource_group" "dns" {
name = "dns"
}
resource "azurerm_role_assignment" "tc-dns" {
scope = data.azurerm_resource_group.dns.id
role_definition_name = "DNS Zone Contributor"
principal_id = azurerm_user_assigned_identity.tc.principal_id
skip_service_principal_aad_check = true
}
resource "azurerm_role_assignment" "tc-vm" {
scope = azurerm_resource_group.rg.id
role_definition_name = "Virtual Machine Contributor"
principal_id = azurerm_user_assigned_identity.tc.principal_id
skip_service_principal_aad_check = true
}
resource "azurerm_role_assignment" "tc-monitoring" {
scope = azurerm_resource_group.rg.id
role_definition_name = "Monitoring Reader"
principal_id = azurerm_user_assigned_identity.traffic-control.principal_id
skip_service_principal_aad_check = true
}
resource "azurerm_role_assignment" "tc-aks" {
scope = azurerm_resource_group.rg.id
role_definition_name = "Azure Kubernetes Service Cluster User Role"
principal_id = azurerm_user_assigned_identity.traffic-control.principal_id
skip_service_principal_aad_check = true
}
resource "azurerm_linux_virtual_machine_scale_set" "tc" {
name = "tc"
identity {
type = "UserAssigned"
identity_ids = [azurerm_user_assigned_identity.tc.id]
}
...
depends_on = [
azurerm_role_assignment.tc-acr,
azurerm_role_assignment.tc-privatedns,
azurerm_role_assignment.tc-dns,
azurerm_role_assignment.tc-vm,
azurerm_role_assignment.tc-monitoring,
azurerm_role_assignment.tc-aks
]
}
resource "azurerm_linux_virtual_machine_scale_set" "te" {
for_each = var.availability_zones
name = "te-${each.value}"
identity {
type = "UserAssigned"
identity_ids = [azurerm_user_assigned_identity.te.id]
}
depends_on = [
azurerm_role_assignment.te-acr
]
}
```
### Debug Output
```
Still creating... [2m0s elapsed]
azurerm_linux_virtual_machine_scale_set.te["3"]:
Still creating... [2m10s elapsed]
azurerm_linux_virtual_machine_scale_set.traffic-envoy["3"]: Creation complete after 2m15s [id=/subscriptions/xxxx/resourceGroups/yyyy/providers/Microsoft.Compute/virtualMachineScaleSets/te-3]
Error: authorization.RoleAssignmentsClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code="RoleAssignmentNotFound" Message="The role assignment '920ffd05-df3c-308b-2e01-c8c58481998e' is not found."
on tc.tf line 8, in resource "azurerm_role_assignment" "tc-acr":
8: resource "azurerm_role_assignment" "tc-acr" {
Error: authorization.RoleAssignmentsClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code="RoleAssignmentNotFound" Message="The role assignment '1fda1763-4e9f-2c9f-b0a3-a2581b68e457' is not found."
on tc.tf line 26, in resource "azurerm_role_assignment" "tc-dns":
26: resource \\\"azurerm_role_assignment\\\" \\\"tc-dns\\\" {
Error: authorization.RoleAssignmentsClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code=\\\"RoleAssignmentNotFound\\\" Message=\\\"The role assignment '8f820931-964c-84bf-ecc7-f3c2385fb6a9' is not found."
on tc.tf line 47, in resource "azurerm_role_assignment" "tc-aks":
47: resource \\\"azurerm_role_assignment\\\" "tc-aks"
```
### Expected Behaviour
Success
### Actual Behaviour
Failure
### Important Factoids
Before I placed the dependency between the VMSS and role assignments, the failure between tc role assignment would occur on the tc VMSS. Once the dependency was added it shifted to the te VMSS.
Not all role assignments fail, its usually 2 or 3, guess its timing related.
* #0000
|
1.0
|
RoleAssignmentNotFound with azurerm_role_assignment - ### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform (and AzureRM Provider) Version
Terraform v0.13.3
+ provider registry.terraform.io/hashicorp/azurerm v2.32.0
+ provider registry.terraform.io/hashicorp/external v2.0.0
### Affected Resource(s)
* `azurerm_role_assigment`
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "azurerm_user_assigned_identity" "tc" {
name = "tc"
resource_group_name = azurerm_resource_group.rg.name
location = azurerm_resource_group.rg.location
tags = local.default_tags
}
resource "azurerm_role_assignment" "tc-acr" {
scope = data.azurerm_resources.acr.resources[0].id
role_definition_name = "AcrPull"
principal_id = azurerm_user_assigned_identity.tc.principal_id
skip_service_principal_aad_check = true
}
resource "azurerm_role_assignment" "tc-privatedns" {
scope = azurerm_resource_group.rg.id
role_definition_name = "Private DNS Zone Contributor"
principal_id = azurerm_user_assigned_identity.tc.principal_id
skip_service_principal_aad_check = true
}
data "azurerm_resource_group" "dns" {
name = "dns"
}
resource "azurerm_role_assignment" "tc-dns" {
scope = data.azurerm_resource_group.dns.id
role_definition_name = "DNS Zone Contributor"
principal_id = azurerm_user_assigned_identity.tc.principal_id
skip_service_principal_aad_check = true
}
resource "azurerm_role_assignment" "tc-vm" {
scope = azurerm_resource_group.rg.id
role_definition_name = "Virtual Machine Contributor"
principal_id = azurerm_user_assigned_identity.tc.principal_id
skip_service_principal_aad_check = true
}
resource "azurerm_role_assignment" "tc-monitoring" {
scope = azurerm_resource_group.rg.id
role_definition_name = "Monitoring Reader"
principal_id = azurerm_user_assigned_identity.traffic-control.principal_id
skip_service_principal_aad_check = true
}
resource "azurerm_role_assignment" "tc-aks" {
scope = azurerm_resource_group.rg.id
role_definition_name = "Azure Kubernetes Service Cluster User Role"
principal_id = azurerm_user_assigned_identity.traffic-control.principal_id
skip_service_principal_aad_check = true
}
resource "azurerm_linux_virtual_machine_scale_set" "tc" {
name = "tc"
identity {
type = "UserAssigned"
identity_ids = [azurerm_user_assigned_identity.tc.id]
}
...
depends_on = [
azurerm_role_assignment.tc-acr,
azurerm_role_assignment.tc-privatedns,
azurerm_role_assignment.tc-dns,
azurerm_role_assignment.tc-vm,
azurerm_role_assignment.tc-monitoring,
azurerm_role_assignment.tc-aks
]
}
resource "azurerm_linux_virtual_machine_scale_set" "te" {
for_each = var.availability_zones
name = "te-${each.value}"
identity {
type = "UserAssigned"
identity_ids = [azurerm_user_assigned_identity.te.id]
}
depends_on = [
azurerm_role_assignment.te-acr
]
}
```
### Debug Output
```
Still creating... [2m0s elapsed]
azurerm_linux_virtual_machine_scale_set.te["3"]:
Still creating... [2m10s elapsed]
azurerm_linux_virtual_machine_scale_set.traffic-envoy["3"]: Creation complete after 2m15s [id=/subscriptions/xxxx/resourceGroups/yyyy/providers/Microsoft.Compute/virtualMachineScaleSets/te-3]
Error: authorization.RoleAssignmentsClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code="RoleAssignmentNotFound" Message="The role assignment '920ffd05-df3c-308b-2e01-c8c58481998e' is not found."
on tc.tf line 8, in resource "azurerm_role_assignment" "tc-acr":
8: resource "azurerm_role_assignment" "tc-acr" {
Error: authorization.RoleAssignmentsClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code="RoleAssignmentNotFound" Message="The role assignment '1fda1763-4e9f-2c9f-b0a3-a2581b68e457' is not found."
on tc.tf line 26, in resource "azurerm_role_assignment" "tc-dns":
26: resource \\\"azurerm_role_assignment\\\" \\\"tc-dns\\\" {
Error: authorization.RoleAssignmentsClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code=\\\"RoleAssignmentNotFound\\\" Message=\\\"The role assignment '8f820931-964c-84bf-ecc7-f3c2385fb6a9' is not found."
on tc.tf line 47, in resource "azurerm_role_assignment" "tc-aks":
47: resource \\\"azurerm_role_assignment\\\" "tc-aks"
```
### Expected Behaviour
Success
### Actual Behaviour
Failure
### Important Factoids
Before I placed the dependency between the VMSS and role assignments, the failure between tc role assignment would occur on the tc VMSS. Once the dependency was added it shifted to the te VMSS.
Not all role assignments fail, its usually 2 or 3, guess its timing related.
* #0000
|
non_process
|
roleassignmentnotfound with azurerm role assignment community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform and azurerm provider version terraform provider registry terraform io hashicorp azurerm provider registry terraform io hashicorp external affected resource s azurerm role assigment terraform configuration files hcl resource azurerm user assigned identity tc name tc resource group name azurerm resource group rg name location azurerm resource group rg location tags local default tags resource azurerm role assignment tc acr scope data azurerm resources acr resources id role definition name acrpull principal id azurerm user assigned identity tc principal id skip service principal aad check true resource azurerm role assignment tc privatedns scope azurerm resource group rg id role definition name private dns zone contributor principal id azurerm user assigned identity tc principal id skip service principal aad check true data azurerm resource group dns name dns resource azurerm role assignment tc dns scope data azurerm resource group dns id role definition name dns zone contributor principal id azurerm user assigned identity tc principal id skip service principal aad check true resource azurerm role assignment tc vm scope azurerm resource group rg id role definition name virtual machine contributor principal id azurerm user assigned identity tc principal id skip service principal aad check true resource azurerm role assignment tc monitoring scope azurerm resource group rg id role definition name monitoring reader principal id azurerm user assigned identity traffic control principal id skip service principal aad check true resource azurerm role assignment tc aks scope azurerm resource group rg id role definition name azure kubernetes service cluster user role principal id azurerm user assigned identity traffic control principal id skip service principal aad check true resource azurerm linux virtual machine scale set tc name tc identity type userassigned identity ids depends on azurerm role assignment tc acr azurerm role assignment tc privatedns azurerm role assignment tc dns azurerm role assignment tc vm azurerm role assignment tc monitoring azurerm role assignment tc aks resource azurerm linux virtual machine scale set te for each var availability zones name te each value identity type userassigned identity ids depends on azurerm role assignment te acr debug output still creating azurerm linux virtual machine scale set te still creating azurerm linux virtual machine scale set traffic envoy creation complete after error authorization roleassignmentsclient get failure responding to request statuscode original error autorest azure service returned an error status code roleassignmentnotfound message the role assignment is not found on tc tf line in resource azurerm role assignment tc acr resource azurerm role assignment tc acr error authorization roleassignmentsclient get failure responding to request statuscode original error autorest azure service returned an error status code roleassignmentnotfound message the role assignment is not found on tc tf line in resource azurerm role assignment tc dns resource azurerm role assignment tc dns error authorization roleassignmentsclient get failure responding to request statuscode original error autorest azure service returned an error status code roleassignmentnotfound message the role assignment is not found on tc tf line in resource azurerm role assignment tc aks resource azurerm role assignment tc aks expected behaviour success actual behaviour failure important factoids before i placed the dependency between the vmss and role assignments the failure between tc role assignment would occur on the tc vmss once the dependency was added it shifted to the te vmss not all role assignments fail its usually or guess its timing related
| 0
|
4,737
| 7,594,724,809
|
IssuesEvent
|
2018-04-27 00:49:35
|
agroportal/agroportal_web_ui
|
https://api.github.com/repos/agroportal/agroportal_web_ui
|
closed
|
Biorefinery & Transmat failed to parse
|
ontology processing problem
|
Error from parsing log file (Biorefinery): Illegal rdf:nodeID value '_:genid259'
there is an equivalent error for Transmat.
This error seems to have been identified in the NCBO BioPortal: see
- [https://github.com/ncbo/bioportal-project/issues/32#event-1226205997](https://github.com/ncbo/bioportal-project/issues/32#event-1226205997)
- [https://github.com/ncbo/bioportal-project/issues/9](https://github.com/ncbo/bioportal-project/issues/9)
- https://sourceforge.net/p/owlapi/mailman/owlapi-developer/thread/CAD2jOMO3yMeQ7Ojk7TtNdMoshqQCQ87ke37aQWJDNeqZzvCAew%40mail.gmail.com/#msg35964428
@jvendetti Did you solve this problem?
|
1.0
|
Biorefinery & Transmat failed to parse - Error from parsing log file (Biorefinery): Illegal rdf:nodeID value '_:genid259'
there is an equivalent error for Transmat.
This error seems to have been identified in the NCBO BioPortal: see
- [https://github.com/ncbo/bioportal-project/issues/32#event-1226205997](https://github.com/ncbo/bioportal-project/issues/32#event-1226205997)
- [https://github.com/ncbo/bioportal-project/issues/9](https://github.com/ncbo/bioportal-project/issues/9)
- https://sourceforge.net/p/owlapi/mailman/owlapi-developer/thread/CAD2jOMO3yMeQ7Ojk7TtNdMoshqQCQ87ke37aQWJDNeqZzvCAew%40mail.gmail.com/#msg35964428
@jvendetti Did you solve this problem?
|
process
|
biorefinery transmat failed to parse error from parsing log file biorefinery illegal rdf nodeid value there is an equivalent error for transmat this error seems to have been identified in the ncbo bioportal see jvendetti did you solve this problem
| 1
|
5,575
| 8,410,084,216
|
IssuesEvent
|
2018-10-12 09:28:16
|
threefoldtech/jumpscale_lib
|
https://api.github.com/repos/threefoldtech/jumpscale_lib
|
closed
|
Use the base token value instead of hastings for the txfee in the j.clients.rivine.wallet creation
|
process_wontfix
|
And rename it to transactionfee.
might be that the docs are wrong, they seem to contradict:
```python
from JumpscaleLib.clients.blockchain.rivine.RivineWallet import RivineWallet
wallet = RivineWallet(seed=seed,
bc_network='https://explorer.testnet.threefoldtoken.com/',
bc_network_password='test123',
nr_keys_per_seed=5,
minerfee=100000000)
# where seed is the seed you have or generated
# bc_network: is the url to the blockchain network explorer node
# bc_network_password: is the password to use while communicating with the chain explorer node
# nr_keys_per_seed: is how many keys to generate per seed
# minerfee: How many hastings should be added as minerfee
```
Or alternatively you can configure the wallet instance using the following code:
```python
client_data = {'bc_address': 'https://explorer.testnet.threefoldtoken.com/',
'password_': 'test123',
'minerfee': 10,
'nr_keys_per_seed': 5,
'seed_': seed}
rivine_client = j.clients.rivine.get('mytestwallet', data=client_data)
rivine_client.config.save()
wallet = rivine_client.wallet
```
while it should be 0.1
|
1.0
|
Use the base token value instead of hastings for the txfee in the j.clients.rivine.wallet creation - And rename it to transactionfee.
might be that the docs are wrong, they seem to contradict:
```python
from JumpscaleLib.clients.blockchain.rivine.RivineWallet import RivineWallet
wallet = RivineWallet(seed=seed,
bc_network='https://explorer.testnet.threefoldtoken.com/',
bc_network_password='test123',
nr_keys_per_seed=5,
minerfee=100000000)
# where seed is the seed you have or generated
# bc_network: is the url to the blockchain network explorer node
# bc_network_password: is the password to use while communicating with the chain explorer node
# nr_keys_per_seed: is how many keys to generate per seed
# minerfee: How many hastings should be added as minerfee
```
Or alternatively you can configure the wallet instance using the following code:
```python
client_data = {'bc_address': 'https://explorer.testnet.threefoldtoken.com/',
'password_': 'test123',
'minerfee': 10,
'nr_keys_per_seed': 5,
'seed_': seed}
rivine_client = j.clients.rivine.get('mytestwallet', data=client_data)
rivine_client.config.save()
wallet = rivine_client.wallet
```
while it should be 0.1
|
process
|
use the base token value instead of hastings for the txfee in the j clients rivine wallet creation and rename it to transactionfee might be that the docs are wrong they seem to contradict python from jumpscalelib clients blockchain rivine rivinewallet import rivinewallet wallet rivinewallet seed seed bc network bc network password nr keys per seed minerfee where seed is the seed you have or generated bc network is the url to the blockchain network explorer node bc network password is the password to use while communicating with the chain explorer node nr keys per seed is how many keys to generate per seed minerfee how many hastings should be added as minerfee or alternatively you can configure the wallet instance using the following code python client data bc address password minerfee nr keys per seed seed seed rivine client j clients rivine get mytestwallet data client data rivine client config save wallet rivine client wallet while it should be
| 1
|
6,164
| 9,049,614,521
|
IssuesEvent
|
2019-02-12 05:34:13
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
NullPointerException when Publishing a Learning Group Map
|
bug needs reproduction preprocess/chunking priority/high
|
I'm using DITA OT 2.4.6 and when I attempt to publish a Learning Group Map to PDF or HTML5, I'm getting this error below. I tested with DITA OT 2.0 and it is working fine.
[LearningMap.zip](https://github.com/dita-ot/dita-ot/files/1019705/LearningMap.zip)
chunk:
[chunk] Processing file:A$L000022.ditamap
BUILD FAILED
The following error occurred while executing this line:
lib\dita-ot\plugins\org.dita.base\build_preprocess.xml:308: java.lang.NullPointerException
at org.dita.dost.writer.AbstractChunkTopicParser.generateOutputFile(AbstractChunkTopicParser.java:340)
at org.dita.dost.writer.ChunkTopicParser.processChunk(ChunkTopicParser.java:149)
at org.dita.dost.writer.ChunkTopicParser.write(ChunkTopicParser.java:60)
at org.dita.dost.reader.ChunkMapReader.processCombineChunk(ChunkMapReader.java:536)
at org.dita.dost.reader.ChunkMapReader.processTopicref(ChunkMapReader.java:341)
at org.dita.dost.reader.ChunkMapReader.processChildTopicref(ChunkMapReader.java:513)
at org.dita.dost.reader.ChunkMapReader.processTopicref(ChunkMapReader.java:366)
at org.dita.dost.reader.ChunkMapReader.process(ChunkMapReader.java:142)
at org.dita.dost.writer.AbstractDomFilter.read(AbstractDomFilter.java:55)
at org.dita.dost.reader.ChunkMapReader.read(ChunkMapReader.java:118)
at org.dita.dost.module.ChunkModule.execute(ChunkModule.java:80)
at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:70)
at org.dita.dost.invoker.ExtensibleAntInvoker.execute(ExtensibleAntInvoker.java:222)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:293)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
at org.apache.tools.ant.Task.perform(Task.java:348)
at org.apache.tools.ant.Target.execute(Target.java:435)
at org.apache.tools.ant.Target.performTasks(Target.java:456)
at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1405)
at org.apache.tools.ant.helper.SingleCheckExecutor.executeTargets(SingleCheckExecutor.java:38)
at org.apache.tools.ant.Project.executeTargets(Project.java:1260)
at org.apache.tools.ant.taskdefs.Ant.execute(Ant.java:441)
at org.apache.tools.ant.taskdefs.CallTarget.execute(CallTarget.java:105)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:293)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
at org.apache.tools.ant.Task.perform(Task.java:348)
at org.apache.tools.ant.Target.execute(Target.java:435)
at org.apache.tools.ant.Target.performTasks(Target.java:456)
at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1405)
at org.apache.tools.ant.Project.executeTarget(Project.java:1376)
at org.apache.tools.ant.helper.DefaultExecutor.executeTargets(DefaultExecutor.java:41)
at org.apache.tools.ant.Project.executeTargets(Project.java:1260)
at org.apache.tools.ant.Main.runBuild(Main.java:854)
at org.apache.tools.ant.Main.startAnt(Main.java:236)
at org.apache.tools.ant.launch.Launcher.run(Launcher.java:285)
at org.apache.tools.ant.launch.Launcher.main(Launcher.java:112)
|
1.0
|
NullPointerException when Publishing a Learning Group Map - I'm using DITA OT 2.4.6 and when I attempt to publish a Learning Group Map to PDF or HTML5, I'm getting this error below. I tested with DITA OT 2.0 and it is working fine.
[LearningMap.zip](https://github.com/dita-ot/dita-ot/files/1019705/LearningMap.zip)
chunk:
[chunk] Processing file:A$L000022.ditamap
BUILD FAILED
The following error occurred while executing this line:
lib\dita-ot\plugins\org.dita.base\build_preprocess.xml:308: java.lang.NullPointerException
at org.dita.dost.writer.AbstractChunkTopicParser.generateOutputFile(AbstractChunkTopicParser.java:340)
at org.dita.dost.writer.ChunkTopicParser.processChunk(ChunkTopicParser.java:149)
at org.dita.dost.writer.ChunkTopicParser.write(ChunkTopicParser.java:60)
at org.dita.dost.reader.ChunkMapReader.processCombineChunk(ChunkMapReader.java:536)
at org.dita.dost.reader.ChunkMapReader.processTopicref(ChunkMapReader.java:341)
at org.dita.dost.reader.ChunkMapReader.processChildTopicref(ChunkMapReader.java:513)
at org.dita.dost.reader.ChunkMapReader.processTopicref(ChunkMapReader.java:366)
at org.dita.dost.reader.ChunkMapReader.process(ChunkMapReader.java:142)
at org.dita.dost.writer.AbstractDomFilter.read(AbstractDomFilter.java:55)
at org.dita.dost.reader.ChunkMapReader.read(ChunkMapReader.java:118)
at org.dita.dost.module.ChunkModule.execute(ChunkModule.java:80)
at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:70)
at org.dita.dost.invoker.ExtensibleAntInvoker.execute(ExtensibleAntInvoker.java:222)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:293)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
at org.apache.tools.ant.Task.perform(Task.java:348)
at org.apache.tools.ant.Target.execute(Target.java:435)
at org.apache.tools.ant.Target.performTasks(Target.java:456)
at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1405)
at org.apache.tools.ant.helper.SingleCheckExecutor.executeTargets(SingleCheckExecutor.java:38)
at org.apache.tools.ant.Project.executeTargets(Project.java:1260)
at org.apache.tools.ant.taskdefs.Ant.execute(Ant.java:441)
at org.apache.tools.ant.taskdefs.CallTarget.execute(CallTarget.java:105)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:293)
at sun.reflect.GeneratedMethodAccessor6.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
at org.apache.tools.ant.Task.perform(Task.java:348)
at org.apache.tools.ant.Target.execute(Target.java:435)
at org.apache.tools.ant.Target.performTasks(Target.java:456)
at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1405)
at org.apache.tools.ant.Project.executeTarget(Project.java:1376)
at org.apache.tools.ant.helper.DefaultExecutor.executeTargets(DefaultExecutor.java:41)
at org.apache.tools.ant.Project.executeTargets(Project.java:1260)
at org.apache.tools.ant.Main.runBuild(Main.java:854)
at org.apache.tools.ant.Main.startAnt(Main.java:236)
at org.apache.tools.ant.launch.Launcher.run(Launcher.java:285)
at org.apache.tools.ant.launch.Launcher.main(Launcher.java:112)
|
process
|
nullpointerexception when publishing a learning group map i m using dita ot and when i attempt to publish a learning group map to pdf or i m getting this error below i tested with dita ot and it is working fine chunk processing file a ditamap build failed the following error occurred while executing this line lib dita ot plugins org dita base build preprocess xml java lang nullpointerexception at org dita dost writer abstractchunktopicparser generateoutputfile abstractchunktopicparser java at org dita dost writer chunktopicparser processchunk chunktopicparser java at org dita dost writer chunktopicparser write chunktopicparser java at org dita dost reader chunkmapreader processcombinechunk chunkmapreader java at org dita dost reader chunkmapreader processtopicref chunkmapreader java at org dita dost reader chunkmapreader processchildtopicref chunkmapreader java at org dita dost reader chunkmapreader processtopicref chunkmapreader java at org dita dost reader chunkmapreader process chunkmapreader java at org dita dost writer abstractdomfilter read abstractdomfilter java at org dita dost reader chunkmapreader read chunkmapreader java at org dita dost module chunkmodule execute chunkmodule java at org dita dost pipeline pipelinefacade execute pipelinefacade java at org dita dost invoker extensibleantinvoker execute extensibleantinvoker java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache tools ant dispatch dispatchutils execute dispatchutils java at org apache tools ant task perform task java at org apache tools ant target execute target java at org apache tools ant target performtasks target java at org apache tools ant project executesortedtargets project java at org apache tools ant helper singlecheckexecutor executetargets singlecheckexecutor java at org apache tools ant project executetargets project java at org apache tools ant taskdefs ant execute ant java at org apache tools ant taskdefs calltarget execute calltarget java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache tools ant dispatch dispatchutils execute dispatchutils java at org apache tools ant task perform task java at org apache tools ant target execute target java at org apache tools ant target performtasks target java at org apache tools ant project executesortedtargets project java at org apache tools ant project executetarget project java at org apache tools ant helper defaultexecutor executetargets defaultexecutor java at org apache tools ant project executetargets project java at org apache tools ant main runbuild main java at org apache tools ant main startant main java at org apache tools ant launch launcher run launcher java at org apache tools ant launch launcher main launcher java
| 1
|
1,951
| 4,773,979,274
|
IssuesEvent
|
2016-10-27 03:49:37
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
opened
|
doc: need better link to signals list in process documentation
|
doc good first contribution process
|
* **Version**: all
* **Platform**: n/a
* **Subsystem**: doc
In the 'Signal Events' section of the `process` documentation, there is a link to a man page that supposedly lists the names of the various signals that can be listened for. However, the page that is currently linked ([sigaction(2)](http://man7.org/linux/man-pages/man2/sigaction.2.html)) doesn't present all of the valid signals and the ones that it does show are not formatted very well. I think a better link would be to [signal(7)](http://man7.org/linux/man-pages/man7/signal.7.html), which *does* list all of the signal names and in an easy to spot/read format.
|
1.0
|
doc: need better link to signals list in process documentation - * **Version**: all
* **Platform**: n/a
* **Subsystem**: doc
In the 'Signal Events' section of the `process` documentation, there is a link to a man page that supposedly lists the names of the various signals that can be listened for. However, the page that is currently linked ([sigaction(2)](http://man7.org/linux/man-pages/man2/sigaction.2.html)) doesn't present all of the valid signals and the ones that it does show are not formatted very well. I think a better link would be to [signal(7)](http://man7.org/linux/man-pages/man7/signal.7.html), which *does* list all of the signal names and in an easy to spot/read format.
|
process
|
doc need better link to signals list in process documentation version all platform n a subsystem doc in the signal events section of the process documentation there is a link to a man page that supposedly lists the names of the various signals that can be listened for however the page that is currently linked doesn t present all of the valid signals and the ones that it does show are not formatted very well i think a better link would be to which does list all of the signal names and in an easy to spot read format
| 1
|
16,455
| 21,334,029,560
|
IssuesEvent
|
2022-04-18 12:26:06
|
km4ack/patmenu2
|
https://api.github.com/repos/km4ack/patmenu2
|
closed
|
Backup/Save current config file
|
enhancement in process
|
There should be a way to save/backup the currently loaded config file.
|
1.0
|
Backup/Save current config file - There should be a way to save/backup the currently loaded config file.
|
process
|
backup save current config file there should be a way to save backup the currently loaded config file
| 1
|
18,886
| 3,091,421,457
|
IssuesEvent
|
2015-08-26 13:07:35
|
kattunga/uib
|
https://api.github.com/repos/kattunga/uib
|
closed
|
Exceptions text better readability
|
auto-migrated Priority-Medium Type-Defect
|
```
UibLib.pas
Line 1528
was: Exception.Message := Exception.Message + 'Error Code: ' +
IntToStr(Exception.FErrorCode);
better: Exception.Message := Exception.Message + '. Error Code: ' +
IntToStr(Exception.FErrorCode);
Error Code is to be separated from previous GDS Code.
```
Original issue reported on code.google.com by `Arioch...@gmail.com` on 31 Oct 2011 at 7:56
|
1.0
|
Exceptions text better readability - ```
UibLib.pas
Line 1528
was: Exception.Message := Exception.Message + 'Error Code: ' +
IntToStr(Exception.FErrorCode);
better: Exception.Message := Exception.Message + '. Error Code: ' +
IntToStr(Exception.FErrorCode);
Error Code is to be separated from previous GDS Code.
```
Original issue reported on code.google.com by `Arioch...@gmail.com` on 31 Oct 2011 at 7:56
|
non_process
|
exceptions text better readability uiblib pas line was exception message exception message error code inttostr exception ferrorcode better exception message exception message error code inttostr exception ferrorcode error code is to be separated from previous gds code original issue reported on code google com by arioch gmail com on oct at
| 0
|
16,353
| 9,784,102,245
|
IssuesEvent
|
2019-06-08 16:15:50
|
alexcorvi/apexo
|
https://api.github.com/repos/alexcorvi/apexo
|
closed
|
CVE-2019-6283 (Medium) detected in node-sass-v4.12.0
|
security vulnerability
|
## CVE-2019-6283 - Medium Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.12.0</b></p></summary>
<p>
<p>:rainbow: Node.js bindings to libsass</p>
<p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Library Source Files (125)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /apexo/node_modules/node-sass/src/libsass/src/expand.hpp
- /apexo/node_modules/node-sass/src/libsass/src/color_maps.cpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_util.hpp
- /apexo/node_modules/node-sass/src/libsass/src/utf8/unchecked.h
- /apexo/node_modules/node-sass/src/libsass/src/output.hpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_values.hpp
- /apexo/node_modules/node-sass/src/libsass/src/util.hpp
- /apexo/node_modules/node-sass/src/libsass/src/emitter.hpp
- /apexo/node_modules/node-sass/src/libsass/src/lexer.cpp
- /apexo/node_modules/node-sass/src/libsass/test/test_node.cpp
- /apexo/node_modules/node-sass/src/libsass/src/plugins.cpp
- /apexo/node_modules/node-sass/src/libsass/include/sass/base.h
- /apexo/node_modules/node-sass/src/libsass/src/position.hpp
- /apexo/node_modules/node-sass/src/libsass/src/subset_map.hpp
- /apexo/node_modules/node-sass/src/libsass/src/operation.hpp
- /apexo/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp
- /apexo/node_modules/node-sass/src/libsass/src/error_handling.hpp
- /apexo/node_modules/node-sass/src/custom_importer_bridge.cpp
- /apexo/node_modules/node-sass/src/libsass/contrib/plugin.cpp
- /apexo/node_modules/node-sass/src/libsass/src/functions.hpp
- /apexo/node_modules/node-sass/src/libsass/test/test_superselector.cpp
- /apexo/node_modules/node-sass/src/libsass/src/eval.hpp
- /apexo/node_modules/node-sass/src/libsass/src/utf8_string.hpp
- /apexo/node_modules/node-sass/src/sass_context_wrapper.h
- /apexo/node_modules/node-sass/src/libsass/src/error_handling.cpp
- /apexo/node_modules/node-sass/src/libsass/src/node.cpp
- /apexo/node_modules/node-sass/src/libsass/src/parser.cpp
- /apexo/node_modules/node-sass/src/libsass/src/subset_map.cpp
- /apexo/node_modules/node-sass/src/libsass/src/emitter.cpp
- /apexo/node_modules/node-sass/src/libsass/src/listize.cpp
- /apexo/node_modules/node-sass/src/libsass/src/ast.hpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_functions.hpp
- /apexo/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp
- /apexo/node_modules/node-sass/src/libsass/src/output.cpp
- /apexo/node_modules/node-sass/src/libsass/src/check_nesting.cpp
- /apexo/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp
- /apexo/node_modules/node-sass/src/libsass/src/functions.cpp
- /apexo/node_modules/node-sass/src/libsass/src/cssize.hpp
- /apexo/node_modules/node-sass/src/libsass/src/prelexer.cpp
- /apexo/node_modules/node-sass/src/libsass/src/paths.hpp
- /apexo/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp
- /apexo/node_modules/node-sass/src/libsass/src/inspect.hpp
- /apexo/node_modules/node-sass/src/sass_types/color.cpp
- /apexo/node_modules/node-sass/src/libsass/test/test_unification.cpp
- /apexo/node_modules/node-sass/src/libsass/src/values.cpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_util.cpp
- /apexo/node_modules/node-sass/src/libsass/src/source_map.hpp
- /apexo/node_modules/node-sass/src/sass_types/list.h
- /apexo/node_modules/node-sass/src/libsass/src/check_nesting.hpp
- /apexo/node_modules/node-sass/src/libsass/src/json.cpp
- /apexo/node_modules/node-sass/src/libsass/src/units.cpp
- /apexo/node_modules/node-sass/src/libsass/src/units.hpp
- /apexo/node_modules/node-sass/src/libsass/src/context.cpp
- /apexo/node_modules/node-sass/src/libsass/src/utf8/checked.h
- /apexo/node_modules/node-sass/src/libsass/src/listize.hpp
- /apexo/node_modules/node-sass/src/sass_types/string.cpp
- /apexo/node_modules/node-sass/src/libsass/src/prelexer.hpp
- /apexo/node_modules/node-sass/src/libsass/src/context.hpp
- /apexo/node_modules/node-sass/src/sass_types/boolean.h
- /apexo/node_modules/node-sass/src/libsass/include/sass2scss.h
- /apexo/node_modules/node-sass/src/libsass/src/eval.cpp
- /apexo/node_modules/node-sass/src/libsass/src/expand.cpp
- /apexo/node_modules/node-sass/src/sass_types/factory.cpp
- /apexo/node_modules/node-sass/src/libsass/src/operators.cpp
- /apexo/node_modules/node-sass/src/sass_types/boolean.cpp
- /apexo/node_modules/node-sass/src/libsass/src/source_map.cpp
- /apexo/node_modules/node-sass/src/sass_types/value.h
- /apexo/node_modules/node-sass/src/libsass/src/utf8_string.cpp
- /apexo/node_modules/node-sass/src/callback_bridge.h
- /apexo/node_modules/node-sass/src/libsass/src/file.cpp
- /apexo/node_modules/node-sass/src/libsass/src/sass.cpp
- /apexo/node_modules/node-sass/src/libsass/src/node.hpp
- /apexo/node_modules/node-sass/src/libsass/src/environment.cpp
- /apexo/node_modules/node-sass/src/libsass/src/extend.hpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_context.hpp
- /apexo/node_modules/node-sass/src/libsass/src/operators.hpp
- /apexo/node_modules/node-sass/src/libsass/src/constants.hpp
- /apexo/node_modules/node-sass/src/libsass/src/sass.hpp
- /apexo/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp
- /apexo/node_modules/node-sass/src/libsass/src/parser.hpp
- /apexo/node_modules/node-sass/src/libsass/src/constants.cpp
- /apexo/node_modules/node-sass/src/sass_types/list.cpp
- /apexo/node_modules/node-sass/src/libsass/src/cssize.cpp
- /apexo/node_modules/node-sass/src/libsass/include/sass/functions.h
- /apexo/node_modules/node-sass/src/libsass/src/util.cpp
- /apexo/node_modules/node-sass/src/custom_function_bridge.cpp
- /apexo/node_modules/node-sass/src/custom_importer_bridge.h
- /apexo/node_modules/node-sass/src/libsass/src/bind.cpp
- /apexo/node_modules/node-sass/src/libsass/src/inspect.cpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_functions.cpp
- /apexo/node_modules/node-sass/src/libsass/src/backtrace.cpp
- /apexo/node_modules/node-sass/src/libsass/src/extend.cpp
- /apexo/node_modules/node-sass/src/sass_types/sass_value_wrapper.h
- /apexo/node_modules/node-sass/src/libsass/src/debugger.hpp
- /apexo/node_modules/node-sass/src/libsass/src/cencode.c
- /apexo/node_modules/node-sass/src/libsass/src/base64vlq.cpp
- /apexo/node_modules/node-sass/src/sass_types/number.cpp
- /apexo/node_modules/node-sass/src/sass_types/color.h
- /apexo/node_modules/node-sass/src/libsass/src/c99func.c
- /apexo/node_modules/node-sass/src/libsass/src/position.cpp
- /apexo/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_values.cpp
- /apexo/node_modules/node-sass/src/libsass/include/sass/values.h
- /apexo/node_modules/node-sass/src/libsass/test/test_subset_map.cpp
- /apexo/node_modules/node-sass/src/libsass/src/sass2scss.cpp
- /apexo/node_modules/node-sass/src/sass_types/null.cpp
- /apexo/node_modules/node-sass/src/libsass/src/ast.cpp
- /apexo/node_modules/node-sass/src/libsass/include/sass/context.h
- /apexo/node_modules/node-sass/src/libsass/src/to_c.cpp
- /apexo/node_modules/node-sass/src/libsass/src/to_value.hpp
- /apexo/node_modules/node-sass/src/libsass/src/color_maps.hpp
- /apexo/node_modules/node-sass/src/sass_context_wrapper.cpp
- /apexo/node_modules/node-sass/src/libsass/script/test-leaks.pl
- /apexo/node_modules/node-sass/src/libsass/src/lexer.hpp
- /apexo/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp
- /apexo/node_modules/node-sass/src/libsass/src/to_c.hpp
- /apexo/node_modules/node-sass/src/sass_types/map.cpp
- /apexo/node_modules/node-sass/src/libsass/src/to_value.cpp
- /apexo/node_modules/node-sass/src/libsass/src/b64/encode.h
- /apexo/node_modules/node-sass/src/libsass/src/file.hpp
- /apexo/node_modules/node-sass/src/libsass/src/environment.hpp
- /apexo/node_modules/node-sass/src/libsass/src/plugins.hpp
- /apexo/node_modules/node-sass/src/binding.cpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_context.cpp
- /apexo/node_modules/node-sass/src/libsass/src/debug.hpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::parenthese_scope in prelexer.hpp.
<p>Publish Date: 2019-01-14
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6283>CVE-2019-6283</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-6283 (Medium) detected in node-sass-v4.12.0 - ## CVE-2019-6283 - Medium Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.12.0</b></p></summary>
<p>
<p>:rainbow: Node.js bindings to libsass</p>
<p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Library Source Files (125)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /apexo/node_modules/node-sass/src/libsass/src/expand.hpp
- /apexo/node_modules/node-sass/src/libsass/src/color_maps.cpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_util.hpp
- /apexo/node_modules/node-sass/src/libsass/src/utf8/unchecked.h
- /apexo/node_modules/node-sass/src/libsass/src/output.hpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_values.hpp
- /apexo/node_modules/node-sass/src/libsass/src/util.hpp
- /apexo/node_modules/node-sass/src/libsass/src/emitter.hpp
- /apexo/node_modules/node-sass/src/libsass/src/lexer.cpp
- /apexo/node_modules/node-sass/src/libsass/test/test_node.cpp
- /apexo/node_modules/node-sass/src/libsass/src/plugins.cpp
- /apexo/node_modules/node-sass/src/libsass/include/sass/base.h
- /apexo/node_modules/node-sass/src/libsass/src/position.hpp
- /apexo/node_modules/node-sass/src/libsass/src/subset_map.hpp
- /apexo/node_modules/node-sass/src/libsass/src/operation.hpp
- /apexo/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp
- /apexo/node_modules/node-sass/src/libsass/src/error_handling.hpp
- /apexo/node_modules/node-sass/src/custom_importer_bridge.cpp
- /apexo/node_modules/node-sass/src/libsass/contrib/plugin.cpp
- /apexo/node_modules/node-sass/src/libsass/src/functions.hpp
- /apexo/node_modules/node-sass/src/libsass/test/test_superselector.cpp
- /apexo/node_modules/node-sass/src/libsass/src/eval.hpp
- /apexo/node_modules/node-sass/src/libsass/src/utf8_string.hpp
- /apexo/node_modules/node-sass/src/sass_context_wrapper.h
- /apexo/node_modules/node-sass/src/libsass/src/error_handling.cpp
- /apexo/node_modules/node-sass/src/libsass/src/node.cpp
- /apexo/node_modules/node-sass/src/libsass/src/parser.cpp
- /apexo/node_modules/node-sass/src/libsass/src/subset_map.cpp
- /apexo/node_modules/node-sass/src/libsass/src/emitter.cpp
- /apexo/node_modules/node-sass/src/libsass/src/listize.cpp
- /apexo/node_modules/node-sass/src/libsass/src/ast.hpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_functions.hpp
- /apexo/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp
- /apexo/node_modules/node-sass/src/libsass/src/output.cpp
- /apexo/node_modules/node-sass/src/libsass/src/check_nesting.cpp
- /apexo/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp
- /apexo/node_modules/node-sass/src/libsass/src/functions.cpp
- /apexo/node_modules/node-sass/src/libsass/src/cssize.hpp
- /apexo/node_modules/node-sass/src/libsass/src/prelexer.cpp
- /apexo/node_modules/node-sass/src/libsass/src/paths.hpp
- /apexo/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp
- /apexo/node_modules/node-sass/src/libsass/src/inspect.hpp
- /apexo/node_modules/node-sass/src/sass_types/color.cpp
- /apexo/node_modules/node-sass/src/libsass/test/test_unification.cpp
- /apexo/node_modules/node-sass/src/libsass/src/values.cpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_util.cpp
- /apexo/node_modules/node-sass/src/libsass/src/source_map.hpp
- /apexo/node_modules/node-sass/src/sass_types/list.h
- /apexo/node_modules/node-sass/src/libsass/src/check_nesting.hpp
- /apexo/node_modules/node-sass/src/libsass/src/json.cpp
- /apexo/node_modules/node-sass/src/libsass/src/units.cpp
- /apexo/node_modules/node-sass/src/libsass/src/units.hpp
- /apexo/node_modules/node-sass/src/libsass/src/context.cpp
- /apexo/node_modules/node-sass/src/libsass/src/utf8/checked.h
- /apexo/node_modules/node-sass/src/libsass/src/listize.hpp
- /apexo/node_modules/node-sass/src/sass_types/string.cpp
- /apexo/node_modules/node-sass/src/libsass/src/prelexer.hpp
- /apexo/node_modules/node-sass/src/libsass/src/context.hpp
- /apexo/node_modules/node-sass/src/sass_types/boolean.h
- /apexo/node_modules/node-sass/src/libsass/include/sass2scss.h
- /apexo/node_modules/node-sass/src/libsass/src/eval.cpp
- /apexo/node_modules/node-sass/src/libsass/src/expand.cpp
- /apexo/node_modules/node-sass/src/sass_types/factory.cpp
- /apexo/node_modules/node-sass/src/libsass/src/operators.cpp
- /apexo/node_modules/node-sass/src/sass_types/boolean.cpp
- /apexo/node_modules/node-sass/src/libsass/src/source_map.cpp
- /apexo/node_modules/node-sass/src/sass_types/value.h
- /apexo/node_modules/node-sass/src/libsass/src/utf8_string.cpp
- /apexo/node_modules/node-sass/src/callback_bridge.h
- /apexo/node_modules/node-sass/src/libsass/src/file.cpp
- /apexo/node_modules/node-sass/src/libsass/src/sass.cpp
- /apexo/node_modules/node-sass/src/libsass/src/node.hpp
- /apexo/node_modules/node-sass/src/libsass/src/environment.cpp
- /apexo/node_modules/node-sass/src/libsass/src/extend.hpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_context.hpp
- /apexo/node_modules/node-sass/src/libsass/src/operators.hpp
- /apexo/node_modules/node-sass/src/libsass/src/constants.hpp
- /apexo/node_modules/node-sass/src/libsass/src/sass.hpp
- /apexo/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp
- /apexo/node_modules/node-sass/src/libsass/src/parser.hpp
- /apexo/node_modules/node-sass/src/libsass/src/constants.cpp
- /apexo/node_modules/node-sass/src/sass_types/list.cpp
- /apexo/node_modules/node-sass/src/libsass/src/cssize.cpp
- /apexo/node_modules/node-sass/src/libsass/include/sass/functions.h
- /apexo/node_modules/node-sass/src/libsass/src/util.cpp
- /apexo/node_modules/node-sass/src/custom_function_bridge.cpp
- /apexo/node_modules/node-sass/src/custom_importer_bridge.h
- /apexo/node_modules/node-sass/src/libsass/src/bind.cpp
- /apexo/node_modules/node-sass/src/libsass/src/inspect.cpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_functions.cpp
- /apexo/node_modules/node-sass/src/libsass/src/backtrace.cpp
- /apexo/node_modules/node-sass/src/libsass/src/extend.cpp
- /apexo/node_modules/node-sass/src/sass_types/sass_value_wrapper.h
- /apexo/node_modules/node-sass/src/libsass/src/debugger.hpp
- /apexo/node_modules/node-sass/src/libsass/src/cencode.c
- /apexo/node_modules/node-sass/src/libsass/src/base64vlq.cpp
- /apexo/node_modules/node-sass/src/sass_types/number.cpp
- /apexo/node_modules/node-sass/src/sass_types/color.h
- /apexo/node_modules/node-sass/src/libsass/src/c99func.c
- /apexo/node_modules/node-sass/src/libsass/src/position.cpp
- /apexo/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_values.cpp
- /apexo/node_modules/node-sass/src/libsass/include/sass/values.h
- /apexo/node_modules/node-sass/src/libsass/test/test_subset_map.cpp
- /apexo/node_modules/node-sass/src/libsass/src/sass2scss.cpp
- /apexo/node_modules/node-sass/src/sass_types/null.cpp
- /apexo/node_modules/node-sass/src/libsass/src/ast.cpp
- /apexo/node_modules/node-sass/src/libsass/include/sass/context.h
- /apexo/node_modules/node-sass/src/libsass/src/to_c.cpp
- /apexo/node_modules/node-sass/src/libsass/src/to_value.hpp
- /apexo/node_modules/node-sass/src/libsass/src/color_maps.hpp
- /apexo/node_modules/node-sass/src/sass_context_wrapper.cpp
- /apexo/node_modules/node-sass/src/libsass/script/test-leaks.pl
- /apexo/node_modules/node-sass/src/libsass/src/lexer.hpp
- /apexo/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp
- /apexo/node_modules/node-sass/src/libsass/src/to_c.hpp
- /apexo/node_modules/node-sass/src/sass_types/map.cpp
- /apexo/node_modules/node-sass/src/libsass/src/to_value.cpp
- /apexo/node_modules/node-sass/src/libsass/src/b64/encode.h
- /apexo/node_modules/node-sass/src/libsass/src/file.hpp
- /apexo/node_modules/node-sass/src/libsass/src/environment.hpp
- /apexo/node_modules/node-sass/src/libsass/src/plugins.hpp
- /apexo/node_modules/node-sass/src/binding.cpp
- /apexo/node_modules/node-sass/src/libsass/src/sass_context.cpp
- /apexo/node_modules/node-sass/src/libsass/src/debug.hpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::parenthese_scope in prelexer.hpp.
<p>Publish Date: 2019-01-14
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6283>CVE-2019-6283</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in node sass cve medium severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries apexo node modules node sass src libsass src expand hpp apexo node modules node sass src libsass src color maps cpp apexo node modules node sass src libsass src sass util hpp apexo node modules node sass src libsass src unchecked h apexo node modules node sass src libsass src output hpp apexo node modules node sass src libsass src sass values hpp apexo node modules node sass src libsass src util hpp apexo node modules node sass src libsass src emitter hpp apexo node modules node sass src libsass src lexer cpp apexo node modules node sass src libsass test test node cpp apexo node modules node sass src libsass src plugins cpp apexo node modules node sass src libsass include sass base h apexo node modules node sass src libsass src position hpp apexo node modules node sass src libsass src subset map hpp apexo node modules node sass src libsass src operation hpp apexo node modules node sass src libsass src remove placeholders cpp apexo node modules node sass src libsass src error handling hpp apexo node modules node sass src custom importer bridge cpp apexo node modules node sass src libsass contrib plugin cpp apexo node modules node sass src libsass src functions hpp apexo node modules node sass src libsass test test superselector cpp apexo node modules node sass src libsass src eval hpp apexo node modules node sass src libsass src string hpp apexo node modules node sass src sass context wrapper h apexo node modules node sass src libsass src error handling cpp apexo node modules node sass src libsass src node cpp apexo node modules node sass src libsass src parser cpp apexo node modules node sass src libsass src subset map cpp apexo node modules node sass src libsass src emitter cpp apexo node modules node sass src libsass src listize cpp apexo node modules node sass src libsass src ast hpp apexo node modules node sass src libsass src sass functions hpp apexo node modules node sass src libsass src memory sharedptr cpp apexo node modules node sass src libsass src output cpp apexo node modules node sass src libsass src check nesting cpp apexo node modules node sass src libsass src ast def macros hpp apexo node modules node sass src libsass src functions cpp apexo node modules node sass src libsass src cssize hpp apexo node modules node sass src libsass src prelexer cpp apexo node modules node sass src libsass src paths hpp apexo node modules node sass src libsass src ast fwd decl hpp apexo node modules node sass src libsass src inspect hpp apexo node modules node sass src sass types color cpp apexo node modules node sass src libsass test test unification cpp apexo node modules node sass src libsass src values cpp apexo node modules node sass src libsass src sass util cpp apexo node modules node sass src libsass src source map hpp apexo node modules node sass src sass types list h apexo node modules node sass src libsass src check nesting hpp apexo node modules node sass src libsass src json cpp apexo node modules node sass src libsass src units cpp apexo node modules node sass src libsass src units hpp apexo node modules node sass src libsass src context cpp apexo node modules node sass src libsass src checked h apexo node modules node sass src libsass src listize hpp apexo node modules node sass src sass types string cpp apexo node modules node sass src libsass src prelexer hpp apexo node modules node sass src libsass src context hpp apexo node modules node sass src sass types boolean h apexo node modules node sass src libsass include h apexo node modules node sass src libsass src eval cpp apexo node modules node sass src libsass src expand cpp apexo node modules node sass src sass types factory cpp apexo node modules node sass src libsass src operators cpp apexo node modules node sass src sass types boolean cpp apexo node modules node sass src libsass src source map cpp apexo node modules node sass src sass types value h apexo node modules node sass src libsass src string cpp apexo node modules node sass src callback bridge h apexo node modules node sass src libsass src file cpp apexo node modules node sass src libsass src sass cpp apexo node modules node sass src libsass src node hpp apexo node modules node sass src libsass src environment cpp apexo node modules node sass src libsass src extend hpp apexo node modules node sass src libsass src sass context hpp apexo node modules node sass src libsass src operators hpp apexo node modules node sass src libsass src constants hpp apexo node modules node sass src libsass src sass hpp apexo node modules node sass src libsass src ast fwd decl cpp apexo node modules node sass src libsass src parser hpp apexo node modules node sass src libsass src constants cpp apexo node modules node sass src sass types list cpp apexo node modules node sass src libsass src cssize cpp apexo node modules node sass src libsass include sass functions h apexo node modules node sass src libsass src util cpp apexo node modules node sass src custom function bridge cpp apexo node modules node sass src custom importer bridge h apexo node modules node sass src libsass src bind cpp apexo node modules node sass src libsass src inspect cpp apexo node modules node sass src libsass src sass functions cpp apexo node modules node sass src libsass src backtrace cpp apexo node modules node sass src libsass src extend cpp apexo node modules node sass src sass types sass value wrapper h apexo node modules node sass src libsass src debugger hpp apexo node modules node sass src libsass src cencode c apexo node modules node sass src libsass src cpp apexo node modules node sass src sass types number cpp apexo node modules node sass src sass types color h apexo node modules node sass src libsass src c apexo node modules node sass src libsass src position cpp apexo node modules node sass src libsass src remove placeholders hpp apexo node modules node sass src libsass src sass values cpp apexo node modules node sass src libsass include sass values h apexo node modules node sass src libsass test test subset map cpp apexo node modules node sass src libsass src cpp apexo node modules node sass src sass types null cpp apexo node modules node sass src libsass src ast cpp apexo node modules node sass src libsass include sass context h apexo node modules node sass src libsass src to c cpp apexo node modules node sass src libsass src to value hpp apexo node modules node sass src libsass src color maps hpp apexo node modules node sass src sass context wrapper cpp apexo node modules node sass src libsass script test leaks pl apexo node modules node sass src libsass src lexer hpp apexo node modules node sass src libsass src memory sharedptr hpp apexo node modules node sass src libsass src to c hpp apexo node modules node sass src sass types map cpp apexo node modules node sass src libsass src to value cpp apexo node modules node sass src libsass src encode h apexo node modules node sass src libsass src file hpp apexo node modules node sass src libsass src environment hpp apexo node modules node sass src libsass src plugins hpp apexo node modules node sass src binding cpp apexo node modules node sass src libsass src sass context cpp apexo node modules node sass src libsass src debug hpp vulnerability details in libsass a heap based buffer over read exists in sass prelexer parenthese scope in prelexer hpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource
| 0
|
252,087
| 21,556,148,023
|
IssuesEvent
|
2022-04-30 13:00:42
|
poikilos/EnlivenMinetest
|
https://api.github.com/repos/poikilos/EnlivenMinetest
|
closed
|
airitems (air tools) textures are blurry
|
fixed - resolved in next Final Minetest release patch available externally
|
They were scaled up using an edge-detecting resize method then scaled back down. airitems 16x16 originals are at https://git.minetest.org/zander999/air_items
mods/coderbuild/coderblocks/textures/
- [x] air_pick.png
- [x] air_pick_upgraded.png
- [x] air_sword.png
Patchable via (set BUCKET_GAME then):
```
cd /tmp
git clone https://git.minetest.org/zander999/air_items.git /tmp/air_items
cd /tmp/air_items
cp textures/air_pick.png $BUCKET_GAME/mods/coderbuild/coderblocks/textures/
cp textures/air_pick_upgraded.png $BUCKET_GAME/mods/coderbuild/coderblocks/textures/
cp textures/air_sword.png $BUCKET_GAME/mods/coderbuild/coderblocks/textures/
```
|
1.0
|
airitems (air tools) textures are blurry - They were scaled up using an edge-detecting resize method then scaled back down. airitems 16x16 originals are at https://git.minetest.org/zander999/air_items
mods/coderbuild/coderblocks/textures/
- [x] air_pick.png
- [x] air_pick_upgraded.png
- [x] air_sword.png
Patchable via (set BUCKET_GAME then):
```
cd /tmp
git clone https://git.minetest.org/zander999/air_items.git /tmp/air_items
cd /tmp/air_items
cp textures/air_pick.png $BUCKET_GAME/mods/coderbuild/coderblocks/textures/
cp textures/air_pick_upgraded.png $BUCKET_GAME/mods/coderbuild/coderblocks/textures/
cp textures/air_sword.png $BUCKET_GAME/mods/coderbuild/coderblocks/textures/
```
|
non_process
|
airitems air tools textures are blurry they were scaled up using an edge detecting resize method then scaled back down airitems originals are at mods coderbuild coderblocks textures air pick png air pick upgraded png air sword png patchable via set bucket game then cd tmp git clone tmp air items cd tmp air items cp textures air pick png bucket game mods coderbuild coderblocks textures cp textures air pick upgraded png bucket game mods coderbuild coderblocks textures cp textures air sword png bucket game mods coderbuild coderblocks textures
| 0
|
36,677
| 2,810,587,032
|
IssuesEvent
|
2015-05-17 01:11:24
|
yearography/web
|
https://api.github.com/repos/yearography/web
|
closed
|
Invalid character found in landing page
|
category: frontend priority: now type: bug
|
It would appear as though some jackass left an 'e' in the header of the landing page. Please fix asap.
|
1.0
|
Invalid character found in landing page - It would appear as though some jackass left an 'e' in the header of the landing page. Please fix asap.
|
non_process
|
invalid character found in landing page it would appear as though some jackass left an e in the header of the landing page please fix asap
| 0
|
197,638
| 15,687,108,527
|
IssuesEvent
|
2021-03-25 13:19:19
|
hdl/awesome
|
https://api.github.com/repos/hdl/awesome
|
opened
|
How to handle BibTeX references?
|
documentation enhancement help wanted theme
|
In the context of #184, @rodrigomelo9 asked https://github.com/verilog-to-routing/vtr-verilog-to-routing/issues/1691. There, several academic/bibliographic references were provided. We don't have a mechanism for showing that content in the site. However, since many of the open source EDA projects are related to academia and research publications, it would be very desirable to support BibTeX.
There are some solutions:
- https://github.com/wowchemy/hugo-academic-cli
- https://gohugo.io/content-management/formats/#external-helper-asciidoctor
Yet, either of them requires some effort. They don't seem to be built-in.
Meanwhile, we might have a `*.bib` file and suggest users to browse it with [JabRef](https://www.jabref.org/), [Zotero](https://www.zotero.org/), etc.
|
1.0
|
How to handle BibTeX references? - In the context of #184, @rodrigomelo9 asked https://github.com/verilog-to-routing/vtr-verilog-to-routing/issues/1691. There, several academic/bibliographic references were provided. We don't have a mechanism for showing that content in the site. However, since many of the open source EDA projects are related to academia and research publications, it would be very desirable to support BibTeX.
There are some solutions:
- https://github.com/wowchemy/hugo-academic-cli
- https://gohugo.io/content-management/formats/#external-helper-asciidoctor
Yet, either of them requires some effort. They don't seem to be built-in.
Meanwhile, we might have a `*.bib` file and suggest users to browse it with [JabRef](https://www.jabref.org/), [Zotero](https://www.zotero.org/), etc.
|
non_process
|
how to handle bibtex references in the context of asked there several academic bibliographic references were provided we don t have a mechanism for showing that content in the site however since many of the open source eda projects are related to academia and research publications it would be very desirable to support bibtex there are some solutions yet either of them requires some effort they don t seem to be built in meanwhile we might have a bib file and suggest users to browse it with etc
| 0
|
57,672
| 24,187,760,967
|
IssuesEvent
|
2022-09-23 14:40:39
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
closed
|
Add Arterial Management & Austin Police Department to the Coordination Partners list
|
Impact: 2-Major Service: Dev Need: 1-Must Have Workgroup: VZ Type: Enhancement
|
Staff would like to add Arterial Management and Austin Police Department as additional options for selections from the Coordination Partners drop down in the Recommendations section.
Please add them to the `atd__coordination_partners_lkp` table.

|
1.0
|
Add Arterial Management & Austin Police Department to the Coordination Partners list - Staff would like to add Arterial Management and Austin Police Department as additional options for selections from the Coordination Partners drop down in the Recommendations section.
Please add them to the `atd__coordination_partners_lkp` table.

|
non_process
|
add arterial management austin police department to the coordination partners list staff would like to add arterial management and austin police department as additional options for selections from the coordination partners drop down in the recommendations section please add them to the atd coordination partners lkp table
| 0
|
2,899
| 5,887,623,343
|
IssuesEvent
|
2017-05-17 08:01:29
|
g8os/core0
|
https://api.github.com/repos/g8os/core0
|
closed
|
Add a way to stream the output of a command
|
process_wontfix type_feature
|
It would be interesting to have a command which stream the output of a command run in a container. When starting a command, there is no way to know the status (stdout, stderr) of that given command right now, until it complete.
|
1.0
|
Add a way to stream the output of a command - It would be interesting to have a command which stream the output of a command run in a container. When starting a command, there is no way to know the status (stdout, stderr) of that given command right now, until it complete.
|
process
|
add a way to stream the output of a command it would be interesting to have a command which stream the output of a command run in a container when starting a command there is no way to know the status stdout stderr of that given command right now until it complete
| 1
|
10,530
| 13,309,453,683
|
IssuesEvent
|
2020-08-26 04:04:08
|
Zjmainstay/blog-comment
|
https://api.github.com/repos/Zjmainstay/blog-comment
|
opened
|
Shell+Crontab实现服务进程监控重启 - Zjmainstay学习笔记 | 是非明于学习,境界升于内省
|
Gitalk _shell-checkprocess
|
http://www.zjmainstay.cn/shell-checkprocess
本文提供一个shell脚本,结合crontab的使用可以实现对服务器服务进程的运行情况进行监控,及时重启服务器关键服务进程,避免关键服务进程长期终止情况发生。
|
1.0
|
Shell+Crontab实现服务进程监控重启 - Zjmainstay学习笔记 | 是非明于学习,境界升于内省 - http://www.zjmainstay.cn/shell-checkprocess
本文提供一个shell脚本,结合crontab的使用可以实现对服务器服务进程的运行情况进行监控,及时重启服务器关键服务进程,避免关键服务进程长期终止情况发生。
|
process
|
shell crontab实现服务进程监控重启 zjmainstay学习笔记 是非明于学习,境界升于内省 本文提供一个shell脚本,结合crontab的使用可以实现对服务器服务进程的运行情况进行监控,及时重启服务器关键服务进程,避免关键服务进程长期终止情况发生。
| 1
|
455
| 2,894,791,486
|
IssuesEvent
|
2015-06-16 03:13:01
|
hammerlab/pileup.js
|
https://api.github.com/repos/hammerlab/pileup.js
|
opened
|
Use async/await
|
process
|
Flow 0.12 added support for the async & await keywords. The Babel transpiler supports them. Using these features of ES6/7 would greatly simplify the async code in pileup.js.
https://github.com/facebook/flow/releases/tag/v0.12.0
|
1.0
|
Use async/await - Flow 0.12 added support for the async & await keywords. The Babel transpiler supports them. Using these features of ES6/7 would greatly simplify the async code in pileup.js.
https://github.com/facebook/flow/releases/tag/v0.12.0
|
process
|
use async await flow added support for the async await keywords the babel transpiler supports them using these features of would greatly simplify the async code in pileup js
| 1
|
10,371
| 6,689,312,033
|
IssuesEvent
|
2017-10-09 00:43:31
|
uqbar-project/wollok
|
https://api.github.com/repos/uqbar-project/wollok
|
closed
|
REPL continues to run after program has been modified
|
component: repl in progress usability
|
If you have a program running in the REPL and you modify it, the REPL continues running the old program.
I would like to give a try to update the program in execution, maybe it is easier than it appears at first sight. But still, if that is not possible, we should take some action when the running program is modified. Otherwise, programmers tend to get confused because they can't see any behavior change after program modifications.
Two simple actions, we should either
a. Stop de REPL, or
b. If we allow to continue running, at least add some kind of visual clue about running an obsolete program: change a color or icon, display a warning, etc.
|
True
|
REPL continues to run after program has been modified - If you have a program running in the REPL and you modify it, the REPL continues running the old program.
I would like to give a try to update the program in execution, maybe it is easier than it appears at first sight. But still, if that is not possible, we should take some action when the running program is modified. Otherwise, programmers tend to get confused because they can't see any behavior change after program modifications.
Two simple actions, we should either
a. Stop de REPL, or
b. If we allow to continue running, at least add some kind of visual clue about running an obsolete program: change a color or icon, display a warning, etc.
|
non_process
|
repl continues to run after program has been modified if you have a program running in the repl and you modify it the repl continues running the old program i would like to give a try to update the program in execution maybe it is easier than it appears at first sight but still if that is not possible we should take some action when the running program is modified otherwise programmers tend to get confused because they can t see any behavior change after program modifications two simple actions we should either a stop de repl or b if we allow to continue running at least add some kind of visual clue about running an obsolete program change a color or icon display a warning etc
| 0
|
14,628
| 17,767,680,194
|
IssuesEvent
|
2021-08-30 09:37:53
|
googleapis/nodejs-dns
|
https://api.github.com/repos/googleapis/nodejs-dns
|
closed
|
Dependency Dashboard
|
type: process api: dns
|
This issue contains a list of Renovate updates and their statuses.
## Awaiting Schedule
These updates are awaiting their schedule. Click on a checkbox to get an update now.
- [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2
## Ignored or Blocked
These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below.
- [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/460) (`mocha`, `@types/mocha`)
- [ ] <!-- recreate-branch=renovate/arrify-3.x -->[fix(deps): update dependency arrify to v3](../pull/448)
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Awaiting Schedule
These updates are awaiting their schedule. Click on a checkbox to get an update now.
- [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2
## Ignored or Blocked
These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below.
- [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/460) (`mocha`, `@types/mocha`)
- [ ] <!-- recreate-branch=renovate/arrify-3.x -->[fix(deps): update dependency arrify to v3](../pull/448)
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue contains a list of renovate updates and their statuses awaiting schedule these updates are awaiting their schedule click on a checkbox to get an update now chore deps update actions setup node action to ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull mocha types mocha pull check this box to trigger a request for renovate to run again on this repository
| 1
|
380,132
| 11,254,160,680
|
IssuesEvent
|
2020-01-11 21:25:04
|
Kaktushose/levelbot2
|
https://api.github.com/repos/Kaktushose/levelbot2
|
opened
|
update Command
|
enhancement low priority member team
|
**Steht dein Feature-Wunsch im Zusammenhang mit einem Problem? Bitte beschreibe es.**
Falls es passiert, dass der Bot einen User nicht erfasst, z.B. durch Ausfälle, fehlt dieser in der DB.
**Beschreibe die Lösung, die du gerne hättest**
Command `!update`, der die Datenbank auf Fehler prüft
|
1.0
|
update Command - **Steht dein Feature-Wunsch im Zusammenhang mit einem Problem? Bitte beschreibe es.**
Falls es passiert, dass der Bot einen User nicht erfasst, z.B. durch Ausfälle, fehlt dieser in der DB.
**Beschreibe die Lösung, die du gerne hättest**
Command `!update`, der die Datenbank auf Fehler prüft
|
non_process
|
update command steht dein feature wunsch im zusammenhang mit einem problem bitte beschreibe es falls es passiert dass der bot einen user nicht erfasst z b durch ausfälle fehlt dieser in der db beschreibe die lösung die du gerne hättest command update der die datenbank auf fehler prüft
| 0
|
34,642
| 12,294,674,151
|
IssuesEvent
|
2020-05-11 01:05:09
|
finos/secref-data
|
https://api.github.com/repos/finos/secref-data
|
opened
|
WS-2020-0070 (High) detected in lodash-4.17.15.tgz
|
security vulnerability
|
## WS-2020-0070 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.15.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/secref-data/website/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/secref-data/website/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- docusaurus-1.14.4.tgz (Root Library)
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
a prototype pollution vulnerability in lodash. It allows an attacker to inject properties on Object.prototype
<p>Publish Date: 2020-04-28
<p>URL: <a href=https://hackerone.com/reports/712065>WS-2020-0070</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.15","isTransitiveDependency":true,"dependencyTree":"docusaurus:1.14.4;lodash:4.17.15","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"WS-2020-0070","vulnerabilityDetails":"a prototype pollution vulnerability in lodash. It allows an attacker to inject properties on Object.prototype","vulnerabilityUrl":"https://hackerone.com/reports/712065","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
WS-2020-0070 (High) detected in lodash-4.17.15.tgz - ## WS-2020-0070 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.15.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/secref-data/website/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/secref-data/website/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- docusaurus-1.14.4.tgz (Root Library)
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
a prototype pollution vulnerability in lodash. It allows an attacker to inject properties on Object.prototype
<p>Publish Date: 2020-04-28
<p>URL: <a href=https://hackerone.com/reports/712065>WS-2020-0070</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.15","isTransitiveDependency":true,"dependencyTree":"docusaurus:1.14.4;lodash:4.17.15","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"WS-2020-0070","vulnerabilityDetails":"a prototype pollution vulnerability in lodash. It allows an attacker to inject properties on Object.prototype","vulnerabilityUrl":"https://hackerone.com/reports/712065","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
ws high detected in lodash tgz ws high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file tmp ws scm secref data website package json path to vulnerable library tmp ws scm secref data website node modules lodash package json dependency hierarchy docusaurus tgz root library x lodash tgz vulnerable library vulnerability details a prototype pollution vulnerability in lodash it allows an attacker to inject properties on object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails a prototype pollution vulnerability in lodash it allows an attacker to inject properties on object prototype vulnerabilityurl
| 0
|
16,193
| 20,674,211,081
|
IssuesEvent
|
2022-03-10 07:28:10
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
CockroachDB: Change the integer native types to the idiomatic ones
|
process/candidate topic: schema topic: native database types topic: schema validation engines/data model parser team/migrations topic: cockroachdb team/psl-wg
|
CockroachDB has two sets of names for integer column types, as described [in their docs](https://www.cockroachlabs.com/docs/v21.2/int).
We started by forking the postgresql connector to support crdb, so we currently work with the aliases. To recap this in a table:
<table>
<tr>
<td>Integer size
<td>Postgres
<td>CockroachDB
<tr>
<td>16 bits
<td>SMALLINT
<td>SMALLINT
<tr>
<td>32 bits
<td>INTEGER
<td>? (depends on session/server settings, may or may not be available, is not expressible by default)
<tr>
<td>64 bits
<td>BIGINT
<td>BIGINT / depends (INTEGER could be 64 bits, depending on settings)
</table>
There is however a solution to avoid ambiguities around integers: use the cockroachdb integer types `INT2`, `INT4` and `INT8`. We already use `INT4` in the migration engine to avoid ambiguities.
This issue is here to propose that we replace the ambiguous/complicated `@db.SmallInt`, `@db.Integer` and `@db.BigInt` native types with the unambiguous/clear `@db.Int2`, `@db.Int4` and `@db.Int8` within the cockroachdb provider.
|
1.0
|
CockroachDB: Change the integer native types to the idiomatic ones - CockroachDB has two sets of names for integer column types, as described [in their docs](https://www.cockroachlabs.com/docs/v21.2/int).
We started by forking the postgresql connector to support crdb, so we currently work with the aliases. To recap this in a table:
<table>
<tr>
<td>Integer size
<td>Postgres
<td>CockroachDB
<tr>
<td>16 bits
<td>SMALLINT
<td>SMALLINT
<tr>
<td>32 bits
<td>INTEGER
<td>? (depends on session/server settings, may or may not be available, is not expressible by default)
<tr>
<td>64 bits
<td>BIGINT
<td>BIGINT / depends (INTEGER could be 64 bits, depending on settings)
</table>
There is however a solution to avoid ambiguities around integers: use the cockroachdb integer types `INT2`, `INT4` and `INT8`. We already use `INT4` in the migration engine to avoid ambiguities.
This issue is here to propose that we replace the ambiguous/complicated `@db.SmallInt`, `@db.Integer` and `@db.BigInt` native types with the unambiguous/clear `@db.Int2`, `@db.Int4` and `@db.Int8` within the cockroachdb provider.
|
process
|
cockroachdb change the integer native types to the idiomatic ones cockroachdb has two sets of names for integer column types as described we started by forking the postgresql connector to support crdb so we currently work with the aliases to recap this in a table integer size postgres cockroachdb bits smallint smallint bits integer depends on session server settings may or may not be available is not expressible by default bits bigint bigint depends integer could be bits depending on settings there is however a solution to avoid ambiguities around integers use the cockroachdb integer types and we already use in the migration engine to avoid ambiguities this issue is here to propose that we replace the ambiguous complicated db smallint db integer and db bigint native types with the unambiguous clear db db and db within the cockroachdb provider
| 1
|
11,763
| 14,594,656,277
|
IssuesEvent
|
2020-12-20 07:11:56
|
rdoddanavar/hpr-sim
|
https://api.github.com/repos/rdoddanavar/hpr-sim
|
closed
|
util_unit.py: Temperature conversion
|
pre-processing
|
Implement lambda functions in `config_unit.yml` w/ corresponding routine in `util_unit.py` to handle more complex unit conversions
|
1.0
|
util_unit.py: Temperature conversion - Implement lambda functions in `config_unit.yml` w/ corresponding routine in `util_unit.py` to handle more complex unit conversions
|
process
|
util unit py temperature conversion implement lambda functions in config unit yml w corresponding routine in util unit py to handle more complex unit conversions
| 1
|
20,112
| 26,652,476,344
|
IssuesEvent
|
2023-01-25 14:38:16
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
How to understand decision_wait of tail_sampling
|
bug question processor/tailsampling
|
### Component(s)
processor/tailsampling
### What happened?
## Description
the decision_wait describe by:
```
decision_wait (default = 30s): Wait time since the first span of a trace before making a sampling decision
````
When I make two policies to get long-request-time traces and the error traces like this:
```
tail_sampling:
decision_wait: 60s
num_traces: 5000
expected_new_traces_per_sec: 10
policies:
[
{
name: errors-policy,
type: numeric_attribute,
numeric_attribute: {key: http.status_code, min_value: 499, max_value: 600}
},
{
name: latency-5s,
type: latency,
latency: {threshold_ms: 5000}
},
]
```
So if the trace is error already, the processor will delivery the trace in 60s later?
In the real world, Microservice cross call make more than 60s long-request. So If I want to catch the threshold_ms>60000, the decision_wait must more than 60s, better about decision_wait=80s. And all the trace will store to the Jaeger at least delay decision_wait=80s later?
Correct me if I'm wrong, Thank you.
## Steps to Reproduce
## Expected Result
## Actual Result
### Collector version
0.68.0
### Environment information
## Environment
OS: (e.g., "Ubuntu 20.04")
Compiler(if manually compiled): (e.g., "go 14.2")
### OpenTelemetry Collector configuration
_No response_
### Log output
_No response_
### Additional context
_No response_
|
1.0
|
How to understand decision_wait of tail_sampling - ### Component(s)
processor/tailsampling
### What happened?
## Description
the decision_wait describe by:
```
decision_wait (default = 30s): Wait time since the first span of a trace before making a sampling decision
````
When I make two policies to get long-request-time traces and the error traces like this:
```
tail_sampling:
decision_wait: 60s
num_traces: 5000
expected_new_traces_per_sec: 10
policies:
[
{
name: errors-policy,
type: numeric_attribute,
numeric_attribute: {key: http.status_code, min_value: 499, max_value: 600}
},
{
name: latency-5s,
type: latency,
latency: {threshold_ms: 5000}
},
]
```
So if the trace is error already, the processor will delivery the trace in 60s later?
In the real world, Microservice cross call make more than 60s long-request. So If I want to catch the threshold_ms>60000, the decision_wait must more than 60s, better about decision_wait=80s. And all the trace will store to the Jaeger at least delay decision_wait=80s later?
Correct me if I'm wrong, Thank you.
## Steps to Reproduce
## Expected Result
## Actual Result
### Collector version
0.68.0
### Environment information
## Environment
OS: (e.g., "Ubuntu 20.04")
Compiler(if manually compiled): (e.g., "go 14.2")
### OpenTelemetry Collector configuration
_No response_
### Log output
_No response_
### Additional context
_No response_
|
process
|
how to understand decision wait of tail sampling component s processor tailsampling what happened description the decision wait describe by decision wait default wait time since the first span of a trace before making a sampling decision when i make two policies to get long request time traces and the error traces like this tail sampling decision wait num traces expected new traces per sec policies name errors policy type numeric attribute numeric attribute key http status code min value max value name latency type latency latency threshold ms so if the trace is error already the processor will delivery the trace in later in the real world microservice cross call make more than long request so if i want to catch the threshold ms the decision wait must more than better about decision wait and all the trace will store to the jaeger at least delay decision wait later correct me if i m wrong thank you steps to reproduce expected result actual result collector version environment information environment os e g ubuntu compiler if manually compiled e g go opentelemetry collector configuration no response log output no response additional context no response
| 1
|
14,153
| 17,064,861,025
|
IssuesEvent
|
2021-07-07 05:38:00
|
ECP-WarpX/WarpX
|
https://api.github.com/repos/ECP-WarpX/WarpX
|
closed
|
yt: nbody species appearing
|
component: post-processing component: tests
|
I've been having the issue when using a newer version of yt, that the extra species nbody appears when reading in plot files. For regular post processing, this is not a problem. However, when dealing with regression tests, it does cause problems. Since that species is not present in the json files the tests fail. The other way around happens when I create a regression test - the nbody species is added to the locally created json file and then the test fails when run on azure since there is no nbody species there. Have others seen this?
I have a hack to the `Regression/Checksum/checksum.py` file that fixes this, adding the line `species_list.discard('nbody')` after line 54 in the file.
|
1.0
|
yt: nbody species appearing - I've been having the issue when using a newer version of yt, that the extra species nbody appears when reading in plot files. For regular post processing, this is not a problem. However, when dealing with regression tests, it does cause problems. Since that species is not present in the json files the tests fail. The other way around happens when I create a regression test - the nbody species is added to the locally created json file and then the test fails when run on azure since there is no nbody species there. Have others seen this?
I have a hack to the `Regression/Checksum/checksum.py` file that fixes this, adding the line `species_list.discard('nbody')` after line 54 in the file.
|
process
|
yt nbody species appearing i ve been having the issue when using a newer version of yt that the extra species nbody appears when reading in plot files for regular post processing this is not a problem however when dealing with regression tests it does cause problems since that species is not present in the json files the tests fail the other way around happens when i create a regression test the nbody species is added to the locally created json file and then the test fails when run on azure since there is no nbody species there have others seen this i have a hack to the regression checksum checksum py file that fixes this adding the line species list discard nbody after line in the file
| 1
|
123,154
| 16,449,477,271
|
IssuesEvent
|
2021-05-21 02:04:01
|
mobigen/IRIS-BigData-Platform
|
https://api.github.com/repos/mobigen/IRIS-BigData-Platform
|
closed
|
보고서 - 콤보박스 안에 스크롤이 가능한 체크박스 형태
|
#Design #IBP Priority: P2 Status: Backlog
|
## 기능 요청 ##
문제가 무엇인지에 대한 명확하고 간결한 설명 부탁드립니다.
보고서 - 콤보박스 안에 스크롤이 가능한 체크박스 형태
콤보박스처럼 접었다가 펼수 있는 체크박스 형태와 스크롤이 가능한 체크박스 형태를 원합니다.
## 원하는 솔루션 설명 ##
{ 원하는 기능에 대한 명확하고 간결한 설명 부탁드립니다 }
psi 시연 과제를 하다보니 한정된 대시보드 화면 사이즈에서
현재 아이리스 체크박스의 형태로는 체크박스의 내용이 많아질 경우(200개 넘어가는경우 등) 화면상의 제약이 있어
화면의 대부분을 체크박스가 차지하는 경우가 생겼습니다.
현재 태블로에서는 체크박스가 콤보박스내부에 체크박스가 존재하고, 스크롤이 가능한 형태로 되어 있는데,
기능 개발이 되었으면 좋겠습니다.
## 고려한 다른 대안 ##
## 기타 ##
기능 요청에 대한 다른 의견 또는 스크린샷을 여기에 모두 추가 부탁드립니다.
<img width="237" alt="스크린샷 2020-03-06 오후 1 34 28" src="https://user-images.githubusercontent.com/36151180/76053151-9a3fa800-5faf-11ea-9bc0-e7469fa927ed.png">
{ 고려한 대체 솔루션이나 기능에 대한 명확하고 간결한 설명 부탁드립니다. }
<img width="245" alt="스크린샷 2020-03-06 오후 1 34 23" src="https://user-images.githubusercontent.com/36151180/76053145-97dd4e00-5faf-11ea-95b9-7ee060df86a2.png">

|
1.0
|
보고서 - 콤보박스 안에 스크롤이 가능한 체크박스 형태 - ## 기능 요청 ##
문제가 무엇인지에 대한 명확하고 간결한 설명 부탁드립니다.
보고서 - 콤보박스 안에 스크롤이 가능한 체크박스 형태
콤보박스처럼 접었다가 펼수 있는 체크박스 형태와 스크롤이 가능한 체크박스 형태를 원합니다.
## 원하는 솔루션 설명 ##
{ 원하는 기능에 대한 명확하고 간결한 설명 부탁드립니다 }
psi 시연 과제를 하다보니 한정된 대시보드 화면 사이즈에서
현재 아이리스 체크박스의 형태로는 체크박스의 내용이 많아질 경우(200개 넘어가는경우 등) 화면상의 제약이 있어
화면의 대부분을 체크박스가 차지하는 경우가 생겼습니다.
현재 태블로에서는 체크박스가 콤보박스내부에 체크박스가 존재하고, 스크롤이 가능한 형태로 되어 있는데,
기능 개발이 되었으면 좋겠습니다.
## 고려한 다른 대안 ##
## 기타 ##
기능 요청에 대한 다른 의견 또는 스크린샷을 여기에 모두 추가 부탁드립니다.
<img width="237" alt="스크린샷 2020-03-06 오후 1 34 28" src="https://user-images.githubusercontent.com/36151180/76053151-9a3fa800-5faf-11ea-9bc0-e7469fa927ed.png">
{ 고려한 대체 솔루션이나 기능에 대한 명확하고 간결한 설명 부탁드립니다. }
<img width="245" alt="스크린샷 2020-03-06 오후 1 34 23" src="https://user-images.githubusercontent.com/36151180/76053145-97dd4e00-5faf-11ea-95b9-7ee060df86a2.png">

|
non_process
|
보고서 콤보박스 안에 스크롤이 가능한 체크박스 형태 기능 요청 문제가 무엇인지에 대한 명확하고 간결한 설명 부탁드립니다 보고서 콤보박스 안에 스크롤이 가능한 체크박스 형태 콤보박스처럼 접었다가 펼수 있는 체크박스 형태와 스크롤이 가능한 체크박스 형태를 원합니다 원하는 솔루션 설명 원하는 기능에 대한 명확하고 간결한 설명 부탁드립니다 psi 시연 과제를 하다보니 한정된 대시보드 화면 사이즈에서 현재 아이리스 체크박스의 형태로는 체크박스의 내용이 많아질 경우 넘어가는경우 등 화면상의 제약이 있어 화면의 대부분을 체크박스가 차지하는 경우가 생겼습니다 현재 태블로에서는 체크박스가 콤보박스내부에 체크박스가 존재하고 스크롤이 가능한 형태로 되어 있는데 기능 개발이 되었으면 좋겠습니다 고려한 다른 대안 기타 기능 요청에 대한 다른 의견 또는 스크린샷을 여기에 모두 추가 부탁드립니다 img width alt 스크린샷 오후 src 고려한 대체 솔루션이나 기능에 대한 명확하고 간결한 설명 부탁드립니다 img width alt 스크린샷 오후 src
| 0
|
10,922
| 3,441,380,871
|
IssuesEvent
|
2015-12-14 18:13:57
|
dotnet/wcf
|
https://api.github.com/repos/dotnet/wcf
|
closed
|
Improve legend for what's supported table and update all cells to reflect RC1
|
documentation
|
2 parts to this task:
1. Improve the granularity of the emoji's to distinguish better between "what works completely" to "what works for the paths we've tested"
2. Go through and choose the right emoji's for each table cell
|
1.0
|
Improve legend for what's supported table and update all cells to reflect RC1 - 2 parts to this task:
1. Improve the granularity of the emoji's to distinguish better between "what works completely" to "what works for the paths we've tested"
2. Go through and choose the right emoji's for each table cell
|
non_process
|
improve legend for what s supported table and update all cells to reflect parts to this task improve the granularity of the emoji s to distinguish better between what works completely to what works for the paths we ve tested go through and choose the right emoji s for each table cell
| 0
|
648,712
| 21,192,587,242
|
IssuesEvent
|
2022-04-08 19:17:21
|
status-im/status-desktop
|
https://api.github.com/repos/status-im/status-desktop
|
opened
|
image missing in get your keys screen
|
bug onboarding priority 2: medium
|
# Bug Report
## Description

## Steps to reproduce
1. use a new data folder
2. start app and choose "I am new to Status"
#### Expected behavior

|
1.0
|
image missing in get your keys screen - # Bug Report
## Description

## Steps to reproduce
1. use a new data folder
2. start app and choose "I am new to Status"
#### Expected behavior

|
non_process
|
image missing in get your keys screen bug report description steps to reproduce use a new data folder start app and choose i am new to status expected behavior
| 0
|
707
| 9,979,386,976
|
IssuesEvent
|
2019-07-09 22:40:57
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
opened
|
Completion crashes (fault) if more than instance of Roslyn completion service (different languages) are attached to the single completion session
|
Area-IDE Tenet-Reliability
|
from activity.log
```
Exception from ActivityLog.xml:
System.ArgumentException: An entry with the same key already exists.
at System.Collections.Specialized.ListDictionary.Add(Object key, Object value)
at System.Collections.Specialized.HybridDictionary.Add(Object key, Object value)
at Microsoft.VisualStudio.Utilities.PropertyCollection.AddProperty(Object key, Object property)
at Microsoft.CodeAnalysis.Editor.Implementation.IntelliSense.AsyncCompletion.CompletionSource.<GetCompletionContextAsync>d__23.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at Microsoft.VisualStudio.Text.Utilities.GuardedOperations.<CallExtensionPointAsync>d__34`1.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at Microsoft.VisualStudio.Telemetry.WindowsErrorReporting.WatsonReport.GetClrWatsonExceptionInfo(Exception exceptionObject)
```
|
True
|
Completion crashes (fault) if more than instance of Roslyn completion service (different languages) are attached to the single completion session - from activity.log
```
Exception from ActivityLog.xml:
System.ArgumentException: An entry with the same key already exists.
at System.Collections.Specialized.ListDictionary.Add(Object key, Object value)
at System.Collections.Specialized.HybridDictionary.Add(Object key, Object value)
at Microsoft.VisualStudio.Utilities.PropertyCollection.AddProperty(Object key, Object property)
at Microsoft.CodeAnalysis.Editor.Implementation.IntelliSense.AsyncCompletion.CompletionSource.<GetCompletionContextAsync>d__23.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at Microsoft.VisualStudio.Text.Utilities.GuardedOperations.<CallExtensionPointAsync>d__34`1.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at Microsoft.VisualStudio.Telemetry.WindowsErrorReporting.WatsonReport.GetClrWatsonExceptionInfo(Exception exceptionObject)
```
|
non_process
|
completion crashes fault if more than instance of roslyn completion service different languages are attached to the single completion session from activity log exception from activitylog xml system argumentexception an entry with the same key already exists at system collections specialized listdictionary add object key object value at system collections specialized hybriddictionary add object key object value at microsoft visualstudio utilities propertycollection addproperty object key object property at microsoft codeanalysis editor implementation intellisense asynccompletion completionsource lt getcompletioncontextasync gt d movenext end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at microsoft visualstudio text utilities guardedoperations lt callextensionpointasync gt d movenext end of stack trace from previous location where exception was thrown at microsoft visualstudio telemetry windowserrorreporting watsonreport getclrwatsonexceptioninfo exception exceptionobject
| 0
|
21,992
| 30,485,991,134
|
IssuesEvent
|
2023-07-18 02:16:17
|
rust-lang/rust
|
https://api.github.com/repos/rust-lang/rust
|
closed
|
Windows: Reduce raciness of `spawn()` vs. `SetCurrentDirectory`
|
O-windows C-bug T-libs A-process
|
I was inspecting `resolve_exe()` specifically looking for racy coding patterns.
The code currently contains this:
```rust
if !path::is_file_name(exe_path) {
if has_exe_suffix {
// The application name is a path to a `.exe` file.
// Let `CreateProcessW` figure out if it exists or not.
return Ok(exe_path.into());
}
let mut path = PathBuf::from(exe_path);
// Append `.exe` if not already there.
path = path::append_suffix(path, EXE_SUFFIX.as_ref());
if path.try_exists().unwrap_or(false) {
return Ok(path);
} else {
// It's ok to use `set_extension` here because the intent is to
// remove the extension that was just added.
path.set_extension("");
return Ok(path);
}
```
Consider what happens when we pass a relative path containing a path separator, e.g. "target/something". This code will check if the file `target/something.exe` exists, relative to the current working directory and if so, return `"target/something.exe"`.
This function is used by the implementation of `spawn()`. Roughly, spawn looks like this:
```rust
let program = resolve_exe(&self.program, || env::var_os("PATH"), child_paths)?;
...
let (dirp, _data) = make_dirp(self.cwd.as_ref())?;
...
let program = to_u16s(&program)?;
unsafe {
cvt(c::CreateProcessW(
program.as_ptr(),
...
dirp, // lpCurrentDirectory
&mut pi,
))
}?;
```
In theory the current working directory could change at any time. To minimize the effect that would have on `spawn()`, it seems like it could instead be written like this:
```rust
let cwd = self.cwd.as_ref();
let program = resolve_exe(cwd, &self.program, || env::var_os("PATH"), child_paths)?;
...
let (dirp, _data) = make_dirp(cwd)?;
...
let program = to_u16s(&program)?;
unsafe {
cvt(c::CreateProcessW(
program.as_ptr(),
...
dirp, // lpCurrentDirectory
&mut pi,
))
}?;
```
And then change `resolve_exe` so that it always returns an absolute path, where it always resolves any relative paths relative to the value of the `cwd` parameter passed in.
`spawn()` and `SetCurrentDirectory` would still be racy, but at least all the path resolution within `spawn` would be internally consistent with respect to the current directory at a point in time.
|
1.0
|
Windows: Reduce raciness of `spawn()` vs. `SetCurrentDirectory` - I was inspecting `resolve_exe()` specifically looking for racy coding patterns.
The code currently contains this:
```rust
if !path::is_file_name(exe_path) {
if has_exe_suffix {
// The application name is a path to a `.exe` file.
// Let `CreateProcessW` figure out if it exists or not.
return Ok(exe_path.into());
}
let mut path = PathBuf::from(exe_path);
// Append `.exe` if not already there.
path = path::append_suffix(path, EXE_SUFFIX.as_ref());
if path.try_exists().unwrap_or(false) {
return Ok(path);
} else {
// It's ok to use `set_extension` here because the intent is to
// remove the extension that was just added.
path.set_extension("");
return Ok(path);
}
```
Consider what happens when we pass a relative path containing a path separator, e.g. "target/something". This code will check if the file `target/something.exe` exists, relative to the current working directory and if so, return `"target/something.exe"`.
This function is used by the implementation of `spawn()`. Roughly, spawn looks like this:
```rust
let program = resolve_exe(&self.program, || env::var_os("PATH"), child_paths)?;
...
let (dirp, _data) = make_dirp(self.cwd.as_ref())?;
...
let program = to_u16s(&program)?;
unsafe {
cvt(c::CreateProcessW(
program.as_ptr(),
...
dirp, // lpCurrentDirectory
&mut pi,
))
}?;
```
In theory the current working directory could change at any time. To minimize the effect that would have on `spawn()`, it seems like it could instead be written like this:
```rust
let cwd = self.cwd.as_ref();
let program = resolve_exe(cwd, &self.program, || env::var_os("PATH"), child_paths)?;
...
let (dirp, _data) = make_dirp(cwd)?;
...
let program = to_u16s(&program)?;
unsafe {
cvt(c::CreateProcessW(
program.as_ptr(),
...
dirp, // lpCurrentDirectory
&mut pi,
))
}?;
```
And then change `resolve_exe` so that it always returns an absolute path, where it always resolves any relative paths relative to the value of the `cwd` parameter passed in.
`spawn()` and `SetCurrentDirectory` would still be racy, but at least all the path resolution within `spawn` would be internally consistent with respect to the current directory at a point in time.
|
process
|
windows reduce raciness of spawn vs setcurrentdirectory i was inspecting resolve exe specifically looking for racy coding patterns the code currently contains this rust if path is file name exe path if has exe suffix the application name is a path to a exe file let createprocessw figure out if it exists or not return ok exe path into let mut path pathbuf from exe path append exe if not already there path path append suffix path exe suffix as ref if path try exists unwrap or false return ok path else it s ok to use set extension here because the intent is to remove the extension that was just added path set extension return ok path consider what happens when we pass a relative path containing a path separator e g target something this code will check if the file target something exe exists relative to the current working directory and if so return target something exe this function is used by the implementation of spawn roughly spawn looks like this rust let program resolve exe self program env var os path child paths let dirp data make dirp self cwd as ref let program to program unsafe cvt c createprocessw program as ptr dirp lpcurrentdirectory mut pi in theory the current working directory could change at any time to minimize the effect that would have on spawn it seems like it could instead be written like this rust let cwd self cwd as ref let program resolve exe cwd self program env var os path child paths let dirp data make dirp cwd let program to program unsafe cvt c createprocessw program as ptr dirp lpcurrentdirectory mut pi and then change resolve exe so that it always returns an absolute path where it always resolves any relative paths relative to the value of the cwd parameter passed in spawn and setcurrentdirectory would still be racy but at least all the path resolution within spawn would be internally consistent with respect to the current directory at a point in time
| 1
|
88,031
| 25,283,521,921
|
IssuesEvent
|
2022-11-16 17:24:58
|
getll/pokemon-web-services
|
https://api.github.com/repos/getll/pokemon-web-services
|
closed
|
TASK 2.b: Populate Games, Generations, Gyms, Locations Tables
|
Build #1 Sub Task
|
Creating the Games, Generations, Gyms, and Locations tables.
### Requirements
- XAMPP
- phpMyAdmin
### Dependencies
- None
|
1.0
|
TASK 2.b: Populate Games, Generations, Gyms, Locations Tables - Creating the Games, Generations, Gyms, and Locations tables.
### Requirements
- XAMPP
- phpMyAdmin
### Dependencies
- None
|
non_process
|
task b populate games generations gyms locations tables creating the games generations gyms and locations tables requirements xampp phpmyadmin dependencies none
| 0
|
413,226
| 12,061,556,032
|
IssuesEvent
|
2020-04-16 00:10:24
|
eclipse-ee4j/glassfish
|
https://api.github.com/repos/eclipse-ee4j/glassfish
|
closed
|
gf4, web console athentication failed with 64 bit jdk7.
|
Component: admin ERR: Assignee Priority: Major Stale Type: Bug
|
The administration web console denies login if glassfish 4.0 ml is started with a JAVA_HOME that poits to a 64 bit jdk. This happens even if there is no passord for the admin user. The asadmin login command always works instead.
The same glassfish installation if run with the corresponding x86 version of the jdk (jdk1.7.0_55 x86) works as advertised and does not exhibit the problem.
#### Environment
win7 x64, jdk1.7.0_55 x64.
#### Affected Versions
[4.0]
|
1.0
|
gf4, web console athentication failed with 64 bit jdk7. - The administration web console denies login if glassfish 4.0 ml is started with a JAVA_HOME that poits to a 64 bit jdk. This happens even if there is no passord for the admin user. The asadmin login command always works instead.
The same glassfish installation if run with the corresponding x86 version of the jdk (jdk1.7.0_55 x86) works as advertised and does not exhibit the problem.
#### Environment
win7 x64, jdk1.7.0_55 x64.
#### Affected Versions
[4.0]
|
non_process
|
web console athentication failed with bit the administration web console denies login if glassfish ml is started with a java home that poits to a bit jdk this happens even if there is no passord for the admin user the asadmin login command always works instead the same glassfish installation if run with the corresponding version of the jdk works as advertised and does not exhibit the problem environment affected versions
| 0
|
5,329
| 8,142,963,715
|
IssuesEvent
|
2018-08-21 09:19:15
|
aiidateam/aiida_core
|
https://api.github.com/repos/aiidateam/aiida_core
|
closed
|
JobCalculation states not properly set when JobProcess is terminated
|
priority/critical-blocking topic/JobCalculationAndProcess topic/Workflows type/bug
|
The `state` attribute of a `JobCalculation` should be set accordingly when the `JobProcess` reaches a terminal state. When it finishes normally, nothing should be done most likely, however, when it reaches `Killed` or `Excepted` it is crucial that the calculation state be updated accordingly
|
1.0
|
JobCalculation states not properly set when JobProcess is terminated - The `state` attribute of a `JobCalculation` should be set accordingly when the `JobProcess` reaches a terminal state. When it finishes normally, nothing should be done most likely, however, when it reaches `Killed` or `Excepted` it is crucial that the calculation state be updated accordingly
|
process
|
jobcalculation states not properly set when jobprocess is terminated the state attribute of a jobcalculation should be set accordingly when the jobprocess reaches a terminal state when it finishes normally nothing should be done most likely however when it reaches killed or excepted it is crucial that the calculation state be updated accordingly
| 1
|
133,989
| 5,218,120,275
|
IssuesEvent
|
2017-01-26 15:44:11
|
michaeljcalkins/rangersteve-ideas
|
https://api.github.com/repos/michaeljcalkins/rangersteve-ideas
|
opened
|
Optimal entity interpolation
|
Priority: Critical
|
I don't know if tweening itself is performance heavy but we need some sort of interpolation.
|
1.0
|
Optimal entity interpolation - I don't know if tweening itself is performance heavy but we need some sort of interpolation.
|
non_process
|
optimal entity interpolation i don t know if tweening itself is performance heavy but we need some sort of interpolation
| 0
|
13,238
| 15,706,655,033
|
IssuesEvent
|
2021-03-26 17:43:44
|
netlify/build
|
https://api.github.com/repos/netlify/build
|
closed
|
Improve lingering processes message color
|
proj/lingering-processes type: chore
|
Background at https://github.com/netlify/build/issues/2371
The lingering processes message is a warning, not an error. Its color should be yellow, not red.
We should use this new warning color for other warnings in `@netlify/build` and `@netlify/config` as well.
|
1.0
|
Improve lingering processes message color - Background at https://github.com/netlify/build/issues/2371
The lingering processes message is a warning, not an error. Its color should be yellow, not red.
We should use this new warning color for other warnings in `@netlify/build` and `@netlify/config` as well.
|
process
|
improve lingering processes message color background at the lingering processes message is a warning not an error its color should be yellow not red we should use this new warning color for other warnings in netlify build and netlify config as well
| 1
|
2,385
| 5,187,641,399
|
IssuesEvent
|
2017-01-20 17:24:41
|
Alfresco/alfresco-ng2-components
|
https://api.github.com/repos/Alfresco/alfresco-ng2-components
|
closed
|
No processes are displayed when starting process within task app
|
browser: all bug comp: activiti-processList
|
1. Go to task app
2. Go to processes
3. Start a process
**Expected results**
List of avaliable processes are displayed
**Actual results**
No processes are displayed
N.b. processes display within individual apps
<img width="576" alt="screen shot 2017-01-09 at 11 50 07" src="https://cloud.githubusercontent.com/assets/13200338/21765758/bb52885c-d662-11e6-865d-0ee686d29adc.png">
|
1.0
|
No processes are displayed when starting process within task app - 1. Go to task app
2. Go to processes
3. Start a process
**Expected results**
List of avaliable processes are displayed
**Actual results**
No processes are displayed
N.b. processes display within individual apps
<img width="576" alt="screen shot 2017-01-09 at 11 50 07" src="https://cloud.githubusercontent.com/assets/13200338/21765758/bb52885c-d662-11e6-865d-0ee686d29adc.png">
|
process
|
no processes are displayed when starting process within task app go to task app go to processes start a process expected results list of avaliable processes are displayed actual results no processes are displayed n b processes display within individual apps img width alt screen shot at src
| 1
|
18,850
| 24,763,987,847
|
IssuesEvent
|
2022-10-22 09:06:48
|
home-climate-control/dz
|
https://api.github.com/repos/home-climate-control/dz
|
closed
|
SingleStageUnitController demand output is wrong
|
annoyance process control
|
### Expected Behavior
`HvacCommand` emitted by `SingleStageUnitController` reflects the actual demand value.
### Actual Behavior
The value is always 1 when the HVAC is requested to be on. This throws off the metrics.
|
1.0
|
SingleStageUnitController demand output is wrong - ### Expected Behavior
`HvacCommand` emitted by `SingleStageUnitController` reflects the actual demand value.
### Actual Behavior
The value is always 1 when the HVAC is requested to be on. This throws off the metrics.
|
process
|
singlestageunitcontroller demand output is wrong expected behavior hvaccommand emitted by singlestageunitcontroller reflects the actual demand value actual behavior the value is always when the hvac is requested to be on this throws off the metrics
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.