Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
218,655
| 7,332,033,375
|
IssuesEvent
|
2018-03-05 15:15:08
|
enviroCar/enviroCar-app
|
https://api.github.com/repos/enviroCar/enviroCar-app
|
closed
|
Simultaneous upload causes track duplicates on server
|
Priority - 1 - High bug
|
The user can hit the "upload all" button and after that upload a single track via a long-click. This serializes the same track twice resulting in a duplication on the server side (as there are no duplication checks).
Implement a mutex (maybe using the UploadManager) and block uploading interaction as long as an upload takes place.
|
1.0
|
Simultaneous upload causes track duplicates on server - The user can hit the "upload all" button and after that upload a single track via a long-click. This serializes the same track twice resulting in a duplication on the server side (as there are no duplication checks).
Implement a mutex (maybe using the UploadManager) and block uploading interaction as long as an upload takes place.
|
non_defect
|
simultaneous upload causes track duplicates on server the user can hit the upload all button and after that upload a single track via a long click this serializes the same track twice resulting in a duplication on the server side as there are no duplication checks implement a mutex maybe using the uploadmanager and block uploading interaction as long as an upload takes place
| 0
|
39,146
| 9,218,651,631
|
IssuesEvent
|
2019-03-11 13:52:29
|
STEllAR-GROUP/phylanx
|
https://api.github.com/repos/STEllAR-GROUP/phylanx
|
closed
|
PhySL interpreter inserts 'random' nil into output
|
category: examples type: defect
|
I should add that `random` sometimes has `nil` somewhere in its output. In physl:
```
random(make_list(5,10), list("binomial", 3, .5))
```
produces:
```
[[2, 2, 3nil,
0, 2, 1, 1, 1, 1, 2], [2, 2, 2, 0, 1, 1, 2, 1, 3, 2], [3, 2, 2, 1, 0, 1, 2, 2, 2, 0], [2, 0, 1, 2, 1, 3, 3, 1, 0, 2], [2, 1, 1, 2, 2, 2, 2, 2, 1, 1]]
```
_Originally posted by @taless474 in https://github.com/STEllAR-GROUP/phylanx/issues/827#issuecomment-471065529_
|
1.0
|
PhySL interpreter inserts 'random' nil into output - I should add that `random` sometimes has `nil` somewhere in its output. In physl:
```
random(make_list(5,10), list("binomial", 3, .5))
```
produces:
```
[[2, 2, 3nil,
0, 2, 1, 1, 1, 1, 2], [2, 2, 2, 0, 1, 1, 2, 1, 3, 2], [3, 2, 2, 1, 0, 1, 2, 2, 2, 0], [2, 0, 1, 2, 1, 3, 3, 1, 0, 2], [2, 1, 1, 2, 2, 2, 2, 2, 1, 1]]
```
_Originally posted by @taless474 in https://github.com/STEllAR-GROUP/phylanx/issues/827#issuecomment-471065529_
|
defect
|
physl interpreter inserts random nil into output i should add that random sometimes has nil somewhere in its output in physl random make list list binomial produces originally posted by in
| 1
|
31,042
| 6,413,752,441
|
IssuesEvent
|
2017-08-08 08:28:41
|
oleg-shilo/cs-script
|
https://api.github.com/repos/oleg-shilo/cs-script
|
closed
|
Ambiguous call in dbg.inject<ID>
|
defect Done: waiting for release
|
While trying to compose a test script environment related to #71 I ran into this issue:
```
<Workspace>\CSScript\Issue #79\Redirect>cscs.exe -cd RedirectRefLib.cs
C# Script execution engine. Version 3.27.0.0.
Copyright (C) 2004-2017 Oleg Shilo.
Error: Specified file could not be compiled.
csscript.CompilerException: c:\Users\klaey-1\AppData\Local\Temp\CSSCRIPT\Cache\dbg.inject.-110101375
1.cs(13,16): error CS0121: The call is ambiguous between the following methods or properties: 'dbg_e
xtensions.print<T>(T, params object[])' and 'dbg_extensions.print<T>(T, params object[])'
at csscript.CSExecutor.ProcessCompilingResult(CompilerResults results, CompilerParameters compile
rParams, ScriptParser parser, String scriptFileName, String assemblyFileName, String[] additionalDep
endencies)
at csscript.CSExecutor.Compile(String scriptFileName)
at csscript.CSExecutor.ExecuteImpl()
<Workspace>\CSScript\Issue #79\Redirect>
```
Can be reproduced by extracting the attached files into a temporary location, place `cscs.exe` into the `\Redirect` folder and run `cscs.exe -cd RedirectRefLib.cs`.
[Issue #79.zip](https://github.com/oleg-shilo/cs-script/files/1162248/Issue.79.zip)
|
1.0
|
Ambiguous call in dbg.inject<ID> - While trying to compose a test script environment related to #71 I ran into this issue:
```
<Workspace>\CSScript\Issue #79\Redirect>cscs.exe -cd RedirectRefLib.cs
C# Script execution engine. Version 3.27.0.0.
Copyright (C) 2004-2017 Oleg Shilo.
Error: Specified file could not be compiled.
csscript.CompilerException: c:\Users\klaey-1\AppData\Local\Temp\CSSCRIPT\Cache\dbg.inject.-110101375
1.cs(13,16): error CS0121: The call is ambiguous between the following methods or properties: 'dbg_e
xtensions.print<T>(T, params object[])' and 'dbg_extensions.print<T>(T, params object[])'
at csscript.CSExecutor.ProcessCompilingResult(CompilerResults results, CompilerParameters compile
rParams, ScriptParser parser, String scriptFileName, String assemblyFileName, String[] additionalDep
endencies)
at csscript.CSExecutor.Compile(String scriptFileName)
at csscript.CSExecutor.ExecuteImpl()
<Workspace>\CSScript\Issue #79\Redirect>
```
Can be reproduced by extracting the attached files into a temporary location, place `cscs.exe` into the `\Redirect` folder and run `cscs.exe -cd RedirectRefLib.cs`.
[Issue #79.zip](https://github.com/oleg-shilo/cs-script/files/1162248/Issue.79.zip)
|
defect
|
ambiguous call in dbg inject while trying to compose a test script environment related to i ran into this issue csscript issue redirect cscs exe cd redirectreflib cs c script execution engine version copyright c oleg shilo error specified file could not be compiled csscript compilerexception c users klaey appdata local temp csscript cache dbg inject cs error the call is ambiguous between the following methods or properties dbg e xtensions print t params object and dbg extensions print t params object at csscript csexecutor processcompilingresult compilerresults results compilerparameters compile rparams scriptparser parser string scriptfilename string assemblyfilename string additionaldep endencies at csscript csexecutor compile string scriptfilename at csscript csexecutor executeimpl csscript issue redirect can be reproduced by extracting the attached files into a temporary location place cscs exe into the redirect folder and run cscs exe cd redirectreflib cs
| 1
|
75,071
| 25,513,554,408
|
IssuesEvent
|
2022-11-28 14:47:45
|
PowerDNS/pdns
|
https://api.github.com/repos/PowerDNS/pdns
|
closed
|
API endpoint /servers/{server_id}/zones/{zone_id}/check should exist but when calling there is an error and the endpoint cannot be found in the source code.
|
auth docs defect
|
<!-- Hi! Thanks for filing an issue. It will be read with care by human beings. Can we ask you to please fill out this template and not simply demand new features or send in complaints? Thanks! -->
<!-- Also please search the existing issues (both open and closed) to see if your report might be duplicate -->
<!-- Please don't file an issue when you have a support question, send support questions to the mailinglist or ask them on IRC (https://www.powerdns.com/opensource.html) -->
<!-- Tell us what is issue is about -->
- Program: Authoritative
- Issue type: Bug report
### Short description
<!-- Explain in a few sentences what the issue/request is -->
GET /servers/{server_id}/zones/{zone_id}/check is a documented endpoint in https://doc.powerdns.com/authoritative/http-api/zone.html. However, I reviewed the sourcecode at I cannot find that this is implemented in code.
In all the versions of wc-auth.cc that I verified, this path is not registered.
### Environment
N/A
### Steps to reproduce
Call the 'check' endpoint as described in https://doc.powerdns.com/authoritative/http-api/zone.html.
### Expected behaviour
That the endpoint works.
I am running 4.1.5 and got an error when calling the `check` endpoint. I checked if a newer version had that endpoint and i checked a lost of branches and tags, but the endpoint was not there.
### Actual behaviour
Calling the check endpoint resulted in:
'{"error": "Could not find domain '<testdomain>./check.'"}'
### Other information
<!-- if you already did more digging into the issue, please provide all the information you gathered -->
|
1.0
|
API endpoint /servers/{server_id}/zones/{zone_id}/check should exist but when calling there is an error and the endpoint cannot be found in the source code. - <!-- Hi! Thanks for filing an issue. It will be read with care by human beings. Can we ask you to please fill out this template and not simply demand new features or send in complaints? Thanks! -->
<!-- Also please search the existing issues (both open and closed) to see if your report might be duplicate -->
<!-- Please don't file an issue when you have a support question, send support questions to the mailinglist or ask them on IRC (https://www.powerdns.com/opensource.html) -->
<!-- Tell us what is issue is about -->
- Program: Authoritative
- Issue type: Bug report
### Short description
<!-- Explain in a few sentences what the issue/request is -->
GET /servers/{server_id}/zones/{zone_id}/check is a documented endpoint in https://doc.powerdns.com/authoritative/http-api/zone.html. However, I reviewed the sourcecode at I cannot find that this is implemented in code.
In all the versions of wc-auth.cc that I verified, this path is not registered.
### Environment
N/A
### Steps to reproduce
Call the 'check' endpoint as described in https://doc.powerdns.com/authoritative/http-api/zone.html.
### Expected behaviour
That the endpoint works.
I am running 4.1.5 and got an error when calling the `check` endpoint. I checked if a newer version had that endpoint and i checked a lost of branches and tags, but the endpoint was not there.
### Actual behaviour
Calling the check endpoint resulted in:
'{"error": "Could not find domain '<testdomain>./check.'"}'
### Other information
<!-- if you already did more digging into the issue, please provide all the information you gathered -->
|
defect
|
api endpoint servers server id zones zone id check should exist but when calling there is an error and the endpoint cannot be found in the source code program authoritative issue type bug report short description get servers server id zones zone id check is a documented endpoint in however i reviewed the sourcecode at i cannot find that this is implemented in code in all the versions of wc auth cc that i verified this path is not registered environment n a steps to reproduce call the check endpoint as described in expected behaviour that the endpoint works i am running and got an error when calling the check endpoint i checked if a newer version had that endpoint and i checked a lost of branches and tags but the endpoint was not there actual behaviour calling the check endpoint resulted in error could not find domain check other information
| 1
|
10,691
| 8,134,797,724
|
IssuesEvent
|
2018-08-19 20:05:23
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
docker exposes containers' ports to the world despite firewall module being enabled
|
1.severity: security
|
## Issue description
Docker has its own iptables chain and somehow¹ bypasses NixOS firewall, the firewall, which I thought would safeguard this from me. This is very dangerous when you have passwordless development containers running.
It seems [that](https://fralef.me/docker-and-iptables.html) `virtualisation.docker.extraOptions = "--iptables=false";` mitigates this, so perhaps it is possible to make the firewall module add a `docker.extraOptions`?
1 — I didn't try understanding why, but I want to start the discussion.
### Steps to reproduce
Start docker and, say, a postgres container with (warning: docker-compose syntax ahead) `ports: ["5432:5432"]`, enable firewall and observe an open 5432 port from another machine.
## Technical details
- system: `"x86_64-linux"`
- host os: `Linux 4.14.32, NixOS, 18.03.132304.45cd6090d9a (Impala)`
- multi-user?: `yes`
- sandbox: `no`
- version: `nix-env (Nix) 2.1pre6148_a4aac7f`
- channels(s): `"unstable-18.03pre130932.cc4677c36ee"`
- channels(root): `"nixos-18.03.132304.45cd6090d9a"`
- nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos/nixpkgs`
|
True
|
docker exposes containers' ports to the world despite firewall module being enabled - ## Issue description
Docker has its own iptables chain and somehow¹ bypasses NixOS firewall, the firewall, which I thought would safeguard this from me. This is very dangerous when you have passwordless development containers running.
It seems [that](https://fralef.me/docker-and-iptables.html) `virtualisation.docker.extraOptions = "--iptables=false";` mitigates this, so perhaps it is possible to make the firewall module add a `docker.extraOptions`?
1 — I didn't try understanding why, but I want to start the discussion.
### Steps to reproduce
Start docker and, say, a postgres container with (warning: docker-compose syntax ahead) `ports: ["5432:5432"]`, enable firewall and observe an open 5432 port from another machine.
## Technical details
- system: `"x86_64-linux"`
- host os: `Linux 4.14.32, NixOS, 18.03.132304.45cd6090d9a (Impala)`
- multi-user?: `yes`
- sandbox: `no`
- version: `nix-env (Nix) 2.1pre6148_a4aac7f`
- channels(s): `"unstable-18.03pre130932.cc4677c36ee"`
- channels(root): `"nixos-18.03.132304.45cd6090d9a"`
- nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos/nixpkgs`
|
non_defect
|
docker exposes containers ports to the world despite firewall module being enabled issue description docker has its own iptables chain and somehow¹ bypasses nixos firewall the firewall which i thought would safeguard this from me this is very dangerous when you have passwordless development containers running it seems virtualisation docker extraoptions iptables false mitigates this so perhaps it is possible to make the firewall module add a docker extraoptions — i didn t try understanding why but i want to start the discussion steps to reproduce start docker and say a postgres container with warning docker compose syntax ahead ports enable firewall and observe an open port from another machine technical details system linux host os linux nixos impala multi user yes sandbox no version nix env nix channels s unstable channels root nixos nixpkgs nix var nix profiles per user root channels nixos nixpkgs
| 0
|
10,783
| 2,622,188,814
|
IssuesEvent
|
2015-03-04 00:22:07
|
byzhang/cudpp
|
https://api.github.com/repos/byzhang/cudpp
|
closed
|
satGL produces a garbled image
|
auto-migrated Milestone-Release1.1 OpSys-Linux Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Build and run the satGL sample app
What is the expected output? What do you see instead?
Correct output can be seen by running the device emulation version. In
release or debug builds, instead the results are a green and blue smear.
```
Original issue reported on code.google.com by `harr...@gmail.com` on 17 Jun 2009 at 1:57
|
1.0
|
satGL produces a garbled image - ```
What steps will reproduce the problem?
1. Build and run the satGL sample app
What is the expected output? What do you see instead?
Correct output can be seen by running the device emulation version. In
release or debug builds, instead the results are a green and blue smear.
```
Original issue reported on code.google.com by `harr...@gmail.com` on 17 Jun 2009 at 1:57
|
defect
|
satgl produces a garbled image what steps will reproduce the problem build and run the satgl sample app what is the expected output what do you see instead correct output can be seen by running the device emulation version in release or debug builds instead the results are a green and blue smear original issue reported on code google com by harr gmail com on jun at
| 1
|
58,553
| 14,432,798,076
|
IssuesEvent
|
2020-12-07 03:01:21
|
tensorflow/tensorflow
|
https://api.github.com/repos/tensorflow/tensorflow
|
opened
|
Whl package not building due with a issue with sed - MacOS 11
|
type:build/install
|
**System information**
- OS Platform and Distribution: Mac OS 11.0.1 Big Sur
- TensorFlow installed from (source or binary): Source - git
- TensorFlow version: 2.4.0-rc4
- Python version: 3.8.6 (macports)
- Installed using virtualenv? pip? conda?: No
- Bazel version (if compiling from source): 3.1.0
- GCC/Compiler version (if compiling from source): Apple clang version 12.0.0 (clang-1200.0.32.27) XCode 12.2
- CUDA/cuDNN version: No
- GPU model and memory: AMD Radeon Pro 555X - 4GB
**Describe the problem**
Tensorflow compiles correctly. However, when using the command:
`bazel-bin/tensorflow/tools/pip_package/build_pip_package`
the whl package is not produced and rather the following error is shown:
`sed: /var/folders/y4/85hx48054tbd2y04jkf3l6880000gn/T/tmp.XXXXXXXXXX.Ty6dzcrU/tensorflow/__init__.py: in-place editing only works for regular files`
**Provide the exact sequence of commands / steps that you executed before running into the problem**
1. Compile from source as indicated in the tensorflow manual.
2. Run the command to build the whl package:
`bazel-bin/tensorflow/tools/pip_package/build_pip_package ..`
**Any other info / logs**
This error is not present in MacOS 10.15. All proceed as it should
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
|
1.0
|
Whl package not building due with a issue with sed - MacOS 11 - **System information**
- OS Platform and Distribution: Mac OS 11.0.1 Big Sur
- TensorFlow installed from (source or binary): Source - git
- TensorFlow version: 2.4.0-rc4
- Python version: 3.8.6 (macports)
- Installed using virtualenv? pip? conda?: No
- Bazel version (if compiling from source): 3.1.0
- GCC/Compiler version (if compiling from source): Apple clang version 12.0.0 (clang-1200.0.32.27) XCode 12.2
- CUDA/cuDNN version: No
- GPU model and memory: AMD Radeon Pro 555X - 4GB
**Describe the problem**
Tensorflow compiles correctly. However, when using the command:
`bazel-bin/tensorflow/tools/pip_package/build_pip_package`
the whl package is not produced and rather the following error is shown:
`sed: /var/folders/y4/85hx48054tbd2y04jkf3l6880000gn/T/tmp.XXXXXXXXXX.Ty6dzcrU/tensorflow/__init__.py: in-place editing only works for regular files`
**Provide the exact sequence of commands / steps that you executed before running into the problem**
1. Compile from source as indicated in the tensorflow manual.
2. Run the command to build the whl package:
`bazel-bin/tensorflow/tools/pip_package/build_pip_package ..`
**Any other info / logs**
This error is not present in MacOS 10.15. All proceed as it should
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
|
non_defect
|
whl package not building due with a issue with sed macos system information os platform and distribution mac os big sur tensorflow installed from source or binary source git tensorflow version python version macports installed using virtualenv pip conda no bazel version if compiling from source gcc compiler version if compiling from source apple clang version clang xcode cuda cudnn version no gpu model and memory amd radeon pro describe the problem tensorflow compiles correctly however when using the command bazel bin tensorflow tools pip package build pip package the whl package is not produced and rather the following error is shown sed var folders t tmp xxxxxxxxxx tensorflow init py in place editing only works for regular files provide the exact sequence of commands steps that you executed before running into the problem compile from source as indicated in the tensorflow manual run the command to build the whl package bazel bin tensorflow tools pip package build pip package any other info logs this error is not present in macos all proceed as it should include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached
| 0
|
216,031
| 16,625,852,139
|
IssuesEvent
|
2021-06-03 09:26:46
|
RotherOSS/doc-otobo-installation
|
https://api.github.com/repos/RotherOSS/doc-otobo-installation
|
opened
|
Update the requirements
|
documentation
|
https://doc.otobo.org/manual/installation/stable/en/content/requirements.html can be improved:
- [ ] No need to mention Node
- [ ] Add Redis as an optional dependency
|
1.0
|
Update the requirements - https://doc.otobo.org/manual/installation/stable/en/content/requirements.html can be improved:
- [ ] No need to mention Node
- [ ] Add Redis as an optional dependency
|
non_defect
|
update the requirements can be improved no need to mention node add redis as an optional dependency
| 0
|
77,577
| 27,058,913,688
|
IssuesEvent
|
2023-02-13 18:08:45
|
fecgov/fecfile-web-app
|
https://api.github.com/repos/fecgov/fecfile-web-app
|
closed
|
Defect - "Earmark Memo" the transactions table is not displaying the transaction type "Earmark Memo"
|
defect
|
This defect is being written for after creating an "Earmark Memo" the transactions table is not displaying the transaction type "Earmark Memo" as shown in screenshot below. This was tested in both DEV and STAGE environments.

### DEV ###
Note: The isDependentChild property in the TransactionType class, when set to true, identifies the transaction type as being the child transaction in a "double entry" transaction screen like AG and AF
Per @toddlees It has no label because it's missing from `ScheduleATransactionTypeLabels` in the front end
|
1.0
|
Defect - "Earmark Memo" the transactions table is not displaying the transaction type "Earmark Memo" - This defect is being written for after creating an "Earmark Memo" the transactions table is not displaying the transaction type "Earmark Memo" as shown in screenshot below. This was tested in both DEV and STAGE environments.

### DEV ###
Note: The isDependentChild property in the TransactionType class, when set to true, identifies the transaction type as being the child transaction in a "double entry" transaction screen like AG and AF
Per @toddlees It has no label because it's missing from `ScheduleATransactionTypeLabels` in the front end
|
defect
|
defect earmark memo the transactions table is not displaying the transaction type earmark memo this defect is being written for after creating an earmark memo the transactions table is not displaying the transaction type earmark memo as shown in screenshot below this was tested in both dev and stage environments dev note the isdependentchild property in the transactiontype class when set to true identifies the transaction type as being the child transaction in a double entry transaction screen like ag and af per toddlees it has no label because it s missing from scheduleatransactiontypelabels in the front end
| 1
|
578,618
| 17,149,083,820
|
IssuesEvent
|
2021-07-13 18:00:03
|
CyanLabs/Syn3Updater
|
https://api.github.com/repos/CyanLabs/Syn3Updater
|
closed
|
2.10.0 - Error: System.IndexOutOfRangeException
|
Priority: Medium Type: Bug
|
Message: Index was outside the bounds of the array.
FileHelper.cs:line 194
FileHelper.cs:line 201
DownloadViewModel.cs:line 310
DownloadViewModel.cs:line 184
|
1.0
|
2.10.0 - Error: System.IndexOutOfRangeException - Message: Index was outside the bounds of the array.
FileHelper.cs:line 194
FileHelper.cs:line 201
DownloadViewModel.cs:line 310
DownloadViewModel.cs:line 184
|
non_defect
|
error system indexoutofrangeexception message index was outside the bounds of the array filehelper cs line filehelper cs line downloadviewmodel cs line downloadviewmodel cs line
| 0
|
345,926
| 24,880,073,001
|
IssuesEvent
|
2022-10-27 23:35:31
|
quacksouls/bitburner
|
https://api.github.com/repos/quacksouls/bitburner
|
closed
|
crime: karma, combat stats, Charisma, Hack, and money per minute
|
documentation
|
Similar to #91, but we exclude the Intelligence stat.
- [x] A save file where we have destroyed BN1.3 (Genesis) and BN4.3 (Singularity). We need access to the Singularity API.
- [x] A script to calculate the following per minute for each crime: karma, combat stats, Charisma, Hack, and money.
- [x] Data on the amount of these per minute for each crime: karma, combat stats, Charisma, Hack, and money.
|
1.0
|
crime: karma, combat stats, Charisma, Hack, and money per minute - Similar to #91, but we exclude the Intelligence stat.
- [x] A save file where we have destroyed BN1.3 (Genesis) and BN4.3 (Singularity). We need access to the Singularity API.
- [x] A script to calculate the following per minute for each crime: karma, combat stats, Charisma, Hack, and money.
- [x] Data on the amount of these per minute for each crime: karma, combat stats, Charisma, Hack, and money.
|
non_defect
|
crime karma combat stats charisma hack and money per minute similar to but we exclude the intelligence stat a save file where we have destroyed genesis and singularity we need access to the singularity api a script to calculate the following per minute for each crime karma combat stats charisma hack and money data on the amount of these per minute for each crime karma combat stats charisma hack and money
| 0
|
54,592
| 13,780,022,739
|
IssuesEvent
|
2020-10-08 14:26:46
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
opened
|
Problem in code generation for entities with @Type(type = "jsonb") from hibernate-types-52 library
|
T: Defect
|
### Expected behavior
Upon executing the plugin _jooq-codegen-maven_ with the correct project structure as indicated in the jOOQ documentation, the expected generated classes from the existing project entities should be the following (e.g. for the sake of demonstration let´s use an entity called User):
1. Under the folder _/tables_ we should have a **User.java**
2. Under folder _/tables/records_ we should have a **UserRecord.java**
### Actual behavior
In my case i wanted to use the Postgres **jsonb** data type for a String field, so using the types provided by the very useful library
```
<dependency>
<groupId>com.vladmihalcea</groupId>
<artifactId>hibernate-types-52</artifactId>
<version>2.9.10</version>
</dependency>
```
we get an example similar to the one below for the User entity. Configured like this, this entity **will not** have the expected classes (User and UserRecord) derived the execution of the _jooq-codegen-maven_ plugin.
```
@Entity
@Table(name = "audit", schema = "dino_operations")
@TypeDef(name = "jsonb", typeClass = JsonBinaryType.class) // By removing this jOOQ generated expected classes
public class User {
// (...)
@Type(type = "jsonb") // By removing this jOOQ generated expected classes
@Column(name = "elements", columnDefinition = "jsonb")
private String elements;
}
```
**Summary**: Let´s say i have 10 entities and 2 annotated like the example above, so these are excluded(or a silent error occurs) from code generation so i get 8 entities being processed correctly and respective classes generated.
When i removed the annotations i got a 10 out of 10.
### Steps to reproduce the problem
- An entity with a String field annotated with the annotations indicated above
- Execute jooq-codegen-maven with the desired configurations and structure (in my case the database was org.jooq.meta.extensions.jpa.JPADatabase)
### Versions
- jOOQ (3.13.5):
- Java (8):
- Database (Postgres 10):
- OS (macOS Catalina:
- JDBC Driver (spring-jdbc from spring Spring boot 2.3.3.RELEASE):
|
1.0
|
Problem in code generation for entities with @Type(type = "jsonb") from hibernate-types-52 library - ### Expected behavior
Upon executing the plugin _jooq-codegen-maven_ with the correct project structure as indicated in the jOOQ documentation, the expected generated classes from the existing project entities should be the following (e.g. for the sake of demonstration let´s use an entity called User):
1. Under the folder _/tables_ we should have a **User.java**
2. Under folder _/tables/records_ we should have a **UserRecord.java**
### Actual behavior
In my case i wanted to use the Postgres **jsonb** data type for a String field, so using the types provided by the very useful library
```
<dependency>
<groupId>com.vladmihalcea</groupId>
<artifactId>hibernate-types-52</artifactId>
<version>2.9.10</version>
</dependency>
```
we get an example similar to the one below for the User entity. Configured like this, this entity **will not** have the expected classes (User and UserRecord) derived the execution of the _jooq-codegen-maven_ plugin.
```
@Entity
@Table(name = "audit", schema = "dino_operations")
@TypeDef(name = "jsonb", typeClass = JsonBinaryType.class) // By removing this jOOQ generated expected classes
public class User {
// (...)
@Type(type = "jsonb") // By removing this jOOQ generated expected classes
@Column(name = "elements", columnDefinition = "jsonb")
private String elements;
}
```
**Summary**: Let´s say i have 10 entities and 2 annotated like the example above, so these are excluded(or a silent error occurs) from code generation so i get 8 entities being processed correctly and respective classes generated.
When i removed the annotations i got a 10 out of 10.
### Steps to reproduce the problem
- An entity with a String field annotated with the annotations indicated above
- Execute jooq-codegen-maven with the desired configurations and structure (in my case the database was org.jooq.meta.extensions.jpa.JPADatabase)
### Versions
- jOOQ (3.13.5):
- Java (8):
- Database (Postgres 10):
- OS (macOS Catalina:
- JDBC Driver (spring-jdbc from spring Spring boot 2.3.3.RELEASE):
|
defect
|
problem in code generation for entities with type type jsonb from hibernate types library expected behavior upon executing the plugin jooq codegen maven with the correct project structure as indicated in the jooq documentation the expected generated classes from the existing project entities should be the following e g for the sake of demonstration let´s use an entity called user under the folder tables we should have a user java under folder tables records we should have a userrecord java actual behavior in my case i wanted to use the postgres jsonb data type for a string field so using the types provided by the very useful library com vladmihalcea hibernate types we get an example similar to the one below for the user entity configured like this this entity will not have the expected classes user and userrecord derived the execution of the jooq codegen maven plugin entity table name audit schema dino operations typedef name jsonb typeclass jsonbinarytype class by removing this jooq generated expected classes public class user type type jsonb by removing this jooq generated expected classes column name elements columndefinition jsonb private string elements summary let´s say i have entities and annotated like the example above so these are excluded or a silent error occurs from code generation so i get entities being processed correctly and respective classes generated when i removed the annotations i got a out of steps to reproduce the problem an entity with a string field annotated with the annotations indicated above execute jooq codegen maven with the desired configurations and structure in my case the database was org jooq meta extensions jpa jpadatabase versions jooq java database postgres os macos catalina jdbc driver spring jdbc from spring spring boot release
| 1
|
33,967
| 7,314,613,137
|
IssuesEvent
|
2018-03-01 08:03:18
|
PowerDNS/pdns
|
https://api.github.com/repos/PowerDNS/pdns
|
closed
|
dnsdist 1.2.0 complains about incorrect option but starts anyway
|
defect dnsdist
|
- Program: dnsdist <!-- delete the ones that do not apply -->
- Issue type: Bug report
### Short description
dnsdist 1.2.0 complains about option; starts anyway
<!--
If this is a bug report, use the following part of the the template and delete the part at the bottom
-->
### Steps to reproduce
```
$ /usr/sbin/dnsdist --check-bla --config /tmp/dddd
/usr/sbin/dnsdist: unrecognized option '--check-bla'
Added downstream server 127.0.0.2:5353
```
### Expected behaviour
diagnostic message and exit (1)
|
1.0
|
dnsdist 1.2.0 complains about incorrect option but starts anyway -
- Program: dnsdist <!-- delete the ones that do not apply -->
- Issue type: Bug report
### Short description
dnsdist 1.2.0 complains about option; starts anyway
<!--
If this is a bug report, use the following part of the the template and delete the part at the bottom
-->
### Steps to reproduce
```
$ /usr/sbin/dnsdist --check-bla --config /tmp/dddd
/usr/sbin/dnsdist: unrecognized option '--check-bla'
Added downstream server 127.0.0.2:5353
```
### Expected behaviour
diagnostic message and exit (1)
|
defect
|
dnsdist complains about incorrect option but starts anyway program dnsdist issue type bug report short description dnsdist complains about option starts anyway if this is a bug report use the following part of the the template and delete the part at the bottom steps to reproduce usr sbin dnsdist check bla config tmp dddd usr sbin dnsdist unrecognized option check bla added downstream server expected behaviour diagnostic message and exit
| 1
|
30,293
| 6,086,427,670
|
IssuesEvent
|
2017-06-18 00:52:57
|
jfabry/LiveRobotProgramming
|
https://api.github.com/repos/jfabry/LiveRobotProgramming
|
opened
|
PhaROS Bridge: context menu is not working
|
Bridge-PhaROS Component-UI Priority-Medium Type-Defect
|
When there is a pub or sub in the pharos bridge window, the context menu for deleting them is not working
|
1.0
|
PhaROS Bridge: context menu is not working - When there is a pub or sub in the pharos bridge window, the context menu for deleting them is not working
|
defect
|
pharos bridge context menu is not working when there is a pub or sub in the pharos bridge window the context menu for deleting them is not working
| 1
|
678,206
| 23,190,668,710
|
IssuesEvent
|
2022-08-01 12:23:27
|
SAP/xsk
|
https://api.github.com/repos/SAP/xsk
|
closed
|
[Core] Reconsider the authentication mechanisms in the XSK
|
wontfix core priority-medium effort-medium security investigation / discussion incomplete
|
Currently, when running the XSK locally, following the suggested way in the documentation, we use Form-based authentication provided by Tomcat. This leads to inconsistencies between the dev and production environments where we use OAuth.
Due to these differences, we have observed several problems:
- websockets are behaving differently on different environments due to different authorization/authentication approaches used
- Tomcat's updates break the existing Form-based authentication but not the OAuth one and this leads to blockers when updating
- difficulties in testing new code that is somehow dependent on the authentication
- maintaining code that does the same thing but in a different way
We should think of a way to use a common authentication mechanism for the environments XSK is running on. This may be Keycloak (or something similar) or at least have an easy way to run XSK locally using the same authentication services as in Kyma/Cloud Foundry, and advise developers to use this approach.
|
1.0
|
[Core] Reconsider the authentication mechanisms in the XSK - Currently, when running the XSK locally, following the suggested way in the documentation, we use Form-based authentication provided by Tomcat. This leads to inconsistencies between the dev and production environments where we use OAuth.
Due to these differences, we have observed several problems:
- websockets are behaving differently on different environments due to different authorization/authentication approaches used
- Tomcat's updates break the existing Form-based authentication but not the OAuth one and this leads to blockers when updating
- difficulties in testing new code that is somehow dependent on the authentication
- maintaining code that does the same thing but in a different way
We should think of a way to use a common authentication mechanism for the environments XSK is running on. This may be Keycloak (or something similar) or at least have an easy way to run XSK locally using the same authentication services as in Kyma/Cloud Foundry, and advise developers to use this approach.
|
non_defect
|
reconsider the authentication mechanisms in the xsk currently when running the xsk locally following the suggested way in the documentation we use form based authentication provided by tomcat this leads to inconsistencies between the dev and production environments where we use oauth due to these differences we have observed several problems websockets are behaving differently on different environments due to different authorization authentication approaches used tomcat s updates break the existing form based authentication but not the oauth one and this leads to blockers when updating difficulties in testing new code that is somehow dependent on the authentication maintaining code that does the same thing but in a different way we should think of a way to use a common authentication mechanism for the environments xsk is running on this may be keycloak or something similar or at least have an easy way to run xsk locally using the same authentication services as in kyma cloud foundry and advise developers to use this approach
| 0
|
17,213
| 2,984,426,524
|
IssuesEvent
|
2015-07-18 00:45:39
|
google/omaha
|
https://api.github.com/repos/google/omaha
|
closed
|
google Chrom or Talk Bundle cannot be installed
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. build omaha with visual studio 2010, successfully pass build and all unit
tests
2. go into staging folder
3. input
GoogleUpdate.exe /install
"bundlename=Google%20Talk%20Bundle&appguid={D0AB2EBC-931B-4013-9FEB-C9C4C2225C8C
}&appname=Google%20Talk%20Plugin&needsadmin=False&lang=en"
or
GoogleUpdate.exe /install
"bundlename=Google%20Chrome%20Bundle&appguid={8A69D345-D564-463C-AFF1-A69D9E530F
96}&appname=Google%20Chrome&needsadmin=True&lang=en"
What is the expected output?
It's supposed to download and install the product.
What do you see instead?
After downloading is completed, GoogleUpdate.exe crash
showing:
Debug Assertion Failed!
Program: ...\AppData\Local\OmahaCompanyName\Update\GoogleUpdate.exe
File:f:\dd\vctools\crt_bld\self_x86\crt\src\isctype.c
Line:56
Expression:(unsigned)(c+1)<=256
What version of the product are you using?
Visual Studio 2010, python-2.4.4, pywin32-216.win32-py2.4, scons-1.3.1.win32,
wix37, WTL80_sf, ATL_Server_Source_and_Headers_9_0_70425_Alpha,
omaha-src-r110.zip
On what operating system?
on Windows 7
Please provide any additional information below.
```
Original issue reported on code.google.com by `mmange...@gmail.com` on 7 Mar 2013 at 1:13
|
1.0
|
google Chrom or Talk Bundle cannot be installed - ```
What steps will reproduce the problem?
1. build omaha with visual studio 2010, successfully pass build and all unit
tests
2. go into staging folder
3. input
GoogleUpdate.exe /install
"bundlename=Google%20Talk%20Bundle&appguid={D0AB2EBC-931B-4013-9FEB-C9C4C2225C8C
}&appname=Google%20Talk%20Plugin&needsadmin=False&lang=en"
or
GoogleUpdate.exe /install
"bundlename=Google%20Chrome%20Bundle&appguid={8A69D345-D564-463C-AFF1-A69D9E530F
96}&appname=Google%20Chrome&needsadmin=True&lang=en"
What is the expected output?
It's supposed to download and install the product.
What do you see instead?
After downloading is completed, GoogleUpdate.exe crash
showing:
Debug Assertion Failed!
Program: ...\AppData\Local\OmahaCompanyName\Update\GoogleUpdate.exe
File:f:\dd\vctools\crt_bld\self_x86\crt\src\isctype.c
Line:56
Expression:(unsigned)(c+1)<=256
What version of the product are you using?
Visual Studio 2010, python-2.4.4, pywin32-216.win32-py2.4, scons-1.3.1.win32,
wix37, WTL80_sf, ATL_Server_Source_and_Headers_9_0_70425_Alpha,
omaha-src-r110.zip
On what operating system?
on Windows 7
Please provide any additional information below.
```
Original issue reported on code.google.com by `mmange...@gmail.com` on 7 Mar 2013 at 1:13
|
defect
|
google chrom or talk bundle cannot be installed what steps will reproduce the problem build omaha with visual studio successfully pass build and all unit tests go into staging folder input googleupdate exe install bundlename google appguid appname google needsadmin false lang en or googleupdate exe install bundlename google appguid appname google needsadmin true lang en what is the expected output it s supposed to download and install the product what do you see instead after downloading is completed googleupdate exe crash showing debug assertion failed program appdata local omahacompanyname update googleupdate exe file f dd vctools crt bld self crt src isctype c line expression unsigned c what version of the product are you using visual studio python scons sf atl server source and headers alpha omaha src zip on what operating system on windows please provide any additional information below original issue reported on code google com by mmange gmail com on mar at
| 1
|
50,596
| 13,187,609,136
|
IssuesEvent
|
2020-08-13 03:58:55
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
Link to PROPOSAL project (Trac #1019)
|
Migrated from Trac cmake defect
|
Hi,
Could you please provide a link to the paper describing how the PROPOSAL icesim meta-project work?
Thanks.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1019">https://code.icecube.wisc.edu/ticket/1019</a>, reported by icecube and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-06-11T18:38:27",
"description": "Hi,\n\nCould you please provide a link to the paper describing how the PROPOSAL icesim meta-project work?\n\nThanks.",
"reporter": "icecube",
"cc": "jtatar@lbl.gov",
"resolution": "wontfix",
"_ts": "1434047907705473",
"component": "cmake",
"summary": "Link to PROPOSAL project",
"priority": "normal",
"keywords": "",
"time": "2015-06-11T17:37:23",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Link to PROPOSAL project (Trac #1019) - Hi,
Could you please provide a link to the paper describing how the PROPOSAL icesim meta-project work?
Thanks.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1019">https://code.icecube.wisc.edu/ticket/1019</a>, reported by icecube and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-06-11T18:38:27",
"description": "Hi,\n\nCould you please provide a link to the paper describing how the PROPOSAL icesim meta-project work?\n\nThanks.",
"reporter": "icecube",
"cc": "jtatar@lbl.gov",
"resolution": "wontfix",
"_ts": "1434047907705473",
"component": "cmake",
"summary": "Link to PROPOSAL project",
"priority": "normal",
"keywords": "",
"time": "2015-06-11T17:37:23",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
|
defect
|
link to proposal project trac hi could you please provide a link to the paper describing how the proposal icesim meta project work thanks migrated from json status closed changetime description hi n ncould you please provide a link to the paper describing how the proposal icesim meta project work n nthanks reporter icecube cc jtatar lbl gov resolution wontfix ts component cmake summary link to proposal project priority normal keywords time milestone owner type defect
| 1
|
62,650
| 17,105,641,678
|
IssuesEvent
|
2021-07-09 17:15:52
|
Gogo1951/GogoLoot
|
https://api.github.com/repos/Gogo1951/GogoLoot
|
closed
|
Remove Lag from "Auto Destroy Grays"
|
GogoLoot - Application Type - Defect
|
There's a lag when you have "auto destroy grays" enabled. every time you loot there's a notable drop in frame rates.
|
1.0
|
Remove Lag from "Auto Destroy Grays" - There's a lag when you have "auto destroy grays" enabled. every time you loot there's a notable drop in frame rates.
|
defect
|
remove lag from auto destroy grays there s a lag when you have auto destroy grays enabled every time you loot there s a notable drop in frame rates
| 1
|
55,376
| 14,410,691,869
|
IssuesEvent
|
2020-12-04 05:33:03
|
AeroScripts/QuestieDev
|
https://api.github.com/repos/AeroScripts/QuestieDev
|
opened
|
Error when moving quest tracker while in combat
|
Type - Defect
|
## Bug description
When trying to move the quest tracker while in combat using "Control + Left Click" i get this error:
`...s\Questie\Modules\Tracker\QuestieTrackerPrivates.lua:24: Frame Questie_BaseFrame is not movable
[C]: in function `StartMoving'
...s\Questie\Modules\Tracker\QuestieTrackerPrivates.lua:24: in function <...s\Questie\Modules\Tracker\QuestieTrackerPrivates.lua:15>
Locals:
(*temporary) = Questie_BaseFrame {
0 = <userdata>
Update = <function> defined @Questie\Modules\Tracker\QuestieTracker.lua:228
sizer = Questie_Sizer {
}
}`
## Questie version
v6.2.3
|
1.0
|
Error when moving quest tracker while in combat - ## Bug description
When trying to move the quest tracker while in combat using "Control + Left Click" i get this error:
`...s\Questie\Modules\Tracker\QuestieTrackerPrivates.lua:24: Frame Questie_BaseFrame is not movable
[C]: in function `StartMoving'
...s\Questie\Modules\Tracker\QuestieTrackerPrivates.lua:24: in function <...s\Questie\Modules\Tracker\QuestieTrackerPrivates.lua:15>
Locals:
(*temporary) = Questie_BaseFrame {
0 = <userdata>
Update = <function> defined @Questie\Modules\Tracker\QuestieTracker.lua:228
sizer = Questie_Sizer {
}
}`
## Questie version
v6.2.3
|
defect
|
error when moving quest tracker while in combat bug description when trying to move the quest tracker while in combat using control left click i get this error s questie modules tracker questietrackerprivates lua frame questie baseframe is not movable in function startmoving s questie modules tracker questietrackerprivates lua in function locals temporary questie baseframe update defined questie modules tracker questietracker lua sizer questie sizer questie version
| 1
|
10,072
| 7,888,843,040
|
IssuesEvent
|
2018-06-28 00:14:41
|
8xprotocol/contracts
|
https://api.github.com/repos/8xprotocol/contracts
|
closed
|
Create nonce field for VolumeSubscription
|
bug in progress security
|
1. In volume subscription, create a new plan with a price and identifier
2. User subscribes to that plan
3. Delete the plan, but recreate it with the same identifier and different price OR interval (it'll have the same hash)
4. Users get charged more.
A simple resolution would be to hash the plan with the amount AND interval.
|
True
|
Create nonce field for VolumeSubscription - 1. In volume subscription, create a new plan with a price and identifier
2. User subscribes to that plan
3. Delete the plan, but recreate it with the same identifier and different price OR interval (it'll have the same hash)
4. Users get charged more.
A simple resolution would be to hash the plan with the amount AND interval.
|
non_defect
|
create nonce field for volumesubscription in volume subscription create a new plan with a price and identifier user subscribes to that plan delete the plan but recreate it with the same identifier and different price or interval it ll have the same hash users get charged more a simple resolution would be to hash the plan with the amount and interval
| 0
|
40,710
| 6,845,818,457
|
IssuesEvent
|
2017-11-13 09:45:48
|
dgraph-io/badger
|
https://api.github.com/repos/dgraph-io/badger
|
closed
|
Need to document the sorting order of the keys
|
documentation
|
There is no document describing how keys are sorted, so I have to test it with the code below.
It seems the ordering should be described as something like "byte-wise prefix order", similar to big endian but the keys are aligned from the least significant byte, and "no byte" is smaller than "byte zero" (the shorter the smaller).
I think the behavior should be clarified in the document so that we can depend on the sorting more confidently.
The test code:
```go
func TestKeyOrder(t *testing.T) {
dir := "test_order"
if err := os.MkdirAll(dir, 0755); err != nil {
t.Fatal(err)
}
defer os.RemoveAll(dir)
db, err := openBadger(dir)
if err != nil {
t.Fatal(err)
}
defer db.Close()
if err := db.Update(func(txn *badger.Txn) error {
if err := txn.Set([]byte{0, 0}, nil); err != nil {
return err
}
if err := txn.Set([]byte{0, 1}, nil); err != nil {
return err
}
if err := txn.Set([]byte{0, 2}, nil); err != nil {
return err
}
if err := txn.Set([]byte{1, 0}, nil); err != nil {
return err
}
if err := txn.Set([]byte{1, 1}, nil); err != nil {
return err
}
if err := txn.Set([]byte{1, 2}, nil); err != nil {
return err
}
if err := txn.Set([]byte{1, 2, 0}, nil); err != nil {
return err
}
if err := txn.Set([]byte{2, 0}, nil); err != nil {
return err
}
if err := txn.Set([]byte{2, 1}, nil); err != nil {
return err
}
if err := txn.Set([]byte{2, 2}, nil); err != nil {
return err
}
return nil
}); err != nil {
t.Fatal(err)
}
if err := db.View(func(txn *badger.Txn) error {
opt := badger.DefaultIteratorOptions
// opt.Reverse = true
it := txn.NewIterator(opt)
for it.Rewind(); it.Valid(); it.Next() {
item := it.Item()
k := item.Key()
fmt.Println(k)
}
return nil
}); err != nil {
t.Fatal(err)
}
}
```
|
1.0
|
Need to document the sorting order of the keys - There is no document describing how keys are sorted, so I have to test it with the code below.
It seems the ordering should be described as something like "byte-wise prefix order", similar to big endian but the keys are aligned from the least significant byte, and "no byte" is smaller than "byte zero" (the shorter the smaller).
I think the behavior should be clarified in the document so that we can depend on the sorting more confidently.
The test code:
```go
func TestKeyOrder(t *testing.T) {
dir := "test_order"
if err := os.MkdirAll(dir, 0755); err != nil {
t.Fatal(err)
}
defer os.RemoveAll(dir)
db, err := openBadger(dir)
if err != nil {
t.Fatal(err)
}
defer db.Close()
if err := db.Update(func(txn *badger.Txn) error {
if err := txn.Set([]byte{0, 0}, nil); err != nil {
return err
}
if err := txn.Set([]byte{0, 1}, nil); err != nil {
return err
}
if err := txn.Set([]byte{0, 2}, nil); err != nil {
return err
}
if err := txn.Set([]byte{1, 0}, nil); err != nil {
return err
}
if err := txn.Set([]byte{1, 1}, nil); err != nil {
return err
}
if err := txn.Set([]byte{1, 2}, nil); err != nil {
return err
}
if err := txn.Set([]byte{1, 2, 0}, nil); err != nil {
return err
}
if err := txn.Set([]byte{2, 0}, nil); err != nil {
return err
}
if err := txn.Set([]byte{2, 1}, nil); err != nil {
return err
}
if err := txn.Set([]byte{2, 2}, nil); err != nil {
return err
}
return nil
}); err != nil {
t.Fatal(err)
}
if err := db.View(func(txn *badger.Txn) error {
opt := badger.DefaultIteratorOptions
// opt.Reverse = true
it := txn.NewIterator(opt)
for it.Rewind(); it.Valid(); it.Next() {
item := it.Item()
k := item.Key()
fmt.Println(k)
}
return nil
}); err != nil {
t.Fatal(err)
}
}
```
|
non_defect
|
need to document the sorting order of the keys there is no document describing how keys are sorted so i have to test it with the code below it seems the ordering should be described as something like byte wise prefix order similar to big endian but the keys are aligned from the least significant byte and no byte is smaller than byte zero the shorter the smaller i think the behavior should be clarified in the document so that we can depend on the sorting more confidently the test code go func testkeyorder t testing t dir test order if err os mkdirall dir err nil t fatal err defer os removeall dir db err openbadger dir if err nil t fatal err defer db close if err db update func txn badger txn error if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err return nil err nil t fatal err if err db view func txn badger txn error opt badger defaultiteratoroptions opt reverse true it txn newiterator opt for it rewind it valid it next item it item k item key fmt println k return nil err nil t fatal err
| 0
|
142,213
| 5,460,265,240
|
IssuesEvent
|
2017-03-09 04:17:59
|
NuGet/Home
|
https://api.github.com/repos/NuGet/Home
|
closed
|
Logging (via status bar, output window, and warning/error window) requires UI thread and slows down Restore greatly.
|
Area: Perf Area: Restore Priority:0 Type:Bug
|
Split off from #4617
|
1.0
|
Logging (via status bar, output window, and warning/error window) requires UI thread and slows down Restore greatly. - Split off from #4617
|
non_defect
|
logging via status bar output window and warning error window requires ui thread and slows down restore greatly split off from
| 0
|
58,672
| 16,678,901,142
|
IssuesEvent
|
2021-06-07 20:05:04
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
sitewide, forms — 508-defect-3 [SCREENREADER]: Consider updating phone number documentation in design system, to include spacing in the aria-label
|
508-defect-3 508-issue-cognition 508-issue-semantic-markup 508/Accessibility components design system forms sitewide triage vsa-benefits-2 vsp-design-system-team
|
# [508-defect-3](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-3)
> Located in [VSA BAM2 MDOT MVP a11y spot check](https://github.com/department-of-veterans-affairs/va.gov-team/issues/5868#issuecomment-627425136)
> Previously documented https://github.com/department-of-veterans-affairs/va.gov-team/issues/6997
**Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
**Finding 2**
There is a new method for coding phone numbers on VA.gov, and Robin Garrison just added a PR for a global component. **Consider** adding the aria-label to ensure the screen reader reads the number empathetically. Using a title attribute on links is not recommended.
This guidance applies to all of the phone numbers within the product.
Work initiated by @Mottie: https://github.com/department-of-veterans-affairs/va.gov-team/issues/8793
## Point of Contact
**VFS Point of Contact:** Jennifer
## Acceptance Criteria
As a screen reader user, I want to read phone numbers spelling out each number with enough clarity to comprehend.
## Environment
* Operating System: all
* Browser: all
* Screenreading device: all
* Server destination: all
## Steps to Recreate
1. Enter https://staging.va.gov/hearing-aid-batteries-and-accessories/introduction in browser
2. Log in as username: vets.gov.user+1@gmail.com
3. Start screen reading device
4. Verify the title text is read, and the numbers read out may vary depending on the screen reader used; sometimes read as hundreds and thousands
## Possible Fixes (optional)
Remove the title attribute, and add an aria-label.
**Current code**
```html
<a title="Dial the telephone number 303-273-6200" href="tel:303-273-6200">303-273-6200</a>
```
**Recommended code**
```html
<a aria-label="3 0 3. 2 7 3. 6 2 0 0." href="tel:303-273-6200">303-273-6200</a>
```
## WCAG or Vendor Guidance (optional)
* [I thought title text improved accessibility. I was wrong.](https://silktide.com/blog/2013/i-thought-title-text-improved-accessibility-i-was-wrong)
* [Accessibility and phone number formatting](https://thatdevgirl.com/blog/accessibility-phone-number-formatting)
## Screenshots or Trace Logs

|
1.0
|
sitewide, forms — 508-defect-3 [SCREENREADER]: Consider updating phone number documentation in design system, to include spacing in the aria-label - # [508-defect-3](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-3)
> Located in [VSA BAM2 MDOT MVP a11y spot check](https://github.com/department-of-veterans-affairs/va.gov-team/issues/5868#issuecomment-627425136)
> Previously documented https://github.com/department-of-veterans-affairs/va.gov-team/issues/6997
**Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
**Finding 2**
There is a new method for coding phone numbers on VA.gov, and Robin Garrison just added a PR for a global component. **Consider** adding the aria-label to ensure the screen reader reads the number empathetically. Using a title attribute on links is not recommended.
This guidance applies to all of the phone numbers within the product.
Work initiated by @Mottie: https://github.com/department-of-veterans-affairs/va.gov-team/issues/8793
## Point of Contact
**VFS Point of Contact:** Jennifer
## Acceptance Criteria
As a screen reader user, I want to read phone numbers spelling out each number with enough clarity to comprehend.
## Environment
* Operating System: all
* Browser: all
* Screenreading device: all
* Server destination: all
## Steps to Recreate
1. Enter https://staging.va.gov/hearing-aid-batteries-and-accessories/introduction in browser
2. Log in as username: vets.gov.user+1@gmail.com
3. Start screen reading device
4. Verify the title text is read, and the numbers read out may vary depending on the screen reader used; sometimes read as hundreds and thousands
## Possible Fixes (optional)
Remove the title attribute, and add an aria-label.
**Current code**
```html
<a title="Dial the telephone number 303-273-6200" href="tel:303-273-6200">303-273-6200</a>
```
**Recommended code**
```html
<a aria-label="3 0 3. 2 7 3. 6 2 0 0." href="tel:303-273-6200">303-273-6200</a>
```
## WCAG or Vendor Guidance (optional)
* [I thought title text improved accessibility. I was wrong.](https://silktide.com/blog/2013/i-thought-title-text-improved-accessibility-i-was-wrong)
* [Accessibility and phone number formatting](https://thatdevgirl.com/blog/accessibility-phone-number-formatting)
## Screenshots or Trace Logs

|
defect
|
sitewide forms — defect consider updating phone number documentation in design system to include spacing in the aria label located in previously documented feedback framework ❗️ must for if the feedback must be applied ⚠️should if the feedback is best practice ✔️ consider for suggestions enhancements description finding there is a new method for coding phone numbers on va gov and robin garrison just added a pr for a global component consider adding the aria label to ensure the screen reader reads the number empathetically using a title attribute on links is not recommended this guidance applies to all of the phone numbers within the product work initiated by mottie point of contact vfs point of contact jennifer acceptance criteria as a screen reader user i want to read phone numbers spelling out each number with enough clarity to comprehend environment operating system all browser all screenreading device all server destination all steps to recreate enter in browser log in as username vets gov user gmail com start screen reading device verify the title text is read and the numbers read out may vary depending on the screen reader used sometimes read as hundreds and thousands possible fixes optional remove the title attribute and add an aria label current code html recommended code html wcag or vendor guidance optional screenshots or trace logs
| 1
|
100,627
| 12,541,716,684
|
IssuesEvent
|
2020-06-05 12:52:11
|
XAMLMarkupExtensions/WPFLocalizationExtension
|
https://api.github.com/repos/XAMLMarkupExtensions/WPFLocalizationExtension
|
opened
|
Check Binding in Binding in Setter & Docu
|
Designtime Problem Docu Enhancement
|
designtime support in setters with non Binding Markupelement
Check if this is working
```xml
<Setter Property="Text" Value="{Binding Source={lex:Loc {Binding test}}}" />
```
Docu the following cool workaround many thanks @karnah
```xml
<TextBlock FontSize="20">
<TextBlock.Style>
<Style TargetType="TextBlock">
<Setter Property="Text" Value="{Binding Source={lex:Loc en}}" />
<Style.Triggers>
<DataTrigger Binding="{Binding ElementName=CheckBox, Path=IsChecked}" Value="True">
<Setter Property="Text" Value="{Binding Source={lex:Loc de}}" />
</DataTrigger>
</Style.Triggers>
</Style>
</TextBlock.Style>
</TextBlock>
```
|
1.0
|
Check Binding in Binding in Setter & Docu - designtime support in setters with non Binding Markupelement
Check if this is working
```xml
<Setter Property="Text" Value="{Binding Source={lex:Loc {Binding test}}}" />
```
Docu the following cool workaround many thanks @karnah
```xml
<TextBlock FontSize="20">
<TextBlock.Style>
<Style TargetType="TextBlock">
<Setter Property="Text" Value="{Binding Source={lex:Loc en}}" />
<Style.Triggers>
<DataTrigger Binding="{Binding ElementName=CheckBox, Path=IsChecked}" Value="True">
<Setter Property="Text" Value="{Binding Source={lex:Loc de}}" />
</DataTrigger>
</Style.Triggers>
</Style>
</TextBlock.Style>
</TextBlock>
```
|
non_defect
|
check binding in binding in setter docu designtime support in setters with non binding markupelement check if this is working xml docu the following cool workaround many thanks karnah xml
| 0
|
53,629
| 13,261,998,102
|
IssuesEvent
|
2020-08-20 20:55:10
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
[clsim] pr2 is never used (Trac #1783)
|
Migrated from Trac cmake defect
|
unused varialbe cought by static analysis http://software.icecube.wisc.edu/static_analysis/00_LATEST/report-0ca23c.html#EndPath
Also should there be an else at line 212?
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1783">https://code.icecube.wisc.edu/projects/icecube/ticket/1783</a>, reported by kjmeagherand owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:08",
"_ts": "1550067308113782",
"description": "unused varialbe cought by static analysis http://software.icecube.wisc.edu/static_analysis/00_LATEST/report-0ca23c.html#EndPath\n\nAlso should there be an else at line 212?",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"time": "2016-07-21T17:16:57",
"component": "cmake",
"summary": "[clsim] pr2 is never used",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[clsim] pr2 is never used (Trac #1783) - unused varialbe cought by static analysis http://software.icecube.wisc.edu/static_analysis/00_LATEST/report-0ca23c.html#EndPath
Also should there be an else at line 212?
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1783">https://code.icecube.wisc.edu/projects/icecube/ticket/1783</a>, reported by kjmeagherand owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:08",
"_ts": "1550067308113782",
"description": "unused varialbe cought by static analysis http://software.icecube.wisc.edu/static_analysis/00_LATEST/report-0ca23c.html#EndPath\n\nAlso should there be an else at line 212?",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"time": "2016-07-21T17:16:57",
"component": "cmake",
"summary": "[clsim] pr2 is never used",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
|
defect
|
is never used trac unused varialbe cought by static analysis also should there be an else at line migrated from json status closed changetime ts description unused varialbe cought by static analysis should there be an else at line reporter kjmeagher cc resolution fixed time component cmake summary is never used priority normal keywords milestone owner kjmeagher type defect
| 1
|
21,817
| 3,561,887,460
|
IssuesEvent
|
2016-01-24 03:29:29
|
ariya/phantomjs
|
https://api.github.com/repos/ariya/phantomjs
|
closed
|
Linux Redhat6 Enterprize version 1. 6 install... Korean site font broken
|
old.Priority-Medium old.Status-New old.Type-Defect
|
_**[yan...@mz.co.kr](http://code.google.com/u/109652123093259324656/) commented:**_
> <b>Which version of PhantomJS are you using? Tip: run 'phantomjs --version'.</b>
> /root/phantomjs-1.6.1/bin/phantomjs /root/phantomjs-1.6.1/examples/rasterize.js http://www.daum.net daum.png
>
> <b>What steps will reproduce the problem?</b>
1. Hangul some broken
> <b>2.</b>
<b>3.</b>
> <b>What is the expected output? What do you see instead?</b>
> <b>Which operating system are you using?</b>
> <b>Did you use binary PhantomJS or did you compile it from source?</b>
> <b>Please provide any additional information below.</b>
**Disclaimer:**
This issue was migrated on 2013-03-15 from the project's former issue tracker on Google Code, [Issue #742](http://code.google.com/p/phantomjs/issues/detail?id=742).
:star2: **3** people had starred this issue at the time of migration.
|
1.0
|
Linux Redhat6 Enterprize version 1. 6 install... Korean site font broken - _**[yan...@mz.co.kr](http://code.google.com/u/109652123093259324656/) commented:**_
> <b>Which version of PhantomJS are you using? Tip: run 'phantomjs --version'.</b>
> /root/phantomjs-1.6.1/bin/phantomjs /root/phantomjs-1.6.1/examples/rasterize.js http://www.daum.net daum.png
>
> <b>What steps will reproduce the problem?</b>
1. Hangul some broken
> <b>2.</b>
<b>3.</b>
> <b>What is the expected output? What do you see instead?</b>
> <b>Which operating system are you using?</b>
> <b>Did you use binary PhantomJS or did you compile it from source?</b>
> <b>Please provide any additional information below.</b>
**Disclaimer:**
This issue was migrated on 2013-03-15 from the project's former issue tracker on Google Code, [Issue #742](http://code.google.com/p/phantomjs/issues/detail?id=742).
:star2: **3** people had starred this issue at the time of migration.
|
defect
|
linux enterprize version install korean site font broken commented which version of phantomjs are you using tip run phantomjs version root phantomjs bin phantomjs root phantomjs examples rasterize js daum png what steps will reproduce the problem hangul some broken what is the expected output what do you see instead which operating system are you using did you use binary phantomjs or did you compile it from source please provide any additional information below disclaimer this issue was migrated on from the project s former issue tracker on google code nbsp people had starred this issue at the time of migration
| 1
|
71,015
| 23,411,386,814
|
IssuesEvent
|
2022-08-12 17:57:06
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
opened
|
508-defect-3: Links that open in a new tab should warn users ahead of time
|
508/Accessibility 508-defect-3 Letters benefits-crew benefits-team-1 squad-1
|
### Point of contact
Josh Kim
### Severity level
3, Moderate. Should be fixed in 1-3 sprints post-launch.
### Details
Several issues exist when forcing a link to open in a new tab or window:
- Screen reader users, screen magnifier users, and users with certain cognitive impairments can become disoriented when they are taken to a new window
- The use of the back button, the most used button within the browser, is broken.
- Certain user agents, like kiosks, are unable to open new windows or tabs.
From WCAG guidance:
> “In general, it is better not to open new windows and tabs since they can be disorienting for people, especially people who have difficulty perceiving visual content.”
[G200: Opening new windows and tabs from a link only when necessary](https://www.w3.org/TR/WCAG20-TECHS/G200.html)
For more, [view our guidance on links (google slides)](https://docs.google.com/presentation/d/14AwUoFOg_Ix1Txl9oePtPD1ntOpylBWwBJMYq9f5vmo/edit?usp=sharing)
<img width="1499" alt="links" src="https://user-images.githubusercontent.com/14154792/184416236-c98697d0-067f-4fbc-b01a-06969ff52a95.png">
<img width="1541" alt="links-1" src="https://user-images.githubusercontent.com/14154792/184416239-81b66f5c-5b31-4ad1-8496-59f177c48970.png">
<img width="1530" alt="links2" src="https://user-images.githubusercontent.com/14154792/184416241-abdf70db-feb6-4d29-a0cf-baa4e52100d2.png">
### Reproduction steps
1. [Sign in as user 299 and go to step 2 of 2 letter list](https://staging.va.gov/records/download-va-letters/letters/letter-list)
2. Click on any of the links, confirm they all open new tabs without warning
### Proposed solution or next steps
We can either:
- Don't open the links in a new tab (which may be undesired if users will be signed out and can't directly return back to it)
- Provide a warning that links will open in a new tab by appending either:
- a text warning (opens in new tab)
- an external link icon with alt text that states the prior
### References, articles, or WCAG support
1.
2.
3.
...
### Type of issue
- [ ] Screenreader
- [ ] Keyboard
- [ ] Focus
- [ ] Headings
- [ ] Color
- [ ] Zoom
- [ ] Semantics
- [ ] Axe-core
- [ ] Something else
|
1.0
|
508-defect-3: Links that open in a new tab should warn users ahead of time - ### Point of contact
Josh Kim
### Severity level
3, Moderate. Should be fixed in 1-3 sprints post-launch.
### Details
Several issues exist when forcing a link to open in a new tab or window:
- Screen reader users, screen magnifier users, and users with certain cognitive impairments can become disoriented when they are taken to a new window
- The use of the back button, the most used button within the browser, is broken.
- Certain user agents, like kiosks, are unable to open new windows or tabs.
From WCAG guidance:
> “In general, it is better not to open new windows and tabs since they can be disorienting for people, especially people who have difficulty perceiving visual content.”
[G200: Opening new windows and tabs from a link only when necessary](https://www.w3.org/TR/WCAG20-TECHS/G200.html)
For more, [view our guidance on links (google slides)](https://docs.google.com/presentation/d/14AwUoFOg_Ix1Txl9oePtPD1ntOpylBWwBJMYq9f5vmo/edit?usp=sharing)
<img width="1499" alt="links" src="https://user-images.githubusercontent.com/14154792/184416236-c98697d0-067f-4fbc-b01a-06969ff52a95.png">
<img width="1541" alt="links-1" src="https://user-images.githubusercontent.com/14154792/184416239-81b66f5c-5b31-4ad1-8496-59f177c48970.png">
<img width="1530" alt="links2" src="https://user-images.githubusercontent.com/14154792/184416241-abdf70db-feb6-4d29-a0cf-baa4e52100d2.png">
### Reproduction steps
1. [Sign in as user 299 and go to step 2 of 2 letter list](https://staging.va.gov/records/download-va-letters/letters/letter-list)
2. Click on any of the links, confirm they all open new tabs without warning
### Proposed solution or next steps
We can either:
- Don't open the links in a new tab (which may be undesired if users will be signed out and can't directly return back to it)
- Provide a warning that links will open in a new tab by appending either:
- a text warning (opens in new tab)
- an external link icon with alt text that states the prior
### References, articles, or WCAG support
1.
2.
3.
...
### Type of issue
- [ ] Screenreader
- [ ] Keyboard
- [ ] Focus
- [ ] Headings
- [ ] Color
- [ ] Zoom
- [ ] Semantics
- [ ] Axe-core
- [ ] Something else
|
defect
|
defect links that open in a new tab should warn users ahead of time point of contact josh kim severity level moderate should be fixed in sprints post launch details several issues exist when forcing a link to open in a new tab or window screen reader users screen magnifier users and users with certain cognitive impairments can become disoriented when they are taken to a new window the use of the back button the most used button within the browser is broken certain user agents like kiosks are unable to open new windows or tabs from wcag guidance “in general it is better not to open new windows and tabs since they can be disorienting for people especially people who have difficulty perceiving visual content ” for more img width alt links src img width alt links src img width alt src reproduction steps click on any of the links confirm they all open new tabs without warning proposed solution or next steps we can either don t open the links in a new tab which may be undesired if users will be signed out and can t directly return back to it provide a warning that links will open in a new tab by appending either a text warning opens in new tab an external link icon with alt text that states the prior references articles or wcag support type of issue screenreader keyboard focus headings color zoom semantics axe core something else
| 1
|
34,372
| 7,447,865,941
|
IssuesEvent
|
2018-03-28 13:48:19
|
kerdokullamae/test_koik_issued
|
https://api.github.com/repos/kerdokullamae/test_koik_issued
|
closed
|
Isikunimed 2 väljana importida Excelist
|
C: AIS P: highest R: fixed T: defect
|
**Reported by aadikaljuvee on 19 Sep 2012 13:32 UTC**
Isiku nimed on korrastamata kujul kuigi on olemas korrastatud andmetega tabelid, kus eraldatud on Eesnimi ja Perenimi.
|
1.0
|
Isikunimed 2 väljana importida Excelist - **Reported by aadikaljuvee on 19 Sep 2012 13:32 UTC**
Isiku nimed on korrastamata kujul kuigi on olemas korrastatud andmetega tabelid, kus eraldatud on Eesnimi ja Perenimi.
|
defect
|
isikunimed väljana importida excelist reported by aadikaljuvee on sep utc isiku nimed on korrastamata kujul kuigi on olemas korrastatud andmetega tabelid kus eraldatud on eesnimi ja perenimi
| 1
|
75,455
| 25,856,419,534
|
IssuesEvent
|
2022-12-13 14:05:17
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
"You can't specify target table '...' for update in FROM clause" when target table has index hint in MySQL
|
T: Defect C: Functionality C: DB: MySQL P: Medium E: All Editions
|
When using the `USE INDEX` clause, or a similar index hint in MySQL, then the fix for #6583 doesn't work.
The reason is the same as #14387. We only traverse the join tree to find "unaliased" tables, not also "unwrapped" ones. Thus, this query is produced in an integration test:
```sql
update `test`.`t_author`
set `test`.`t_author`.`LAST_NAME` = (
select min(`test`.`t_author`.`FIRST_NAME`)
from `test`.`t_author` use index (`i_author_name`)
)
where `test`.`t_author`.`ID` = (
select max(`test`.`t_author`.`ID`)
from `test`.`t_author` use index (`i_author_name`)
)
```
Instead of:
```sql
update `test`.`t_author`
set `test`.`t_author`.`LAST_NAME` = (
select *
from (
select min(`test`.`t_author`.`FIRST_NAME`)
from `test`.`t_author` use index (`i_author_name`)
) t
)
where `test`.`t_author`.`ID` = (
select *
from (
select max(`test`.`t_author`.`ID`)
from `test`.`t_author` use index (`i_author_name`)
) t
)
```
Exception:
```
org.jooq.exception.DataAccessException: SQL [update `test`.`t_author` set `test`.`t_author`.`LAST_NAME` = (select min(`test`.`t_author`.`FIRST_NAME`) from `test`.`t_author` use index (`i_author_name`)) where `test`.`t_author`.`ID` = (select max(`test`.`t_author`.`ID`) from `test`.`t_author` use index (`i_author_name`))]; You can't specify target table 't_author' for update in FROM clause
at org.jooq_3.18.0-SNAPSHOT.MYSQL.debug(Unknown Source)
at org.jooq.impl.Tools.translate(Tools.java:3389)
at org.jooq.impl.DefaultExecuteContext.sqlException(DefaultExecuteContext.java:746)
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:357)
at org.jooq.impl.AbstractDelegatingQuery.execute(AbstractDelegatingQuery.java:115)
at org.jooq.test.all.testcases.InsertUpdateTests.testDMLIntoTWithSubqueryFromT0(InsertUpdateTests.java:390)
at org.jooq.test.all.testcases.InsertUpdateTests.testDMLIntoTWithSubqueryFromTWithTableHints(InsertUpdateTests.java:359)
at org.jooq.test.jOOQAbstractTest.testDMLIntoTWithSubqueryFromTWithTableHints(jOOQAbstractTest.java:2391)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:568)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:93)
at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:40)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:529)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:756)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:452)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:210)
Caused by: java.sql.SQLException: You can't specify target table 't_author' for update in FROM clause
at com.mysql.cj.jdbc.exceptions.SQLError.createSQLException(SQLError.java:129)
at com.mysql.cj.jdbc.exceptions.SQLExceptionsMapping.translateException(SQLExceptionsMapping.java:122)
at com.mysql.cj.jdbc.ClientPreparedStatement.executeInternal(ClientPreparedStatement.java:916)
at com.mysql.cj.jdbc.ClientPreparedStatement.execute(ClientPreparedStatement.java:354)
at org.jooq.tools.jdbc.DefaultPreparedStatement.execute(DefaultPreparedStatement.java:219)
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:422)
at org.jooq.impl.AbstractDMLQuery.execute(AbstractDMLQuery.java:980)
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:343)
... 36 more
```
|
1.0
|
"You can't specify target table '...' for update in FROM clause" when target table has index hint in MySQL - When using the `USE INDEX` clause, or a similar index hint in MySQL, then the fix for #6583 doesn't work.
The reason is the same as #14387. We only traverse the join tree to find "unaliased" tables, not also "unwrapped" ones. Thus, this query is produced in an integration test:
```sql
update `test`.`t_author`
set `test`.`t_author`.`LAST_NAME` = (
select min(`test`.`t_author`.`FIRST_NAME`)
from `test`.`t_author` use index (`i_author_name`)
)
where `test`.`t_author`.`ID` = (
select max(`test`.`t_author`.`ID`)
from `test`.`t_author` use index (`i_author_name`)
)
```
Instead of:
```sql
update `test`.`t_author`
set `test`.`t_author`.`LAST_NAME` = (
select *
from (
select min(`test`.`t_author`.`FIRST_NAME`)
from `test`.`t_author` use index (`i_author_name`)
) t
)
where `test`.`t_author`.`ID` = (
select *
from (
select max(`test`.`t_author`.`ID`)
from `test`.`t_author` use index (`i_author_name`)
) t
)
```
Exception:
```
org.jooq.exception.DataAccessException: SQL [update `test`.`t_author` set `test`.`t_author`.`LAST_NAME` = (select min(`test`.`t_author`.`FIRST_NAME`) from `test`.`t_author` use index (`i_author_name`)) where `test`.`t_author`.`ID` = (select max(`test`.`t_author`.`ID`) from `test`.`t_author` use index (`i_author_name`))]; You can't specify target table 't_author' for update in FROM clause
at org.jooq_3.18.0-SNAPSHOT.MYSQL.debug(Unknown Source)
at org.jooq.impl.Tools.translate(Tools.java:3389)
at org.jooq.impl.DefaultExecuteContext.sqlException(DefaultExecuteContext.java:746)
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:357)
at org.jooq.impl.AbstractDelegatingQuery.execute(AbstractDelegatingQuery.java:115)
at org.jooq.test.all.testcases.InsertUpdateTests.testDMLIntoTWithSubqueryFromT0(InsertUpdateTests.java:390)
at org.jooq.test.all.testcases.InsertUpdateTests.testDMLIntoTWithSubqueryFromTWithTableHints(InsertUpdateTests.java:359)
at org.jooq.test.jOOQAbstractTest.testDMLIntoTWithSubqueryFromTWithTableHints(jOOQAbstractTest.java:2391)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:568)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:93)
at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:40)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:529)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:756)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:452)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:210)
Caused by: java.sql.SQLException: You can't specify target table 't_author' for update in FROM clause
at com.mysql.cj.jdbc.exceptions.SQLError.createSQLException(SQLError.java:129)
at com.mysql.cj.jdbc.exceptions.SQLExceptionsMapping.translateException(SQLExceptionsMapping.java:122)
at com.mysql.cj.jdbc.ClientPreparedStatement.executeInternal(ClientPreparedStatement.java:916)
at com.mysql.cj.jdbc.ClientPreparedStatement.execute(ClientPreparedStatement.java:354)
at org.jooq.tools.jdbc.DefaultPreparedStatement.execute(DefaultPreparedStatement.java:219)
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:422)
at org.jooq.impl.AbstractDMLQuery.execute(AbstractDMLQuery.java:980)
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:343)
... 36 more
```
|
defect
|
you can t specify target table for update in from clause when target table has index hint in mysql when using the use index clause or a similar index hint in mysql then the fix for doesn t work the reason is the same as we only traverse the join tree to find unaliased tables not also unwrapped ones thus this query is produced in an integration test sql update test t author set test t author last name select min test t author first name from test t author use index i author name where test t author id select max test t author id from test t author use index i author name instead of sql update test t author set test t author last name select from select min test t author first name from test t author use index i author name t where test t author id select from select max test t author id from test t author use index i author name t exception org jooq exception dataaccessexception sql you can t specify target table t author for update in from clause at org jooq snapshot mysql debug unknown source at org jooq impl tools translate tools java at org jooq impl defaultexecutecontext sqlexception defaultexecutecontext java at org jooq impl abstractquery execute abstractquery java at org jooq impl abstractdelegatingquery execute abstractdelegatingquery java at org jooq test all testcases insertupdatetests insertupdatetests java at org jooq test all testcases insertupdatetests testdmlintotwithsubqueryfromtwithtablehints insertupdatetests java at org jooq test jooqabstracttest testdmlintotwithsubqueryfromtwithtablehints jooqabstracttest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit rules testwatcher evaluate testwatcher java at org junit rules testwatcher evaluate testwatcher java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org eclipse jdt internal runner run java at org eclipse jdt internal junit runner testexecution run testexecution java at org eclipse jdt internal junit runner remotetestrunner runtests remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner runtests remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner run remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner main remotetestrunner java caused by java sql sqlexception you can t specify target table t author for update in from clause at com mysql cj jdbc exceptions sqlerror createsqlexception sqlerror java at com mysql cj jdbc exceptions sqlexceptionsmapping translateexception sqlexceptionsmapping java at com mysql cj jdbc clientpreparedstatement executeinternal clientpreparedstatement java at com mysql cj jdbc clientpreparedstatement execute clientpreparedstatement java at org jooq tools jdbc defaultpreparedstatement execute defaultpreparedstatement java at org jooq impl abstractquery execute abstractquery java at org jooq impl abstractdmlquery execute abstractdmlquery java at org jooq impl abstractquery execute abstractquery java more
| 1
|
65,791
| 19,694,627,944
|
IssuesEvent
|
2022-01-12 10:48:06
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Message bubbles have too much padding on the left of the timeline
|
T-Defect S-Minor A-Message-Bubbles O-Occasional
|
### Steps to reproduce
1. Switch from Modern layout to Message Bubbles in the settings
### Outcome
#### What did you expect?

#### What happened instead?

### Operating system
Arch Linux
### Browser information
Chromium 96.0.4664.45 (Official Build) Arch Linux (64-bit)
### URL for webapp
develop.element.io
### Application version
Element version: ef9335697e05-react-b8355883313e-js-6fc586598a98 Olm version: 3.2.8
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Message bubbles have too much padding on the left of the timeline - ### Steps to reproduce
1. Switch from Modern layout to Message Bubbles in the settings
### Outcome
#### What did you expect?

#### What happened instead?

### Operating system
Arch Linux
### Browser information
Chromium 96.0.4664.45 (Official Build) Arch Linux (64-bit)
### URL for webapp
develop.element.io
### Application version
Element version: ef9335697e05-react-b8355883313e-js-6fc586598a98 Olm version: 3.2.8
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
message bubbles have too much padding on the left of the timeline steps to reproduce switch from modern layout to message bubbles in the settings outcome what did you expect what happened instead operating system arch linux browser information chromium official build arch linux bit url for webapp develop element io application version element version react js olm version homeserver no response will you send logs no
| 1
|
786,863
| 27,696,715,900
|
IssuesEvent
|
2023-03-14 03:11:46
|
mudkipdev/leaf
|
https://api.github.com/repos/mudkipdev/leaf
|
opened
|
Tag Aliases
|
Type: Feature Status: Planned Priority: Low
|
**I have not figured out how this feature should work exactly,** but users should be able to create tags that point to another tag. I'm thinking aliases will be independent from the tag, and can be separated from the main tag entirely and have its own content.
This will probably add a `/tags alias <alias: str> <tag: str>` that points the alias tag to the main tag. In addition, to create an alias, you can use a new option in the `/tags create` command named "alias_to" or something similar that will skip the content and create the tag instantly.
Part of #1.
|
1.0
|
Tag Aliases - **I have not figured out how this feature should work exactly,** but users should be able to create tags that point to another tag. I'm thinking aliases will be independent from the tag, and can be separated from the main tag entirely and have its own content.
This will probably add a `/tags alias <alias: str> <tag: str>` that points the alias tag to the main tag. In addition, to create an alias, you can use a new option in the `/tags create` command named "alias_to" or something similar that will skip the content and create the tag instantly.
Part of #1.
|
non_defect
|
tag aliases i have not figured out how this feature should work exactly but users should be able to create tags that point to another tag i m thinking aliases will be independent from the tag and can be separated from the main tag entirely and have its own content this will probably add a tags alias that points the alias tag to the main tag in addition to create an alias you can use a new option in the tags create command named alias to or something similar that will skip the content and create the tag instantly part of
| 0
|
9,429
| 2,615,149,688
|
IssuesEvent
|
2015-03-01 06:26:47
|
chrsmith/reaver-wps
|
https://api.github.com/repos/chrsmith/reaver-wps
|
closed
|
WARNING: Receive timeout occurred - WPS transaction failed, re-trying last pin
|
auto-migrated Priority-Low Type-Defect
|
```
Hi guys,
Playing about with my router trying to get this working. It was trying PINs up
until 2 days ago, it seems to have a permanent lock out or something? I have
pasted the console commands and have attached a cap - Anyone got any ideas?
root@bt:~# reaver -i mon0 -b 00:22:75:B8:98:41 -w -r 5:60 -l 300 -c 6 -L -vv
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Waiting for beacon from 00:22:75:B8:98:41
[+] Associated with 00:22:75:B8:98:41
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending identity response
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[!] WPS transaction failed, re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending identity response
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[!] WPS transaction failed, re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[!] WPS transaction failed, re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[!] WPS transaction failed, re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending identity response
[+] Sending identity response
[+] Sending identity response
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[!] WPS transaction failed, re-trying last pin
[+] Entering recurring delay of 60 seconds
^C
[+] Nothing done, nothing to save.
root@bt:~#
```
Original issue reported on code.google.com by `themosts...@googlemail.com` on 12 Jan 2012 at 7:39
Attachments:
* [ReaverDUMP-02.cap](https://storage.googleapis.com/google-code-attachments/reaver-wps/issue-132/comment-0/ReaverDUMP-02.cap)
|
1.0
|
WARNING: Receive timeout occurred - WPS transaction failed, re-trying last pin - ```
Hi guys,
Playing about with my router trying to get this working. It was trying PINs up
until 2 days ago, it seems to have a permanent lock out or something? I have
pasted the console commands and have attached a cap - Anyone got any ideas?
root@bt:~# reaver -i mon0 -b 00:22:75:B8:98:41 -w -r 5:60 -l 300 -c 6 -L -vv
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Waiting for beacon from 00:22:75:B8:98:41
[+] Associated with 00:22:75:B8:98:41
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending identity response
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[!] WPS transaction failed, re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending identity response
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[!] WPS transaction failed, re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[!] WPS transaction failed, re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[!] WPS transaction failed, re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Sending identity response
[+] Sending identity response
[+] Sending identity response
[+] Sending identity response
[!] WARNING: Receive timeout occurred
[!] WPS transaction failed, re-trying last pin
[+] Entering recurring delay of 60 seconds
^C
[+] Nothing done, nothing to save.
root@bt:~#
```
Original issue reported on code.google.com by `themosts...@googlemail.com` on 12 Jan 2012 at 7:39
Attachments:
* [ReaverDUMP-02.cap](https://storage.googleapis.com/google-code-attachments/reaver-wps/issue-132/comment-0/ReaverDUMP-02.cap)
|
defect
|
warning receive timeout occurred wps transaction failed re trying last pin hi guys playing about with my router trying to get this working it was trying pins up until days ago it seems to have a permanent lock out or something i have pasted the console commands and have attached a cap anyone got any ideas root bt reaver i b w r l c l vv reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner waiting for beacon from associated with trying pin sending eapol start request sending identity response sending identity response warning receive timeout occurred wps transaction failed re trying last pin trying pin sending eapol start request sending identity response sending identity response warning receive timeout occurred wps transaction failed re trying last pin trying pin sending eapol start request sending identity response warning receive timeout occurred wps transaction failed re trying last pin trying pin sending eapol start request sending identity response warning receive timeout occurred wps transaction failed re trying last pin trying pin sending eapol start request sending identity response sending identity response sending identity response sending identity response warning receive timeout occurred wps transaction failed re trying last pin entering recurring delay of seconds c nothing done nothing to save root bt original issue reported on code google com by themosts googlemail com on jan at attachments
| 1
|
192,593
| 14,620,732,538
|
IssuesEvent
|
2020-12-22 20:16:40
|
backend-br/vagas
|
https://api.github.com/repos/backend-br/vagas
|
closed
|
[Maceió/Remoto] Back-end Developer @Ilhasoft
|
Django Docker Kubernetes Linux PJ PostgreSQL Testes automatizados
|
## Nossa empresa
Somos especialistas em desenvolver chatbots que utilizam os mais sofisticados recursos da Inteligência Artificial para automatizar a comunicação entre organizações e pessoas ao redor do mundo.
### Onde queremos chegar?
Crescer com sustentabilidade, permanecendo com nossas raízes.
Queremos conectar comunidades e fazer com que o uso da tecnologia seja democratizado: alcance a todos, onde quer que seja. Queremos inovar, revolucionar, provocar novos comportamentos. Queremos facilitar o acesso aos melhores serviços de um jeito rápido e prático, sem burocracias.
Queremos criar soluções para problemas, otimizar a interação entre empresas e consumidores, levar a igualdade para os lugares mais remotos, trabalhar em parceria com ONGs e Governos para conscientizar as pessoas e dar-lhes mais qualidade de vida.
Queremos fazer parte de uma nova era sem sair da nossa terra!
### Panorama atual
A pandemia do COVID-19 tem exigido bastante do nosso time. Nossos produtos assumiram um papel de destaque na luta contra o vírus, atualmente estamos ajudando a salvar milhares de vidas em estados e municípios no cenário nacional e internacional. Conheça alguns projetos:
* Health Buddy: Chatbot informativo multi-idiomas criado para UNICEF e para a OMS (http://healthbuddy.info/);
* Governo do Ceará: Chatbot informativo, com atendimento ao paciente por profissionais de saúde em casos graves (https://www.ceara.gov.br/)
* Governo da Bahia: Chatbot informativo, com atendimento ao paciente por profissionais de saúde em casos graves (WhatsApp: [71 3118-8499](tel:71 3118-8499))
* Governo de Alagoas: Chatbot informativo, com atendimento ao paciente por profissionais de saúde em casos graves (WhatsApp: [82 3315-1532](tel:82 3315-1532))
## Descrição da vaga
Como Django Back-end Developer, você integrará nosso time no desenvolvimento de software para correção de bugs, desenvolvimento de novas funcionalidades nas plataformas de criação de gestão de Chatbots da Ilhasoft.
No fundo, nós queremos você no time para:
* Crescer conosco à medida que desbravamos novos projetos
* Ter orgulho do trabalho que desenvolve
* Melhorar a cada dia
* Ter a certeza de ter contribuído para problemas de nossos clientes após um dia de trabalho desafiador
* Ajudar todos seus colegas de trabalho a completar o check list acima
## Local
Após a pandemia a Ilhasoft se tornou uma empresa 100% remota com uma equipe distribuída em vários países.
## Requisitos
**Obrigatórios:**
* 1 ano de experiência com Django
* 1 ano de experiência com PostgreSQL
* Experiência prévia com Gitflow
**Diferenciais:**
* Um projeto próprio desenvolvido em Django
* Projetos de código aberto.
* Conhecimento em servidores linux, Docker, Kubernetes
* Fluência em Inglês
* Testes Automatizados
## Contratação
CLT ou PJ a combinar
## Como se candidatar
Cadastre-se no seguinte link: https://ilhasoft.gupy.io/jobs/600315
## Tempo médio de feedbacks
Costumamos enviar feedbacks em até 3 dias após cada processo.
E-mail para contato em caso de não haver resposta: [rh@ilhasoft.com.br](mailto:rh@ilhasoft.com.br)
#### Nível
* Júnior
|
1.0
|
[Maceió/Remoto] Back-end Developer @Ilhasoft - ## Nossa empresa
Somos especialistas em desenvolver chatbots que utilizam os mais sofisticados recursos da Inteligência Artificial para automatizar a comunicação entre organizações e pessoas ao redor do mundo.
### Onde queremos chegar?
Crescer com sustentabilidade, permanecendo com nossas raízes.
Queremos conectar comunidades e fazer com que o uso da tecnologia seja democratizado: alcance a todos, onde quer que seja. Queremos inovar, revolucionar, provocar novos comportamentos. Queremos facilitar o acesso aos melhores serviços de um jeito rápido e prático, sem burocracias.
Queremos criar soluções para problemas, otimizar a interação entre empresas e consumidores, levar a igualdade para os lugares mais remotos, trabalhar em parceria com ONGs e Governos para conscientizar as pessoas e dar-lhes mais qualidade de vida.
Queremos fazer parte de uma nova era sem sair da nossa terra!
### Panorama atual
A pandemia do COVID-19 tem exigido bastante do nosso time. Nossos produtos assumiram um papel de destaque na luta contra o vírus, atualmente estamos ajudando a salvar milhares de vidas em estados e municípios no cenário nacional e internacional. Conheça alguns projetos:
* Health Buddy: Chatbot informativo multi-idiomas criado para UNICEF e para a OMS (http://healthbuddy.info/);
* Governo do Ceará: Chatbot informativo, com atendimento ao paciente por profissionais de saúde em casos graves (https://www.ceara.gov.br/)
* Governo da Bahia: Chatbot informativo, com atendimento ao paciente por profissionais de saúde em casos graves (WhatsApp: [71 3118-8499](tel:71 3118-8499))
* Governo de Alagoas: Chatbot informativo, com atendimento ao paciente por profissionais de saúde em casos graves (WhatsApp: [82 3315-1532](tel:82 3315-1532))
## Descrição da vaga
Como Django Back-end Developer, você integrará nosso time no desenvolvimento de software para correção de bugs, desenvolvimento de novas funcionalidades nas plataformas de criação de gestão de Chatbots da Ilhasoft.
No fundo, nós queremos você no time para:
* Crescer conosco à medida que desbravamos novos projetos
* Ter orgulho do trabalho que desenvolve
* Melhorar a cada dia
* Ter a certeza de ter contribuído para problemas de nossos clientes após um dia de trabalho desafiador
* Ajudar todos seus colegas de trabalho a completar o check list acima
## Local
Após a pandemia a Ilhasoft se tornou uma empresa 100% remota com uma equipe distribuída em vários países.
## Requisitos
**Obrigatórios:**
* 1 ano de experiência com Django
* 1 ano de experiência com PostgreSQL
* Experiência prévia com Gitflow
**Diferenciais:**
* Um projeto próprio desenvolvido em Django
* Projetos de código aberto.
* Conhecimento em servidores linux, Docker, Kubernetes
* Fluência em Inglês
* Testes Automatizados
## Contratação
CLT ou PJ a combinar
## Como se candidatar
Cadastre-se no seguinte link: https://ilhasoft.gupy.io/jobs/600315
## Tempo médio de feedbacks
Costumamos enviar feedbacks em até 3 dias após cada processo.
E-mail para contato em caso de não haver resposta: [rh@ilhasoft.com.br](mailto:rh@ilhasoft.com.br)
#### Nível
* Júnior
|
non_defect
|
back end developer ilhasoft nossa empresa somos especialistas em desenvolver chatbots que utilizam os mais sofisticados recursos da inteligência artificial para automatizar a comunicação entre organizações e pessoas ao redor do mundo onde queremos chegar crescer com sustentabilidade permanecendo com nossas raízes queremos conectar comunidades e fazer com que o uso da tecnologia seja democratizado alcance a todos onde quer que seja queremos inovar revolucionar provocar novos comportamentos queremos facilitar o acesso aos melhores serviços de um jeito rápido e prático sem burocracias queremos criar soluções para problemas otimizar a interação entre empresas e consumidores levar a igualdade para os lugares mais remotos trabalhar em parceria com ongs e governos para conscientizar as pessoas e dar lhes mais qualidade de vida queremos fazer parte de uma nova era sem sair da nossa terra panorama atual a pandemia do covid tem exigido bastante do nosso time nossos produtos assumiram um papel de destaque na luta contra o vírus atualmente estamos ajudando a salvar milhares de vidas em estados e municípios no cenário nacional e internacional conheça alguns projetos health buddy chatbot informativo multi idiomas criado para unicef e para a oms governo do ceará chatbot informativo com atendimento ao paciente por profissionais de saúde em casos graves governo da bahia chatbot informativo com atendimento ao paciente por profissionais de saúde em casos graves whatsapp tel governo de alagoas chatbot informativo com atendimento ao paciente por profissionais de saúde em casos graves whatsapp tel descrição da vaga como django back end developer você integrará nosso time no desenvolvimento de software para correção de bugs desenvolvimento de novas funcionalidades nas plataformas de criação de gestão de chatbots da ilhasoft no fundo nós queremos você no time para crescer conosco à medida que desbravamos novos projetos ter orgulho do trabalho que desenvolve melhorar a cada dia ter a certeza de ter contribuído para problemas de nossos clientes após um dia de trabalho desafiador ajudar todos seus colegas de trabalho a completar o check list acima local após a pandemia a ilhasoft se tornou uma empresa remota com uma equipe distribuída em vários países requisitos obrigatórios ano de experiência com django ano de experiência com postgresql experiência prévia com gitflow diferenciais um projeto próprio desenvolvido em django projetos de código aberto conhecimento em servidores linux docker kubernetes fluência em inglês testes automatizados contratação clt ou pj a combinar como se candidatar cadastre se no seguinte link tempo médio de feedbacks costumamos enviar feedbacks em até dias após cada processo e mail para contato em caso de não haver resposta mailto rh ilhasoft com br nível júnior
| 0
|
37,998
| 5,157,156,125
|
IssuesEvent
|
2017-01-16 04:46:36
|
containenv/rapcom
|
https://api.github.com/repos/containenv/rapcom
|
closed
|
Test all major functionality.
|
test
|
Rapcom needs tests to cover all expected functionality. This is not intended to be exhaustive, but should at least handle the happy paths.
|
1.0
|
Test all major functionality. - Rapcom needs tests to cover all expected functionality. This is not intended to be exhaustive, but should at least handle the happy paths.
|
non_defect
|
test all major functionality rapcom needs tests to cover all expected functionality this is not intended to be exhaustive but should at least handle the happy paths
| 0
|
3,452
| 6,542,647,946
|
IssuesEvent
|
2017-09-02 10:26:27
|
pwittchen/ReactiveNetwork
|
https://api.github.com/repos/pwittchen/ReactiveNetwork
|
opened
|
Relase 0.12.1 (RxJava2.x)
|
release process RxJava2.x
|
**Initial release notes**:
Fixed memory leak in `PreLollipopNetworkObservingStrategy` during disposing of an `Observable` - issue #219.
**Things to do**:
TBD.
|
1.0
|
Relase 0.12.1 (RxJava2.x) - **Initial release notes**:
Fixed memory leak in `PreLollipopNetworkObservingStrategy` during disposing of an `Observable` - issue #219.
**Things to do**:
TBD.
|
non_defect
|
relase x initial release notes fixed memory leak in prelollipopnetworkobservingstrategy during disposing of an observable issue things to do tbd
| 0
|
47,116
| 13,056,034,204
|
IssuesEvent
|
2020-07-30 03:27:14
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
glshovel docs and plugin examples (Trac #15)
|
Migrated from Trac defect glshovel
|
Migrated from https://code.icecube.wisc.edu/ticket/15
```json
{
"status": "closed",
"changetime": "2007-11-11T03:51:18",
"description": "\n",
"reporter": "troy",
"cc": "",
"resolution": "duplicate",
"_ts": "1194753078000000",
"component": "glshovel",
"summary": "glshovel docs and plugin examples",
"priority": "normal",
"keywords": "",
"time": "2007-06-03T16:33:00",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
|
1.0
|
glshovel docs and plugin examples (Trac #15) -
Migrated from https://code.icecube.wisc.edu/ticket/15
```json
{
"status": "closed",
"changetime": "2007-11-11T03:51:18",
"description": "\n",
"reporter": "troy",
"cc": "",
"resolution": "duplicate",
"_ts": "1194753078000000",
"component": "glshovel",
"summary": "glshovel docs and plugin examples",
"priority": "normal",
"keywords": "",
"time": "2007-06-03T16:33:00",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
|
defect
|
glshovel docs and plugin examples trac migrated from json status closed changetime description n reporter troy cc resolution duplicate ts component glshovel summary glshovel docs and plugin examples priority normal keywords time milestone owner troy type defect
| 1
|
24,436
| 3,980,349,103
|
IssuesEvent
|
2016-05-06 06:59:47
|
Quantum64/Arcade-Issues
|
https://api.github.com/repos/Quantum64/Arcade-Issues
|
opened
|
Kill Effect not working
|
cosmeticsmenu defect effects killeffect
|
The "Notes Kill Effect" does not work.
With the other kill effects it shows a piece of redstone but with the notes one it shows paper, Like what you would see with a title.
http://i.imgur.com/o7hHeYs.png
|
1.0
|
Kill Effect not working - The "Notes Kill Effect" does not work.
With the other kill effects it shows a piece of redstone but with the notes one it shows paper, Like what you would see with a title.
http://i.imgur.com/o7hHeYs.png
|
defect
|
kill effect not working the notes kill effect does not work with the other kill effects it shows a piece of redstone but with the notes one it shows paper like what you would see with a title
| 1
|
40,417
| 9,984,513,603
|
IssuesEvent
|
2019-07-10 14:40:07
|
telus/tds-core
|
https://api.github.com/repos/telus/tds-core
|
closed
|
TDS Tooltip - Accessibility & Language issues
|
accessibility :wheelchair: priority: medium status: in progress type: defect :bug:
|
## Description
- The `aria-label` inside the icon "Reveal additional information" does not seem to be translatable in French, we can add more text after it but not actually change it. Additionally, the Wave extension seems to flag the button as an error ("Empty label") because it doesn't see any content inside it.
- If possible, please translate "Reveal additional information" if it's used in a `/fr/` page to "Afficher l’information supplémentaire". If the "Empty label" issue is not a false flag, there needs to be an actual label inside of the button for screen readers, etc.
## Reproduction Steps
1. View the Tooltip component on a `/fr/` page.
2. `aria-label` remains in English.
3. Scan the page with Wave and the "Empty label" issue pops (might be a false flag?).
## Meta
- TDS component version: @tds/core-tooltip”: “^2.0.4
- Willing to develop solution: No
- Has workaround: No
- High impact: No
## Screenshots

|
1.0
|
TDS Tooltip - Accessibility & Language issues - ## Description
- The `aria-label` inside the icon "Reveal additional information" does not seem to be translatable in French, we can add more text after it but not actually change it. Additionally, the Wave extension seems to flag the button as an error ("Empty label") because it doesn't see any content inside it.
- If possible, please translate "Reveal additional information" if it's used in a `/fr/` page to "Afficher l’information supplémentaire". If the "Empty label" issue is not a false flag, there needs to be an actual label inside of the button for screen readers, etc.
## Reproduction Steps
1. View the Tooltip component on a `/fr/` page.
2. `aria-label` remains in English.
3. Scan the page with Wave and the "Empty label" issue pops (might be a false flag?).
## Meta
- TDS component version: @tds/core-tooltip”: “^2.0.4
- Willing to develop solution: No
- Has workaround: No
- High impact: No
## Screenshots

|
defect
|
tds tooltip accessibility language issues description the aria label inside the icon reveal additional information does not seem to be translatable in french we can add more text after it but not actually change it additionally the wave extension seems to flag the button as an error empty label because it doesn t see any content inside it if possible please translate reveal additional information if it s used in a fr page to afficher l’information supplémentaire if the empty label issue is not a false flag there needs to be an actual label inside of the button for screen readers etc reproduction steps view the tooltip component on a fr page aria label remains in english scan the page with wave and the empty label issue pops might be a false flag meta tds component version tds core tooltip” “ willing to develop solution no has workaround no high impact no screenshots
| 1
|
14,198
| 8,506,478,649
|
IssuesEvent
|
2018-10-30 16:39:57
|
CompletelyFairGames/dwarfcorp
|
https://api.github.com/repos/CompletelyFairGames/dwarfcorp
|
opened
|
Performance: As soon as ten dwarves are in game, performance dies
|
A Bug Performance
|


[Borderland_413_131853894854936125.zip](https://github.com/CompletelyFairGames/dwarfcorp/files/2530519/Borderland_413_131853894854936125.zip)
[Borderland.zip](https://github.com/CompletelyFairGames/dwarfcorp/files/2530522/Borderland.zip)
[DwarfCorp181030_IPSummary.zip](https://github.com/CompletelyFairGames/dwarfcorp/files/2530542/DwarfCorp181030_IPSummary.zip)
Performance is measurably worse. It went from spikes of 60% before to a sustained 80% CPU. On 1x with ten dwarves or so, FPS is staying at 0 FPS.
I'm guessing something in our refactor missed something because things were running totally fine until I switched to 64f01d10.
This a repro, as far as I can tell:
1. New game.
2. Set a dig task to the bottom
3. God/hire five more miners (for some reason reaching 10 hires appears to be exactly when perf drops. Not 9. Not 8. If this is actually perf related, I'm guessing it'll take more for you all to repro the bug.)
4. Observe game perf drop to 0 FPS after about a minute or two.
|
True
|
Performance: As soon as ten dwarves are in game, performance dies - 

[Borderland_413_131853894854936125.zip](https://github.com/CompletelyFairGames/dwarfcorp/files/2530519/Borderland_413_131853894854936125.zip)
[Borderland.zip](https://github.com/CompletelyFairGames/dwarfcorp/files/2530522/Borderland.zip)
[DwarfCorp181030_IPSummary.zip](https://github.com/CompletelyFairGames/dwarfcorp/files/2530542/DwarfCorp181030_IPSummary.zip)
Performance is measurably worse. It went from spikes of 60% before to a sustained 80% CPU. On 1x with ten dwarves or so, FPS is staying at 0 FPS.
I'm guessing something in our refactor missed something because things were running totally fine until I switched to 64f01d10.
This a repro, as far as I can tell:
1. New game.
2. Set a dig task to the bottom
3. God/hire five more miners (for some reason reaching 10 hires appears to be exactly when perf drops. Not 9. Not 8. If this is actually perf related, I'm guessing it'll take more for you all to repro the bug.)
4. Observe game perf drop to 0 FPS after about a minute or two.
|
non_defect
|
performance as soon as ten dwarves are in game performance dies performance is measurably worse it went from spikes of before to a sustained cpu on with ten dwarves or so fps is staying at fps i m guessing something in our refactor missed something because things were running totally fine until i switched to this a repro as far as i can tell new game set a dig task to the bottom god hire five more miners for some reason reaching hires appears to be exactly when perf drops not not if this is actually perf related i m guessing it ll take more for you all to repro the bug observe game perf drop to fps after about a minute or two
| 0
|
264,908
| 28,214,112,133
|
IssuesEvent
|
2023-04-05 07:39:38
|
hshivhare67/platform_device_renesas_kernel_v4.19.72
|
https://api.github.com/repos/hshivhare67/platform_device_renesas_kernel_v4.19.72
|
closed
|
CVE-2022-0847 (High) detected in linuxlinux-4.19.279 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2022-0847 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/iov_iter.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/iov_iter.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the way the "flags" member of the new pipe buffer structure was lacking proper initialization in copy_page_to_iter_pipe and push_pipe functions in the Linux kernel and could thus contain stale values. An unprivileged local user could use this flaw to write to pages in the page cache backed by read only files and as such escalate their privileges on the system.
<p>Publish Date: 2022-03-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0847>CVE-2022-0847</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-0847">https://www.linuxkernelcves.com/cves/CVE-2022-0847</a></p>
<p>Release Date: 2022-03-10</p>
<p>Fix Resolution: v5.10.102,v5.15.25,v5.16.11,v5.17-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-0847 (High) detected in linuxlinux-4.19.279 - autoclosed - ## CVE-2022-0847 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/iov_iter.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/iov_iter.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the way the "flags" member of the new pipe buffer structure was lacking proper initialization in copy_page_to_iter_pipe and push_pipe functions in the Linux kernel and could thus contain stale values. An unprivileged local user could use this flaw to write to pages in the page cache backed by read only files and as such escalate their privileges on the system.
<p>Publish Date: 2022-03-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0847>CVE-2022-0847</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-0847">https://www.linuxkernelcves.com/cves/CVE-2022-0847</a></p>
<p>Release Date: 2022-03-10</p>
<p>Fix Resolution: v5.10.102,v5.15.25,v5.16.11,v5.17-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch main vulnerable source files lib iov iter c lib iov iter c vulnerability details a flaw was found in the way the flags member of the new pipe buffer structure was lacking proper initialization in copy page to iter pipe and push pipe functions in the linux kernel and could thus contain stale values an unprivileged local user could use this flaw to write to pages in the page cache backed by read only files and as such escalate their privileges on the system publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
176,490
| 28,102,202,746
|
IssuesEvent
|
2023-03-30 20:27:36
|
WordPress/openverse
|
https://api.github.com/repos/WordPress/openverse
|
opened
|
Add instructions for navigating sub-categories in Figma
|
🟩 priority: low 🌟 goal: addition 🖼️ aspect: design 🧱 stack: frontend
|
## Description
<!-- Describe the feature and how it solves the problem. -->
Originally reported by @sarayourfriend
> Could figma page have instructions for where to find the sub-categories? I forgot how to find the other pages in Figma and had to click around a bit before I could find it
Link: https://www.figma.com/file/GIIQ4sDbaToCfFQyKMvzr8/Openverse-Design-Library?node-id=0-1
## Alternatives
<!-- Describe any alternative solutions or features you have considered. How is this feature better? -->
## Additional context
<!-- Add any other context about the feature here; or delete the section entirely. -->
<!-- If you would like to work on this, please comment below separately. -->
|
1.0
|
Add instructions for navigating sub-categories in Figma - ## Description
<!-- Describe the feature and how it solves the problem. -->
Originally reported by @sarayourfriend
> Could figma page have instructions for where to find the sub-categories? I forgot how to find the other pages in Figma and had to click around a bit before I could find it
Link: https://www.figma.com/file/GIIQ4sDbaToCfFQyKMvzr8/Openverse-Design-Library?node-id=0-1
## Alternatives
<!-- Describe any alternative solutions or features you have considered. How is this feature better? -->
## Additional context
<!-- Add any other context about the feature here; or delete the section entirely. -->
<!-- If you would like to work on this, please comment below separately. -->
|
non_defect
|
add instructions for navigating sub categories in figma description originally reported by sarayourfriend could figma page have instructions for where to find the sub categories i forgot how to find the other pages in figma and had to click around a bit before i could find it link alternatives additional context
| 0
|
862
| 2,594,241,063
|
IssuesEvent
|
2015-02-20 01:02:03
|
BALL-Project/ball
|
https://api.github.com/repos/BALL-Project/ball
|
closed
|
BALLView on Windows does not move light sources correctly
|
C: VIEW P: major R: fixed T: defect
|
**Reported by akdehof on 26 Aug 39272847 02:13 UTC**
In the 1.3-beta1 release on windows, light sources are not correctly adapted when moving the camera around.
|
1.0
|
BALLView on Windows does not move light sources correctly - **Reported by akdehof on 26 Aug 39272847 02:13 UTC**
In the 1.3-beta1 release on windows, light sources are not correctly adapted when moving the camera around.
|
defect
|
ballview on windows does not move light sources correctly reported by akdehof on aug utc in the release on windows light sources are not correctly adapted when moving the camera around
| 1
|
53,916
| 13,262,511,048
|
IssuesEvent
|
2020-08-20 21:57:10
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
[PROPOSAL] Don't auto-generate tables. (Trac #2329)
|
Migrated from Trac combo simulation defect
|
When running a test in sim-services (propagator_state_storage.py) that test takes an unusually long time to finish. I strongly suspect PROPOSAL table generation is the culprit. This test runs for over 10 minutes on my machine.
Attached is the output.
PROPOSAL should not ever try to auto-generate tables. Demand that the user point to the tables they want to use and throw immediately if it doesn't like what it sees (e.g. empty tables).
This is causing problems for me and it's still causing problems for production.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2329">https://code.icecube.wisc.edu/projects/icecube/ticket/2329</a>, reported by olivasand owned by jsoedingrekso</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-06-28T14:57:28",
"_ts": "1561733848005867",
"description": "When running a test in sim-services (propagator_state_storage.py) that test takes an unusually long time to finish. I strongly suspect PROPOSAL table generation is the culprit. This test runs for over 10 minutes on my machine.\n\nAttached is the output.\n\nPROPOSAL should not ever try to auto-generate tables. Demand that the user point to the tables they want to use and throw immediately if it doesn't like what it sees (e.g. empty tables).\n\nThis is causing problems for me and it's still causing problems for production.\n",
"reporter": "olivas",
"cc": "juancarlos",
"resolution": "fixed",
"time": "2019-06-20T14:05:00",
"component": "combo simulation",
"summary": "[PROPOSAL] Don't auto-generate tables.",
"priority": "blocker",
"keywords": "",
"milestone": "Autumnal Equinox 2019",
"owner": "jsoedingrekso",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[PROPOSAL] Don't auto-generate tables. (Trac #2329) - When running a test in sim-services (propagator_state_storage.py) that test takes an unusually long time to finish. I strongly suspect PROPOSAL table generation is the culprit. This test runs for over 10 minutes on my machine.
Attached is the output.
PROPOSAL should not ever try to auto-generate tables. Demand that the user point to the tables they want to use and throw immediately if it doesn't like what it sees (e.g. empty tables).
This is causing problems for me and it's still causing problems for production.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2329">https://code.icecube.wisc.edu/projects/icecube/ticket/2329</a>, reported by olivasand owned by jsoedingrekso</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-06-28T14:57:28",
"_ts": "1561733848005867",
"description": "When running a test in sim-services (propagator_state_storage.py) that test takes an unusually long time to finish. I strongly suspect PROPOSAL table generation is the culprit. This test runs for over 10 minutes on my machine.\n\nAttached is the output.\n\nPROPOSAL should not ever try to auto-generate tables. Demand that the user point to the tables they want to use and throw immediately if it doesn't like what it sees (e.g. empty tables).\n\nThis is causing problems for me and it's still causing problems for production.\n",
"reporter": "olivas",
"cc": "juancarlos",
"resolution": "fixed",
"time": "2019-06-20T14:05:00",
"component": "combo simulation",
"summary": "[PROPOSAL] Don't auto-generate tables.",
"priority": "blocker",
"keywords": "",
"milestone": "Autumnal Equinox 2019",
"owner": "jsoedingrekso",
"type": "defect"
}
```
</p>
</details>
|
defect
|
don t auto generate tables trac when running a test in sim services propagator state storage py that test takes an unusually long time to finish i strongly suspect proposal table generation is the culprit this test runs for over minutes on my machine attached is the output proposal should not ever try to auto generate tables demand that the user point to the tables they want to use and throw immediately if it doesn t like what it sees e g empty tables this is causing problems for me and it s still causing problems for production migrated from json status closed changetime ts description when running a test in sim services propagator state storage py that test takes an unusually long time to finish i strongly suspect proposal table generation is the culprit this test runs for over minutes on my machine n nattached is the output n nproposal should not ever try to auto generate tables demand that the user point to the tables they want to use and throw immediately if it doesn t like what it sees e g empty tables n nthis is causing problems for me and it s still causing problems for production n reporter olivas cc juancarlos resolution fixed time component combo simulation summary don t auto generate tables priority blocker keywords milestone autumnal equinox owner jsoedingrekso type defect
| 1
|
32,666
| 7,569,910,129
|
IssuesEvent
|
2018-04-23 07:10:17
|
pywbem/pywbem
|
https://api.github.com/repos/pywbem/pywbem
|
closed
|
Deprecate tocimobj()
|
area: code release: optional resolution: fixed type: cleanup
|
cim_obj.py contains this TODO:
```
W:5450, 0: TODO: Move remaining internal uses of tocimobj() to cimvalue() and deprecate (fixme)
```
`cimvalue()` is a new function added in 0.12 that does pretty much what `tocimobj()` does, but with a cleaner approach to handling the different types. Because `tocimobj()` is an external functions and there are several subtle differences in type handling, a new function was created instead of risking an incompatibility for users.
The `tocimobj()` function should be deprecated, and all remaining internal uses of it should migrate to using `cimvalue()`.
|
1.0
|
Deprecate tocimobj() - cim_obj.py contains this TODO:
```
W:5450, 0: TODO: Move remaining internal uses of tocimobj() to cimvalue() and deprecate (fixme)
```
`cimvalue()` is a new function added in 0.12 that does pretty much what `tocimobj()` does, but with a cleaner approach to handling the different types. Because `tocimobj()` is an external functions and there are several subtle differences in type handling, a new function was created instead of risking an incompatibility for users.
The `tocimobj()` function should be deprecated, and all remaining internal uses of it should migrate to using `cimvalue()`.
|
non_defect
|
deprecate tocimobj cim obj py contains this todo w todo move remaining internal uses of tocimobj to cimvalue and deprecate fixme cimvalue is a new function added in that does pretty much what tocimobj does but with a cleaner approach to handling the different types because tocimobj is an external functions and there are several subtle differences in type handling a new function was created instead of risking an incompatibility for users the tocimobj function should be deprecated and all remaining internal uses of it should migrate to using cimvalue
| 0
|
7,012
| 2,610,321,952
|
IssuesEvent
|
2015-02-26 19:43:50
|
chrsmith/republic-at-war
|
https://api.github.com/repos/chrsmith/republic-at-war
|
closed
|
Gameplay Error
|
auto-migrated Priority-Medium Type-Defect
|
```
I believe there are different observations already made by other testers, some
said it works, some said it not works.
For me, if i activate this ability of the Light Assault Cruiser, it gives me
the Beep sound as if the area is not accesible or it is out of range etc.
However, the Cruiser responds with a standard "attack phrase". After the Beep,
nothing happens , the ability has no cooldown and no rockets seem to barrage
the area.
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 10 May 2011 at 12:49
|
1.0
|
Gameplay Error - ```
I believe there are different observations already made by other testers, some
said it works, some said it not works.
For me, if i activate this ability of the Light Assault Cruiser, it gives me
the Beep sound as if the area is not accesible or it is out of range etc.
However, the Cruiser responds with a standard "attack phrase". After the Beep,
nothing happens , the ability has no cooldown and no rockets seem to barrage
the area.
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 10 May 2011 at 12:49
|
defect
|
gameplay error i believe there are different observations already made by other testers some said it works some said it not works for me if i activate this ability of the light assault cruiser it gives me the beep sound as if the area is not accesible or it is out of range etc however the cruiser responds with a standard attack phrase after the beep nothing happens the ability has no cooldown and no rockets seem to barrage the area original issue reported on code google com by gmail com on may at
| 1
|
327,642
| 28,075,494,706
|
IssuesEvent
|
2023-03-29 23:02:00
|
ray-project/ray
|
https://api.github.com/repos/ray-project/ray
|
closed
|
[Release][Data] Migrate `dataset_shuffle_random_shuffle_1tb` test to v2 stack and Anyscale Jobs
|
P1 datasets release-test ray-team-created
|
### What happened + What you expected to happen
Migrate `dataset_shuffle_random_shuffle_1tb` test to v2 stack and Anyscale Jobs. If migration is impossible, consider rewriting the test or removing it. If there are features missing in v2 stack or/and Anyscale Jobs that make the migration impossible, please outline them in this issue.
### Versions / Dependencies
master
### Reproduction script
-
### Issue Severity
None
|
1.0
|
[Release][Data] Migrate `dataset_shuffle_random_shuffle_1tb` test to v2 stack and Anyscale Jobs - ### What happened + What you expected to happen
Migrate `dataset_shuffle_random_shuffle_1tb` test to v2 stack and Anyscale Jobs. If migration is impossible, consider rewriting the test or removing it. If there are features missing in v2 stack or/and Anyscale Jobs that make the migration impossible, please outline them in this issue.
### Versions / Dependencies
master
### Reproduction script
-
### Issue Severity
None
|
non_defect
|
migrate dataset shuffle random shuffle test to stack and anyscale jobs what happened what you expected to happen migrate dataset shuffle random shuffle test to stack and anyscale jobs if migration is impossible consider rewriting the test or removing it if there are features missing in stack or and anyscale jobs that make the migration impossible please outline them in this issue versions dependencies master reproduction script issue severity none
| 0
|
64,513
| 18,722,540,599
|
IssuesEvent
|
2021-11-03 13:22:38
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
DataTable: after filtering and changing value during ajax old filtered data is still shown
|
defect
|
**Describe the defect**
When you filter a Datatable and updates the value using ajax, the filtered old value will be still applied even if you set the filteredValue to null
**Reproducer**
https://github.com/neXus1987/primefaces-datatable-test.git
**Environment:**
- PF Version: _10.0_
- JSF + version: Mojarra 2.2.20
- Affected browsers: FireFox
**To Reproduce**
Filter for: Data 1
click on button toggle
**Expected behavior**
no records should been shown, because the filter do not match the new data (Data 2 *)
**Example XHTML**
```html
<!DOCTYPE html>
<html xmlns="http://www.w3.org/1999/xhtml"
xmlns:h="http://java.sun.com/jsf/html"
xmlns:f="http://java.sun.com/jsf/core"
xmlns:ui="http://java.sun.com/jsf/facelets"
xmlns:p="http://primefaces.org/ui" lang="de">
<h:head>
</h:head>
<h:body>
<h:form>
<p:dataTable value="#{kundenbereichModel.listStringData}" var="d">
<p:column filterBy="#{d}" filterMatchMode="contains">#{d}</p:column>
</p:dataTable>
<p:commandLink update="@form" process="@this"
actionListener="#{kundebereichTestController.toggleList()}">toggle</p:commandLink>
</h:form>
</h:body>
</html>
```
**Example Bean**
```java
@ManagedBean
@RequestScoped
public class KundebereichTestController {
private final static Log logger = LogFactory
.getLog(KundebereichTestController.class);
@ManagedProperty("#{kundenbereichModel}")
private KundenbereichModel kundenbereichModel;
public void setKundenbereichModel(KundenbereichModel kundenbereichModel) {
this.kundenbereichModel = kundenbereichModel;
}
private List<String> getData1() {
List<String> t = new ArrayList<>();
t.add("Data 1 1");
t.add("Data 1 2");
t.add("Data 1 3");
t.add("Data 1 4");
return t;
}
private List<String> getData2() {
List<String> t = new ArrayList<>();
t.add("Data 2 1");
t.add("Data 2 2");
t.add("Data 2 3");
t.add("Data 2 4");
return t;
}
public void toggleList() {
if (kundenbereichModel.getViewState() == 0) {
kundenbereichModel.setListStringData(getData1());
kundenbereichModel.setViewState(1);
} else {
kundenbereichModel.setListStringData(getData2());
kundenbereichModel.setViewState(0);
}
}
}
@ManagedBean
@ViewScoped
public class KundenbereichModel implements Serializable, {
private List<String> listStringData;
public List<String> getListStringData() {
return listStringData;
}
public void setListStringData(List<String> listStringData) {
this.listStringData = listStringData;
}
private int viewState;
public int getViewState() {
return viewState;
}
public void setViewState(int viewState) {
this.viewState = viewState;
}
}
```
|
1.0
|
DataTable: after filtering and changing value during ajax old filtered data is still shown - **Describe the defect**
When you filter a Datatable and updates the value using ajax, the filtered old value will be still applied even if you set the filteredValue to null
**Reproducer**
https://github.com/neXus1987/primefaces-datatable-test.git
**Environment:**
- PF Version: _10.0_
- JSF + version: Mojarra 2.2.20
- Affected browsers: FireFox
**To Reproduce**
Filter for: Data 1
click on button toggle
**Expected behavior**
no records should been shown, because the filter do not match the new data (Data 2 *)
**Example XHTML**
```html
<!DOCTYPE html>
<html xmlns="http://www.w3.org/1999/xhtml"
xmlns:h="http://java.sun.com/jsf/html"
xmlns:f="http://java.sun.com/jsf/core"
xmlns:ui="http://java.sun.com/jsf/facelets"
xmlns:p="http://primefaces.org/ui" lang="de">
<h:head>
</h:head>
<h:body>
<h:form>
<p:dataTable value="#{kundenbereichModel.listStringData}" var="d">
<p:column filterBy="#{d}" filterMatchMode="contains">#{d}</p:column>
</p:dataTable>
<p:commandLink update="@form" process="@this"
actionListener="#{kundebereichTestController.toggleList()}">toggle</p:commandLink>
</h:form>
</h:body>
</html>
```
**Example Bean**
```java
@ManagedBean
@RequestScoped
public class KundebereichTestController {
private final static Log logger = LogFactory
.getLog(KundebereichTestController.class);
@ManagedProperty("#{kundenbereichModel}")
private KundenbereichModel kundenbereichModel;
public void setKundenbereichModel(KundenbereichModel kundenbereichModel) {
this.kundenbereichModel = kundenbereichModel;
}
private List<String> getData1() {
List<String> t = new ArrayList<>();
t.add("Data 1 1");
t.add("Data 1 2");
t.add("Data 1 3");
t.add("Data 1 4");
return t;
}
private List<String> getData2() {
List<String> t = new ArrayList<>();
t.add("Data 2 1");
t.add("Data 2 2");
t.add("Data 2 3");
t.add("Data 2 4");
return t;
}
public void toggleList() {
if (kundenbereichModel.getViewState() == 0) {
kundenbereichModel.setListStringData(getData1());
kundenbereichModel.setViewState(1);
} else {
kundenbereichModel.setListStringData(getData2());
kundenbereichModel.setViewState(0);
}
}
}
@ManagedBean
@ViewScoped
public class KundenbereichModel implements Serializable, {
private List<String> listStringData;
public List<String> getListStringData() {
return listStringData;
}
public void setListStringData(List<String> listStringData) {
this.listStringData = listStringData;
}
private int viewState;
public int getViewState() {
return viewState;
}
public void setViewState(int viewState) {
this.viewState = viewState;
}
}
```
|
defect
|
datatable after filtering and changing value during ajax old filtered data is still shown describe the defect when you filter a datatable and updates the value using ajax the filtered old value will be still applied even if you set the filteredvalue to null reproducer environment pf version jsf version mojarra affected browsers firefox to reproduce filter for data click on button toggle expected behavior no records should been shown because the filter do not match the new data data example xhtml html html xmlns xmlns h xmlns f xmlns ui xmlns p lang de d p commandlink update form process this actionlistener kundebereichtestcontroller togglelist toggle example bean java managedbean requestscoped public class kundebereichtestcontroller private final static log logger logfactory getlog kundebereichtestcontroller class managedproperty kundenbereichmodel private kundenbereichmodel kundenbereichmodel public void setkundenbereichmodel kundenbereichmodel kundenbereichmodel this kundenbereichmodel kundenbereichmodel private list list t new arraylist t add data t add data t add data t add data return t private list list t new arraylist t add data t add data t add data t add data return t public void togglelist if kundenbereichmodel getviewstate kundenbereichmodel setliststringdata kundenbereichmodel setviewstate else kundenbereichmodel setliststringdata kundenbereichmodel setviewstate managedbean viewscoped public class kundenbereichmodel implements serializable private list liststringdata public list getliststringdata return liststringdata public void setliststringdata list liststringdata this liststringdata liststringdata private int viewstate public int getviewstate return viewstate public void setviewstate int viewstate this viewstate viewstate
| 1
|
76,728
| 15,496,181,874
|
IssuesEvent
|
2021-03-11 02:12:40
|
mwilliams7197/zendo
|
https://api.github.com/repos/mwilliams7197/zendo
|
closed
|
WS-2019-0032 (Medium) detected in js-yaml-3.7.0.tgz - autoclosed
|
security vulnerability
|
## WS-2019-0032 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p>
<p>Path to dependency file: zendo/package.json</p>
<p>Path to vulnerable library: zendo/node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.7.tgz (Root Library)
- eslint-3.19.0.tgz
- :x: **js-yaml-3.7.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mwilliams7197/zendo/commit/cf147eff03fa219eea73721a78a106371bf2f0bc">cf147eff03fa219eea73721a78a106371bf2f0bc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service.
<p>Publish Date: 2019-03-20
<p>URL: <a href=https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235>WS-2019-0032</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/788/versions">https://www.npmjs.com/advisories/788/versions</a></p>
<p>Release Date: 2019-03-20</p>
<p>Fix Resolution: js-yaml - 3.13.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"js-yaml","packageVersion":"3.7.0","isTransitiveDependency":true,"dependencyTree":"react-scripts:1.0.7;eslint:3.19.0;js-yaml:3.7.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"js-yaml - 3.13.0"}],"vulnerabilityIdentifier":"WS-2019-0032","vulnerabilityDetails":"Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service.","vulnerabilityUrl":"https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
|
True
|
WS-2019-0032 (Medium) detected in js-yaml-3.7.0.tgz - autoclosed - ## WS-2019-0032 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p>
<p>Path to dependency file: zendo/package.json</p>
<p>Path to vulnerable library: zendo/node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.7.tgz (Root Library)
- eslint-3.19.0.tgz
- :x: **js-yaml-3.7.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mwilliams7197/zendo/commit/cf147eff03fa219eea73721a78a106371bf2f0bc">cf147eff03fa219eea73721a78a106371bf2f0bc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service.
<p>Publish Date: 2019-03-20
<p>URL: <a href=https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235>WS-2019-0032</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/788/versions">https://www.npmjs.com/advisories/788/versions</a></p>
<p>Release Date: 2019-03-20</p>
<p>Fix Resolution: js-yaml - 3.13.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"js-yaml","packageVersion":"3.7.0","isTransitiveDependency":true,"dependencyTree":"react-scripts:1.0.7;eslint:3.19.0;js-yaml:3.7.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"js-yaml - 3.13.0"}],"vulnerabilityIdentifier":"WS-2019-0032","vulnerabilityDetails":"Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service.","vulnerabilityUrl":"https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
|
non_defect
|
ws medium detected in js yaml tgz autoclosed ws medium severity vulnerability vulnerable library js yaml tgz yaml parser and serializer library home page a href path to dependency file zendo package json path to vulnerable library zendo node modules js yaml package json dependency hierarchy react scripts tgz root library eslint tgz x js yaml tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions js yaml prior to are vulnerable to denial of service by parsing a carefully crafted yaml file the node process stalls and may exhaust system resources leading to a denial of service publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution js yaml isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails versions js yaml prior to are vulnerable to denial of service by parsing a carefully crafted yaml file the node process stalls and may exhaust system resources leading to a denial of service vulnerabilityurl
| 0
|
64,871
| 18,949,777,713
|
IssuesEvent
|
2021-11-18 14:06:56
|
vector-im/element-android
|
https://api.github.com/repos/vector-im/element-android
|
closed
|
Push notifications sometimes disappear immediately
|
T-Defect A-Notifications
|
In a 1:1 encrypted chat, I've noticed that often (but not always, maybe 2/3 of the time), notifications disappear shortly after they appear and my phone vibrates. When this happens, the dot signifying new notifications on the element app icon also disappears.
I'm currently using version 1.2.1 on a Pixel 5 running Android 11, but I've noticed this issue for a few months on both this device and a Pixel 3, also running Android 11.
[Here's a video](https://streamable.com/focnm4)
|
1.0
|
Push notifications sometimes disappear immediately - In a 1:1 encrypted chat, I've noticed that often (but not always, maybe 2/3 of the time), notifications disappear shortly after they appear and my phone vibrates. When this happens, the dot signifying new notifications on the element app icon also disappears.
I'm currently using version 1.2.1 on a Pixel 5 running Android 11, but I've noticed this issue for a few months on both this device and a Pixel 3, also running Android 11.
[Here's a video](https://streamable.com/focnm4)
|
defect
|
push notifications sometimes disappear immediately in a encrypted chat i ve noticed that often but not always maybe of the time notifications disappear shortly after they appear and my phone vibrates when this happens the dot signifying new notifications on the element app icon also disappears i m currently using version on a pixel running android but i ve noticed this issue for a few months on both this device and a pixel also running android
| 1
|
3,448
| 2,610,062,965
|
IssuesEvent
|
2015-02-26 18:18:30
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
黄岩治不育一般需要多少钱
|
auto-migrated Priority-Medium Type-Defect
|
```
黄岩治不育一般需要多少钱【台州五洲生殖医院】24小时健康
咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台
州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、1
08、118、198及椒江一金清公交车直达枫南小区,乘坐107、105、
109、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:45
|
1.0
|
黄岩治不育一般需要多少钱 - ```
黄岩治不育一般需要多少钱【台州五洲生殖医院】24小时健康
咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台
州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、1
08、118、198及椒江一金清公交车直达枫南小区,乘坐107、105、
109、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:45
|
defect
|
黄岩治不育一般需要多少钱 黄岩治不育一般需要多少钱【台州五洲生殖医院】 咨询热线 微信号tzwzszyy 医院地址 台 (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 1
|
50,617
| 13,187,626,198
|
IssuesEvent
|
2020-08-13 04:02:01
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
muex option 'detail' is broken in trunk r135229 (Trac #1055)
|
Migrated from Trac combo reconstruction defect
|
If the detail option of the muex module is set to True it will seg fault.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1055">https://code.icecube.wisc.edu/ticket/1055</a>, reported by sflis and owned by dima</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "If the detail option of the muex module is set to True it will seg fault.",
"reporter": "sflis",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "muex option 'detail' is broken in trunk r135229",
"priority": "blocker",
"keywords": "",
"time": "2015-07-16T09:47:18",
"milestone": "",
"owner": "dima",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
muex option 'detail' is broken in trunk r135229 (Trac #1055) - If the detail option of the muex module is set to True it will seg fault.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1055">https://code.icecube.wisc.edu/ticket/1055</a>, reported by sflis and owned by dima</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "If the detail option of the muex module is set to True it will seg fault.",
"reporter": "sflis",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "muex option 'detail' is broken in trunk r135229",
"priority": "blocker",
"keywords": "",
"time": "2015-07-16T09:47:18",
"milestone": "",
"owner": "dima",
"type": "defect"
}
```
</p>
</details>
|
defect
|
muex option detail is broken in trunk trac if the detail option of the muex module is set to true it will seg fault migrated from json status closed changetime description if the detail option of the muex module is set to true it will seg fault reporter sflis cc resolution fixed ts component combo reconstruction summary muex option detail is broken in trunk priority blocker keywords time milestone owner dima type defect
| 1
|
26,213
| 4,622,767,866
|
IssuesEvent
|
2016-09-27 08:46:29
|
siddhartha-gadgil/ProvingGround
|
https://api.github.com/repos/siddhartha-gadgil/ProvingGround
|
closed
|
Very poor performance due to cascade of replacements.
|
defect
|
In generating from _Monoids_, but not from just _logic_, there is a huge slowdown, with most of the time spent in replace methods. This appears to be because:
* for the sake of safety, lambdas:
* create an inner variable
* in the case of lambda-fixed, check independence.
* both these involve replacements for lambdas
* if the replaced values are `IdentityTyp` etc., then the terms themselves contain a lot of lambdas, so we get a cascade of replacements.
* We can avoid creating inner variables and checking for dependence if we are using a new synthesized variable (and are careful about dependence).
|
1.0
|
Very poor performance due to cascade of replacements. - In generating from _Monoids_, but not from just _logic_, there is a huge slowdown, with most of the time spent in replace methods. This appears to be because:
* for the sake of safety, lambdas:
* create an inner variable
* in the case of lambda-fixed, check independence.
* both these involve replacements for lambdas
* if the replaced values are `IdentityTyp` etc., then the terms themselves contain a lot of lambdas, so we get a cascade of replacements.
* We can avoid creating inner variables and checking for dependence if we are using a new synthesized variable (and are careful about dependence).
|
defect
|
very poor performance due to cascade of replacements in generating from monoids but not from just logic there is a huge slowdown with most of the time spent in replace methods this appears to be because for the sake of safety lambdas create an inner variable in the case of lambda fixed check independence both these involve replacements for lambdas if the replaced values are identitytyp etc then the terms themselves contain a lot of lambdas so we get a cascade of replacements we can avoid creating inner variables and checking for dependence if we are using a new synthesized variable and are careful about dependence
| 1
|
18,629
| 3,077,631,643
|
IssuesEvent
|
2015-08-21 02:27:04
|
martingkelly/imms
|
https://api.github.com/repos/martingkelly/imms
|
closed
|
imms takes a while to record played songs in its database
|
auto-migrated Priority-Medium Type-Defect
|
```
If you finish playing a song, imms does not immediately record it in its
database. I spent a while debugging this and tracked it to the buffering
behavior of the GIOChannel used on the immsd server. Basically, GIOChannel has
an internal buffer, and data doesn't get sent until that buffer fills up. The
size of the buffer can vary across platforms, and thus the time until a song is
recorded can vary a lot too. To add to this, if audacious exits before the
buffer fills up, the song data from that session is completely lost :(.
I wrote a patch to fix this behavior by disabling buffering in the GIOChannel.
With this patch, song data is recorded immediately. Given that the amount of
data being sent between the audacious client and the immsd server is relatively
low, I believe it's OK to skip buffering.
My patch can be found below (based on John Lindgren's fixes for Audacious 3.6).
https://github.com/martingkelly/imms/commit/c3f99566fe98459f0c7436804039190c4101
cfd9
```
Original issue reported on code.google.com by `mar...@martingkelly.com` on 12 Aug 2015 at 3:32
|
1.0
|
imms takes a while to record played songs in its database - ```
If you finish playing a song, imms does not immediately record it in its
database. I spent a while debugging this and tracked it to the buffering
behavior of the GIOChannel used on the immsd server. Basically, GIOChannel has
an internal buffer, and data doesn't get sent until that buffer fills up. The
size of the buffer can vary across platforms, and thus the time until a song is
recorded can vary a lot too. To add to this, if audacious exits before the
buffer fills up, the song data from that session is completely lost :(.
I wrote a patch to fix this behavior by disabling buffering in the GIOChannel.
With this patch, song data is recorded immediately. Given that the amount of
data being sent between the audacious client and the immsd server is relatively
low, I believe it's OK to skip buffering.
My patch can be found below (based on John Lindgren's fixes for Audacious 3.6).
https://github.com/martingkelly/imms/commit/c3f99566fe98459f0c7436804039190c4101
cfd9
```
Original issue reported on code.google.com by `mar...@martingkelly.com` on 12 Aug 2015 at 3:32
|
defect
|
imms takes a while to record played songs in its database if you finish playing a song imms does not immediately record it in its database i spent a while debugging this and tracked it to the buffering behavior of the giochannel used on the immsd server basically giochannel has an internal buffer and data doesn t get sent until that buffer fills up the size of the buffer can vary across platforms and thus the time until a song is recorded can vary a lot too to add to this if audacious exits before the buffer fills up the song data from that session is completely lost i wrote a patch to fix this behavior by disabling buffering in the giochannel with this patch song data is recorded immediately given that the amount of data being sent between the audacious client and the immsd server is relatively low i believe it s ok to skip buffering my patch can be found below based on john lindgren s fixes for audacious original issue reported on code google com by mar martingkelly com on aug at
| 1
|
41,668
| 10,563,396,765
|
IssuesEvent
|
2019-10-04 20:51:37
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
[COGNITION]: Recommend showing all secondary specialties instead of show more/show less button
|
508-defect-1 508/Accessibility facility locator frontend vsa-global-ux
|
## Issue
The facility detail views sometimes have second-level lists that include a show more button. These buttons prepend additional `<li>` before the button, which causes an issue for assistive device users. This practice was flagged as an SC 1.3.1 issue. Screenshot attached below.
## Audit Finding
* Note 1, Defect 3 of 3
* Note 5, Defect 2 of 2
## Related Issues
* https://app.zenhub.com/workspaces/vsp-5cedc9cce6e3335dc5a49fc4/issues/department-of-veterans-affairs/va.gov-team/484
## Acceptance Criteria
* As an assistive device user, I want to hear the nested list items all read out, or understand that list items have been added farther down in the DOM, after the show more button.
## Environment
* MacOS Mojave
* Chrome latest
* https://staging.va.gov/find-locations/facility/vha_674GD
## Possible Fixes (optional)
The simplest way to remediate this error is to remove the show more button and return all of the nested list items on first render. The other option is to remove the button from the list, and create a new list below the button. This feels less useful from a user flow, but could be an option if needed.
## WCAG or Vendor Guidance (optional)
* [Info and Relationships: Understanding SC 1.3.1](https://www.w3.org/TR/UNDERSTANDING-WCAG20/content-structure-separation-programmatic.html)
## Screenshots or Trace Logs

|
1.0
|
[COGNITION]: Recommend showing all secondary specialties instead of show more/show less button - ## Issue
The facility detail views sometimes have second-level lists that include a show more button. These buttons prepend additional `<li>` before the button, which causes an issue for assistive device users. This practice was flagged as an SC 1.3.1 issue. Screenshot attached below.
## Audit Finding
* Note 1, Defect 3 of 3
* Note 5, Defect 2 of 2
## Related Issues
* https://app.zenhub.com/workspaces/vsp-5cedc9cce6e3335dc5a49fc4/issues/department-of-veterans-affairs/va.gov-team/484
## Acceptance Criteria
* As an assistive device user, I want to hear the nested list items all read out, or understand that list items have been added farther down in the DOM, after the show more button.
## Environment
* MacOS Mojave
* Chrome latest
* https://staging.va.gov/find-locations/facility/vha_674GD
## Possible Fixes (optional)
The simplest way to remediate this error is to remove the show more button and return all of the nested list items on first render. The other option is to remove the button from the list, and create a new list below the button. This feels less useful from a user flow, but could be an option if needed.
## WCAG or Vendor Guidance (optional)
* [Info and Relationships: Understanding SC 1.3.1](https://www.w3.org/TR/UNDERSTANDING-WCAG20/content-structure-separation-programmatic.html)
## Screenshots or Trace Logs

|
defect
|
recommend showing all secondary specialties instead of show more show less button issue the facility detail views sometimes have second level lists that include a show more button these buttons prepend additional before the button which causes an issue for assistive device users this practice was flagged as an sc issue screenshot attached below audit finding note defect of note defect of related issues acceptance criteria as an assistive device user i want to hear the nested list items all read out or understand that list items have been added farther down in the dom after the show more button environment macos mojave chrome latest possible fixes optional the simplest way to remediate this error is to remove the show more button and return all of the nested list items on first render the other option is to remove the button from the list and create a new list below the button this feels less useful from a user flow but could be an option if needed wcag or vendor guidance optional screenshots or trace logs
| 1
|
34,481
| 7,452,016,728
|
IssuesEvent
|
2018-03-29 06:38:26
|
kerdokullamae/test_koik_issued
|
https://api.github.com/repos/kerdokullamae/test_koik_issued
|
closed
|
Seostamine: Org/Isiku seostamine teise Org/Isikuga ning Kirjeldusüksusega
|
P: highest R: fixed T: defect
|
**Reported by maiu pevkur on 9 May 2013 13:13 UTC**
Organisatsiooni ja Isiku lisainfo vormidel on sakid Seotud kirjeldusüksused ja Seotud isikud/org, kus on vanast AISist kanud andmed (osaliselt need andmed küll veel puuduvad). Puudub aga võimalus lisada/kustutada seotud kirjeldusüksusi ja seotud isikuid/orge. Kui vajutada Muuda, mis avab andmete muutmise vormi (nii Org-idele kui Isikutele), siis seal sellist võimalust ei paista.
Samuti oleks hea, kui saaks ka vormil Kirjeldusüksuse andmed (ehk andmete muutmise vormil) lisada seotud isikut/organisatsiooni ja ka vastavat seost kustutada. St et seostamine toimiks ka vastupidi. Oleks võimalik seoseid lisada/kustutada nii Org/Isiku kui Kirjeldusüksuse andmete muutmise vormidel.
|
1.0
|
Seostamine: Org/Isiku seostamine teise Org/Isikuga ning Kirjeldusüksusega - **Reported by maiu pevkur on 9 May 2013 13:13 UTC**
Organisatsiooni ja Isiku lisainfo vormidel on sakid Seotud kirjeldusüksused ja Seotud isikud/org, kus on vanast AISist kanud andmed (osaliselt need andmed küll veel puuduvad). Puudub aga võimalus lisada/kustutada seotud kirjeldusüksusi ja seotud isikuid/orge. Kui vajutada Muuda, mis avab andmete muutmise vormi (nii Org-idele kui Isikutele), siis seal sellist võimalust ei paista.
Samuti oleks hea, kui saaks ka vormil Kirjeldusüksuse andmed (ehk andmete muutmise vormil) lisada seotud isikut/organisatsiooni ja ka vastavat seost kustutada. St et seostamine toimiks ka vastupidi. Oleks võimalik seoseid lisada/kustutada nii Org/Isiku kui Kirjeldusüksuse andmete muutmise vormidel.
|
defect
|
seostamine org isiku seostamine teise org isikuga ning kirjeldusüksusega reported by maiu pevkur on may utc organisatsiooni ja isiku lisainfo vormidel on sakid seotud kirjeldusüksused ja seotud isikud org kus on vanast aisist kanud andmed osaliselt need andmed küll veel puuduvad puudub aga võimalus lisada kustutada seotud kirjeldusüksusi ja seotud isikuid orge kui vajutada muuda mis avab andmete muutmise vormi nii org idele kui isikutele siis seal sellist võimalust ei paista samuti oleks hea kui saaks ka vormil kirjeldusüksuse andmed ehk andmete muutmise vormil lisada seotud isikut organisatsiooni ja ka vastavat seost kustutada st et seostamine toimiks ka vastupidi oleks võimalik seoseid lisada kustutada nii org isiku kui kirjeldusüksuse andmete muutmise vormidel
| 1
|
48,126
| 13,067,466,136
|
IssuesEvent
|
2020-07-31 00:32:42
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
L1 filter for 2014 and 2015 (Trac #1828)
|
Migrated from Trac cmake defect
|
Running offline filter for 2014 and 2015:
/cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/metaprojects/icerec/IC2014-L2_V14-02-00/lib/icecube/filterscripts/offlineL2/level1_SimulationFiltering.py
I receive the following error:
`RuntimeError: dlopen() dynamic loading error: /data/user/saxani/environments/buildfwd/lib/libpfauxiliary.so: cannot open shared object file: No such file or directory`
Migrated from https://code.icecube.wisc.edu/ticket/1828
```json
{
"status": "closed",
"changetime": "2016-08-18T21:39:38",
"description": "Running offline filter for 2014 and 2015:\n/cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/metaprojects/icerec/IC2014-L2_V14-02-00/lib/icecube/filterscripts/offlineL2/level1_SimulationFiltering.py\n\nI receive the following error:\n`RuntimeError: dlopen() dynamic loading error: /data/user/saxani/environments/buildfwd/lib/libpfauxiliary.so: cannot open shared object file: No such file or directory`\n",
"reporter": "saxani",
"cc": "",
"resolution": "fixed",
"_ts": "1471556378950850",
"component": "cmake",
"summary": "L1 filter for 2014 and 2015",
"priority": "normal",
"keywords": "",
"time": "2016-08-18T21:29:33",
"milestone": "",
"owner": "",
"type": "defect"
}
```
|
1.0
|
L1 filter for 2014 and 2015 (Trac #1828) - Running offline filter for 2014 and 2015:
/cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/metaprojects/icerec/IC2014-L2_V14-02-00/lib/icecube/filterscripts/offlineL2/level1_SimulationFiltering.py
I receive the following error:
`RuntimeError: dlopen() dynamic loading error: /data/user/saxani/environments/buildfwd/lib/libpfauxiliary.so: cannot open shared object file: No such file or directory`
Migrated from https://code.icecube.wisc.edu/ticket/1828
```json
{
"status": "closed",
"changetime": "2016-08-18T21:39:38",
"description": "Running offline filter for 2014 and 2015:\n/cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/metaprojects/icerec/IC2014-L2_V14-02-00/lib/icecube/filterscripts/offlineL2/level1_SimulationFiltering.py\n\nI receive the following error:\n`RuntimeError: dlopen() dynamic loading error: /data/user/saxani/environments/buildfwd/lib/libpfauxiliary.so: cannot open shared object file: No such file or directory`\n",
"reporter": "saxani",
"cc": "",
"resolution": "fixed",
"_ts": "1471556378950850",
"component": "cmake",
"summary": "L1 filter for 2014 and 2015",
"priority": "normal",
"keywords": "",
"time": "2016-08-18T21:29:33",
"milestone": "",
"owner": "",
"type": "defect"
}
```
|
defect
|
filter for and trac running offline filter for and cvmfs icecube opensciencegrid org rhel metaprojects icerec lib icecube filterscripts simulationfiltering py i receive the following error runtimeerror dlopen dynamic loading error data user saxani environments buildfwd lib libpfauxiliary so cannot open shared object file no such file or directory migrated from json status closed changetime description running offline filter for and n cvmfs icecube opensciencegrid org rhel metaprojects icerec lib icecube filterscripts simulationfiltering py n ni receive the following error n runtimeerror dlopen dynamic loading error data user saxani environments buildfwd lib libpfauxiliary so cannot open shared object file no such file or directory n reporter saxani cc resolution fixed ts component cmake summary filter for and priority normal keywords time milestone owner type defect
| 1
|
47,769
| 19,716,000,457
|
IssuesEvent
|
2022-01-13 11:02:10
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
How should the AZ CLI script for importing the Helm images be run
|
container-service/svc triaged cxp needs-more-info product-issue Pri1
|
I tried running those commands in PowerShell and get errors all over the place. I then figured I'd put them into a powershell script file and run that. Still no dice. What am I missing?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 1bf9c2f3-caeb-0fb1-f089-3f7ca4fad729
* Version Independent ID: 929253d7-c763-6429-561f-529d39b5284d
* Content: [Ingress controller on internal network - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/ingress-internal-ip)
* Content Source: [articles/aks/ingress-internal-ip.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/ingress-internal-ip.md)
* Service: **container-service**
* GitHub Login: @zr-msft
* Microsoft Alias: **zarhoads**
|
1.0
|
How should the AZ CLI script for importing the Helm images be run - I tried running those commands in PowerShell and get errors all over the place. I then figured I'd put them into a powershell script file and run that. Still no dice. What am I missing?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 1bf9c2f3-caeb-0fb1-f089-3f7ca4fad729
* Version Independent ID: 929253d7-c763-6429-561f-529d39b5284d
* Content: [Ingress controller on internal network - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/ingress-internal-ip)
* Content Source: [articles/aks/ingress-internal-ip.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/ingress-internal-ip.md)
* Service: **container-service**
* GitHub Login: @zr-msft
* Microsoft Alias: **zarhoads**
|
non_defect
|
how should the az cli script for importing the helm images be run i tried running those commands in powershell and get errors all over the place i then figured i d put them into a powershell script file and run that still no dice what am i missing document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id caeb version independent id content content source service container service github login zr msft microsoft alias zarhoads
| 0
|
354,378
| 25,163,572,768
|
IssuesEvent
|
2022-11-10 18:45:13
|
netbox-community/netbox
|
https://api.github.com/repos/netbox-community/netbox
|
opened
|
'Installation & Upgrade' Netbox documentation page - Add note to 'Warning' section at bottom of the page to have user confirm virtual environment is still activated
|
type: documentation
|
### Change Type
Addition
### Area
Installation/upgrade
### Proposed Changes
Add a basic troubleshooting note to the "Warning" section at the bottom of the "3. Netbox" page in the "Installation & Upgrade" section of the documentation.
Have user ensure that the virtual environment is still activated before performing additional troubleshooting measures.
It's a basic enough step that could help newer users avoid unnecessary confusion, while still keeping the documentation concise.
Suggested addition:
"If you receive errors regarding missing modules/packages, be sure that you still have the virtual environment activated before performing any additional troubleshooting measures."
|
1.0
|
'Installation & Upgrade' Netbox documentation page - Add note to 'Warning' section at bottom of the page to have user confirm virtual environment is still activated - ### Change Type
Addition
### Area
Installation/upgrade
### Proposed Changes
Add a basic troubleshooting note to the "Warning" section at the bottom of the "3. Netbox" page in the "Installation & Upgrade" section of the documentation.
Have user ensure that the virtual environment is still activated before performing additional troubleshooting measures.
It's a basic enough step that could help newer users avoid unnecessary confusion, while still keeping the documentation concise.
Suggested addition:
"If you receive errors regarding missing modules/packages, be sure that you still have the virtual environment activated before performing any additional troubleshooting measures."
|
non_defect
|
installation upgrade netbox documentation page add note to warning section at bottom of the page to have user confirm virtual environment is still activated change type addition area installation upgrade proposed changes add a basic troubleshooting note to the warning section at the bottom of the netbox page in the installation upgrade section of the documentation have user ensure that the virtual environment is still activated before performing additional troubleshooting measures it s a basic enough step that could help newer users avoid unnecessary confusion while still keeping the documentation concise suggested addition if you receive errors regarding missing modules packages be sure that you still have the virtual environment activated before performing any additional troubleshooting measures
| 0
|
447,142
| 31,624,355,073
|
IssuesEvent
|
2023-09-06 03:21:17
|
Virtue-Digital-Indonesia/frontend-examples
|
https://api.github.com/repos/Virtue-Digital-Indonesia/frontend-examples
|
opened
|
Projects Tracker
|
documentation
|
This issue is used to list and track the progress of each project.
Here are the projects:
- [ ] Data Grid Filters and Data Grid Table
- [ ] Object List with MUI Data Grid
- [ ] Markers, Popups, and Clusters with Leaflet
- [ ] Date Range and Time Picker
- [ ] Create, show, edit, and delete Geofences with Leaflet
|
1.0
|
Projects Tracker - This issue is used to list and track the progress of each project.
Here are the projects:
- [ ] Data Grid Filters and Data Grid Table
- [ ] Object List with MUI Data Grid
- [ ] Markers, Popups, and Clusters with Leaflet
- [ ] Date Range and Time Picker
- [ ] Create, show, edit, and delete Geofences with Leaflet
|
non_defect
|
projects tracker this issue is used to list and track the progress of each project here are the projects data grid filters and data grid table object list with mui data grid markers popups and clusters with leaflet date range and time picker create show edit and delete geofences with leaflet
| 0
|
171,141
| 27,066,919,191
|
IssuesEvent
|
2023-02-14 01:36:52
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
Expose IsInExpressionTree
|
Area-IDE Concept-API Need Design Review
|
Useful in code fixes as the rules are slightly different in expressions.
Preferable via a static mehtod and not via some service locator. Don't know if `SemanticFacts` is a thing.
|
1.0
|
Expose IsInExpressionTree - Useful in code fixes as the rules are slightly different in expressions.
Preferable via a static mehtod and not via some service locator. Don't know if `SemanticFacts` is a thing.
|
non_defect
|
expose isinexpressiontree useful in code fixes as the rules are slightly different in expressions preferable via a static mehtod and not via some service locator don t know if semanticfacts is a thing
| 0
|
277,880
| 24,108,146,369
|
IssuesEvent
|
2022-09-20 09:09:21
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
Test failure: BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles
|
ci-concern bot/type/test bot/channel/nightly bot/platform/android bot/arch/x86-mono bot/branch/v1.45
|
Greetings human!
Bad news. `BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles` [failed on android x86-mono nightly v1.45.58](https://ci.brave.com/job/brave-browser-build-android-variant/5971/testReport/junit/(root)/BraveProfileManagerTest/test___test_browser___ExcludeServicesInOTRAndGuestProfiles).
<details>
<summary>Stack trace</summary>
```
[ RUN ] BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles
tracing_subscriber - init success
[INFO:mod.rs(72)] skus::sdk: skus sdk initialized, environment: production, version: 4fe49d14c2
at /home/ubuntu/workspace/brave-browser-build-android-variant-nightly/src/brave/components/skus/browser/rs/lib/src/sdk/mod.rs:72
[VERBOSE1:kv.rs(123)] skus::storage::kv: got orders, orders: None
at /home/ubuntu/workspace/brave-browser-build-android-variant-nightly/src/brave/components/skus/browser/rs/lib/src/storage/kv.rs:123
in skus::storage::kv::get_orders with self=NativeClient
in skus::sdk::credentials::matching_credential_summary with self=SDK domain="vpn.brave.com"
../../brave/browser/profiles/brave_profile_manager_browsertest.cc:175: Failure
Expected: (brave_rewards::RewardsServiceFactory::GetForProfile(profile)) != (nullptr), actual: NULL vs (nullptr)
Stack trace:
../../brave/browser/profiles/brave_profile_manager_browsertest.cc:180: Failure
Expected: (brave_ads::AdsServiceFactory::GetForProfile(profile)) != (nullptr), actual: NULL vs (nullptr)
Stack trace:
[WARNING:task_environment.cc(800)] ParallelExecutionFence is ineffective when ThreadPoolInstance is not managed by a TaskEnvironment.
Test fixtures should use a TaskEnvironment member or statically invoke TaskEnvironment::CreateThreadPool() + ThreadPoolInstance::Get()->StartWithDefaultParams() when the former is not possible.
[WARNING:task_environment.cc(800)] ParallelExecutionFence is ineffective when ThreadPoolInstance is not managed by a TaskEnvironment.
Test fixtures should use a TaskEnvironment member or statically invoke TaskEnvironment::CreateThreadPool() + ThreadPoolInstance::Get()->StartWithDefaultParams() when the former is not possible.
[ FAILED ] BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles, where TypeParam = and GetParam() = (671 ms)
[----------] 1 test from BraveProfileManagerTest (672 ms total)
[----------] Global test environment tear-down
[==========] 1 test from 1 test suite ran. (672 ms total)
[ PASSED ] 0 tests.
[ FAILED ] 1 test, listed below:
[ FAILED ] BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles, where TypeParam = and GetParam() =
Stack Trace:
RELADDR FUNCTION FILE:LINE
02ec85db BraveProfileManagerTest_ExcludeServicesInOTRAndGuestProfiles_Test::RunTestOnMainThread() ../../brave/browser/profiles/brave_profile_manager_browsertest.cc:174:3
08814c44 content::BrowserTestBase::ProxyRunTestOnMainThreadLoop() ../../content/public/test/browser_test_base.cc:909:7
0881423e content::BrowserTestBase::SetUp() ../../content/public/test/browser_test_base.cc:707:5
0517b84a AndroidBrowserTest::SetUp() ../../chrome/test/base/android/android_browser_test.cc:25:20
0509892d bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) ../../third_party/googletest/src/googletest/src/gtest.cc:0:0
-----------------------------------------------------
Stack Trace:
RELADDR FUNCTION FILE:LINE
02ec8692 BraveProfileManagerTest_ExcludeServicesInOTRAndGuestProfiles_Test::RunTestOnMainThread() ../../brave/browser/profiles/brave_profile_manager_browsertest.cc:180:3
08814c44 content::BrowserTestBase::ProxyRunTestOnMainThreadLoop() ../../content/public/test/browser_test_base.cc:909:7
0881423e content::BrowserTestBase::SetUp() ../../content/public/test/browser_test_base.cc:707:5
0517b84a AndroidBrowserTest::SetUp() ../../chrome/test/base/android/android_browser_test.cc:25:20
0509892d bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) ../../third_party/googletest/src/googletest/src/gtest.cc:0:0
failed
```
</details>
|
1.0
|
Test failure: BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles - Greetings human!
Bad news. `BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles` [failed on android x86-mono nightly v1.45.58](https://ci.brave.com/job/brave-browser-build-android-variant/5971/testReport/junit/(root)/BraveProfileManagerTest/test___test_browser___ExcludeServicesInOTRAndGuestProfiles).
<details>
<summary>Stack trace</summary>
```
[ RUN ] BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles
tracing_subscriber - init success
[INFO:mod.rs(72)] skus::sdk: skus sdk initialized, environment: production, version: 4fe49d14c2
at /home/ubuntu/workspace/brave-browser-build-android-variant-nightly/src/brave/components/skus/browser/rs/lib/src/sdk/mod.rs:72
[VERBOSE1:kv.rs(123)] skus::storage::kv: got orders, orders: None
at /home/ubuntu/workspace/brave-browser-build-android-variant-nightly/src/brave/components/skus/browser/rs/lib/src/storage/kv.rs:123
in skus::storage::kv::get_orders with self=NativeClient
in skus::sdk::credentials::matching_credential_summary with self=SDK domain="vpn.brave.com"
../../brave/browser/profiles/brave_profile_manager_browsertest.cc:175: Failure
Expected: (brave_rewards::RewardsServiceFactory::GetForProfile(profile)) != (nullptr), actual: NULL vs (nullptr)
Stack trace:
../../brave/browser/profiles/brave_profile_manager_browsertest.cc:180: Failure
Expected: (brave_ads::AdsServiceFactory::GetForProfile(profile)) != (nullptr), actual: NULL vs (nullptr)
Stack trace:
[WARNING:task_environment.cc(800)] ParallelExecutionFence is ineffective when ThreadPoolInstance is not managed by a TaskEnvironment.
Test fixtures should use a TaskEnvironment member or statically invoke TaskEnvironment::CreateThreadPool() + ThreadPoolInstance::Get()->StartWithDefaultParams() when the former is not possible.
[WARNING:task_environment.cc(800)] ParallelExecutionFence is ineffective when ThreadPoolInstance is not managed by a TaskEnvironment.
Test fixtures should use a TaskEnvironment member or statically invoke TaskEnvironment::CreateThreadPool() + ThreadPoolInstance::Get()->StartWithDefaultParams() when the former is not possible.
[ FAILED ] BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles, where TypeParam = and GetParam() = (671 ms)
[----------] 1 test from BraveProfileManagerTest (672 ms total)
[----------] Global test environment tear-down
[==========] 1 test from 1 test suite ran. (672 ms total)
[ PASSED ] 0 tests.
[ FAILED ] 1 test, listed below:
[ FAILED ] BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles, where TypeParam = and GetParam() =
Stack Trace:
RELADDR FUNCTION FILE:LINE
02ec85db BraveProfileManagerTest_ExcludeServicesInOTRAndGuestProfiles_Test::RunTestOnMainThread() ../../brave/browser/profiles/brave_profile_manager_browsertest.cc:174:3
08814c44 content::BrowserTestBase::ProxyRunTestOnMainThreadLoop() ../../content/public/test/browser_test_base.cc:909:7
0881423e content::BrowserTestBase::SetUp() ../../content/public/test/browser_test_base.cc:707:5
0517b84a AndroidBrowserTest::SetUp() ../../chrome/test/base/android/android_browser_test.cc:25:20
0509892d bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) ../../third_party/googletest/src/googletest/src/gtest.cc:0:0
-----------------------------------------------------
Stack Trace:
RELADDR FUNCTION FILE:LINE
02ec8692 BraveProfileManagerTest_ExcludeServicesInOTRAndGuestProfiles_Test::RunTestOnMainThread() ../../brave/browser/profiles/brave_profile_manager_browsertest.cc:180:3
08814c44 content::BrowserTestBase::ProxyRunTestOnMainThreadLoop() ../../content/public/test/browser_test_base.cc:909:7
0881423e content::BrowserTestBase::SetUp() ../../content/public/test/browser_test_base.cc:707:5
0517b84a AndroidBrowserTest::SetUp() ../../chrome/test/base/android/android_browser_test.cc:25:20
0509892d bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) ../../third_party/googletest/src/googletest/src/gtest.cc:0:0
failed
```
</details>
|
non_defect
|
test failure braveprofilemanagertest excludeservicesinotrandguestprofiles greetings human bad news braveprofilemanagertest excludeservicesinotrandguestprofiles stack trace braveprofilemanagertest excludeservicesinotrandguestprofiles tracing subscriber init success skus sdk skus sdk initialized environment production version at home ubuntu workspace brave browser build android variant nightly src brave components skus browser rs lib src sdk mod rs skus storage kv got orders orders none at home ubuntu workspace brave browser build android variant nightly src brave components skus browser rs lib src storage kv rs in skus storage kv get orders with self nativeclient in skus sdk credentials matching credential summary with self sdk domain vpn brave com brave browser profiles brave profile manager browsertest cc failure expected brave rewards rewardsservicefactory getforprofile profile nullptr actual null vs nullptr stack trace brave browser profiles brave profile manager browsertest cc failure expected brave ads adsservicefactory getforprofile profile nullptr actual null vs nullptr stack trace parallelexecutionfence is ineffective when threadpoolinstance is not managed by a taskenvironment test fixtures should use a taskenvironment member or statically invoke taskenvironment createthreadpool threadpoolinstance get startwithdefaultparams when the former is not possible parallelexecutionfence is ineffective when threadpoolinstance is not managed by a taskenvironment test fixtures should use a taskenvironment member or statically invoke taskenvironment createthreadpool threadpoolinstance get startwithdefaultparams when the former is not possible braveprofilemanagertest excludeservicesinotrandguestprofiles where typeparam and getparam ms test from braveprofilemanagertest ms total global test environment tear down test from test suite ran ms total tests test listed below braveprofilemanagertest excludeservicesinotrandguestprofiles where typeparam and getparam stack trace reladdr function file line braveprofilemanagertest excludeservicesinotrandguestprofiles test runtestonmainthread brave browser profiles brave profile manager browsertest cc content browsertestbase proxyruntestonmainthreadloop content public test browser test base cc content browsertestbase setup content public test browser test base cc androidbrowsertest setup chrome test base android android browser test cc bool testing internal handleexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const third party googletest src googletest src gtest cc stack trace reladdr function file line braveprofilemanagertest excludeservicesinotrandguestprofiles test runtestonmainthread brave browser profiles brave profile manager browsertest cc content browsertestbase proxyruntestonmainthreadloop content public test browser test base cc content browsertestbase setup content public test browser test base cc androidbrowsertest setup chrome test base android android browser test cc bool testing internal handleexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const third party googletest src googletest src gtest cc failed
| 0
|
166,790
| 12,972,020,216
|
IssuesEvent
|
2020-07-21 11:55:28
|
prisma/prisma-client-js
|
https://api.github.com/repos/prisma/prisma-client-js
|
closed
|
Enum value 'false' breaks Prisma client create
|
bug/2-confirmed kind/bug status/needs-fix-confirmation team/engines topic: test-utils
|
## Bug description
If an enum has 'false' as one of the possible values. Prisma client fails at runtime while providing 'false' as the value for that enum.
## How to reproduce
1. Run the following SQL in a MySQL (untested with Postgres, SQLite) database
```
CREATE TABLE `e` (
`id` bigint(20) unsigned NOT NULL,
`visible` enum('true','false') NOT NULL DEFAULT 'true',
PRIMARY KEY (`id`)
) ENGINE=MyISAM DEFAULT CHARSET=utf8;
```
2. Introspect and generate client
```
model e {
id Int @id
visible e_visible @default(true)
}
enum e_visible {
true
false
}
```
3. Run the following Prisma client code
```
const data = await prisma.e.create({
data: {
id: 1,
visible: 'false',
},
})
```
4. It fails with
```
divyendusingh [p2-enum-bool]$ node index.js 130 ↵
(node:21526) UnhandledPromiseRejectionWarning: Error:
Invalid `prisma.e.create()` invocation in
/Users/divyendusingh/Documents/prisma/p2-enum-bool/index.js:6:31
2
3 const prisma = new PrismaClient()
4
5 async function main() {
→ 6 const data = await prisma.e.create(
Failed to validate the query `Error occurred during query validation & transformation:
Mutation (object)
↳ createOnee (field)
↳ data (argument)
↳ eCreateInput (object)
↳ visible (field)
↳ Error parsing value: Enum value 'false ' is invalid for enum type e_visible.` at `.Mutation.createOnee.data.eCreateInput.visible`
at PrismaClientFetcher._request (/Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/client/runtime/index.js:1:206978)
at processTicksAndRejections (internal/process/task_queues.js:97:5)
(node:21526) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag `--unhandled-rejections=strict` (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 1)
(node:21526) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
```
## Expected behavior
It should work
## Prisma information
```
yarn run v1.22.4
$ /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/.bin/prisma --version
@prisma/cli : 2.3.0-dev.16
Current platform : darwin
Query Engine : query-engine 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/query-engine-darwin)
Migration Engine : migration-engine-cli 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/migration-engine-darwin)
Introspection Engine : introspection-core 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/introspection-engine-darwin)
Format Binary : prisma-fmt 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/prisma-fmt-darwin)
Done in 1.42s.
```
Test utils database affected:
- mysql/piwigo
|
1.0
|
Enum value 'false' breaks Prisma client create - ## Bug description
If an enum has 'false' as one of the possible values. Prisma client fails at runtime while providing 'false' as the value for that enum.
## How to reproduce
1. Run the following SQL in a MySQL (untested with Postgres, SQLite) database
```
CREATE TABLE `e` (
`id` bigint(20) unsigned NOT NULL,
`visible` enum('true','false') NOT NULL DEFAULT 'true',
PRIMARY KEY (`id`)
) ENGINE=MyISAM DEFAULT CHARSET=utf8;
```
2. Introspect and generate client
```
model e {
id Int @id
visible e_visible @default(true)
}
enum e_visible {
true
false
}
```
3. Run the following Prisma client code
```
const data = await prisma.e.create({
data: {
id: 1,
visible: 'false',
},
})
```
4. It fails with
```
divyendusingh [p2-enum-bool]$ node index.js 130 ↵
(node:21526) UnhandledPromiseRejectionWarning: Error:
Invalid `prisma.e.create()` invocation in
/Users/divyendusingh/Documents/prisma/p2-enum-bool/index.js:6:31
2
3 const prisma = new PrismaClient()
4
5 async function main() {
→ 6 const data = await prisma.e.create(
Failed to validate the query `Error occurred during query validation & transformation:
Mutation (object)
↳ createOnee (field)
↳ data (argument)
↳ eCreateInput (object)
↳ visible (field)
↳ Error parsing value: Enum value 'false ' is invalid for enum type e_visible.` at `.Mutation.createOnee.data.eCreateInput.visible`
at PrismaClientFetcher._request (/Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/client/runtime/index.js:1:206978)
at processTicksAndRejections (internal/process/task_queues.js:97:5)
(node:21526) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag `--unhandled-rejections=strict` (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 1)
(node:21526) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
```
## Expected behavior
It should work
## Prisma information
```
yarn run v1.22.4
$ /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/.bin/prisma --version
@prisma/cli : 2.3.0-dev.16
Current platform : darwin
Query Engine : query-engine 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/query-engine-darwin)
Migration Engine : migration-engine-cli 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/migration-engine-darwin)
Introspection Engine : introspection-core 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/introspection-engine-darwin)
Format Binary : prisma-fmt 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/prisma-fmt-darwin)
Done in 1.42s.
```
Test utils database affected:
- mysql/piwigo
|
non_defect
|
enum value false breaks prisma client create bug description if an enum has false as one of the possible values prisma client fails at runtime while providing false as the value for that enum how to reproduce run the following sql in a mysql untested with postgres sqlite database create table e id bigint unsigned not null visible enum true false not null default true primary key id engine myisam default charset introspect and generate client model e id int id visible e visible default true enum e visible true false run the following prisma client code const data await prisma e create data id visible false it fails with divyendusingh node index js ↵ node unhandledpromiserejectionwarning error invalid prisma e create invocation in users divyendusingh documents prisma enum bool index js const prisma new prismaclient async function main → const data await prisma e create failed to validate the query error occurred during query validation transformation mutation object ↳ createonee field ↳ data argument ↳ ecreateinput object ↳ visible field ↳ error parsing value enum value false is invalid for enum type e visible at mutation createonee data ecreateinput visible at prismaclientfetcher request users divyendusingh documents prisma enum bool node modules prisma client runtime index js at processticksandrejections internal process task queues js node unhandledpromiserejectionwarning unhandled promise rejection this error originated either by throwing inside of an async function without a catch block or by rejecting a promise which was not handled with catch to terminate the node process on unhandled promise rejection use the cli flag unhandled rejections strict see rejection id node deprecationwarning unhandled promise rejections are deprecated in the future promise rejections that are not handled will terminate the node js process with a non zero exit code expected behavior it should work prisma information yarn run users divyendusingh documents prisma enum bool node modules bin prisma version prisma cli dev current platform darwin query engine query engine at users divyendusingh documents prisma enum bool node modules prisma cli query engine darwin migration engine migration engine cli at users divyendusingh documents prisma enum bool node modules prisma cli migration engine darwin introspection engine introspection core at users divyendusingh documents prisma enum bool node modules prisma cli introspection engine darwin format binary prisma fmt at users divyendusingh documents prisma enum bool node modules prisma cli prisma fmt darwin done in test utils database affected mysql piwigo
| 0
|
696,847
| 23,918,512,171
|
IssuesEvent
|
2022-09-09 14:42:18
|
camsaul/methodical
|
https://api.github.com/repos/camsaul/methodical
|
closed
|
`trace!` and `untrace!` facilities to trace existing existing code without changing it
|
enhancement high-priority!
|
- It would be good to be able to trace all usages of a certain method without touching existing code
- It would be good to be able to trace method calls in an external library
Maybe we can use `alter-var-root!` or something swap out the untraced multimethod with a traced one
|
1.0
|
`trace!` and `untrace!` facilities to trace existing existing code without changing it - - It would be good to be able to trace all usages of a certain method without touching existing code
- It would be good to be able to trace method calls in an external library
Maybe we can use `alter-var-root!` or something swap out the untraced multimethod with a traced one
|
non_defect
|
trace and untrace facilities to trace existing existing code without changing it it would be good to be able to trace all usages of a certain method without touching existing code it would be good to be able to trace method calls in an external library maybe we can use alter var root or something swap out the untraced multimethod with a traced one
| 0
|
567,896
| 16,918,839,507
|
IssuesEvent
|
2021-06-25 00:12:22
|
googleapis/gapic-generator-python
|
https://api.github.com/repos/googleapis/gapic-generator-python
|
opened
|
`name 'warnings' is not defined` raised for clients with deprecated methods
|
priority: p2 type: bug
|
See [log](https://source.cloud.google.com/results/invocations/2f31c792-2c4e-453f-943b-08403508b66f/targets/github%2Fpython-container/tests) from https://github.com/googleapis/python-container/pull/115 for an example
|
1.0
|
`name 'warnings' is not defined` raised for clients with deprecated methods - See [log](https://source.cloud.google.com/results/invocations/2f31c792-2c4e-453f-943b-08403508b66f/targets/github%2Fpython-container/tests) from https://github.com/googleapis/python-container/pull/115 for an example
|
non_defect
|
name warnings is not defined raised for clients with deprecated methods see from for an example
| 0
|
218,669
| 7,332,097,136
|
IssuesEvent
|
2018-03-05 15:25:31
|
NCEAS/metacat
|
https://api.github.com/repos/NCEAS/metacat
|
closed
|
continue updating user documentation
|
Category: metacat Component: Bugzilla-Id Priority: Normal Status: Resolved Tracker: Bug
|
---
Author Name: **Matt Jones** (Matt Jones)
Original Redmine Issue: 5516, https://projects.ecoinformatics.org/ecoinfo/issues/5516
Original Date: 2011-10-26
Original Assignee: ben leinfelder
---
Documentation needs editing to describe new 2.0.0 features, including support for new DataONE APIs, deprecation of older servlet APIs, and general cleanup.
If possible, moving the admin guide to the sphinx system now would be good.
|
1.0
|
continue updating user documentation - ---
Author Name: **Matt Jones** (Matt Jones)
Original Redmine Issue: 5516, https://projects.ecoinformatics.org/ecoinfo/issues/5516
Original Date: 2011-10-26
Original Assignee: ben leinfelder
---
Documentation needs editing to describe new 2.0.0 features, including support for new DataONE APIs, deprecation of older servlet APIs, and general cleanup.
If possible, moving the admin guide to the sphinx system now would be good.
|
non_defect
|
continue updating user documentation author name matt jones matt jones original redmine issue original date original assignee ben leinfelder documentation needs editing to describe new features including support for new dataone apis deprecation of older servlet apis and general cleanup if possible moving the admin guide to the sphinx system now would be good
| 0
|
292,688
| 25,229,496,582
|
IssuesEvent
|
2022-11-14 18:33:13
|
spack/spack
|
https://api.github.com/repos/spack/spack
|
closed
|
Testing issue: ginkgo@1.4.0%gcc@11.1.0
|
test-error
|
### Steps to reproduce the failure(s) or link(s) to test output(s)
@hartwiganzt @tcojean
Stand-alone tests failed during a Spack PR pipeline run. More information can be found at:
- Build: https://cdash.spack.io/build/1816848
- Stand-alone tests: https://cdash.spack.io/viewTest.php?buildid=1816848
### Error message
Build test software:
```
[ 25%] Building CXX object CMakeFiles/test_install.dir/test_install.cpp.o
[ 50%] Linking CXX executable test_install
/usr/bin/ld: warning: librocrand.so.1, needed by /home/software/spack/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/linux-ubuntu20.04-x86_64/gcc-11.1.0/rocrand-5.2.3-srqqcyn4wxm4uwfo3ulghfzxia2zvuq5/lib/libhiprand.so.1, not found (try using -rpath or -rpath-link)
/usr/bin/ld: /home/software/spack/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/linux-ubuntu20.04-x86_64/gcc-11.1.0/rocrand-5.2.3-srqqcyn4wxm4uwfo3ulghfzxia2zvuq5/lib/libhiprand.so.1: undefined reference to `rocrand_generate_normal_half'
/usr/bin/ld: /home/software/spack/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/linux-ubuntu20.04-x86_64/gcc-11.1.0/rocrand-5.2.3-srqqcyn4wxm4uwfo3ulghfzxia2zvuq5/lib/libhiprand.so.1: undefined reference to `rocrand_generate_normal'
...
```
### Information on your system or the test runner
Runner: uo-reptar
OS Release: 4.18.0-372.9.1.el8.x86_64
Spack version: spack-0.19.0.dev0 (10720fbb4e277c97a262dbf8d51f6b93c27dab2a)
Repository revision: [10720fbb4e277c97a262dbf8d51f6b93c27dab2a](https://github.com/spack/spack/commit/10720fbb4e277c97a262dbf8d51f6b93c27dab2a)
### Additional information
_No response_
### General information
- [X] I have reported the version of Spack/Python/Platform/Runner
- [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers
- [X] I have uploaded any available logs
- [X] I have searched the issues of this repo and believe this is not a duplicate
|
1.0
|
Testing issue: ginkgo@1.4.0%gcc@11.1.0 - ### Steps to reproduce the failure(s) or link(s) to test output(s)
@hartwiganzt @tcojean
Stand-alone tests failed during a Spack PR pipeline run. More information can be found at:
- Build: https://cdash.spack.io/build/1816848
- Stand-alone tests: https://cdash.spack.io/viewTest.php?buildid=1816848
### Error message
Build test software:
```
[ 25%] Building CXX object CMakeFiles/test_install.dir/test_install.cpp.o
[ 50%] Linking CXX executable test_install
/usr/bin/ld: warning: librocrand.so.1, needed by /home/software/spack/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/linux-ubuntu20.04-x86_64/gcc-11.1.0/rocrand-5.2.3-srqqcyn4wxm4uwfo3ulghfzxia2zvuq5/lib/libhiprand.so.1, not found (try using -rpath or -rpath-link)
/usr/bin/ld: /home/software/spack/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/linux-ubuntu20.04-x86_64/gcc-11.1.0/rocrand-5.2.3-srqqcyn4wxm4uwfo3ulghfzxia2zvuq5/lib/libhiprand.so.1: undefined reference to `rocrand_generate_normal_half'
/usr/bin/ld: /home/software/spack/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/linux-ubuntu20.04-x86_64/gcc-11.1.0/rocrand-5.2.3-srqqcyn4wxm4uwfo3ulghfzxia2zvuq5/lib/libhiprand.so.1: undefined reference to `rocrand_generate_normal'
...
```
### Information on your system or the test runner
Runner: uo-reptar
OS Release: 4.18.0-372.9.1.el8.x86_64
Spack version: spack-0.19.0.dev0 (10720fbb4e277c97a262dbf8d51f6b93c27dab2a)
Repository revision: [10720fbb4e277c97a262dbf8d51f6b93c27dab2a](https://github.com/spack/spack/commit/10720fbb4e277c97a262dbf8d51f6b93c27dab2a)
### Additional information
_No response_
### General information
- [X] I have reported the version of Spack/Python/Platform/Runner
- [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers
- [X] I have uploaded any available logs
- [X] I have searched the issues of this repo and believe this is not a duplicate
|
non_defect
|
testing issue ginkgo gcc steps to reproduce the failure s or link s to test output s hartwiganzt tcojean stand alone tests failed during a spack pr pipeline run more information can be found at build stand alone tests error message build test software building cxx object cmakefiles test install dir test install cpp o linking cxx executable test install usr bin ld warning librocrand so needed by home software spack spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spa linux gcc rocrand lib libhiprand so not found try using rpath or rpath link usr bin ld home software spack spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spa linux gcc rocrand lib libhiprand so undefined reference to rocrand generate normal half usr bin ld home software spack spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spa linux gcc rocrand lib libhiprand so undefined reference to rocrand generate normal information on your system or the test runner runner uo reptar os release spack version spack repository revision additional information no response general information i have reported the version of spack python platform runner i have run spack maintainers and mentioned any maintainers i have uploaded any available logs i have searched the issues of this repo and believe this is not a duplicate
| 0
|
58,630
| 16,662,876,925
|
IssuesEvent
|
2021-06-06 16:49:36
|
Questie/Questie
|
https://api.github.com/repos/Questie/Questie
|
opened
|
Questie error for the quest Gorgrom the dragon-Eater
|
Type - Defect
|
Hi, using questie 6.3.14
Got mass error when accepting Gorgrom the dragon-Eater in blades edge.
Questie: [ERROR] [QuestieQuest]: v6.3.14 - There was an error populating objectives for Gorgrom the Dragon-Eater 10723 1 No error
...erface\AddOns\Questie\Modules\Quest\QuestieQuest.lua:690: bad argument #1 to 'next' (table expected, got nil)
[string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:556: in function <...erface\AddOns\Questie\Modules\Quest\QuestieQuest.lua:554>
[string "=[C]"]: in function `next'
[string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:690: in function <...erface\AddOns\Questie\Modules\Quest\QuestieQuest.lua:683>
[string "=[C]"]: in function `xpcall'
[string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:554: in function `UpdateObjectiveNotes'
[string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:469: in function `UpdateQuest'
[string "@Interface\AddOns\Questie\Modules\Libs\QuestieHash.lua"]:108: in function <...nterface\AddOns\Questie\Modules\Libs\QuestieHash.lua:103>
[string "@Interface\AddOns\Questie\Modules\Libs\QuestieHash.lua"]:91: in function `CompareQuestHashes'
[string "@Interface\AddOns\Questie\Modules\QuestieEventHandler.lua"]:296: in function <...rface\AddOns\Questie\Modules\QuestieEventHandler.lua:278>
[string "=[C]"]: ?
[string "@Interface\AddOns\Quartz\libs\AceBucket-3.0\AceBucket-3.0.lua"]:73: in function <...e\AddOns\Quartz\libs\AceBucket-3.0\AceBucket-3.0.lua:71>
[string "@Interface\AddOns\Quartz\libs\AceBucket-3.0\AceBucket-3.0.lua"]:89: in function `func'
[string "@Interface\AddOns\Details\Libs\AceTimer-3.0\AceTimer-3.0.lua"]:57: in function <...ce\AddOns\Details\Libs\AceTimer-3.0\AceTimer-3.0.lua:50>
|
1.0
|
Questie error for the quest Gorgrom the dragon-Eater - Hi, using questie 6.3.14
Got mass error when accepting Gorgrom the dragon-Eater in blades edge.
Questie: [ERROR] [QuestieQuest]: v6.3.14 - There was an error populating objectives for Gorgrom the Dragon-Eater 10723 1 No error
...erface\AddOns\Questie\Modules\Quest\QuestieQuest.lua:690: bad argument #1 to 'next' (table expected, got nil)
[string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:556: in function <...erface\AddOns\Questie\Modules\Quest\QuestieQuest.lua:554>
[string "=[C]"]: in function `next'
[string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:690: in function <...erface\AddOns\Questie\Modules\Quest\QuestieQuest.lua:683>
[string "=[C]"]: in function `xpcall'
[string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:554: in function `UpdateObjectiveNotes'
[string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:469: in function `UpdateQuest'
[string "@Interface\AddOns\Questie\Modules\Libs\QuestieHash.lua"]:108: in function <...nterface\AddOns\Questie\Modules\Libs\QuestieHash.lua:103>
[string "@Interface\AddOns\Questie\Modules\Libs\QuestieHash.lua"]:91: in function `CompareQuestHashes'
[string "@Interface\AddOns\Questie\Modules\QuestieEventHandler.lua"]:296: in function <...rface\AddOns\Questie\Modules\QuestieEventHandler.lua:278>
[string "=[C]"]: ?
[string "@Interface\AddOns\Quartz\libs\AceBucket-3.0\AceBucket-3.0.lua"]:73: in function <...e\AddOns\Quartz\libs\AceBucket-3.0\AceBucket-3.0.lua:71>
[string "@Interface\AddOns\Quartz\libs\AceBucket-3.0\AceBucket-3.0.lua"]:89: in function `func'
[string "@Interface\AddOns\Details\Libs\AceTimer-3.0\AceTimer-3.0.lua"]:57: in function <...ce\AddOns\Details\Libs\AceTimer-3.0\AceTimer-3.0.lua:50>
|
defect
|
questie error for the quest gorgrom the dragon eater hi using questie got mass error when accepting gorgrom the dragon eater in blades edge questie there was an error populating objectives for gorgrom the dragon eater no error erface addons questie modules quest questiequest lua bad argument to next table expected got nil in function in function next in function in function xpcall in function updateobjectivenotes in function updatequest in function in function comparequesthashes in function in function in function func in function
| 1
|
69,057
| 22,098,352,015
|
IssuesEvent
|
2022-06-01 11:58:24
|
DEVA9N/FlexTimeMonitor
|
https://api.github.com/repos/DEVA9N/FlexTimeMonitor
|
closed
|
The options (break time) are gone after reinstalling the programm
|
Priority-Medium Type-Defect auto-migrated
|
```
What steps will reproduce the problem?
1. Reinstall or install a new version of Flex Time Monitor
2. Notice how the default options are used
What is the expected output? What do you see instead?
expected: the options will stay the same as before the new installation
instead: the default options are set
```
Original issue reported on code.google.com by `aj9...@gmail.com` on 6 Aug 2012 at 3:25
|
1.0
|
The options (break time) are gone after reinstalling the programm - ```
What steps will reproduce the problem?
1. Reinstall or install a new version of Flex Time Monitor
2. Notice how the default options are used
What is the expected output? What do you see instead?
expected: the options will stay the same as before the new installation
instead: the default options are set
```
Original issue reported on code.google.com by `aj9...@gmail.com` on 6 Aug 2012 at 3:25
|
defect
|
the options break time are gone after reinstalling the programm what steps will reproduce the problem reinstall or install a new version of flex time monitor notice how the default options are used what is the expected output what do you see instead expected the options will stay the same as before the new installation instead the default options are set original issue reported on code google com by gmail com on aug at
| 1
|
38,599
| 8,924,971,976
|
IssuesEvent
|
2019-01-21 20:44:11
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
opened
|
Python utils check file size can fail (race condition)
|
C: MOOSE Scripts C: TestHarness P: normal T: defect
|
## Rationale
<!--What is the reason for this enhancement or what error are you reporting?-->
The python/mooseutils/tests.check_file_size test can fail due to a race condition. The problem is that this test runs concurrently with several other tests on the system and it does an os.walk() looking for files sizes to sum up. It's possible for the file list to change (e.g. files can be removed) while this test is running causing the `os.stat()` command to fail. This test should be cleaned up to work in either a clean directory where other test outputs are not being written, or the test should be serialized and isolated.
## Description
<!--Provide details of the enhancement or instructions for reproducing the error.-->
Race condition in python testing (mooseutils.check_file_size). This failure is difficult to replicate, but it should be fairly easy to clean up. See description above.
## Impact
<!--How will the changes impact the code, developers, and users? Discuss changes to the
internal interfaces and public API.-->
minor: Does not affect MOOSE, but can impede testing.
|
1.0
|
Python utils check file size can fail (race condition) - ## Rationale
<!--What is the reason for this enhancement or what error are you reporting?-->
The python/mooseutils/tests.check_file_size test can fail due to a race condition. The problem is that this test runs concurrently with several other tests on the system and it does an os.walk() looking for files sizes to sum up. It's possible for the file list to change (e.g. files can be removed) while this test is running causing the `os.stat()` command to fail. This test should be cleaned up to work in either a clean directory where other test outputs are not being written, or the test should be serialized and isolated.
## Description
<!--Provide details of the enhancement or instructions for reproducing the error.-->
Race condition in python testing (mooseutils.check_file_size). This failure is difficult to replicate, but it should be fairly easy to clean up. See description above.
## Impact
<!--How will the changes impact the code, developers, and users? Discuss changes to the
internal interfaces and public API.-->
minor: Does not affect MOOSE, but can impede testing.
|
defect
|
python utils check file size can fail race condition rationale the python mooseutils tests check file size test can fail due to a race condition the problem is that this test runs concurrently with several other tests on the system and it does an os walk looking for files sizes to sum up it s possible for the file list to change e g files can be removed while this test is running causing the os stat command to fail this test should be cleaned up to work in either a clean directory where other test outputs are not being written or the test should be serialized and isolated description race condition in python testing mooseutils check file size this failure is difficult to replicate but it should be fairly easy to clean up see description above impact how will the changes impact the code developers and users discuss changes to the internal interfaces and public api minor does not affect moose but can impede testing
| 1
|
56,058
| 14,916,170,434
|
IssuesEvent
|
2021-01-22 17:47:39
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
[CI/CD]: Review coverage of accessibility checks in 996 end-to-end tests
|
508-defect-3 508/Accessibility HLR testing vsa vsa-benefits
|
**Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
Applications **must** have thorough end-to-end tests that run in our continuous integration/continuous deployment (CI/CD) pipeline. These tests should include thorough axe checks. While auditing the `/disability-benefits/996` app folder, I wasn't sure if there were modals or hidden content that needed axe checks. I'd like the front-end engineering team to review this application, and add end-to-end tests as needed. Definition of done in acceptance criteria below.
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket.
-->
**VFS Point of Contact:** _Jennifer_
## Environment
* `vets-website/src/applications/disability-benefits/996/tests`
* `$ yarn test:e2e src/applications/disability-benefits/996/tests`
## Acceptance Criteria
<!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. -->
**Definition of done:**
- [ ] Front-end team member(s) have reviewed end-to-end tests for axe checks
- [ ] axe checks are run for hidden content like modal windows, accordions
- [ ] FE team has consulted with accessibility specialist in cases where there are high numbers of modals, accordions, other hidden content that could slow down e2e test runs.
- [ ] No axe `violations` appear in the plugin console. These will break the CI/CD build.
## WCAG or Vendor Guidance (optional)
* [Custom axeCheck helper method](https://github.com/department-of-veterans-affairs/vets-website/blob/master/src/platform/testing/e2e/nightwatch-commands/axeCheck.js)
* [VSP guidance on writing end-to-end tests](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/quality-assurance/e2e-testing)
|
1.0
|
[CI/CD]: Review coverage of accessibility checks in 996 end-to-end tests - **Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
Applications **must** have thorough end-to-end tests that run in our continuous integration/continuous deployment (CI/CD) pipeline. These tests should include thorough axe checks. While auditing the `/disability-benefits/996` app folder, I wasn't sure if there were modals or hidden content that needed axe checks. I'd like the front-end engineering team to review this application, and add end-to-end tests as needed. Definition of done in acceptance criteria below.
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket.
-->
**VFS Point of Contact:** _Jennifer_
## Environment
* `vets-website/src/applications/disability-benefits/996/tests`
* `$ yarn test:e2e src/applications/disability-benefits/996/tests`
## Acceptance Criteria
<!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. -->
**Definition of done:**
- [ ] Front-end team member(s) have reviewed end-to-end tests for axe checks
- [ ] axe checks are run for hidden content like modal windows, accordions
- [ ] FE team has consulted with accessibility specialist in cases where there are high numbers of modals, accordions, other hidden content that could slow down e2e test runs.
- [ ] No axe `violations` appear in the plugin console. These will break the CI/CD build.
## WCAG or Vendor Guidance (optional)
* [Custom axeCheck helper method](https://github.com/department-of-veterans-affairs/vets-website/blob/master/src/platform/testing/e2e/nightwatch-commands/axeCheck.js)
* [VSP guidance on writing end-to-end tests](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/quality-assurance/e2e-testing)
|
defect
|
review coverage of accessibility checks in end to end tests feedback framework ❗️ must for if the feedback must be applied ⚠️should if the feedback is best practice ✔️ consider for suggestions enhancements description applications must have thorough end to end tests that run in our continuous integration continuous deployment ci cd pipeline these tests should include thorough axe checks while auditing the disability benefits app folder i wasn t sure if there were modals or hidden content that needed axe checks i d like the front end engineering team to review this application and add end to end tests as needed definition of done in acceptance criteria below point of contact if this issue is being opened by a vfs team member please add a point of contact usually this is the same person who enters the issue ticket vfs point of contact jennifer environment vets website src applications disability benefits tests yarn test src applications disability benefits tests acceptance criteria definition of done front end team member s have reviewed end to end tests for axe checks axe checks are run for hidden content like modal windows accordions fe team has consulted with accessibility specialist in cases where there are high numbers of modals accordions other hidden content that could slow down test runs no axe violations appear in the plugin console these will break the ci cd build wcag or vendor guidance optional
| 1
|
29,023
| 8,250,827,653
|
IssuesEvent
|
2018-09-12 05:11:34
|
avast-tl/retdec
|
https://api.github.com/repos/avast-tl/retdec
|
closed
|
Could NOT find PythonInterp: Found unsuitable version "2.7.14", but required is at least "3.4" (found /usr/local/bin/python)
|
C-build-system O-macos P-build
|
Someone please help me.
OSX, I have both python 2.7 installed and python 3.7 installed.
My PATH is /usr/local/opt/python3.7:/Library/Frameworks/Python.framework/Versions/3.7/bin/python3:/usr/local/opt/flex/bin:/usr/local/opt/bison/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin
When I try to cmake it always says:
Could NOT find PythonInterp: Found unsuitable version "2.7.14", but
required is at least "3.4" (found /usr/local/bin/python)
How do I fix this?
|
2.0
|
Could NOT find PythonInterp: Found unsuitable version "2.7.14", but required is at least "3.4" (found /usr/local/bin/python) - Someone please help me.
OSX, I have both python 2.7 installed and python 3.7 installed.
My PATH is /usr/local/opt/python3.7:/Library/Frameworks/Python.framework/Versions/3.7/bin/python3:/usr/local/opt/flex/bin:/usr/local/opt/bison/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin
When I try to cmake it always says:
Could NOT find PythonInterp: Found unsuitable version "2.7.14", but
required is at least "3.4" (found /usr/local/bin/python)
How do I fix this?
|
non_defect
|
could not find pythoninterp found unsuitable version but required is at least found usr local bin python someone please help me osx i have both python installed and python installed my path is usr local opt library frameworks python framework versions bin usr local opt flex bin usr local opt bison bin usr local bin usr bin bin usr sbin sbin when i try to cmake it always says could not find pythoninterp found unsuitable version but required is at least found usr local bin python how do i fix this
| 0
|
7,155
| 2,610,329,582
|
IssuesEvent
|
2015-02-26 19:46:12
|
chrsmith/republic-at-war
|
https://api.github.com/repos/chrsmith/republic-at-war
|
closed
|
Map Issue
|
auto-migrated Priority-Low Type-Defect
|
```
I've got a few bugs to report. I've noticed that both Sluis Van and Raxus Prime
do not have Reinforcement points when I play on them for some reason. When I
originally fought on the planets they had reinforcement points but they
dissapeared after I conquered them. Then, the CIS will launch assaults on these
worlds and will have no Reinforcement points to drop reinforcements from.
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 15 Jun 2011 at 12:45
|
1.0
|
Map Issue - ```
I've got a few bugs to report. I've noticed that both Sluis Van and Raxus Prime
do not have Reinforcement points when I play on them for some reason. When I
originally fought on the planets they had reinforcement points but they
dissapeared after I conquered them. Then, the CIS will launch assaults on these
worlds and will have no Reinforcement points to drop reinforcements from.
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 15 Jun 2011 at 12:45
|
defect
|
map issue i ve got a few bugs to report i ve noticed that both sluis van and raxus prime do not have reinforcement points when i play on them for some reason when i originally fought on the planets they had reinforcement points but they dissapeared after i conquered them then the cis will launch assaults on these worlds and will have no reinforcement points to drop reinforcements from original issue reported on code google com by gmail com on jun at
| 1
|
40,380
| 9,977,380,804
|
IssuesEvent
|
2019-07-09 17:07:37
|
mozilla/experimenter
|
https://api.github.com/repos/mozilla/experimenter
|
closed
|
The "Help" button for the "Lightning Advisory (Optional)" Sign-off is wrongly linked to the mana page/"Dependent Sign-offs"
|
Defect P1 - High Priority [QA]:Normal issue
|
**[Action Needed]:**
Please link the "Lightning Advising" help here: https://mana.mozilla.org/wiki/display/FIREFOX/Pref-Flip+and+Add-On+Experiments#Pref-FlipandAdd-OnExperiments-LightningAdvising
**[Notes]:**
- I logged this issue because the "Lightning Advisory (Optional)" Sign-off doesn't have any dependencies with the Risk questions and also, it's description on the mana page is displayed in the "Optional Sign-off" section.
**[Affected versions]:**
- Firefox Release 67.0.1 and above
- Chrome 75 and above.
**[Affected Platforms]:**
- All Windows
- All Mac
- All Linux
**[Prerequisites]:**
- You are logged in your LDAP account on Experimenter stage website, "https://experimenter.stage.mozaws.net/".
- Have an experiment in the "Ready for Sign-Off state, created on Experimenter stage website.
**[Steps to reproduce]:**
1. Open the browser with the profile from prerequisites and navigate to "https://experimenter.stage.mozaws.net/".
2. Open the experiment from prerequisites.
3. Click the "Help" button for the "Lightning Advisory (Optional)" Sign-Off.
4. Observe the behavior.
**[Expected result]:**
- The mana page is opened on the "Lightning Advising" section.
**[Actual result]:**
- The mana page is opened on the "Dependent Sign-offs" section.
**[Additional Notes]:**
- Attached is a screen recording of the issue:

|
1.0
|
The "Help" button for the "Lightning Advisory (Optional)" Sign-off is wrongly linked to the mana page/"Dependent Sign-offs" - **[Action Needed]:**
Please link the "Lightning Advising" help here: https://mana.mozilla.org/wiki/display/FIREFOX/Pref-Flip+and+Add-On+Experiments#Pref-FlipandAdd-OnExperiments-LightningAdvising
**[Notes]:**
- I logged this issue because the "Lightning Advisory (Optional)" Sign-off doesn't have any dependencies with the Risk questions and also, it's description on the mana page is displayed in the "Optional Sign-off" section.
**[Affected versions]:**
- Firefox Release 67.0.1 and above
- Chrome 75 and above.
**[Affected Platforms]:**
- All Windows
- All Mac
- All Linux
**[Prerequisites]:**
- You are logged in your LDAP account on Experimenter stage website, "https://experimenter.stage.mozaws.net/".
- Have an experiment in the "Ready for Sign-Off state, created on Experimenter stage website.
**[Steps to reproduce]:**
1. Open the browser with the profile from prerequisites and navigate to "https://experimenter.stage.mozaws.net/".
2. Open the experiment from prerequisites.
3. Click the "Help" button for the "Lightning Advisory (Optional)" Sign-Off.
4. Observe the behavior.
**[Expected result]:**
- The mana page is opened on the "Lightning Advising" section.
**[Actual result]:**
- The mana page is opened on the "Dependent Sign-offs" section.
**[Additional Notes]:**
- Attached is a screen recording of the issue:

|
defect
|
the help button for the lightning advisory optional sign off is wrongly linked to the mana page dependent sign offs please link the lightning advising help here i logged this issue because the lightning advisory optional sign off doesn t have any dependencies with the risk questions and also it s description on the mana page is displayed in the optional sign off section firefox release and above chrome and above all windows all mac all linux you are logged in your ldap account on experimenter stage website have an experiment in the ready for sign off state created on experimenter stage website open the browser with the profile from prerequisites and navigate to open the experiment from prerequisites click the help button for the lightning advisory optional sign off observe the behavior the mana page is opened on the lightning advising section the mana page is opened on the dependent sign offs section attached is a screen recording of the issue
| 1
|
187,095
| 14,426,956,619
|
IssuesEvent
|
2020-12-06 01:00:41
|
kalexmills/github-vet-tests-dec2020
|
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
|
closed
|
giantswarm/kvm-operator-node-controller: vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go; 30 LoC
|
fresh small test vendored
|
Found a possible issue in [giantswarm/kvm-operator-node-controller](https://www.github.com/giantswarm/kvm-operator-node-controller) at [vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go](https://github.com/giantswarm/kvm-operator-node-controller/blob/7146561e54142d4f986daee0206336ebee3ceb18/vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go#L1064-L1093)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to svc at line 1067 may start a goroutine
[Click here to see the code in its original context.](https://github.com/giantswarm/kvm-operator-node-controller/blob/7146561e54142d4f986daee0206336ebee3ceb18/vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go#L1064-L1093)
<details>
<summary>Click here to show the 30 line(s) of Go which triggered the analyzer.</summary>
```go
for _, svc := range tests {
for _, wide := range []bool{false, true} {
buff := bytes.Buffer{}
printService(&svc, &buff, printers.PrintOptions{Wide: wide})
output := string(buff.Bytes())
ip := svc.Spec.ClusterIP
if !strings.Contains(output, ip) {
t.Errorf("expected to contain ClusterIP %s, but doesn't: %s", ip, output)
}
for n, ingress := range svc.Status.LoadBalancer.Ingress {
ip = ingress.IP
// For non-wide output, we only guarantee the first IP to be printed
if (n == 0 || wide) && !strings.Contains(output, ip) {
t.Errorf("expected to contain ingress ip %s with wide=%v, but doesn't: %s", ip, wide, output)
}
}
for _, port := range svc.Spec.Ports {
portSpec := fmt.Sprintf("%d/%s", port.Port, port.Protocol)
if !strings.Contains(output, portSpec) {
t.Errorf("expected to contain port: %s, but doesn't: %s", portSpec, output)
}
}
// Each service should print on one line
if 1 != strings.Count(output, "\n") {
t.Errorf("expected a single newline, found %d", strings.Count(output, "\n"))
}
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 7146561e54142d4f986daee0206336ebee3ceb18
|
1.0
|
giantswarm/kvm-operator-node-controller: vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go; 30 LoC -
Found a possible issue in [giantswarm/kvm-operator-node-controller](https://www.github.com/giantswarm/kvm-operator-node-controller) at [vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go](https://github.com/giantswarm/kvm-operator-node-controller/blob/7146561e54142d4f986daee0206336ebee3ceb18/vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go#L1064-L1093)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to svc at line 1067 may start a goroutine
[Click here to see the code in its original context.](https://github.com/giantswarm/kvm-operator-node-controller/blob/7146561e54142d4f986daee0206336ebee3ceb18/vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go#L1064-L1093)
<details>
<summary>Click here to show the 30 line(s) of Go which triggered the analyzer.</summary>
```go
for _, svc := range tests {
for _, wide := range []bool{false, true} {
buff := bytes.Buffer{}
printService(&svc, &buff, printers.PrintOptions{Wide: wide})
output := string(buff.Bytes())
ip := svc.Spec.ClusterIP
if !strings.Contains(output, ip) {
t.Errorf("expected to contain ClusterIP %s, but doesn't: %s", ip, output)
}
for n, ingress := range svc.Status.LoadBalancer.Ingress {
ip = ingress.IP
// For non-wide output, we only guarantee the first IP to be printed
if (n == 0 || wide) && !strings.Contains(output, ip) {
t.Errorf("expected to contain ingress ip %s with wide=%v, but doesn't: %s", ip, wide, output)
}
}
for _, port := range svc.Spec.Ports {
portSpec := fmt.Sprintf("%d/%s", port.Port, port.Protocol)
if !strings.Contains(output, portSpec) {
t.Errorf("expected to contain port: %s, but doesn't: %s", portSpec, output)
}
}
// Each service should print on one line
if 1 != strings.Count(output, "\n") {
t.Errorf("expected a single newline, found %d", strings.Count(output, "\n"))
}
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 7146561e54142d4f986daee0206336ebee3ceb18
|
non_defect
|
giantswarm kvm operator node controller vendor io kubernetes pkg printers internalversion printers test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to svc at line may start a goroutine click here to show the line s of go which triggered the analyzer go for svc range tests for wide range bool false true buff bytes buffer printservice svc buff printers printoptions wide wide output string buff bytes ip svc spec clusterip if strings contains output ip t errorf expected to contain clusterip s but doesn t s ip output for n ingress range svc status loadbalancer ingress ip ingress ip for non wide output we only guarantee the first ip to be printed if n wide strings contains output ip t errorf expected to contain ingress ip s with wide v but doesn t s ip wide output for port range svc spec ports portspec fmt sprintf d s port port port protocol if strings contains output portspec t errorf expected to contain port s but doesn t s portspec output each service should print on one line if strings count output n t errorf expected a single newline found d strings count output n leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 0
|
60,816
| 14,919,928,882
|
IssuesEvent
|
2021-01-23 02:01:44
|
microsoft/fluentui
|
https://api.github.com/repos/microsoft/fluentui
|
closed
|
Migrate e2e / accessibility tests to Playwright
|
Area: Build System Area: Testing Fluent UI react-northstar Needs: Discussion 🙋 Type: Feature
|
Currently Puppeteer will be used to perform basic e2e and accessibility tests which is based on Chromium. What about migrating it over to [Playwright](https://github.com/microsoft/playwright) which supports Chromium / WebKit / Firefox and running / checking it with all of them in a row?
I already played a bit around with the [e2e tests](https://github.com/microsoft/fluentui/tree/master/packages/fluentui/e2e/tests) locally and these tests are working mostly almost directly. But I wanted to ask first, what you think about it. I'm open to contribute to it.
|
1.0
|
Migrate e2e / accessibility tests to Playwright - Currently Puppeteer will be used to perform basic e2e and accessibility tests which is based on Chromium. What about migrating it over to [Playwright](https://github.com/microsoft/playwright) which supports Chromium / WebKit / Firefox and running / checking it with all of them in a row?
I already played a bit around with the [e2e tests](https://github.com/microsoft/fluentui/tree/master/packages/fluentui/e2e/tests) locally and these tests are working mostly almost directly. But I wanted to ask first, what you think about it. I'm open to contribute to it.
|
non_defect
|
migrate accessibility tests to playwright currently puppeteer will be used to perform basic and accessibility tests which is based on chromium what about migrating it over to which supports chromium webkit firefox and running checking it with all of them in a row i already played a bit around with the locally and these tests are working mostly almost directly but i wanted to ask first what you think about it i m open to contribute to it
| 0
|
785,757
| 27,624,232,421
|
IssuesEvent
|
2023-03-10 04:36:48
|
AY2223S2-CS2113-W12-1/tp
|
https://api.github.com/repos/AY2223S2-CS2113-W12-1/tp
|
opened
|
Create a remove appointment feature.
|
type.Story priority.High
|
As a user, I am able to remove appointments if necessary so that the appointment list is not clogged up.
|
1.0
|
Create a remove appointment feature. - As a user, I am able to remove appointments if necessary so that the appointment list is not clogged up.
|
non_defect
|
create a remove appointment feature as a user i am able to remove appointments if necessary so that the appointment list is not clogged up
| 0
|
4,287
| 2,610,090,946
|
IssuesEvent
|
2015-02-26 18:27:32
|
chrsmith/dsdsdaadf
|
https://api.github.com/repos/chrsmith/dsdsdaadf
|
opened
|
深圳痘印方法
|
auto-migrated Priority-Medium Type-Defect
|
```
深圳痘印方法【深圳韩方科颜全国热线400-869-1818,24小时QQ4008
691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方—��
�韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科颜�
��业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康祛
痘技术并结合先进“先进豪华彩光”仪,开创国内专业治疗��
�刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:45
|
1.0
|
深圳痘印方法 - ```
深圳痘印方法【深圳韩方科颜全国热线400-869-1818,24小时QQ4008
691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方—��
�韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科颜�
��业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康祛
痘技术并结合先进“先进豪华彩光”仪,开创国内专业治疗��
�刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:45
|
defect
|
深圳痘印方法 深圳痘印方法【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方—�� �韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科颜� ��业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康祛 痘技术并结合先进“先进豪华彩光”仪,开创国内专业治疗�� �刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘。 original issue reported on code google com by szft com on may at
| 1
|
120,250
| 4,787,175,689
|
IssuesEvent
|
2016-10-29 21:05:26
|
minj/foxtrick
|
https://api.github.com/repos/minj/foxtrick
|
closed
|
CurrentTransfers caches stale CHPP response if player has just been TLed
|
bug Priority-Low started Transfers
|
**From:** LA-MJ
**PostID:** [16898912.142](https://www.hattrick.org/goto.ashx?path=%2FForum%2FRead.aspx%3Ft%3D16898912%26n%3D142%26v%3D0)
**To:** kosio
**Re:** [16898912.140](https://www.hattrick.org/goto.ashx?path=%2FForum%2FRead.aspx%3Ft%3D16898912%26n%3D140%26v%3D0)
**Datetime:** 2016-01-17 12:50
**Message:**
> kosio rašė:
>
> Hi,
>
> Just got the problem with missing opening price again.
>
> [(http://screencast.com/t/l7g9UNzNc)](http://screencast.com/t/l7g9UNzNc)
> [(http://pastebin.com/QMYkcLBX)](http://pastebin.com/QMYkcLBX)
>
> If you need something else, just let me know.
PlayerID would be useful in these situations, otherwise I need to look for it manually.
This seems to have been a temporary CHPP hickup. FT log indicates 'AskingPrice' field was missing in the CHPP response. I guess this is because you were so quick to look at the transfers page (screenshot says 2d 23h 57min) that CHPP returned a cached response without transfer data. I guess I need to guard against these somehow...
|
1.0
|
CurrentTransfers caches stale CHPP response if player has just been TLed - **From:** LA-MJ
**PostID:** [16898912.142](https://www.hattrick.org/goto.ashx?path=%2FForum%2FRead.aspx%3Ft%3D16898912%26n%3D142%26v%3D0)
**To:** kosio
**Re:** [16898912.140](https://www.hattrick.org/goto.ashx?path=%2FForum%2FRead.aspx%3Ft%3D16898912%26n%3D140%26v%3D0)
**Datetime:** 2016-01-17 12:50
**Message:**
> kosio rašė:
>
> Hi,
>
> Just got the problem with missing opening price again.
>
> [(http://screencast.com/t/l7g9UNzNc)](http://screencast.com/t/l7g9UNzNc)
> [(http://pastebin.com/QMYkcLBX)](http://pastebin.com/QMYkcLBX)
>
> If you need something else, just let me know.
PlayerID would be useful in these situations, otherwise I need to look for it manually.
This seems to have been a temporary CHPP hickup. FT log indicates 'AskingPrice' field was missing in the CHPP response. I guess this is because you were so quick to look at the transfers page (screenshot says 2d 23h 57min) that CHPP returned a cached response without transfer data. I guess I need to guard against these somehow...
|
non_defect
|
currenttransfers caches stale chpp response if player has just been tled from la mj postid to kosio re datetime message kosio rašė hi just got the problem with missing opening price again if you need something else just let me know playerid would be useful in these situations otherwise i need to look for it manually this seems to have been a temporary chpp hickup ft log indicates askingprice field was missing in the chpp response i guess this is because you were so quick to look at the transfers page screenshot says that chpp returned a cached response without transfer data i guess i need to guard against these somehow
| 0
|
460,752
| 13,217,743,422
|
IssuesEvent
|
2020-08-17 07:25:03
|
netdata/netdata
|
https://api.github.com/repos/netdata/netdata
|
closed
|
New exporting module is missing metrics compared to old when sending to graphite
|
area/backends bug priority/medium
|
##### Bug report summary
I have migrated the exporting config from the old way (using `[backend]`) to the new way by using `exporting.conf`
However, there is something strange, or I have put wrong config, because we do not get all the metrics.
##### OS / Environment
Running netdata v1.23.1 on docker, together with grafana latest (v7.0.6) in docker and graphite 1.1.7-3 also in docker.
I am running them on a macOS host, but I have also seen missing stats on our ubuntu 18.04 servers.
##### Component Name
The metrics send to graphite have gaps, which did not happen before. I have made a test by running a single netdata with the new config and old config.

The yellow and green uptime are when I run netdata with the new exporting config
The red and blue uptime are when I run netdata with the old backend config
##### Steps To Reproduce
1. Run netdata with new backend and old backend and send to graphite
Here is my `[backend]` config:
```
[backend]
enabled = yes
type = graphite
destination = docker.for.mac.localhost:2023
data source = as collected
prefix = netdata
update every = 10
buffer on failures = 10
timeout ms = 20000
send charts matching = system.uptime* system.cpu* system.ram* system.net* disk_space.* !*
send hosts matching = localhost *
send names instead of ids = yes
```
Here is my `exporting.conf` file:
```
[graphite:netdata]
enabled = yes
destination = docker.for.mac.localhost:2023
data source = as collected
prefix = netdata
update every = 10
buffer on failures = 10
timeout ms = 20000
send charts matching = system.uptime* system.cpu* system.ram* system.net* disk_space.* !*
send hosts matching = localhost *
send names instead of ids = yes
```
This is the only config which I have changed when testing this behaviour
##### Expected behavior
I expect that I do not see gaps in the dashboard when using the new style.
It could be that I have configured the new one in a wrong way, in such a case I would like to hear any suggestions.
|
1.0
|
New exporting module is missing metrics compared to old when sending to graphite - ##### Bug report summary
I have migrated the exporting config from the old way (using `[backend]`) to the new way by using `exporting.conf`
However, there is something strange, or I have put wrong config, because we do not get all the metrics.
##### OS / Environment
Running netdata v1.23.1 on docker, together with grafana latest (v7.0.6) in docker and graphite 1.1.7-3 also in docker.
I am running them on a macOS host, but I have also seen missing stats on our ubuntu 18.04 servers.
##### Component Name
The metrics send to graphite have gaps, which did not happen before. I have made a test by running a single netdata with the new config and old config.

The yellow and green uptime are when I run netdata with the new exporting config
The red and blue uptime are when I run netdata with the old backend config
##### Steps To Reproduce
1. Run netdata with new backend and old backend and send to graphite
Here is my `[backend]` config:
```
[backend]
enabled = yes
type = graphite
destination = docker.for.mac.localhost:2023
data source = as collected
prefix = netdata
update every = 10
buffer on failures = 10
timeout ms = 20000
send charts matching = system.uptime* system.cpu* system.ram* system.net* disk_space.* !*
send hosts matching = localhost *
send names instead of ids = yes
```
Here is my `exporting.conf` file:
```
[graphite:netdata]
enabled = yes
destination = docker.for.mac.localhost:2023
data source = as collected
prefix = netdata
update every = 10
buffer on failures = 10
timeout ms = 20000
send charts matching = system.uptime* system.cpu* system.ram* system.net* disk_space.* !*
send hosts matching = localhost *
send names instead of ids = yes
```
This is the only config which I have changed when testing this behaviour
##### Expected behavior
I expect that I do not see gaps in the dashboard when using the new style.
It could be that I have configured the new one in a wrong way, in such a case I would like to hear any suggestions.
|
non_defect
|
new exporting module is missing metrics compared to old when sending to graphite bug report summary i have migrated the exporting config from the old way using to the new way by using exporting conf however there is something strange or i have put wrong config because we do not get all the metrics os environment running netdata on docker together with grafana latest in docker and graphite also in docker i am running them on a macos host but i have also seen missing stats on our ubuntu servers component name the metrics send to graphite have gaps which did not happen before i have made a test by running a single netdata with the new config and old config the yellow and green uptime are when i run netdata with the new exporting config the red and blue uptime are when i run netdata with the old backend config steps to reproduce run netdata with new backend and old backend and send to graphite here is my config enabled yes type graphite destination docker for mac localhost data source as collected prefix netdata update every buffer on failures timeout ms send charts matching system uptime system cpu system ram system net disk space send hosts matching localhost send names instead of ids yes here is my exporting conf file enabled yes destination docker for mac localhost data source as collected prefix netdata update every buffer on failures timeout ms send charts matching system uptime system cpu system ram system net disk space send hosts matching localhost send names instead of ids yes this is the only config which i have changed when testing this behaviour expected behavior i expect that i do not see gaps in the dashboard when using the new style it could be that i have configured the new one in a wrong way in such a case i would like to hear any suggestions
| 0
|
1,518
| 2,603,966,891
|
IssuesEvent
|
2015-02-24 18:59:15
|
chrsmith/nishazi6
|
https://api.github.com/repos/chrsmith/nishazi6
|
opened
|
沈阳冠状沟长肉芽
|
auto-migrated Priority-Medium Type-Defect
|
```
沈阳冠状沟长肉芽〓沈陽軍區政治部醫院性病〓TEL:024-3102330
8〓成立于1946年,68年專注于性傳播疾病的研究和治療。位于�
��陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷�
��悠久、設備精良、技術權威、專家云集,是預防、保健、醫
療、科研康復為一體的綜合性醫院。是國家首批公立甲等部��
�醫院、全國首批醫療規范定點單位,是第四軍醫大學、東南�
��學等知名高等院校的教學醫院。曾被中國人民解放軍空軍后
勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二等��
�。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:07
|
1.0
|
沈阳冠状沟长肉芽 - ```
沈阳冠状沟长肉芽〓沈陽軍區政治部醫院性病〓TEL:024-3102330
8〓成立于1946年,68年專注于性傳播疾病的研究和治療。位于�
��陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷�
��悠久、設備精良、技術權威、專家云集,是預防、保健、醫
療、科研康復為一體的綜合性醫院。是國家首批公立甲等部��
�醫院、全國首批醫療規范定點單位,是第四軍醫大學、東南�
��學等知名高等院校的教學醫院。曾被中國人民解放軍空軍后
勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二等��
�。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:07
|
defect
|
沈阳冠状沟长肉芽 沈阳冠状沟长肉芽〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位于� �� 。是一所與新中國同建立共輝煌的歷� ��悠久、設備精良、技術權威、專家云集,是預防、保健、醫 療、科研康復為一體的綜合性醫院。是國家首批公立甲等部�� �醫院、全國首批醫療規范定點單位,是第四軍醫大學、東南� ��學等知名高等院校的教學醫院。曾被中國人民解放軍空軍后 勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二等�� �。 original issue reported on code google com by gmail com on jun at
| 1
|
126,854
| 5,006,746,674
|
IssuesEvent
|
2016-12-12 15:01:12
|
storyteller/Storyteller
|
https://api.github.com/repos/storyteller/Storyteller
|
opened
|
Server should respond to set-breakpoints and remember for later
|
low priority
|
Just so you can more easily refresh the screen.
|
1.0
|
Server should respond to set-breakpoints and remember for later - Just so you can more easily refresh the screen.
|
non_defect
|
server should respond to set breakpoints and remember for later just so you can more easily refresh the screen
| 0
|
46,521
| 13,055,926,240
|
IssuesEvent
|
2020-07-30 03:08:28
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
[cmake] toplevel-parasite tool detection is backwards (Trac #1358)
|
Incomplete Migration Migrated from Trac cmake defect
|
Migrated from https://code.icecube.wisc.edu/ticket/1358
```json
{
"status": "closed",
"changetime": "2019-01-12T00:06:17",
"description": "The parasitic build system has no explicit tool detection phase. Instead, the parasitic version of `i3_project()` loads tool-related variables from the parent CMakeCache.txt based on the values passed to `USE_TOOLS`. Some projects with uncommon dependencies such as `astro`, however, wrap the body of their CMakeLists.txt in an `if (${TOOL}_FOUND)` clause, which will fail because the corresponding variable is not defined until explicitly requested.\n\nTwo solutions come to mind:\n1. Eagerly load everything from the parent CMakeCache.txt, filtering out tool-related variables by looking for names that share a prefix with variables of the form `${TOOL}_FOUND`. This is icky but simple.\n2. Replace the `if (${TOOL}_FOUND)` pattern with a function call, e.g. `if (tool_found(${TOOL}))`. This is more work and involves touching multiple projects, but is probably clearer in the long run.",
"reporter": "jvansanten",
"cc": "nega, david.schultz",
"resolution": "fixed",
"_ts": "1547251577858784",
"component": "cmake",
"summary": "[cmake] toplevel-parasite tool detection is backwards",
"priority": "major",
"keywords": "",
"time": "2015-09-21T12:36:01",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
|
1.0
|
[cmake] toplevel-parasite tool detection is backwards (Trac #1358) - Migrated from https://code.icecube.wisc.edu/ticket/1358
```json
{
"status": "closed",
"changetime": "2019-01-12T00:06:17",
"description": "The parasitic build system has no explicit tool detection phase. Instead, the parasitic version of `i3_project()` loads tool-related variables from the parent CMakeCache.txt based on the values passed to `USE_TOOLS`. Some projects with uncommon dependencies such as `astro`, however, wrap the body of their CMakeLists.txt in an `if (${TOOL}_FOUND)` clause, which will fail because the corresponding variable is not defined until explicitly requested.\n\nTwo solutions come to mind:\n1. Eagerly load everything from the parent CMakeCache.txt, filtering out tool-related variables by looking for names that share a prefix with variables of the form `${TOOL}_FOUND`. This is icky but simple.\n2. Replace the `if (${TOOL}_FOUND)` pattern with a function call, e.g. `if (tool_found(${TOOL}))`. This is more work and involves touching multiple projects, but is probably clearer in the long run.",
"reporter": "jvansanten",
"cc": "nega, david.schultz",
"resolution": "fixed",
"_ts": "1547251577858784",
"component": "cmake",
"summary": "[cmake] toplevel-parasite tool detection is backwards",
"priority": "major",
"keywords": "",
"time": "2015-09-21T12:36:01",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
|
defect
|
toplevel parasite tool detection is backwards trac migrated from json status closed changetime description the parasitic build system has no explicit tool detection phase instead the parasitic version of project loads tool related variables from the parent cmakecache txt based on the values passed to use tools some projects with uncommon dependencies such as astro however wrap the body of their cmakelists txt in an if tool found clause which will fail because the corresponding variable is not defined until explicitly requested n ntwo solutions come to mind eagerly load everything from the parent cmakecache txt filtering out tool related variables by looking for names that share a prefix with variables of the form tool found this is icky but simple replace the if tool found pattern with a function call e g if tool found tool this is more work and involves touching multiple projects but is probably clearer in the long run reporter jvansanten cc nega david schultz resolution fixed ts component cmake summary toplevel parasite tool detection is backwards priority major keywords time milestone owner jvansanten type defect
| 1
|
44,329
| 12,101,450,515
|
IssuesEvent
|
2020-04-20 15:13:55
|
codesmithtools/Templates
|
https://api.github.com/repos/codesmithtools/Templates
|
closed
|
possibility to force join tables being entities from the GUI
|
Framework-NHibernate Type-Defect auto-migrated
|
```
What steps will reproduce the problem?
1. more details here:
http://community.codesmithtools.com/forums/t/10459.aspx
What is the expected output? What do you see instead?
A table should be generated as an entity instead of being simply not
generated at all when the table is a join table linked to another join
table.
What version of the product are you using?
5.2.0
Please provide any additional information below.
Maybe you can add to the csp file the option to specify wich tables should
be entities. In the NHibernateMaster.cst output edition mode, we could have
a system similar to "excludedTables" called "entityForcedTables" where we
could select which tables to force being entities.... which is the same as
adding extended properties to tables without needing to modify the
database.
```
Original issue reported on code.google.com by `metash...@gmail.com` on 13 Jan 2010 at 4:21
|
1.0
|
possibility to force join tables being entities from the GUI - ```
What steps will reproduce the problem?
1. more details here:
http://community.codesmithtools.com/forums/t/10459.aspx
What is the expected output? What do you see instead?
A table should be generated as an entity instead of being simply not
generated at all when the table is a join table linked to another join
table.
What version of the product are you using?
5.2.0
Please provide any additional information below.
Maybe you can add to the csp file the option to specify wich tables should
be entities. In the NHibernateMaster.cst output edition mode, we could have
a system similar to "excludedTables" called "entityForcedTables" where we
could select which tables to force being entities.... which is the same as
adding extended properties to tables without needing to modify the
database.
```
Original issue reported on code.google.com by `metash...@gmail.com` on 13 Jan 2010 at 4:21
|
defect
|
possibility to force join tables being entities from the gui what steps will reproduce the problem more details here what is the expected output what do you see instead a table should be generated as an entity instead of being simply not generated at all when the table is a join table linked to another join table what version of the product are you using please provide any additional information below maybe you can add to the csp file the option to specify wich tables should be entities in the nhibernatemaster cst output edition mode we could have a system similar to excludedtables called entityforcedtables where we could select which tables to force being entities which is the same as adding extended properties to tables without needing to modify the database original issue reported on code google com by metash gmail com on jan at
| 1
|
22,741
| 11,726,647,468
|
IssuesEvent
|
2020-03-10 14:48:59
|
golang/go
|
https://api.github.com/repos/golang/go
|
closed
|
cmd/compile: favour UDIV over UMULH + LSR on arm64 for 64 bit integer division by a constant
|
NeedsInvestigation Performance
|
### What version of Go are you using (`go version`)?
<pre>
$ go version
go version go1.14 darwin/amd64 # but (cross compiling to linux/arm64)
</pre>
### Does this issue reproduce with the latest release?
Yes
### What operating system and processor architecture are you using (`go env`)?
<details><summary><code>go env</code> Output</summary><br><pre>
$ go env
GO111MODULE=""
GOARCH="amd64"
GOBIN=""
GOCACHE="/Users/pmoore/Library/Caches/go-build"
GOENV="/Users/pmoore/Library/Application Support/go/env"
GOEXE=""
GOFLAGS=""
GOHOSTARCH="amd64"
GOHOSTOS="darwin"
GOINSECURE=""
GONOPROXY=""
GONOSUMDB=""
GOOS="darwin"
GOPATH="/Users/pmoore/.gvm/pkgsets/go1.14/global"
GOPRIVATE=""
GOPROXY="https://proxy.golang.org,direct"
GOROOT="/Users/pmoore/.gvm/gos/go1.14"
GOSUMDB="sum.golang.org"
GOTMPDIR=""
GOTOOLDIR="/Users/pmoore/.gvm/gos/go1.14/pkg/tool/darwin_amd64"
GCCGO="gccgo"
AR="ar"
CC="clang"
CXX="clang++"
CGO_ENABLED="1"
GOMOD="/Users/pmoore/git/go/src/go.mod"
CGO_CFLAGS="-g -O2"
CGO_CPPFLAGS=""
CGO_CXXFLAGS="-g -O2"
CGO_FFLAGS="-g -O2"
CGO_LDFLAGS="-g -O2"
PKG_CONFIG="pkg-config"
GOGCCFLAGS="-fPIC -m64 -pthread -fno-caret-diagnostics -Qunused-arguments -fmessage-length=0 -fdebug-prefix-map=/var/folders/v9/mll6p_rj5h94dt_m5m8j0f9c0000gn/T/go-build535314509=/tmp/go-build -gno-record-gcc-switches -fno-common"
</pre></details>
### What did you do?
```
pmoore@Petes-iMac:~/udiv $ cat udiv.go
package main
import (
"fmt"
)
func main() {
for x := uint64(0); x < 123456; x++ {
fmt.Printf("%v\n", x/216)
}
}
pmoore@Petes-iMac:~/udiv $ GOARCH=arm64 GOOS=linux go build -gcflags -S udiv.go 2>&1 | grep -F udiv.go:9
0x0028 00040 (/Users/pmoore/udiv/udiv.go:9) MOVD $-7515340178177965473, R1
0x0038 00056 (/Users/pmoore/udiv/udiv.go:9) UMULH R0, R1, R2
0x003c 00060 (/Users/pmoore/udiv/udiv.go:9) LSR $7, R2, R2
0x0040 00064 (/Users/pmoore/udiv/udiv.go:9) MOVD R2, 8(RSP)
0x0044 00068 (/Users/pmoore/udiv/udiv.go:9) CALL runtime.convT64(SB)
0x0048 00072 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, $1
0x0048 00072 (/Users/pmoore/udiv/udiv.go:9) MOVD 16(RSP), R0
0x004c 00076 (/Users/pmoore/udiv/udiv.go:9) PCDATA $1, $1
0x004c 00076 (/Users/pmoore/udiv/udiv.go:9) STP (ZR, ZR), ""..autotmp_15-16(SP)
0x0050 00080 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, $2
0x0050 00080 (/Users/pmoore/udiv/udiv.go:9) MOVD $type.uint64(SB), R1
0x0058 00088 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, $1
0x0058 00088 (/Users/pmoore/udiv/udiv.go:9) MOVD R1, ""..autotmp_15-16(SP)
0x005c 00092 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, ZR
0x005c 00092 (/Users/pmoore/udiv/udiv.go:9) MOVD R0, ""..autotmp_15-8(SP)
pmoore@Petes-iMac:~/udiv $
```
### What did you expect to see?
I expected the compiler to generate the UDIV instruction rather than `UMULH` + `LSR`, as empirical testing (below) shows it to be twice as fast on an arm64 Cortex A53 (BCM2837) SoC.
The final assembly instructions are:
```
pmoore@Petes-iMac:~/udiv $ aarch64-unknown-linux-gnu-objdump -d udiv | sed -n '/\<main\.main\>/,$p'
000000000009bf50 <main.main>:
9bf50: f9400b81 ldr x1, [x28, #16]
9bf54: 910003e2 mov x2, sp
9bf58: eb01005f cmp x2, x1
9bf5c: 54000609 b.ls 9c01c <main.main+0xcc> // b.plast
9bf60: f8180ffe str x30, [sp, #-128]!
9bf64: f81f83fd stur x29, [sp, #-8]
9bf68: d10023fd sub x29, sp, #0x8
9bf6c: d2800000 mov x0, #0x0 // #0
9bf70: 14000024 b 9c000 <main.main+0xb0>
9bf74: f90033e0 str x0, [sp, #96]
9bf78: d2884be1 mov x1, #0x425f // #16991
9bf7c: f2a12f61 movk x1, #0x97b, lsl #16
9bf80: f2c4bda1 movk x1, #0x25ed, lsl #32
9bf84: f2f2f681 movk x1, #0x97b4, lsl #48
9bf88: 9bc07c22 umulh x2, x1, x0
9bf8c: d347fc42 lsr x2, x2, #7
9bf90: f90007e2 str x2, [sp, #8]
9bf94: 97fdf237 bl 18870 <runtime.convT64>
9bf98: f9400be0 ldr x0, [sp, #16]
9bf9c: a906ffff stp xzr, xzr, [sp, #104]
9bfa0: d0000081 adrp x1, ad000 <type.*+0xd000>
9bfa4: 91348021 add x1, x1, #0xd20
9bfa8: f90037e1 str x1, [sp, #104]
9bfac: f9003be0 str x0, [sp, #112]
9bfb0: d00006db adrp x27, 175000 <runtime.itabTableInit+0xee0>
9bfb4: 9106237b add x27, x27, #0x188
9bfb8: f9400360 ldr x0, [x27]
9bfbc: f0000262 adrp x2, ea000 <runtime.vdsoauxv.stkobj>
9bfc0: 91060042 add x2, x2, #0x180
9bfc4: f90007e2 str x2, [sp, #8]
9bfc8: f9000be0 str x0, [sp, #16]
9bfcc: d0000180 adrp x0, cd000 <type.*+0x2d000>
9bfd0: 91300800 add x0, x0, #0xc02
9bfd4: f9000fe0 str x0, [sp, #24]
9bfd8: b24007e3 orr x3, xzr, #0x3
9bfdc: f90013e3 str x3, [sp, #32]
9bfe0: 9101a3e4 add x4, sp, #0x68
9bfe4: f90017e4 str x4, [sp, #40]
9bfe8: b24003e4 orr x4, xzr, #0x1
9bfec: f9001be4 str x4, [sp, #48]
9bff0: f9001fe4 str x4, [sp, #56]
9bff4: 97ffe363 bl 94d80 <fmt.Fprintf>
9bff8: f94033e0 ldr x0, [sp, #96]
9bffc: 91000400 add x0, x0, #0x1
9c000: d29c481b mov x27, #0xe240 // #57920
9c004: f2a0003b movk x27, #0x1, lsl #16
9c008: eb1b001f cmp x0, x27
9c00c: 54fffb43 b.cc 9bf74 <main.main+0x24> // b.lo, b.ul, b.last
9c010: f85f83fd ldur x29, [sp, #-8]
9c014: f84807fe ldr x30, [sp], #128
9c018: d65f03c0 ret
9c01c: aa1e03e3 mov x3, x30
9c020: 97ff2a48 bl 66940 <runtime.morestack_noctxt>
9c024: 17ffffcb b 9bf50 <main.main>
...
```
The division by 216 is comprised of the six instructions:
```
9bf78: d2884be1 mov x1, #0x425f // #16991
9bf7c: f2a12f61 movk x1, #0x97b, lsl #16
9bf80: f2c4bda1 movk x1, #0x25ed, lsl #32
9bf84: f2f2f681 movk x1, #0x97b4, lsl #48
9bf88: 9bc07c22 umulh x2, x1, x0
9bf8c: d347fc42 lsr x2, x2, #7
```
### Benchmark results
I created two simple executables to compare the generated `mov/movk/movk/movk/umulh/lsr` instructions with the equivalent `mov/udiv` instructions that I had expected to see, and compared results. Here we see that `mov/udiv` is consistently twice as fast:
```
ubuntu@ubuntu:~/git/udiv$ cat umulh+lsr.s
/*
Benchmark with: as -o umulh+lsr.o umulh+lsr.s && gcc -o umulh+lsr umulh+lsr.o && for ((i=0; i<5; i++)); do time ./umulh+lsr; done
*/
.global main
.data
fmt:
.asciz "%d / %d = %d\n"
.text
main:
mov x1, #0xec73
movk x1, #0x0018, lsl #16
mov x2, #216
mov x11, #0x10000000
1:
mov x4, #0x425f
movk x4, #0x97b, lsl #16
movk x4, #0x25ed, lsl #32
movk x4, #0x97b4, lsl #48
umulh x3, x4, x1
lsr x3, x3, #7
subs x11, x11, #1
b.ne 1b
ldr x0,=fmt
bl printf
mov x8, #93
svc 0
ubuntu@ubuntu:~/git/udiv$ as -o umulh+lsr.o umulh+lsr.s && gcc -o umulh+lsr umulh+lsr.o && for ((i=0; i<5; i++)); do time ./umulh+lsr; done
1633395 / 216 = 7562
real 0m2.720s
user 0m2.713s
sys 0m0.004s
1633395 / 216 = 7562
real 0m2.713s
user 0m2.706s
sys 0m0.004s
1633395 / 216 = 7562
real 0m2.713s
user 0m2.706s
sys 0m0.005s
1633395 / 216 = 7562
real 0m2.713s
user 0m2.710s
sys 0m0.001s
1633395 / 216 = 7562
real 0m2.713s
user 0m2.709s
sys 0m0.001s
ubuntu@ubuntu:~/git/udiv$ cat udiv.s
/*
Benchmark with: as -o udiv.o udiv.s && gcc -o udiv udiv.o && for ((i=0; i<5; i++)); do time ./udiv; done
*/
.global main
.data
fmt:
.asciz "%d / %d = %d\n"
.text
main:
mov x1, #0xec73
movk x1, #0x0018, lsl #16
mov x2, #216
mov x11, #0x10000000
1:
mov x4, #216
udiv x3, x1, x4
subs x11, x11, #1
b.ne 1b
ldr x0,=fmt
bl printf
mov x8, #93
svc 0
ubuntu@ubuntu:~/git/udiv$ as -o udiv.o udiv.s && gcc -o udiv udiv.o && for ((i=0; i<5; i++)); do time ./udiv; done
1633395 / 216 = 7562
real 0m1.358s
user 0m1.357s
sys 0m0.000s
1633395 / 216 = 7562
real 0m1.358s
user 0m1.353s
sys 0m0.004s
1633395 / 216 = 7562
real 0m1.358s
user 0m1.357s
sys 0m0.000s
1633395 / 216 = 7562
real 0m1.358s
user 0m1.352s
sys 0m0.005s
1633395 / 216 = 7562
real 0m1.358s
user 0m1.352s
sys 0m0.005s
ubuntu@ubuntu:~/git/udiv$
```
|
True
|
cmd/compile: favour UDIV over UMULH + LSR on arm64 for 64 bit integer division by a constant - ### What version of Go are you using (`go version`)?
<pre>
$ go version
go version go1.14 darwin/amd64 # but (cross compiling to linux/arm64)
</pre>
### Does this issue reproduce with the latest release?
Yes
### What operating system and processor architecture are you using (`go env`)?
<details><summary><code>go env</code> Output</summary><br><pre>
$ go env
GO111MODULE=""
GOARCH="amd64"
GOBIN=""
GOCACHE="/Users/pmoore/Library/Caches/go-build"
GOENV="/Users/pmoore/Library/Application Support/go/env"
GOEXE=""
GOFLAGS=""
GOHOSTARCH="amd64"
GOHOSTOS="darwin"
GOINSECURE=""
GONOPROXY=""
GONOSUMDB=""
GOOS="darwin"
GOPATH="/Users/pmoore/.gvm/pkgsets/go1.14/global"
GOPRIVATE=""
GOPROXY="https://proxy.golang.org,direct"
GOROOT="/Users/pmoore/.gvm/gos/go1.14"
GOSUMDB="sum.golang.org"
GOTMPDIR=""
GOTOOLDIR="/Users/pmoore/.gvm/gos/go1.14/pkg/tool/darwin_amd64"
GCCGO="gccgo"
AR="ar"
CC="clang"
CXX="clang++"
CGO_ENABLED="1"
GOMOD="/Users/pmoore/git/go/src/go.mod"
CGO_CFLAGS="-g -O2"
CGO_CPPFLAGS=""
CGO_CXXFLAGS="-g -O2"
CGO_FFLAGS="-g -O2"
CGO_LDFLAGS="-g -O2"
PKG_CONFIG="pkg-config"
GOGCCFLAGS="-fPIC -m64 -pthread -fno-caret-diagnostics -Qunused-arguments -fmessage-length=0 -fdebug-prefix-map=/var/folders/v9/mll6p_rj5h94dt_m5m8j0f9c0000gn/T/go-build535314509=/tmp/go-build -gno-record-gcc-switches -fno-common"
</pre></details>
### What did you do?
```
pmoore@Petes-iMac:~/udiv $ cat udiv.go
package main
import (
"fmt"
)
func main() {
for x := uint64(0); x < 123456; x++ {
fmt.Printf("%v\n", x/216)
}
}
pmoore@Petes-iMac:~/udiv $ GOARCH=arm64 GOOS=linux go build -gcflags -S udiv.go 2>&1 | grep -F udiv.go:9
0x0028 00040 (/Users/pmoore/udiv/udiv.go:9) MOVD $-7515340178177965473, R1
0x0038 00056 (/Users/pmoore/udiv/udiv.go:9) UMULH R0, R1, R2
0x003c 00060 (/Users/pmoore/udiv/udiv.go:9) LSR $7, R2, R2
0x0040 00064 (/Users/pmoore/udiv/udiv.go:9) MOVD R2, 8(RSP)
0x0044 00068 (/Users/pmoore/udiv/udiv.go:9) CALL runtime.convT64(SB)
0x0048 00072 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, $1
0x0048 00072 (/Users/pmoore/udiv/udiv.go:9) MOVD 16(RSP), R0
0x004c 00076 (/Users/pmoore/udiv/udiv.go:9) PCDATA $1, $1
0x004c 00076 (/Users/pmoore/udiv/udiv.go:9) STP (ZR, ZR), ""..autotmp_15-16(SP)
0x0050 00080 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, $2
0x0050 00080 (/Users/pmoore/udiv/udiv.go:9) MOVD $type.uint64(SB), R1
0x0058 00088 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, $1
0x0058 00088 (/Users/pmoore/udiv/udiv.go:9) MOVD R1, ""..autotmp_15-16(SP)
0x005c 00092 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, ZR
0x005c 00092 (/Users/pmoore/udiv/udiv.go:9) MOVD R0, ""..autotmp_15-8(SP)
pmoore@Petes-iMac:~/udiv $
```
### What did you expect to see?
I expected the compiler to generate the UDIV instruction rather than `UMULH` + `LSR`, as empirical testing (below) shows it to be twice as fast on an arm64 Cortex A53 (BCM2837) SoC.
The final assembly instructions are:
```
pmoore@Petes-iMac:~/udiv $ aarch64-unknown-linux-gnu-objdump -d udiv | sed -n '/\<main\.main\>/,$p'
000000000009bf50 <main.main>:
9bf50: f9400b81 ldr x1, [x28, #16]
9bf54: 910003e2 mov x2, sp
9bf58: eb01005f cmp x2, x1
9bf5c: 54000609 b.ls 9c01c <main.main+0xcc> // b.plast
9bf60: f8180ffe str x30, [sp, #-128]!
9bf64: f81f83fd stur x29, [sp, #-8]
9bf68: d10023fd sub x29, sp, #0x8
9bf6c: d2800000 mov x0, #0x0 // #0
9bf70: 14000024 b 9c000 <main.main+0xb0>
9bf74: f90033e0 str x0, [sp, #96]
9bf78: d2884be1 mov x1, #0x425f // #16991
9bf7c: f2a12f61 movk x1, #0x97b, lsl #16
9bf80: f2c4bda1 movk x1, #0x25ed, lsl #32
9bf84: f2f2f681 movk x1, #0x97b4, lsl #48
9bf88: 9bc07c22 umulh x2, x1, x0
9bf8c: d347fc42 lsr x2, x2, #7
9bf90: f90007e2 str x2, [sp, #8]
9bf94: 97fdf237 bl 18870 <runtime.convT64>
9bf98: f9400be0 ldr x0, [sp, #16]
9bf9c: a906ffff stp xzr, xzr, [sp, #104]
9bfa0: d0000081 adrp x1, ad000 <type.*+0xd000>
9bfa4: 91348021 add x1, x1, #0xd20
9bfa8: f90037e1 str x1, [sp, #104]
9bfac: f9003be0 str x0, [sp, #112]
9bfb0: d00006db adrp x27, 175000 <runtime.itabTableInit+0xee0>
9bfb4: 9106237b add x27, x27, #0x188
9bfb8: f9400360 ldr x0, [x27]
9bfbc: f0000262 adrp x2, ea000 <runtime.vdsoauxv.stkobj>
9bfc0: 91060042 add x2, x2, #0x180
9bfc4: f90007e2 str x2, [sp, #8]
9bfc8: f9000be0 str x0, [sp, #16]
9bfcc: d0000180 adrp x0, cd000 <type.*+0x2d000>
9bfd0: 91300800 add x0, x0, #0xc02
9bfd4: f9000fe0 str x0, [sp, #24]
9bfd8: b24007e3 orr x3, xzr, #0x3
9bfdc: f90013e3 str x3, [sp, #32]
9bfe0: 9101a3e4 add x4, sp, #0x68
9bfe4: f90017e4 str x4, [sp, #40]
9bfe8: b24003e4 orr x4, xzr, #0x1
9bfec: f9001be4 str x4, [sp, #48]
9bff0: f9001fe4 str x4, [sp, #56]
9bff4: 97ffe363 bl 94d80 <fmt.Fprintf>
9bff8: f94033e0 ldr x0, [sp, #96]
9bffc: 91000400 add x0, x0, #0x1
9c000: d29c481b mov x27, #0xe240 // #57920
9c004: f2a0003b movk x27, #0x1, lsl #16
9c008: eb1b001f cmp x0, x27
9c00c: 54fffb43 b.cc 9bf74 <main.main+0x24> // b.lo, b.ul, b.last
9c010: f85f83fd ldur x29, [sp, #-8]
9c014: f84807fe ldr x30, [sp], #128
9c018: d65f03c0 ret
9c01c: aa1e03e3 mov x3, x30
9c020: 97ff2a48 bl 66940 <runtime.morestack_noctxt>
9c024: 17ffffcb b 9bf50 <main.main>
...
```
The division by 216 is comprised of the six instructions:
```
9bf78: d2884be1 mov x1, #0x425f // #16991
9bf7c: f2a12f61 movk x1, #0x97b, lsl #16
9bf80: f2c4bda1 movk x1, #0x25ed, lsl #32
9bf84: f2f2f681 movk x1, #0x97b4, lsl #48
9bf88: 9bc07c22 umulh x2, x1, x0
9bf8c: d347fc42 lsr x2, x2, #7
```
### Benchmark results
I created two simple executables to compare the generated `mov/movk/movk/movk/umulh/lsr` instructions with the equivalent `mov/udiv` instructions that I had expected to see, and compared results. Here we see that `mov/udiv` is consistently twice as fast:
```
ubuntu@ubuntu:~/git/udiv$ cat umulh+lsr.s
/*
Benchmark with: as -o umulh+lsr.o umulh+lsr.s && gcc -o umulh+lsr umulh+lsr.o && for ((i=0; i<5; i++)); do time ./umulh+lsr; done
*/
.global main
.data
fmt:
.asciz "%d / %d = %d\n"
.text
main:
mov x1, #0xec73
movk x1, #0x0018, lsl #16
mov x2, #216
mov x11, #0x10000000
1:
mov x4, #0x425f
movk x4, #0x97b, lsl #16
movk x4, #0x25ed, lsl #32
movk x4, #0x97b4, lsl #48
umulh x3, x4, x1
lsr x3, x3, #7
subs x11, x11, #1
b.ne 1b
ldr x0,=fmt
bl printf
mov x8, #93
svc 0
ubuntu@ubuntu:~/git/udiv$ as -o umulh+lsr.o umulh+lsr.s && gcc -o umulh+lsr umulh+lsr.o && for ((i=0; i<5; i++)); do time ./umulh+lsr; done
1633395 / 216 = 7562
real 0m2.720s
user 0m2.713s
sys 0m0.004s
1633395 / 216 = 7562
real 0m2.713s
user 0m2.706s
sys 0m0.004s
1633395 / 216 = 7562
real 0m2.713s
user 0m2.706s
sys 0m0.005s
1633395 / 216 = 7562
real 0m2.713s
user 0m2.710s
sys 0m0.001s
1633395 / 216 = 7562
real 0m2.713s
user 0m2.709s
sys 0m0.001s
ubuntu@ubuntu:~/git/udiv$ cat udiv.s
/*
Benchmark with: as -o udiv.o udiv.s && gcc -o udiv udiv.o && for ((i=0; i<5; i++)); do time ./udiv; done
*/
.global main
.data
fmt:
.asciz "%d / %d = %d\n"
.text
main:
mov x1, #0xec73
movk x1, #0x0018, lsl #16
mov x2, #216
mov x11, #0x10000000
1:
mov x4, #216
udiv x3, x1, x4
subs x11, x11, #1
b.ne 1b
ldr x0,=fmt
bl printf
mov x8, #93
svc 0
ubuntu@ubuntu:~/git/udiv$ as -o udiv.o udiv.s && gcc -o udiv udiv.o && for ((i=0; i<5; i++)); do time ./udiv; done
1633395 / 216 = 7562
real 0m1.358s
user 0m1.357s
sys 0m0.000s
1633395 / 216 = 7562
real 0m1.358s
user 0m1.353s
sys 0m0.004s
1633395 / 216 = 7562
real 0m1.358s
user 0m1.357s
sys 0m0.000s
1633395 / 216 = 7562
real 0m1.358s
user 0m1.352s
sys 0m0.005s
1633395 / 216 = 7562
real 0m1.358s
user 0m1.352s
sys 0m0.005s
ubuntu@ubuntu:~/git/udiv$
```
|
non_defect
|
cmd compile favour udiv over umulh lsr on for bit integer division by a constant what version of go are you using go version go version go version darwin but cross compiling to linux does this issue reproduce with the latest release yes what operating system and processor architecture are you using go env go env output go env goarch gobin gocache users pmoore library caches go build goenv users pmoore library application support go env goexe goflags gohostarch gohostos darwin goinsecure gonoproxy gonosumdb goos darwin gopath users pmoore gvm pkgsets global goprivate goproxy goroot users pmoore gvm gos gosumdb sum golang org gotmpdir gotooldir users pmoore gvm gos pkg tool darwin gccgo gccgo ar ar cc clang cxx clang cgo enabled gomod users pmoore git go src go mod cgo cflags g cgo cppflags cgo cxxflags g cgo fflags g cgo ldflags g pkg config pkg config gogccflags fpic pthread fno caret diagnostics qunused arguments fmessage length fdebug prefix map var folders t go tmp go build gno record gcc switches fno common what did you do pmoore petes imac udiv cat udiv go package main import fmt func main for x x x fmt printf v n x pmoore petes imac udiv goarch goos linux go build gcflags s udiv go grep f udiv go users pmoore udiv udiv go movd users pmoore udiv udiv go umulh users pmoore udiv udiv go lsr users pmoore udiv udiv go movd rsp users pmoore udiv udiv go call runtime sb users pmoore udiv udiv go pcdata zr users pmoore udiv udiv go movd rsp users pmoore udiv udiv go pcdata users pmoore udiv udiv go stp zr zr autotmp sp users pmoore udiv udiv go pcdata zr users pmoore udiv udiv go movd type sb users pmoore udiv udiv go pcdata zr users pmoore udiv udiv go movd autotmp sp users pmoore udiv udiv go pcdata zr zr users pmoore udiv udiv go movd autotmp sp pmoore petes imac udiv what did you expect to see i expected the compiler to generate the udiv instruction rather than umulh lsr as empirical testing below shows it to be twice as fast on an cortex soc the final assembly instructions are pmoore petes imac udiv unknown linux gnu objdump d udiv sed n p ldr mov sp cmp b ls b plast str stur sub sp mov b str mov movk lsl movk lsl movk lsl umulh lsr str bl ldr stp xzr xzr adrp add str str adrp add ldr adrp add str str adrp add str orr xzr str add sp str orr xzr str str bl ldr add mov movk lsl cmp b cc b lo b ul b last ldur ldr ret mov bl b the division by is comprised of the six instructions mov movk lsl movk lsl movk lsl umulh lsr benchmark results i created two simple executables to compare the generated mov movk movk movk umulh lsr instructions with the equivalent mov udiv instructions that i had expected to see and compared results here we see that mov udiv is consistently twice as fast ubuntu ubuntu git udiv cat umulh lsr s benchmark with as o umulh lsr o umulh lsr s gcc o umulh lsr umulh lsr o for i i i do time umulh lsr done global main data fmt asciz d d d n text main mov movk lsl mov mov mov movk lsl movk lsl movk lsl umulh lsr subs b ne ldr fmt bl printf mov svc ubuntu ubuntu git udiv as o umulh lsr o umulh lsr s gcc o umulh lsr umulh lsr o for i i i do time umulh lsr done real user sys real user sys real user sys real user sys real user sys ubuntu ubuntu git udiv cat udiv s benchmark with as o udiv o udiv s gcc o udiv udiv o for i i i do time udiv done global main data fmt asciz d d d n text main mov movk lsl mov mov mov udiv subs b ne ldr fmt bl printf mov svc ubuntu ubuntu git udiv as o udiv o udiv s gcc o udiv udiv o for i i i do time udiv done real user sys real user sys real user sys real user sys real user sys ubuntu ubuntu git udiv
| 0
|
69,438
| 22,352,446,612
|
IssuesEvent
|
2022-06-15 13:12:15
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
opened
|
Cannot parse quoted database link
|
T: Defect P: Medium E: Professional Edition E: Enterprise Edition C: Parser
|
This parses fine:
```sql
select c from t@link
```
So does this, though there's some weird extra whitespace in the translation output:
```sql
-- input
select c from "t"@link
-- output
select C
from "t" @LINK
```
But this doesn't parse:
```sql
select c from "t"@"link"
```
The output is:
> Unsupported query type: [1:19] select c from "t"@[*]"link"
|
1.0
|
Cannot parse quoted database link - This parses fine:
```sql
select c from t@link
```
So does this, though there's some weird extra whitespace in the translation output:
```sql
-- input
select c from "t"@link
-- output
select C
from "t" @LINK
```
But this doesn't parse:
```sql
select c from "t"@"link"
```
The output is:
> Unsupported query type: [1:19] select c from "t"@[*]"link"
|
defect
|
cannot parse quoted database link this parses fine sql select c from t link so does this though there s some weird extra whitespace in the translation output sql input select c from t link output select c from t link but this doesn t parse sql select c from t link the output is unsupported query type select c from t link
| 1
|
25,179
| 4,232,038,341
|
IssuesEvent
|
2016-07-04 19:39:46
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
reopened
|
slsqp fails to detect infeasible problem
|
defect scipy.optimize
|
From mailing list:
```
x = [1,2]
optimize.minimize(
lambda x: x[0]**2+x[1]**2, x,
constraints = (
{'type':'eq','fun': lambda x: x[0]+x[1]-1},
{'type':'ineq','fun': lambda x: x[0]-2}
),
bounds = ((0,None),(0,None)),
method='SLSQP')
```
returns (0.18.0.dev0+1eeef28)
```
fun: 2.77777777777712
jac: array([ 3.33333334e+00, 2.98023224e-08, 0.00000000e+00])
message: 'Optimization terminated successfully.'
nfev: 8
nit: 6
njev: 2
status: 0
success: True
x: array([ 1.66666667e+00, 1.39888101e-14])
```
|
1.0
|
slsqp fails to detect infeasible problem - From mailing list:
```
x = [1,2]
optimize.minimize(
lambda x: x[0]**2+x[1]**2, x,
constraints = (
{'type':'eq','fun': lambda x: x[0]+x[1]-1},
{'type':'ineq','fun': lambda x: x[0]-2}
),
bounds = ((0,None),(0,None)),
method='SLSQP')
```
returns (0.18.0.dev0+1eeef28)
```
fun: 2.77777777777712
jac: array([ 3.33333334e+00, 2.98023224e-08, 0.00000000e+00])
message: 'Optimization terminated successfully.'
nfev: 8
nit: 6
njev: 2
status: 0
success: True
x: array([ 1.66666667e+00, 1.39888101e-14])
```
|
defect
|
slsqp fails to detect infeasible problem from mailing list x optimize minimize lambda x x x x constraints type eq fun lambda x x x type ineq fun lambda x x bounds none none method slsqp returns fun jac array message optimization terminated successfully nfev nit njev status success true x array
| 1
|
470,630
| 13,541,873,705
|
IssuesEvent
|
2020-09-16 16:28:38
|
hotosm/tasking-manager
|
https://api.github.com/repos/hotosm/tasking-manager
|
opened
|
Define workflows for final data check
|
Component: Frontend Priority: High Status: Needs implementation Type: Enhancement
|
Validators and project managers generally needs to do a final consistency check to evaluate the consistence of the mapping and validation around a batch of tasks, so we need to define some workflows for that.
|
1.0
|
Define workflows for final data check - Validators and project managers generally needs to do a final consistency check to evaluate the consistence of the mapping and validation around a batch of tasks, so we need to define some workflows for that.
|
non_defect
|
define workflows for final data check validators and project managers generally needs to do a final consistency check to evaluate the consistence of the mapping and validation around a batch of tasks so we need to define some workflows for that
| 0
|
51,095
| 13,188,110,710
|
IssuesEvent
|
2020-08-13 05:35:02
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
[icetray] Memory tracking blows the stack on exit (Trac #1973)
|
Migrated from Trac combo core defect
|
If I run this script with memory tracking enabled on SL6 with devtoolset-2:
```text
from icecube import icetray, dataclasses, dataio
from I3Tray import I3Tray
tray = I3Tray()
tray.Add("I3InfiniteSource")
t = icetray.memory.MemoryTimeline()
t.start(10000)
tray.Execute(1)
t.stop
tray.Finish()
```
it runs to completion, and I can do things like plot the memory timeline. Then, after the interpreter exits, it segfaults:
```text
https://code.icecube.wisc.edu/ticket/0 0x00007ffff6e38b80 in _int_malloc () from /lib64/libc.so.6
#1 0x00007ffff6e39b1c in malloc () from /lib64/libc.so.6
#2 0x00007ffff048d3c4 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#3 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#4 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()
from /usr/lib64/libstdc++.so.6
#5 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6
#6 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6
#7 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#8 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#9 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#10 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()
from /usr/lib64/libstdc++.so.6
#11 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6
#12 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6
#13 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#14 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#15 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#16 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()
from /usr/lib64/libstdc++.so.6
```
and so on for 248174 frames.
I suspect that global objects may be getting destroyed in an unexpected order.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1973">https://code.icecube.wisc.edu/ticket/1973</a>, reported by jvansanten and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:44",
"description": "If I run this script with memory tracking enabled on SL6 with devtoolset-2:\n{{{\nfrom icecube import icetray, dataclasses, dataio\nfrom I3Tray import I3Tray\n\ntray = I3Tray()\n\ntray.Add(\"I3InfiniteSource\")\n\nt = icetray.memory.MemoryTimeline()\nt.start(10000)\ntray.Execute(1)\nt.stop\n\ntray.Finish()\n\n}}}\n\nit runs to completion, and I can do things like plot the memory timeline. Then, after the interpreter exits, it segfaults:\n\n{{{\n#0 0x00007ffff6e38b80 in _int_malloc () from /lib64/libc.so.6\n#1 0x00007ffff6e39b1c in malloc () from /lib64/libc.so.6\n#2 0x00007ffff048d3c4 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#3 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#4 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()\n from /usr/lib64/libstdc++.so.6\n#5 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6\n#6 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6\n#7 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#8 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#9 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#10 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()\n from /usr/lib64/libstdc++.so.6\n#11 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6\n#12 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6\n#13 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#14 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#15 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#16 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()\n from /usr/lib64/libstdc++.so.6\n}}}\nand so on for 248174 frames.\n\nI suspect that global objects may be getting destroyed in an unexpected order.",
"reporter": "jvansanten",
"cc": "cweaver, david.schultz",
"resolution": "fixed",
"_ts": "1550067284370534",
"component": "combo core",
"summary": "[icetray] Memory tracking blows the stack on exit",
"priority": "normal",
"keywords": "",
"time": "2017-03-23T10:04:27",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[icetray] Memory tracking blows the stack on exit (Trac #1973) - If I run this script with memory tracking enabled on SL6 with devtoolset-2:
```text
from icecube import icetray, dataclasses, dataio
from I3Tray import I3Tray
tray = I3Tray()
tray.Add("I3InfiniteSource")
t = icetray.memory.MemoryTimeline()
t.start(10000)
tray.Execute(1)
t.stop
tray.Finish()
```
it runs to completion, and I can do things like plot the memory timeline. Then, after the interpreter exits, it segfaults:
```text
https://code.icecube.wisc.edu/ticket/0 0x00007ffff6e38b80 in _int_malloc () from /lib64/libc.so.6
#1 0x00007ffff6e39b1c in malloc () from /lib64/libc.so.6
#2 0x00007ffff048d3c4 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#3 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#4 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()
from /usr/lib64/libstdc++.so.6
#5 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6
#6 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6
#7 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#8 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#9 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#10 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()
from /usr/lib64/libstdc++.so.6
#11 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6
#12 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6
#13 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#14 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#15 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so
#16 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()
from /usr/lib64/libstdc++.so.6
```
and so on for 248174 frames.
I suspect that global objects may be getting destroyed in an unexpected order.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1973">https://code.icecube.wisc.edu/ticket/1973</a>, reported by jvansanten and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:44",
"description": "If I run this script with memory tracking enabled on SL6 with devtoolset-2:\n{{{\nfrom icecube import icetray, dataclasses, dataio\nfrom I3Tray import I3Tray\n\ntray = I3Tray()\n\ntray.Add(\"I3InfiniteSource\")\n\nt = icetray.memory.MemoryTimeline()\nt.start(10000)\ntray.Execute(1)\nt.stop\n\ntray.Finish()\n\n}}}\n\nit runs to completion, and I can do things like plot the memory timeline. Then, after the interpreter exits, it segfaults:\n\n{{{\n#0 0x00007ffff6e38b80 in _int_malloc () from /lib64/libc.so.6\n#1 0x00007ffff6e39b1c in malloc () from /lib64/libc.so.6\n#2 0x00007ffff048d3c4 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#3 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#4 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()\n from /usr/lib64/libstdc++.so.6\n#5 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6\n#6 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6\n#7 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#8 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#9 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#10 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()\n from /usr/lib64/libstdc++.so.6\n#11 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6\n#12 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6\n#13 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#14 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#15 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#16 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()\n from /usr/lib64/libstdc++.so.6\n}}}\nand so on for 248174 frames.\n\nI suspect that global objects may be getting destroyed in an unexpected order.",
"reporter": "jvansanten",
"cc": "cweaver, david.schultz",
"resolution": "fixed",
"_ts": "1550067284370534",
"component": "combo core",
"summary": "[icetray] Memory tracking blows the stack on exit",
"priority": "normal",
"keywords": "",
"time": "2017-03-23T10:04:27",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
defect
|
memory tracking blows the stack on exit trac if i run this script with memory tracking enabled on with devtoolset text from icecube import icetray dataclasses dataio from import tray tray add t icetray memory memorytimeline t start tray execute t stop tray finish it runs to completion and i can do things like plot the memory timeline then after the interpreter exits it segfaults text in int malloc from libc so in malloc from libc so in memory malloc override unsigned long from scratch jvs metaprojects simulation build lib libicetray so in operator new unsigned long from scratch jvs metaprojects simulation build lib libicetray so in std basic string std allocator rep s create unsigned long unsigned long std allocator const from usr libstdc so in from usr libstdc so in std basic string std allocator basic string char const std allocator const from usr libstdc so in boost thread exception thread exception int char const from scratch jvs metaprojects simulation build lib libicetray so in memory malloc override unsigned long from scratch jvs metaprojects simulation build lib libicetray so in operator new unsigned long from scratch jvs metaprojects simulation build lib libicetray so in std basic string std allocator rep s create unsigned long unsigned long std allocator const from usr libstdc so in from usr libstdc so in std basic string std allocator basic string char const std allocator const from usr libstdc so in boost thread exception thread exception int char const from scratch jvs metaprojects simulation build lib libicetray so in memory malloc override unsigned long from scratch jvs metaprojects simulation build lib libicetray so in operator new unsigned long from scratch jvs metaprojects simulation build lib libicetray so in std basic string std allocator rep s create unsigned long unsigned long std allocator const from usr libstdc so and so on for frames i suspect that global objects may be getting destroyed in an unexpected order migrated from json status closed changetime description if i run this script with memory tracking enabled on with devtoolset n nfrom icecube import icetray dataclasses dataio nfrom import n ntray n ntray add n nt icetray memory memorytimeline nt start ntray execute nt stop n ntray finish n n n nit runs to completion and i can do things like plot the memory timeline then after the interpreter exits it segfaults n n n in int malloc from libc so n in malloc from libc so n in memory malloc override unsigned long from scratch jvs metaprojects simulation build lib libicetray so n in operator new unsigned long from scratch jvs metaprojects simulation build lib libicetray so n in std basic string std allocator rep s create unsigned long unsigned long std allocator const n from usr libstdc so n in from usr libstdc so n in std basic string std allocator basic string char const std allocator const from usr libstdc so n in boost thread exception thread exception int char const from scratch jvs metaprojects simulation build lib libicetray so n in memory malloc override unsigned long from scratch jvs metaprojects simulation build lib libicetray so n in operator new unsigned long from scratch jvs metaprojects simulation build lib libicetray so n in std basic string std allocator rep s create unsigned long unsigned long std allocator const n from usr libstdc so n in from usr libstdc so n in std basic string std allocator basic string char const std allocator const from usr libstdc so n in boost thread exception thread exception int char const from scratch jvs metaprojects simulation build lib libicetray so n in memory malloc override unsigned long from scratch jvs metaprojects simulation build lib libicetray so n in operator new unsigned long from scratch jvs metaprojects simulation build lib libicetray so n in std basic string std allocator rep s create unsigned long unsigned long std allocator const n from usr libstdc so n nand so on for frames n ni suspect that global objects may be getting destroyed in an unexpected order reporter jvansanten cc cweaver david schultz resolution fixed ts component combo core summary memory tracking blows the stack on exit priority normal keywords time milestone owner nega type defect
| 1
|
6,796
| 2,610,279,869
|
IssuesEvent
|
2015-02-26 19:29:29
|
chrsmith/scribefire-chrome
|
https://api.github.com/repos/chrsmith/scribefire-chrome
|
closed
|
Import Not Working
|
auto-migrated Priority-Medium Type-Defect
|
```
What's the problem?
Can't import blogs and settings from earlier export. There is a button to
browse to the file but appears to be nothing to begin the import other than an
"X"button
What browser are you using?
Firefox
What version of ScribeFire are you running?
1.7.1
```
-----
Original issue reported on code.google.com by `stme...@gmail.com` on 26 Jun 2011 at 3:07
Attachments:
* [scribefire.tiff](https://storage.googleapis.com/google-code-attachments/scribefire-chrome/issue-371/comment-0/scribefire.tiff)
|
1.0
|
Import Not Working - ```
What's the problem?
Can't import blogs and settings from earlier export. There is a button to
browse to the file but appears to be nothing to begin the import other than an
"X"button
What browser are you using?
Firefox
What version of ScribeFire are you running?
1.7.1
```
-----
Original issue reported on code.google.com by `stme...@gmail.com` on 26 Jun 2011 at 3:07
Attachments:
* [scribefire.tiff](https://storage.googleapis.com/google-code-attachments/scribefire-chrome/issue-371/comment-0/scribefire.tiff)
|
defect
|
import not working what s the problem can t import blogs and settings from earlier export there is a button to browse to the file but appears to be nothing to begin the import other than an x button what browser are you using firefox what version of scribefire are you running original issue reported on code google com by stme gmail com on jun at attachments
| 1
|
19,362
| 3,194,962,709
|
IssuesEvent
|
2015-09-30 14:40:35
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
opened
|
Improper Restriction of XML External Entity References ('XXE') in XMLasDOMBinding
|
C: Functionality P: Medium T: Defect
|
`XMLasDOMBinding` exposes a vulnerability:
**Attack Vector**: javax.xml.parsers.DocumentBuilder.parse
**Description**: The resolution of external entity references is enabled. If attacker-controlled XML can be submitted to the XML parser here, then the attacker could gain access to information about an internal network, local filesystem, or other sensitive data. This is known as an XML eXternal Entity (XXE) attack.
**Remediation**: Configure the XML parser to disable external entity resolution.
**The fix**:
The fix for this issue is as outlined on the OWASP site at this URL:
https://www.owasp.org/index.php/XML_External_Entity_%28XXE%29_Processing
It is recommended adding all of the fixes mentioned, with the resulting code looking similar to the following:
```
DocumentBuilderFactory docFactory = DocumentBuilderFactory.newInstance();
String feature = EMPTY;
try {
// BEGIN FIXES
feature = "http://apache.org/xml/features/disallow-doctype-decl";
docFactory.setFeature(feature, true);
feature = "http://xml.org/sax/features/external-general-entities";
docFactory.setFeature(feature, false);
docFactory.setXIncludeAware(false);
docFactory.setExpandEntityReferences(false);
feature = "http://xml.org/sax/features/external-parameter-entities";
docFactory.setFeature(feature, false);
DocumentBuilder docBuilder = docFactory.newDocumentBuilder();
// END FIXES
...
InputStream is = new ByteArrayInputStream(xml.getBytes());
document_ = docBuilder.parse(is);
...
```
|
1.0
|
Improper Restriction of XML External Entity References ('XXE') in XMLasDOMBinding - `XMLasDOMBinding` exposes a vulnerability:
**Attack Vector**: javax.xml.parsers.DocumentBuilder.parse
**Description**: The resolution of external entity references is enabled. If attacker-controlled XML can be submitted to the XML parser here, then the attacker could gain access to information about an internal network, local filesystem, or other sensitive data. This is known as an XML eXternal Entity (XXE) attack.
**Remediation**: Configure the XML parser to disable external entity resolution.
**The fix**:
The fix for this issue is as outlined on the OWASP site at this URL:
https://www.owasp.org/index.php/XML_External_Entity_%28XXE%29_Processing
It is recommended adding all of the fixes mentioned, with the resulting code looking similar to the following:
```
DocumentBuilderFactory docFactory = DocumentBuilderFactory.newInstance();
String feature = EMPTY;
try {
// BEGIN FIXES
feature = "http://apache.org/xml/features/disallow-doctype-decl";
docFactory.setFeature(feature, true);
feature = "http://xml.org/sax/features/external-general-entities";
docFactory.setFeature(feature, false);
docFactory.setXIncludeAware(false);
docFactory.setExpandEntityReferences(false);
feature = "http://xml.org/sax/features/external-parameter-entities";
docFactory.setFeature(feature, false);
DocumentBuilder docBuilder = docFactory.newDocumentBuilder();
// END FIXES
...
InputStream is = new ByteArrayInputStream(xml.getBytes());
document_ = docBuilder.parse(is);
...
```
|
defect
|
improper restriction of xml external entity references xxe in xmlasdombinding xmlasdombinding exposes a vulnerability attack vector javax xml parsers documentbuilder parse description the resolution of external entity references is enabled if attacker controlled xml can be submitted to the xml parser here then the attacker could gain access to information about an internal network local filesystem or other sensitive data this is known as an xml external entity xxe attack remediation configure the xml parser to disable external entity resolution the fix the fix for this issue is as outlined on the owasp site at this url it is recommended adding all of the fixes mentioned with the resulting code looking similar to the following documentbuilderfactory docfactory documentbuilderfactory newinstance string feature empty try begin fixes feature docfactory setfeature feature true feature docfactory setfeature feature false docfactory setxincludeaware false docfactory setexpandentityreferences false feature docfactory setfeature feature false documentbuilder docbuilder docfactory newdocumentbuilder end fixes inputstream is new bytearrayinputstream xml getbytes document docbuilder parse is
| 1
|
90,469
| 3,816,033,933
|
IssuesEvent
|
2016-03-28 20:07:32
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
opened
|
Add missing description of panels and toolbar
|
Low Priority Text
|
Adding a text under http://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html#panels-and-toolbars section + link from another part toward this chapter.
|
1.0
|
Add missing description of panels and toolbar - Adding a text under http://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html#panels-and-toolbars section + link from another part toward this chapter.
|
non_defect
|
add missing description of panels and toolbar adding a text under section link from another part toward this chapter
| 0
|
67,629
| 21,035,403,238
|
IssuesEvent
|
2022-03-31 07:20:29
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Input new message is sometimes very slow (blocking for some seconds)
|
T-Defect
|
### Steps to reproduce
Type a new message.
### Outcome
#### What did you expect?
The message appears immediately while typing.
#### What happened instead?
The page blocks and the message appears seconds after typing.
The keydown event handler ist blocking for several seconds, caused by getDevicesInRoom(). Please find attached a screenshot of the performance inspector.

### Operating system
Linux
### Browser information
Version 99.0.4844.74 (Official Build) (64-bit)
### URL for webapp
_No response_
### Application version
Element Version: 1.10.7 Version von Olm: 3.2.8
### Homeserver
Synapse 1.55.0
### Will you send logs?
No
|
1.0
|
Input new message is sometimes very slow (blocking for some seconds) - ### Steps to reproduce
Type a new message.
### Outcome
#### What did you expect?
The message appears immediately while typing.
#### What happened instead?
The page blocks and the message appears seconds after typing.
The keydown event handler ist blocking for several seconds, caused by getDevicesInRoom(). Please find attached a screenshot of the performance inspector.

### Operating system
Linux
### Browser information
Version 99.0.4844.74 (Official Build) (64-bit)
### URL for webapp
_No response_
### Application version
Element Version: 1.10.7 Version von Olm: 3.2.8
### Homeserver
Synapse 1.55.0
### Will you send logs?
No
|
defect
|
input new message is sometimes very slow blocking for some seconds steps to reproduce type a new message outcome what did you expect the message appears immediately while typing what happened instead the page blocks and the message appears seconds after typing the keydown event handler ist blocking for several seconds caused by getdevicesinroom please find attached a screenshot of the performance inspector operating system linux browser information version official build bit url for webapp no response application version element version version von olm homeserver synapse will you send logs no
| 1
|
28,375
| 12,834,849,913
|
IssuesEvent
|
2020-07-07 11:50:27
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
kubectl apply -f hello-world-ingress.yaml --namespace ingress-basic
|
Pri2 container-service/svc
|
Hi,
isn't the --namespace ingress-basic missing for the apply ingress command?
so
<== kubectl apply -f hello-world-ingress.yaml
==> kubectl apply -f hello-world-ingress.yaml --namespace ingress-basic
Cheers,
Andre
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: fd0e7d5a-37e9-07dc-f395-9cb5bb580e45
* Version Independent ID: 7e4faf70-9724-7e7b-832b-1cd99a974920
* Content: [Create an ingress controller - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/ingress-basic)
* Content Source: [articles/aks/ingress-basic.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/ingress-basic.md)
* Service: **container-service**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned**
|
1.0
|
kubectl apply -f hello-world-ingress.yaml --namespace ingress-basic -
Hi,
isn't the --namespace ingress-basic missing for the apply ingress command?
so
<== kubectl apply -f hello-world-ingress.yaml
==> kubectl apply -f hello-world-ingress.yaml --namespace ingress-basic
Cheers,
Andre
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: fd0e7d5a-37e9-07dc-f395-9cb5bb580e45
* Version Independent ID: 7e4faf70-9724-7e7b-832b-1cd99a974920
* Content: [Create an ingress controller - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/ingress-basic)
* Content Source: [articles/aks/ingress-basic.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/ingress-basic.md)
* Service: **container-service**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned**
|
non_defect
|
kubectl apply f hello world ingress yaml namespace ingress basic hi isn t the namespace ingress basic missing for the apply ingress command so kubectl apply f hello world ingress yaml kubectl apply f hello world ingress yaml namespace ingress basic cheers andre document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login mlearned microsoft alias mlearned
| 0
|
581,943
| 17,348,392,906
|
IssuesEvent
|
2021-07-29 04:39:23
|
DeFiCh/whale
|
https://api.github.com/repos/DeFiCh/whale
|
closed
|
block.controller.ts
|
area/module-api area/module-database area/whale-api-client kind/feature priority/important-soon triage/accepted
|
<!-- Please only use this template for submitting enhancement/feature requests -->
#### What would you like to be added:
Block Controller for blocks.
`id` can be block hash or block height.
```
GET v1/:network/blocks
GET v1/:network/blocks/:id
GET v1/:network/blocks/:id/transactions
GET v1/:network/blocks/:id/transactions/:txId
GET v1/:network/blocks/:id/transactions/:txId/vins
GET v1/:network/blocks/:id/transactions/:txId/vouts
```
|
1.0
|
block.controller.ts - <!-- Please only use this template for submitting enhancement/feature requests -->
#### What would you like to be added:
Block Controller for blocks.
`id` can be block hash or block height.
```
GET v1/:network/blocks
GET v1/:network/blocks/:id
GET v1/:network/blocks/:id/transactions
GET v1/:network/blocks/:id/transactions/:txId
GET v1/:network/blocks/:id/transactions/:txId/vins
GET v1/:network/blocks/:id/transactions/:txId/vouts
```
|
non_defect
|
block controller ts what would you like to be added block controller for blocks id can be block hash or block height get network blocks get network blocks id get network blocks id transactions get network blocks id transactions txid get network blocks id transactions txid vins get network blocks id transactions txid vouts
| 0
|
33,405
| 7,110,954,868
|
IssuesEvent
|
2018-01-17 12:36:18
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
Context Menu stay when window is resized
|
6.1.12 defect
|
Reported By PRO User;
> p:contextMenu on p:dataTable stay on same place on responsive design when resize or change orientation of Navigator.
Can reproduce it on showcase !
Open contextMenu (right click) and resize window.
It should be better to hide or move it when Window size change.
|
1.0
|
Context Menu stay when window is resized - Reported By PRO User;
> p:contextMenu on p:dataTable stay on same place on responsive design when resize or change orientation of Navigator.
Can reproduce it on showcase !
Open contextMenu (right click) and resize window.
It should be better to hide or move it when Window size change.
|
defect
|
context menu stay when window is resized reported by pro user p contextmenu on p datatable stay on same place on responsive design when resize or change orientation of navigator can reproduce it on showcase open contextmenu right click and resize window it should be better to hide or move it when window size change
| 1
|
808,939
| 30,117,818,010
|
IssuesEvent
|
2023-06-30 12:58:46
|
Three-s-A-Crowd-Games/Gerda
|
https://api.github.com/repos/Three-s-A-Crowd-Games/Gerda
|
opened
|
better damage feedback
|
enhancement high priority
|
We need better an more feedback for the player when he receives damage.
- red flash at the screen border
- screenshake
- ...
also better poisoned feedback:
- more green tint
- green flash at screen border
- ...
|
1.0
|
better damage feedback - We need better an more feedback for the player when he receives damage.
- red flash at the screen border
- screenshake
- ...
also better poisoned feedback:
- more green tint
- green flash at screen border
- ...
|
non_defect
|
better damage feedback we need better an more feedback for the player when he receives damage red flash at the screen border screenshake also better poisoned feedback more green tint green flash at screen border
| 0
|
143,421
| 19,180,298,271
|
IssuesEvent
|
2021-12-04 09:03:50
|
AlexRogalskiy/android-patterns
|
https://api.github.com/repos/AlexRogalskiy/android-patterns
|
opened
|
CVE-2019-1010266 (Medium) detected in lodash-2.4.2.tgz
|
security vulnerability
|
## CVE-2019-1010266 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-2.4.2.tgz</b></p></summary>
<p>A utility library delivering consistency, customization, performance, & extras.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p>
<p>Path to dependency file: android-patterns/package.json</p>
<p>Path to vulnerable library: android-patterns/node_modules/dockerfile_lint/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- dockerfile_lint-0.3.4.tgz (Root Library)
- :x: **lodash-2.4.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/android-patterns/commit/8d0a10402c32fdfadb3cf3eed011045fb6bdc1fb">8d0a10402c32fdfadb3cf3eed011045fb6bdc1fb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.
<p>Publish Date: 2019-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p>
<p>Release Date: 2020-09-30</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-1010266 (Medium) detected in lodash-2.4.2.tgz - ## CVE-2019-1010266 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-2.4.2.tgz</b></p></summary>
<p>A utility library delivering consistency, customization, performance, & extras.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p>
<p>Path to dependency file: android-patterns/package.json</p>
<p>Path to vulnerable library: android-patterns/node_modules/dockerfile_lint/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- dockerfile_lint-0.3.4.tgz (Root Library)
- :x: **lodash-2.4.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/android-patterns/commit/8d0a10402c32fdfadb3cf3eed011045fb6bdc1fb">8d0a10402c32fdfadb3cf3eed011045fb6bdc1fb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11.
<p>Publish Date: 2019-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p>
<p>Release Date: 2020-09-30</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file android patterns package json path to vulnerable library android patterns node modules dockerfile lint node modules lodash package json dependency hierarchy dockerfile lint tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details lodash prior to is affected by cwe uncontrolled resource consumption the impact is denial of service the component is date handler the attack vector is attacker provides very long strings which the library attempts to match using a regular expression the fixed version is publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
72,577
| 7,302,020,508
|
IssuesEvent
|
2018-02-27 08:13:50
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
hack/test-cmd.sh referencing gcr.io/google-containers/perl which cannot be docker pull-ed
|
area/test-infra lifecycle/rotten sig/testing
|
[hack/test-cmd.sh](https://github.com/kubernetes/kubernetes/blob/bfc953cd4b0e1e1d0b0f4d341baa5b5537920f09/hack/test-cmd.sh#L142) is referencing `gcr.io/google-containers/perl` which cannot be found atm.
@erictune fyi
|
2.0
|
hack/test-cmd.sh referencing gcr.io/google-containers/perl which cannot be docker pull-ed - [hack/test-cmd.sh](https://github.com/kubernetes/kubernetes/blob/bfc953cd4b0e1e1d0b0f4d341baa5b5537920f09/hack/test-cmd.sh#L142) is referencing `gcr.io/google-containers/perl` which cannot be found atm.
@erictune fyi
|
non_defect
|
hack test cmd sh referencing gcr io google containers perl which cannot be docker pull ed is referencing gcr io google containers perl which cannot be found atm erictune fyi
| 0
|
61,670
| 17,023,753,018
|
IssuesEvent
|
2021-07-03 03:39:38
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
rpmlint error (no-return-in-nonvoid-function) on building osm2pgsql (r26957)
|
Component: osm2pgsql Priority: critical Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 10.27am, Monday, 24th October 2011]**
Rpmlint blocked to build and deploy osm2pgsql package in openSUSE Build Service with error:
I: Program returns random data in a function
E: osm2pgsql no-return-in-nonvoid-function middle-pgsql.c:167
E: osm2pgsql no-return-in-nonvoid-function output-pgsql.c:1155
|
1.0
|
rpmlint error (no-return-in-nonvoid-function) on building osm2pgsql (r26957) - **[Submitted to the original trac issue database at 10.27am, Monday, 24th October 2011]**
Rpmlint blocked to build and deploy osm2pgsql package in openSUSE Build Service with error:
I: Program returns random data in a function
E: osm2pgsql no-return-in-nonvoid-function middle-pgsql.c:167
E: osm2pgsql no-return-in-nonvoid-function output-pgsql.c:1155
|
defect
|
rpmlint error no return in nonvoid function on building rpmlint blocked to build and deploy package in opensuse build service with error i program returns random data in a function e no return in nonvoid function middle pgsql c e no return in nonvoid function output pgsql c
| 1
|
107,615
| 13,490,917,218
|
IssuesEvent
|
2020-09-11 15:45:45
|
JoscaWij/Spacey
|
https://api.github.com/repos/JoscaWij/Spacey
|
closed
|
Plattform graphic
|
Design User Story
|
# User Story
As a player
I want the plattforms to match the space astetic of the game
# Description
- scalable SVG of a asteroid in comic style
# Material
First Design:

Design-Overview:
Design:

# To do's:
- [x] create SVG in Photoshop
|
1.0
|
Plattform graphic - # User Story
As a player
I want the plattforms to match the space astetic of the game
# Description
- scalable SVG of a asteroid in comic style
# Material
First Design:

Design-Overview:
Design:

# To do's:
- [x] create SVG in Photoshop
|
non_defect
|
plattform graphic user story as a player i want the plattforms to match the space astetic of the game description scalable svg of a asteroid in comic style material first design design overview design to do s create svg in photoshop
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.