Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
218,655
7,332,033,375
IssuesEvent
2018-03-05 15:15:08
enviroCar/enviroCar-app
https://api.github.com/repos/enviroCar/enviroCar-app
closed
Simultaneous upload causes track duplicates on server
Priority - 1 - High bug
The user can hit the "upload all" button and after that upload a single track via a long-click. This serializes the same track twice resulting in a duplication on the server side (as there are no duplication checks). Implement a mutex (maybe using the UploadManager) and block uploading interaction as long as an upload takes place.
1.0
Simultaneous upload causes track duplicates on server - The user can hit the "upload all" button and after that upload a single track via a long-click. This serializes the same track twice resulting in a duplication on the server side (as there are no duplication checks). Implement a mutex (maybe using the UploadManager) and block uploading interaction as long as an upload takes place.
non_defect
simultaneous upload causes track duplicates on server the user can hit the upload all button and after that upload a single track via a long click this serializes the same track twice resulting in a duplication on the server side as there are no duplication checks implement a mutex maybe using the uploadmanager and block uploading interaction as long as an upload takes place
0
39,146
9,218,651,631
IssuesEvent
2019-03-11 13:52:29
STEllAR-GROUP/phylanx
https://api.github.com/repos/STEllAR-GROUP/phylanx
closed
PhySL interpreter inserts 'random' nil into output
category: examples type: defect
I should add that `random` sometimes has `nil` somewhere in its output. In physl: ``` random(make_list(5,10), list("binomial", 3, .5)) ``` produces: ``` [[2, 2, 3nil, 0, 2, 1, 1, 1, 1, 2], [2, 2, 2, 0, 1, 1, 2, 1, 3, 2], [3, 2, 2, 1, 0, 1, 2, 2, 2, 0], [2, 0, 1, 2, 1, 3, 3, 1, 0, 2], [2, 1, 1, 2, 2, 2, 2, 2, 1, 1]] ``` _Originally posted by @taless474 in https://github.com/STEllAR-GROUP/phylanx/issues/827#issuecomment-471065529_
1.0
PhySL interpreter inserts 'random' nil into output - I should add that `random` sometimes has `nil` somewhere in its output. In physl: ``` random(make_list(5,10), list("binomial", 3, .5)) ``` produces: ``` [[2, 2, 3nil, 0, 2, 1, 1, 1, 1, 2], [2, 2, 2, 0, 1, 1, 2, 1, 3, 2], [3, 2, 2, 1, 0, 1, 2, 2, 2, 0], [2, 0, 1, 2, 1, 3, 3, 1, 0, 2], [2, 1, 1, 2, 2, 2, 2, 2, 1, 1]] ``` _Originally posted by @taless474 in https://github.com/STEllAR-GROUP/phylanx/issues/827#issuecomment-471065529_
defect
physl interpreter inserts random nil into output i should add that random sometimes has nil somewhere in its output in physl random make list list binomial produces originally posted by in
1
31,042
6,413,752,441
IssuesEvent
2017-08-08 08:28:41
oleg-shilo/cs-script
https://api.github.com/repos/oleg-shilo/cs-script
closed
Ambiguous call in dbg.inject<ID>
defect Done: waiting for release
While trying to compose a test script environment related to #71 I ran into this issue: ``` <Workspace>\CSScript\Issue #79\Redirect>cscs.exe -cd RedirectRefLib.cs C# Script execution engine. Version 3.27.0.0. Copyright (C) 2004-2017 Oleg Shilo. Error: Specified file could not be compiled. csscript.CompilerException: c:\Users\klaey-1\AppData\Local\Temp\CSSCRIPT\Cache\dbg.inject.-110101375 1.cs(13,16): error CS0121: The call is ambiguous between the following methods or properties: 'dbg_e xtensions.print<T>(T, params object[])' and 'dbg_extensions.print<T>(T, params object[])' at csscript.CSExecutor.ProcessCompilingResult(CompilerResults results, CompilerParameters compile rParams, ScriptParser parser, String scriptFileName, String assemblyFileName, String[] additionalDep endencies) at csscript.CSExecutor.Compile(String scriptFileName) at csscript.CSExecutor.ExecuteImpl() <Workspace>\CSScript\Issue #79\Redirect> ``` Can be reproduced by extracting the attached files into a temporary location, place `cscs.exe` into the `\Redirect` folder and run `cscs.exe -cd RedirectRefLib.cs`. [Issue #79.zip](https://github.com/oleg-shilo/cs-script/files/1162248/Issue.79.zip)
1.0
Ambiguous call in dbg.inject<ID> - While trying to compose a test script environment related to #71 I ran into this issue: ``` <Workspace>\CSScript\Issue #79\Redirect>cscs.exe -cd RedirectRefLib.cs C# Script execution engine. Version 3.27.0.0. Copyright (C) 2004-2017 Oleg Shilo. Error: Specified file could not be compiled. csscript.CompilerException: c:\Users\klaey-1\AppData\Local\Temp\CSSCRIPT\Cache\dbg.inject.-110101375 1.cs(13,16): error CS0121: The call is ambiguous between the following methods or properties: 'dbg_e xtensions.print<T>(T, params object[])' and 'dbg_extensions.print<T>(T, params object[])' at csscript.CSExecutor.ProcessCompilingResult(CompilerResults results, CompilerParameters compile rParams, ScriptParser parser, String scriptFileName, String assemblyFileName, String[] additionalDep endencies) at csscript.CSExecutor.Compile(String scriptFileName) at csscript.CSExecutor.ExecuteImpl() <Workspace>\CSScript\Issue #79\Redirect> ``` Can be reproduced by extracting the attached files into a temporary location, place `cscs.exe` into the `\Redirect` folder and run `cscs.exe -cd RedirectRefLib.cs`. [Issue #79.zip](https://github.com/oleg-shilo/cs-script/files/1162248/Issue.79.zip)
defect
ambiguous call in dbg inject while trying to compose a test script environment related to i ran into this issue csscript issue redirect cscs exe cd redirectreflib cs c script execution engine version copyright c oleg shilo error specified file could not be compiled csscript compilerexception c users klaey appdata local temp csscript cache dbg inject cs error the call is ambiguous between the following methods or properties dbg e xtensions print t params object and dbg extensions print t params object at csscript csexecutor processcompilingresult compilerresults results compilerparameters compile rparams scriptparser parser string scriptfilename string assemblyfilename string additionaldep endencies at csscript csexecutor compile string scriptfilename at csscript csexecutor executeimpl csscript issue redirect can be reproduced by extracting the attached files into a temporary location place cscs exe into the redirect folder and run cscs exe cd redirectreflib cs
1
75,071
25,513,554,408
IssuesEvent
2022-11-28 14:47:45
PowerDNS/pdns
https://api.github.com/repos/PowerDNS/pdns
closed
API endpoint /servers/{server_id}/zones/{zone_id}/check should exist but when calling there is an error and the endpoint cannot be found in the source code.
auth docs defect
<!-- Hi! Thanks for filing an issue. It will be read with care by human beings. Can we ask you to please fill out this template and not simply demand new features or send in complaints? Thanks! --> <!-- Also please search the existing issues (both open and closed) to see if your report might be duplicate --> <!-- Please don't file an issue when you have a support question, send support questions to the mailinglist or ask them on IRC (https://www.powerdns.com/opensource.html) --> <!-- Tell us what is issue is about --> - Program: Authoritative - Issue type: Bug report ### Short description <!-- Explain in a few sentences what the issue/request is --> GET /servers/{server_id}/zones/{zone_id}/check is a documented endpoint in https://doc.powerdns.com/authoritative/http-api/zone.html. However, I reviewed the sourcecode at I cannot find that this is implemented in code. In all the versions of wc-auth.cc that I verified, this path is not registered. ### Environment N/A ### Steps to reproduce Call the 'check' endpoint as described in https://doc.powerdns.com/authoritative/http-api/zone.html. ### Expected behaviour That the endpoint works. I am running 4.1.5 and got an error when calling the `check` endpoint. I checked if a newer version had that endpoint and i checked a lost of branches and tags, but the endpoint was not there. ### Actual behaviour Calling the check endpoint resulted in: '{"error": "Could not find domain '<testdomain>./check.'"}' ### Other information <!-- if you already did more digging into the issue, please provide all the information you gathered -->
1.0
API endpoint /servers/{server_id}/zones/{zone_id}/check should exist but when calling there is an error and the endpoint cannot be found in the source code. - <!-- Hi! Thanks for filing an issue. It will be read with care by human beings. Can we ask you to please fill out this template and not simply demand new features or send in complaints? Thanks! --> <!-- Also please search the existing issues (both open and closed) to see if your report might be duplicate --> <!-- Please don't file an issue when you have a support question, send support questions to the mailinglist or ask them on IRC (https://www.powerdns.com/opensource.html) --> <!-- Tell us what is issue is about --> - Program: Authoritative - Issue type: Bug report ### Short description <!-- Explain in a few sentences what the issue/request is --> GET /servers/{server_id}/zones/{zone_id}/check is a documented endpoint in https://doc.powerdns.com/authoritative/http-api/zone.html. However, I reviewed the sourcecode at I cannot find that this is implemented in code. In all the versions of wc-auth.cc that I verified, this path is not registered. ### Environment N/A ### Steps to reproduce Call the 'check' endpoint as described in https://doc.powerdns.com/authoritative/http-api/zone.html. ### Expected behaviour That the endpoint works. I am running 4.1.5 and got an error when calling the `check` endpoint. I checked if a newer version had that endpoint and i checked a lost of branches and tags, but the endpoint was not there. ### Actual behaviour Calling the check endpoint resulted in: '{"error": "Could not find domain '<testdomain>./check.'"}' ### Other information <!-- if you already did more digging into the issue, please provide all the information you gathered -->
defect
api endpoint servers server id zones zone id check should exist but when calling there is an error and the endpoint cannot be found in the source code program authoritative issue type bug report short description get servers server id zones zone id check is a documented endpoint in however i reviewed the sourcecode at i cannot find that this is implemented in code in all the versions of wc auth cc that i verified this path is not registered environment n a steps to reproduce call the check endpoint as described in expected behaviour that the endpoint works i am running and got an error when calling the check endpoint i checked if a newer version had that endpoint and i checked a lost of branches and tags but the endpoint was not there actual behaviour calling the check endpoint resulted in error could not find domain check other information
1
10,691
8,134,797,724
IssuesEvent
2018-08-19 20:05:23
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
docker exposes containers' ports to the world despite firewall module being enabled
1.severity: security
## Issue description Docker has its own iptables chain and somehow¹ bypasses NixOS firewall, the firewall, which I thought would safeguard this from me. This is very dangerous when you have passwordless development containers running. It seems [that](https://fralef.me/docker-and-iptables.html) `virtualisation.docker.extraOptions = "--iptables=false";` mitigates this, so perhaps it is possible to make the firewall module add a `docker.extraOptions`? 1 — I didn't try understanding why, but I want to start the discussion. ### Steps to reproduce Start docker and, say, a postgres container with (warning: docker-compose syntax ahead) `ports: ["5432:5432"]`, enable firewall and observe an open 5432 port from another machine. ## Technical details - system: `"x86_64-linux"` - host os: `Linux 4.14.32, NixOS, 18.03.132304.45cd6090d9a (Impala)` - multi-user?: `yes` - sandbox: `no` - version: `nix-env (Nix) 2.1pre6148_a4aac7f` - channels(s): `"unstable-18.03pre130932.cc4677c36ee"` - channels(root): `"nixos-18.03.132304.45cd6090d9a"` - nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos/nixpkgs`
True
docker exposes containers' ports to the world despite firewall module being enabled - ## Issue description Docker has its own iptables chain and somehow¹ bypasses NixOS firewall, the firewall, which I thought would safeguard this from me. This is very dangerous when you have passwordless development containers running. It seems [that](https://fralef.me/docker-and-iptables.html) `virtualisation.docker.extraOptions = "--iptables=false";` mitigates this, so perhaps it is possible to make the firewall module add a `docker.extraOptions`? 1 — I didn't try understanding why, but I want to start the discussion. ### Steps to reproduce Start docker and, say, a postgres container with (warning: docker-compose syntax ahead) `ports: ["5432:5432"]`, enable firewall and observe an open 5432 port from another machine. ## Technical details - system: `"x86_64-linux"` - host os: `Linux 4.14.32, NixOS, 18.03.132304.45cd6090d9a (Impala)` - multi-user?: `yes` - sandbox: `no` - version: `nix-env (Nix) 2.1pre6148_a4aac7f` - channels(s): `"unstable-18.03pre130932.cc4677c36ee"` - channels(root): `"nixos-18.03.132304.45cd6090d9a"` - nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos/nixpkgs`
non_defect
docker exposes containers ports to the world despite firewall module being enabled issue description docker has its own iptables chain and somehow¹ bypasses nixos firewall the firewall which i thought would safeguard this from me this is very dangerous when you have passwordless development containers running it seems virtualisation docker extraoptions iptables false mitigates this so perhaps it is possible to make the firewall module add a docker extraoptions — i didn t try understanding why but i want to start the discussion steps to reproduce start docker and say a postgres container with warning docker compose syntax ahead ports enable firewall and observe an open port from another machine technical details system linux host os linux nixos impala multi user yes sandbox no version nix env nix channels s unstable channels root nixos nixpkgs nix var nix profiles per user root channels nixos nixpkgs
0
10,783
2,622,188,814
IssuesEvent
2015-03-04 00:22:07
byzhang/cudpp
https://api.github.com/repos/byzhang/cudpp
closed
satGL produces a garbled image
auto-migrated Milestone-Release1.1 OpSys-Linux Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Build and run the satGL sample app What is the expected output? What do you see instead? Correct output can be seen by running the device emulation version. In release or debug builds, instead the results are a green and blue smear. ``` Original issue reported on code.google.com by `harr...@gmail.com` on 17 Jun 2009 at 1:57
1.0
satGL produces a garbled image - ``` What steps will reproduce the problem? 1. Build and run the satGL sample app What is the expected output? What do you see instead? Correct output can be seen by running the device emulation version. In release or debug builds, instead the results are a green and blue smear. ``` Original issue reported on code.google.com by `harr...@gmail.com` on 17 Jun 2009 at 1:57
defect
satgl produces a garbled image what steps will reproduce the problem build and run the satgl sample app what is the expected output what do you see instead correct output can be seen by running the device emulation version in release or debug builds instead the results are a green and blue smear original issue reported on code google com by harr gmail com on jun at
1
58,553
14,432,798,076
IssuesEvent
2020-12-07 03:01:21
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
opened
Whl package not building due with a issue with sed - MacOS 11
type:build/install
**System information** - OS Platform and Distribution: Mac OS 11.0.1 Big Sur - TensorFlow installed from (source or binary): Source - git - TensorFlow version: 2.4.0-rc4 - Python version: 3.8.6 (macports) - Installed using virtualenv? pip? conda?: No - Bazel version (if compiling from source): 3.1.0 - GCC/Compiler version (if compiling from source): Apple clang version 12.0.0 (clang-1200.0.32.27) XCode 12.2 - CUDA/cuDNN version: No - GPU model and memory: AMD Radeon Pro 555X - 4GB **Describe the problem** Tensorflow compiles correctly. However, when using the command: `bazel-bin/tensorflow/tools/pip_package/build_pip_package` the whl package is not produced and rather the following error is shown: `sed: /var/folders/y4/85hx48054tbd2y04jkf3l6880000gn/T/tmp.XXXXXXXXXX.Ty6dzcrU/tensorflow/__init__.py: in-place editing only works for regular files` **Provide the exact sequence of commands / steps that you executed before running into the problem** 1. Compile from source as indicated in the tensorflow manual. 2. Run the command to build the whl package: `bazel-bin/tensorflow/tools/pip_package/build_pip_package ..` **Any other info / logs** This error is not present in MacOS 10.15. All proceed as it should Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
1.0
Whl package not building due with a issue with sed - MacOS 11 - **System information** - OS Platform and Distribution: Mac OS 11.0.1 Big Sur - TensorFlow installed from (source or binary): Source - git - TensorFlow version: 2.4.0-rc4 - Python version: 3.8.6 (macports) - Installed using virtualenv? pip? conda?: No - Bazel version (if compiling from source): 3.1.0 - GCC/Compiler version (if compiling from source): Apple clang version 12.0.0 (clang-1200.0.32.27) XCode 12.2 - CUDA/cuDNN version: No - GPU model and memory: AMD Radeon Pro 555X - 4GB **Describe the problem** Tensorflow compiles correctly. However, when using the command: `bazel-bin/tensorflow/tools/pip_package/build_pip_package` the whl package is not produced and rather the following error is shown: `sed: /var/folders/y4/85hx48054tbd2y04jkf3l6880000gn/T/tmp.XXXXXXXXXX.Ty6dzcrU/tensorflow/__init__.py: in-place editing only works for regular files` **Provide the exact sequence of commands / steps that you executed before running into the problem** 1. Compile from source as indicated in the tensorflow manual. 2. Run the command to build the whl package: `bazel-bin/tensorflow/tools/pip_package/build_pip_package ..` **Any other info / logs** This error is not present in MacOS 10.15. All proceed as it should Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
non_defect
whl package not building due with a issue with sed macos system information os platform and distribution mac os big sur tensorflow installed from source or binary source git tensorflow version python version macports installed using virtualenv pip conda no bazel version if compiling from source gcc compiler version if compiling from source apple clang version clang xcode cuda cudnn version no gpu model and memory amd radeon pro describe the problem tensorflow compiles correctly however when using the command bazel bin tensorflow tools pip package build pip package the whl package is not produced and rather the following error is shown sed var folders t tmp xxxxxxxxxx tensorflow init py in place editing only works for regular files provide the exact sequence of commands steps that you executed before running into the problem compile from source as indicated in the tensorflow manual run the command to build the whl package bazel bin tensorflow tools pip package build pip package any other info logs this error is not present in macos all proceed as it should include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached
0
216,031
16,625,852,139
IssuesEvent
2021-06-03 09:26:46
RotherOSS/doc-otobo-installation
https://api.github.com/repos/RotherOSS/doc-otobo-installation
opened
Update the requirements
documentation
https://doc.otobo.org/manual/installation/stable/en/content/requirements.html can be improved: - [ ] No need to mention Node - [ ] Add Redis as an optional dependency
1.0
Update the requirements - https://doc.otobo.org/manual/installation/stable/en/content/requirements.html can be improved: - [ ] No need to mention Node - [ ] Add Redis as an optional dependency
non_defect
update the requirements can be improved no need to mention node add redis as an optional dependency
0
77,577
27,058,913,688
IssuesEvent
2023-02-13 18:08:45
fecgov/fecfile-web-app
https://api.github.com/repos/fecgov/fecfile-web-app
closed
Defect - "Earmark Memo" the transactions table is not displaying the transaction type "Earmark Memo"
defect
This defect is being written for after creating an "Earmark Memo" the transactions table is not displaying the transaction type "Earmark Memo" as shown in screenshot below. This was tested in both DEV and STAGE environments. ![image.png](https://images.zenhubusercontent.com/61ba01e428a658b4eb0ca758/8a8e125d-ec3a-4fce-b960-ef454748c2f3) ### DEV ### Note: The isDependentChild property in the TransactionType class, when set to true, identifies the transaction type as being the child transaction in a "double entry" transaction screen like AG and AF Per @toddlees It has no label because it's missing from `ScheduleATransactionTypeLabels` in the front end
1.0
Defect - "Earmark Memo" the transactions table is not displaying the transaction type "Earmark Memo" - This defect is being written for after creating an "Earmark Memo" the transactions table is not displaying the transaction type "Earmark Memo" as shown in screenshot below. This was tested in both DEV and STAGE environments. ![image.png](https://images.zenhubusercontent.com/61ba01e428a658b4eb0ca758/8a8e125d-ec3a-4fce-b960-ef454748c2f3) ### DEV ### Note: The isDependentChild property in the TransactionType class, when set to true, identifies the transaction type as being the child transaction in a "double entry" transaction screen like AG and AF Per @toddlees It has no label because it's missing from `ScheduleATransactionTypeLabels` in the front end
defect
defect earmark memo the transactions table is not displaying the transaction type earmark memo this defect is being written for after creating an earmark memo the transactions table is not displaying the transaction type earmark memo as shown in screenshot below this was tested in both dev and stage environments dev note the isdependentchild property in the transactiontype class when set to true identifies the transaction type as being the child transaction in a double entry transaction screen like ag and af per toddlees it has no label because it s missing from scheduleatransactiontypelabels in the front end
1
578,618
17,149,083,820
IssuesEvent
2021-07-13 18:00:03
CyanLabs/Syn3Updater
https://api.github.com/repos/CyanLabs/Syn3Updater
closed
2.10.0 - Error: System.IndexOutOfRangeException
Priority: Medium Type: Bug
Message: Index was outside the bounds of the array. FileHelper.cs:line 194 FileHelper.cs:line 201 DownloadViewModel.cs:line 310 DownloadViewModel.cs:line 184
1.0
2.10.0 - Error: System.IndexOutOfRangeException - Message: Index was outside the bounds of the array. FileHelper.cs:line 194 FileHelper.cs:line 201 DownloadViewModel.cs:line 310 DownloadViewModel.cs:line 184
non_defect
error system indexoutofrangeexception message index was outside the bounds of the array filehelper cs line filehelper cs line downloadviewmodel cs line downloadviewmodel cs line
0
345,926
24,880,073,001
IssuesEvent
2022-10-27 23:35:31
quacksouls/bitburner
https://api.github.com/repos/quacksouls/bitburner
closed
crime: karma, combat stats, Charisma, Hack, and money per minute
documentation
Similar to #91, but we exclude the Intelligence stat. - [x] A save file where we have destroyed BN1.3 (Genesis) and BN4.3 (Singularity). We need access to the Singularity API. - [x] A script to calculate the following per minute for each crime: karma, combat stats, Charisma, Hack, and money. - [x] Data on the amount of these per minute for each crime: karma, combat stats, Charisma, Hack, and money.
1.0
crime: karma, combat stats, Charisma, Hack, and money per minute - Similar to #91, but we exclude the Intelligence stat. - [x] A save file where we have destroyed BN1.3 (Genesis) and BN4.3 (Singularity). We need access to the Singularity API. - [x] A script to calculate the following per minute for each crime: karma, combat stats, Charisma, Hack, and money. - [x] Data on the amount of these per minute for each crime: karma, combat stats, Charisma, Hack, and money.
non_defect
crime karma combat stats charisma hack and money per minute similar to but we exclude the intelligence stat a save file where we have destroyed genesis and singularity we need access to the singularity api a script to calculate the following per minute for each crime karma combat stats charisma hack and money data on the amount of these per minute for each crime karma combat stats charisma hack and money
0
54,592
13,780,022,739
IssuesEvent
2020-10-08 14:26:46
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
Problem in code generation for entities with @Type(type = "jsonb") from hibernate-types-52 library
T: Defect
### Expected behavior Upon executing the plugin _jooq-codegen-maven_ with the correct project structure as indicated in the jOOQ documentation, the expected generated classes from the existing project entities should be the following (e.g. for the sake of demonstration let´s use an entity called User): 1. Under the folder _/tables_ we should have a **User.java** 2. Under folder _/tables/records_ we should have a **UserRecord.java** ### Actual behavior In my case i wanted to use the Postgres **jsonb** data type for a String field, so using the types provided by the very useful library ``` <dependency> <groupId>com.vladmihalcea</groupId> <artifactId>hibernate-types-52</artifactId> <version>2.9.10</version> </dependency> ``` we get an example similar to the one below for the User entity. Configured like this, this entity **will not** have the expected classes (User and UserRecord) derived the execution of the _jooq-codegen-maven_ plugin. ``` @Entity @Table(name = "audit", schema = "dino_operations") @TypeDef(name = "jsonb", typeClass = JsonBinaryType.class) // By removing this jOOQ generated expected classes public class User { // (...) @Type(type = "jsonb") // By removing this jOOQ generated expected classes @Column(name = "elements", columnDefinition = "jsonb") private String elements; } ``` **Summary**: Let´s say i have 10 entities and 2 annotated like the example above, so these are excluded(or a silent error occurs) from code generation so i get 8 entities being processed correctly and respective classes generated. When i removed the annotations i got a 10 out of 10. ### Steps to reproduce the problem - An entity with a String field annotated with the annotations indicated above - Execute jooq-codegen-maven with the desired configurations and structure (in my case the database was org.jooq.meta.extensions.jpa.JPADatabase) ### Versions - jOOQ (3.13.5): - Java (8): - Database (Postgres 10): - OS (macOS Catalina: - JDBC Driver (spring-jdbc from spring Spring boot 2.3.3.RELEASE):
1.0
Problem in code generation for entities with @Type(type = "jsonb") from hibernate-types-52 library - ### Expected behavior Upon executing the plugin _jooq-codegen-maven_ with the correct project structure as indicated in the jOOQ documentation, the expected generated classes from the existing project entities should be the following (e.g. for the sake of demonstration let´s use an entity called User): 1. Under the folder _/tables_ we should have a **User.java** 2. Under folder _/tables/records_ we should have a **UserRecord.java** ### Actual behavior In my case i wanted to use the Postgres **jsonb** data type for a String field, so using the types provided by the very useful library ``` <dependency> <groupId>com.vladmihalcea</groupId> <artifactId>hibernate-types-52</artifactId> <version>2.9.10</version> </dependency> ``` we get an example similar to the one below for the User entity. Configured like this, this entity **will not** have the expected classes (User and UserRecord) derived the execution of the _jooq-codegen-maven_ plugin. ``` @Entity @Table(name = "audit", schema = "dino_operations") @TypeDef(name = "jsonb", typeClass = JsonBinaryType.class) // By removing this jOOQ generated expected classes public class User { // (...) @Type(type = "jsonb") // By removing this jOOQ generated expected classes @Column(name = "elements", columnDefinition = "jsonb") private String elements; } ``` **Summary**: Let´s say i have 10 entities and 2 annotated like the example above, so these are excluded(or a silent error occurs) from code generation so i get 8 entities being processed correctly and respective classes generated. When i removed the annotations i got a 10 out of 10. ### Steps to reproduce the problem - An entity with a String field annotated with the annotations indicated above - Execute jooq-codegen-maven with the desired configurations and structure (in my case the database was org.jooq.meta.extensions.jpa.JPADatabase) ### Versions - jOOQ (3.13.5): - Java (8): - Database (Postgres 10): - OS (macOS Catalina: - JDBC Driver (spring-jdbc from spring Spring boot 2.3.3.RELEASE):
defect
problem in code generation for entities with type type jsonb from hibernate types library expected behavior upon executing the plugin jooq codegen maven with the correct project structure as indicated in the jooq documentation the expected generated classes from the existing project entities should be the following e g for the sake of demonstration let´s use an entity called user under the folder tables we should have a user java under folder tables records we should have a userrecord java actual behavior in my case i wanted to use the postgres jsonb data type for a string field so using the types provided by the very useful library com vladmihalcea hibernate types we get an example similar to the one below for the user entity configured like this this entity will not have the expected classes user and userrecord derived the execution of the jooq codegen maven plugin entity table name audit schema dino operations typedef name jsonb typeclass jsonbinarytype class by removing this jooq generated expected classes public class user type type jsonb by removing this jooq generated expected classes column name elements columndefinition jsonb private string elements summary let´s say i have entities and annotated like the example above so these are excluded or a silent error occurs from code generation so i get entities being processed correctly and respective classes generated when i removed the annotations i got a out of steps to reproduce the problem an entity with a string field annotated with the annotations indicated above execute jooq codegen maven with the desired configurations and structure in my case the database was org jooq meta extensions jpa jpadatabase versions jooq java database postgres os macos catalina jdbc driver spring jdbc from spring spring boot release
1
33,967
7,314,613,137
IssuesEvent
2018-03-01 08:03:18
PowerDNS/pdns
https://api.github.com/repos/PowerDNS/pdns
closed
dnsdist 1.2.0 complains about incorrect option but starts anyway
defect dnsdist
- Program: dnsdist <!-- delete the ones that do not apply --> - Issue type: Bug report ### Short description dnsdist 1.2.0 complains about option; starts anyway <!-- If this is a bug report, use the following part of the the template and delete the part at the bottom --> ### Steps to reproduce ``` $ /usr/sbin/dnsdist --check-bla --config /tmp/dddd /usr/sbin/dnsdist: unrecognized option '--check-bla' Added downstream server 127.0.0.2:5353 ``` ### Expected behaviour diagnostic message and exit (1)
1.0
dnsdist 1.2.0 complains about incorrect option but starts anyway - - Program: dnsdist <!-- delete the ones that do not apply --> - Issue type: Bug report ### Short description dnsdist 1.2.0 complains about option; starts anyway <!-- If this is a bug report, use the following part of the the template and delete the part at the bottom --> ### Steps to reproduce ``` $ /usr/sbin/dnsdist --check-bla --config /tmp/dddd /usr/sbin/dnsdist: unrecognized option '--check-bla' Added downstream server 127.0.0.2:5353 ``` ### Expected behaviour diagnostic message and exit (1)
defect
dnsdist complains about incorrect option but starts anyway program dnsdist issue type bug report short description dnsdist complains about option starts anyway if this is a bug report use the following part of the the template and delete the part at the bottom steps to reproduce usr sbin dnsdist check bla config tmp dddd usr sbin dnsdist unrecognized option check bla added downstream server expected behaviour diagnostic message and exit
1
30,293
6,086,427,670
IssuesEvent
2017-06-18 00:52:57
jfabry/LiveRobotProgramming
https://api.github.com/repos/jfabry/LiveRobotProgramming
opened
PhaROS Bridge: context menu is not working
Bridge-PhaROS Component-UI Priority-Medium Type-Defect
When there is a pub or sub in the pharos bridge window, the context menu for deleting them is not working
1.0
PhaROS Bridge: context menu is not working - When there is a pub or sub in the pharos bridge window, the context menu for deleting them is not working
defect
pharos bridge context menu is not working when there is a pub or sub in the pharos bridge window the context menu for deleting them is not working
1
678,206
23,190,668,710
IssuesEvent
2022-08-01 12:23:27
SAP/xsk
https://api.github.com/repos/SAP/xsk
closed
[Core] Reconsider the authentication mechanisms in the XSK
wontfix core priority-medium effort-medium security investigation / discussion incomplete
Currently, when running the XSK locally, following the suggested way in the documentation, we use Form-based authentication provided by Tomcat. This leads to inconsistencies between the dev and production environments where we use OAuth. Due to these differences, we have observed several problems: - websockets are behaving differently on different environments due to different authorization/authentication approaches used - Tomcat's updates break the existing Form-based authentication but not the OAuth one and this leads to blockers when updating - difficulties in testing new code that is somehow dependent on the authentication - maintaining code that does the same thing but in a different way We should think of a way to use a common authentication mechanism for the environments XSK is running on. This may be Keycloak (or something similar) or at least have an easy way to run XSK locally using the same authentication services as in Kyma/Cloud Foundry, and advise developers to use this approach.
1.0
[Core] Reconsider the authentication mechanisms in the XSK - Currently, when running the XSK locally, following the suggested way in the documentation, we use Form-based authentication provided by Tomcat. This leads to inconsistencies between the dev and production environments where we use OAuth. Due to these differences, we have observed several problems: - websockets are behaving differently on different environments due to different authorization/authentication approaches used - Tomcat's updates break the existing Form-based authentication but not the OAuth one and this leads to blockers when updating - difficulties in testing new code that is somehow dependent on the authentication - maintaining code that does the same thing but in a different way We should think of a way to use a common authentication mechanism for the environments XSK is running on. This may be Keycloak (or something similar) or at least have an easy way to run XSK locally using the same authentication services as in Kyma/Cloud Foundry, and advise developers to use this approach.
non_defect
reconsider the authentication mechanisms in the xsk currently when running the xsk locally following the suggested way in the documentation we use form based authentication provided by tomcat this leads to inconsistencies between the dev and production environments where we use oauth due to these differences we have observed several problems websockets are behaving differently on different environments due to different authorization authentication approaches used tomcat s updates break the existing form based authentication but not the oauth one and this leads to blockers when updating difficulties in testing new code that is somehow dependent on the authentication maintaining code that does the same thing but in a different way we should think of a way to use a common authentication mechanism for the environments xsk is running on this may be keycloak or something similar or at least have an easy way to run xsk locally using the same authentication services as in kyma cloud foundry and advise developers to use this approach
0
17,213
2,984,426,524
IssuesEvent
2015-07-18 00:45:39
google/omaha
https://api.github.com/repos/google/omaha
closed
google Chrom or Talk Bundle cannot be installed
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. build omaha with visual studio 2010, successfully pass build and all unit tests 2. go into staging folder 3. input GoogleUpdate.exe /install "bundlename=Google%20Talk%20Bundle&appguid={D0AB2EBC-931B-4013-9FEB-C9C4C2225C8C }&appname=Google%20Talk%20Plugin&needsadmin=False&lang=en" or GoogleUpdate.exe /install "bundlename=Google%20Chrome%20Bundle&appguid={8A69D345-D564-463C-AFF1-A69D9E530F 96}&appname=Google%20Chrome&needsadmin=True&lang=en" What is the expected output? It's supposed to download and install the product. What do you see instead? After downloading is completed, GoogleUpdate.exe crash showing: Debug Assertion Failed! Program: ...\AppData\Local\OmahaCompanyName\Update\GoogleUpdate.exe File:f:\dd\vctools\crt_bld\self_x86\crt\src\isctype.c Line:56 Expression:(unsigned)(c+1)<=256 What version of the product are you using? Visual Studio 2010, python-2.4.4, pywin32-216.win32-py2.4, scons-1.3.1.win32, wix37, WTL80_sf, ATL_Server_Source_and_Headers_9_0_70425_Alpha, omaha-src-r110.zip On what operating system? on Windows 7 Please provide any additional information below. ``` Original issue reported on code.google.com by `mmange...@gmail.com` on 7 Mar 2013 at 1:13
1.0
google Chrom or Talk Bundle cannot be installed - ``` What steps will reproduce the problem? 1. build omaha with visual studio 2010, successfully pass build and all unit tests 2. go into staging folder 3. input GoogleUpdate.exe /install "bundlename=Google%20Talk%20Bundle&appguid={D0AB2EBC-931B-4013-9FEB-C9C4C2225C8C }&appname=Google%20Talk%20Plugin&needsadmin=False&lang=en" or GoogleUpdate.exe /install "bundlename=Google%20Chrome%20Bundle&appguid={8A69D345-D564-463C-AFF1-A69D9E530F 96}&appname=Google%20Chrome&needsadmin=True&lang=en" What is the expected output? It's supposed to download and install the product. What do you see instead? After downloading is completed, GoogleUpdate.exe crash showing: Debug Assertion Failed! Program: ...\AppData\Local\OmahaCompanyName\Update\GoogleUpdate.exe File:f:\dd\vctools\crt_bld\self_x86\crt\src\isctype.c Line:56 Expression:(unsigned)(c+1)<=256 What version of the product are you using? Visual Studio 2010, python-2.4.4, pywin32-216.win32-py2.4, scons-1.3.1.win32, wix37, WTL80_sf, ATL_Server_Source_and_Headers_9_0_70425_Alpha, omaha-src-r110.zip On what operating system? on Windows 7 Please provide any additional information below. ``` Original issue reported on code.google.com by `mmange...@gmail.com` on 7 Mar 2013 at 1:13
defect
google chrom or talk bundle cannot be installed what steps will reproduce the problem build omaha with visual studio successfully pass build and all unit tests go into staging folder input googleupdate exe install bundlename google appguid appname google needsadmin false lang en or googleupdate exe install bundlename google appguid appname google needsadmin true lang en what is the expected output it s supposed to download and install the product what do you see instead after downloading is completed googleupdate exe crash showing debug assertion failed program appdata local omahacompanyname update googleupdate exe file f dd vctools crt bld self crt src isctype c line expression unsigned c what version of the product are you using visual studio python scons sf atl server source and headers alpha omaha src zip on what operating system on windows please provide any additional information below original issue reported on code google com by mmange gmail com on mar at
1
50,596
13,187,609,136
IssuesEvent
2020-08-13 03:58:55
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
Link to PROPOSAL project (Trac #1019)
Migrated from Trac cmake defect
Hi, Could you please provide a link to the paper describing how the PROPOSAL icesim meta-project work? Thanks. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1019">https://code.icecube.wisc.edu/ticket/1019</a>, reported by icecube and owned by </em></summary> <p> ```json { "status": "closed", "changetime": "2015-06-11T18:38:27", "description": "Hi,\n\nCould you please provide a link to the paper describing how the PROPOSAL icesim meta-project work?\n\nThanks.", "reporter": "icecube", "cc": "jtatar@lbl.gov", "resolution": "wontfix", "_ts": "1434047907705473", "component": "cmake", "summary": "Link to PROPOSAL project", "priority": "normal", "keywords": "", "time": "2015-06-11T17:37:23", "milestone": "", "owner": "", "type": "defect" } ``` </p> </details>
1.0
Link to PROPOSAL project (Trac #1019) - Hi, Could you please provide a link to the paper describing how the PROPOSAL icesim meta-project work? Thanks. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1019">https://code.icecube.wisc.edu/ticket/1019</a>, reported by icecube and owned by </em></summary> <p> ```json { "status": "closed", "changetime": "2015-06-11T18:38:27", "description": "Hi,\n\nCould you please provide a link to the paper describing how the PROPOSAL icesim meta-project work?\n\nThanks.", "reporter": "icecube", "cc": "jtatar@lbl.gov", "resolution": "wontfix", "_ts": "1434047907705473", "component": "cmake", "summary": "Link to PROPOSAL project", "priority": "normal", "keywords": "", "time": "2015-06-11T17:37:23", "milestone": "", "owner": "", "type": "defect" } ``` </p> </details>
defect
link to proposal project trac hi could you please provide a link to the paper describing how the proposal icesim meta project work thanks migrated from json status closed changetime description hi n ncould you please provide a link to the paper describing how the proposal icesim meta project work n nthanks reporter icecube cc jtatar lbl gov resolution wontfix ts component cmake summary link to proposal project priority normal keywords time milestone owner type defect
1
62,650
17,105,641,678
IssuesEvent
2021-07-09 17:15:52
Gogo1951/GogoLoot
https://api.github.com/repos/Gogo1951/GogoLoot
closed
Remove Lag from "Auto Destroy Grays"
GogoLoot - Application Type - Defect
There's a lag when you have "auto destroy grays" enabled. every time you loot there's a notable drop in frame rates.
1.0
Remove Lag from "Auto Destroy Grays" - There's a lag when you have "auto destroy grays" enabled. every time you loot there's a notable drop in frame rates.
defect
remove lag from auto destroy grays there s a lag when you have auto destroy grays enabled every time you loot there s a notable drop in frame rates
1
55,376
14,410,691,869
IssuesEvent
2020-12-04 05:33:03
AeroScripts/QuestieDev
https://api.github.com/repos/AeroScripts/QuestieDev
opened
Error when moving quest tracker while in combat
Type - Defect
## Bug description When trying to move the quest tracker while in combat using "Control + Left Click" i get this error: `...s\Questie\Modules\Tracker\QuestieTrackerPrivates.lua:24: Frame Questie_BaseFrame is not movable [C]: in function `StartMoving' ...s\Questie\Modules\Tracker\QuestieTrackerPrivates.lua:24: in function <...s\Questie\Modules\Tracker\QuestieTrackerPrivates.lua:15> Locals: (*temporary) = Questie_BaseFrame { 0 = <userdata> Update = <function> defined @Questie\Modules\Tracker\QuestieTracker.lua:228 sizer = Questie_Sizer { } }` ## Questie version v6.2.3
1.0
Error when moving quest tracker while in combat - ## Bug description When trying to move the quest tracker while in combat using "Control + Left Click" i get this error: `...s\Questie\Modules\Tracker\QuestieTrackerPrivates.lua:24: Frame Questie_BaseFrame is not movable [C]: in function `StartMoving' ...s\Questie\Modules\Tracker\QuestieTrackerPrivates.lua:24: in function <...s\Questie\Modules\Tracker\QuestieTrackerPrivates.lua:15> Locals: (*temporary) = Questie_BaseFrame { 0 = <userdata> Update = <function> defined @Questie\Modules\Tracker\QuestieTracker.lua:228 sizer = Questie_Sizer { } }` ## Questie version v6.2.3
defect
error when moving quest tracker while in combat bug description when trying to move the quest tracker while in combat using control left click i get this error s questie modules tracker questietrackerprivates lua frame questie baseframe is not movable in function startmoving s questie modules tracker questietrackerprivates lua in function locals temporary questie baseframe update defined questie modules tracker questietracker lua sizer questie sizer questie version
1
10,072
7,888,843,040
IssuesEvent
2018-06-28 00:14:41
8xprotocol/contracts
https://api.github.com/repos/8xprotocol/contracts
closed
Create nonce field for VolumeSubscription
bug in progress security
1. In volume subscription, create a new plan with a price and identifier 2. User subscribes to that plan 3. Delete the plan, but recreate it with the same identifier and different price OR interval (it'll have the same hash) 4. Users get charged more. A simple resolution would be to hash the plan with the amount AND interval.
True
Create nonce field for VolumeSubscription - 1. In volume subscription, create a new plan with a price and identifier 2. User subscribes to that plan 3. Delete the plan, but recreate it with the same identifier and different price OR interval (it'll have the same hash) 4. Users get charged more. A simple resolution would be to hash the plan with the amount AND interval.
non_defect
create nonce field for volumesubscription in volume subscription create a new plan with a price and identifier user subscribes to that plan delete the plan but recreate it with the same identifier and different price or interval it ll have the same hash users get charged more a simple resolution would be to hash the plan with the amount and interval
0
40,710
6,845,818,457
IssuesEvent
2017-11-13 09:45:48
dgraph-io/badger
https://api.github.com/repos/dgraph-io/badger
closed
Need to document the sorting order of the keys
documentation
There is no document describing how keys are sorted, so I have to test it with the code below. It seems the ordering should be described as something like "byte-wise prefix order", similar to big endian but the keys are aligned from the least significant byte, and "no byte" is smaller than "byte zero" (the shorter the smaller). I think the behavior should be clarified in the document so that we can depend on the sorting more confidently. The test code: ```go func TestKeyOrder(t *testing.T) { dir := "test_order" if err := os.MkdirAll(dir, 0755); err != nil { t.Fatal(err) } defer os.RemoveAll(dir) db, err := openBadger(dir) if err != nil { t.Fatal(err) } defer db.Close() if err := db.Update(func(txn *badger.Txn) error { if err := txn.Set([]byte{0, 0}, nil); err != nil { return err } if err := txn.Set([]byte{0, 1}, nil); err != nil { return err } if err := txn.Set([]byte{0, 2}, nil); err != nil { return err } if err := txn.Set([]byte{1, 0}, nil); err != nil { return err } if err := txn.Set([]byte{1, 1}, nil); err != nil { return err } if err := txn.Set([]byte{1, 2}, nil); err != nil { return err } if err := txn.Set([]byte{1, 2, 0}, nil); err != nil { return err } if err := txn.Set([]byte{2, 0}, nil); err != nil { return err } if err := txn.Set([]byte{2, 1}, nil); err != nil { return err } if err := txn.Set([]byte{2, 2}, nil); err != nil { return err } return nil }); err != nil { t.Fatal(err) } if err := db.View(func(txn *badger.Txn) error { opt := badger.DefaultIteratorOptions // opt.Reverse = true it := txn.NewIterator(opt) for it.Rewind(); it.Valid(); it.Next() { item := it.Item() k := item.Key() fmt.Println(k) } return nil }); err != nil { t.Fatal(err) } } ```
1.0
Need to document the sorting order of the keys - There is no document describing how keys are sorted, so I have to test it with the code below. It seems the ordering should be described as something like "byte-wise prefix order", similar to big endian but the keys are aligned from the least significant byte, and "no byte" is smaller than "byte zero" (the shorter the smaller). I think the behavior should be clarified in the document so that we can depend on the sorting more confidently. The test code: ```go func TestKeyOrder(t *testing.T) { dir := "test_order" if err := os.MkdirAll(dir, 0755); err != nil { t.Fatal(err) } defer os.RemoveAll(dir) db, err := openBadger(dir) if err != nil { t.Fatal(err) } defer db.Close() if err := db.Update(func(txn *badger.Txn) error { if err := txn.Set([]byte{0, 0}, nil); err != nil { return err } if err := txn.Set([]byte{0, 1}, nil); err != nil { return err } if err := txn.Set([]byte{0, 2}, nil); err != nil { return err } if err := txn.Set([]byte{1, 0}, nil); err != nil { return err } if err := txn.Set([]byte{1, 1}, nil); err != nil { return err } if err := txn.Set([]byte{1, 2}, nil); err != nil { return err } if err := txn.Set([]byte{1, 2, 0}, nil); err != nil { return err } if err := txn.Set([]byte{2, 0}, nil); err != nil { return err } if err := txn.Set([]byte{2, 1}, nil); err != nil { return err } if err := txn.Set([]byte{2, 2}, nil); err != nil { return err } return nil }); err != nil { t.Fatal(err) } if err := db.View(func(txn *badger.Txn) error { opt := badger.DefaultIteratorOptions // opt.Reverse = true it := txn.NewIterator(opt) for it.Rewind(); it.Valid(); it.Next() { item := it.Item() k := item.Key() fmt.Println(k) } return nil }); err != nil { t.Fatal(err) } } ```
non_defect
need to document the sorting order of the keys there is no document describing how keys are sorted so i have to test it with the code below it seems the ordering should be described as something like byte wise prefix order similar to big endian but the keys are aligned from the least significant byte and no byte is smaller than byte zero the shorter the smaller i think the behavior should be clarified in the document so that we can depend on the sorting more confidently the test code go func testkeyorder t testing t dir test order if err os mkdirall dir err nil t fatal err defer os removeall dir db err openbadger dir if err nil t fatal err defer db close if err db update func txn badger txn error if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err if err txn set byte nil err nil return err return nil err nil t fatal err if err db view func txn badger txn error opt badger defaultiteratoroptions opt reverse true it txn newiterator opt for it rewind it valid it next item it item k item key fmt println k return nil err nil t fatal err
0
142,213
5,460,265,240
IssuesEvent
2017-03-09 04:17:59
NuGet/Home
https://api.github.com/repos/NuGet/Home
closed
Logging (via status bar, output window, and warning/error window) requires UI thread and slows down Restore greatly.
Area: Perf Area: Restore Priority:0 Type:Bug
Split off from #4617
1.0
Logging (via status bar, output window, and warning/error window) requires UI thread and slows down Restore greatly. - Split off from #4617
non_defect
logging via status bar output window and warning error window requires ui thread and slows down restore greatly split off from
0
58,672
16,678,901,142
IssuesEvent
2021-06-07 20:05:04
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
sitewide, forms — 508-defect-3 [SCREENREADER]: Consider updating phone number documentation in design system, to include spacing in the aria-label
508-defect-3 508-issue-cognition 508-issue-semantic-markup 508/Accessibility components design system forms sitewide triage vsa-benefits-2 vsp-design-system-team
# [508-defect-3](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-3) > Located in [VSA BAM2 MDOT MVP a11y spot check](https://github.com/department-of-veterans-affairs/va.gov-team/issues/5868#issuecomment-627425136) > Previously documented https://github.com/department-of-veterans-affairs/va.gov-team/issues/6997 **Feedback framework** - **❗️ Must** for if the feedback must be applied - **⚠️Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Description **Finding 2** There is a new method for coding phone numbers on VA.gov, and Robin Garrison just added a PR for a global component. **Consider** adding the aria-label to ensure the screen reader reads the number empathetically. Using a title attribute on links is not recommended. This guidance applies to all of the phone numbers within the product. Work initiated by @Mottie: https://github.com/department-of-veterans-affairs/va.gov-team/issues/8793 ## Point of Contact **VFS Point of Contact:** Jennifer ## Acceptance Criteria As a screen reader user, I want to read phone numbers spelling out each number with enough clarity to comprehend. ## Environment * Operating System: all * Browser: all * Screenreading device: all * Server destination: all ## Steps to Recreate 1. Enter https://staging.va.gov/hearing-aid-batteries-and-accessories/introduction in browser 2. Log in as username: vets.gov.user+1@gmail.com 3. Start screen reading device 4. Verify the title text is read, and the numbers read out may vary depending on the screen reader used; sometimes read as hundreds and thousands ## Possible Fixes (optional) Remove the title attribute, and add an aria-label. **Current code** ```html <a title="Dial the telephone number 303-273-6200" href="tel:303-273-6200">303-273-6200</a> ``` **Recommended code** ```html <a aria-label="3 0 3. 2 7 3. 6 2 0 0." href="tel:303-273-6200">303-273-6200</a> ``` ## WCAG or Vendor Guidance (optional) * [I thought title text improved accessibility. I was wrong.](https://silktide.com/blog/2013/i-thought-title-text-improved-accessibility-i-was-wrong) * [Accessibility and phone number formatting](https://thatdevgirl.com/blog/accessibility-phone-number-formatting) ## Screenshots or Trace Logs ![Screen Shot 2020-05-12 at 11 32 31 AM](https://user-images.githubusercontent.com/57469/81713793-500df380-9444-11ea-8884-47e3514b831e.png)
1.0
sitewide, forms — 508-defect-3 [SCREENREADER]: Consider updating phone number documentation in design system, to include spacing in the aria-label - # [508-defect-3](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-3) > Located in [VSA BAM2 MDOT MVP a11y spot check](https://github.com/department-of-veterans-affairs/va.gov-team/issues/5868#issuecomment-627425136) > Previously documented https://github.com/department-of-veterans-affairs/va.gov-team/issues/6997 **Feedback framework** - **❗️ Must** for if the feedback must be applied - **⚠️Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Description **Finding 2** There is a new method for coding phone numbers on VA.gov, and Robin Garrison just added a PR for a global component. **Consider** adding the aria-label to ensure the screen reader reads the number empathetically. Using a title attribute on links is not recommended. This guidance applies to all of the phone numbers within the product. Work initiated by @Mottie: https://github.com/department-of-veterans-affairs/va.gov-team/issues/8793 ## Point of Contact **VFS Point of Contact:** Jennifer ## Acceptance Criteria As a screen reader user, I want to read phone numbers spelling out each number with enough clarity to comprehend. ## Environment * Operating System: all * Browser: all * Screenreading device: all * Server destination: all ## Steps to Recreate 1. Enter https://staging.va.gov/hearing-aid-batteries-and-accessories/introduction in browser 2. Log in as username: vets.gov.user+1@gmail.com 3. Start screen reading device 4. Verify the title text is read, and the numbers read out may vary depending on the screen reader used; sometimes read as hundreds and thousands ## Possible Fixes (optional) Remove the title attribute, and add an aria-label. **Current code** ```html <a title="Dial the telephone number 303-273-6200" href="tel:303-273-6200">303-273-6200</a> ``` **Recommended code** ```html <a aria-label="3 0 3. 2 7 3. 6 2 0 0." href="tel:303-273-6200">303-273-6200</a> ``` ## WCAG or Vendor Guidance (optional) * [I thought title text improved accessibility. I was wrong.](https://silktide.com/blog/2013/i-thought-title-text-improved-accessibility-i-was-wrong) * [Accessibility and phone number formatting](https://thatdevgirl.com/blog/accessibility-phone-number-formatting) ## Screenshots or Trace Logs ![Screen Shot 2020-05-12 at 11 32 31 AM](https://user-images.githubusercontent.com/57469/81713793-500df380-9444-11ea-8884-47e3514b831e.png)
defect
sitewide forms — defect consider updating phone number documentation in design system to include spacing in the aria label located in previously documented feedback framework ❗️ must for if the feedback must be applied ⚠️should if the feedback is best practice ✔️ consider for suggestions enhancements description finding there is a new method for coding phone numbers on va gov and robin garrison just added a pr for a global component consider adding the aria label to ensure the screen reader reads the number empathetically using a title attribute on links is not recommended this guidance applies to all of the phone numbers within the product work initiated by mottie point of contact vfs point of contact jennifer acceptance criteria as a screen reader user i want to read phone numbers spelling out each number with enough clarity to comprehend environment operating system all browser all screenreading device all server destination all steps to recreate enter in browser log in as username vets gov user gmail com start screen reading device verify the title text is read and the numbers read out may vary depending on the screen reader used sometimes read as hundreds and thousands possible fixes optional remove the title attribute and add an aria label current code html recommended code html wcag or vendor guidance optional screenshots or trace logs
1
100,627
12,541,716,684
IssuesEvent
2020-06-05 12:52:11
XAMLMarkupExtensions/WPFLocalizationExtension
https://api.github.com/repos/XAMLMarkupExtensions/WPFLocalizationExtension
opened
Check Binding in Binding in Setter & Docu
Designtime Problem Docu Enhancement
designtime support in setters with non Binding Markupelement Check if this is working ```xml <Setter Property="Text" Value="{Binding Source={lex:Loc {Binding test}}}" /> ``` Docu the following cool workaround many thanks @karnah ```xml <TextBlock FontSize="20"> <TextBlock.Style> <Style TargetType="TextBlock"> <Setter Property="Text" Value="{Binding Source={lex:Loc en}}" /> <Style.Triggers> <DataTrigger Binding="{Binding ElementName=CheckBox, Path=IsChecked}" Value="True"> <Setter Property="Text" Value="{Binding Source={lex:Loc de}}" /> </DataTrigger> </Style.Triggers> </Style> </TextBlock.Style> </TextBlock> ```
1.0
Check Binding in Binding in Setter & Docu - designtime support in setters with non Binding Markupelement Check if this is working ```xml <Setter Property="Text" Value="{Binding Source={lex:Loc {Binding test}}}" /> ``` Docu the following cool workaround many thanks @karnah ```xml <TextBlock FontSize="20"> <TextBlock.Style> <Style TargetType="TextBlock"> <Setter Property="Text" Value="{Binding Source={lex:Loc en}}" /> <Style.Triggers> <DataTrigger Binding="{Binding ElementName=CheckBox, Path=IsChecked}" Value="True"> <Setter Property="Text" Value="{Binding Source={lex:Loc de}}" /> </DataTrigger> </Style.Triggers> </Style> </TextBlock.Style> </TextBlock> ```
non_defect
check binding in binding in setter docu designtime support in setters with non binding markupelement check if this is working xml docu the following cool workaround many thanks karnah xml
0
53,629
13,261,998,102
IssuesEvent
2020-08-20 20:55:10
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[clsim] pr2 is never used (Trac #1783)
Migrated from Trac cmake defect
unused varialbe cought by static analysis http://software.icecube.wisc.edu/static_analysis/00_LATEST/report-0ca23c.html#EndPath Also should there be an else at line 212? <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1783">https://code.icecube.wisc.edu/projects/icecube/ticket/1783</a>, reported by kjmeagherand owned by kjmeagher</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:15:08", "_ts": "1550067308113782", "description": "unused varialbe cought by static analysis http://software.icecube.wisc.edu/static_analysis/00_LATEST/report-0ca23c.html#EndPath\n\nAlso should there be an else at line 212?", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2016-07-21T17:16:57", "component": "cmake", "summary": "[clsim] pr2 is never used", "priority": "normal", "keywords": "", "milestone": "", "owner": "kjmeagher", "type": "defect" } ``` </p> </details>
1.0
[clsim] pr2 is never used (Trac #1783) - unused varialbe cought by static analysis http://software.icecube.wisc.edu/static_analysis/00_LATEST/report-0ca23c.html#EndPath Also should there be an else at line 212? <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1783">https://code.icecube.wisc.edu/projects/icecube/ticket/1783</a>, reported by kjmeagherand owned by kjmeagher</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:15:08", "_ts": "1550067308113782", "description": "unused varialbe cought by static analysis http://software.icecube.wisc.edu/static_analysis/00_LATEST/report-0ca23c.html#EndPath\n\nAlso should there be an else at line 212?", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2016-07-21T17:16:57", "component": "cmake", "summary": "[clsim] pr2 is never used", "priority": "normal", "keywords": "", "milestone": "", "owner": "kjmeagher", "type": "defect" } ``` </p> </details>
defect
is never used trac unused varialbe cought by static analysis also should there be an else at line migrated from json status closed changetime ts description unused varialbe cought by static analysis should there be an else at line reporter kjmeagher cc resolution fixed time component cmake summary is never used priority normal keywords milestone owner kjmeagher type defect
1
21,817
3,561,887,460
IssuesEvent
2016-01-24 03:29:29
ariya/phantomjs
https://api.github.com/repos/ariya/phantomjs
closed
Linux Redhat6 Enterprize version 1. 6 install... Korean site font broken
old.Priority-Medium old.Status-New old.Type-Defect
_**[yan...@mz.co.kr](http://code.google.com/u/109652123093259324656/) commented:**_ > <b>Which version of PhantomJS are you using? Tip: run 'phantomjs --version'.</b> > /root/phantomjs-1.6.1/bin/phantomjs /root/phantomjs-1.6.1/examples/rasterize.js http://www.daum.net daum.png > > <b>What steps will reproduce the problem?</b> 1. Hangul some broken > <b>2.</b> <b>3.</b> > <b>What is the expected output? What do you see instead?</b> > <b>Which operating system are you using?</b> > <b>Did you use binary PhantomJS or did you compile it from source?</b> > <b>Please provide any additional information below.</b> **Disclaimer:** This issue was migrated on 2013-03-15 from the project's former issue tracker on Google Code, [Issue #742](http://code.google.com/p/phantomjs/issues/detail?id=742). :star2: &nbsp; **3** people had starred this issue at the time of migration.
1.0
Linux Redhat6 Enterprize version 1. 6 install... Korean site font broken - _**[yan...@mz.co.kr](http://code.google.com/u/109652123093259324656/) commented:**_ > <b>Which version of PhantomJS are you using? Tip: run 'phantomjs --version'.</b> > /root/phantomjs-1.6.1/bin/phantomjs /root/phantomjs-1.6.1/examples/rasterize.js http://www.daum.net daum.png > > <b>What steps will reproduce the problem?</b> 1. Hangul some broken > <b>2.</b> <b>3.</b> > <b>What is the expected output? What do you see instead?</b> > <b>Which operating system are you using?</b> > <b>Did you use binary PhantomJS or did you compile it from source?</b> > <b>Please provide any additional information below.</b> **Disclaimer:** This issue was migrated on 2013-03-15 from the project's former issue tracker on Google Code, [Issue #742](http://code.google.com/p/phantomjs/issues/detail?id=742). :star2: &nbsp; **3** people had starred this issue at the time of migration.
defect
linux enterprize version install korean site font broken commented which version of phantomjs are you using tip run phantomjs version root phantomjs bin phantomjs root phantomjs examples rasterize js daum png what steps will reproduce the problem hangul some broken what is the expected output what do you see instead which operating system are you using did you use binary phantomjs or did you compile it from source please provide any additional information below disclaimer this issue was migrated on from the project s former issue tracker on google code nbsp people had starred this issue at the time of migration
1
71,015
23,411,386,814
IssuesEvent
2022-08-12 17:57:06
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
508-defect-3: Links that open in a new tab should warn users ahead of time
508/Accessibility 508-defect-3 Letters benefits-crew benefits-team-1 squad-1
### Point of contact Josh Kim ### Severity level 3, Moderate. Should be fixed in 1-3 sprints post-launch. ### Details Several issues exist when forcing a link to open in a new tab or window: - Screen reader users, screen magnifier users, and users with certain cognitive impairments can become disoriented when they are taken to a new window - The use of the back button, the most used button within the browser, is broken. - Certain user agents, like kiosks, are unable to open new windows or tabs. From WCAG guidance: > “In general, it is better not to open new windows and tabs since they can be disorienting for people, especially people who have difficulty perceiving visual content.” [G200: Opening new windows and tabs from a link only when necessary](https://www.w3.org/TR/WCAG20-TECHS/G200.html) For more, [view our guidance on links (google slides)](https://docs.google.com/presentation/d/14AwUoFOg_Ix1Txl9oePtPD1ntOpylBWwBJMYq9f5vmo/edit?usp=sharing) <img width="1499" alt="links" src="https://user-images.githubusercontent.com/14154792/184416236-c98697d0-067f-4fbc-b01a-06969ff52a95.png"> <img width="1541" alt="links-1" src="https://user-images.githubusercontent.com/14154792/184416239-81b66f5c-5b31-4ad1-8496-59f177c48970.png"> <img width="1530" alt="links2" src="https://user-images.githubusercontent.com/14154792/184416241-abdf70db-feb6-4d29-a0cf-baa4e52100d2.png"> ### Reproduction steps 1. [Sign in as user 299 and go to step 2 of 2 letter list](https://staging.va.gov/records/download-va-letters/letters/letter-list) 2. Click on any of the links, confirm they all open new tabs without warning ### Proposed solution or next steps We can either: - Don't open the links in a new tab (which may be undesired if users will be signed out and can't directly return back to it) - Provide a warning that links will open in a new tab by appending either: - a text warning (opens in new tab) - an external link icon with alt text that states the prior ### References, articles, or WCAG support 1. 2. 3. ... ### Type of issue - [ ] Screenreader - [ ] Keyboard - [ ] Focus - [ ] Headings - [ ] Color - [ ] Zoom - [ ] Semantics - [ ] Axe-core - [ ] Something else
1.0
508-defect-3: Links that open in a new tab should warn users ahead of time - ### Point of contact Josh Kim ### Severity level 3, Moderate. Should be fixed in 1-3 sprints post-launch. ### Details Several issues exist when forcing a link to open in a new tab or window: - Screen reader users, screen magnifier users, and users with certain cognitive impairments can become disoriented when they are taken to a new window - The use of the back button, the most used button within the browser, is broken. - Certain user agents, like kiosks, are unable to open new windows or tabs. From WCAG guidance: > “In general, it is better not to open new windows and tabs since they can be disorienting for people, especially people who have difficulty perceiving visual content.” [G200: Opening new windows and tabs from a link only when necessary](https://www.w3.org/TR/WCAG20-TECHS/G200.html) For more, [view our guidance on links (google slides)](https://docs.google.com/presentation/d/14AwUoFOg_Ix1Txl9oePtPD1ntOpylBWwBJMYq9f5vmo/edit?usp=sharing) <img width="1499" alt="links" src="https://user-images.githubusercontent.com/14154792/184416236-c98697d0-067f-4fbc-b01a-06969ff52a95.png"> <img width="1541" alt="links-1" src="https://user-images.githubusercontent.com/14154792/184416239-81b66f5c-5b31-4ad1-8496-59f177c48970.png"> <img width="1530" alt="links2" src="https://user-images.githubusercontent.com/14154792/184416241-abdf70db-feb6-4d29-a0cf-baa4e52100d2.png"> ### Reproduction steps 1. [Sign in as user 299 and go to step 2 of 2 letter list](https://staging.va.gov/records/download-va-letters/letters/letter-list) 2. Click on any of the links, confirm they all open new tabs without warning ### Proposed solution or next steps We can either: - Don't open the links in a new tab (which may be undesired if users will be signed out and can't directly return back to it) - Provide a warning that links will open in a new tab by appending either: - a text warning (opens in new tab) - an external link icon with alt text that states the prior ### References, articles, or WCAG support 1. 2. 3. ... ### Type of issue - [ ] Screenreader - [ ] Keyboard - [ ] Focus - [ ] Headings - [ ] Color - [ ] Zoom - [ ] Semantics - [ ] Axe-core - [ ] Something else
defect
defect links that open in a new tab should warn users ahead of time point of contact josh kim severity level moderate should be fixed in sprints post launch details several issues exist when forcing a link to open in a new tab or window screen reader users screen magnifier users and users with certain cognitive impairments can become disoriented when they are taken to a new window the use of the back button the most used button within the browser is broken certain user agents like kiosks are unable to open new windows or tabs from wcag guidance “in general it is better not to open new windows and tabs since they can be disorienting for people especially people who have difficulty perceiving visual content ” for more img width alt links src img width alt links src img width alt src reproduction steps click on any of the links confirm they all open new tabs without warning proposed solution or next steps we can either don t open the links in a new tab which may be undesired if users will be signed out and can t directly return back to it provide a warning that links will open in a new tab by appending either a text warning opens in new tab an external link icon with alt text that states the prior references articles or wcag support type of issue screenreader keyboard focus headings color zoom semantics axe core something else
1
34,372
7,447,865,941
IssuesEvent
2018-03-28 13:48:19
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
Isikunimed 2 väljana importida Excelist
C: AIS P: highest R: fixed T: defect
**Reported by aadikaljuvee on 19 Sep 2012 13:32 UTC** Isiku nimed on korrastamata kujul kuigi on olemas korrastatud andmetega tabelid, kus eraldatud on Eesnimi ja Perenimi.
1.0
Isikunimed 2 väljana importida Excelist - **Reported by aadikaljuvee on 19 Sep 2012 13:32 UTC** Isiku nimed on korrastamata kujul kuigi on olemas korrastatud andmetega tabelid, kus eraldatud on Eesnimi ja Perenimi.
defect
isikunimed väljana importida excelist reported by aadikaljuvee on sep utc isiku nimed on korrastamata kujul kuigi on olemas korrastatud andmetega tabelid kus eraldatud on eesnimi ja perenimi
1
75,455
25,856,419,534
IssuesEvent
2022-12-13 14:05:17
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
"You can't specify target table '...' for update in FROM clause" when target table has index hint in MySQL
T: Defect C: Functionality C: DB: MySQL P: Medium E: All Editions
When using the `USE INDEX` clause, or a similar index hint in MySQL, then the fix for #6583 doesn't work. The reason is the same as #14387. We only traverse the join tree to find "unaliased" tables, not also "unwrapped" ones. Thus, this query is produced in an integration test: ```sql update `test`.`t_author` set `test`.`t_author`.`LAST_NAME` = ( select min(`test`.`t_author`.`FIRST_NAME`) from `test`.`t_author` use index (`i_author_name`) ) where `test`.`t_author`.`ID` = ( select max(`test`.`t_author`.`ID`) from `test`.`t_author` use index (`i_author_name`) ) ``` Instead of: ```sql update `test`.`t_author` set `test`.`t_author`.`LAST_NAME` = ( select * from ( select min(`test`.`t_author`.`FIRST_NAME`) from `test`.`t_author` use index (`i_author_name`) ) t ) where `test`.`t_author`.`ID` = ( select * from ( select max(`test`.`t_author`.`ID`) from `test`.`t_author` use index (`i_author_name`) ) t ) ``` Exception: ``` org.jooq.exception.DataAccessException: SQL [update `test`.`t_author` set `test`.`t_author`.`LAST_NAME` = (select min(`test`.`t_author`.`FIRST_NAME`) from `test`.`t_author` use index (`i_author_name`)) where `test`.`t_author`.`ID` = (select max(`test`.`t_author`.`ID`) from `test`.`t_author` use index (`i_author_name`))]; You can't specify target table 't_author' for update in FROM clause at org.jooq_3.18.0-SNAPSHOT.MYSQL.debug(Unknown Source) at org.jooq.impl.Tools.translate(Tools.java:3389) at org.jooq.impl.DefaultExecuteContext.sqlException(DefaultExecuteContext.java:746) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:357) at org.jooq.impl.AbstractDelegatingQuery.execute(AbstractDelegatingQuery.java:115) at org.jooq.test.all.testcases.InsertUpdateTests.testDMLIntoTWithSubqueryFromT0(InsertUpdateTests.java:390) at org.jooq.test.all.testcases.InsertUpdateTests.testDMLIntoTWithSubqueryFromTWithTableHints(InsertUpdateTests.java:359) at org.jooq.test.jOOQAbstractTest.testDMLIntoTWithSubqueryFromTWithTableHints(jOOQAbstractTest.java:2391) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:93) at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:40) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:529) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:756) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:452) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:210) Caused by: java.sql.SQLException: You can't specify target table 't_author' for update in FROM clause at com.mysql.cj.jdbc.exceptions.SQLError.createSQLException(SQLError.java:129) at com.mysql.cj.jdbc.exceptions.SQLExceptionsMapping.translateException(SQLExceptionsMapping.java:122) at com.mysql.cj.jdbc.ClientPreparedStatement.executeInternal(ClientPreparedStatement.java:916) at com.mysql.cj.jdbc.ClientPreparedStatement.execute(ClientPreparedStatement.java:354) at org.jooq.tools.jdbc.DefaultPreparedStatement.execute(DefaultPreparedStatement.java:219) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:422) at org.jooq.impl.AbstractDMLQuery.execute(AbstractDMLQuery.java:980) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:343) ... 36 more ```
1.0
"You can't specify target table '...' for update in FROM clause" when target table has index hint in MySQL - When using the `USE INDEX` clause, or a similar index hint in MySQL, then the fix for #6583 doesn't work. The reason is the same as #14387. We only traverse the join tree to find "unaliased" tables, not also "unwrapped" ones. Thus, this query is produced in an integration test: ```sql update `test`.`t_author` set `test`.`t_author`.`LAST_NAME` = ( select min(`test`.`t_author`.`FIRST_NAME`) from `test`.`t_author` use index (`i_author_name`) ) where `test`.`t_author`.`ID` = ( select max(`test`.`t_author`.`ID`) from `test`.`t_author` use index (`i_author_name`) ) ``` Instead of: ```sql update `test`.`t_author` set `test`.`t_author`.`LAST_NAME` = ( select * from ( select min(`test`.`t_author`.`FIRST_NAME`) from `test`.`t_author` use index (`i_author_name`) ) t ) where `test`.`t_author`.`ID` = ( select * from ( select max(`test`.`t_author`.`ID`) from `test`.`t_author` use index (`i_author_name`) ) t ) ``` Exception: ``` org.jooq.exception.DataAccessException: SQL [update `test`.`t_author` set `test`.`t_author`.`LAST_NAME` = (select min(`test`.`t_author`.`FIRST_NAME`) from `test`.`t_author` use index (`i_author_name`)) where `test`.`t_author`.`ID` = (select max(`test`.`t_author`.`ID`) from `test`.`t_author` use index (`i_author_name`))]; You can't specify target table 't_author' for update in FROM clause at org.jooq_3.18.0-SNAPSHOT.MYSQL.debug(Unknown Source) at org.jooq.impl.Tools.translate(Tools.java:3389) at org.jooq.impl.DefaultExecuteContext.sqlException(DefaultExecuteContext.java:746) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:357) at org.jooq.impl.AbstractDelegatingQuery.execute(AbstractDelegatingQuery.java:115) at org.jooq.test.all.testcases.InsertUpdateTests.testDMLIntoTWithSubqueryFromT0(InsertUpdateTests.java:390) at org.jooq.test.all.testcases.InsertUpdateTests.testDMLIntoTWithSubqueryFromTWithTableHints(InsertUpdateTests.java:359) at org.jooq.test.jOOQAbstractTest.testDMLIntoTWithSubqueryFromTWithTableHints(jOOQAbstractTest.java:2391) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:93) at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:40) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:529) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:756) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:452) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:210) Caused by: java.sql.SQLException: You can't specify target table 't_author' for update in FROM clause at com.mysql.cj.jdbc.exceptions.SQLError.createSQLException(SQLError.java:129) at com.mysql.cj.jdbc.exceptions.SQLExceptionsMapping.translateException(SQLExceptionsMapping.java:122) at com.mysql.cj.jdbc.ClientPreparedStatement.executeInternal(ClientPreparedStatement.java:916) at com.mysql.cj.jdbc.ClientPreparedStatement.execute(ClientPreparedStatement.java:354) at org.jooq.tools.jdbc.DefaultPreparedStatement.execute(DefaultPreparedStatement.java:219) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:422) at org.jooq.impl.AbstractDMLQuery.execute(AbstractDMLQuery.java:980) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:343) ... 36 more ```
defect
you can t specify target table for update in from clause when target table has index hint in mysql when using the use index clause or a similar index hint in mysql then the fix for doesn t work the reason is the same as we only traverse the join tree to find unaliased tables not also unwrapped ones thus this query is produced in an integration test sql update test t author set test t author last name select min test t author first name from test t author use index i author name where test t author id select max test t author id from test t author use index i author name instead of sql update test t author set test t author last name select from select min test t author first name from test t author use index i author name t where test t author id select from select max test t author id from test t author use index i author name t exception org jooq exception dataaccessexception sql you can t specify target table t author for update in from clause at org jooq snapshot mysql debug unknown source at org jooq impl tools translate tools java at org jooq impl defaultexecutecontext sqlexception defaultexecutecontext java at org jooq impl abstractquery execute abstractquery java at org jooq impl abstractdelegatingquery execute abstractdelegatingquery java at org jooq test all testcases insertupdatetests insertupdatetests java at org jooq test all testcases insertupdatetests testdmlintotwithsubqueryfromtwithtablehints insertupdatetests java at org jooq test jooqabstracttest testdmlintotwithsubqueryfromtwithtablehints jooqabstracttest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit rules testwatcher evaluate testwatcher java at org junit rules testwatcher evaluate testwatcher java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org eclipse jdt internal runner run java at org eclipse jdt internal junit runner testexecution run testexecution java at org eclipse jdt internal junit runner remotetestrunner runtests remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner runtests remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner run remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner main remotetestrunner java caused by java sql sqlexception you can t specify target table t author for update in from clause at com mysql cj jdbc exceptions sqlerror createsqlexception sqlerror java at com mysql cj jdbc exceptions sqlexceptionsmapping translateexception sqlexceptionsmapping java at com mysql cj jdbc clientpreparedstatement executeinternal clientpreparedstatement java at com mysql cj jdbc clientpreparedstatement execute clientpreparedstatement java at org jooq tools jdbc defaultpreparedstatement execute defaultpreparedstatement java at org jooq impl abstractquery execute abstractquery java at org jooq impl abstractdmlquery execute abstractdmlquery java at org jooq impl abstractquery execute abstractquery java more
1
65,791
19,694,627,944
IssuesEvent
2022-01-12 10:48:06
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Message bubbles have too much padding on the left of the timeline
T-Defect S-Minor A-Message-Bubbles O-Occasional
### Steps to reproduce 1. Switch from Modern layout to Message Bubbles in the settings ### Outcome #### What did you expect? ![Screenshot from 2022-01-12 10-47-05](https://user-images.githubusercontent.com/51663/149126164-a77640a8-cf2c-4686-b011-672650224eaf.png) #### What happened instead? ![Screenshot from 2022-01-12 10-44-41](https://user-images.githubusercontent.com/51663/149126053-5f942c1f-361c-452f-8eee-5d1786f9348a.png) ### Operating system Arch Linux ### Browser information Chromium 96.0.4664.45 (Official Build) Arch Linux (64-bit) ### URL for webapp develop.element.io ### Application version Element version: ef9335697e05-react-b8355883313e-js-6fc586598a98 Olm version: 3.2.8 ### Homeserver _No response_ ### Will you send logs? No
1.0
Message bubbles have too much padding on the left of the timeline - ### Steps to reproduce 1. Switch from Modern layout to Message Bubbles in the settings ### Outcome #### What did you expect? ![Screenshot from 2022-01-12 10-47-05](https://user-images.githubusercontent.com/51663/149126164-a77640a8-cf2c-4686-b011-672650224eaf.png) #### What happened instead? ![Screenshot from 2022-01-12 10-44-41](https://user-images.githubusercontent.com/51663/149126053-5f942c1f-361c-452f-8eee-5d1786f9348a.png) ### Operating system Arch Linux ### Browser information Chromium 96.0.4664.45 (Official Build) Arch Linux (64-bit) ### URL for webapp develop.element.io ### Application version Element version: ef9335697e05-react-b8355883313e-js-6fc586598a98 Olm version: 3.2.8 ### Homeserver _No response_ ### Will you send logs? No
defect
message bubbles have too much padding on the left of the timeline steps to reproduce switch from modern layout to message bubbles in the settings outcome what did you expect what happened instead operating system arch linux browser information chromium official build arch linux bit url for webapp develop element io application version element version react js olm version homeserver no response will you send logs no
1
786,863
27,696,715,900
IssuesEvent
2023-03-14 03:11:46
mudkipdev/leaf
https://api.github.com/repos/mudkipdev/leaf
opened
Tag Aliases
Type: Feature Status: Planned Priority: Low
**I have not figured out how this feature should work exactly,** but users should be able to create tags that point to another tag. I'm thinking aliases will be independent from the tag, and can be separated from the main tag entirely and have its own content. This will probably add a `/tags alias <alias: str> <tag: str>` that points the alias tag to the main tag. In addition, to create an alias, you can use a new option in the `/tags create` command named "alias_to" or something similar that will skip the content and create the tag instantly. Part of #1.
1.0
Tag Aliases - **I have not figured out how this feature should work exactly,** but users should be able to create tags that point to another tag. I'm thinking aliases will be independent from the tag, and can be separated from the main tag entirely and have its own content. This will probably add a `/tags alias <alias: str> <tag: str>` that points the alias tag to the main tag. In addition, to create an alias, you can use a new option in the `/tags create` command named "alias_to" or something similar that will skip the content and create the tag instantly. Part of #1.
non_defect
tag aliases i have not figured out how this feature should work exactly but users should be able to create tags that point to another tag i m thinking aliases will be independent from the tag and can be separated from the main tag entirely and have its own content this will probably add a tags alias that points the alias tag to the main tag in addition to create an alias you can use a new option in the tags create command named alias to or something similar that will skip the content and create the tag instantly part of
0
9,429
2,615,149,688
IssuesEvent
2015-03-01 06:26:47
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
closed
WARNING: Receive timeout occurred - WPS transaction failed, re-trying last pin
auto-migrated Priority-Low Type-Defect
``` Hi guys, Playing about with my router trying to get this working. It was trying PINs up until 2 days ago, it seems to have a permanent lock out or something? I have pasted the console commands and have attached a cap - Anyone got any ideas? root@bt:~# reaver -i mon0 -b 00:22:75:B8:98:41 -w -r 5:60 -l 300 -c 6 -L -vv Reaver v1.4 WiFi Protected Setup Attack Tool Copyright (c) 2011, Tactical Network Solutions, Craig Heffner <cheffner@tacnetsol.com> [+] Waiting for beacon from 00:22:75:B8:98:41 [+] Associated with 00:22:75:B8:98:41 [+] Trying pin 12345670 [+] Sending EAPOL START request [+] Sending identity response [+] Sending identity response [!] WARNING: Receive timeout occurred [!] WPS transaction failed, re-trying last pin [+] Trying pin 12345670 [+] Sending EAPOL START request [+] Sending identity response [+] Sending identity response [!] WARNING: Receive timeout occurred [!] WPS transaction failed, re-trying last pin [+] Trying pin 12345670 [+] Sending EAPOL START request [+] Sending identity response [!] WARNING: Receive timeout occurred [!] WPS transaction failed, re-trying last pin [+] Trying pin 12345670 [+] Sending EAPOL START request [+] Sending identity response [!] WARNING: Receive timeout occurred [!] WPS transaction failed, re-trying last pin [+] Trying pin 12345670 [+] Sending EAPOL START request [+] Sending identity response [+] Sending identity response [+] Sending identity response [+] Sending identity response [!] WARNING: Receive timeout occurred [!] WPS transaction failed, re-trying last pin [+] Entering recurring delay of 60 seconds ^C [+] Nothing done, nothing to save. root@bt:~# ``` Original issue reported on code.google.com by `themosts...@googlemail.com` on 12 Jan 2012 at 7:39 Attachments: * [ReaverDUMP-02.cap](https://storage.googleapis.com/google-code-attachments/reaver-wps/issue-132/comment-0/ReaverDUMP-02.cap)
1.0
WARNING: Receive timeout occurred - WPS transaction failed, re-trying last pin - ``` Hi guys, Playing about with my router trying to get this working. It was trying PINs up until 2 days ago, it seems to have a permanent lock out or something? I have pasted the console commands and have attached a cap - Anyone got any ideas? root@bt:~# reaver -i mon0 -b 00:22:75:B8:98:41 -w -r 5:60 -l 300 -c 6 -L -vv Reaver v1.4 WiFi Protected Setup Attack Tool Copyright (c) 2011, Tactical Network Solutions, Craig Heffner <cheffner@tacnetsol.com> [+] Waiting for beacon from 00:22:75:B8:98:41 [+] Associated with 00:22:75:B8:98:41 [+] Trying pin 12345670 [+] Sending EAPOL START request [+] Sending identity response [+] Sending identity response [!] WARNING: Receive timeout occurred [!] WPS transaction failed, re-trying last pin [+] Trying pin 12345670 [+] Sending EAPOL START request [+] Sending identity response [+] Sending identity response [!] WARNING: Receive timeout occurred [!] WPS transaction failed, re-trying last pin [+] Trying pin 12345670 [+] Sending EAPOL START request [+] Sending identity response [!] WARNING: Receive timeout occurred [!] WPS transaction failed, re-trying last pin [+] Trying pin 12345670 [+] Sending EAPOL START request [+] Sending identity response [!] WARNING: Receive timeout occurred [!] WPS transaction failed, re-trying last pin [+] Trying pin 12345670 [+] Sending EAPOL START request [+] Sending identity response [+] Sending identity response [+] Sending identity response [+] Sending identity response [!] WARNING: Receive timeout occurred [!] WPS transaction failed, re-trying last pin [+] Entering recurring delay of 60 seconds ^C [+] Nothing done, nothing to save. root@bt:~# ``` Original issue reported on code.google.com by `themosts...@googlemail.com` on 12 Jan 2012 at 7:39 Attachments: * [ReaverDUMP-02.cap](https://storage.googleapis.com/google-code-attachments/reaver-wps/issue-132/comment-0/ReaverDUMP-02.cap)
defect
warning receive timeout occurred wps transaction failed re trying last pin hi guys playing about with my router trying to get this working it was trying pins up until days ago it seems to have a permanent lock out or something i have pasted the console commands and have attached a cap anyone got any ideas root bt reaver i b w r l c l vv reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner waiting for beacon from associated with trying pin sending eapol start request sending identity response sending identity response warning receive timeout occurred wps transaction failed re trying last pin trying pin sending eapol start request sending identity response sending identity response warning receive timeout occurred wps transaction failed re trying last pin trying pin sending eapol start request sending identity response warning receive timeout occurred wps transaction failed re trying last pin trying pin sending eapol start request sending identity response warning receive timeout occurred wps transaction failed re trying last pin trying pin sending eapol start request sending identity response sending identity response sending identity response sending identity response warning receive timeout occurred wps transaction failed re trying last pin entering recurring delay of seconds c nothing done nothing to save root bt original issue reported on code google com by themosts googlemail com on jan at attachments
1
192,593
14,620,732,538
IssuesEvent
2020-12-22 20:16:40
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[Maceió/Remoto] Back-end Developer @Ilhasoft
Django Docker Kubernetes Linux PJ PostgreSQL Testes automatizados
## Nossa empresa Somos especialistas em desenvolver chatbots que utilizam os mais sofisticados recursos da Inteligência Artificial para automatizar a comunicação entre organizações e pessoas ao redor do mundo. ### Onde queremos chegar? Crescer com sustentabilidade, permanecendo com nossas raízes. Queremos conectar comunidades e fazer com que o uso da tecnologia seja democratizado: alcance a todos, onde quer que seja. Queremos inovar, revolucionar, provocar novos comportamentos. Queremos facilitar o acesso aos melhores serviços de um jeito rápido e prático, sem burocracias. Queremos criar soluções para problemas, otimizar a interação entre empresas e consumidores, levar a igualdade para os lugares mais remotos, trabalhar em parceria com ONGs e Governos para conscientizar as pessoas e dar-lhes mais qualidade de vida. Queremos fazer parte de uma nova era sem sair da nossa terra! ### Panorama atual A pandemia do COVID-19 tem exigido bastante do nosso time. Nossos produtos assumiram um papel de destaque na luta contra o vírus, atualmente estamos ajudando a salvar milhares de vidas em estados e municípios no cenário nacional e internacional. Conheça alguns projetos: * Health Buddy: Chatbot informativo multi-idiomas criado para UNICEF e para a OMS (http://healthbuddy.info/); * Governo do Ceará: Chatbot informativo, com atendimento ao paciente por profissionais de saúde em casos graves (https://www.ceara.gov.br/) * Governo da Bahia: Chatbot informativo, com atendimento ao paciente por profissionais de saúde em casos graves (WhatsApp: [71 3118-8499](tel:71 3118-8499)) * Governo de Alagoas: Chatbot informativo, com atendimento ao paciente por profissionais de saúde em casos graves (WhatsApp: [82 3315-1532](tel:82 3315-1532)) ## Descrição da vaga Como Django Back-end Developer, você integrará nosso time no desenvolvimento de software para correção de bugs, desenvolvimento de novas funcionalidades nas plataformas de criação de gestão de Chatbots da Ilhasoft. No fundo, nós queremos você no time para: * Crescer conosco à medida que desbravamos novos projetos * Ter orgulho do trabalho que desenvolve * Melhorar a cada dia * Ter a certeza de ter contribuído para problemas de nossos clientes após um dia de trabalho desafiador * Ajudar todos seus colegas de trabalho a completar o check list acima ## Local Após a pandemia a Ilhasoft se tornou uma empresa 100% remota com uma equipe distribuída em vários países. ## Requisitos **Obrigatórios:** * 1 ano de experiência com Django * 1 ano de experiência com PostgreSQL * Experiência prévia com Gitflow **Diferenciais:** * Um projeto próprio desenvolvido em Django * Projetos de código aberto. * Conhecimento em servidores linux, Docker, Kubernetes * Fluência em Inglês * Testes Automatizados ## Contratação CLT ou PJ a combinar ## Como se candidatar Cadastre-se no seguinte link: https://ilhasoft.gupy.io/jobs/600315 ## Tempo médio de feedbacks Costumamos enviar feedbacks em até 3 dias após cada processo. E-mail para contato em caso de não haver resposta: [rh@ilhasoft.com.br](mailto:rh@ilhasoft.com.br) #### Nível * Júnior
1.0
[Maceió/Remoto] Back-end Developer @Ilhasoft - ## Nossa empresa Somos especialistas em desenvolver chatbots que utilizam os mais sofisticados recursos da Inteligência Artificial para automatizar a comunicação entre organizações e pessoas ao redor do mundo. ### Onde queremos chegar? Crescer com sustentabilidade, permanecendo com nossas raízes. Queremos conectar comunidades e fazer com que o uso da tecnologia seja democratizado: alcance a todos, onde quer que seja. Queremos inovar, revolucionar, provocar novos comportamentos. Queremos facilitar o acesso aos melhores serviços de um jeito rápido e prático, sem burocracias. Queremos criar soluções para problemas, otimizar a interação entre empresas e consumidores, levar a igualdade para os lugares mais remotos, trabalhar em parceria com ONGs e Governos para conscientizar as pessoas e dar-lhes mais qualidade de vida. Queremos fazer parte de uma nova era sem sair da nossa terra! ### Panorama atual A pandemia do COVID-19 tem exigido bastante do nosso time. Nossos produtos assumiram um papel de destaque na luta contra o vírus, atualmente estamos ajudando a salvar milhares de vidas em estados e municípios no cenário nacional e internacional. Conheça alguns projetos: * Health Buddy: Chatbot informativo multi-idiomas criado para UNICEF e para a OMS (http://healthbuddy.info/); * Governo do Ceará: Chatbot informativo, com atendimento ao paciente por profissionais de saúde em casos graves (https://www.ceara.gov.br/) * Governo da Bahia: Chatbot informativo, com atendimento ao paciente por profissionais de saúde em casos graves (WhatsApp: [71 3118-8499](tel:71 3118-8499)) * Governo de Alagoas: Chatbot informativo, com atendimento ao paciente por profissionais de saúde em casos graves (WhatsApp: [82 3315-1532](tel:82 3315-1532)) ## Descrição da vaga Como Django Back-end Developer, você integrará nosso time no desenvolvimento de software para correção de bugs, desenvolvimento de novas funcionalidades nas plataformas de criação de gestão de Chatbots da Ilhasoft. No fundo, nós queremos você no time para: * Crescer conosco à medida que desbravamos novos projetos * Ter orgulho do trabalho que desenvolve * Melhorar a cada dia * Ter a certeza de ter contribuído para problemas de nossos clientes após um dia de trabalho desafiador * Ajudar todos seus colegas de trabalho a completar o check list acima ## Local Após a pandemia a Ilhasoft se tornou uma empresa 100% remota com uma equipe distribuída em vários países. ## Requisitos **Obrigatórios:** * 1 ano de experiência com Django * 1 ano de experiência com PostgreSQL * Experiência prévia com Gitflow **Diferenciais:** * Um projeto próprio desenvolvido em Django * Projetos de código aberto. * Conhecimento em servidores linux, Docker, Kubernetes * Fluência em Inglês * Testes Automatizados ## Contratação CLT ou PJ a combinar ## Como se candidatar Cadastre-se no seguinte link: https://ilhasoft.gupy.io/jobs/600315 ## Tempo médio de feedbacks Costumamos enviar feedbacks em até 3 dias após cada processo. E-mail para contato em caso de não haver resposta: [rh@ilhasoft.com.br](mailto:rh@ilhasoft.com.br) #### Nível * Júnior
non_defect
back end developer ilhasoft nossa empresa somos especialistas em desenvolver chatbots que utilizam os mais sofisticados recursos da inteligência artificial para automatizar a comunicação entre organizações e pessoas ao redor do mundo onde queremos chegar crescer com sustentabilidade permanecendo com nossas raízes queremos conectar comunidades e fazer com que o uso da tecnologia seja democratizado alcance a todos onde quer que seja queremos inovar revolucionar provocar novos comportamentos queremos facilitar o acesso aos melhores serviços de um jeito rápido e prático sem burocracias queremos criar soluções para problemas otimizar a interação entre empresas e consumidores levar a igualdade para os lugares mais remotos trabalhar em parceria com ongs e governos para conscientizar as pessoas e dar lhes mais qualidade de vida queremos fazer parte de uma nova era sem sair da nossa terra panorama atual a pandemia do covid tem exigido bastante do nosso time nossos produtos assumiram um papel de destaque na luta contra o vírus atualmente estamos ajudando a salvar milhares de vidas em estados e municípios no cenário nacional e internacional conheça alguns projetos health buddy chatbot informativo multi idiomas criado para unicef e para a oms governo do ceará chatbot informativo com atendimento ao paciente por profissionais de saúde em casos graves governo da bahia chatbot informativo com atendimento ao paciente por profissionais de saúde em casos graves whatsapp tel governo de alagoas chatbot informativo com atendimento ao paciente por profissionais de saúde em casos graves whatsapp tel descrição da vaga como django back end developer você integrará nosso time no desenvolvimento de software para correção de bugs desenvolvimento de novas funcionalidades nas plataformas de criação de gestão de chatbots da ilhasoft no fundo nós queremos você no time para crescer conosco à medida que desbravamos novos projetos ter orgulho do trabalho que desenvolve melhorar a cada dia ter a certeza de ter contribuído para problemas de nossos clientes após um dia de trabalho desafiador ajudar todos seus colegas de trabalho a completar o check list acima local após a pandemia a ilhasoft se tornou uma empresa remota com uma equipe distribuída em vários países requisitos obrigatórios ano de experiência com django ano de experiência com postgresql experiência prévia com gitflow diferenciais um projeto próprio desenvolvido em django projetos de código aberto conhecimento em servidores linux docker kubernetes fluência em inglês testes automatizados contratação clt ou pj a combinar como se candidatar cadastre se no seguinte link tempo médio de feedbacks costumamos enviar feedbacks em até dias após cada processo e mail para contato em caso de não haver resposta mailto rh ilhasoft com br nível júnior
0
37,998
5,157,156,125
IssuesEvent
2017-01-16 04:46:36
containenv/rapcom
https://api.github.com/repos/containenv/rapcom
closed
Test all major functionality.
test
Rapcom needs tests to cover all expected functionality. This is not intended to be exhaustive, but should at least handle the happy paths.
1.0
Test all major functionality. - Rapcom needs tests to cover all expected functionality. This is not intended to be exhaustive, but should at least handle the happy paths.
non_defect
test all major functionality rapcom needs tests to cover all expected functionality this is not intended to be exhaustive but should at least handle the happy paths
0
3,452
6,542,647,946
IssuesEvent
2017-09-02 10:26:27
pwittchen/ReactiveNetwork
https://api.github.com/repos/pwittchen/ReactiveNetwork
opened
Relase 0.12.1 (RxJava2.x)
release process RxJava2.x
**Initial release notes**: Fixed memory leak in `PreLollipopNetworkObservingStrategy` during disposing of an `Observable` - issue #219. **Things to do**: TBD.
1.0
Relase 0.12.1 (RxJava2.x) - **Initial release notes**: Fixed memory leak in `PreLollipopNetworkObservingStrategy` during disposing of an `Observable` - issue #219. **Things to do**: TBD.
non_defect
relase x initial release notes fixed memory leak in prelollipopnetworkobservingstrategy during disposing of an observable issue things to do tbd
0
47,116
13,056,034,204
IssuesEvent
2020-07-30 03:27:14
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
glshovel docs and plugin examples (Trac #15)
Migrated from Trac defect glshovel
Migrated from https://code.icecube.wisc.edu/ticket/15 ```json { "status": "closed", "changetime": "2007-11-11T03:51:18", "description": "\n", "reporter": "troy", "cc": "", "resolution": "duplicate", "_ts": "1194753078000000", "component": "glshovel", "summary": "glshovel docs and plugin examples", "priority": "normal", "keywords": "", "time": "2007-06-03T16:33:00", "milestone": "", "owner": "troy", "type": "defect" } ```
1.0
glshovel docs and plugin examples (Trac #15) - Migrated from https://code.icecube.wisc.edu/ticket/15 ```json { "status": "closed", "changetime": "2007-11-11T03:51:18", "description": "\n", "reporter": "troy", "cc": "", "resolution": "duplicate", "_ts": "1194753078000000", "component": "glshovel", "summary": "glshovel docs and plugin examples", "priority": "normal", "keywords": "", "time": "2007-06-03T16:33:00", "milestone": "", "owner": "troy", "type": "defect" } ```
defect
glshovel docs and plugin examples trac migrated from json status closed changetime description n reporter troy cc resolution duplicate ts component glshovel summary glshovel docs and plugin examples priority normal keywords time milestone owner troy type defect
1
24,436
3,980,349,103
IssuesEvent
2016-05-06 06:59:47
Quantum64/Arcade-Issues
https://api.github.com/repos/Quantum64/Arcade-Issues
opened
Kill Effect not working
cosmeticsmenu defect effects killeffect
The "Notes Kill Effect" does not work. With the other kill effects it shows a piece of redstone but with the notes one it shows paper, Like what you would see with a title. http://i.imgur.com/o7hHeYs.png
1.0
Kill Effect not working - The "Notes Kill Effect" does not work. With the other kill effects it shows a piece of redstone but with the notes one it shows paper, Like what you would see with a title. http://i.imgur.com/o7hHeYs.png
defect
kill effect not working the notes kill effect does not work with the other kill effects it shows a piece of redstone but with the notes one it shows paper like what you would see with a title
1
40,417
9,984,513,603
IssuesEvent
2019-07-10 14:40:07
telus/tds-core
https://api.github.com/repos/telus/tds-core
closed
TDS Tooltip - Accessibility & Language issues
accessibility :wheelchair: priority: medium status: in progress type: defect :bug:
## Description - The `aria-label` inside the icon "Reveal additional information" does not seem to be translatable in French, we can add more text after it but not actually change it. Additionally, the Wave extension seems to flag the button as an error ("Empty label") because it doesn't see any content inside it. - If possible, please translate "Reveal additional information" if it's used in a `/fr/` page to "Afficher l’information supplémentaire". If the "Empty label" issue is not a false flag, there needs to be an actual label inside of the button for screen readers, etc. ## Reproduction Steps 1. View the Tooltip component on a `/fr/` page. 2. `aria-label` remains in English. 3. Scan the page with Wave and the "Empty label" issue pops (might be a false flag?). ## Meta - TDS component version: @tds/core-tooltip”: “^2.0.4 - Willing to develop solution: No - Has workaround: No - High impact: No ## Screenshots ![1](https://user-images.githubusercontent.com/8562738/51937905-c74b7f80-23d9-11e9-9188-292dcb63281a.png)
1.0
TDS Tooltip - Accessibility & Language issues - ## Description - The `aria-label` inside the icon "Reveal additional information" does not seem to be translatable in French, we can add more text after it but not actually change it. Additionally, the Wave extension seems to flag the button as an error ("Empty label") because it doesn't see any content inside it. - If possible, please translate "Reveal additional information" if it's used in a `/fr/` page to "Afficher l’information supplémentaire". If the "Empty label" issue is not a false flag, there needs to be an actual label inside of the button for screen readers, etc. ## Reproduction Steps 1. View the Tooltip component on a `/fr/` page. 2. `aria-label` remains in English. 3. Scan the page with Wave and the "Empty label" issue pops (might be a false flag?). ## Meta - TDS component version: @tds/core-tooltip”: “^2.0.4 - Willing to develop solution: No - Has workaround: No - High impact: No ## Screenshots ![1](https://user-images.githubusercontent.com/8562738/51937905-c74b7f80-23d9-11e9-9188-292dcb63281a.png)
defect
tds tooltip accessibility language issues description the aria label inside the icon reveal additional information does not seem to be translatable in french we can add more text after it but not actually change it additionally the wave extension seems to flag the button as an error empty label because it doesn t see any content inside it if possible please translate reveal additional information if it s used in a fr page to afficher l’information supplémentaire if the empty label issue is not a false flag there needs to be an actual label inside of the button for screen readers etc reproduction steps view the tooltip component on a fr page aria label remains in english scan the page with wave and the empty label issue pops might be a false flag meta tds component version tds core tooltip” “ willing to develop solution no has workaround no high impact no screenshots
1
14,198
8,506,478,649
IssuesEvent
2018-10-30 16:39:57
CompletelyFairGames/dwarfcorp
https://api.github.com/repos/CompletelyFairGames/dwarfcorp
opened
Performance: As soon as ten dwarves are in game, performance dies
A Bug Performance
![image](https://user-images.githubusercontent.com/6562812/47733236-2e706100-dc3e-11e8-9d2b-df8ccce4cdc2.png) ![image](https://user-images.githubusercontent.com/6562812/47733254-34fed880-dc3e-11e8-8d71-06af7973e3ae.png) [Borderland_413_131853894854936125.zip](https://github.com/CompletelyFairGames/dwarfcorp/files/2530519/Borderland_413_131853894854936125.zip) [Borderland.zip](https://github.com/CompletelyFairGames/dwarfcorp/files/2530522/Borderland.zip) [DwarfCorp181030_IPSummary.zip](https://github.com/CompletelyFairGames/dwarfcorp/files/2530542/DwarfCorp181030_IPSummary.zip) Performance is measurably worse. It went from spikes of 60% before to a sustained 80% CPU. On 1x with ten dwarves or so, FPS is staying at 0 FPS. I'm guessing something in our refactor missed something because things were running totally fine until I switched to 64f01d10. This a repro, as far as I can tell: 1. New game. 2. Set a dig task to the bottom 3. God/hire five more miners (for some reason reaching 10 hires appears to be exactly when perf drops. Not 9. Not 8. If this is actually perf related, I'm guessing it'll take more for you all to repro the bug.) 4. Observe game perf drop to 0 FPS after about a minute or two.
True
Performance: As soon as ten dwarves are in game, performance dies - ![image](https://user-images.githubusercontent.com/6562812/47733236-2e706100-dc3e-11e8-9d2b-df8ccce4cdc2.png) ![image](https://user-images.githubusercontent.com/6562812/47733254-34fed880-dc3e-11e8-8d71-06af7973e3ae.png) [Borderland_413_131853894854936125.zip](https://github.com/CompletelyFairGames/dwarfcorp/files/2530519/Borderland_413_131853894854936125.zip) [Borderland.zip](https://github.com/CompletelyFairGames/dwarfcorp/files/2530522/Borderland.zip) [DwarfCorp181030_IPSummary.zip](https://github.com/CompletelyFairGames/dwarfcorp/files/2530542/DwarfCorp181030_IPSummary.zip) Performance is measurably worse. It went from spikes of 60% before to a sustained 80% CPU. On 1x with ten dwarves or so, FPS is staying at 0 FPS. I'm guessing something in our refactor missed something because things were running totally fine until I switched to 64f01d10. This a repro, as far as I can tell: 1. New game. 2. Set a dig task to the bottom 3. God/hire five more miners (for some reason reaching 10 hires appears to be exactly when perf drops. Not 9. Not 8. If this is actually perf related, I'm guessing it'll take more for you all to repro the bug.) 4. Observe game perf drop to 0 FPS after about a minute or two.
non_defect
performance as soon as ten dwarves are in game performance dies performance is measurably worse it went from spikes of before to a sustained cpu on with ten dwarves or so fps is staying at fps i m guessing something in our refactor missed something because things were running totally fine until i switched to this a repro as far as i can tell new game set a dig task to the bottom god hire five more miners for some reason reaching hires appears to be exactly when perf drops not not if this is actually perf related i m guessing it ll take more for you all to repro the bug observe game perf drop to fps after about a minute or two
0
264,908
28,214,112,133
IssuesEvent
2023-04-05 07:39:38
hshivhare67/platform_device_renesas_kernel_v4.19.72
https://api.github.com/repos/hshivhare67/platform_device_renesas_kernel_v4.19.72
closed
CVE-2022-0847 (High) detected in linuxlinux-4.19.279 - autoclosed
Mend: dependency security vulnerability
## CVE-2022-0847 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/iov_iter.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/iov_iter.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the way the "flags" member of the new pipe buffer structure was lacking proper initialization in copy_page_to_iter_pipe and push_pipe functions in the Linux kernel and could thus contain stale values. An unprivileged local user could use this flaw to write to pages in the page cache backed by read only files and as such escalate their privileges on the system. <p>Publish Date: 2022-03-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0847>CVE-2022-0847</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-0847">https://www.linuxkernelcves.com/cves/CVE-2022-0847</a></p> <p>Release Date: 2022-03-10</p> <p>Fix Resolution: v5.10.102,v5.15.25,v5.16.11,v5.17-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-0847 (High) detected in linuxlinux-4.19.279 - autoclosed - ## CVE-2022-0847 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/iov_iter.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/iov_iter.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the way the "flags" member of the new pipe buffer structure was lacking proper initialization in copy_page_to_iter_pipe and push_pipe functions in the Linux kernel and could thus contain stale values. An unprivileged local user could use this flaw to write to pages in the page cache backed by read only files and as such escalate their privileges on the system. <p>Publish Date: 2022-03-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0847>CVE-2022-0847</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-0847">https://www.linuxkernelcves.com/cves/CVE-2022-0847</a></p> <p>Release Date: 2022-03-10</p> <p>Fix Resolution: v5.10.102,v5.15.25,v5.16.11,v5.17-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch main vulnerable source files lib iov iter c lib iov iter c vulnerability details a flaw was found in the way the flags member of the new pipe buffer structure was lacking proper initialization in copy page to iter pipe and push pipe functions in the linux kernel and could thus contain stale values an unprivileged local user could use this flaw to write to pages in the page cache backed by read only files and as such escalate their privileges on the system publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
176,490
28,102,202,746
IssuesEvent
2023-03-30 20:27:36
WordPress/openverse
https://api.github.com/repos/WordPress/openverse
opened
Add instructions for navigating sub-categories in Figma
🟩 priority: low 🌟 goal: addition 🖼️ aspect: design 🧱 stack: frontend
## Description <!-- Describe the feature and how it solves the problem. --> Originally reported by @sarayourfriend > Could figma page have instructions for where to find the sub-categories? I forgot how to find the other pages in Figma and had to click around a bit before I could find it Link: https://www.figma.com/file/GIIQ4sDbaToCfFQyKMvzr8/Openverse-Design-Library?node-id=0-1 ## Alternatives <!-- Describe any alternative solutions or features you have considered. How is this feature better? --> ## Additional context <!-- Add any other context about the feature here; or delete the section entirely. --> <!-- If you would like to work on this, please comment below separately. -->
1.0
Add instructions for navigating sub-categories in Figma - ## Description <!-- Describe the feature and how it solves the problem. --> Originally reported by @sarayourfriend > Could figma page have instructions for where to find the sub-categories? I forgot how to find the other pages in Figma and had to click around a bit before I could find it Link: https://www.figma.com/file/GIIQ4sDbaToCfFQyKMvzr8/Openverse-Design-Library?node-id=0-1 ## Alternatives <!-- Describe any alternative solutions or features you have considered. How is this feature better? --> ## Additional context <!-- Add any other context about the feature here; or delete the section entirely. --> <!-- If you would like to work on this, please comment below separately. -->
non_defect
add instructions for navigating sub categories in figma description originally reported by sarayourfriend could figma page have instructions for where to find the sub categories i forgot how to find the other pages in figma and had to click around a bit before i could find it link alternatives additional context
0
862
2,594,241,063
IssuesEvent
2015-02-20 01:02:03
BALL-Project/ball
https://api.github.com/repos/BALL-Project/ball
closed
BALLView on Windows does not move light sources correctly
C: VIEW P: major R: fixed T: defect
**Reported by akdehof on 26 Aug 39272847 02:13 UTC** In the 1.3-beta1 release on windows, light sources are not correctly adapted when moving the camera around.
1.0
BALLView on Windows does not move light sources correctly - **Reported by akdehof on 26 Aug 39272847 02:13 UTC** In the 1.3-beta1 release on windows, light sources are not correctly adapted when moving the camera around.
defect
ballview on windows does not move light sources correctly reported by akdehof on aug utc in the release on windows light sources are not correctly adapted when moving the camera around
1
53,916
13,262,511,048
IssuesEvent
2020-08-20 21:57:10
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[PROPOSAL] Don't auto-generate tables. (Trac #2329)
Migrated from Trac combo simulation defect
When running a test in sim-services (propagator_state_storage.py) that test takes an unusually long time to finish. I strongly suspect PROPOSAL table generation is the culprit. This test runs for over 10 minutes on my machine. Attached is the output. PROPOSAL should not ever try to auto-generate tables. Demand that the user point to the tables they want to use and throw immediately if it doesn't like what it sees (e.g. empty tables). This is causing problems for me and it's still causing problems for production. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2329">https://code.icecube.wisc.edu/projects/icecube/ticket/2329</a>, reported by olivasand owned by jsoedingrekso</em></summary> <p> ```json { "status": "closed", "changetime": "2019-06-28T14:57:28", "_ts": "1561733848005867", "description": "When running a test in sim-services (propagator_state_storage.py) that test takes an unusually long time to finish. I strongly suspect PROPOSAL table generation is the culprit. This test runs for over 10 minutes on my machine.\n\nAttached is the output.\n\nPROPOSAL should not ever try to auto-generate tables. Demand that the user point to the tables they want to use and throw immediately if it doesn't like what it sees (e.g. empty tables).\n\nThis is causing problems for me and it's still causing problems for production.\n", "reporter": "olivas", "cc": "juancarlos", "resolution": "fixed", "time": "2019-06-20T14:05:00", "component": "combo simulation", "summary": "[PROPOSAL] Don't auto-generate tables.", "priority": "blocker", "keywords": "", "milestone": "Autumnal Equinox 2019", "owner": "jsoedingrekso", "type": "defect" } ``` </p> </details>
1.0
[PROPOSAL] Don't auto-generate tables. (Trac #2329) - When running a test in sim-services (propagator_state_storage.py) that test takes an unusually long time to finish. I strongly suspect PROPOSAL table generation is the culprit. This test runs for over 10 minutes on my machine. Attached is the output. PROPOSAL should not ever try to auto-generate tables. Demand that the user point to the tables they want to use and throw immediately if it doesn't like what it sees (e.g. empty tables). This is causing problems for me and it's still causing problems for production. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2329">https://code.icecube.wisc.edu/projects/icecube/ticket/2329</a>, reported by olivasand owned by jsoedingrekso</em></summary> <p> ```json { "status": "closed", "changetime": "2019-06-28T14:57:28", "_ts": "1561733848005867", "description": "When running a test in sim-services (propagator_state_storage.py) that test takes an unusually long time to finish. I strongly suspect PROPOSAL table generation is the culprit. This test runs for over 10 minutes on my machine.\n\nAttached is the output.\n\nPROPOSAL should not ever try to auto-generate tables. Demand that the user point to the tables they want to use and throw immediately if it doesn't like what it sees (e.g. empty tables).\n\nThis is causing problems for me and it's still causing problems for production.\n", "reporter": "olivas", "cc": "juancarlos", "resolution": "fixed", "time": "2019-06-20T14:05:00", "component": "combo simulation", "summary": "[PROPOSAL] Don't auto-generate tables.", "priority": "blocker", "keywords": "", "milestone": "Autumnal Equinox 2019", "owner": "jsoedingrekso", "type": "defect" } ``` </p> </details>
defect
don t auto generate tables trac when running a test in sim services propagator state storage py that test takes an unusually long time to finish i strongly suspect proposal table generation is the culprit this test runs for over minutes on my machine attached is the output proposal should not ever try to auto generate tables demand that the user point to the tables they want to use and throw immediately if it doesn t like what it sees e g empty tables this is causing problems for me and it s still causing problems for production migrated from json status closed changetime ts description when running a test in sim services propagator state storage py that test takes an unusually long time to finish i strongly suspect proposal table generation is the culprit this test runs for over minutes on my machine n nattached is the output n nproposal should not ever try to auto generate tables demand that the user point to the tables they want to use and throw immediately if it doesn t like what it sees e g empty tables n nthis is causing problems for me and it s still causing problems for production n reporter olivas cc juancarlos resolution fixed time component combo simulation summary don t auto generate tables priority blocker keywords milestone autumnal equinox owner jsoedingrekso type defect
1
32,666
7,569,910,129
IssuesEvent
2018-04-23 07:10:17
pywbem/pywbem
https://api.github.com/repos/pywbem/pywbem
closed
Deprecate tocimobj()
area: code release: optional resolution: fixed type: cleanup
cim_obj.py contains this TODO: ``` W:5450, 0: TODO: Move remaining internal uses of tocimobj() to cimvalue() and deprecate (fixme) ``` `cimvalue()` is a new function added in 0.12 that does pretty much what `tocimobj()` does, but with a cleaner approach to handling the different types. Because `tocimobj()` is an external functions and there are several subtle differences in type handling, a new function was created instead of risking an incompatibility for users. The `tocimobj()` function should be deprecated, and all remaining internal uses of it should migrate to using `cimvalue()`.
1.0
Deprecate tocimobj() - cim_obj.py contains this TODO: ``` W:5450, 0: TODO: Move remaining internal uses of tocimobj() to cimvalue() and deprecate (fixme) ``` `cimvalue()` is a new function added in 0.12 that does pretty much what `tocimobj()` does, but with a cleaner approach to handling the different types. Because `tocimobj()` is an external functions and there are several subtle differences in type handling, a new function was created instead of risking an incompatibility for users. The `tocimobj()` function should be deprecated, and all remaining internal uses of it should migrate to using `cimvalue()`.
non_defect
deprecate tocimobj cim obj py contains this todo w todo move remaining internal uses of tocimobj to cimvalue and deprecate fixme cimvalue is a new function added in that does pretty much what tocimobj does but with a cleaner approach to handling the different types because tocimobj is an external functions and there are several subtle differences in type handling a new function was created instead of risking an incompatibility for users the tocimobj function should be deprecated and all remaining internal uses of it should migrate to using cimvalue
0
7,012
2,610,321,952
IssuesEvent
2015-02-26 19:43:50
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Gameplay Error
auto-migrated Priority-Medium Type-Defect
``` I believe there are different observations already made by other testers, some said it works, some said it not works. For me, if i activate this ability of the Light Assault Cruiser, it gives me the Beep sound as if the area is not accesible or it is out of range etc. However, the Cruiser responds with a standard "attack phrase". After the Beep, nothing happens , the ability has no cooldown and no rockets seem to barrage the area. ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 10 May 2011 at 12:49
1.0
Gameplay Error - ``` I believe there are different observations already made by other testers, some said it works, some said it not works. For me, if i activate this ability of the Light Assault Cruiser, it gives me the Beep sound as if the area is not accesible or it is out of range etc. However, the Cruiser responds with a standard "attack phrase". After the Beep, nothing happens , the ability has no cooldown and no rockets seem to barrage the area. ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 10 May 2011 at 12:49
defect
gameplay error i believe there are different observations already made by other testers some said it works some said it not works for me if i activate this ability of the light assault cruiser it gives me the beep sound as if the area is not accesible or it is out of range etc however the cruiser responds with a standard attack phrase after the beep nothing happens the ability has no cooldown and no rockets seem to barrage the area original issue reported on code google com by gmail com on may at
1
327,642
28,075,494,706
IssuesEvent
2023-03-29 23:02:00
ray-project/ray
https://api.github.com/repos/ray-project/ray
closed
[Release][Data] Migrate `dataset_shuffle_random_shuffle_1tb` test to v2 stack and Anyscale Jobs
P1 datasets release-test ray-team-created
### What happened + What you expected to happen Migrate `dataset_shuffle_random_shuffle_1tb` test to v2 stack and Anyscale Jobs. If migration is impossible, consider rewriting the test or removing it. If there are features missing in v2 stack or/and Anyscale Jobs that make the migration impossible, please outline them in this issue. ### Versions / Dependencies master ### Reproduction script - ### Issue Severity None
1.0
[Release][Data] Migrate `dataset_shuffle_random_shuffle_1tb` test to v2 stack and Anyscale Jobs - ### What happened + What you expected to happen Migrate `dataset_shuffle_random_shuffle_1tb` test to v2 stack and Anyscale Jobs. If migration is impossible, consider rewriting the test or removing it. If there are features missing in v2 stack or/and Anyscale Jobs that make the migration impossible, please outline them in this issue. ### Versions / Dependencies master ### Reproduction script - ### Issue Severity None
non_defect
migrate dataset shuffle random shuffle test to stack and anyscale jobs what happened what you expected to happen migrate dataset shuffle random shuffle test to stack and anyscale jobs if migration is impossible consider rewriting the test or removing it if there are features missing in stack or and anyscale jobs that make the migration impossible please outline them in this issue versions dependencies master reproduction script issue severity none
0
64,513
18,722,540,599
IssuesEvent
2021-11-03 13:22:38
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
DataTable: after filtering and changing value during ajax old filtered data is still shown
defect
**Describe the defect** When you filter a Datatable and updates the value using ajax, the filtered old value will be still applied even if you set the filteredValue to null **Reproducer** https://github.com/neXus1987/primefaces-datatable-test.git **Environment:** - PF Version: _10.0_ - JSF + version: Mojarra 2.2.20 - Affected browsers: FireFox **To Reproduce** Filter for: Data 1 click on button toggle **Expected behavior** no records should been shown, because the filter do not match the new data (Data 2 *) **Example XHTML** ```html <!DOCTYPE html> <html xmlns="http://www.w3.org/1999/xhtml" xmlns:h="http://java.sun.com/jsf/html" xmlns:f="http://java.sun.com/jsf/core" xmlns:ui="http://java.sun.com/jsf/facelets" xmlns:p="http://primefaces.org/ui" lang="de"> <h:head> </h:head> <h:body> <h:form> <p:dataTable value="#{kundenbereichModel.listStringData}" var="d"> <p:column filterBy="#{d}" filterMatchMode="contains">#{d}</p:column> </p:dataTable> <p:commandLink update="@form" process="@this" actionListener="#{kundebereichTestController.toggleList()}">toggle</p:commandLink> </h:form> </h:body> </html> ``` **Example Bean** ```java @ManagedBean @RequestScoped public class KundebereichTestController { private final static Log logger = LogFactory .getLog(KundebereichTestController.class); @ManagedProperty("#{kundenbereichModel}") private KundenbereichModel kundenbereichModel; public void setKundenbereichModel(KundenbereichModel kundenbereichModel) { this.kundenbereichModel = kundenbereichModel; } private List<String> getData1() { List<String> t = new ArrayList<>(); t.add("Data 1 1"); t.add("Data 1 2"); t.add("Data 1 3"); t.add("Data 1 4"); return t; } private List<String> getData2() { List<String> t = new ArrayList<>(); t.add("Data 2 1"); t.add("Data 2 2"); t.add("Data 2 3"); t.add("Data 2 4"); return t; } public void toggleList() { if (kundenbereichModel.getViewState() == 0) { kundenbereichModel.setListStringData(getData1()); kundenbereichModel.setViewState(1); } else { kundenbereichModel.setListStringData(getData2()); kundenbereichModel.setViewState(0); } } } @ManagedBean @ViewScoped public class KundenbereichModel implements Serializable, { private List<String> listStringData; public List<String> getListStringData() { return listStringData; } public void setListStringData(List<String> listStringData) { this.listStringData = listStringData; } private int viewState; public int getViewState() { return viewState; } public void setViewState(int viewState) { this.viewState = viewState; } } ```
1.0
DataTable: after filtering and changing value during ajax old filtered data is still shown - **Describe the defect** When you filter a Datatable and updates the value using ajax, the filtered old value will be still applied even if you set the filteredValue to null **Reproducer** https://github.com/neXus1987/primefaces-datatable-test.git **Environment:** - PF Version: _10.0_ - JSF + version: Mojarra 2.2.20 - Affected browsers: FireFox **To Reproduce** Filter for: Data 1 click on button toggle **Expected behavior** no records should been shown, because the filter do not match the new data (Data 2 *) **Example XHTML** ```html <!DOCTYPE html> <html xmlns="http://www.w3.org/1999/xhtml" xmlns:h="http://java.sun.com/jsf/html" xmlns:f="http://java.sun.com/jsf/core" xmlns:ui="http://java.sun.com/jsf/facelets" xmlns:p="http://primefaces.org/ui" lang="de"> <h:head> </h:head> <h:body> <h:form> <p:dataTable value="#{kundenbereichModel.listStringData}" var="d"> <p:column filterBy="#{d}" filterMatchMode="contains">#{d}</p:column> </p:dataTable> <p:commandLink update="@form" process="@this" actionListener="#{kundebereichTestController.toggleList()}">toggle</p:commandLink> </h:form> </h:body> </html> ``` **Example Bean** ```java @ManagedBean @RequestScoped public class KundebereichTestController { private final static Log logger = LogFactory .getLog(KundebereichTestController.class); @ManagedProperty("#{kundenbereichModel}") private KundenbereichModel kundenbereichModel; public void setKundenbereichModel(KundenbereichModel kundenbereichModel) { this.kundenbereichModel = kundenbereichModel; } private List<String> getData1() { List<String> t = new ArrayList<>(); t.add("Data 1 1"); t.add("Data 1 2"); t.add("Data 1 3"); t.add("Data 1 4"); return t; } private List<String> getData2() { List<String> t = new ArrayList<>(); t.add("Data 2 1"); t.add("Data 2 2"); t.add("Data 2 3"); t.add("Data 2 4"); return t; } public void toggleList() { if (kundenbereichModel.getViewState() == 0) { kundenbereichModel.setListStringData(getData1()); kundenbereichModel.setViewState(1); } else { kundenbereichModel.setListStringData(getData2()); kundenbereichModel.setViewState(0); } } } @ManagedBean @ViewScoped public class KundenbereichModel implements Serializable, { private List<String> listStringData; public List<String> getListStringData() { return listStringData; } public void setListStringData(List<String> listStringData) { this.listStringData = listStringData; } private int viewState; public int getViewState() { return viewState; } public void setViewState(int viewState) { this.viewState = viewState; } } ```
defect
datatable after filtering and changing value during ajax old filtered data is still shown describe the defect when you filter a datatable and updates the value using ajax the filtered old value will be still applied even if you set the filteredvalue to null reproducer environment pf version jsf version mojarra affected browsers firefox to reproduce filter for data click on button toggle expected behavior no records should been shown because the filter do not match the new data data example xhtml html html xmlns xmlns h xmlns f xmlns ui xmlns p lang de d p commandlink update form process this actionlistener kundebereichtestcontroller togglelist toggle example bean java managedbean requestscoped public class kundebereichtestcontroller private final static log logger logfactory getlog kundebereichtestcontroller class managedproperty kundenbereichmodel private kundenbereichmodel kundenbereichmodel public void setkundenbereichmodel kundenbereichmodel kundenbereichmodel this kundenbereichmodel kundenbereichmodel private list list t new arraylist t add data t add data t add data t add data return t private list list t new arraylist t add data t add data t add data t add data return t public void togglelist if kundenbereichmodel getviewstate kundenbereichmodel setliststringdata kundenbereichmodel setviewstate else kundenbereichmodel setliststringdata kundenbereichmodel setviewstate managedbean viewscoped public class kundenbereichmodel implements serializable private list liststringdata public list getliststringdata return liststringdata public void setliststringdata list liststringdata this liststringdata liststringdata private int viewstate public int getviewstate return viewstate public void setviewstate int viewstate this viewstate viewstate
1
76,728
15,496,181,874
IssuesEvent
2021-03-11 02:12:40
mwilliams7197/zendo
https://api.github.com/repos/mwilliams7197/zendo
closed
WS-2019-0032 (Medium) detected in js-yaml-3.7.0.tgz - autoclosed
security vulnerability
## WS-2019-0032 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p></summary> <p>YAML 1.2 parser and serializer</p> <p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p> <p>Path to dependency file: zendo/package.json</p> <p>Path to vulnerable library: zendo/node_modules/js-yaml/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.0.7.tgz (Root Library) - eslint-3.19.0.tgz - :x: **js-yaml-3.7.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mwilliams7197/zendo/commit/cf147eff03fa219eea73721a78a106371bf2f0bc">cf147eff03fa219eea73721a78a106371bf2f0bc</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service. <p>Publish Date: 2019-03-20 <p>URL: <a href=https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235>WS-2019-0032</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/788/versions">https://www.npmjs.com/advisories/788/versions</a></p> <p>Release Date: 2019-03-20</p> <p>Fix Resolution: js-yaml - 3.13.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"js-yaml","packageVersion":"3.7.0","isTransitiveDependency":true,"dependencyTree":"react-scripts:1.0.7;eslint:3.19.0;js-yaml:3.7.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"js-yaml - 3.13.0"}],"vulnerabilityIdentifier":"WS-2019-0032","vulnerabilityDetails":"Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service.","vulnerabilityUrl":"https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
True
WS-2019-0032 (Medium) detected in js-yaml-3.7.0.tgz - autoclosed - ## WS-2019-0032 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p></summary> <p>YAML 1.2 parser and serializer</p> <p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p> <p>Path to dependency file: zendo/package.json</p> <p>Path to vulnerable library: zendo/node_modules/js-yaml/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.0.7.tgz (Root Library) - eslint-3.19.0.tgz - :x: **js-yaml-3.7.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mwilliams7197/zendo/commit/cf147eff03fa219eea73721a78a106371bf2f0bc">cf147eff03fa219eea73721a78a106371bf2f0bc</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service. <p>Publish Date: 2019-03-20 <p>URL: <a href=https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235>WS-2019-0032</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/788/versions">https://www.npmjs.com/advisories/788/versions</a></p> <p>Release Date: 2019-03-20</p> <p>Fix Resolution: js-yaml - 3.13.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"js-yaml","packageVersion":"3.7.0","isTransitiveDependency":true,"dependencyTree":"react-scripts:1.0.7;eslint:3.19.0;js-yaml:3.7.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"js-yaml - 3.13.0"}],"vulnerabilityIdentifier":"WS-2019-0032","vulnerabilityDetails":"Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service.","vulnerabilityUrl":"https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
non_defect
ws medium detected in js yaml tgz autoclosed ws medium severity vulnerability vulnerable library js yaml tgz yaml parser and serializer library home page a href path to dependency file zendo package json path to vulnerable library zendo node modules js yaml package json dependency hierarchy react scripts tgz root library eslint tgz x js yaml tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions js yaml prior to are vulnerable to denial of service by parsing a carefully crafted yaml file the node process stalls and may exhaust system resources leading to a denial of service publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution js yaml isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails versions js yaml prior to are vulnerable to denial of service by parsing a carefully crafted yaml file the node process stalls and may exhaust system resources leading to a denial of service vulnerabilityurl
0
64,871
18,949,777,713
IssuesEvent
2021-11-18 14:06:56
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
closed
Push notifications sometimes disappear immediately
T-Defect A-Notifications
In a 1:1 encrypted chat, I've noticed that often (but not always, maybe 2/3 of the time), notifications disappear shortly after they appear and my phone vibrates. When this happens, the dot signifying new notifications on the element app icon also disappears. I'm currently using version 1.2.1 on a Pixel 5 running Android 11, but I've noticed this issue for a few months on both this device and a Pixel 3, also running Android 11. [Here's a video](https://streamable.com/focnm4)
1.0
Push notifications sometimes disappear immediately - In a 1:1 encrypted chat, I've noticed that often (but not always, maybe 2/3 of the time), notifications disappear shortly after they appear and my phone vibrates. When this happens, the dot signifying new notifications on the element app icon also disappears. I'm currently using version 1.2.1 on a Pixel 5 running Android 11, but I've noticed this issue for a few months on both this device and a Pixel 3, also running Android 11. [Here's a video](https://streamable.com/focnm4)
defect
push notifications sometimes disappear immediately in a encrypted chat i ve noticed that often but not always maybe of the time notifications disappear shortly after they appear and my phone vibrates when this happens the dot signifying new notifications on the element app icon also disappears i m currently using version on a pixel running android but i ve noticed this issue for a few months on both this device and a pixel also running android
1
3,448
2,610,062,965
IssuesEvent
2015-02-26 18:18:30
chrsmith/jsjsj122
https://api.github.com/repos/chrsmith/jsjsj122
opened
黄岩治不育一般需要多少钱
auto-migrated Priority-Medium Type-Defect
``` 黄岩治不育一般需要多少钱【台州五洲生殖医院】24小时健康 咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台 州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、1 08、118、198及椒江一金清公交车直达枫南小区,乘坐107、105、 109、112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:45
1.0
黄岩治不育一般需要多少钱 - ``` 黄岩治不育一般需要多少钱【台州五洲生殖医院】24小时健康 咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台 州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、1 08、118、198及椒江一金清公交车直达枫南小区,乘坐107、105、 109、112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:45
defect
黄岩治不育一般需要多少钱 黄岩治不育一般需要多少钱【台州五洲生殖医院】 咨询热线 微信号tzwzszyy 医院地址 台 (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
1
50,617
13,187,626,198
IssuesEvent
2020-08-13 04:02:01
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
muex option 'detail' is broken in trunk r135229 (Trac #1055)
Migrated from Trac combo reconstruction defect
If the detail option of the muex module is set to True it will seg fault. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1055">https://code.icecube.wisc.edu/ticket/1055</a>, reported by sflis and owned by dima</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "description": "If the detail option of the muex module is set to True it will seg fault.", "reporter": "sflis", "cc": "", "resolution": "fixed", "_ts": "1550067117911749", "component": "combo reconstruction", "summary": "muex option 'detail' is broken in trunk r135229", "priority": "blocker", "keywords": "", "time": "2015-07-16T09:47:18", "milestone": "", "owner": "dima", "type": "defect" } ``` </p> </details>
1.0
muex option 'detail' is broken in trunk r135229 (Trac #1055) - If the detail option of the muex module is set to True it will seg fault. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1055">https://code.icecube.wisc.edu/ticket/1055</a>, reported by sflis and owned by dima</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "description": "If the detail option of the muex module is set to True it will seg fault.", "reporter": "sflis", "cc": "", "resolution": "fixed", "_ts": "1550067117911749", "component": "combo reconstruction", "summary": "muex option 'detail' is broken in trunk r135229", "priority": "blocker", "keywords": "", "time": "2015-07-16T09:47:18", "milestone": "", "owner": "dima", "type": "defect" } ``` </p> </details>
defect
muex option detail is broken in trunk trac if the detail option of the muex module is set to true it will seg fault migrated from json status closed changetime description if the detail option of the muex module is set to true it will seg fault reporter sflis cc resolution fixed ts component combo reconstruction summary muex option detail is broken in trunk priority blocker keywords time milestone owner dima type defect
1
26,213
4,622,767,866
IssuesEvent
2016-09-27 08:46:29
siddhartha-gadgil/ProvingGround
https://api.github.com/repos/siddhartha-gadgil/ProvingGround
closed
Very poor performance due to cascade of replacements.
defect
In generating from _Monoids_, but not from just _logic_, there is a huge slowdown, with most of the time spent in replace methods. This appears to be because: * for the sake of safety, lambdas: * create an inner variable * in the case of lambda-fixed, check independence. * both these involve replacements for lambdas * if the replaced values are `IdentityTyp` etc., then the terms themselves contain a lot of lambdas, so we get a cascade of replacements. * We can avoid creating inner variables and checking for dependence if we are using a new synthesized variable (and are careful about dependence).
1.0
Very poor performance due to cascade of replacements. - In generating from _Monoids_, but not from just _logic_, there is a huge slowdown, with most of the time spent in replace methods. This appears to be because: * for the sake of safety, lambdas: * create an inner variable * in the case of lambda-fixed, check independence. * both these involve replacements for lambdas * if the replaced values are `IdentityTyp` etc., then the terms themselves contain a lot of lambdas, so we get a cascade of replacements. * We can avoid creating inner variables and checking for dependence if we are using a new synthesized variable (and are careful about dependence).
defect
very poor performance due to cascade of replacements in generating from monoids but not from just logic there is a huge slowdown with most of the time spent in replace methods this appears to be because for the sake of safety lambdas create an inner variable in the case of lambda fixed check independence both these involve replacements for lambdas if the replaced values are identitytyp etc then the terms themselves contain a lot of lambdas so we get a cascade of replacements we can avoid creating inner variables and checking for dependence if we are using a new synthesized variable and are careful about dependence
1
18,629
3,077,631,643
IssuesEvent
2015-08-21 02:27:04
martingkelly/imms
https://api.github.com/repos/martingkelly/imms
closed
imms takes a while to record played songs in its database
auto-migrated Priority-Medium Type-Defect
``` If you finish playing a song, imms does not immediately record it in its database. I spent a while debugging this and tracked it to the buffering behavior of the GIOChannel used on the immsd server. Basically, GIOChannel has an internal buffer, and data doesn't get sent until that buffer fills up. The size of the buffer can vary across platforms, and thus the time until a song is recorded can vary a lot too. To add to this, if audacious exits before the buffer fills up, the song data from that session is completely lost :(. I wrote a patch to fix this behavior by disabling buffering in the GIOChannel. With this patch, song data is recorded immediately. Given that the amount of data being sent between the audacious client and the immsd server is relatively low, I believe it's OK to skip buffering. My patch can be found below (based on John Lindgren's fixes for Audacious 3.6). https://github.com/martingkelly/imms/commit/c3f99566fe98459f0c7436804039190c4101 cfd9 ``` Original issue reported on code.google.com by `mar...@martingkelly.com` on 12 Aug 2015 at 3:32
1.0
imms takes a while to record played songs in its database - ``` If you finish playing a song, imms does not immediately record it in its database. I spent a while debugging this and tracked it to the buffering behavior of the GIOChannel used on the immsd server. Basically, GIOChannel has an internal buffer, and data doesn't get sent until that buffer fills up. The size of the buffer can vary across platforms, and thus the time until a song is recorded can vary a lot too. To add to this, if audacious exits before the buffer fills up, the song data from that session is completely lost :(. I wrote a patch to fix this behavior by disabling buffering in the GIOChannel. With this patch, song data is recorded immediately. Given that the amount of data being sent between the audacious client and the immsd server is relatively low, I believe it's OK to skip buffering. My patch can be found below (based on John Lindgren's fixes for Audacious 3.6). https://github.com/martingkelly/imms/commit/c3f99566fe98459f0c7436804039190c4101 cfd9 ``` Original issue reported on code.google.com by `mar...@martingkelly.com` on 12 Aug 2015 at 3:32
defect
imms takes a while to record played songs in its database if you finish playing a song imms does not immediately record it in its database i spent a while debugging this and tracked it to the buffering behavior of the giochannel used on the immsd server basically giochannel has an internal buffer and data doesn t get sent until that buffer fills up the size of the buffer can vary across platforms and thus the time until a song is recorded can vary a lot too to add to this if audacious exits before the buffer fills up the song data from that session is completely lost i wrote a patch to fix this behavior by disabling buffering in the giochannel with this patch song data is recorded immediately given that the amount of data being sent between the audacious client and the immsd server is relatively low i believe it s ok to skip buffering my patch can be found below based on john lindgren s fixes for audacious original issue reported on code google com by mar martingkelly com on aug at
1
41,668
10,563,396,765
IssuesEvent
2019-10-04 20:51:37
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
[COGNITION]: Recommend showing all secondary specialties instead of show more/show less button
508-defect-1 508/Accessibility facility locator frontend vsa-global-ux
## Issue The facility detail views sometimes have second-level lists that include a show more button. These buttons prepend additional `<li>` before the button, which causes an issue for assistive device users. This practice was flagged as an SC 1.3.1 issue. Screenshot attached below. ## Audit Finding * Note 1, Defect 3 of 3 * Note 5, Defect 2 of 2 ## Related Issues * https://app.zenhub.com/workspaces/vsp-5cedc9cce6e3335dc5a49fc4/issues/department-of-veterans-affairs/va.gov-team/484 ## Acceptance Criteria * As an assistive device user, I want to hear the nested list items all read out, or understand that list items have been added farther down in the DOM, after the show more button. ## Environment * MacOS Mojave * Chrome latest * https://staging.va.gov/find-locations/facility/vha_674GD ## Possible Fixes (optional) The simplest way to remediate this error is to remove the show more button and return all of the nested list items on first render. The other option is to remove the button from the list, and create a new list below the button. This feels less useful from a user flow, but could be an option if needed. ## WCAG or Vendor Guidance (optional) * [Info and Relationships: Understanding SC 1.3.1](https://www.w3.org/TR/UNDERSTANDING-WCAG20/content-structure-separation-programmatic.html) ## Screenshots or Trace Logs ![Screen Shot 2019-07-03 at 5.37.06 PM.png](https://images.zenhubusercontent.com/5ac217b74b5806bc2bcd3fc8/faccdc9f-bc2c-42f0-a8ad-75fe085bfb90)
1.0
[COGNITION]: Recommend showing all secondary specialties instead of show more/show less button - ## Issue The facility detail views sometimes have second-level lists that include a show more button. These buttons prepend additional `<li>` before the button, which causes an issue for assistive device users. This practice was flagged as an SC 1.3.1 issue. Screenshot attached below. ## Audit Finding * Note 1, Defect 3 of 3 * Note 5, Defect 2 of 2 ## Related Issues * https://app.zenhub.com/workspaces/vsp-5cedc9cce6e3335dc5a49fc4/issues/department-of-veterans-affairs/va.gov-team/484 ## Acceptance Criteria * As an assistive device user, I want to hear the nested list items all read out, or understand that list items have been added farther down in the DOM, after the show more button. ## Environment * MacOS Mojave * Chrome latest * https://staging.va.gov/find-locations/facility/vha_674GD ## Possible Fixes (optional) The simplest way to remediate this error is to remove the show more button and return all of the nested list items on first render. The other option is to remove the button from the list, and create a new list below the button. This feels less useful from a user flow, but could be an option if needed. ## WCAG or Vendor Guidance (optional) * [Info and Relationships: Understanding SC 1.3.1](https://www.w3.org/TR/UNDERSTANDING-WCAG20/content-structure-separation-programmatic.html) ## Screenshots or Trace Logs ![Screen Shot 2019-07-03 at 5.37.06 PM.png](https://images.zenhubusercontent.com/5ac217b74b5806bc2bcd3fc8/faccdc9f-bc2c-42f0-a8ad-75fe085bfb90)
defect
recommend showing all secondary specialties instead of show more show less button issue the facility detail views sometimes have second level lists that include a show more button these buttons prepend additional before the button which causes an issue for assistive device users this practice was flagged as an sc issue screenshot attached below audit finding note defect of note defect of related issues acceptance criteria as an assistive device user i want to hear the nested list items all read out or understand that list items have been added farther down in the dom after the show more button environment macos mojave chrome latest possible fixes optional the simplest way to remediate this error is to remove the show more button and return all of the nested list items on first render the other option is to remove the button from the list and create a new list below the button this feels less useful from a user flow but could be an option if needed wcag or vendor guidance optional screenshots or trace logs
1
34,481
7,452,016,728
IssuesEvent
2018-03-29 06:38:26
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
Seostamine: Org/Isiku seostamine teise Org/Isikuga ning Kirjeldusüksusega
P: highest R: fixed T: defect
**Reported by maiu pevkur on 9 May 2013 13:13 UTC** Organisatsiooni ja Isiku lisainfo vormidel on sakid Seotud kirjeldusüksused ja Seotud isikud/org, kus on vanast AISist kanud andmed (osaliselt need andmed küll veel puuduvad). Puudub aga võimalus lisada/kustutada seotud kirjeldusüksusi ja seotud isikuid/orge. Kui vajutada Muuda, mis avab andmete muutmise vormi (nii Org-idele kui Isikutele), siis seal sellist võimalust ei paista. Samuti oleks hea, kui saaks ka vormil Kirjeldusüksuse andmed (ehk andmete muutmise vormil) lisada seotud isikut/organisatsiooni ja ka vastavat seost kustutada. St et seostamine toimiks ka vastupidi. Oleks võimalik seoseid lisada/kustutada nii Org/Isiku kui Kirjeldusüksuse andmete muutmise vormidel.
1.0
Seostamine: Org/Isiku seostamine teise Org/Isikuga ning Kirjeldusüksusega - **Reported by maiu pevkur on 9 May 2013 13:13 UTC** Organisatsiooni ja Isiku lisainfo vormidel on sakid Seotud kirjeldusüksused ja Seotud isikud/org, kus on vanast AISist kanud andmed (osaliselt need andmed küll veel puuduvad). Puudub aga võimalus lisada/kustutada seotud kirjeldusüksusi ja seotud isikuid/orge. Kui vajutada Muuda, mis avab andmete muutmise vormi (nii Org-idele kui Isikutele), siis seal sellist võimalust ei paista. Samuti oleks hea, kui saaks ka vormil Kirjeldusüksuse andmed (ehk andmete muutmise vormil) lisada seotud isikut/organisatsiooni ja ka vastavat seost kustutada. St et seostamine toimiks ka vastupidi. Oleks võimalik seoseid lisada/kustutada nii Org/Isiku kui Kirjeldusüksuse andmete muutmise vormidel.
defect
seostamine org isiku seostamine teise org isikuga ning kirjeldusüksusega reported by maiu pevkur on may utc organisatsiooni ja isiku lisainfo vormidel on sakid seotud kirjeldusüksused ja seotud isikud org kus on vanast aisist kanud andmed osaliselt need andmed küll veel puuduvad puudub aga võimalus lisada kustutada seotud kirjeldusüksusi ja seotud isikuid orge kui vajutada muuda mis avab andmete muutmise vormi nii org idele kui isikutele siis seal sellist võimalust ei paista samuti oleks hea kui saaks ka vormil kirjeldusüksuse andmed ehk andmete muutmise vormil lisada seotud isikut organisatsiooni ja ka vastavat seost kustutada st et seostamine toimiks ka vastupidi oleks võimalik seoseid lisada kustutada nii org isiku kui kirjeldusüksuse andmete muutmise vormidel
1
48,126
13,067,466,136
IssuesEvent
2020-07-31 00:32:42
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
L1 filter for 2014 and 2015 (Trac #1828)
Migrated from Trac cmake defect
Running offline filter for 2014 and 2015: /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/metaprojects/icerec/IC2014-L2_V14-02-00/lib/icecube/filterscripts/offlineL2/level1_SimulationFiltering.py I receive the following error: `RuntimeError: dlopen() dynamic loading error: /data/user/saxani/environments/buildfwd/lib/libpfauxiliary.so: cannot open shared object file: No such file or directory` Migrated from https://code.icecube.wisc.edu/ticket/1828 ```json { "status": "closed", "changetime": "2016-08-18T21:39:38", "description": "Running offline filter for 2014 and 2015:\n/cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/metaprojects/icerec/IC2014-L2_V14-02-00/lib/icecube/filterscripts/offlineL2/level1_SimulationFiltering.py\n\nI receive the following error:\n`RuntimeError: dlopen() dynamic loading error: /data/user/saxani/environments/buildfwd/lib/libpfauxiliary.so: cannot open shared object file: No such file or directory`\n", "reporter": "saxani", "cc": "", "resolution": "fixed", "_ts": "1471556378950850", "component": "cmake", "summary": "L1 filter for 2014 and 2015", "priority": "normal", "keywords": "", "time": "2016-08-18T21:29:33", "milestone": "", "owner": "", "type": "defect" } ```
1.0
L1 filter for 2014 and 2015 (Trac #1828) - Running offline filter for 2014 and 2015: /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/metaprojects/icerec/IC2014-L2_V14-02-00/lib/icecube/filterscripts/offlineL2/level1_SimulationFiltering.py I receive the following error: `RuntimeError: dlopen() dynamic loading error: /data/user/saxani/environments/buildfwd/lib/libpfauxiliary.so: cannot open shared object file: No such file or directory` Migrated from https://code.icecube.wisc.edu/ticket/1828 ```json { "status": "closed", "changetime": "2016-08-18T21:39:38", "description": "Running offline filter for 2014 and 2015:\n/cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/metaprojects/icerec/IC2014-L2_V14-02-00/lib/icecube/filterscripts/offlineL2/level1_SimulationFiltering.py\n\nI receive the following error:\n`RuntimeError: dlopen() dynamic loading error: /data/user/saxani/environments/buildfwd/lib/libpfauxiliary.so: cannot open shared object file: No such file or directory`\n", "reporter": "saxani", "cc": "", "resolution": "fixed", "_ts": "1471556378950850", "component": "cmake", "summary": "L1 filter for 2014 and 2015", "priority": "normal", "keywords": "", "time": "2016-08-18T21:29:33", "milestone": "", "owner": "", "type": "defect" } ```
defect
filter for and trac running offline filter for and cvmfs icecube opensciencegrid org rhel metaprojects icerec lib icecube filterscripts simulationfiltering py i receive the following error runtimeerror dlopen dynamic loading error data user saxani environments buildfwd lib libpfauxiliary so cannot open shared object file no such file or directory migrated from json status closed changetime description running offline filter for and n cvmfs icecube opensciencegrid org rhel metaprojects icerec lib icecube filterscripts simulationfiltering py n ni receive the following error n runtimeerror dlopen dynamic loading error data user saxani environments buildfwd lib libpfauxiliary so cannot open shared object file no such file or directory n reporter saxani cc resolution fixed ts component cmake summary filter for and priority normal keywords time milestone owner type defect
1
47,769
19,716,000,457
IssuesEvent
2022-01-13 11:02:10
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
How should the AZ CLI script for importing the Helm images be run
container-service/svc triaged cxp needs-more-info product-issue Pri1
I tried running those commands in PowerShell and get errors all over the place. I then figured I'd put them into a powershell script file and run that. Still no dice. What am I missing? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 1bf9c2f3-caeb-0fb1-f089-3f7ca4fad729 * Version Independent ID: 929253d7-c763-6429-561f-529d39b5284d * Content: [Ingress controller on internal network - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/ingress-internal-ip) * Content Source: [articles/aks/ingress-internal-ip.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/ingress-internal-ip.md) * Service: **container-service** * GitHub Login: @zr-msft * Microsoft Alias: **zarhoads**
1.0
How should the AZ CLI script for importing the Helm images be run - I tried running those commands in PowerShell and get errors all over the place. I then figured I'd put them into a powershell script file and run that. Still no dice. What am I missing? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 1bf9c2f3-caeb-0fb1-f089-3f7ca4fad729 * Version Independent ID: 929253d7-c763-6429-561f-529d39b5284d * Content: [Ingress controller on internal network - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/ingress-internal-ip) * Content Source: [articles/aks/ingress-internal-ip.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/ingress-internal-ip.md) * Service: **container-service** * GitHub Login: @zr-msft * Microsoft Alias: **zarhoads**
non_defect
how should the az cli script for importing the helm images be run i tried running those commands in powershell and get errors all over the place i then figured i d put them into a powershell script file and run that still no dice what am i missing document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id caeb version independent id content content source service container service github login zr msft microsoft alias zarhoads
0
354,378
25,163,572,768
IssuesEvent
2022-11-10 18:45:13
netbox-community/netbox
https://api.github.com/repos/netbox-community/netbox
opened
'Installation & Upgrade' Netbox documentation page - Add note to 'Warning' section at bottom of the page to have user confirm virtual environment is still activated
type: documentation
### Change Type Addition ### Area Installation/upgrade ### Proposed Changes Add a basic troubleshooting note to the "Warning" section at the bottom of the "3. Netbox" page in the "Installation & Upgrade" section of the documentation. Have user ensure that the virtual environment is still activated before performing additional troubleshooting measures. It's a basic enough step that could help newer users avoid unnecessary confusion, while still keeping the documentation concise. Suggested addition: "If you receive errors regarding missing modules/packages, be sure that you still have the virtual environment activated before performing any additional troubleshooting measures."
1.0
'Installation & Upgrade' Netbox documentation page - Add note to 'Warning' section at bottom of the page to have user confirm virtual environment is still activated - ### Change Type Addition ### Area Installation/upgrade ### Proposed Changes Add a basic troubleshooting note to the "Warning" section at the bottom of the "3. Netbox" page in the "Installation & Upgrade" section of the documentation. Have user ensure that the virtual environment is still activated before performing additional troubleshooting measures. It's a basic enough step that could help newer users avoid unnecessary confusion, while still keeping the documentation concise. Suggested addition: "If you receive errors regarding missing modules/packages, be sure that you still have the virtual environment activated before performing any additional troubleshooting measures."
non_defect
installation upgrade netbox documentation page add note to warning section at bottom of the page to have user confirm virtual environment is still activated change type addition area installation upgrade proposed changes add a basic troubleshooting note to the warning section at the bottom of the netbox page in the installation upgrade section of the documentation have user ensure that the virtual environment is still activated before performing additional troubleshooting measures it s a basic enough step that could help newer users avoid unnecessary confusion while still keeping the documentation concise suggested addition if you receive errors regarding missing modules packages be sure that you still have the virtual environment activated before performing any additional troubleshooting measures
0
447,142
31,624,355,073
IssuesEvent
2023-09-06 03:21:17
Virtue-Digital-Indonesia/frontend-examples
https://api.github.com/repos/Virtue-Digital-Indonesia/frontend-examples
opened
Projects Tracker
documentation
This issue is used to list and track the progress of each project. Here are the projects: - [ ] Data Grid Filters and Data Grid Table - [ ] Object List with MUI Data Grid - [ ] Markers, Popups, and Clusters with Leaflet - [ ] Date Range and Time Picker - [ ] Create, show, edit, and delete Geofences with Leaflet
1.0
Projects Tracker - This issue is used to list and track the progress of each project. Here are the projects: - [ ] Data Grid Filters and Data Grid Table - [ ] Object List with MUI Data Grid - [ ] Markers, Popups, and Clusters with Leaflet - [ ] Date Range and Time Picker - [ ] Create, show, edit, and delete Geofences with Leaflet
non_defect
projects tracker this issue is used to list and track the progress of each project here are the projects data grid filters and data grid table object list with mui data grid markers popups and clusters with leaflet date range and time picker create show edit and delete geofences with leaflet
0
171,141
27,066,919,191
IssuesEvent
2023-02-14 01:36:52
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
Expose IsInExpressionTree
Area-IDE Concept-API Need Design Review
Useful in code fixes as the rules are slightly different in expressions. Preferable via a static mehtod and not via some service locator. Don't know if `SemanticFacts` is a thing.
1.0
Expose IsInExpressionTree - Useful in code fixes as the rules are slightly different in expressions. Preferable via a static mehtod and not via some service locator. Don't know if `SemanticFacts` is a thing.
non_defect
expose isinexpressiontree useful in code fixes as the rules are slightly different in expressions preferable via a static mehtod and not via some service locator don t know if semanticfacts is a thing
0
277,880
24,108,146,369
IssuesEvent
2022-09-20 09:09:21
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Test failure: BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles
ci-concern bot/type/test bot/channel/nightly bot/platform/android bot/arch/x86-mono bot/branch/v1.45
Greetings human! Bad news. `BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles` [failed on android x86-mono nightly v1.45.58](https://ci.brave.com/job/brave-browser-build-android-variant/5971/testReport/junit/(root)/BraveProfileManagerTest/test___test_browser___ExcludeServicesInOTRAndGuestProfiles). <details> <summary>Stack trace</summary> ``` [ RUN ] BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles tracing_subscriber - init success [INFO:mod.rs(72)] skus::sdk: skus sdk initialized, environment: production, version: 4fe49d14c2 at /home/ubuntu/workspace/brave-browser-build-android-variant-nightly/src/brave/components/skus/browser/rs/lib/src/sdk/mod.rs:72 [VERBOSE1:kv.rs(123)] skus::storage::kv: got orders, orders: None at /home/ubuntu/workspace/brave-browser-build-android-variant-nightly/src/brave/components/skus/browser/rs/lib/src/storage/kv.rs:123 in skus::storage::kv::get_orders with self=NativeClient in skus::sdk::credentials::matching_credential_summary with self=SDK domain="vpn.brave.com" ../../brave/browser/profiles/brave_profile_manager_browsertest.cc:175: Failure Expected: (brave_rewards::RewardsServiceFactory::GetForProfile(profile)) != (nullptr), actual: NULL vs (nullptr) Stack trace: ../../brave/browser/profiles/brave_profile_manager_browsertest.cc:180: Failure Expected: (brave_ads::AdsServiceFactory::GetForProfile(profile)) != (nullptr), actual: NULL vs (nullptr) Stack trace: [WARNING:task_environment.cc(800)] ParallelExecutionFence is ineffective when ThreadPoolInstance is not managed by a TaskEnvironment. Test fixtures should use a TaskEnvironment member or statically invoke TaskEnvironment::CreateThreadPool() + ThreadPoolInstance::Get()->StartWithDefaultParams() when the former is not possible. [WARNING:task_environment.cc(800)] ParallelExecutionFence is ineffective when ThreadPoolInstance is not managed by a TaskEnvironment. Test fixtures should use a TaskEnvironment member or statically invoke TaskEnvironment::CreateThreadPool() + ThreadPoolInstance::Get()->StartWithDefaultParams() when the former is not possible. [ FAILED ] BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles, where TypeParam = and GetParam() = (671 ms) [----------] 1 test from BraveProfileManagerTest (672 ms total) [----------] Global test environment tear-down [==========] 1 test from 1 test suite ran. (672 ms total) [ PASSED ] 0 tests. [ FAILED ] 1 test, listed below: [ FAILED ] BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles, where TypeParam = and GetParam() = Stack Trace: RELADDR FUNCTION FILE:LINE 02ec85db BraveProfileManagerTest_ExcludeServicesInOTRAndGuestProfiles_Test::RunTestOnMainThread() ../../brave/browser/profiles/brave_profile_manager_browsertest.cc:174:3 08814c44 content::BrowserTestBase::ProxyRunTestOnMainThreadLoop() ../../content/public/test/browser_test_base.cc:909:7 0881423e content::BrowserTestBase::SetUp() ../../content/public/test/browser_test_base.cc:707:5 0517b84a AndroidBrowserTest::SetUp() ../../chrome/test/base/android/android_browser_test.cc:25:20 0509892d bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) ../../third_party/googletest/src/googletest/src/gtest.cc:0:0 ----------------------------------------------------- Stack Trace: RELADDR FUNCTION FILE:LINE 02ec8692 BraveProfileManagerTest_ExcludeServicesInOTRAndGuestProfiles_Test::RunTestOnMainThread() ../../brave/browser/profiles/brave_profile_manager_browsertest.cc:180:3 08814c44 content::BrowserTestBase::ProxyRunTestOnMainThreadLoop() ../../content/public/test/browser_test_base.cc:909:7 0881423e content::BrowserTestBase::SetUp() ../../content/public/test/browser_test_base.cc:707:5 0517b84a AndroidBrowserTest::SetUp() ../../chrome/test/base/android/android_browser_test.cc:25:20 0509892d bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) ../../third_party/googletest/src/googletest/src/gtest.cc:0:0 failed ``` </details>
1.0
Test failure: BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles - Greetings human! Bad news. `BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles` [failed on android x86-mono nightly v1.45.58](https://ci.brave.com/job/brave-browser-build-android-variant/5971/testReport/junit/(root)/BraveProfileManagerTest/test___test_browser___ExcludeServicesInOTRAndGuestProfiles). <details> <summary>Stack trace</summary> ``` [ RUN ] BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles tracing_subscriber - init success [INFO:mod.rs(72)] skus::sdk: skus sdk initialized, environment: production, version: 4fe49d14c2 at /home/ubuntu/workspace/brave-browser-build-android-variant-nightly/src/brave/components/skus/browser/rs/lib/src/sdk/mod.rs:72 [VERBOSE1:kv.rs(123)] skus::storage::kv: got orders, orders: None at /home/ubuntu/workspace/brave-browser-build-android-variant-nightly/src/brave/components/skus/browser/rs/lib/src/storage/kv.rs:123 in skus::storage::kv::get_orders with self=NativeClient in skus::sdk::credentials::matching_credential_summary with self=SDK domain="vpn.brave.com" ../../brave/browser/profiles/brave_profile_manager_browsertest.cc:175: Failure Expected: (brave_rewards::RewardsServiceFactory::GetForProfile(profile)) != (nullptr), actual: NULL vs (nullptr) Stack trace: ../../brave/browser/profiles/brave_profile_manager_browsertest.cc:180: Failure Expected: (brave_ads::AdsServiceFactory::GetForProfile(profile)) != (nullptr), actual: NULL vs (nullptr) Stack trace: [WARNING:task_environment.cc(800)] ParallelExecutionFence is ineffective when ThreadPoolInstance is not managed by a TaskEnvironment. Test fixtures should use a TaskEnvironment member or statically invoke TaskEnvironment::CreateThreadPool() + ThreadPoolInstance::Get()->StartWithDefaultParams() when the former is not possible. [WARNING:task_environment.cc(800)] ParallelExecutionFence is ineffective when ThreadPoolInstance is not managed by a TaskEnvironment. Test fixtures should use a TaskEnvironment member or statically invoke TaskEnvironment::CreateThreadPool() + ThreadPoolInstance::Get()->StartWithDefaultParams() when the former is not possible. [ FAILED ] BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles, where TypeParam = and GetParam() = (671 ms) [----------] 1 test from BraveProfileManagerTest (672 ms total) [----------] Global test environment tear-down [==========] 1 test from 1 test suite ran. (672 ms total) [ PASSED ] 0 tests. [ FAILED ] 1 test, listed below: [ FAILED ] BraveProfileManagerTest.ExcludeServicesInOTRAndGuestProfiles, where TypeParam = and GetParam() = Stack Trace: RELADDR FUNCTION FILE:LINE 02ec85db BraveProfileManagerTest_ExcludeServicesInOTRAndGuestProfiles_Test::RunTestOnMainThread() ../../brave/browser/profiles/brave_profile_manager_browsertest.cc:174:3 08814c44 content::BrowserTestBase::ProxyRunTestOnMainThreadLoop() ../../content/public/test/browser_test_base.cc:909:7 0881423e content::BrowserTestBase::SetUp() ../../content/public/test/browser_test_base.cc:707:5 0517b84a AndroidBrowserTest::SetUp() ../../chrome/test/base/android/android_browser_test.cc:25:20 0509892d bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) ../../third_party/googletest/src/googletest/src/gtest.cc:0:0 ----------------------------------------------------- Stack Trace: RELADDR FUNCTION FILE:LINE 02ec8692 BraveProfileManagerTest_ExcludeServicesInOTRAndGuestProfiles_Test::RunTestOnMainThread() ../../brave/browser/profiles/brave_profile_manager_browsertest.cc:180:3 08814c44 content::BrowserTestBase::ProxyRunTestOnMainThreadLoop() ../../content/public/test/browser_test_base.cc:909:7 0881423e content::BrowserTestBase::SetUp() ../../content/public/test/browser_test_base.cc:707:5 0517b84a AndroidBrowserTest::SetUp() ../../chrome/test/base/android/android_browser_test.cc:25:20 0509892d bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) ../../third_party/googletest/src/googletest/src/gtest.cc:0:0 failed ``` </details>
non_defect
test failure braveprofilemanagertest excludeservicesinotrandguestprofiles greetings human bad news braveprofilemanagertest excludeservicesinotrandguestprofiles stack trace braveprofilemanagertest excludeservicesinotrandguestprofiles tracing subscriber init success skus sdk skus sdk initialized environment production version at home ubuntu workspace brave browser build android variant nightly src brave components skus browser rs lib src sdk mod rs skus storage kv got orders orders none at home ubuntu workspace brave browser build android variant nightly src brave components skus browser rs lib src storage kv rs in skus storage kv get orders with self nativeclient in skus sdk credentials matching credential summary with self sdk domain vpn brave com brave browser profiles brave profile manager browsertest cc failure expected brave rewards rewardsservicefactory getforprofile profile nullptr actual null vs nullptr stack trace brave browser profiles brave profile manager browsertest cc failure expected brave ads adsservicefactory getforprofile profile nullptr actual null vs nullptr stack trace parallelexecutionfence is ineffective when threadpoolinstance is not managed by a taskenvironment test fixtures should use a taskenvironment member or statically invoke taskenvironment createthreadpool threadpoolinstance get startwithdefaultparams when the former is not possible parallelexecutionfence is ineffective when threadpoolinstance is not managed by a taskenvironment test fixtures should use a taskenvironment member or statically invoke taskenvironment createthreadpool threadpoolinstance get startwithdefaultparams when the former is not possible braveprofilemanagertest excludeservicesinotrandguestprofiles where typeparam and getparam ms test from braveprofilemanagertest ms total global test environment tear down test from test suite ran ms total tests test listed below braveprofilemanagertest excludeservicesinotrandguestprofiles where typeparam and getparam stack trace reladdr function file line braveprofilemanagertest excludeservicesinotrandguestprofiles test runtestonmainthread brave browser profiles brave profile manager browsertest cc content browsertestbase proxyruntestonmainthreadloop content public test browser test base cc content browsertestbase setup content public test browser test base cc androidbrowsertest setup chrome test base android android browser test cc bool testing internal handleexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const third party googletest src googletest src gtest cc stack trace reladdr function file line braveprofilemanagertest excludeservicesinotrandguestprofiles test runtestonmainthread brave browser profiles brave profile manager browsertest cc content browsertestbase proxyruntestonmainthreadloop content public test browser test base cc content browsertestbase setup content public test browser test base cc androidbrowsertest setup chrome test base android android browser test cc bool testing internal handleexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const third party googletest src googletest src gtest cc failed
0
166,790
12,972,020,216
IssuesEvent
2020-07-21 11:55:28
prisma/prisma-client-js
https://api.github.com/repos/prisma/prisma-client-js
closed
Enum value 'false' breaks Prisma client create
bug/2-confirmed kind/bug status/needs-fix-confirmation team/engines topic: test-utils
## Bug description If an enum has 'false' as one of the possible values. Prisma client fails at runtime while providing 'false' as the value for that enum. ## How to reproduce 1. Run the following SQL in a MySQL (untested with Postgres, SQLite) database ``` CREATE TABLE `e` ( `id` bigint(20) unsigned NOT NULL, `visible` enum('true','false') NOT NULL DEFAULT 'true', PRIMARY KEY (`id`) ) ENGINE=MyISAM DEFAULT CHARSET=utf8; ``` 2. Introspect and generate client ``` model e { id Int @id visible e_visible @default(true) } enum e_visible { true false } ``` 3. Run the following Prisma client code ``` const data = await prisma.e.create({ data: { id: 1, visible: 'false', }, }) ``` 4. It fails with ``` divyendusingh [p2-enum-bool]$ node index.js 130 ↵ (node:21526) UnhandledPromiseRejectionWarning: Error: Invalid `prisma.e.create()` invocation in /Users/divyendusingh/Documents/prisma/p2-enum-bool/index.js:6:31 2 3 const prisma = new PrismaClient() 4 5 async function main() { → 6 const data = await prisma.e.create( Failed to validate the query `Error occurred during query validation & transformation: Mutation (object) ↳ createOnee (field) ↳ data (argument) ↳ eCreateInput (object) ↳ visible (field) ↳ Error parsing value: Enum value 'false ' is invalid for enum type e_visible.` at `.Mutation.createOnee.data.eCreateInput.visible` at PrismaClientFetcher._request (/Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/client/runtime/index.js:1:206978) at processTicksAndRejections (internal/process/task_queues.js:97:5) (node:21526) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag `--unhandled-rejections=strict` (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 1) (node:21526) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code. ``` ## Expected behavior It should work ## Prisma information ``` yarn run v1.22.4 $ /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/.bin/prisma --version @prisma/cli : 2.3.0-dev.16 Current platform : darwin Query Engine : query-engine 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/query-engine-darwin) Migration Engine : migration-engine-cli 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/migration-engine-darwin) Introspection Engine : introspection-core 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/introspection-engine-darwin) Format Binary : prisma-fmt 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/prisma-fmt-darwin) Done in 1.42s. ``` Test utils database affected: - mysql/piwigo
1.0
Enum value 'false' breaks Prisma client create - ## Bug description If an enum has 'false' as one of the possible values. Prisma client fails at runtime while providing 'false' as the value for that enum. ## How to reproduce 1. Run the following SQL in a MySQL (untested with Postgres, SQLite) database ``` CREATE TABLE `e` ( `id` bigint(20) unsigned NOT NULL, `visible` enum('true','false') NOT NULL DEFAULT 'true', PRIMARY KEY (`id`) ) ENGINE=MyISAM DEFAULT CHARSET=utf8; ``` 2. Introspect and generate client ``` model e { id Int @id visible e_visible @default(true) } enum e_visible { true false } ``` 3. Run the following Prisma client code ``` const data = await prisma.e.create({ data: { id: 1, visible: 'false', }, }) ``` 4. It fails with ``` divyendusingh [p2-enum-bool]$ node index.js 130 ↵ (node:21526) UnhandledPromiseRejectionWarning: Error: Invalid `prisma.e.create()` invocation in /Users/divyendusingh/Documents/prisma/p2-enum-bool/index.js:6:31 2 3 const prisma = new PrismaClient() 4 5 async function main() { → 6 const data = await prisma.e.create( Failed to validate the query `Error occurred during query validation & transformation: Mutation (object) ↳ createOnee (field) ↳ data (argument) ↳ eCreateInput (object) ↳ visible (field) ↳ Error parsing value: Enum value 'false ' is invalid for enum type e_visible.` at `.Mutation.createOnee.data.eCreateInput.visible` at PrismaClientFetcher._request (/Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/client/runtime/index.js:1:206978) at processTicksAndRejections (internal/process/task_queues.js:97:5) (node:21526) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag `--unhandled-rejections=strict` (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 1) (node:21526) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code. ``` ## Expected behavior It should work ## Prisma information ``` yarn run v1.22.4 $ /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/.bin/prisma --version @prisma/cli : 2.3.0-dev.16 Current platform : darwin Query Engine : query-engine 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/query-engine-darwin) Migration Engine : migration-engine-cli 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/migration-engine-darwin) Introspection Engine : introspection-core 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/introspection-engine-darwin) Format Binary : prisma-fmt 21bd701db858e37b8f3157190a86da93e21ec5df (at /Users/divyendusingh/Documents/prisma/p2-enum-bool/node_modules/@prisma/cli/prisma-fmt-darwin) Done in 1.42s. ``` Test utils database affected: - mysql/piwigo
non_defect
enum value false breaks prisma client create bug description if an enum has false as one of the possible values prisma client fails at runtime while providing false as the value for that enum how to reproduce run the following sql in a mysql untested with postgres sqlite database create table e id bigint unsigned not null visible enum true false not null default true primary key id engine myisam default charset introspect and generate client model e id int id visible e visible default true enum e visible true false run the following prisma client code const data await prisma e create data id visible false it fails with divyendusingh node index js ↵ node unhandledpromiserejectionwarning error invalid prisma e create invocation in users divyendusingh documents prisma enum bool index js const prisma new prismaclient async function main → const data await prisma e create failed to validate the query error occurred during query validation transformation mutation object ↳ createonee field ↳ data argument ↳ ecreateinput object ↳ visible field ↳ error parsing value enum value false is invalid for enum type e visible at mutation createonee data ecreateinput visible at prismaclientfetcher request users divyendusingh documents prisma enum bool node modules prisma client runtime index js at processticksandrejections internal process task queues js node unhandledpromiserejectionwarning unhandled promise rejection this error originated either by throwing inside of an async function without a catch block or by rejecting a promise which was not handled with catch to terminate the node process on unhandled promise rejection use the cli flag unhandled rejections strict see rejection id node deprecationwarning unhandled promise rejections are deprecated in the future promise rejections that are not handled will terminate the node js process with a non zero exit code expected behavior it should work prisma information yarn run users divyendusingh documents prisma enum bool node modules bin prisma version prisma cli dev current platform darwin query engine query engine at users divyendusingh documents prisma enum bool node modules prisma cli query engine darwin migration engine migration engine cli at users divyendusingh documents prisma enum bool node modules prisma cli migration engine darwin introspection engine introspection core at users divyendusingh documents prisma enum bool node modules prisma cli introspection engine darwin format binary prisma fmt at users divyendusingh documents prisma enum bool node modules prisma cli prisma fmt darwin done in test utils database affected mysql piwigo
0
696,847
23,918,512,171
IssuesEvent
2022-09-09 14:42:18
camsaul/methodical
https://api.github.com/repos/camsaul/methodical
closed
`trace!` and `untrace!` facilities to trace existing existing code without changing it
enhancement high-priority!
- It would be good to be able to trace all usages of a certain method without touching existing code - It would be good to be able to trace method calls in an external library Maybe we can use `alter-var-root!` or something swap out the untraced multimethod with a traced one
1.0
`trace!` and `untrace!` facilities to trace existing existing code without changing it - - It would be good to be able to trace all usages of a certain method without touching existing code - It would be good to be able to trace method calls in an external library Maybe we can use `alter-var-root!` or something swap out the untraced multimethod with a traced one
non_defect
trace and untrace facilities to trace existing existing code without changing it it would be good to be able to trace all usages of a certain method without touching existing code it would be good to be able to trace method calls in an external library maybe we can use alter var root or something swap out the untraced multimethod with a traced one
0
567,896
16,918,839,507
IssuesEvent
2021-06-25 00:12:22
googleapis/gapic-generator-python
https://api.github.com/repos/googleapis/gapic-generator-python
opened
`name 'warnings' is not defined` raised for clients with deprecated methods
priority: p2 type: bug
See [log](https://source.cloud.google.com/results/invocations/2f31c792-2c4e-453f-943b-08403508b66f/targets/github%2Fpython-container/tests) from https://github.com/googleapis/python-container/pull/115 for an example
1.0
`name 'warnings' is not defined` raised for clients with deprecated methods - See [log](https://source.cloud.google.com/results/invocations/2f31c792-2c4e-453f-943b-08403508b66f/targets/github%2Fpython-container/tests) from https://github.com/googleapis/python-container/pull/115 for an example
non_defect
name warnings is not defined raised for clients with deprecated methods see from for an example
0
218,669
7,332,097,136
IssuesEvent
2018-03-05 15:25:31
NCEAS/metacat
https://api.github.com/repos/NCEAS/metacat
closed
continue updating user documentation
Category: metacat Component: Bugzilla-Id Priority: Normal Status: Resolved Tracker: Bug
--- Author Name: **Matt Jones** (Matt Jones) Original Redmine Issue: 5516, https://projects.ecoinformatics.org/ecoinfo/issues/5516 Original Date: 2011-10-26 Original Assignee: ben leinfelder --- Documentation needs editing to describe new 2.0.0 features, including support for new DataONE APIs, deprecation of older servlet APIs, and general cleanup. If possible, moving the admin guide to the sphinx system now would be good.
1.0
continue updating user documentation - --- Author Name: **Matt Jones** (Matt Jones) Original Redmine Issue: 5516, https://projects.ecoinformatics.org/ecoinfo/issues/5516 Original Date: 2011-10-26 Original Assignee: ben leinfelder --- Documentation needs editing to describe new 2.0.0 features, including support for new DataONE APIs, deprecation of older servlet APIs, and general cleanup. If possible, moving the admin guide to the sphinx system now would be good.
non_defect
continue updating user documentation author name matt jones matt jones original redmine issue original date original assignee ben leinfelder documentation needs editing to describe new features including support for new dataone apis deprecation of older servlet apis and general cleanup if possible moving the admin guide to the sphinx system now would be good
0
292,688
25,229,496,582
IssuesEvent
2022-11-14 18:33:13
spack/spack
https://api.github.com/repos/spack/spack
closed
Testing issue: ginkgo@1.4.0%gcc@11.1.0
test-error
### Steps to reproduce the failure(s) or link(s) to test output(s) @hartwiganzt @tcojean Stand-alone tests failed during a Spack PR pipeline run. More information can be found at: - Build: https://cdash.spack.io/build/1816848 - Stand-alone tests: https://cdash.spack.io/viewTest.php?buildid=1816848 ### Error message Build test software: ``` [ 25%] Building CXX object CMakeFiles/test_install.dir/test_install.cpp.o [ 50%] Linking CXX executable test_install /usr/bin/ld: warning: librocrand.so.1, needed by /home/software/spack/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/linux-ubuntu20.04-x86_64/gcc-11.1.0/rocrand-5.2.3-srqqcyn4wxm4uwfo3ulghfzxia2zvuq5/lib/libhiprand.so.1, not found (try using -rpath or -rpath-link) /usr/bin/ld: /home/software/spack/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/linux-ubuntu20.04-x86_64/gcc-11.1.0/rocrand-5.2.3-srqqcyn4wxm4uwfo3ulghfzxia2zvuq5/lib/libhiprand.so.1: undefined reference to `rocrand_generate_normal_half' /usr/bin/ld: /home/software/spack/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/linux-ubuntu20.04-x86_64/gcc-11.1.0/rocrand-5.2.3-srqqcyn4wxm4uwfo3ulghfzxia2zvuq5/lib/libhiprand.so.1: undefined reference to `rocrand_generate_normal' ... ``` ### Information on your system or the test runner Runner: uo-reptar OS Release: 4.18.0-372.9.1.el8.x86_64 Spack version: spack-0.19.0.dev0 (10720fbb4e277c97a262dbf8d51f6b93c27dab2a) Repository revision: [10720fbb4e277c97a262dbf8d51f6b93c27dab2a](https://github.com/spack/spack/commit/10720fbb4e277c97a262dbf8d51f6b93c27dab2a) ### Additional information _No response_ ### General information - [X] I have reported the version of Spack/Python/Platform/Runner - [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers - [X] I have uploaded any available logs - [X] I have searched the issues of this repo and believe this is not a duplicate
1.0
Testing issue: ginkgo@1.4.0%gcc@11.1.0 - ### Steps to reproduce the failure(s) or link(s) to test output(s) @hartwiganzt @tcojean Stand-alone tests failed during a Spack PR pipeline run. More information can be found at: - Build: https://cdash.spack.io/build/1816848 - Stand-alone tests: https://cdash.spack.io/viewTest.php?buildid=1816848 ### Error message Build test software: ``` [ 25%] Building CXX object CMakeFiles/test_install.dir/test_install.cpp.o [ 50%] Linking CXX executable test_install /usr/bin/ld: warning: librocrand.so.1, needed by /home/software/spack/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/linux-ubuntu20.04-x86_64/gcc-11.1.0/rocrand-5.2.3-srqqcyn4wxm4uwfo3ulghfzxia2zvuq5/lib/libhiprand.so.1, not found (try using -rpath or -rpath-link) /usr/bin/ld: /home/software/spack/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/linux-ubuntu20.04-x86_64/gcc-11.1.0/rocrand-5.2.3-srqqcyn4wxm4uwfo3ulghfzxia2zvuq5/lib/libhiprand.so.1: undefined reference to `rocrand_generate_normal_half' /usr/bin/ld: /home/software/spack/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spack_path_placeholder__/__spa/linux-ubuntu20.04-x86_64/gcc-11.1.0/rocrand-5.2.3-srqqcyn4wxm4uwfo3ulghfzxia2zvuq5/lib/libhiprand.so.1: undefined reference to `rocrand_generate_normal' ... ``` ### Information on your system or the test runner Runner: uo-reptar OS Release: 4.18.0-372.9.1.el8.x86_64 Spack version: spack-0.19.0.dev0 (10720fbb4e277c97a262dbf8d51f6b93c27dab2a) Repository revision: [10720fbb4e277c97a262dbf8d51f6b93c27dab2a](https://github.com/spack/spack/commit/10720fbb4e277c97a262dbf8d51f6b93c27dab2a) ### Additional information _No response_ ### General information - [X] I have reported the version of Spack/Python/Platform/Runner - [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers - [X] I have uploaded any available logs - [X] I have searched the issues of this repo and believe this is not a duplicate
non_defect
testing issue ginkgo gcc steps to reproduce the failure s or link s to test output s hartwiganzt tcojean stand alone tests failed during a spack pr pipeline run more information can be found at build stand alone tests error message build test software building cxx object cmakefiles test install dir test install cpp o linking cxx executable test install usr bin ld warning librocrand so needed by home software spack spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spa linux gcc rocrand lib libhiprand so not found try using rpath or rpath link usr bin ld home software spack spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spa linux gcc rocrand lib libhiprand so undefined reference to rocrand generate normal half usr bin ld home software spack spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spack path placeholder spa linux gcc rocrand lib libhiprand so undefined reference to rocrand generate normal information on your system or the test runner runner uo reptar os release spack version spack repository revision additional information no response general information i have reported the version of spack python platform runner i have run spack maintainers and mentioned any maintainers i have uploaded any available logs i have searched the issues of this repo and believe this is not a duplicate
0
58,630
16,662,876,925
IssuesEvent
2021-06-06 16:49:36
Questie/Questie
https://api.github.com/repos/Questie/Questie
opened
Questie error for the quest Gorgrom the dragon-Eater
Type - Defect
Hi, using questie 6.3.14 Got mass error when accepting Gorgrom the dragon-Eater in blades edge. Questie: [ERROR] [QuestieQuest]: v6.3.14 - There was an error populating objectives for Gorgrom the Dragon-Eater 10723 1 No error ...erface\AddOns\Questie\Modules\Quest\QuestieQuest.lua:690: bad argument #1 to 'next' (table expected, got nil) [string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:556: in function <...erface\AddOns\Questie\Modules\Quest\QuestieQuest.lua:554> [string "=[C]"]: in function `next' [string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:690: in function <...erface\AddOns\Questie\Modules\Quest\QuestieQuest.lua:683> [string "=[C]"]: in function `xpcall' [string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:554: in function `UpdateObjectiveNotes' [string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:469: in function `UpdateQuest' [string "@Interface\AddOns\Questie\Modules\Libs\QuestieHash.lua"]:108: in function <...nterface\AddOns\Questie\Modules\Libs\QuestieHash.lua:103> [string "@Interface\AddOns\Questie\Modules\Libs\QuestieHash.lua"]:91: in function `CompareQuestHashes' [string "@Interface\AddOns\Questie\Modules\QuestieEventHandler.lua"]:296: in function <...rface\AddOns\Questie\Modules\QuestieEventHandler.lua:278> [string "=[C]"]: ? [string "@Interface\AddOns\Quartz\libs\AceBucket-3.0\AceBucket-3.0.lua"]:73: in function <...e\AddOns\Quartz\libs\AceBucket-3.0\AceBucket-3.0.lua:71> [string "@Interface\AddOns\Quartz\libs\AceBucket-3.0\AceBucket-3.0.lua"]:89: in function `func' [string "@Interface\AddOns\Details\Libs\AceTimer-3.0\AceTimer-3.0.lua"]:57: in function <...ce\AddOns\Details\Libs\AceTimer-3.0\AceTimer-3.0.lua:50>
1.0
Questie error for the quest Gorgrom the dragon-Eater - Hi, using questie 6.3.14 Got mass error when accepting Gorgrom the dragon-Eater in blades edge. Questie: [ERROR] [QuestieQuest]: v6.3.14 - There was an error populating objectives for Gorgrom the Dragon-Eater 10723 1 No error ...erface\AddOns\Questie\Modules\Quest\QuestieQuest.lua:690: bad argument #1 to 'next' (table expected, got nil) [string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:556: in function <...erface\AddOns\Questie\Modules\Quest\QuestieQuest.lua:554> [string "=[C]"]: in function `next' [string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:690: in function <...erface\AddOns\Questie\Modules\Quest\QuestieQuest.lua:683> [string "=[C]"]: in function `xpcall' [string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:554: in function `UpdateObjectiveNotes' [string "@Interface\AddOns\Questie\Modules\Quest\QuestieQuest.lua"]:469: in function `UpdateQuest' [string "@Interface\AddOns\Questie\Modules\Libs\QuestieHash.lua"]:108: in function <...nterface\AddOns\Questie\Modules\Libs\QuestieHash.lua:103> [string "@Interface\AddOns\Questie\Modules\Libs\QuestieHash.lua"]:91: in function `CompareQuestHashes' [string "@Interface\AddOns\Questie\Modules\QuestieEventHandler.lua"]:296: in function <...rface\AddOns\Questie\Modules\QuestieEventHandler.lua:278> [string "=[C]"]: ? [string "@Interface\AddOns\Quartz\libs\AceBucket-3.0\AceBucket-3.0.lua"]:73: in function <...e\AddOns\Quartz\libs\AceBucket-3.0\AceBucket-3.0.lua:71> [string "@Interface\AddOns\Quartz\libs\AceBucket-3.0\AceBucket-3.0.lua"]:89: in function `func' [string "@Interface\AddOns\Details\Libs\AceTimer-3.0\AceTimer-3.0.lua"]:57: in function <...ce\AddOns\Details\Libs\AceTimer-3.0\AceTimer-3.0.lua:50>
defect
questie error for the quest gorgrom the dragon eater hi using questie got mass error when accepting gorgrom the dragon eater in blades edge questie there was an error populating objectives for gorgrom the dragon eater no error erface addons questie modules quest questiequest lua bad argument to next table expected got nil in function in function next in function in function xpcall in function updateobjectivenotes in function updatequest in function in function comparequesthashes in function in function in function func in function
1
69,057
22,098,352,015
IssuesEvent
2022-06-01 11:58:24
DEVA9N/FlexTimeMonitor
https://api.github.com/repos/DEVA9N/FlexTimeMonitor
closed
The options (break time) are gone after reinstalling the programm
Priority-Medium Type-Defect auto-migrated
``` What steps will reproduce the problem? 1. Reinstall or install a new version of Flex Time Monitor 2. Notice how the default options are used What is the expected output? What do you see instead? expected: the options will stay the same as before the new installation instead: the default options are set ``` Original issue reported on code.google.com by `aj9...@gmail.com` on 6 Aug 2012 at 3:25
1.0
The options (break time) are gone after reinstalling the programm - ``` What steps will reproduce the problem? 1. Reinstall or install a new version of Flex Time Monitor 2. Notice how the default options are used What is the expected output? What do you see instead? expected: the options will stay the same as before the new installation instead: the default options are set ``` Original issue reported on code.google.com by `aj9...@gmail.com` on 6 Aug 2012 at 3:25
defect
the options break time are gone after reinstalling the programm what steps will reproduce the problem reinstall or install a new version of flex time monitor notice how the default options are used what is the expected output what do you see instead expected the options will stay the same as before the new installation instead the default options are set original issue reported on code google com by gmail com on aug at
1
38,599
8,924,971,976
IssuesEvent
2019-01-21 20:44:11
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
Python utils check file size can fail (race condition)
C: MOOSE Scripts C: TestHarness P: normal T: defect
## Rationale <!--What is the reason for this enhancement or what error are you reporting?--> The python/mooseutils/tests.check_file_size test can fail due to a race condition. The problem is that this test runs concurrently with several other tests on the system and it does an os.walk() looking for files sizes to sum up. It's possible for the file list to change (e.g. files can be removed) while this test is running causing the `os.stat()` command to fail. This test should be cleaned up to work in either a clean directory where other test outputs are not being written, or the test should be serialized and isolated. ## Description <!--Provide details of the enhancement or instructions for reproducing the error.--> Race condition in python testing (mooseutils.check_file_size). This failure is difficult to replicate, but it should be fairly easy to clean up. See description above. ## Impact <!--How will the changes impact the code, developers, and users? Discuss changes to the internal interfaces and public API.--> minor: Does not affect MOOSE, but can impede testing.
1.0
Python utils check file size can fail (race condition) - ## Rationale <!--What is the reason for this enhancement or what error are you reporting?--> The python/mooseutils/tests.check_file_size test can fail due to a race condition. The problem is that this test runs concurrently with several other tests on the system and it does an os.walk() looking for files sizes to sum up. It's possible for the file list to change (e.g. files can be removed) while this test is running causing the `os.stat()` command to fail. This test should be cleaned up to work in either a clean directory where other test outputs are not being written, or the test should be serialized and isolated. ## Description <!--Provide details of the enhancement or instructions for reproducing the error.--> Race condition in python testing (mooseutils.check_file_size). This failure is difficult to replicate, but it should be fairly easy to clean up. See description above. ## Impact <!--How will the changes impact the code, developers, and users? Discuss changes to the internal interfaces and public API.--> minor: Does not affect MOOSE, but can impede testing.
defect
python utils check file size can fail race condition rationale the python mooseutils tests check file size test can fail due to a race condition the problem is that this test runs concurrently with several other tests on the system and it does an os walk looking for files sizes to sum up it s possible for the file list to change e g files can be removed while this test is running causing the os stat command to fail this test should be cleaned up to work in either a clean directory where other test outputs are not being written or the test should be serialized and isolated description race condition in python testing mooseutils check file size this failure is difficult to replicate but it should be fairly easy to clean up see description above impact how will the changes impact the code developers and users discuss changes to the internal interfaces and public api minor does not affect moose but can impede testing
1
56,058
14,916,170,434
IssuesEvent
2021-01-22 17:47:39
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
[CI/CD]: Review coverage of accessibility checks in 996 end-to-end tests
508-defect-3 508/Accessibility HLR testing vsa vsa-benefits
**Feedback framework** - **❗️ Must** for if the feedback must be applied - **⚠️Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Description Applications **must** have thorough end-to-end tests that run in our continuous integration/continuous deployment (CI/CD) pipeline. These tests should include thorough axe checks. While auditing the `/disability-benefits/996` app folder, I wasn't sure if there were modals or hidden content that needed axe checks. I'd like the front-end engineering team to review this application, and add end-to-end tests as needed. Definition of done in acceptance criteria below. ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** _Jennifer_ ## Environment * `vets-website/src/applications/disability-benefits/996/tests` * `$ yarn test:e2e src/applications/disability-benefits/996/tests` ## Acceptance Criteria <!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. --> **Definition of done:** - [ ] Front-end team member(s) have reviewed end-to-end tests for axe checks - [ ] axe checks are run for hidden content like modal windows, accordions - [ ] FE team has consulted with accessibility specialist in cases where there are high numbers of modals, accordions, other hidden content that could slow down e2e test runs. - [ ] No axe `violations` appear in the plugin console. These will break the CI/CD build. ## WCAG or Vendor Guidance (optional) * [Custom axeCheck helper method](https://github.com/department-of-veterans-affairs/vets-website/blob/master/src/platform/testing/e2e/nightwatch-commands/axeCheck.js) * [VSP guidance on writing end-to-end tests](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/quality-assurance/e2e-testing)
1.0
[CI/CD]: Review coverage of accessibility checks in 996 end-to-end tests - **Feedback framework** - **❗️ Must** for if the feedback must be applied - **⚠️Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Description Applications **must** have thorough end-to-end tests that run in our continuous integration/continuous deployment (CI/CD) pipeline. These tests should include thorough axe checks. While auditing the `/disability-benefits/996` app folder, I wasn't sure if there were modals or hidden content that needed axe checks. I'd like the front-end engineering team to review this application, and add end-to-end tests as needed. Definition of done in acceptance criteria below. ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** _Jennifer_ ## Environment * `vets-website/src/applications/disability-benefits/996/tests` * `$ yarn test:e2e src/applications/disability-benefits/996/tests` ## Acceptance Criteria <!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. --> **Definition of done:** - [ ] Front-end team member(s) have reviewed end-to-end tests for axe checks - [ ] axe checks are run for hidden content like modal windows, accordions - [ ] FE team has consulted with accessibility specialist in cases where there are high numbers of modals, accordions, other hidden content that could slow down e2e test runs. - [ ] No axe `violations` appear in the plugin console. These will break the CI/CD build. ## WCAG or Vendor Guidance (optional) * [Custom axeCheck helper method](https://github.com/department-of-veterans-affairs/vets-website/blob/master/src/platform/testing/e2e/nightwatch-commands/axeCheck.js) * [VSP guidance on writing end-to-end tests](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/quality-assurance/e2e-testing)
defect
review coverage of accessibility checks in end to end tests feedback framework ❗️ must for if the feedback must be applied ⚠️should if the feedback is best practice ✔️ consider for suggestions enhancements description applications must have thorough end to end tests that run in our continuous integration continuous deployment ci cd pipeline these tests should include thorough axe checks while auditing the disability benefits app folder i wasn t sure if there were modals or hidden content that needed axe checks i d like the front end engineering team to review this application and add end to end tests as needed definition of done in acceptance criteria below point of contact if this issue is being opened by a vfs team member please add a point of contact usually this is the same person who enters the issue ticket vfs point of contact jennifer environment vets website src applications disability benefits tests yarn test src applications disability benefits tests acceptance criteria definition of done front end team member s have reviewed end to end tests for axe checks axe checks are run for hidden content like modal windows accordions fe team has consulted with accessibility specialist in cases where there are high numbers of modals accordions other hidden content that could slow down test runs no axe violations appear in the plugin console these will break the ci cd build wcag or vendor guidance optional
1
29,023
8,250,827,653
IssuesEvent
2018-09-12 05:11:34
avast-tl/retdec
https://api.github.com/repos/avast-tl/retdec
closed
Could NOT find PythonInterp: Found unsuitable version "2.7.14", but required is at least "3.4" (found /usr/local/bin/python)
C-build-system O-macos P-build
Someone please help me. OSX, I have both python 2.7 installed and python 3.7 installed. My PATH is /usr/local/opt/python3.7:/Library/Frameworks/Python.framework/Versions/3.7/bin/python3:/usr/local/opt/flex/bin:/usr/local/opt/bison/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin When I try to cmake it always says: Could NOT find PythonInterp: Found unsuitable version "2.7.14", but required is at least "3.4" (found /usr/local/bin/python) How do I fix this?
2.0
Could NOT find PythonInterp: Found unsuitable version "2.7.14", but required is at least "3.4" (found /usr/local/bin/python) - Someone please help me. OSX, I have both python 2.7 installed and python 3.7 installed. My PATH is /usr/local/opt/python3.7:/Library/Frameworks/Python.framework/Versions/3.7/bin/python3:/usr/local/opt/flex/bin:/usr/local/opt/bison/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin When I try to cmake it always says: Could NOT find PythonInterp: Found unsuitable version "2.7.14", but required is at least "3.4" (found /usr/local/bin/python) How do I fix this?
non_defect
could not find pythoninterp found unsuitable version but required is at least found usr local bin python someone please help me osx i have both python installed and python installed my path is usr local opt library frameworks python framework versions bin usr local opt flex bin usr local opt bison bin usr local bin usr bin bin usr sbin sbin when i try to cmake it always says could not find pythoninterp found unsuitable version but required is at least found usr local bin python how do i fix this
0
7,155
2,610,329,582
IssuesEvent
2015-02-26 19:46:12
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Map Issue
auto-migrated Priority-Low Type-Defect
``` I've got a few bugs to report. I've noticed that both Sluis Van and Raxus Prime do not have Reinforcement points when I play on them for some reason. When I originally fought on the planets they had reinforcement points but they dissapeared after I conquered them. Then, the CIS will launch assaults on these worlds and will have no Reinforcement points to drop reinforcements from. ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 15 Jun 2011 at 12:45
1.0
Map Issue - ``` I've got a few bugs to report. I've noticed that both Sluis Van and Raxus Prime do not have Reinforcement points when I play on them for some reason. When I originally fought on the planets they had reinforcement points but they dissapeared after I conquered them. Then, the CIS will launch assaults on these worlds and will have no Reinforcement points to drop reinforcements from. ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 15 Jun 2011 at 12:45
defect
map issue i ve got a few bugs to report i ve noticed that both sluis van and raxus prime do not have reinforcement points when i play on them for some reason when i originally fought on the planets they had reinforcement points but they dissapeared after i conquered them then the cis will launch assaults on these worlds and will have no reinforcement points to drop reinforcements from original issue reported on code google com by gmail com on jun at
1
40,380
9,977,380,804
IssuesEvent
2019-07-09 17:07:37
mozilla/experimenter
https://api.github.com/repos/mozilla/experimenter
closed
The "Help" button for the "Lightning Advisory (Optional)" Sign-off is wrongly linked to the mana page/"Dependent Sign-offs"
Defect P1 - High Priority [QA]:Normal issue
**[Action Needed]:** Please link the "Lightning Advising" help here: https://mana.mozilla.org/wiki/display/FIREFOX/Pref-Flip+and+Add-On+Experiments#Pref-FlipandAdd-OnExperiments-LightningAdvising **[Notes]:** - I logged this issue because the "Lightning Advisory (Optional)" Sign-off doesn't have any dependencies with the Risk questions and also, it's description on the mana page is displayed in the "Optional Sign-off" section. **[Affected versions]:** - Firefox Release 67.0.1 and above - Chrome 75 and above. **[Affected Platforms]:** - All Windows - All Mac - All Linux **[Prerequisites]:** - You are logged in your LDAP account on Experimenter stage website, "https://experimenter.stage.mozaws.net/". - Have an experiment in the "Ready for Sign-Off state, created on Experimenter stage website. **[Steps to reproduce]:** 1. Open the browser with the profile from prerequisites and navigate to "https://experimenter.stage.mozaws.net/". 2. Open the experiment from prerequisites. 3. Click the "Help" button for the "Lightning Advisory (Optional)" Sign-Off. 4. Observe the behavior. **[Expected result]:** - The mana page is opened on the "Lightning Advising" section. **[Actual result]:** - The mana page is opened on the "Dependent Sign-offs" section. **[Additional Notes]:** - Attached is a screen recording of the issue: ![LigAdv](https://user-images.githubusercontent.com/40061643/59273966-003a1180-8c62-11e9-856c-17cfa1600706.gif)
1.0
The "Help" button for the "Lightning Advisory (Optional)" Sign-off is wrongly linked to the mana page/"Dependent Sign-offs" - **[Action Needed]:** Please link the "Lightning Advising" help here: https://mana.mozilla.org/wiki/display/FIREFOX/Pref-Flip+and+Add-On+Experiments#Pref-FlipandAdd-OnExperiments-LightningAdvising **[Notes]:** - I logged this issue because the "Lightning Advisory (Optional)" Sign-off doesn't have any dependencies with the Risk questions and also, it's description on the mana page is displayed in the "Optional Sign-off" section. **[Affected versions]:** - Firefox Release 67.0.1 and above - Chrome 75 and above. **[Affected Platforms]:** - All Windows - All Mac - All Linux **[Prerequisites]:** - You are logged in your LDAP account on Experimenter stage website, "https://experimenter.stage.mozaws.net/". - Have an experiment in the "Ready for Sign-Off state, created on Experimenter stage website. **[Steps to reproduce]:** 1. Open the browser with the profile from prerequisites and navigate to "https://experimenter.stage.mozaws.net/". 2. Open the experiment from prerequisites. 3. Click the "Help" button for the "Lightning Advisory (Optional)" Sign-Off. 4. Observe the behavior. **[Expected result]:** - The mana page is opened on the "Lightning Advising" section. **[Actual result]:** - The mana page is opened on the "Dependent Sign-offs" section. **[Additional Notes]:** - Attached is a screen recording of the issue: ![LigAdv](https://user-images.githubusercontent.com/40061643/59273966-003a1180-8c62-11e9-856c-17cfa1600706.gif)
defect
the help button for the lightning advisory optional sign off is wrongly linked to the mana page dependent sign offs please link the lightning advising help here i logged this issue because the lightning advisory optional sign off doesn t have any dependencies with the risk questions and also it s description on the mana page is displayed in the optional sign off section firefox release and above chrome and above all windows all mac all linux you are logged in your ldap account on experimenter stage website have an experiment in the ready for sign off state created on experimenter stage website open the browser with the profile from prerequisites and navigate to open the experiment from prerequisites click the help button for the lightning advisory optional sign off observe the behavior the mana page is opened on the lightning advising section the mana page is opened on the dependent sign offs section attached is a screen recording of the issue
1
187,095
14,426,956,619
IssuesEvent
2020-12-06 01:00:41
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
giantswarm/kvm-operator-node-controller: vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go; 30 LoC
fresh small test vendored
Found a possible issue in [giantswarm/kvm-operator-node-controller](https://www.github.com/giantswarm/kvm-operator-node-controller) at [vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go](https://github.com/giantswarm/kvm-operator-node-controller/blob/7146561e54142d4f986daee0206336ebee3ceb18/vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go#L1064-L1093) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to svc at line 1067 may start a goroutine [Click here to see the code in its original context.](https://github.com/giantswarm/kvm-operator-node-controller/blob/7146561e54142d4f986daee0206336ebee3ceb18/vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go#L1064-L1093) <details> <summary>Click here to show the 30 line(s) of Go which triggered the analyzer.</summary> ```go for _, svc := range tests { for _, wide := range []bool{false, true} { buff := bytes.Buffer{} printService(&svc, &buff, printers.PrintOptions{Wide: wide}) output := string(buff.Bytes()) ip := svc.Spec.ClusterIP if !strings.Contains(output, ip) { t.Errorf("expected to contain ClusterIP %s, but doesn't: %s", ip, output) } for n, ingress := range svc.Status.LoadBalancer.Ingress { ip = ingress.IP // For non-wide output, we only guarantee the first IP to be printed if (n == 0 || wide) && !strings.Contains(output, ip) { t.Errorf("expected to contain ingress ip %s with wide=%v, but doesn't: %s", ip, wide, output) } } for _, port := range svc.Spec.Ports { portSpec := fmt.Sprintf("%d/%s", port.Port, port.Protocol) if !strings.Contains(output, portSpec) { t.Errorf("expected to contain port: %s, but doesn't: %s", portSpec, output) } } // Each service should print on one line if 1 != strings.Count(output, "\n") { t.Errorf("expected a single newline, found %d", strings.Count(output, "\n")) } } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 7146561e54142d4f986daee0206336ebee3ceb18
1.0
giantswarm/kvm-operator-node-controller: vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go; 30 LoC - Found a possible issue in [giantswarm/kvm-operator-node-controller](https://www.github.com/giantswarm/kvm-operator-node-controller) at [vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go](https://github.com/giantswarm/kvm-operator-node-controller/blob/7146561e54142d4f986daee0206336ebee3ceb18/vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go#L1064-L1093) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to svc at line 1067 may start a goroutine [Click here to see the code in its original context.](https://github.com/giantswarm/kvm-operator-node-controller/blob/7146561e54142d4f986daee0206336ebee3ceb18/vendor/k8s.io/kubernetes/pkg/printers/internalversion/printers_test.go#L1064-L1093) <details> <summary>Click here to show the 30 line(s) of Go which triggered the analyzer.</summary> ```go for _, svc := range tests { for _, wide := range []bool{false, true} { buff := bytes.Buffer{} printService(&svc, &buff, printers.PrintOptions{Wide: wide}) output := string(buff.Bytes()) ip := svc.Spec.ClusterIP if !strings.Contains(output, ip) { t.Errorf("expected to contain ClusterIP %s, but doesn't: %s", ip, output) } for n, ingress := range svc.Status.LoadBalancer.Ingress { ip = ingress.IP // For non-wide output, we only guarantee the first IP to be printed if (n == 0 || wide) && !strings.Contains(output, ip) { t.Errorf("expected to contain ingress ip %s with wide=%v, but doesn't: %s", ip, wide, output) } } for _, port := range svc.Spec.Ports { portSpec := fmt.Sprintf("%d/%s", port.Port, port.Protocol) if !strings.Contains(output, portSpec) { t.Errorf("expected to contain port: %s, but doesn't: %s", portSpec, output) } } // Each service should print on one line if 1 != strings.Count(output, "\n") { t.Errorf("expected a single newline, found %d", strings.Count(output, "\n")) } } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 7146561e54142d4f986daee0206336ebee3ceb18
non_defect
giantswarm kvm operator node controller vendor io kubernetes pkg printers internalversion printers test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to svc at line may start a goroutine click here to show the line s of go which triggered the analyzer go for svc range tests for wide range bool false true buff bytes buffer printservice svc buff printers printoptions wide wide output string buff bytes ip svc spec clusterip if strings contains output ip t errorf expected to contain clusterip s but doesn t s ip output for n ingress range svc status loadbalancer ingress ip ingress ip for non wide output we only guarantee the first ip to be printed if n wide strings contains output ip t errorf expected to contain ingress ip s with wide v but doesn t s ip wide output for port range svc spec ports portspec fmt sprintf d s port port port protocol if strings contains output portspec t errorf expected to contain port s but doesn t s portspec output each service should print on one line if strings count output n t errorf expected a single newline found d strings count output n leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
60,816
14,919,928,882
IssuesEvent
2021-01-23 02:01:44
microsoft/fluentui
https://api.github.com/repos/microsoft/fluentui
closed
Migrate e2e / accessibility tests to Playwright
Area: Build System Area: Testing Fluent UI react-northstar Needs: Discussion 🙋 Type: Feature
Currently Puppeteer will be used to perform basic e2e and accessibility tests which is based on Chromium. What about migrating it over to [Playwright](https://github.com/microsoft/playwright) which supports Chromium / WebKit / Firefox and running / checking it with all of them in a row? I already played a bit around with the [e2e tests](https://github.com/microsoft/fluentui/tree/master/packages/fluentui/e2e/tests) locally and these tests are working mostly almost directly. But I wanted to ask first, what you think about it. I'm open to contribute to it.
1.0
Migrate e2e / accessibility tests to Playwright - Currently Puppeteer will be used to perform basic e2e and accessibility tests which is based on Chromium. What about migrating it over to [Playwright](https://github.com/microsoft/playwright) which supports Chromium / WebKit / Firefox and running / checking it with all of them in a row? I already played a bit around with the [e2e tests](https://github.com/microsoft/fluentui/tree/master/packages/fluentui/e2e/tests) locally and these tests are working mostly almost directly. But I wanted to ask first, what you think about it. I'm open to contribute to it.
non_defect
migrate accessibility tests to playwright currently puppeteer will be used to perform basic and accessibility tests which is based on chromium what about migrating it over to which supports chromium webkit firefox and running checking it with all of them in a row i already played a bit around with the locally and these tests are working mostly almost directly but i wanted to ask first what you think about it i m open to contribute to it
0
785,757
27,624,232,421
IssuesEvent
2023-03-10 04:36:48
AY2223S2-CS2113-W12-1/tp
https://api.github.com/repos/AY2223S2-CS2113-W12-1/tp
opened
Create a remove appointment feature.
type.Story priority.High
As a user, I am able to remove appointments if necessary so that the appointment list is not clogged up.
1.0
Create a remove appointment feature. - As a user, I am able to remove appointments if necessary so that the appointment list is not clogged up.
non_defect
create a remove appointment feature as a user i am able to remove appointments if necessary so that the appointment list is not clogged up
0
4,287
2,610,090,946
IssuesEvent
2015-02-26 18:27:32
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
深圳痘印方法
auto-migrated Priority-Medium Type-Defect
``` 深圳痘印方法【深圳韩方科颜全国热线400-869-1818,24小时QQ4008 691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方—�� �韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科颜� ��业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康祛 痘技术并结合先进“先进豪华彩光”仪,开创国内专业治疗�� �刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:45
1.0
深圳痘印方法 - ``` 深圳痘印方法【深圳韩方科颜全国热线400-869-1818,24小时QQ4008 691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方—�� �韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科颜� ��业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康祛 痘技术并结合先进“先进豪华彩光”仪,开创国内专业治疗�� �刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:45
defect
深圳痘印方法 深圳痘印方法【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方—�� �韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科颜� ��业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康祛 痘技术并结合先进“先进豪华彩光”仪,开创国内专业治疗�� �刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘。 original issue reported on code google com by szft com on may at
1
120,250
4,787,175,689
IssuesEvent
2016-10-29 21:05:26
minj/foxtrick
https://api.github.com/repos/minj/foxtrick
closed
CurrentTransfers caches stale CHPP response if player has just been TLed
bug Priority-Low started Transfers
**From:** LA-MJ **PostID:** [16898912.142](https://www.hattrick.org/goto.ashx?path=%2FForum%2FRead.aspx%3Ft%3D16898912%26n%3D142%26v%3D0) **To:** kosio **Re:** [16898912.140](https://www.hattrick.org/goto.ashx?path=%2FForum%2FRead.aspx%3Ft%3D16898912%26n%3D140%26v%3D0) **Datetime:** 2016-01-17 12:50 **Message:** > kosio rašė: > > Hi, > > Just got the problem with missing opening price again. > > [(http://screencast.com/t/l7g9UNzNc)](http://screencast.com/t/l7g9UNzNc) > [(http://pastebin.com/QMYkcLBX)](http://pastebin.com/QMYkcLBX) > > If you need something else, just let me know. PlayerID would be useful in these situations, otherwise I need to look for it manually. This seems to have been a temporary CHPP hickup. FT log indicates 'AskingPrice' field was missing in the CHPP response. I guess this is because you were so quick to look at the transfers page (screenshot says 2d 23h 57min) that CHPP returned a cached response without transfer data. I guess I need to guard against these somehow...
1.0
CurrentTransfers caches stale CHPP response if player has just been TLed - **From:** LA-MJ **PostID:** [16898912.142](https://www.hattrick.org/goto.ashx?path=%2FForum%2FRead.aspx%3Ft%3D16898912%26n%3D142%26v%3D0) **To:** kosio **Re:** [16898912.140](https://www.hattrick.org/goto.ashx?path=%2FForum%2FRead.aspx%3Ft%3D16898912%26n%3D140%26v%3D0) **Datetime:** 2016-01-17 12:50 **Message:** > kosio rašė: > > Hi, > > Just got the problem with missing opening price again. > > [(http://screencast.com/t/l7g9UNzNc)](http://screencast.com/t/l7g9UNzNc) > [(http://pastebin.com/QMYkcLBX)](http://pastebin.com/QMYkcLBX) > > If you need something else, just let me know. PlayerID would be useful in these situations, otherwise I need to look for it manually. This seems to have been a temporary CHPP hickup. FT log indicates 'AskingPrice' field was missing in the CHPP response. I guess this is because you were so quick to look at the transfers page (screenshot says 2d 23h 57min) that CHPP returned a cached response without transfer data. I guess I need to guard against these somehow...
non_defect
currenttransfers caches stale chpp response if player has just been tled from la mj postid to kosio re datetime message kosio rašė hi just got the problem with missing opening price again if you need something else just let me know playerid would be useful in these situations otherwise i need to look for it manually this seems to have been a temporary chpp hickup ft log indicates askingprice field was missing in the chpp response i guess this is because you were so quick to look at the transfers page screenshot says that chpp returned a cached response without transfer data i guess i need to guard against these somehow
0
460,752
13,217,743,422
IssuesEvent
2020-08-17 07:25:03
netdata/netdata
https://api.github.com/repos/netdata/netdata
closed
New exporting module is missing metrics compared to old when sending to graphite
area/backends bug priority/medium
##### Bug report summary I have migrated the exporting config from the old way (using `[backend]`) to the new way by using `exporting.conf` However, there is something strange, or I have put wrong config, because we do not get all the metrics. ##### OS / Environment Running netdata v1.23.1 on docker, together with grafana latest (v7.0.6) in docker and graphite 1.1.7-3 also in docker. I am running them on a macOS host, but I have also seen missing stats on our ubuntu 18.04 servers. ##### Component Name The metrics send to graphite have gaps, which did not happen before. I have made a test by running a single netdata with the new config and old config. ![Screenshot 2020-07-10 at 0 09 11](https://user-images.githubusercontent.com/8693064/87086242-5cdf7580-c242-11ea-9557-2f605e4209c2.png) The yellow and green uptime are when I run netdata with the new exporting config The red and blue uptime are when I run netdata with the old backend config ##### Steps To Reproduce 1. Run netdata with new backend and old backend and send to graphite Here is my `[backend]` config: ``` [backend] enabled = yes type = graphite destination = docker.for.mac.localhost:2023 data source = as collected prefix = netdata update every = 10 buffer on failures = 10 timeout ms = 20000 send charts matching = system.uptime* system.cpu* system.ram* system.net* disk_space.* !* send hosts matching = localhost * send names instead of ids = yes ``` Here is my `exporting.conf` file: ``` [graphite:netdata] enabled = yes destination = docker.for.mac.localhost:2023 data source = as collected prefix = netdata update every = 10 buffer on failures = 10 timeout ms = 20000 send charts matching = system.uptime* system.cpu* system.ram* system.net* disk_space.* !* send hosts matching = localhost * send names instead of ids = yes ``` This is the only config which I have changed when testing this behaviour ##### Expected behavior I expect that I do not see gaps in the dashboard when using the new style. It could be that I have configured the new one in a wrong way, in such a case I would like to hear any suggestions.
1.0
New exporting module is missing metrics compared to old when sending to graphite - ##### Bug report summary I have migrated the exporting config from the old way (using `[backend]`) to the new way by using `exporting.conf` However, there is something strange, or I have put wrong config, because we do not get all the metrics. ##### OS / Environment Running netdata v1.23.1 on docker, together with grafana latest (v7.0.6) in docker and graphite 1.1.7-3 also in docker. I am running them on a macOS host, but I have also seen missing stats on our ubuntu 18.04 servers. ##### Component Name The metrics send to graphite have gaps, which did not happen before. I have made a test by running a single netdata with the new config and old config. ![Screenshot 2020-07-10 at 0 09 11](https://user-images.githubusercontent.com/8693064/87086242-5cdf7580-c242-11ea-9557-2f605e4209c2.png) The yellow and green uptime are when I run netdata with the new exporting config The red and blue uptime are when I run netdata with the old backend config ##### Steps To Reproduce 1. Run netdata with new backend and old backend and send to graphite Here is my `[backend]` config: ``` [backend] enabled = yes type = graphite destination = docker.for.mac.localhost:2023 data source = as collected prefix = netdata update every = 10 buffer on failures = 10 timeout ms = 20000 send charts matching = system.uptime* system.cpu* system.ram* system.net* disk_space.* !* send hosts matching = localhost * send names instead of ids = yes ``` Here is my `exporting.conf` file: ``` [graphite:netdata] enabled = yes destination = docker.for.mac.localhost:2023 data source = as collected prefix = netdata update every = 10 buffer on failures = 10 timeout ms = 20000 send charts matching = system.uptime* system.cpu* system.ram* system.net* disk_space.* !* send hosts matching = localhost * send names instead of ids = yes ``` This is the only config which I have changed when testing this behaviour ##### Expected behavior I expect that I do not see gaps in the dashboard when using the new style. It could be that I have configured the new one in a wrong way, in such a case I would like to hear any suggestions.
non_defect
new exporting module is missing metrics compared to old when sending to graphite bug report summary i have migrated the exporting config from the old way using to the new way by using exporting conf however there is something strange or i have put wrong config because we do not get all the metrics os environment running netdata on docker together with grafana latest in docker and graphite also in docker i am running them on a macos host but i have also seen missing stats on our ubuntu servers component name the metrics send to graphite have gaps which did not happen before i have made a test by running a single netdata with the new config and old config the yellow and green uptime are when i run netdata with the new exporting config the red and blue uptime are when i run netdata with the old backend config steps to reproduce run netdata with new backend and old backend and send to graphite here is my config enabled yes type graphite destination docker for mac localhost data source as collected prefix netdata update every buffer on failures timeout ms send charts matching system uptime system cpu system ram system net disk space send hosts matching localhost send names instead of ids yes here is my exporting conf file enabled yes destination docker for mac localhost data source as collected prefix netdata update every buffer on failures timeout ms send charts matching system uptime system cpu system ram system net disk space send hosts matching localhost send names instead of ids yes this is the only config which i have changed when testing this behaviour expected behavior i expect that i do not see gaps in the dashboard when using the new style it could be that i have configured the new one in a wrong way in such a case i would like to hear any suggestions
0
1,518
2,603,966,891
IssuesEvent
2015-02-24 18:59:15
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳冠状沟长肉芽
auto-migrated Priority-Medium Type-Defect
``` 沈阳冠状沟长肉芽〓沈陽軍區政治部醫院性病〓TEL:024-3102330 8〓成立于1946年,68年專注于性傳播疾病的研究和治療。位于� ��陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷� ��悠久、設備精良、技術權威、專家云集,是預防、保健、醫 療、科研康復為一體的綜合性醫院。是國家首批公立甲等部�� �醫院、全國首批醫療規范定點單位,是第四軍醫大學、東南� ��學等知名高等院校的教學醫院。曾被中國人民解放軍空軍后 勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二等�� �。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:07
1.0
沈阳冠状沟长肉芽 - ``` 沈阳冠状沟长肉芽〓沈陽軍區政治部醫院性病〓TEL:024-3102330 8〓成立于1946年,68年專注于性傳播疾病的研究和治療。位于� ��陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷� ��悠久、設備精良、技術權威、專家云集,是預防、保健、醫 療、科研康復為一體的綜合性醫院。是國家首批公立甲等部�� �醫院、全國首批醫療規范定點單位,是第四軍醫大學、東南� ��學等知名高等院校的教學醫院。曾被中國人民解放軍空軍后 勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二等�� �。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:07
defect
沈阳冠状沟长肉芽 沈阳冠状沟长肉芽〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位于� �� 。是一所與新中國同建立共輝煌的歷� ��悠久、設備精良、技術權威、專家云集,是預防、保健、醫 療、科研康復為一體的綜合性醫院。是國家首批公立甲等部�� �醫院、全國首批醫療規范定點單位,是第四軍醫大學、東南� ��學等知名高等院校的教學醫院。曾被中國人民解放軍空軍后 勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二等�� �。 original issue reported on code google com by gmail com on jun at
1
126,854
5,006,746,674
IssuesEvent
2016-12-12 15:01:12
storyteller/Storyteller
https://api.github.com/repos/storyteller/Storyteller
opened
Server should respond to set-breakpoints and remember for later
low priority
Just so you can more easily refresh the screen.
1.0
Server should respond to set-breakpoints and remember for later - Just so you can more easily refresh the screen.
non_defect
server should respond to set breakpoints and remember for later just so you can more easily refresh the screen
0
46,521
13,055,926,240
IssuesEvent
2020-07-30 03:08:28
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
[cmake] toplevel-parasite tool detection is backwards (Trac #1358)
Incomplete Migration Migrated from Trac cmake defect
Migrated from https://code.icecube.wisc.edu/ticket/1358 ```json { "status": "closed", "changetime": "2019-01-12T00:06:17", "description": "The parasitic build system has no explicit tool detection phase. Instead, the parasitic version of `i3_project()` loads tool-related variables from the parent CMakeCache.txt based on the values passed to `USE_TOOLS`. Some projects with uncommon dependencies such as `astro`, however, wrap the body of their CMakeLists.txt in an `if (${TOOL}_FOUND)` clause, which will fail because the corresponding variable is not defined until explicitly requested.\n\nTwo solutions come to mind:\n1. Eagerly load everything from the parent CMakeCache.txt, filtering out tool-related variables by looking for names that share a prefix with variables of the form `${TOOL}_FOUND`. This is icky but simple.\n2. Replace the `if (${TOOL}_FOUND)` pattern with a function call, e.g. `if (tool_found(${TOOL}))`. This is more work and involves touching multiple projects, but is probably clearer in the long run.", "reporter": "jvansanten", "cc": "nega, david.schultz", "resolution": "fixed", "_ts": "1547251577858784", "component": "cmake", "summary": "[cmake] toplevel-parasite tool detection is backwards", "priority": "major", "keywords": "", "time": "2015-09-21T12:36:01", "milestone": "", "owner": "jvansanten", "type": "defect" } ```
1.0
[cmake] toplevel-parasite tool detection is backwards (Trac #1358) - Migrated from https://code.icecube.wisc.edu/ticket/1358 ```json { "status": "closed", "changetime": "2019-01-12T00:06:17", "description": "The parasitic build system has no explicit tool detection phase. Instead, the parasitic version of `i3_project()` loads tool-related variables from the parent CMakeCache.txt based on the values passed to `USE_TOOLS`. Some projects with uncommon dependencies such as `astro`, however, wrap the body of their CMakeLists.txt in an `if (${TOOL}_FOUND)` clause, which will fail because the corresponding variable is not defined until explicitly requested.\n\nTwo solutions come to mind:\n1. Eagerly load everything from the parent CMakeCache.txt, filtering out tool-related variables by looking for names that share a prefix with variables of the form `${TOOL}_FOUND`. This is icky but simple.\n2. Replace the `if (${TOOL}_FOUND)` pattern with a function call, e.g. `if (tool_found(${TOOL}))`. This is more work and involves touching multiple projects, but is probably clearer in the long run.", "reporter": "jvansanten", "cc": "nega, david.schultz", "resolution": "fixed", "_ts": "1547251577858784", "component": "cmake", "summary": "[cmake] toplevel-parasite tool detection is backwards", "priority": "major", "keywords": "", "time": "2015-09-21T12:36:01", "milestone": "", "owner": "jvansanten", "type": "defect" } ```
defect
toplevel parasite tool detection is backwards trac migrated from json status closed changetime description the parasitic build system has no explicit tool detection phase instead the parasitic version of project loads tool related variables from the parent cmakecache txt based on the values passed to use tools some projects with uncommon dependencies such as astro however wrap the body of their cmakelists txt in an if tool found clause which will fail because the corresponding variable is not defined until explicitly requested n ntwo solutions come to mind eagerly load everything from the parent cmakecache txt filtering out tool related variables by looking for names that share a prefix with variables of the form tool found this is icky but simple replace the if tool found pattern with a function call e g if tool found tool this is more work and involves touching multiple projects but is probably clearer in the long run reporter jvansanten cc nega david schultz resolution fixed ts component cmake summary toplevel parasite tool detection is backwards priority major keywords time milestone owner jvansanten type defect
1
44,329
12,101,450,515
IssuesEvent
2020-04-20 15:13:55
codesmithtools/Templates
https://api.github.com/repos/codesmithtools/Templates
closed
possibility to force join tables being entities from the GUI
Framework-NHibernate Type-Defect auto-migrated
``` What steps will reproduce the problem? 1. more details here: http://community.codesmithtools.com/forums/t/10459.aspx What is the expected output? What do you see instead? A table should be generated as an entity instead of being simply not generated at all when the table is a join table linked to another join table. What version of the product are you using? 5.2.0 Please provide any additional information below. Maybe you can add to the csp file the option to specify wich tables should be entities. In the NHibernateMaster.cst output edition mode, we could have a system similar to "excludedTables" called "entityForcedTables" where we could select which tables to force being entities.... which is the same as adding extended properties to tables without needing to modify the database. ``` Original issue reported on code.google.com by `metash...@gmail.com` on 13 Jan 2010 at 4:21
1.0
possibility to force join tables being entities from the GUI - ``` What steps will reproduce the problem? 1. more details here: http://community.codesmithtools.com/forums/t/10459.aspx What is the expected output? What do you see instead? A table should be generated as an entity instead of being simply not generated at all when the table is a join table linked to another join table. What version of the product are you using? 5.2.0 Please provide any additional information below. Maybe you can add to the csp file the option to specify wich tables should be entities. In the NHibernateMaster.cst output edition mode, we could have a system similar to "excludedTables" called "entityForcedTables" where we could select which tables to force being entities.... which is the same as adding extended properties to tables without needing to modify the database. ``` Original issue reported on code.google.com by `metash...@gmail.com` on 13 Jan 2010 at 4:21
defect
possibility to force join tables being entities from the gui what steps will reproduce the problem more details here what is the expected output what do you see instead a table should be generated as an entity instead of being simply not generated at all when the table is a join table linked to another join table what version of the product are you using please provide any additional information below maybe you can add to the csp file the option to specify wich tables should be entities in the nhibernatemaster cst output edition mode we could have a system similar to excludedtables called entityforcedtables where we could select which tables to force being entities which is the same as adding extended properties to tables without needing to modify the database original issue reported on code google com by metash gmail com on jan at
1
22,741
11,726,647,468
IssuesEvent
2020-03-10 14:48:59
golang/go
https://api.github.com/repos/golang/go
closed
cmd/compile: favour UDIV over UMULH + LSR on arm64 for 64 bit integer division by a constant
NeedsInvestigation Performance
### What version of Go are you using (`go version`)? <pre> $ go version go version go1.14 darwin/amd64 # but (cross compiling to linux/arm64) </pre> ### Does this issue reproduce with the latest release? Yes ### What operating system and processor architecture are you using (`go env`)? <details><summary><code>go env</code> Output</summary><br><pre> $ go env GO111MODULE="" GOARCH="amd64" GOBIN="" GOCACHE="/Users/pmoore/Library/Caches/go-build" GOENV="/Users/pmoore/Library/Application Support/go/env" GOEXE="" GOFLAGS="" GOHOSTARCH="amd64" GOHOSTOS="darwin" GOINSECURE="" GONOPROXY="" GONOSUMDB="" GOOS="darwin" GOPATH="/Users/pmoore/.gvm/pkgsets/go1.14/global" GOPRIVATE="" GOPROXY="https://proxy.golang.org,direct" GOROOT="/Users/pmoore/.gvm/gos/go1.14" GOSUMDB="sum.golang.org" GOTMPDIR="" GOTOOLDIR="/Users/pmoore/.gvm/gos/go1.14/pkg/tool/darwin_amd64" GCCGO="gccgo" AR="ar" CC="clang" CXX="clang++" CGO_ENABLED="1" GOMOD="/Users/pmoore/git/go/src/go.mod" CGO_CFLAGS="-g -O2" CGO_CPPFLAGS="" CGO_CXXFLAGS="-g -O2" CGO_FFLAGS="-g -O2" CGO_LDFLAGS="-g -O2" PKG_CONFIG="pkg-config" GOGCCFLAGS="-fPIC -m64 -pthread -fno-caret-diagnostics -Qunused-arguments -fmessage-length=0 -fdebug-prefix-map=/var/folders/v9/mll6p_rj5h94dt_m5m8j0f9c0000gn/T/go-build535314509=/tmp/go-build -gno-record-gcc-switches -fno-common" </pre></details> ### What did you do? ``` pmoore@Petes-iMac:~/udiv $ cat udiv.go package main import ( "fmt" ) func main() { for x := uint64(0); x < 123456; x++ { fmt.Printf("%v\n", x/216) } } pmoore@Petes-iMac:~/udiv $ GOARCH=arm64 GOOS=linux go build -gcflags -S udiv.go 2>&1 | grep -F udiv.go:9 0x0028 00040 (/Users/pmoore/udiv/udiv.go:9) MOVD $-7515340178177965473, R1 0x0038 00056 (/Users/pmoore/udiv/udiv.go:9) UMULH R0, R1, R2 0x003c 00060 (/Users/pmoore/udiv/udiv.go:9) LSR $7, R2, R2 0x0040 00064 (/Users/pmoore/udiv/udiv.go:9) MOVD R2, 8(RSP) 0x0044 00068 (/Users/pmoore/udiv/udiv.go:9) CALL runtime.convT64(SB) 0x0048 00072 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, $1 0x0048 00072 (/Users/pmoore/udiv/udiv.go:9) MOVD 16(RSP), R0 0x004c 00076 (/Users/pmoore/udiv/udiv.go:9) PCDATA $1, $1 0x004c 00076 (/Users/pmoore/udiv/udiv.go:9) STP (ZR, ZR), ""..autotmp_15-16(SP) 0x0050 00080 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, $2 0x0050 00080 (/Users/pmoore/udiv/udiv.go:9) MOVD $type.uint64(SB), R1 0x0058 00088 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, $1 0x0058 00088 (/Users/pmoore/udiv/udiv.go:9) MOVD R1, ""..autotmp_15-16(SP) 0x005c 00092 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, ZR 0x005c 00092 (/Users/pmoore/udiv/udiv.go:9) MOVD R0, ""..autotmp_15-8(SP) pmoore@Petes-iMac:~/udiv $ ``` ### What did you expect to see? I expected the compiler to generate the UDIV instruction rather than `UMULH` + `LSR`, as empirical testing (below) shows it to be twice as fast on an arm64 Cortex A53 (BCM2837) SoC. The final assembly instructions are: ``` pmoore@Petes-iMac:~/udiv $ aarch64-unknown-linux-gnu-objdump -d udiv | sed -n '/\<main\.main\>/,$p' 000000000009bf50 <main.main>: 9bf50: f9400b81 ldr x1, [x28, #16] 9bf54: 910003e2 mov x2, sp 9bf58: eb01005f cmp x2, x1 9bf5c: 54000609 b.ls 9c01c <main.main+0xcc> // b.plast 9bf60: f8180ffe str x30, [sp, #-128]! 9bf64: f81f83fd stur x29, [sp, #-8] 9bf68: d10023fd sub x29, sp, #0x8 9bf6c: d2800000 mov x0, #0x0 // #0 9bf70: 14000024 b 9c000 <main.main+0xb0> 9bf74: f90033e0 str x0, [sp, #96] 9bf78: d2884be1 mov x1, #0x425f // #16991 9bf7c: f2a12f61 movk x1, #0x97b, lsl #16 9bf80: f2c4bda1 movk x1, #0x25ed, lsl #32 9bf84: f2f2f681 movk x1, #0x97b4, lsl #48 9bf88: 9bc07c22 umulh x2, x1, x0 9bf8c: d347fc42 lsr x2, x2, #7 9bf90: f90007e2 str x2, [sp, #8] 9bf94: 97fdf237 bl 18870 <runtime.convT64> 9bf98: f9400be0 ldr x0, [sp, #16] 9bf9c: a906ffff stp xzr, xzr, [sp, #104] 9bfa0: d0000081 adrp x1, ad000 <type.*+0xd000> 9bfa4: 91348021 add x1, x1, #0xd20 9bfa8: f90037e1 str x1, [sp, #104] 9bfac: f9003be0 str x0, [sp, #112] 9bfb0: d00006db adrp x27, 175000 <runtime.itabTableInit+0xee0> 9bfb4: 9106237b add x27, x27, #0x188 9bfb8: f9400360 ldr x0, [x27] 9bfbc: f0000262 adrp x2, ea000 <runtime.vdsoauxv.stkobj> 9bfc0: 91060042 add x2, x2, #0x180 9bfc4: f90007e2 str x2, [sp, #8] 9bfc8: f9000be0 str x0, [sp, #16] 9bfcc: d0000180 adrp x0, cd000 <type.*+0x2d000> 9bfd0: 91300800 add x0, x0, #0xc02 9bfd4: f9000fe0 str x0, [sp, #24] 9bfd8: b24007e3 orr x3, xzr, #0x3 9bfdc: f90013e3 str x3, [sp, #32] 9bfe0: 9101a3e4 add x4, sp, #0x68 9bfe4: f90017e4 str x4, [sp, #40] 9bfe8: b24003e4 orr x4, xzr, #0x1 9bfec: f9001be4 str x4, [sp, #48] 9bff0: f9001fe4 str x4, [sp, #56] 9bff4: 97ffe363 bl 94d80 <fmt.Fprintf> 9bff8: f94033e0 ldr x0, [sp, #96] 9bffc: 91000400 add x0, x0, #0x1 9c000: d29c481b mov x27, #0xe240 // #57920 9c004: f2a0003b movk x27, #0x1, lsl #16 9c008: eb1b001f cmp x0, x27 9c00c: 54fffb43 b.cc 9bf74 <main.main+0x24> // b.lo, b.ul, b.last 9c010: f85f83fd ldur x29, [sp, #-8] 9c014: f84807fe ldr x30, [sp], #128 9c018: d65f03c0 ret 9c01c: aa1e03e3 mov x3, x30 9c020: 97ff2a48 bl 66940 <runtime.morestack_noctxt> 9c024: 17ffffcb b 9bf50 <main.main> ... ``` The division by 216 is comprised of the six instructions: ``` 9bf78: d2884be1 mov x1, #0x425f // #16991 9bf7c: f2a12f61 movk x1, #0x97b, lsl #16 9bf80: f2c4bda1 movk x1, #0x25ed, lsl #32 9bf84: f2f2f681 movk x1, #0x97b4, lsl #48 9bf88: 9bc07c22 umulh x2, x1, x0 9bf8c: d347fc42 lsr x2, x2, #7 ``` ### Benchmark results I created two simple executables to compare the generated `mov/movk/movk/movk/umulh/lsr` instructions with the equivalent `mov/udiv` instructions that I had expected to see, and compared results. Here we see that `mov/udiv` is consistently twice as fast: ``` ubuntu@ubuntu:~/git/udiv$ cat umulh+lsr.s /* Benchmark with: as -o umulh+lsr.o umulh+lsr.s && gcc -o umulh+lsr umulh+lsr.o && for ((i=0; i<5; i++)); do time ./umulh+lsr; done */ .global main .data fmt: .asciz "%d / %d = %d\n" .text main: mov x1, #0xec73 movk x1, #0x0018, lsl #16 mov x2, #216 mov x11, #0x10000000 1: mov x4, #0x425f movk x4, #0x97b, lsl #16 movk x4, #0x25ed, lsl #32 movk x4, #0x97b4, lsl #48 umulh x3, x4, x1 lsr x3, x3, #7 subs x11, x11, #1 b.ne 1b ldr x0,=fmt bl printf mov x8, #93 svc 0 ubuntu@ubuntu:~/git/udiv$ as -o umulh+lsr.o umulh+lsr.s && gcc -o umulh+lsr umulh+lsr.o && for ((i=0; i<5; i++)); do time ./umulh+lsr; done 1633395 / 216 = 7562 real 0m2.720s user 0m2.713s sys 0m0.004s 1633395 / 216 = 7562 real 0m2.713s user 0m2.706s sys 0m0.004s 1633395 / 216 = 7562 real 0m2.713s user 0m2.706s sys 0m0.005s 1633395 / 216 = 7562 real 0m2.713s user 0m2.710s sys 0m0.001s 1633395 / 216 = 7562 real 0m2.713s user 0m2.709s sys 0m0.001s ubuntu@ubuntu:~/git/udiv$ cat udiv.s /* Benchmark with: as -o udiv.o udiv.s && gcc -o udiv udiv.o && for ((i=0; i<5; i++)); do time ./udiv; done */ .global main .data fmt: .asciz "%d / %d = %d\n" .text main: mov x1, #0xec73 movk x1, #0x0018, lsl #16 mov x2, #216 mov x11, #0x10000000 1: mov x4, #216 udiv x3, x1, x4 subs x11, x11, #1 b.ne 1b ldr x0,=fmt bl printf mov x8, #93 svc 0 ubuntu@ubuntu:~/git/udiv$ as -o udiv.o udiv.s && gcc -o udiv udiv.o && for ((i=0; i<5; i++)); do time ./udiv; done 1633395 / 216 = 7562 real 0m1.358s user 0m1.357s sys 0m0.000s 1633395 / 216 = 7562 real 0m1.358s user 0m1.353s sys 0m0.004s 1633395 / 216 = 7562 real 0m1.358s user 0m1.357s sys 0m0.000s 1633395 / 216 = 7562 real 0m1.358s user 0m1.352s sys 0m0.005s 1633395 / 216 = 7562 real 0m1.358s user 0m1.352s sys 0m0.005s ubuntu@ubuntu:~/git/udiv$ ```
True
cmd/compile: favour UDIV over UMULH + LSR on arm64 for 64 bit integer division by a constant - ### What version of Go are you using (`go version`)? <pre> $ go version go version go1.14 darwin/amd64 # but (cross compiling to linux/arm64) </pre> ### Does this issue reproduce with the latest release? Yes ### What operating system and processor architecture are you using (`go env`)? <details><summary><code>go env</code> Output</summary><br><pre> $ go env GO111MODULE="" GOARCH="amd64" GOBIN="" GOCACHE="/Users/pmoore/Library/Caches/go-build" GOENV="/Users/pmoore/Library/Application Support/go/env" GOEXE="" GOFLAGS="" GOHOSTARCH="amd64" GOHOSTOS="darwin" GOINSECURE="" GONOPROXY="" GONOSUMDB="" GOOS="darwin" GOPATH="/Users/pmoore/.gvm/pkgsets/go1.14/global" GOPRIVATE="" GOPROXY="https://proxy.golang.org,direct" GOROOT="/Users/pmoore/.gvm/gos/go1.14" GOSUMDB="sum.golang.org" GOTMPDIR="" GOTOOLDIR="/Users/pmoore/.gvm/gos/go1.14/pkg/tool/darwin_amd64" GCCGO="gccgo" AR="ar" CC="clang" CXX="clang++" CGO_ENABLED="1" GOMOD="/Users/pmoore/git/go/src/go.mod" CGO_CFLAGS="-g -O2" CGO_CPPFLAGS="" CGO_CXXFLAGS="-g -O2" CGO_FFLAGS="-g -O2" CGO_LDFLAGS="-g -O2" PKG_CONFIG="pkg-config" GOGCCFLAGS="-fPIC -m64 -pthread -fno-caret-diagnostics -Qunused-arguments -fmessage-length=0 -fdebug-prefix-map=/var/folders/v9/mll6p_rj5h94dt_m5m8j0f9c0000gn/T/go-build535314509=/tmp/go-build -gno-record-gcc-switches -fno-common" </pre></details> ### What did you do? ``` pmoore@Petes-iMac:~/udiv $ cat udiv.go package main import ( "fmt" ) func main() { for x := uint64(0); x < 123456; x++ { fmt.Printf("%v\n", x/216) } } pmoore@Petes-iMac:~/udiv $ GOARCH=arm64 GOOS=linux go build -gcflags -S udiv.go 2>&1 | grep -F udiv.go:9 0x0028 00040 (/Users/pmoore/udiv/udiv.go:9) MOVD $-7515340178177965473, R1 0x0038 00056 (/Users/pmoore/udiv/udiv.go:9) UMULH R0, R1, R2 0x003c 00060 (/Users/pmoore/udiv/udiv.go:9) LSR $7, R2, R2 0x0040 00064 (/Users/pmoore/udiv/udiv.go:9) MOVD R2, 8(RSP) 0x0044 00068 (/Users/pmoore/udiv/udiv.go:9) CALL runtime.convT64(SB) 0x0048 00072 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, $1 0x0048 00072 (/Users/pmoore/udiv/udiv.go:9) MOVD 16(RSP), R0 0x004c 00076 (/Users/pmoore/udiv/udiv.go:9) PCDATA $1, $1 0x004c 00076 (/Users/pmoore/udiv/udiv.go:9) STP (ZR, ZR), ""..autotmp_15-16(SP) 0x0050 00080 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, $2 0x0050 00080 (/Users/pmoore/udiv/udiv.go:9) MOVD $type.uint64(SB), R1 0x0058 00088 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, $1 0x0058 00088 (/Users/pmoore/udiv/udiv.go:9) MOVD R1, ""..autotmp_15-16(SP) 0x005c 00092 (/Users/pmoore/udiv/udiv.go:9) PCDATA ZR, ZR 0x005c 00092 (/Users/pmoore/udiv/udiv.go:9) MOVD R0, ""..autotmp_15-8(SP) pmoore@Petes-iMac:~/udiv $ ``` ### What did you expect to see? I expected the compiler to generate the UDIV instruction rather than `UMULH` + `LSR`, as empirical testing (below) shows it to be twice as fast on an arm64 Cortex A53 (BCM2837) SoC. The final assembly instructions are: ``` pmoore@Petes-iMac:~/udiv $ aarch64-unknown-linux-gnu-objdump -d udiv | sed -n '/\<main\.main\>/,$p' 000000000009bf50 <main.main>: 9bf50: f9400b81 ldr x1, [x28, #16] 9bf54: 910003e2 mov x2, sp 9bf58: eb01005f cmp x2, x1 9bf5c: 54000609 b.ls 9c01c <main.main+0xcc> // b.plast 9bf60: f8180ffe str x30, [sp, #-128]! 9bf64: f81f83fd stur x29, [sp, #-8] 9bf68: d10023fd sub x29, sp, #0x8 9bf6c: d2800000 mov x0, #0x0 // #0 9bf70: 14000024 b 9c000 <main.main+0xb0> 9bf74: f90033e0 str x0, [sp, #96] 9bf78: d2884be1 mov x1, #0x425f // #16991 9bf7c: f2a12f61 movk x1, #0x97b, lsl #16 9bf80: f2c4bda1 movk x1, #0x25ed, lsl #32 9bf84: f2f2f681 movk x1, #0x97b4, lsl #48 9bf88: 9bc07c22 umulh x2, x1, x0 9bf8c: d347fc42 lsr x2, x2, #7 9bf90: f90007e2 str x2, [sp, #8] 9bf94: 97fdf237 bl 18870 <runtime.convT64> 9bf98: f9400be0 ldr x0, [sp, #16] 9bf9c: a906ffff stp xzr, xzr, [sp, #104] 9bfa0: d0000081 adrp x1, ad000 <type.*+0xd000> 9bfa4: 91348021 add x1, x1, #0xd20 9bfa8: f90037e1 str x1, [sp, #104] 9bfac: f9003be0 str x0, [sp, #112] 9bfb0: d00006db adrp x27, 175000 <runtime.itabTableInit+0xee0> 9bfb4: 9106237b add x27, x27, #0x188 9bfb8: f9400360 ldr x0, [x27] 9bfbc: f0000262 adrp x2, ea000 <runtime.vdsoauxv.stkobj> 9bfc0: 91060042 add x2, x2, #0x180 9bfc4: f90007e2 str x2, [sp, #8] 9bfc8: f9000be0 str x0, [sp, #16] 9bfcc: d0000180 adrp x0, cd000 <type.*+0x2d000> 9bfd0: 91300800 add x0, x0, #0xc02 9bfd4: f9000fe0 str x0, [sp, #24] 9bfd8: b24007e3 orr x3, xzr, #0x3 9bfdc: f90013e3 str x3, [sp, #32] 9bfe0: 9101a3e4 add x4, sp, #0x68 9bfe4: f90017e4 str x4, [sp, #40] 9bfe8: b24003e4 orr x4, xzr, #0x1 9bfec: f9001be4 str x4, [sp, #48] 9bff0: f9001fe4 str x4, [sp, #56] 9bff4: 97ffe363 bl 94d80 <fmt.Fprintf> 9bff8: f94033e0 ldr x0, [sp, #96] 9bffc: 91000400 add x0, x0, #0x1 9c000: d29c481b mov x27, #0xe240 // #57920 9c004: f2a0003b movk x27, #0x1, lsl #16 9c008: eb1b001f cmp x0, x27 9c00c: 54fffb43 b.cc 9bf74 <main.main+0x24> // b.lo, b.ul, b.last 9c010: f85f83fd ldur x29, [sp, #-8] 9c014: f84807fe ldr x30, [sp], #128 9c018: d65f03c0 ret 9c01c: aa1e03e3 mov x3, x30 9c020: 97ff2a48 bl 66940 <runtime.morestack_noctxt> 9c024: 17ffffcb b 9bf50 <main.main> ... ``` The division by 216 is comprised of the six instructions: ``` 9bf78: d2884be1 mov x1, #0x425f // #16991 9bf7c: f2a12f61 movk x1, #0x97b, lsl #16 9bf80: f2c4bda1 movk x1, #0x25ed, lsl #32 9bf84: f2f2f681 movk x1, #0x97b4, lsl #48 9bf88: 9bc07c22 umulh x2, x1, x0 9bf8c: d347fc42 lsr x2, x2, #7 ``` ### Benchmark results I created two simple executables to compare the generated `mov/movk/movk/movk/umulh/lsr` instructions with the equivalent `mov/udiv` instructions that I had expected to see, and compared results. Here we see that `mov/udiv` is consistently twice as fast: ``` ubuntu@ubuntu:~/git/udiv$ cat umulh+lsr.s /* Benchmark with: as -o umulh+lsr.o umulh+lsr.s && gcc -o umulh+lsr umulh+lsr.o && for ((i=0; i<5; i++)); do time ./umulh+lsr; done */ .global main .data fmt: .asciz "%d / %d = %d\n" .text main: mov x1, #0xec73 movk x1, #0x0018, lsl #16 mov x2, #216 mov x11, #0x10000000 1: mov x4, #0x425f movk x4, #0x97b, lsl #16 movk x4, #0x25ed, lsl #32 movk x4, #0x97b4, lsl #48 umulh x3, x4, x1 lsr x3, x3, #7 subs x11, x11, #1 b.ne 1b ldr x0,=fmt bl printf mov x8, #93 svc 0 ubuntu@ubuntu:~/git/udiv$ as -o umulh+lsr.o umulh+lsr.s && gcc -o umulh+lsr umulh+lsr.o && for ((i=0; i<5; i++)); do time ./umulh+lsr; done 1633395 / 216 = 7562 real 0m2.720s user 0m2.713s sys 0m0.004s 1633395 / 216 = 7562 real 0m2.713s user 0m2.706s sys 0m0.004s 1633395 / 216 = 7562 real 0m2.713s user 0m2.706s sys 0m0.005s 1633395 / 216 = 7562 real 0m2.713s user 0m2.710s sys 0m0.001s 1633395 / 216 = 7562 real 0m2.713s user 0m2.709s sys 0m0.001s ubuntu@ubuntu:~/git/udiv$ cat udiv.s /* Benchmark with: as -o udiv.o udiv.s && gcc -o udiv udiv.o && for ((i=0; i<5; i++)); do time ./udiv; done */ .global main .data fmt: .asciz "%d / %d = %d\n" .text main: mov x1, #0xec73 movk x1, #0x0018, lsl #16 mov x2, #216 mov x11, #0x10000000 1: mov x4, #216 udiv x3, x1, x4 subs x11, x11, #1 b.ne 1b ldr x0,=fmt bl printf mov x8, #93 svc 0 ubuntu@ubuntu:~/git/udiv$ as -o udiv.o udiv.s && gcc -o udiv udiv.o && for ((i=0; i<5; i++)); do time ./udiv; done 1633395 / 216 = 7562 real 0m1.358s user 0m1.357s sys 0m0.000s 1633395 / 216 = 7562 real 0m1.358s user 0m1.353s sys 0m0.004s 1633395 / 216 = 7562 real 0m1.358s user 0m1.357s sys 0m0.000s 1633395 / 216 = 7562 real 0m1.358s user 0m1.352s sys 0m0.005s 1633395 / 216 = 7562 real 0m1.358s user 0m1.352s sys 0m0.005s ubuntu@ubuntu:~/git/udiv$ ```
non_defect
cmd compile favour udiv over umulh lsr on for bit integer division by a constant what version of go are you using go version go version go version darwin but cross compiling to linux does this issue reproduce with the latest release yes what operating system and processor architecture are you using go env go env output go env goarch gobin gocache users pmoore library caches go build goenv users pmoore library application support go env goexe goflags gohostarch gohostos darwin goinsecure gonoproxy gonosumdb goos darwin gopath users pmoore gvm pkgsets global goprivate goproxy goroot users pmoore gvm gos gosumdb sum golang org gotmpdir gotooldir users pmoore gvm gos pkg tool darwin gccgo gccgo ar ar cc clang cxx clang cgo enabled gomod users pmoore git go src go mod cgo cflags g cgo cppflags cgo cxxflags g cgo fflags g cgo ldflags g pkg config pkg config gogccflags fpic pthread fno caret diagnostics qunused arguments fmessage length fdebug prefix map var folders t go tmp go build gno record gcc switches fno common what did you do pmoore petes imac udiv cat udiv go package main import fmt func main for x x x fmt printf v n x pmoore petes imac udiv goarch goos linux go build gcflags s udiv go grep f udiv go users pmoore udiv udiv go movd users pmoore udiv udiv go umulh users pmoore udiv udiv go lsr users pmoore udiv udiv go movd rsp users pmoore udiv udiv go call runtime sb users pmoore udiv udiv go pcdata zr users pmoore udiv udiv go movd rsp users pmoore udiv udiv go pcdata users pmoore udiv udiv go stp zr zr autotmp sp users pmoore udiv udiv go pcdata zr users pmoore udiv udiv go movd type sb users pmoore udiv udiv go pcdata zr users pmoore udiv udiv go movd autotmp sp users pmoore udiv udiv go pcdata zr zr users pmoore udiv udiv go movd autotmp sp pmoore petes imac udiv what did you expect to see i expected the compiler to generate the udiv instruction rather than umulh lsr as empirical testing below shows it to be twice as fast on an cortex soc the final assembly instructions are pmoore petes imac udiv unknown linux gnu objdump d udiv sed n p ldr mov sp cmp b ls b plast str stur sub sp mov b str mov movk lsl movk lsl movk lsl umulh lsr str bl ldr stp xzr xzr adrp add str str adrp add ldr adrp add str str adrp add str orr xzr str add sp str orr xzr str str bl ldr add mov movk lsl cmp b cc b lo b ul b last ldur ldr ret mov bl b the division by is comprised of the six instructions mov movk lsl movk lsl movk lsl umulh lsr benchmark results i created two simple executables to compare the generated mov movk movk movk umulh lsr instructions with the equivalent mov udiv instructions that i had expected to see and compared results here we see that mov udiv is consistently twice as fast ubuntu ubuntu git udiv cat umulh lsr s benchmark with as o umulh lsr o umulh lsr s gcc o umulh lsr umulh lsr o for i i i do time umulh lsr done global main data fmt asciz d d d n text main mov movk lsl mov mov mov movk lsl movk lsl movk lsl umulh lsr subs b ne ldr fmt bl printf mov svc ubuntu ubuntu git udiv as o umulh lsr o umulh lsr s gcc o umulh lsr umulh lsr o for i i i do time umulh lsr done real user sys real user sys real user sys real user sys real user sys ubuntu ubuntu git udiv cat udiv s benchmark with as o udiv o udiv s gcc o udiv udiv o for i i i do time udiv done global main data fmt asciz d d d n text main mov movk lsl mov mov mov udiv subs b ne ldr fmt bl printf mov svc ubuntu ubuntu git udiv as o udiv o udiv s gcc o udiv udiv o for i i i do time udiv done real user sys real user sys real user sys real user sys real user sys ubuntu ubuntu git udiv
0
69,438
22,352,446,612
IssuesEvent
2022-06-15 13:12:15
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
Cannot parse quoted database link
T: Defect P: Medium E: Professional Edition E: Enterprise Edition C: Parser
This parses fine: ```sql select c from t@link ``` So does this, though there's some weird extra whitespace in the translation output: ```sql -- input select c from "t"@link -- output select C from "t" @LINK ``` But this doesn't parse: ```sql select c from "t"@"link" ``` The output is: > Unsupported query type: [1:19] select c from "t"@[*]"link"
1.0
Cannot parse quoted database link - This parses fine: ```sql select c from t@link ``` So does this, though there's some weird extra whitespace in the translation output: ```sql -- input select c from "t"@link -- output select C from "t" @LINK ``` But this doesn't parse: ```sql select c from "t"@"link" ``` The output is: > Unsupported query type: [1:19] select c from "t"@[*]"link"
defect
cannot parse quoted database link this parses fine sql select c from t link so does this though there s some weird extra whitespace in the translation output sql input select c from t link output select c from t link but this doesn t parse sql select c from t link the output is unsupported query type select c from t link
1
25,179
4,232,038,341
IssuesEvent
2016-07-04 19:39:46
scipy/scipy
https://api.github.com/repos/scipy/scipy
reopened
slsqp fails to detect infeasible problem
defect scipy.optimize
From mailing list: ``` x = [1,2] optimize.minimize( lambda x: x[0]**2+x[1]**2, x, constraints = ( {'type':'eq','fun': lambda x: x[0]+x[1]-1}, {'type':'ineq','fun': lambda x: x[0]-2} ), bounds = ((0,None),(0,None)), method='SLSQP') ``` returns (0.18.0.dev0+1eeef28) ``` fun: 2.77777777777712 jac: array([ 3.33333334e+00, 2.98023224e-08, 0.00000000e+00]) message: 'Optimization terminated successfully.' nfev: 8 nit: 6 njev: 2 status: 0 success: True x: array([ 1.66666667e+00, 1.39888101e-14]) ```
1.0
slsqp fails to detect infeasible problem - From mailing list: ``` x = [1,2] optimize.minimize( lambda x: x[0]**2+x[1]**2, x, constraints = ( {'type':'eq','fun': lambda x: x[0]+x[1]-1}, {'type':'ineq','fun': lambda x: x[0]-2} ), bounds = ((0,None),(0,None)), method='SLSQP') ``` returns (0.18.0.dev0+1eeef28) ``` fun: 2.77777777777712 jac: array([ 3.33333334e+00, 2.98023224e-08, 0.00000000e+00]) message: 'Optimization terminated successfully.' nfev: 8 nit: 6 njev: 2 status: 0 success: True x: array([ 1.66666667e+00, 1.39888101e-14]) ```
defect
slsqp fails to detect infeasible problem from mailing list x optimize minimize lambda x x x x constraints type eq fun lambda x x x type ineq fun lambda x x bounds none none method slsqp returns fun jac array message optimization terminated successfully nfev nit njev status success true x array
1
470,630
13,541,873,705
IssuesEvent
2020-09-16 16:28:38
hotosm/tasking-manager
https://api.github.com/repos/hotosm/tasking-manager
opened
Define workflows for final data check
Component: Frontend Priority: High Status: Needs implementation Type: Enhancement
Validators and project managers generally needs to do a final consistency check to evaluate the consistence of the mapping and validation around a batch of tasks, so we need to define some workflows for that.
1.0
Define workflows for final data check - Validators and project managers generally needs to do a final consistency check to evaluate the consistence of the mapping and validation around a batch of tasks, so we need to define some workflows for that.
non_defect
define workflows for final data check validators and project managers generally needs to do a final consistency check to evaluate the consistence of the mapping and validation around a batch of tasks so we need to define some workflows for that
0
51,095
13,188,110,710
IssuesEvent
2020-08-13 05:35:02
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
[icetray] Memory tracking blows the stack on exit (Trac #1973)
Migrated from Trac combo core defect
If I run this script with memory tracking enabled on SL6 with devtoolset-2: ```text from icecube import icetray, dataclasses, dataio from I3Tray import I3Tray tray = I3Tray() tray.Add("I3InfiniteSource") t = icetray.memory.MemoryTimeline() t.start(10000) tray.Execute(1) t.stop tray.Finish() ``` it runs to completion, and I can do things like plot the memory timeline. Then, after the interpreter exits, it segfaults: ```text https://code.icecube.wisc.edu/ticket/0 0x00007ffff6e38b80 in _int_malloc () from /lib64/libc.so.6 #1 0x00007ffff6e39b1c in malloc () from /lib64/libc.so.6 #2 0x00007ffff048d3c4 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #3 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #4 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6 #5 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6 #6 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6 #7 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #8 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #9 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #10 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6 #11 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6 #12 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6 #13 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #14 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #15 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #16 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6 ``` and so on for 248174 frames. I suspect that global objects may be getting destroyed in an unexpected order. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1973">https://code.icecube.wisc.edu/ticket/1973</a>, reported by jvansanten and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:14:44", "description": "If I run this script with memory tracking enabled on SL6 with devtoolset-2:\n{{{\nfrom icecube import icetray, dataclasses, dataio\nfrom I3Tray import I3Tray\n\ntray = I3Tray()\n\ntray.Add(\"I3InfiniteSource\")\n\nt = icetray.memory.MemoryTimeline()\nt.start(10000)\ntray.Execute(1)\nt.stop\n\ntray.Finish()\n\n}}}\n\nit runs to completion, and I can do things like plot the memory timeline. Then, after the interpreter exits, it segfaults:\n\n{{{\n#0 0x00007ffff6e38b80 in _int_malloc () from /lib64/libc.so.6\n#1 0x00007ffff6e39b1c in malloc () from /lib64/libc.so.6\n#2 0x00007ffff048d3c4 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#3 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#4 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()\n from /usr/lib64/libstdc++.so.6\n#5 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6\n#6 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6\n#7 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#8 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#9 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#10 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()\n from /usr/lib64/libstdc++.so.6\n#11 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6\n#12 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6\n#13 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#14 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#15 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#16 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()\n from /usr/lib64/libstdc++.so.6\n}}}\nand so on for 248174 frames.\n\nI suspect that global objects may be getting destroyed in an unexpected order.", "reporter": "jvansanten", "cc": "cweaver, david.schultz", "resolution": "fixed", "_ts": "1550067284370534", "component": "combo core", "summary": "[icetray] Memory tracking blows the stack on exit", "priority": "normal", "keywords": "", "time": "2017-03-23T10:04:27", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
[icetray] Memory tracking blows the stack on exit (Trac #1973) - If I run this script with memory tracking enabled on SL6 with devtoolset-2: ```text from icecube import icetray, dataclasses, dataio from I3Tray import I3Tray tray = I3Tray() tray.Add("I3InfiniteSource") t = icetray.memory.MemoryTimeline() t.start(10000) tray.Execute(1) t.stop tray.Finish() ``` it runs to completion, and I can do things like plot the memory timeline. Then, after the interpreter exits, it segfaults: ```text https://code.icecube.wisc.edu/ticket/0 0x00007ffff6e38b80 in _int_malloc () from /lib64/libc.so.6 #1 0x00007ffff6e39b1c in malloc () from /lib64/libc.so.6 #2 0x00007ffff048d3c4 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #3 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #4 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6 #5 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6 #6 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6 #7 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #8 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #9 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #10 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6 #11 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6 #12 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6 #13 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #14 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #15 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so #16 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6 ``` and so on for 248174 frames. I suspect that global objects may be getting destroyed in an unexpected order. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1973">https://code.icecube.wisc.edu/ticket/1973</a>, reported by jvansanten and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:14:44", "description": "If I run this script with memory tracking enabled on SL6 with devtoolset-2:\n{{{\nfrom icecube import icetray, dataclasses, dataio\nfrom I3Tray import I3Tray\n\ntray = I3Tray()\n\ntray.Add(\"I3InfiniteSource\")\n\nt = icetray.memory.MemoryTimeline()\nt.start(10000)\ntray.Execute(1)\nt.stop\n\ntray.Finish()\n\n}}}\n\nit runs to completion, and I can do things like plot the memory timeline. Then, after the interpreter exits, it segfaults:\n\n{{{\n#0 0x00007ffff6e38b80 in _int_malloc () from /lib64/libc.so.6\n#1 0x00007ffff6e39b1c in malloc () from /lib64/libc.so.6\n#2 0x00007ffff048d3c4 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#3 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#4 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()\n from /usr/lib64/libstdc++.so.6\n#5 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6\n#6 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6\n#7 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#8 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#9 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#10 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()\n from /usr/lib64/libstdc++.so.6\n#11 0x00007fffee530de5 in ?? () from /usr/lib64/libstdc++.so.6\n#12 0x00007fffee530f33 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, std::allocator<char> const&) () from /usr/lib64/libstdc++.so.6\n#13 0x00007ffff048bcc2 in boost::thread_exception::thread_exception(int, char const*) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#14 0x00007ffff048d4b5 in memory::malloc_override(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#15 0x00007ffff048d509 in operator new(unsigned long) () from /scratch/jvs/metaprojects/simulation/build/lib/libicetray.so\n#16 0x00007fffee5303c9 in std::basic_string<char, std::char_traits<char>, std::allocator<char> >::_Rep::_S_create(unsigned long, unsigned long, std::allocator<char> const&) ()\n from /usr/lib64/libstdc++.so.6\n}}}\nand so on for 248174 frames.\n\nI suspect that global objects may be getting destroyed in an unexpected order.", "reporter": "jvansanten", "cc": "cweaver, david.schultz", "resolution": "fixed", "_ts": "1550067284370534", "component": "combo core", "summary": "[icetray] Memory tracking blows the stack on exit", "priority": "normal", "keywords": "", "time": "2017-03-23T10:04:27", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
memory tracking blows the stack on exit trac if i run this script with memory tracking enabled on with devtoolset text from icecube import icetray dataclasses dataio from import tray tray add t icetray memory memorytimeline t start tray execute t stop tray finish it runs to completion and i can do things like plot the memory timeline then after the interpreter exits it segfaults text in int malloc from libc so in malloc from libc so in memory malloc override unsigned long from scratch jvs metaprojects simulation build lib libicetray so in operator new unsigned long from scratch jvs metaprojects simulation build lib libicetray so in std basic string std allocator rep s create unsigned long unsigned long std allocator const from usr libstdc so in from usr libstdc so in std basic string std allocator basic string char const std allocator const from usr libstdc so in boost thread exception thread exception int char const from scratch jvs metaprojects simulation build lib libicetray so in memory malloc override unsigned long from scratch jvs metaprojects simulation build lib libicetray so in operator new unsigned long from scratch jvs metaprojects simulation build lib libicetray so in std basic string std allocator rep s create unsigned long unsigned long std allocator const from usr libstdc so in from usr libstdc so in std basic string std allocator basic string char const std allocator const from usr libstdc so in boost thread exception thread exception int char const from scratch jvs metaprojects simulation build lib libicetray so in memory malloc override unsigned long from scratch jvs metaprojects simulation build lib libicetray so in operator new unsigned long from scratch jvs metaprojects simulation build lib libicetray so in std basic string std allocator rep s create unsigned long unsigned long std allocator const from usr libstdc so and so on for frames i suspect that global objects may be getting destroyed in an unexpected order migrated from json status closed changetime description if i run this script with memory tracking enabled on with devtoolset n nfrom icecube import icetray dataclasses dataio nfrom import n ntray n ntray add n nt icetray memory memorytimeline nt start ntray execute nt stop n ntray finish n n n nit runs to completion and i can do things like plot the memory timeline then after the interpreter exits it segfaults n n n in int malloc from libc so n in malloc from libc so n in memory malloc override unsigned long from scratch jvs metaprojects simulation build lib libicetray so n in operator new unsigned long from scratch jvs metaprojects simulation build lib libicetray so n in std basic string std allocator rep s create unsigned long unsigned long std allocator const n from usr libstdc so n in from usr libstdc so n in std basic string std allocator basic string char const std allocator const from usr libstdc so n in boost thread exception thread exception int char const from scratch jvs metaprojects simulation build lib libicetray so n in memory malloc override unsigned long from scratch jvs metaprojects simulation build lib libicetray so n in operator new unsigned long from scratch jvs metaprojects simulation build lib libicetray so n in std basic string std allocator rep s create unsigned long unsigned long std allocator const n from usr libstdc so n in from usr libstdc so n in std basic string std allocator basic string char const std allocator const from usr libstdc so n in boost thread exception thread exception int char const from scratch jvs metaprojects simulation build lib libicetray so n in memory malloc override unsigned long from scratch jvs metaprojects simulation build lib libicetray so n in operator new unsigned long from scratch jvs metaprojects simulation build lib libicetray so n in std basic string std allocator rep s create unsigned long unsigned long std allocator const n from usr libstdc so n nand so on for frames n ni suspect that global objects may be getting destroyed in an unexpected order reporter jvansanten cc cweaver david schultz resolution fixed ts component combo core summary memory tracking blows the stack on exit priority normal keywords time milestone owner nega type defect
1
6,796
2,610,279,869
IssuesEvent
2015-02-26 19:29:29
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
Import Not Working
auto-migrated Priority-Medium Type-Defect
``` What's the problem? Can't import blogs and settings from earlier export. There is a button to browse to the file but appears to be nothing to begin the import other than an "X"button What browser are you using? Firefox What version of ScribeFire are you running? 1.7.1 ``` ----- Original issue reported on code.google.com by `stme...@gmail.com` on 26 Jun 2011 at 3:07 Attachments: * [scribefire.tiff](https://storage.googleapis.com/google-code-attachments/scribefire-chrome/issue-371/comment-0/scribefire.tiff)
1.0
Import Not Working - ``` What's the problem? Can't import blogs and settings from earlier export. There is a button to browse to the file but appears to be nothing to begin the import other than an "X"button What browser are you using? Firefox What version of ScribeFire are you running? 1.7.1 ``` ----- Original issue reported on code.google.com by `stme...@gmail.com` on 26 Jun 2011 at 3:07 Attachments: * [scribefire.tiff](https://storage.googleapis.com/google-code-attachments/scribefire-chrome/issue-371/comment-0/scribefire.tiff)
defect
import not working what s the problem can t import blogs and settings from earlier export there is a button to browse to the file but appears to be nothing to begin the import other than an x button what browser are you using firefox what version of scribefire are you running original issue reported on code google com by stme gmail com on jun at attachments
1
19,362
3,194,962,709
IssuesEvent
2015-09-30 14:40:35
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
Improper Restriction of XML External Entity References ('XXE') in XMLasDOMBinding
C: Functionality P: Medium T: Defect
`XMLasDOMBinding` exposes a vulnerability: **Attack Vector**: javax.xml.parsers.DocumentBuilder.parse **Description**: The resolution of external entity references is enabled. If attacker-controlled XML can be submitted to the XML parser here, then the attacker could gain access to information about an internal network, local filesystem, or other sensitive data. This is known as an XML eXternal Entity (XXE) attack. **Remediation**: Configure the XML parser to disable external entity resolution. **The fix**: The fix for this issue is as outlined on the OWASP site at this URL: https://www.owasp.org/index.php/XML_External_Entity_%28XXE%29_Processing It is recommended adding all of the fixes mentioned, with the resulting code looking similar to the following: ``` DocumentBuilderFactory docFactory = DocumentBuilderFactory.newInstance(); String feature = EMPTY; try { // BEGIN FIXES feature = "http://apache.org/xml/features/disallow-doctype-decl"; docFactory.setFeature(feature, true); feature = "http://xml.org/sax/features/external-general-entities"; docFactory.setFeature(feature, false); docFactory.setXIncludeAware(false); docFactory.setExpandEntityReferences(false); feature = "http://xml.org/sax/features/external-parameter-entities"; docFactory.setFeature(feature, false); DocumentBuilder docBuilder = docFactory.newDocumentBuilder(); // END FIXES ... InputStream is = new ByteArrayInputStream(xml.getBytes()); document_ = docBuilder.parse(is); ... ```
1.0
Improper Restriction of XML External Entity References ('XXE') in XMLasDOMBinding - `XMLasDOMBinding` exposes a vulnerability: **Attack Vector**: javax.xml.parsers.DocumentBuilder.parse **Description**: The resolution of external entity references is enabled. If attacker-controlled XML can be submitted to the XML parser here, then the attacker could gain access to information about an internal network, local filesystem, or other sensitive data. This is known as an XML eXternal Entity (XXE) attack. **Remediation**: Configure the XML parser to disable external entity resolution. **The fix**: The fix for this issue is as outlined on the OWASP site at this URL: https://www.owasp.org/index.php/XML_External_Entity_%28XXE%29_Processing It is recommended adding all of the fixes mentioned, with the resulting code looking similar to the following: ``` DocumentBuilderFactory docFactory = DocumentBuilderFactory.newInstance(); String feature = EMPTY; try { // BEGIN FIXES feature = "http://apache.org/xml/features/disallow-doctype-decl"; docFactory.setFeature(feature, true); feature = "http://xml.org/sax/features/external-general-entities"; docFactory.setFeature(feature, false); docFactory.setXIncludeAware(false); docFactory.setExpandEntityReferences(false); feature = "http://xml.org/sax/features/external-parameter-entities"; docFactory.setFeature(feature, false); DocumentBuilder docBuilder = docFactory.newDocumentBuilder(); // END FIXES ... InputStream is = new ByteArrayInputStream(xml.getBytes()); document_ = docBuilder.parse(is); ... ```
defect
improper restriction of xml external entity references xxe in xmlasdombinding xmlasdombinding exposes a vulnerability attack vector javax xml parsers documentbuilder parse description the resolution of external entity references is enabled if attacker controlled xml can be submitted to the xml parser here then the attacker could gain access to information about an internal network local filesystem or other sensitive data this is known as an xml external entity xxe attack remediation configure the xml parser to disable external entity resolution the fix the fix for this issue is as outlined on the owasp site at this url it is recommended adding all of the fixes mentioned with the resulting code looking similar to the following documentbuilderfactory docfactory documentbuilderfactory newinstance string feature empty try begin fixes feature docfactory setfeature feature true feature docfactory setfeature feature false docfactory setxincludeaware false docfactory setexpandentityreferences false feature docfactory setfeature feature false documentbuilder docbuilder docfactory newdocumentbuilder end fixes inputstream is new bytearrayinputstream xml getbytes document docbuilder parse is
1
90,469
3,816,033,933
IssuesEvent
2016-03-28 20:07:32
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
opened
Add missing description of panels and toolbar
Low Priority Text
Adding a text under http://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html#panels-and-toolbars section + link from another part toward this chapter.
1.0
Add missing description of panels and toolbar - Adding a text under http://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html#panels-and-toolbars section + link from another part toward this chapter.
non_defect
add missing description of panels and toolbar adding a text under section link from another part toward this chapter
0
67,629
21,035,403,238
IssuesEvent
2022-03-31 07:20:29
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Input new message is sometimes very slow (blocking for some seconds)
T-Defect
### Steps to reproduce Type a new message. ### Outcome #### What did you expect? The message appears immediately while typing. #### What happened instead? The page blocks and the message appears seconds after typing. The keydown event handler ist blocking for several seconds, caused by getDevicesInRoom(). Please find attached a screenshot of the performance inspector. ![element-input-slow](https://user-images.githubusercontent.com/91094750/160999757-d0d9c360-725d-4c7f-87c9-0ca5632b3455.png) ### Operating system Linux ### Browser information Version 99.0.4844.74 (Official Build) (64-bit) ### URL for webapp _No response_ ### Application version Element Version: 1.10.7 Version von Olm: 3.2.8 ### Homeserver Synapse 1.55.0 ### Will you send logs? No
1.0
Input new message is sometimes very slow (blocking for some seconds) - ### Steps to reproduce Type a new message. ### Outcome #### What did you expect? The message appears immediately while typing. #### What happened instead? The page blocks and the message appears seconds after typing. The keydown event handler ist blocking for several seconds, caused by getDevicesInRoom(). Please find attached a screenshot of the performance inspector. ![element-input-slow](https://user-images.githubusercontent.com/91094750/160999757-d0d9c360-725d-4c7f-87c9-0ca5632b3455.png) ### Operating system Linux ### Browser information Version 99.0.4844.74 (Official Build) (64-bit) ### URL for webapp _No response_ ### Application version Element Version: 1.10.7 Version von Olm: 3.2.8 ### Homeserver Synapse 1.55.0 ### Will you send logs? No
defect
input new message is sometimes very slow blocking for some seconds steps to reproduce type a new message outcome what did you expect the message appears immediately while typing what happened instead the page blocks and the message appears seconds after typing the keydown event handler ist blocking for several seconds caused by getdevicesinroom please find attached a screenshot of the performance inspector operating system linux browser information version official build bit url for webapp no response application version element version version von olm homeserver synapse will you send logs no
1
28,375
12,834,849,913
IssuesEvent
2020-07-07 11:50:27
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
kubectl apply -f hello-world-ingress.yaml --namespace ingress-basic
Pri2 container-service/svc
Hi, isn't the --namespace ingress-basic missing for the apply ingress command? so <== kubectl apply -f hello-world-ingress.yaml ==> kubectl apply -f hello-world-ingress.yaml --namespace ingress-basic Cheers, Andre --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: fd0e7d5a-37e9-07dc-f395-9cb5bb580e45 * Version Independent ID: 7e4faf70-9724-7e7b-832b-1cd99a974920 * Content: [Create an ingress controller - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/ingress-basic) * Content Source: [articles/aks/ingress-basic.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/ingress-basic.md) * Service: **container-service** * GitHub Login: @mlearned * Microsoft Alias: **mlearned**
1.0
kubectl apply -f hello-world-ingress.yaml --namespace ingress-basic - Hi, isn't the --namespace ingress-basic missing for the apply ingress command? so <== kubectl apply -f hello-world-ingress.yaml ==> kubectl apply -f hello-world-ingress.yaml --namespace ingress-basic Cheers, Andre --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: fd0e7d5a-37e9-07dc-f395-9cb5bb580e45 * Version Independent ID: 7e4faf70-9724-7e7b-832b-1cd99a974920 * Content: [Create an ingress controller - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/ingress-basic) * Content Source: [articles/aks/ingress-basic.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/ingress-basic.md) * Service: **container-service** * GitHub Login: @mlearned * Microsoft Alias: **mlearned**
non_defect
kubectl apply f hello world ingress yaml namespace ingress basic hi isn t the namespace ingress basic missing for the apply ingress command so kubectl apply f hello world ingress yaml kubectl apply f hello world ingress yaml namespace ingress basic cheers andre document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login mlearned microsoft alias mlearned
0
581,943
17,348,392,906
IssuesEvent
2021-07-29 04:39:23
DeFiCh/whale
https://api.github.com/repos/DeFiCh/whale
closed
block.controller.ts
area/module-api area/module-database area/whale-api-client kind/feature priority/important-soon triage/accepted
<!-- Please only use this template for submitting enhancement/feature requests --> #### What would you like to be added: Block Controller for blocks. `id` can be block hash or block height. ``` GET v1/:network/blocks GET v1/:network/blocks/:id GET v1/:network/blocks/:id/transactions GET v1/:network/blocks/:id/transactions/:txId GET v1/:network/blocks/:id/transactions/:txId/vins GET v1/:network/blocks/:id/transactions/:txId/vouts ```
1.0
block.controller.ts - <!-- Please only use this template for submitting enhancement/feature requests --> #### What would you like to be added: Block Controller for blocks. `id` can be block hash or block height. ``` GET v1/:network/blocks GET v1/:network/blocks/:id GET v1/:network/blocks/:id/transactions GET v1/:network/blocks/:id/transactions/:txId GET v1/:network/blocks/:id/transactions/:txId/vins GET v1/:network/blocks/:id/transactions/:txId/vouts ```
non_defect
block controller ts what would you like to be added block controller for blocks id can be block hash or block height get network blocks get network blocks id get network blocks id transactions get network blocks id transactions txid get network blocks id transactions txid vins get network blocks id transactions txid vouts
0
33,405
7,110,954,868
IssuesEvent
2018-01-17 12:36:18
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
Context Menu stay when window is resized
6.1.12 defect
Reported By PRO User; > p:contextMenu on p:dataTable stay on same place on responsive design when resize or change orientation of Navigator. Can reproduce it on showcase ! Open contextMenu (right click) and resize window. It should be better to hide or move it when Window size change.
1.0
Context Menu stay when window is resized - Reported By PRO User; > p:contextMenu on p:dataTable stay on same place on responsive design when resize or change orientation of Navigator. Can reproduce it on showcase ! Open contextMenu (right click) and resize window. It should be better to hide or move it when Window size change.
defect
context menu stay when window is resized reported by pro user p contextmenu on p datatable stay on same place on responsive design when resize or change orientation of navigator can reproduce it on showcase open contextmenu right click and resize window it should be better to hide or move it when window size change
1
808,939
30,117,818,010
IssuesEvent
2023-06-30 12:58:46
Three-s-A-Crowd-Games/Gerda
https://api.github.com/repos/Three-s-A-Crowd-Games/Gerda
opened
better damage feedback
enhancement high priority
We need better an more feedback for the player when he receives damage. - red flash at the screen border - screenshake - ... also better poisoned feedback: - more green tint - green flash at screen border - ...
1.0
better damage feedback - We need better an more feedback for the player when he receives damage. - red flash at the screen border - screenshake - ... also better poisoned feedback: - more green tint - green flash at screen border - ...
non_defect
better damage feedback we need better an more feedback for the player when he receives damage red flash at the screen border screenshake also better poisoned feedback more green tint green flash at screen border
0
143,421
19,180,298,271
IssuesEvent
2021-12-04 09:03:50
AlexRogalskiy/android-patterns
https://api.github.com/repos/AlexRogalskiy/android-patterns
opened
CVE-2019-1010266 (Medium) detected in lodash-2.4.2.tgz
security vulnerability
## CVE-2019-1010266 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: android-patterns/package.json</p> <p>Path to vulnerable library: android-patterns/node_modules/dockerfile_lint/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - dockerfile_lint-0.3.4.tgz (Root Library) - :x: **lodash-2.4.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/android-patterns/commit/8d0a10402c32fdfadb3cf3eed011045fb6bdc1fb">8d0a10402c32fdfadb3cf3eed011045fb6bdc1fb</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11. <p>Publish Date: 2019-07-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p> <p>Release Date: 2020-09-30</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-1010266 (Medium) detected in lodash-2.4.2.tgz - ## CVE-2019-1010266 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: android-patterns/package.json</p> <p>Path to vulnerable library: android-patterns/node_modules/dockerfile_lint/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - dockerfile_lint-0.3.4.tgz (Root Library) - :x: **lodash-2.4.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/android-patterns/commit/8d0a10402c32fdfadb3cf3eed011045fb6bdc1fb">8d0a10402c32fdfadb3cf3eed011045fb6bdc1fb</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11. <p>Publish Date: 2019-07-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1010266>CVE-2019-1010266</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p> <p>Release Date: 2020-09-30</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file android patterns package json path to vulnerable library android patterns node modules dockerfile lint node modules lodash package json dependency hierarchy dockerfile lint tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details lodash prior to is affected by cwe uncontrolled resource consumption the impact is denial of service the component is date handler the attack vector is attacker provides very long strings which the library attempts to match using a regular expression the fixed version is publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
72,577
7,302,020,508
IssuesEvent
2018-02-27 08:13:50
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
hack/test-cmd.sh referencing gcr.io/google-containers/perl which cannot be docker pull-ed
area/test-infra lifecycle/rotten sig/testing
[hack/test-cmd.sh](https://github.com/kubernetes/kubernetes/blob/bfc953cd4b0e1e1d0b0f4d341baa5b5537920f09/hack/test-cmd.sh#L142) is referencing `gcr.io/google-containers/perl` which cannot be found atm. @erictune fyi
2.0
hack/test-cmd.sh referencing gcr.io/google-containers/perl which cannot be docker pull-ed - [hack/test-cmd.sh](https://github.com/kubernetes/kubernetes/blob/bfc953cd4b0e1e1d0b0f4d341baa5b5537920f09/hack/test-cmd.sh#L142) is referencing `gcr.io/google-containers/perl` which cannot be found atm. @erictune fyi
non_defect
hack test cmd sh referencing gcr io google containers perl which cannot be docker pull ed is referencing gcr io google containers perl which cannot be found atm erictune fyi
0
61,670
17,023,753,018
IssuesEvent
2021-07-03 03:39:38
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
rpmlint error (no-return-in-nonvoid-function) on building osm2pgsql (r26957)
Component: osm2pgsql Priority: critical Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 10.27am, Monday, 24th October 2011]** Rpmlint blocked to build and deploy osm2pgsql package in openSUSE Build Service with error: I: Program returns random data in a function E: osm2pgsql no-return-in-nonvoid-function middle-pgsql.c:167 E: osm2pgsql no-return-in-nonvoid-function output-pgsql.c:1155
1.0
rpmlint error (no-return-in-nonvoid-function) on building osm2pgsql (r26957) - **[Submitted to the original trac issue database at 10.27am, Monday, 24th October 2011]** Rpmlint blocked to build and deploy osm2pgsql package in openSUSE Build Service with error: I: Program returns random data in a function E: osm2pgsql no-return-in-nonvoid-function middle-pgsql.c:167 E: osm2pgsql no-return-in-nonvoid-function output-pgsql.c:1155
defect
rpmlint error no return in nonvoid function on building rpmlint blocked to build and deploy package in opensuse build service with error i program returns random data in a function e no return in nonvoid function middle pgsql c e no return in nonvoid function output pgsql c
1
107,615
13,490,917,218
IssuesEvent
2020-09-11 15:45:45
JoscaWij/Spacey
https://api.github.com/repos/JoscaWij/Spacey
closed
Plattform graphic
Design User Story
# User Story As a player I want the plattforms to match the space astetic of the game # Description - scalable SVG of a asteroid in comic style # Material First Design: ![image](https://user-images.githubusercontent.com/68224111/92366550-9c68f780-f0f5-11ea-98f8-2ecea9a533db.png) Design-Overview: Design: ![Asteroid-gif](https://user-images.githubusercontent.com/68224111/92402662-377ec300-f130-11ea-979a-bc054eb4c308.gif) # To do's: - [x] create SVG in Photoshop
1.0
Plattform graphic - # User Story As a player I want the plattforms to match the space astetic of the game # Description - scalable SVG of a asteroid in comic style # Material First Design: ![image](https://user-images.githubusercontent.com/68224111/92366550-9c68f780-f0f5-11ea-98f8-2ecea9a533db.png) Design-Overview: Design: ![Asteroid-gif](https://user-images.githubusercontent.com/68224111/92402662-377ec300-f130-11ea-979a-bc054eb4c308.gif) # To do's: - [x] create SVG in Photoshop
non_defect
plattform graphic user story as a player i want the plattforms to match the space astetic of the game description scalable svg of a asteroid in comic style material first design design overview design to do s create svg in photoshop
0