Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
29,521
5,640,841,258
IssuesEvent
2017-04-06 17:20:16
DanwareCreations/DotKEGG
https://api.github.com/repos/DanwareCreations/DotKEGG
closed
Add a "Release Process" Document
documentation enhancement
Add a simple MarkDown file to the root of the repo, explaining the steps to take on a release branch. These should include the steps that were taken during the v0.2.0-alpha release, as well as updating documentation in the S3 Bucket, and updating the latest version numbers on the main README. Then, it will be easy to see what should be included in a future publish script!
1.0
Add a "Release Process" Document - Add a simple MarkDown file to the root of the repo, explaining the steps to take on a release branch. These should include the steps that were taken during the v0.2.0-alpha release, as well as updating documentation in the S3 Bucket, and updating the latest version numbers on the main README. Then, it will be easy to see what should be included in a future publish script!
non_defect
add a release process document add a simple markdown file to the root of the repo explaining the steps to take on a release branch these should include the steps that were taken during the alpha release as well as updating documentation in the bucket and updating the latest version numbers on the main readme then it will be easy to see what should be included in a future publish script
0
53,903
13,262,479,614
IssuesEvent
2020-08-20 21:53:04
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[icetray] svn revision info required (Trac #2284)
Migrated from Trac combo core defect
If cmake can't detect the svn revision, the build fails: ```text In file included from <command-line>:0:0: /cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx: In member function 'I3TrayInfo I3TrayInfoService::GetConfig()': /cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:22: error: 'SVN_REVISION' was not declared in this scope #define SVN_REVISION SVN_REVISION-NOTFOUND ^ /cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION' the_config.svn_revision = SVN_REVISION; ^ /cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:35: error: 'NOTFOUND' was not declared in this scope #define SVN_REVISION SVN_REVISION-NOTFOUND ^ /cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION' the_config.svn_revision = SVN_REVISION; ^ ``` This is blocking the cvmfs build for combo. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2284">https://code.icecube.wisc.edu/projects/icecube/ticket/2284</a>, reported by david.schultzand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-05-16T22:09:18", "_ts": "1558044558386076", "description": "If cmake can't detect the svn revision, the build fails:\n\n{{{\nIn file included from <command-line>:0:0:\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx: In member function 'I3TrayInfo I3TrayInfoService::GetConfig()':\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:22: error: 'SVN_REVISION' was not declared in this scope\n#define SVN_REVISION SVN_REVISION-NOTFOUND\n ^\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION'\n the_config.svn_revision = SVN_REVISION;\n ^\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:35: error: 'NOTFOUND' was not declared in this scope\n#define SVN_REVISION SVN_REVISION-NOTFOUND\n ^\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION'\n the_config.svn_revision = SVN_REVISION;\n ^\n}}}\n\nThis is blocking the cvmfs build for combo.", "reporter": "david.schultz", "cc": "olivas", "resolution": "fixed", "time": "2019-05-16T20:06:50", "component": "combo core", "summary": "[icetray] svn revision info required", "priority": "critical", "keywords": "", "milestone": "Vernal Equinox 2019", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
[icetray] svn revision info required (Trac #2284) - If cmake can't detect the svn revision, the build fails: ```text In file included from <command-line>:0:0: /cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx: In member function 'I3TrayInfo I3TrayInfoService::GetConfig()': /cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:22: error: 'SVN_REVISION' was not declared in this scope #define SVN_REVISION SVN_REVISION-NOTFOUND ^ /cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION' the_config.svn_revision = SVN_REVISION; ^ /cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:35: error: 'NOTFOUND' was not declared in this scope #define SVN_REVISION SVN_REVISION-NOTFOUND ^ /cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION' the_config.svn_revision = SVN_REVISION; ^ ``` This is blocking the cvmfs build for combo. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2284">https://code.icecube.wisc.edu/projects/icecube/ticket/2284</a>, reported by david.schultzand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-05-16T22:09:18", "_ts": "1558044558386076", "description": "If cmake can't detect the svn revision, the build fails:\n\n{{{\nIn file included from <command-line>:0:0:\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx: In member function 'I3TrayInfo I3TrayInfoService::GetConfig()':\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:22: error: 'SVN_REVISION' was not declared in this scope\n#define SVN_REVISION SVN_REVISION-NOTFOUND\n ^\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION'\n the_config.svn_revision = SVN_REVISION;\n ^\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:35: error: 'NOTFOUND' was not declared in this scope\n#define SVN_REVISION SVN_REVISION-NOTFOUND\n ^\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION'\n the_config.svn_revision = SVN_REVISION;\n ^\n}}}\n\nThis is blocking the cvmfs build for combo.", "reporter": "david.schultz", "cc": "olivas", "resolution": "fixed", "time": "2019-05-16T20:06:50", "component": "combo core", "summary": "[icetray] svn revision info required", "priority": "critical", "keywords": "", "milestone": "Vernal Equinox 2019", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
svn revision info required trac if cmake can t detect the svn revision the build fails text in file included from cvmfs icecube opensciencegrid org metaprojects combo icetray private icetray cxx in member function getconfig cvmfs icecube opensciencegrid org rhel metaprojects combo icetray cmakefiles workspace config h error svn revision was not declared in this scope define svn revision svn revision notfound cvmfs icecube opensciencegrid org metaprojects combo icetray private icetray cxx note in expansion of macro svn revision the config svn revision svn revision cvmfs icecube opensciencegrid org rhel metaprojects combo icetray cmakefiles workspace config h error notfound was not declared in this scope define svn revision svn revision notfound cvmfs icecube opensciencegrid org metaprojects combo icetray private icetray cxx note in expansion of macro svn revision the config svn revision svn revision this is blocking the cvmfs build for combo migrated from json status closed changetime ts description if cmake can t detect the svn revision the build fails n n nin file included from n cvmfs icecube opensciencegrid org metaprojects combo icetray private icetray cxx in member function getconfig n cvmfs icecube opensciencegrid org rhel metaprojects combo icetray cmakefiles workspace config h error svn revision was not declared in this scope n define svn revision svn revision notfound n n cvmfs icecube opensciencegrid org metaprojects combo icetray private icetray cxx note in expansion of macro svn revision n the config svn revision svn revision n n cvmfs icecube opensciencegrid org rhel metaprojects combo icetray cmakefiles workspace config h error notfound was not declared in this scope n define svn revision svn revision notfound n n cvmfs icecube opensciencegrid org metaprojects combo icetray private icetray cxx note in expansion of macro svn revision n the config svn revision svn revision n n n nthis is blocking the cvmfs build for combo reporter david schultz cc olivas resolution fixed time component combo core summary svn revision info required priority critical keywords milestone vernal equinox owner nega type defect
1
269,658
23,456,864,875
IssuesEvent
2022-08-16 09:37:49
oracle/helidon
https://api.github.com/repos/oracle/helidon
opened
JPA integration tests should run on every build
jpa/jta testing
This should be fixed for: - helidon-2.x - helidon-3.x - main The `tests/integration/jpa` module is only executed when profile `jpa` is enabled. They do not run in pipeline or during local builds. Currently the tests seem to fail. The test should be changed to be runnable in Java without any changes (e.g. if I do `mvn install -Ptests`, the test should run against an in-memory database. When run with additional profiles, they can be used to validate other databases and/or native image (this should be pipeline specific and done in the appropriate scripts). We should not have code that is not compiled and executed in the repository.
1.0
JPA integration tests should run on every build - This should be fixed for: - helidon-2.x - helidon-3.x - main The `tests/integration/jpa` module is only executed when profile `jpa` is enabled. They do not run in pipeline or during local builds. Currently the tests seem to fail. The test should be changed to be runnable in Java without any changes (e.g. if I do `mvn install -Ptests`, the test should run against an in-memory database. When run with additional profiles, they can be used to validate other databases and/or native image (this should be pipeline specific and done in the appropriate scripts). We should not have code that is not compiled and executed in the repository.
non_defect
jpa integration tests should run on every build this should be fixed for helidon x helidon x main the tests integration jpa module is only executed when profile jpa is enabled they do not run in pipeline or during local builds currently the tests seem to fail the test should be changed to be runnable in java without any changes e g if i do mvn install ptests the test should run against an in memory database when run with additional profiles they can be used to validate other databases and or native image this should be pipeline specific and done in the appropriate scripts we should not have code that is not compiled and executed in the repository
0
66,781
20,672,094,028
IssuesEvent
2022-03-10 04:07:56
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
/rainbow breaks compound emojis
T-Defect S-Minor A-Emoji A-Slash-Commands O-Uncommon
### Steps to reproduce Send `/rainbow 🐻‍❄️` ### Outcome #### What did you expect? ![Screenshot 2022-03-09 at 23-03-22 Element Test room](https://user-images.githubusercontent.com/48614497/157586246-ba6109ad-8b88-4606-9982-93a0166e0f80.png) #### What happened instead? ![broken](https://user-images.githubusercontent.com/48614497/157586348-73fb6a32-a45b-433d-8741-d51bc39ae1c1.png) ### Operating system NixOS unstable ### Browser information Firefox 97.0.1 ### URL for webapp develop.element.io ### Application version Element version: 2d3757c37f20-react-7b71a9fb6543-js-35a375e3d290 Olm version: 3.2.8 ### Homeserver Synapse 1.54.0 ### Will you send logs? No
1.0
/rainbow breaks compound emojis - ### Steps to reproduce Send `/rainbow 🐻‍❄️` ### Outcome #### What did you expect? ![Screenshot 2022-03-09 at 23-03-22 Element Test room](https://user-images.githubusercontent.com/48614497/157586246-ba6109ad-8b88-4606-9982-93a0166e0f80.png) #### What happened instead? ![broken](https://user-images.githubusercontent.com/48614497/157586348-73fb6a32-a45b-433d-8741-d51bc39ae1c1.png) ### Operating system NixOS unstable ### Browser information Firefox 97.0.1 ### URL for webapp develop.element.io ### Application version Element version: 2d3757c37f20-react-7b71a9fb6543-js-35a375e3d290 Olm version: 3.2.8 ### Homeserver Synapse 1.54.0 ### Will you send logs? No
defect
rainbow breaks compound emojis steps to reproduce send rainbow 🐻‍❄️ outcome what did you expect what happened instead operating system nixos unstable browser information firefox url for webapp develop element io application version element version react js olm version homeserver synapse will you send logs no
1
60,305
17,023,392,216
IssuesEvent
2021-07-03 01:47:27
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
The apidb tasks are not documented.
Component: osmosis Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 12.47am, Sunday, 26th April 2009]** To support the new PostgreSQL database schema, a new set of tasks with an apidb prefix have been created which obsolete the existing mysql tasks. These new tasks have not been documented on the osmosis wiki page(s) yet.
1.0
The apidb tasks are not documented. - **[Submitted to the original trac issue database at 12.47am, Sunday, 26th April 2009]** To support the new PostgreSQL database schema, a new set of tasks with an apidb prefix have been created which obsolete the existing mysql tasks. These new tasks have not been documented on the osmosis wiki page(s) yet.
defect
the apidb tasks are not documented to support the new postgresql database schema a new set of tasks with an apidb prefix have been created which obsolete the existing mysql tasks these new tasks have not been documented on the osmosis wiki page s yet
1
625,621
19,758,790,491
IssuesEvent
2022-01-16 03:19:28
input-output-hk/cardano-db-sync
https://api.github.com/repos/input-output-hk/cardano-db-sync
closed
Empty TX inputs/outputs
bug priority high
**OS** Debian **Versions** The `db-sync` version (eg `cardano-db-sync --version`): cardano-db-sync-extended 12.0.0 - linux-x86_64 - ghc-8.10 git revision 9d0180571482ee4c6acb6fbc6bf55b5a4e2ee833 PostgreSQL version: 14.1 **Build/Install Method** The method you use to build or install `cardano-db-sync`: Hydra binaries **Run method** The method you used to run `cardano-db-sync` (eg Nix/Docker/systemd/none): systemd **Additional context** The database restored from snapshot db-sync-snapshot-schema-12-block-6647999-x86_64.tgz **Problem Report** There are no rows for tx f9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234 in the tx_out table in DB Sync 12 ``` SELECT id FROM tx WHERE hash = '\xf9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234'; id ---------- 25318047 (1 row) ``` ``` SELECT id FROM tx_out WHERE tx_id = (SELECT id FROM tx WHERE hash = '\xf9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234'); id ---- (0 rows) ``` DB Sync 11 (cardano-db-sync-extended 11.0.4 - linux-x86_64 - ghc-8.10 git revision b5cd53d2e86f45127d98de429e8a2b6879fd8a56) feels good ``` SELECT id FROM tx_out WHERE tx_id = (SELECT id FROM tx WHERE hash = '\xf9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234'); id ---------- 67617947 (1 row) ```
1.0
Empty TX inputs/outputs - **OS** Debian **Versions** The `db-sync` version (eg `cardano-db-sync --version`): cardano-db-sync-extended 12.0.0 - linux-x86_64 - ghc-8.10 git revision 9d0180571482ee4c6acb6fbc6bf55b5a4e2ee833 PostgreSQL version: 14.1 **Build/Install Method** The method you use to build or install `cardano-db-sync`: Hydra binaries **Run method** The method you used to run `cardano-db-sync` (eg Nix/Docker/systemd/none): systemd **Additional context** The database restored from snapshot db-sync-snapshot-schema-12-block-6647999-x86_64.tgz **Problem Report** There are no rows for tx f9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234 in the tx_out table in DB Sync 12 ``` SELECT id FROM tx WHERE hash = '\xf9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234'; id ---------- 25318047 (1 row) ``` ``` SELECT id FROM tx_out WHERE tx_id = (SELECT id FROM tx WHERE hash = '\xf9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234'); id ---- (0 rows) ``` DB Sync 11 (cardano-db-sync-extended 11.0.4 - linux-x86_64 - ghc-8.10 git revision b5cd53d2e86f45127d98de429e8a2b6879fd8a56) feels good ``` SELECT id FROM tx_out WHERE tx_id = (SELECT id FROM tx WHERE hash = '\xf9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234'); id ---------- 67617947 (1 row) ```
non_defect
empty tx inputs outputs os debian versions the db sync version eg cardano db sync version cardano db sync extended linux ghc git revision postgresql version build install method the method you use to build or install cardano db sync hydra binaries run method the method you used to run cardano db sync eg nix docker systemd none systemd additional context the database restored from snapshot db sync snapshot schema block tgz problem report there are no rows for tx in the tx out table in db sync select id from tx where hash id row select id from tx out where tx id select id from tx where hash id rows db sync cardano db sync extended linux ghc git revision feels good select id from tx out where tx id select id from tx where hash id row
0
34,598
12,294,008,724
IssuesEvent
2020-05-10 21:40:24
universalturtles/Lisibilite
https://api.github.com/repos/universalturtles/Lisibilite
opened
[STORY]: Authorize Rest calls using API
Security
**A single line user story** Ability to authorize rest calls using API **Details about the user story** Ability to authorize using the below header Authorization: api-key base64(api-id:api-key) **Definition of Done** If token is valid used should be able to user the API If token in invalid, user should not be able to use the API If token is revoked, user should not be able to use the API Log the user usage into a DB with timestamp (For future rate limiting) **Additional Comments/Suggestions/Considerations if any**
True
[STORY]: Authorize Rest calls using API - **A single line user story** Ability to authorize rest calls using API **Details about the user story** Ability to authorize using the below header Authorization: api-key base64(api-id:api-key) **Definition of Done** If token is valid used should be able to user the API If token in invalid, user should not be able to use the API If token is revoked, user should not be able to use the API Log the user usage into a DB with timestamp (For future rate limiting) **Additional Comments/Suggestions/Considerations if any**
non_defect
authorize rest calls using api a single line user story ability to authorize rest calls using api details about the user story ability to authorize using the below header authorization api key api id api key definition of done if token is valid used should be able to user the api if token in invalid user should not be able to use the api if token is revoked user should not be able to use the api log the user usage into a db with timestamp for future rate limiting additional comments suggestions considerations if any
0
9,049
2,615,125,354
IssuesEvent
2015-03-01 05:53:26
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
closed
NullPointerException in GoogleJsonResponseExecption
auto-migrated Milestone-Version1.8.0 Priority-Critical Type-Defect
``` Version of google-api-java-client (e.g. 1.5.0-beta)? 1.7.0-beta Java environment (e.g. Java 6, Android 2.3, App Engine)? All Describe the problem. It appears that Google APIs sometimes return an HTTP error response with Content-Type "application/json; charset=UTF-8" but with no actual content. It results in this stack trace: Thread [main] (Suspended (exception NullPointerException)) InputStreamReader(Reader).<init>(Object) line: 78 InputStreamReader.<init>(InputStream, Charset) line: 113 GsonFactory.createJsonParser(InputStream) line: 48 JsonHttpParser.parserForResponse(JsonFactory, HttpResponse) line: 117 GoogleJsonResponseException.from(JsonFactory, HttpResponse) line: 121 GoogleJsonResponseException.execute(JsonFactory, HttpRequest) line: 187 How would you expect it to be fixed? We should be robust to this kind of problem. That said, this is incorrect behavior from the Google servers. I'll bring that up in a separate bug list. ``` Original issue reported on code.google.com by `yan...@google.com` on 27 Mar 2012 at 3:47
1.0
NullPointerException in GoogleJsonResponseExecption - ``` Version of google-api-java-client (e.g. 1.5.0-beta)? 1.7.0-beta Java environment (e.g. Java 6, Android 2.3, App Engine)? All Describe the problem. It appears that Google APIs sometimes return an HTTP error response with Content-Type "application/json; charset=UTF-8" but with no actual content. It results in this stack trace: Thread [main] (Suspended (exception NullPointerException)) InputStreamReader(Reader).<init>(Object) line: 78 InputStreamReader.<init>(InputStream, Charset) line: 113 GsonFactory.createJsonParser(InputStream) line: 48 JsonHttpParser.parserForResponse(JsonFactory, HttpResponse) line: 117 GoogleJsonResponseException.from(JsonFactory, HttpResponse) line: 121 GoogleJsonResponseException.execute(JsonFactory, HttpRequest) line: 187 How would you expect it to be fixed? We should be robust to this kind of problem. That said, this is incorrect behavior from the Google servers. I'll bring that up in a separate bug list. ``` Original issue reported on code.google.com by `yan...@google.com` on 27 Mar 2012 at 3:47
defect
nullpointerexception in googlejsonresponseexecption version of google api java client e g beta beta java environment e g java android app engine all describe the problem it appears that google apis sometimes return an http error response with content type application json charset utf but with no actual content it results in this stack trace thread suspended exception nullpointerexception inputstreamreader reader object line inputstreamreader inputstream charset line gsonfactory createjsonparser inputstream line jsonhttpparser parserforresponse jsonfactory httpresponse line googlejsonresponseexception from jsonfactory httpresponse line googlejsonresponseexception execute jsonfactory httprequest line how would you expect it to be fixed we should be robust to this kind of problem that said this is incorrect behavior from the google servers i ll bring that up in a separate bug list original issue reported on code google com by yan google com on mar at
1
39,045
15,861,534,008
IssuesEvent
2021-04-08 10:26:25
terraform-providers/terraform-provider-azurerm
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
closed
Support for Service Bus Topic and Queue Data Sources
enhancement service/servicebus
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description <!--- Please leave a helpful description of the feature request here. ---> I would like a way to refer to a Service Bus Topic or Queue as a Data Source ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * azurerm_servicebus_topic * azurerm_servicebus_queue ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl data "azurerm_servicebus_topic" "example" { name = "tfex_servicebus_topic" namespace_name = "tfex-servicebus-namespace" resource_group_name = "terraform-servicebus" } data "azurerm_servicebus_queue" "example" { name = "tfex_servicebus_queue" namespace_name = "tfex-servicebus-namespace" resource_group_name = "terraform-servicebus" } ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/ --->
2.0
Support for Service Bus Topic and Queue Data Sources - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description <!--- Please leave a helpful description of the feature request here. ---> I would like a way to refer to a Service Bus Topic or Queue as a Data Source ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * azurerm_servicebus_topic * azurerm_servicebus_queue ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl data "azurerm_servicebus_topic" "example" { name = "tfex_servicebus_topic" namespace_name = "tfex-servicebus-namespace" resource_group_name = "terraform-servicebus" } data "azurerm_servicebus_queue" "example" { name = "tfex_servicebus_queue" namespace_name = "tfex-servicebus-namespace" resource_group_name = "terraform-servicebus" } ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/ --->
non_defect
support for service bus topic and queue data sources community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description i would like a way to refer to a service bus topic or queue as a data source new or affected resource s azurerm servicebus topic azurerm servicebus queue potential terraform configuration hcl data azurerm servicebus topic example name tfex servicebus topic namespace name tfex servicebus namespace resource group name terraform servicebus data azurerm servicebus queue example name tfex servicebus queue namespace name tfex servicebus namespace resource group name terraform servicebus references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example
0
66,037
19,906,569,005
IssuesEvent
2022-01-25 13:24:45
data4development/iati-workbench
https://api.github.com/repos/data4development/iati-workbench
opened
port narratives merge logic from iati-activities to iati-organisations
Defect Spreadsheets2IATI
The current merge logic for organisation files still produces for instance <description> without <narratives>. Port the fixes in iati-activities to iati-organisations for merging.
1.0
port narratives merge logic from iati-activities to iati-organisations - The current merge logic for organisation files still produces for instance <description> without <narratives>. Port the fixes in iati-activities to iati-organisations for merging.
defect
port narratives merge logic from iati activities to iati organisations the current merge logic for organisation files still produces for instance without port the fixes in iati activities to iati organisations for merging
1
7,192
2,610,357,154
IssuesEvent
2015-02-26 19:55:40
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
UNUSABLE
auto-migrated Priority-Medium Type-Defect
``` What's the problem? The side bar is on the wrong side What browser are you using? Firefox 7.01 What version of ScribeFire are you running? I dunno. It crashes FF when I try to open it, but when I manage to get it started, the sidebar will be on the WRONG SIDE. PUT IT THE F*CK BACK. ``` ----- Original issue reported on code.google.com by `twstdgen...@gmail.com` on 5 Nov 2011 at 5:59
1.0
UNUSABLE - ``` What's the problem? The side bar is on the wrong side What browser are you using? Firefox 7.01 What version of ScribeFire are you running? I dunno. It crashes FF when I try to open it, but when I manage to get it started, the sidebar will be on the WRONG SIDE. PUT IT THE F*CK BACK. ``` ----- Original issue reported on code.google.com by `twstdgen...@gmail.com` on 5 Nov 2011 at 5:59
defect
unusable what s the problem the side bar is on the wrong side what browser are you using firefox what version of scribefire are you running i dunno it crashes ff when i try to open it but when i manage to get it started the sidebar will be on the wrong side put it the f ck back original issue reported on code google com by twstdgen gmail com on nov at
1
84,293
10,368,983,206
IssuesEvent
2019-09-07 21:45:19
SuperCowPowers/bat
https://api.github.com/repos/SuperCowPowers/bat
closed
Switch docs to markdown instead of RST
documentation usability
Going to just use Github's documentation system instead of readthedocs...
1.0
Switch docs to markdown instead of RST - Going to just use Github's documentation system instead of readthedocs...
non_defect
switch docs to markdown instead of rst going to just use github s documentation system instead of readthedocs
0
18,035
3,021,549,513
IssuesEvent
2015-07-31 15:21:14
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
fetchMany() doesn't work with CUBRID
C: DB: CUBRID P: Medium R: Fixed T: Defect
There is an exception when calling `fetchMany()` on CUBRID: ``` java.lang.UnsupportedOperationException at cubrid.jdbc.driver.CUBRIDStatement.getMoreResults(CUBRIDStatement.java:649) at org.jooq.tools.jdbc.DefaultStatement.getMoreResults(DefaultStatement.java:257) at org.jooq.tools.jdbc.DefaultStatement.getMoreResults(DefaultStatement.java:257) at org.jooq.impl.Utils.consumeResultSets(Utils.java:2548) at org.jooq.impl.AbstractResultQuery.execute(AbstractResultQuery.java:282) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:342) at org.jooq.impl.AbstractResultQuery.fetchMany(AbstractResultQuery.java:353) at org.jooq.impl.DefaultDSLContext.fetchMany(DefaultDSLContext.java:540) at org.jooq.test.all.testcases.FetchTests.testFetchMany(FetchTests.java:544) at org.jooq.test.jOOQAbstractTest.testFetchMany(jOOQAbstractTest.java:2106) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner.run(ParentRunner.java:309) at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86) at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192) ``` The problem is in jOOQ's calling `Statement.getMoreResults(int)`, which isn't implemented by CUBRID's JDBC driver: ```java // Call this only when there was at least one ResultSet. // Otherwise, this call is not supported by ojdbc... if (anyResults) ctx.statement().getMoreResults(Statement.CLOSE_ALL_RESULTS); ```
1.0
fetchMany() doesn't work with CUBRID - There is an exception when calling `fetchMany()` on CUBRID: ``` java.lang.UnsupportedOperationException at cubrid.jdbc.driver.CUBRIDStatement.getMoreResults(CUBRIDStatement.java:649) at org.jooq.tools.jdbc.DefaultStatement.getMoreResults(DefaultStatement.java:257) at org.jooq.tools.jdbc.DefaultStatement.getMoreResults(DefaultStatement.java:257) at org.jooq.impl.Utils.consumeResultSets(Utils.java:2548) at org.jooq.impl.AbstractResultQuery.execute(AbstractResultQuery.java:282) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:342) at org.jooq.impl.AbstractResultQuery.fetchMany(AbstractResultQuery.java:353) at org.jooq.impl.DefaultDSLContext.fetchMany(DefaultDSLContext.java:540) at org.jooq.test.all.testcases.FetchTests.testFetchMany(FetchTests.java:544) at org.jooq.test.jOOQAbstractTest.testFetchMany(jOOQAbstractTest.java:2106) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner.run(ParentRunner.java:309) at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86) at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192) ``` The problem is in jOOQ's calling `Statement.getMoreResults(int)`, which isn't implemented by CUBRID's JDBC driver: ```java // Call this only when there was at least one ResultSet. // Otherwise, this call is not supported by ojdbc... if (anyResults) ctx.statement().getMoreResults(Statement.CLOSE_ALL_RESULTS); ```
defect
fetchmany doesn t work with cubrid there is an exception when calling fetchmany on cubrid java lang unsupportedoperationexception at cubrid jdbc driver cubridstatement getmoreresults cubridstatement java at org jooq tools jdbc defaultstatement getmoreresults defaultstatement java at org jooq tools jdbc defaultstatement getmoreresults defaultstatement java at org jooq impl utils consumeresultsets utils java at org jooq impl abstractresultquery execute abstractresultquery java at org jooq impl abstractquery execute abstractquery java at org jooq impl abstractresultquery fetchmany abstractresultquery java at org jooq impl defaultdslcontext fetchmany defaultdslcontext java at org jooq test all testcases fetchtests testfetchmany fetchtests java at org jooq test jooqabstracttest testfetchmany jooqabstracttest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner run parentrunner java at org eclipse jdt internal runner run java at org eclipse jdt internal junit runner testexecution run testexecution java at org eclipse jdt internal junit runner remotetestrunner runtests remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner runtests remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner run remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner main remotetestrunner java the problem is in jooq s calling statement getmoreresults int which isn t implemented by cubrid s jdbc driver java call this only when there was at least one resultset otherwise this call is not supported by ojdbc if anyresults ctx statement getmoreresults statement close all results
1
80,529
30,316,309,354
IssuesEvent
2023-07-10 15:48:08
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
make-icons.sh doesn't make the right icons
T-Defect P3 S-Minor A-Packaging
It's unclear where the icons are coming from if not the script
1.0
make-icons.sh doesn't make the right icons - It's unclear where the icons are coming from if not the script
defect
make icons sh doesn t make the right icons it s unclear where the icons are coming from if not the script
1
54,930
14,046,874,913
IssuesEvent
2020-11-02 05:55:55
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
Unboxing may produce NPE warning in intellij when using count(*)
T: Defect
### Expected behavior No warning ### Actual behavior Unboxing may produce NPE warning in intellij ### Steps to reproduce the problem Code like in stackoverflow [question](https://stackoverflow.com/a/19676187/411965): ``` int count = DSL.using(configuration) .select(DSL.count()) .from(Table) .fetchOne(0, int.class); ``` When the generated code has nullable annotations intellij yields the above warning. The relevant gradle configuration is: ``` generate { generatedAnnotation = true nullableAnnotation = true nullableAnnotationType = 'javax.annotation.Nullable' } ``` ### Versions - nu.studer.jooq: 4.2 - jOOQ: 3.14
1.0
Unboxing may produce NPE warning in intellij when using count(*) - ### Expected behavior No warning ### Actual behavior Unboxing may produce NPE warning in intellij ### Steps to reproduce the problem Code like in stackoverflow [question](https://stackoverflow.com/a/19676187/411965): ``` int count = DSL.using(configuration) .select(DSL.count()) .from(Table) .fetchOne(0, int.class); ``` When the generated code has nullable annotations intellij yields the above warning. The relevant gradle configuration is: ``` generate { generatedAnnotation = true nullableAnnotation = true nullableAnnotationType = 'javax.annotation.Nullable' } ``` ### Versions - nu.studer.jooq: 4.2 - jOOQ: 3.14
defect
unboxing may produce npe warning in intellij when using count expected behavior no warning actual behavior unboxing may produce npe warning in intellij steps to reproduce the problem code like in stackoverflow int count dsl using configuration select dsl count from table fetchone int class when the generated code has nullable annotations intellij yields the above warning the relevant gradle configuration is generate generatedannotation true nullableannotation true nullableannotationtype javax annotation nullable versions nu studer jooq jooq
1
38,409
8,799,194,185
IssuesEvent
2018-12-24 12:36:40
vim/vim
https://api.github.com/repos/vim/vim
closed
terminal resize during file recovery
Priority-Medium auto-migrated defect
``` What steps will reproduce the problem? 1. open a file with vim 2. in a new terminal, open the same file with another vim instance 3. resize your terminal What is the expected output? What do you see instead? I expect the text to reformat while the terminal resizes. Instead, the text gets garbled and never formats properly again. In some cases, the output remains affected after dismissing the prompt, during editing. In either case, the problem can sometimes be fixed by resizing the terminal window once again while editing. What version of the product are you using? On what operating system? $;vim --version VIM - Vi IMproved 7.3 (2010 Aug 15, compiled Nov 16 2010 17:05:25) Included patches: 1-56 Modified by <bugzilla@redhat.com> Compiled by <bugzilla@redhat.com> Huge version without GUI. Features included (+) or not (-): +arabic +autocmd -balloon_eval -browse ++builtin_terms +byte_offset +cindent -clientserver -clipboard +cmdline_compl +cmdline_hist +cmdline_info +comments +conceal +cryptv +cscope +cursorbind +cursorshape +dialog_con +diff +digraphs -dnd -ebcdic +emacs_tags +eval +ex_extra +extra_search +farsi +file_in_path +find_in_path +float +folding -footer +fork() +gettext -hangul_input +iconv +insert_expand +jumplist +keymap +langmap +libcall +linebreak +lispindent +listcmds +localmap -lua +menu +mksession +modify_fname +mouse -mouseshape +mouse_dec +mouse_gpm -mouse_jsbterm +mouse_netterm -mouse_sysmouse +mouse_xterm +multi_byte +multi_lang -mzscheme +netbeans_intg -osfiletype +path_extra +perl +persistent_undo +postscript +printer +profile +python -python3 +quickfix +reltime +rightleft +ruby +scrollbind +signs +smartindent -sniff +startuptime +statusline -sun_workshop +syntax +tag_binary +tag_old_static -tag_any_white -tcl +terminfo +termresponse +textobjects +title -toolbar +user_commands +vertsplit +virtualedit +visual +visualextra +viminfo +vreplace +wildignore +wildmenu +windows +writebackup -X11 -xfontset -xim -xsmp -xterm_clipboard -xterm_save system vimrc file: "/etc/vimrc" user vimrc file: "$HOME/.vimrc" user exrc file: "$HOME/.exrc" fall-back for $VIM: "/etc" f-b for $VIMRUNTIME: "/usr/share/vim/vim73" Compilation: gcc -c -I. -Iproto -DHAVE_CONFIG_H -O2 -g -pipe -Wall -fexceptions -fstack-protector --param=ssp-buffer-size=4 -m64 -mtune=generic -D_GNU_SOURCE -D_FILE_OFFSET_BITS=64 -D_FORTIFY_SOURCE=1 Linking: gcc -L. -rdynamic -Wl,-export-dynamic -Wl,--enable-new-dtags -Wl,-rpath,/usr/lib64/perl5/CORE -L/usr/local/lib -Wl,--as-needed -o vim -lm -lnsl -lselinux -lncurses -lacl -lattr -lgpm -ldl -Wl,--enable-new-dtags -Wl,-rpath,/usr/lib64/perl5/CORE -fstack-protector -L/usr/lib64/perl5/CORE -lperl -lresolv -lnsl -ldl -lm -lcrypt -lutil -lpthread -lc -L/usr/lib64/python2.7/config -lpython2.7 -lpthread -ldl -lutil -lm -Xlinker -export-dynamic -lruby -lpthread -lrt -ldl -lcrypt -lm Fedora 14, XFCE window manager, and xfce4-terminal Please provide any additional information below. ``` Original issue reported on code.google.com by `michaelj...@gmail.com` on 7 Jun 2011 at 7:35
1.0
terminal resize during file recovery - ``` What steps will reproduce the problem? 1. open a file with vim 2. in a new terminal, open the same file with another vim instance 3. resize your terminal What is the expected output? What do you see instead? I expect the text to reformat while the terminal resizes. Instead, the text gets garbled and never formats properly again. In some cases, the output remains affected after dismissing the prompt, during editing. In either case, the problem can sometimes be fixed by resizing the terminal window once again while editing. What version of the product are you using? On what operating system? $;vim --version VIM - Vi IMproved 7.3 (2010 Aug 15, compiled Nov 16 2010 17:05:25) Included patches: 1-56 Modified by <bugzilla@redhat.com> Compiled by <bugzilla@redhat.com> Huge version without GUI. Features included (+) or not (-): +arabic +autocmd -balloon_eval -browse ++builtin_terms +byte_offset +cindent -clientserver -clipboard +cmdline_compl +cmdline_hist +cmdline_info +comments +conceal +cryptv +cscope +cursorbind +cursorshape +dialog_con +diff +digraphs -dnd -ebcdic +emacs_tags +eval +ex_extra +extra_search +farsi +file_in_path +find_in_path +float +folding -footer +fork() +gettext -hangul_input +iconv +insert_expand +jumplist +keymap +langmap +libcall +linebreak +lispindent +listcmds +localmap -lua +menu +mksession +modify_fname +mouse -mouseshape +mouse_dec +mouse_gpm -mouse_jsbterm +mouse_netterm -mouse_sysmouse +mouse_xterm +multi_byte +multi_lang -mzscheme +netbeans_intg -osfiletype +path_extra +perl +persistent_undo +postscript +printer +profile +python -python3 +quickfix +reltime +rightleft +ruby +scrollbind +signs +smartindent -sniff +startuptime +statusline -sun_workshop +syntax +tag_binary +tag_old_static -tag_any_white -tcl +terminfo +termresponse +textobjects +title -toolbar +user_commands +vertsplit +virtualedit +visual +visualextra +viminfo +vreplace +wildignore +wildmenu +windows +writebackup -X11 -xfontset -xim -xsmp -xterm_clipboard -xterm_save system vimrc file: "/etc/vimrc" user vimrc file: "$HOME/.vimrc" user exrc file: "$HOME/.exrc" fall-back for $VIM: "/etc" f-b for $VIMRUNTIME: "/usr/share/vim/vim73" Compilation: gcc -c -I. -Iproto -DHAVE_CONFIG_H -O2 -g -pipe -Wall -fexceptions -fstack-protector --param=ssp-buffer-size=4 -m64 -mtune=generic -D_GNU_SOURCE -D_FILE_OFFSET_BITS=64 -D_FORTIFY_SOURCE=1 Linking: gcc -L. -rdynamic -Wl,-export-dynamic -Wl,--enable-new-dtags -Wl,-rpath,/usr/lib64/perl5/CORE -L/usr/local/lib -Wl,--as-needed -o vim -lm -lnsl -lselinux -lncurses -lacl -lattr -lgpm -ldl -Wl,--enable-new-dtags -Wl,-rpath,/usr/lib64/perl5/CORE -fstack-protector -L/usr/lib64/perl5/CORE -lperl -lresolv -lnsl -ldl -lm -lcrypt -lutil -lpthread -lc -L/usr/lib64/python2.7/config -lpython2.7 -lpthread -ldl -lutil -lm -Xlinker -export-dynamic -lruby -lpthread -lrt -ldl -lcrypt -lm Fedora 14, XFCE window manager, and xfce4-terminal Please provide any additional information below. ``` Original issue reported on code.google.com by `michaelj...@gmail.com` on 7 Jun 2011 at 7:35
defect
terminal resize during file recovery what steps will reproduce the problem open a file with vim in a new terminal open the same file with another vim instance resize your terminal what is the expected output what do you see instead i expect the text to reformat while the terminal resizes instead the text gets garbled and never formats properly again in some cases the output remains affected after dismissing the prompt during editing in either case the problem can sometimes be fixed by resizing the terminal window once again while editing what version of the product are you using on what operating system vim version vim vi improved aug compiled nov included patches modified by compiled by huge version without gui features included or not arabic autocmd balloon eval browse builtin terms byte offset cindent clientserver clipboard cmdline compl cmdline hist cmdline info comments conceal cryptv cscope cursorbind cursorshape dialog con diff digraphs dnd ebcdic emacs tags eval ex extra extra search farsi file in path find in path float folding footer fork gettext hangul input iconv insert expand jumplist keymap langmap libcall linebreak lispindent listcmds localmap lua menu mksession modify fname mouse mouseshape mouse dec mouse gpm mouse jsbterm mouse netterm mouse sysmouse mouse xterm multi byte multi lang mzscheme netbeans intg osfiletype path extra perl persistent undo postscript printer profile python quickfix reltime rightleft ruby scrollbind signs smartindent sniff startuptime statusline sun workshop syntax tag binary tag old static tag any white tcl terminfo termresponse textobjects title toolbar user commands vertsplit virtualedit visual visualextra viminfo vreplace wildignore wildmenu windows writebackup xfontset xim xsmp xterm clipboard xterm save system vimrc file etc vimrc user vimrc file home vimrc user exrc file home exrc fall back for vim etc f b for vimruntime usr share vim compilation gcc c i iproto dhave config h g pipe wall fexceptions fstack protector param ssp buffer size mtune generic d gnu source d file offset bits d fortify source linking gcc l rdynamic wl export dynamic wl enable new dtags wl rpath usr core l usr local lib wl as needed o vim lm lnsl lselinux lncurses lacl lattr lgpm ldl wl enable new dtags wl rpath usr core fstack protector l usr core lperl lresolv lnsl ldl lm lcrypt lutil lpthread lc l usr config lpthread ldl lutil lm xlinker export dynamic lruby lpthread lrt ldl lcrypt lm fedora xfce window manager and terminal please provide any additional information below original issue reported on code google com by michaelj gmail com on jun at
1
62,192
7,551,047,845
IssuesEvent
2018-04-18 18:50:32
phetsims/area-model-common
https://api.github.com/repos/phetsims/area-model-common
opened
Keypad has no way to be dismissed
design:general
@EthanWJohnson brought up the concern that there is no way to dismiss a keypad without pressing "Enter". This means that the user would be forced to overwrite an entry with a new value whenever they wish to dismiss the keypad. Here are 4 quick brainstorms to avoid this: (1) Make the keypad modal. (Means you couldn't jump between edit boxes on the fly). (2) Add a "Cancel" button to the keypad. (3) Dismiss the keypad by clicking on the box/pencil button you are currently editing. (4) Have blank entries NOT actually overwrite the current value with a blank value. Instead, submitting blank entries does nothing.
1.0
Keypad has no way to be dismissed - @EthanWJohnson brought up the concern that there is no way to dismiss a keypad without pressing "Enter". This means that the user would be forced to overwrite an entry with a new value whenever they wish to dismiss the keypad. Here are 4 quick brainstorms to avoid this: (1) Make the keypad modal. (Means you couldn't jump between edit boxes on the fly). (2) Add a "Cancel" button to the keypad. (3) Dismiss the keypad by clicking on the box/pencil button you are currently editing. (4) Have blank entries NOT actually overwrite the current value with a blank value. Instead, submitting blank entries does nothing.
non_defect
keypad has no way to be dismissed ethanwjohnson brought up the concern that there is no way to dismiss a keypad without pressing enter this means that the user would be forced to overwrite an entry with a new value whenever they wish to dismiss the keypad here are quick brainstorms to avoid this make the keypad modal means you couldn t jump between edit boxes on the fly add a cancel button to the keypad dismiss the keypad by clicking on the box pencil button you are currently editing have blank entries not actually overwrite the current value with a blank value instead submitting blank entries does nothing
0
38,978
9,103,965,110
IssuesEvent
2019-02-20 17:00:31
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
Error check in FieldSplit Preconditioner
C: MOOSE P: minor T: defect
## Rationale <!--What is the reason for this enhancement or what error are you reporting?--> No error is reported when you use invalid variable names with the FSP system. ## Description <!--Provide details of the enhancement or instructions for reproducing the error.--> See the discussion here: https://groups.google.com/forum/?utm_medium=email&utm_source=footer#!msg/moose-users/RuMWaoi14Ho/9fjoneVrBQAJ ## Impact <!--How will the changes impact the code, developers, and users? Discuss changes to the internal interfaces and public API.--> Minor: Bug but not a widely used system.
1.0
Error check in FieldSplit Preconditioner - ## Rationale <!--What is the reason for this enhancement or what error are you reporting?--> No error is reported when you use invalid variable names with the FSP system. ## Description <!--Provide details of the enhancement or instructions for reproducing the error.--> See the discussion here: https://groups.google.com/forum/?utm_medium=email&utm_source=footer#!msg/moose-users/RuMWaoi14Ho/9fjoneVrBQAJ ## Impact <!--How will the changes impact the code, developers, and users? Discuss changes to the internal interfaces and public API.--> Minor: Bug but not a widely used system.
defect
error check in fieldsplit preconditioner rationale no error is reported when you use invalid variable names with the fsp system description see the discussion here impact how will the changes impact the code developers and users discuss changes to the internal interfaces and public api minor bug but not a widely used system
1
41,116
5,308,759,160
IssuesEvent
2017-02-12 01:07:03
RestComm/Restcomm-Connect
https://api.github.com/repos/RestComm/Restcomm-Connect
opened
Assuming Subaccount identity to help a newbie
2. Enhancement UI UX Visual App Designer
Its common to have to help a new user who is learning how to use Restcomm. It would be practical to be able to assume a subaccount identity without the subaccount user disclosing their password. Restcomm should log for audit purpose who performed actions for any given account - the account holder themselves or a superaccount holder. <img width="1124" alt="screen shot 2017-02-11 at 5 03 25 pm" src="https://cloud.githubusercontent.com/assets/2234901/22858789/5a8dfb9e-f07c-11e6-9dc0-33426fef40e6.png">
1.0
Assuming Subaccount identity to help a newbie - Its common to have to help a new user who is learning how to use Restcomm. It would be practical to be able to assume a subaccount identity without the subaccount user disclosing their password. Restcomm should log for audit purpose who performed actions for any given account - the account holder themselves or a superaccount holder. <img width="1124" alt="screen shot 2017-02-11 at 5 03 25 pm" src="https://cloud.githubusercontent.com/assets/2234901/22858789/5a8dfb9e-f07c-11e6-9dc0-33426fef40e6.png">
non_defect
assuming subaccount identity to help a newbie its common to have to help a new user who is learning how to use restcomm it would be practical to be able to assume a subaccount identity without the subaccount user disclosing their password restcomm should log for audit purpose who performed actions for any given account the account holder themselves or a superaccount holder img width alt screen shot at pm src
0
75,195
7,463,475,959
IssuesEvent
2018-04-01 05:56:46
slackertracker/slackertracker
https://api.github.com/repos/slackertracker/slackertracker
closed
Test Models
Models Tests
Create some simple tests for the models to make sure they save correctly to the database. - User - Comment - Reaction
1.0
Test Models - Create some simple tests for the models to make sure they save correctly to the database. - User - Comment - Reaction
non_defect
test models create some simple tests for the models to make sure they save correctly to the database user comment reaction
0
128,493
10,540,656,001
IssuesEvent
2019-10-02 08:55:45
ckeditor/ckeditor4
https://api.github.com/repos/ckeditor/ckeditor4
closed
Unstable Promise AMD test
status:confirmed target:minor type:failingtest
## Type of report Failing test ## Provide detailed reproduction steps (if any) Sometimes Promise AMD loading test fails randomly, for example: ![image](https://user-images.githubusercontent.com/1061942/65035356-41208100-d949-11e9-97b1-a2000e7d252a.png) It sometimes fails on Chrome sometimes on Firefox so it's not any specific browser fault.
1.0
Unstable Promise AMD test - ## Type of report Failing test ## Provide detailed reproduction steps (if any) Sometimes Promise AMD loading test fails randomly, for example: ![image](https://user-images.githubusercontent.com/1061942/65035356-41208100-d949-11e9-97b1-a2000e7d252a.png) It sometimes fails on Chrome sometimes on Firefox so it's not any specific browser fault.
non_defect
unstable promise amd test type of report failing test provide detailed reproduction steps if any sometimes promise amd loading test fails randomly for example it sometimes fails on chrome sometimes on firefox so it s not any specific browser fault
0
78,824
27,773,922,282
IssuesEvent
2023-03-16 16:01:10
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Desktop client loses keys when disk is full
T-Defect S-Critical A-E2EE A-Storage O-Occasional
<!-- This is a bug report template. By following the instructions below and filling out the sections with your information, you will help the us to get all the necessary data to fix your issue. You can also preview your report before submitting it. You may remove sections that aren't relevant to your particular case. Text between <!-- and --​> marks will be invisible in the report. --> ### Description I ran out of disk space and I believe I rebooted in a disk full condition in order to lose swap files. After reboot, I get vector-im/element-web#2996. I'm not sure if vector-im/element-web#3660 is related or if desktop uses a real file API where you can avoid writing to a file when there is no disk space. ### Version information <!-- IMPORTANT: please answer the following questions, to help us narrow down the problem --> - OSX Desktop 0.9.9
1.0
Desktop client loses keys when disk is full - <!-- This is a bug report template. By following the instructions below and filling out the sections with your information, you will help the us to get all the necessary data to fix your issue. You can also preview your report before submitting it. You may remove sections that aren't relevant to your particular case. Text between <!-- and --​> marks will be invisible in the report. --> ### Description I ran out of disk space and I believe I rebooted in a disk full condition in order to lose swap files. After reboot, I get vector-im/element-web#2996. I'm not sure if vector-im/element-web#3660 is related or if desktop uses a real file API where you can avoid writing to a file when there is no disk space. ### Version information <!-- IMPORTANT: please answer the following questions, to help us narrow down the problem --> - OSX Desktop 0.9.9
defect
desktop client loses keys when disk is full this is a bug report template by following the instructions below and filling out the sections with your information you will help the us to get all the necessary data to fix your issue you can also preview your report before submitting it you may remove sections that aren t relevant to your particular case text between marks will be invisible in the report description i ran out of disk space and i believe i rebooted in a disk full condition in order to lose swap files after reboot i get vector im element web i m not sure if vector im element web is related or if desktop uses a real file api where you can avoid writing to a file when there is no disk space version information osx desktop
1
18,399
3,054,490,588
IssuesEvent
2015-08-13 03:11:00
eczarny/spectacle
https://api.github.com/repos/eczarny/spectacle
closed
Alternate keyboard settings not working
defect ★★
I nearly always have the "Dvorak – Qwerty ⌘" keyboard setting enabled, which allows me to type in Dvorak (all the vowels on the left side of the home row, all the common consonants on the right side of the home row—google Dvorak if you haven't encountered this before) but still lets me use all the standard shortcuts I am used to. For example, the "V" key types a "J" when Dvorak is enabled, but with "Dvorak – Qwerty ⌘", typing ⌘ - "V key" will still do "Paste". (As opposed to just straight up Dvorak layout, wherein I would have to type "⌘ - Period key", since the period key types a "V" in Dvorak.) Not so with Spectacle. For example, to enter the shortcut "⌥⌘J" to center a window, I have to type "⌥⌘C", since Spectacle doesn't recognize the "– Qwerty ⌘" aspect of the keyboard layout. (The "C" key types "J" in standard Dvorak layout.) I know this is somewhat esoteric, but all other apps deal with this keyboard setting okay, with very few exceptions. (LibreOffice mostly works with it, with a few quirks; Acrobat Reader is the only major app I've found that flatly refuses to recognize the shortcuts correctly.)
1.0
Alternate keyboard settings not working - I nearly always have the "Dvorak – Qwerty ⌘" keyboard setting enabled, which allows me to type in Dvorak (all the vowels on the left side of the home row, all the common consonants on the right side of the home row—google Dvorak if you haven't encountered this before) but still lets me use all the standard shortcuts I am used to. For example, the "V" key types a "J" when Dvorak is enabled, but with "Dvorak – Qwerty ⌘", typing ⌘ - "V key" will still do "Paste". (As opposed to just straight up Dvorak layout, wherein I would have to type "⌘ - Period key", since the period key types a "V" in Dvorak.) Not so with Spectacle. For example, to enter the shortcut "⌥⌘J" to center a window, I have to type "⌥⌘C", since Spectacle doesn't recognize the "– Qwerty ⌘" aspect of the keyboard layout. (The "C" key types "J" in standard Dvorak layout.) I know this is somewhat esoteric, but all other apps deal with this keyboard setting okay, with very few exceptions. (LibreOffice mostly works with it, with a few quirks; Acrobat Reader is the only major app I've found that flatly refuses to recognize the shortcuts correctly.)
defect
alternate keyboard settings not working i nearly always have the dvorak – qwerty ⌘ keyboard setting enabled which allows me to type in dvorak all the vowels on the left side of the home row all the common consonants on the right side of the home row—google dvorak if you haven t encountered this before but still lets me use all the standard shortcuts i am used to for example the v key types a j when dvorak is enabled but with dvorak – qwerty ⌘ typing ⌘ v key will still do paste as opposed to just straight up dvorak layout wherein i would have to type ⌘ period key since the period key types a v in dvorak not so with spectacle for example to enter the shortcut ⌥⌘j to center a window i have to type ⌥⌘c since spectacle doesn t recognize the – qwerty ⌘ aspect of the keyboard layout the c key types j in standard dvorak layout i know this is somewhat esoteric but all other apps deal with this keyboard setting okay with very few exceptions libreoffice mostly works with it with a few quirks acrobat reader is the only major app i ve found that flatly refuses to recognize the shortcuts correctly
1
81,448
30,853,449,970
IssuesEvent
2023-08-02 18:34:30
dotCMS/core
https://api.github.com/repos/dotCMS/core
closed
We are not cleaning the cache Page correctly
Type : Defect Merged QA : Passed Internal Team : Falcon dotCMS : Experiments
### Parent Issue https://github.com/dotCMS/core/issues/22222 ### Problem Statement Right now we are cleaning the Page Cache when any change come over the Page, but we are cleaning this Cache using the MultiTree of the page but a Page can be render on any Variant even when it does not have any MultiTree specific for that Variant (it is render by fallback but the Cache is full with the result). Maybe we need to clean up the Page's Cache for all the existing no archived Variants ### Steps to Reproduce Pending ### Acceptance Criteria The cache must be clean for all the Variant where the Page was render no matter the existing MultiTree ### dotCMS Version Lates in master ### Proposed Objective Core Features ### Proposed Priority Priority 2 - Important ### External Links... Slack Conversations, Support Tickets, Figma Designs, etc. _No response_ ### Assumptions & Initiation Needs _No response_ ### Quality Assurance Notes & Workarounds _No response_ ### Sub-Tasks & Estimates _No response_
1.0
We are not cleaning the cache Page correctly - ### Parent Issue https://github.com/dotCMS/core/issues/22222 ### Problem Statement Right now we are cleaning the Page Cache when any change come over the Page, but we are cleaning this Cache using the MultiTree of the page but a Page can be render on any Variant even when it does not have any MultiTree specific for that Variant (it is render by fallback but the Cache is full with the result). Maybe we need to clean up the Page's Cache for all the existing no archived Variants ### Steps to Reproduce Pending ### Acceptance Criteria The cache must be clean for all the Variant where the Page was render no matter the existing MultiTree ### dotCMS Version Lates in master ### Proposed Objective Core Features ### Proposed Priority Priority 2 - Important ### External Links... Slack Conversations, Support Tickets, Figma Designs, etc. _No response_ ### Assumptions & Initiation Needs _No response_ ### Quality Assurance Notes & Workarounds _No response_ ### Sub-Tasks & Estimates _No response_
defect
we are not cleaning the cache page correctly parent issue problem statement right now we are cleaning the page cache when any change come over the page but we are cleaning this cache using the multitree of the page but a page can be render on any variant even when it does not have any multitree specific for that variant it is render by fallback but the cache is full with the result maybe we need to clean up the page s cache for all the existing no archived variants steps to reproduce pending acceptance criteria the cache must be clean for all the variant where the page was render no matter the existing multitree dotcms version lates in master proposed objective core features proposed priority priority important external links slack conversations support tickets figma designs etc no response assumptions initiation needs no response quality assurance notes workarounds no response sub tasks estimates no response
1
13,122
9,882,899,989
IssuesEvent
2019-06-24 18:02:22
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
https for image computer-vision
cognitive-services/svc computer-vision/subsvc cxp doc-enhancement triaged
Hello, The documentation says that we can access our local endpoint with https://localhost:5000 - it appears that only http is supported. Is there a configuration setting to force https? Thanks --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 439c70ec-4484-25de-4942-939112188d58 * Version Independent ID: 08f0bb9d-c6a0-d35b-8b3e-e46cca613487 * Content: [How to install and run containers - Computer Vision](https://docs.microsoft.com/en-us/azure/cognitive-services/computer-vision/computer-vision-how-to-install-containers#feedback) * Content Source: [articles/cognitive-services/Computer-vision/computer-vision-how-to-install-containers.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Computer-vision/computer-vision-how-to-install-containers.md) * Service: **cognitive-services** * Sub-service: **computer-vision** * GitHub Login: @IEvangelist * Microsoft Alias: **IEvangelist**
1.0
https for image computer-vision - Hello, The documentation says that we can access our local endpoint with https://localhost:5000 - it appears that only http is supported. Is there a configuration setting to force https? Thanks --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 439c70ec-4484-25de-4942-939112188d58 * Version Independent ID: 08f0bb9d-c6a0-d35b-8b3e-e46cca613487 * Content: [How to install and run containers - Computer Vision](https://docs.microsoft.com/en-us/azure/cognitive-services/computer-vision/computer-vision-how-to-install-containers#feedback) * Content Source: [articles/cognitive-services/Computer-vision/computer-vision-how-to-install-containers.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Computer-vision/computer-vision-how-to-install-containers.md) * Service: **cognitive-services** * Sub-service: **computer-vision** * GitHub Login: @IEvangelist * Microsoft Alias: **IEvangelist**
non_defect
https for image computer vision hello the documentation says that we can access our local endpoint with it appears that only http is supported is there a configuration setting to force https thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service cognitive services sub service computer vision github login ievangelist microsoft alias ievangelist
0
22,621
3,670,923,412
IssuesEvent
2016-02-22 02:39:24
gperftools/gperftools
https://api.github.com/repos/gperftools/gperftools
closed
tcmalloc crash at CheckCachedSizeClass
Priority-Medium Status-New Type-Defect
Originally reported on Google Code with ID 519 ``` #3 0x0000000000dc00a8 in google::FlushLogFilesUnsafe (min_severity=0) at thirdparty/glog-0.3.2/src/logging.cc:1511 #4 0x0000000000dc618c in google::(anonymous namespace)::FailureSignalHandler (signal_number=11, signal_info=0x63712270, ucontext=0x63712140) at thirdparty/glog-0.3.2/src/signalhandler.cc:324 #5 <signal handler called> #6 0x00000000009d1bd4 in CheckCachedSizeClass (ptr=0x7fbdf92f28a0) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:946 #7 0x00000000009d1c0e in CheckedMallocResult (result=0x7fbdf92f28a0) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:950 #8 0x00000000009d225b in (anonymous namespace)::do_malloc (size=96) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:1084 #9 0x00000000009d2ea6 in (anonymous namespace)::cpp_alloc (size=96, nothrow=false) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:1360 #10 0x00000000009d5ff7 in MallocBlock::Allocate (size=48, type=-21308287) at thirdparty/google-perftools-1.7/src/debugallocation.cc:537 #11 0x00000000009d3942 in DebugAllocate (size=48, type=-21308287) at thirdparty/google-perftools-1.7/src/debugallocation.cc:970 #12 0x00000000009d6b26 in cpp_debug_alloc (size=48, new_type=-21308287, nothrow=false) at thirdparty/google-perftools-1.7/src/debugallocation.cc:1203 #13 0x0000000000e3d0e7 in operator new (size=48) at thirdparty/google-perftools-1.7/src/debugallocation.cc:1250 #14 0x000000000095a001 in __gnu_cxx::new_allocator<std::_Rb_tree_node<std::pair<torca::master::Job const* const, unsigned int> > >::allocate (this=0x7fbe1b61d020, __n=1) at /usr/local/lib/gcc/x86_64-unknown-linux-gnu/4.5.1/../../../../include/c++/4.5.1/ext/new_allocator.h:89 #15 0x0000000000959be0 in std::_Rb_tree<torca::master::Job const*, std::pair<torca::master::Job const* const, unsigned int>, std::_Select1st<std::pair<torca::master::Job const* const, unsigned int> >, torca::master::JobCompare, std::allocator<std::pair<torca::master::Job const* const, unsigned int> > >::_M_get_node (this=0x7fbe1b61d020) at /usr/local/lib/gcc/x86_64-unknown-linux-gnu/4.5.1/../../../../include/c++/4.5.1/bits/stl_tree.h:359 #16 0x00000000009590ef in std::_Rb_tree<torca::master::Job const*, std::pair<torca::master::Job const* const, unsigned int>, std::_Select1st<std::pair<torca::master::Job const* const, unsigned int> >, torca::master::JobCompare, std::allocator<std::pair<torca::master::Job const* const, unsigned int> > >::_M_create_node (this=0x7fbe1b61d020, __x=...) What steps will reproduce the problem? I don't know how to reproduce it, it happened just once. What is the expected output? What do you see instead? My program crash at return reinterpret_cast<Leaf*>(root_->ptrs[i1]->ptrs[i2])->values[i3]; values[i3] = 0xFFFFFFFFDEADBEAF。 What version of the product are you using? On what operating system? I use google-perftools-1.7 on SUSE. Please provide any additional information below. ``` Reported by `tyrionqq` on 2013-04-10 06:17:18
1.0
tcmalloc crash at CheckCachedSizeClass - Originally reported on Google Code with ID 519 ``` #3 0x0000000000dc00a8 in google::FlushLogFilesUnsafe (min_severity=0) at thirdparty/glog-0.3.2/src/logging.cc:1511 #4 0x0000000000dc618c in google::(anonymous namespace)::FailureSignalHandler (signal_number=11, signal_info=0x63712270, ucontext=0x63712140) at thirdparty/glog-0.3.2/src/signalhandler.cc:324 #5 <signal handler called> #6 0x00000000009d1bd4 in CheckCachedSizeClass (ptr=0x7fbdf92f28a0) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:946 #7 0x00000000009d1c0e in CheckedMallocResult (result=0x7fbdf92f28a0) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:950 #8 0x00000000009d225b in (anonymous namespace)::do_malloc (size=96) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:1084 #9 0x00000000009d2ea6 in (anonymous namespace)::cpp_alloc (size=96, nothrow=false) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:1360 #10 0x00000000009d5ff7 in MallocBlock::Allocate (size=48, type=-21308287) at thirdparty/google-perftools-1.7/src/debugallocation.cc:537 #11 0x00000000009d3942 in DebugAllocate (size=48, type=-21308287) at thirdparty/google-perftools-1.7/src/debugallocation.cc:970 #12 0x00000000009d6b26 in cpp_debug_alloc (size=48, new_type=-21308287, nothrow=false) at thirdparty/google-perftools-1.7/src/debugallocation.cc:1203 #13 0x0000000000e3d0e7 in operator new (size=48) at thirdparty/google-perftools-1.7/src/debugallocation.cc:1250 #14 0x000000000095a001 in __gnu_cxx::new_allocator<std::_Rb_tree_node<std::pair<torca::master::Job const* const, unsigned int> > >::allocate (this=0x7fbe1b61d020, __n=1) at /usr/local/lib/gcc/x86_64-unknown-linux-gnu/4.5.1/../../../../include/c++/4.5.1/ext/new_allocator.h:89 #15 0x0000000000959be0 in std::_Rb_tree<torca::master::Job const*, std::pair<torca::master::Job const* const, unsigned int>, std::_Select1st<std::pair<torca::master::Job const* const, unsigned int> >, torca::master::JobCompare, std::allocator<std::pair<torca::master::Job const* const, unsigned int> > >::_M_get_node (this=0x7fbe1b61d020) at /usr/local/lib/gcc/x86_64-unknown-linux-gnu/4.5.1/../../../../include/c++/4.5.1/bits/stl_tree.h:359 #16 0x00000000009590ef in std::_Rb_tree<torca::master::Job const*, std::pair<torca::master::Job const* const, unsigned int>, std::_Select1st<std::pair<torca::master::Job const* const, unsigned int> >, torca::master::JobCompare, std::allocator<std::pair<torca::master::Job const* const, unsigned int> > >::_M_create_node (this=0x7fbe1b61d020, __x=...) What steps will reproduce the problem? I don't know how to reproduce it, it happened just once. What is the expected output? What do you see instead? My program crash at return reinterpret_cast<Leaf*>(root_->ptrs[i1]->ptrs[i2])->values[i3]; values[i3] = 0xFFFFFFFFDEADBEAF。 What version of the product are you using? On what operating system? I use google-perftools-1.7 on SUSE. Please provide any additional information below. ``` Reported by `tyrionqq` on 2013-04-10 06:17:18
defect
tcmalloc crash at checkcachedsizeclass originally reported on google code with id in google flushlogfilesunsafe min severity at thirdparty glog src logging cc in google anonymous namespace failuresignalhandler signal number signal info ucontext at thirdparty glog src signalhandler cc in checkcachedsizeclass ptr at thirdparty google perftools src tcmalloc cc in checkedmallocresult result at thirdparty google perftools src tcmalloc cc in anonymous namespace do malloc size at thirdparty google perftools src tcmalloc cc in anonymous namespace cpp alloc size nothrow false at thirdparty google perftools src tcmalloc cc in mallocblock allocate size type at thirdparty google perftools src debugallocation cc in debugallocate size type at thirdparty google perftools src debugallocation cc in cpp debug alloc size new type nothrow false at thirdparty google perftools src debugallocation cc in operator new size at thirdparty google perftools src debugallocation cc in gnu cxx new allocator std rb tree node std pair torca master job const const unsigned int allocate this n at usr local lib gcc unknown linux gnu include c ext new allocator h in std rb tree torca master job const std pair torca master job const const unsigned int std std pair torca master job const const unsigned int torca master jobcompare std allocator std pair torca master job const const unsigned int m get node this at usr local lib gcc unknown linux gnu include c bits stl tree h in std rb tree torca master job const std pair torca master job const const unsigned int std std pair torca master job const const unsigned int torca master jobcompare std allocator std pair torca master job const const unsigned int m create node this x what steps will reproduce the problem i don t know how to reproduce it it happened just once what is the expected output what do you see instead my program crash at return reinterpret cast root ptrs ptrs values values 。 what version of the product are you using on what operating system i use google perftools on suse please provide any additional information below reported by tyrionqq on
1
329,446
10,019,946,478
IssuesEvent
2019-07-16 11:23:27
aiidateam/aiida-core
https://api.github.com/repos/aiidateam/aiida-core
closed
Update requirements for `plumpy` dependency
priority/important topic/dependencies
New version has just been released `plumpy==0.14.2`
1.0
Update requirements for `plumpy` dependency - New version has just been released `plumpy==0.14.2`
non_defect
update requirements for plumpy dependency new version has just been released plumpy
0
102,165
31,847,506,646
IssuesEvent
2023-09-14 21:13:34
orbeon/orbeon-forms
https://api.github.com/repos/orbeon/orbeon-forms
closed
JavaScript error in Form Builder when copying a Formatted Text Area
Module: Form Builder
This happens when doing the copy while the menu is displayed. Somehow I don't get the error when I just move it. Below, I first try to move it from left to right, which works fine, but then I get the error when copying from right to left. This is with the latest code on the `2022.1-pe` branch. ![screencast](https://github.com/orbeon/orbeon-forms/assets/106961/8788cbe5-d23a-4689-b490-da749e760428)
1.0
JavaScript error in Form Builder when copying a Formatted Text Area - This happens when doing the copy while the menu is displayed. Somehow I don't get the error when I just move it. Below, I first try to move it from left to right, which works fine, but then I get the error when copying from right to left. This is with the latest code on the `2022.1-pe` branch. ![screencast](https://github.com/orbeon/orbeon-forms/assets/106961/8788cbe5-d23a-4689-b490-da749e760428)
non_defect
javascript error in form builder when copying a formatted text area this happens when doing the copy while the menu is displayed somehow i don t get the error when i just move it below i first try to move it from left to right which works fine but then i get the error when copying from right to left this is with the latest code on the pe branch
0
37,605
8,468,380,240
IssuesEvent
2018-10-23 19:34:33
idaholab/raven
https://api.github.com/repos/idaholab/raven
closed
MOOSE Input Drops Text After Second '='
defect external interfaces priority_critical
-------- Issue Description -------- ##### What did you expect to see happen? I expected `function = 'alpha := total_volume * current_power_density * (1 - tau); beta := tau * current_power_density; alpha * beta'` in my moose input file to be parsed to the same thing when it spawned all the jobs. ##### What did you see instead? Instead it dropped all the text after the second `=` putting in the new input file `function = 'alpha :` causing the run to error. ##### Do you have a suggested fix for the development team? I do! Expect a merge request soon. :smiley: And it is running now with the fix. ##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or improvement? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [x] 1. If the issue is a defect, is the defect fixed? - [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [x] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)? - [x] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
1.0
MOOSE Input Drops Text After Second '=' - -------- Issue Description -------- ##### What did you expect to see happen? I expected `function = 'alpha := total_volume * current_power_density * (1 - tau); beta := tau * current_power_density; alpha * beta'` in my moose input file to be parsed to the same thing when it spawned all the jobs. ##### What did you see instead? Instead it dropped all the text after the second `=` putting in the new input file `function = 'alpha :` causing the run to error. ##### Do you have a suggested fix for the development team? I do! Expect a merge request soon. :smiley: And it is running now with the fix. ##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or improvement? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [x] 1. If the issue is a defect, is the defect fixed? - [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [x] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)? - [x] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
defect
moose input drops text after second issue description what did you expect to see happen i expected function alpha total volume current power density tau beta tau current power density alpha beta in my moose input file to be parsed to the same thing when it spawned all the jobs what did you see instead instead it dropped all the text after the second putting in the new input file function alpha causing the run to error do you have a suggested fix for the development team i do expect a merge request soon smiley and it is running now with the fix please attach the input file s that generate this error the simpler the input the faster we can find the issue for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided
1
73,482
24,652,420,724
IssuesEvent
2022-10-17 19:51:44
shedskin/shedskin
https://api.github.com/repos/shedskin/shedskin
closed
Crash when using string.split()
auto-migrated Priority-Medium Type-Defect platform-osx
**Issue by inn...@gmail.com** _13 Sep 2011 at 3:27 GMT_ _Originally opened on Google Code_ --- ``` What steps will reproduce the problem? 1. Compile and run the following code: a = '7,0,1,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,1,0,2,1,0,0,0,0,0,0,0,0,0,0 ,0,0,0,0,0,0,0,0,2,2,1,0,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,3,1,0,0 ,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0 ,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0' b = '7,0,1,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,1,0,2,1,0,0,0,0,0,0,0,0,0,0 ,0,0,0,0,0,0,0,0,2,2,1,0,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,3,1,0,0 ,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0 ,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0' c = '7,0,1,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,1,0,2,1,0,0,0,0,0,0,0,0,0,0 ,0,0,0,0,0,0,0,0,2,2,1,0,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,3,1,0,0 ,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0 ,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0' while True: d = a.split(',') d = b.split(',') d = c.split(',') What is the expected output? What do you see instead? Segfaults for some reason. Doesn't happen under gdb, which makes the problem even trickier. What version of the product are you using? On what operating system? 0.9, OS X 10.7.0, gcc 4.2.1 ```
1.0
Crash when using string.split() - **Issue by inn...@gmail.com** _13 Sep 2011 at 3:27 GMT_ _Originally opened on Google Code_ --- ``` What steps will reproduce the problem? 1. Compile and run the following code: a = '7,0,1,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,1,0,2,1,0,0,0,0,0,0,0,0,0,0 ,0,0,0,0,0,0,0,0,2,2,1,0,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,3,1,0,0 ,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0 ,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0' b = '7,0,1,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,1,0,2,1,0,0,0,0,0,0,0,0,0,0 ,0,0,0,0,0,0,0,0,2,2,1,0,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,3,1,0,0 ,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0 ,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0' c = '7,0,1,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,1,0,2,1,0,0,0,0,0,0,0,0,0,0 ,0,0,0,0,0,0,0,0,2,2,1,0,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,3,1,0,0 ,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0 ,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0' while True: d = a.split(',') d = b.split(',') d = c.split(',') What is the expected output? What do you see instead? Segfaults for some reason. Doesn't happen under gdb, which makes the problem even trickier. What version of the product are you using? On what operating system? 0.9, OS X 10.7.0, gcc 4.2.1 ```
defect
crash when using string split issue by inn gmail com sep at gmt originally opened on google code what steps will reproduce the problem compile and run the following code a b c while true d a split d b split d c split what is the expected output what do you see instead segfaults for some reason doesn t happen under gdb which makes the problem even trickier what version of the product are you using on what operating system os x gcc
1
19,951
3,283,913,298
IssuesEvent
2015-10-28 14:46:34
ARMmbed/uvisor
https://api.github.com/repos/ARMmbed/uvisor
opened
Hardware floating point support
defect
- add hardware floating point support for the main box as a first step - extend floating point support to other secure boxes later
1.0
Hardware floating point support - - add hardware floating point support for the main box as a first step - extend floating point support to other secure boxes later
defect
hardware floating point support add hardware floating point support for the main box as a first step extend floating point support to other secure boxes later
1
10,165
2,618,939,869
IssuesEvent
2015-03-03 00:03:40
marmarek/test
https://api.github.com/repos/marmarek/test
closed
unbind_all_network_devices: fix regexp to cover all pci devs
C: core P: major R: fixed T: defect
**Reported by joanna on 13 May 40287457 01:20 UTC** The regexp used by unbind_all_network_devices script ignores devices that have hex digits (a-f) in their BDF addresses. As reported here: http://groups.google.com/group/qubes-devel/browse_thread/thread/66d517a6c67085f9
1.0
unbind_all_network_devices: fix regexp to cover all pci devs - **Reported by joanna on 13 May 40287457 01:20 UTC** The regexp used by unbind_all_network_devices script ignores devices that have hex digits (a-f) in their BDF addresses. As reported here: http://groups.google.com/group/qubes-devel/browse_thread/thread/66d517a6c67085f9
defect
unbind all network devices fix regexp to cover all pci devs reported by joanna on may utc the regexp used by unbind all network devices script ignores devices that have hex digits a f in their bdf addresses as reported here
1
110,172
13,906,623,893
IssuesEvent
2020-10-20 11:32:02
httpwg/httpbis-issues
https://api.github.com/repos/httpwg/httpbis-issues
opened
List-type header fields vs Set-Cookie (Trac #129)
Incomplete Migration Migrated from Trac design p1-messaging
Migrated from https://trac.ietf.org/ticket/129 ```json { "status": "closed", "changetime": "2012-07-14T13:35:46", "_ts": "1342272946296005", "description": "Part 1, Section 4.2 ([http://tools.ietf.org/html/draft-ietf-httpbis-p1-messaging-03#section-4.2]) states:\n\n\"''Multiple message-header fields with the same field-name MAY be present in a message if and only if the entire field-value for that header field is defined as a comma-separated list [i.e., #(values)]. It MUST be possible to combine the multiple header fields into one \"field-name: field-value\" pair, without changing the semantics of the message, by appending each subsequent field-value to the first, each separated by a comma. The order in which header fields with the same field-name are received is therefore significant to the interpretation of the combined field value, and thus a proxy MUST NOT change the order of these field values when a message is forwarded.''\"\n\nThis is known to be incompatible with Set-Cookie as implemented in practice, as opposed as defined in RFC 2109. \n\nIn particular, Jamie Lokier points out:\n\n> RFC2109 is not implemented by anybody as far as I know.\n> \n> Firstly, cookie _values_ in Set-Cookie may contain a comma which\n> _mustn't_ be quoted because quotes are considered part of the value.\n> When a value is unquoted, RFC2109 says it must match token syntax, but\n> even today that's not conformed to. And RFC2109 doesn't describe an\n> \"expires=\" attribute, but of course nearly all cookies have one, and\n> they don't have the \"max-age=\" attribute with RFC2109 recommands.\n> Finally, as you note, unquoted comma in expires attributes - in fact\n> quoting is not allowed historically for that either.\n> \n> See how many RFC2109 non-compliances you can find in this header I got\n> today from Google, for example.\n> \n>\n{{{\nSet-Cookie: PREF=ID=823cb075fecf6437:TM=1195776675:LM=1195776675:S=WADqk8jBntt5y3gk; expires=Sun, 22-Nov-2009 00:11:15\n}}}\n> \n> (That nobody implements RFC2109 is implied in RFC2965, which obsoletes\n> RFC2109 and in section 9 talks about using Set-Cookie2 alongside\n> Netscape style Set-Cookies, not mentioning RFC2109 style Set-Cookiess.\n> I think this reflects the observation at the time that the change of\n> Set-Cookie syntax promoted in RFC2109 wasn't taken up, probably\n> because it's not backward compatible.)\n\n\nIt seems to me that it would be a service to implementors to minimally add a Note pointing out this special case.\n", "reporter": "julian.reschke@gmx.de", "cc": "", "resolution": "fixed", "time": "2008-08-13T20:05:12", "component": "p1-messaging", "summary": "List-type header fields vs Set-Cookie", "priority": "normal", "keywords": "", "milestone": "unassigned", "owner": "", "type": "design", "severity": "Active WG Document" } ```
1.0
List-type header fields vs Set-Cookie (Trac #129) - Migrated from https://trac.ietf.org/ticket/129 ```json { "status": "closed", "changetime": "2012-07-14T13:35:46", "_ts": "1342272946296005", "description": "Part 1, Section 4.2 ([http://tools.ietf.org/html/draft-ietf-httpbis-p1-messaging-03#section-4.2]) states:\n\n\"''Multiple message-header fields with the same field-name MAY be present in a message if and only if the entire field-value for that header field is defined as a comma-separated list [i.e., #(values)]. It MUST be possible to combine the multiple header fields into one \"field-name: field-value\" pair, without changing the semantics of the message, by appending each subsequent field-value to the first, each separated by a comma. The order in which header fields with the same field-name are received is therefore significant to the interpretation of the combined field value, and thus a proxy MUST NOT change the order of these field values when a message is forwarded.''\"\n\nThis is known to be incompatible with Set-Cookie as implemented in practice, as opposed as defined in RFC 2109. \n\nIn particular, Jamie Lokier points out:\n\n> RFC2109 is not implemented by anybody as far as I know.\n> \n> Firstly, cookie _values_ in Set-Cookie may contain a comma which\n> _mustn't_ be quoted because quotes are considered part of the value.\n> When a value is unquoted, RFC2109 says it must match token syntax, but\n> even today that's not conformed to. And RFC2109 doesn't describe an\n> \"expires=\" attribute, but of course nearly all cookies have one, and\n> they don't have the \"max-age=\" attribute with RFC2109 recommands.\n> Finally, as you note, unquoted comma in expires attributes - in fact\n> quoting is not allowed historically for that either.\n> \n> See how many RFC2109 non-compliances you can find in this header I got\n> today from Google, for example.\n> \n>\n{{{\nSet-Cookie: PREF=ID=823cb075fecf6437:TM=1195776675:LM=1195776675:S=WADqk8jBntt5y3gk; expires=Sun, 22-Nov-2009 00:11:15\n}}}\n> \n> (That nobody implements RFC2109 is implied in RFC2965, which obsoletes\n> RFC2109 and in section 9 talks about using Set-Cookie2 alongside\n> Netscape style Set-Cookies, not mentioning RFC2109 style Set-Cookiess.\n> I think this reflects the observation at the time that the change of\n> Set-Cookie syntax promoted in RFC2109 wasn't taken up, probably\n> because it's not backward compatible.)\n\n\nIt seems to me that it would be a service to implementors to minimally add a Note pointing out this special case.\n", "reporter": "julian.reschke@gmx.de", "cc": "", "resolution": "fixed", "time": "2008-08-13T20:05:12", "component": "p1-messaging", "summary": "List-type header fields vs Set-Cookie", "priority": "normal", "keywords": "", "milestone": "unassigned", "owner": "", "type": "design", "severity": "Active WG Document" } ```
non_defect
list type header fields vs set cookie trac migrated from json status closed changetime ts description part section states n n multiple message header fields with the same field name may be present in a message if and only if the entire field value for that header field is defined as a comma separated list it must be possible to combine the multiple header fields into one field name field value pair without changing the semantics of the message by appending each subsequent field value to the first each separated by a comma the order in which header fields with the same field name are received is therefore significant to the interpretation of the combined field value and thus a proxy must not change the order of these field values when a message is forwarded n nthis is known to be incompatible with set cookie as implemented in practice as opposed as defined in rfc n nin particular jamie lokier points out n n is not implemented by anybody as far as i know n n firstly cookie values in set cookie may contain a comma which n mustn t be quoted because quotes are considered part of the value n when a value is unquoted says it must match token syntax but n even today that s not conformed to and doesn t describe an n expires attribute but of course nearly all cookies have one and n they don t have the max age attribute with recommands n finally as you note unquoted comma in expires attributes in fact n quoting is not allowed historically for that either n n see how many non compliances you can find in this header i got n today from google for example n n n nset cookie pref id tm lm s expires sun nov n n n that nobody implements is implied in which obsoletes n and in section talks about using set alongside n netscape style set cookies not mentioning style set cookiess n i think this reflects the observation at the time that the change of n set cookie syntax promoted in wasn t taken up probably n because it s not backward compatible n n nit seems to me that it would be a service to implementors to minimally add a note pointing out this special case n reporter julian reschke gmx de cc resolution fixed time component messaging summary list type header fields vs set cookie priority normal keywords milestone unassigned owner type design severity active wg document
0
28,355
2,701,067,469
IssuesEvent
2015-04-04 21:38:50
TrinityCore/TrinityCore
https://api.github.com/repos/TrinityCore/TrinityCore
closed
[DB|NPC] Vhel'kur - Missing Waypoints
Comp-Database Feedback-PatchFix Priority-Cosmetic
http://de.wowhead.com/npc=21801 As you can see, this npc should have some waypoints and 'flys' in a circle. If someoone profide some sniffs for this creature, i would copypaste it. <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/4246515-db-npc-vhel-kur-missing-waypoints?utm_campaign=plugin&utm_content=tracker%2F1310&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F1310&utm_medium=issues&utm_source=github). </bountysource-plugin>
1.0
[DB|NPC] Vhel'kur - Missing Waypoints - http://de.wowhead.com/npc=21801 As you can see, this npc should have some waypoints and 'flys' in a circle. If someoone profide some sniffs for this creature, i would copypaste it. <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/4246515-db-npc-vhel-kur-missing-waypoints?utm_campaign=plugin&utm_content=tracker%2F1310&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F1310&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_defect
vhel kur missing waypoints as you can see this npc should have some waypoints and flys in a circle if someoone profide some sniffs for this creature i would copypaste it want to back this issue we accept bounties via
0
17,023
2,966,754,139
IssuesEvent
2015-07-12 07:03:58
ElvisChiang/welly
https://api.github.com/repos/ElvisChiang/welly
closed
站台無法編輯
auto-migrated Priority-Medium Type-Defect
``` 如圖所示 原因不明的無法編輯站台 圖片中可以看出目前有四個儲存的站台 但是站台列表當中卻無列出任何東西,也無法選取 這個問題從OSX 10.8~10.9都一樣(我從10.8開始加入MAC家族的) 希望能解決這個BUG 謝謝 ``` Original issue reported on code.google.com by `CHCOO...@gmail.com` on 6 Dec 2013 at 8:58 Attachments: * [螢幕快照 2013-12-06 下午3.57.05.png](https://storage.googleapis.com/google-code-attachments/welly/issue-69/comment-0/螢幕快照 2013-12-06 下午3.57.05.png)
1.0
站台無法編輯 - ``` 如圖所示 原因不明的無法編輯站台 圖片中可以看出目前有四個儲存的站台 但是站台列表當中卻無列出任何東西,也無法選取 這個問題從OSX 10.8~10.9都一樣(我從10.8開始加入MAC家族的) 希望能解決這個BUG 謝謝 ``` Original issue reported on code.google.com by `CHCOO...@gmail.com` on 6 Dec 2013 at 8:58 Attachments: * [螢幕快照 2013-12-06 下午3.57.05.png](https://storage.googleapis.com/google-code-attachments/welly/issue-69/comment-0/螢幕快照 2013-12-06 下午3.57.05.png)
defect
站台無法編輯 如圖所示 原因不明的無法編輯站台 圖片中可以看出目前有四個儲存的站台 但是站台列表當中卻無列出任何東西,也無法選取 這個問題從osx ( ) 希望能解決這個bug 謝謝 original issue reported on code google com by chcoo gmail com on dec at attachments png
1
46,097
13,055,851,745
IssuesEvent
2020-07-30 02:55:35
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
reported issues with return values of tray.Execute(). (Trac #579)
IceTray Incomplete Migration Migrated from Trac defect
Migrated from https://code.icecube.wisc.edu/ticket/579 ```json { "status": "closed", "changetime": "2009-11-24T14:58:23", "description": "As reported by David and Martin:\n\nHi Erik,\n\nNo, Martin's script ran without fatals. And without fatals, he should get normal, complete output files, and that is not what was happening, because flat-ntuple (the module he added to his script in order to study the results) calls TTree::Write() during the Finish() method, not in its destructor. So that is why it is important to call tray.Finish() after tray.Execute() terminated normally.\n\nDavid\n\nOn Tue, 24 Nov 2009, Erik Blaufuss wrote:\n\nHi David, Martin\n\nThat's a good suggestion, but why he's only seeing\nincomplete files, etc when tray.Execute() has a fatal call?\n\nWhy should there be any fatal calls during normal processing and\nwhy would you expect sensible content in a file after a log_fatal()\nwas generated somewhere?\n\nI didn't put too much thought into the bottom of that\nscript (copied from Robert F in reality), since I\nwas more concerned about the settings for the modules...\n\nLater\n-Erik\nOn Nov 24, 2009, at 8:13 AM, David Boersma wrote:\n\nHi Erik,\n\nMartin Bissok, PhD student here in Aachen, ran into some weird issues with incomplete/empty ROOT files, which we eventually traced down to this construction at the end of his python script:\n\ntry:\n tray.Execute()\nexcept Exception, ex:\n print ex\n tray.Finish()\n sys.exit(1)\n\nThis was copied from the official IC77 script:\nhttp://code.icecube.wisc.edu/svn/sandbox/blaufuss/ic77-retrig/base_filter_modules.py\nwhich Martin had used as a starting point to develop his GC filter work.\n\nThe error here is that tray.Finish() *should* also be run when tray.Execute() returns normally. Otherwise the Finish() method of modules does not get called, potentially leading to unfinished business like not properly closed files, unflushed buffers and what not. When we copied a tray.Finish() immediately after the Execute() line, Martin's ROOT files got the expected content. So, like this:\n\ntry:\n tray.Execute()\n tray.Finish()\nexcept Exception, ex:\n print ex\n tray.Finish()\n sys.exit(1)\n\nThough actually I'm not entirely certain if tray.Finish() is really useful after e.g. a log_fatal() was thrown or a segfault occurred. Maybe the tray.Finish() really wasn't meant to be in the except: leg, but intended to have been copied into the try: leg all along?\n\nDavid\n", "reporter": "blaufuss", "cc": "", "resolution": "invalid", "_ts": "1259074703000000", "component": "IceTray", "summary": "reported issues with return values of tray.Execute().", "priority": "normal", "keywords": "", "time": "2009-11-24T14:55:13", "milestone": "", "owner": "troy", "type": "defect" } ```
1.0
reported issues with return values of tray.Execute(). (Trac #579) - Migrated from https://code.icecube.wisc.edu/ticket/579 ```json { "status": "closed", "changetime": "2009-11-24T14:58:23", "description": "As reported by David and Martin:\n\nHi Erik,\n\nNo, Martin's script ran without fatals. And without fatals, he should get normal, complete output files, and that is not what was happening, because flat-ntuple (the module he added to his script in order to study the results) calls TTree::Write() during the Finish() method, not in its destructor. So that is why it is important to call tray.Finish() after tray.Execute() terminated normally.\n\nDavid\n\nOn Tue, 24 Nov 2009, Erik Blaufuss wrote:\n\nHi David, Martin\n\nThat's a good suggestion, but why he's only seeing\nincomplete files, etc when tray.Execute() has a fatal call?\n\nWhy should there be any fatal calls during normal processing and\nwhy would you expect sensible content in a file after a log_fatal()\nwas generated somewhere?\n\nI didn't put too much thought into the bottom of that\nscript (copied from Robert F in reality), since I\nwas more concerned about the settings for the modules...\n\nLater\n-Erik\nOn Nov 24, 2009, at 8:13 AM, David Boersma wrote:\n\nHi Erik,\n\nMartin Bissok, PhD student here in Aachen, ran into some weird issues with incomplete/empty ROOT files, which we eventually traced down to this construction at the end of his python script:\n\ntry:\n tray.Execute()\nexcept Exception, ex:\n print ex\n tray.Finish()\n sys.exit(1)\n\nThis was copied from the official IC77 script:\nhttp://code.icecube.wisc.edu/svn/sandbox/blaufuss/ic77-retrig/base_filter_modules.py\nwhich Martin had used as a starting point to develop his GC filter work.\n\nThe error here is that tray.Finish() *should* also be run when tray.Execute() returns normally. Otherwise the Finish() method of modules does not get called, potentially leading to unfinished business like not properly closed files, unflushed buffers and what not. When we copied a tray.Finish() immediately after the Execute() line, Martin's ROOT files got the expected content. So, like this:\n\ntry:\n tray.Execute()\n tray.Finish()\nexcept Exception, ex:\n print ex\n tray.Finish()\n sys.exit(1)\n\nThough actually I'm not entirely certain if tray.Finish() is really useful after e.g. a log_fatal() was thrown or a segfault occurred. Maybe the tray.Finish() really wasn't meant to be in the except: leg, but intended to have been copied into the try: leg all along?\n\nDavid\n", "reporter": "blaufuss", "cc": "", "resolution": "invalid", "_ts": "1259074703000000", "component": "IceTray", "summary": "reported issues with return values of tray.Execute().", "priority": "normal", "keywords": "", "time": "2009-11-24T14:55:13", "milestone": "", "owner": "troy", "type": "defect" } ```
defect
reported issues with return values of tray execute trac migrated from json status closed changetime description as reported by david and martin n nhi erik n nno martin s script ran without fatals and without fatals he should get normal complete output files and that is not what was happening because flat ntuple the module he added to his script in order to study the results calls ttree write during the finish method not in its destructor so that is why it is important to call tray finish after tray execute terminated normally n ndavid n non tue nov erik blaufuss wrote n nhi david martin n nthat s a good suggestion but why he s only seeing nincomplete files etc when tray execute has a fatal call n nwhy should there be any fatal calls during normal processing and nwhy would you expect sensible content in a file after a log fatal nwas generated somewhere n ni didn t put too much thought into the bottom of that nscript copied from robert f in reality since i nwas more concerned about the settings for the modules n nlater n erik non nov at am david boersma wrote n nhi erik n nmartin bissok phd student here in aachen ran into some weird issues with incomplete empty root files which we eventually traced down to this construction at the end of his python script n ntry n tray execute nexcept exception ex n print ex n tray finish n sys exit n nthis was copied from the official script n martin had used as a starting point to develop his gc filter work n nthe error here is that tray finish should also be run when tray execute returns normally otherwise the finish method of modules does not get called potentially leading to unfinished business like not properly closed files unflushed buffers and what not when we copied a tray finish immediately after the execute line martin s root files got the expected content so like this n ntry n tray execute n tray finish nexcept exception ex n print ex n tray finish n sys exit n nthough actually i m not entirely certain if tray finish is really useful after e g a log fatal was thrown or a segfault occurred maybe the tray finish really wasn t meant to be in the except leg but intended to have been copied into the try leg all along n ndavid n reporter blaufuss cc resolution invalid ts component icetray summary reported issues with return values of tray execute priority normal keywords time milestone owner troy type defect
1
23,406
3,813,104,726
IssuesEvent
2016-03-28 01:52:41
huboard/huboard
https://api.github.com/repos/huboard/huboard
closed
Important display problems since we installed HB:E v1.6.107
3 - In Review Defect Enterprise ★★★
We are experiencing the following problems: - [ ] Tasks columns display is not fitting the screen - [ ] Columns width are not always the same (this cause the problem above) - [ ] The X that was visible on a filter when selected is not anymore displayed We'll have to roll-out to the previous release as these display problems are too much disrupting unless this can be fixed asap. <!--- @huboard:{"order":3.161309661618673e-10,"milestone_order":631,"custom_state":""} --> <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/30589749-important-display-problems-since-we-installed-hb-e-v1-6-107?utm_campaign=plugin&utm_content=tracker%2F282362&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F282362&utm_medium=issues&utm_source=github). </bountysource-plugin>
1.0
Important display problems since we installed HB:E v1.6.107 - We are experiencing the following problems: - [ ] Tasks columns display is not fitting the screen - [ ] Columns width are not always the same (this cause the problem above) - [ ] The X that was visible on a filter when selected is not anymore displayed We'll have to roll-out to the previous release as these display problems are too much disrupting unless this can be fixed asap. <!--- @huboard:{"order":3.161309661618673e-10,"milestone_order":631,"custom_state":""} --> <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/30589749-important-display-problems-since-we-installed-hb-e-v1-6-107?utm_campaign=plugin&utm_content=tracker%2F282362&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F282362&utm_medium=issues&utm_source=github). </bountysource-plugin>
defect
important display problems since we installed hb e we are experiencing the following problems tasks columns display is not fitting the screen columns width are not always the same this cause the problem above the x that was visible on a filter when selected is not anymore displayed we ll have to roll out to the previous release as these display problems are too much disrupting unless this can be fixed asap huboard order milestone order custom state want to back this issue we accept bounties via
1
1,108
2,856,001,064
IssuesEvent
2015-06-02 13:00:32
zaphoyd/websocketpp
https://api.github.com/repos/zaphoyd/websocketpp
closed
Two SSL records for one message.
iostream Transport Performance Issue
Hello, I wrote a programm using Websocket++, what exchange mesages between client and server. I didn't use any embeded transports, but use library in iostream manner, handling send and receive by myself. Whenewer I call client.send(), I got two writes in my write handle, set by set_write_handler(). First for hrader, and second - for message body. As a result, I got two SSL records in channel. What produces unnecessary overhead. Two bytes of header results in may bytes of SSL record. I, as a library user, would prefer to get in write handle complete message, with header, body and footer (if any), so I can wrap it into single SSL record.
True
Two SSL records for one message. - Hello, I wrote a programm using Websocket++, what exchange mesages between client and server. I didn't use any embeded transports, but use library in iostream manner, handling send and receive by myself. Whenewer I call client.send(), I got two writes in my write handle, set by set_write_handler(). First for hrader, and second - for message body. As a result, I got two SSL records in channel. What produces unnecessary overhead. Two bytes of header results in may bytes of SSL record. I, as a library user, would prefer to get in write handle complete message, with header, body and footer (if any), so I can wrap it into single SSL record.
non_defect
two ssl records for one message hello i wrote a programm using websocket what exchange mesages between client and server i didn t use any embeded transports but use library in iostream manner handling send and receive by myself whenewer i call client send i got two writes in my write handle set by set write handler first for hrader and second for message body as a result i got two ssl records in channel what produces unnecessary overhead two bytes of header results in may bytes of ssl record i as a library user would prefer to get in write handle complete message with header body and footer if any so i can wrap it into single ssl record
0
100,562
30,724,406,268
IssuesEvent
2023-07-27 18:25:14
apache/beam
https://api.github.com/repos/apache/beam
closed
./gradlew :beam-runners-google-cloud-dataflow-java:validatesRunner fails due to invalid GCS path
build P3 bug
The default root is gs://temp-storage-for-end-to-end-tests/ and the code adds a slash and then the test case name, so it results in two slashes in a row, rejected by the filesystem implementation. Imported from Jira [BEAM-6715](https://issues.apache.org/jira/browse/BEAM-6715). Original Jira may contain additional context. Reported by: kenn.
1.0
./gradlew :beam-runners-google-cloud-dataflow-java:validatesRunner fails due to invalid GCS path - The default root is gs://temp-storage-for-end-to-end-tests/ and the code adds a slash and then the test case name, so it results in two slashes in a row, rejected by the filesystem implementation. Imported from Jira [BEAM-6715](https://issues.apache.org/jira/browse/BEAM-6715). Original Jira may contain additional context. Reported by: kenn.
non_defect
gradlew beam runners google cloud dataflow java validatesrunner fails due to invalid gcs path the default root is gs temp storage for end to end tests and the code adds a slash and then the test case name so it results in two slashes in a row rejected by the filesystem implementation imported from jira original jira may contain additional context reported by kenn
0
217,689
24,348,934,982
IssuesEvent
2022-10-02 17:48:47
venkateshreddypala/AngOCR
https://api.github.com/repos/venkateshreddypala/AngOCR
closed
CVE-2020-7598 (Medium) detected in minimist-0.0.8.tgz, minimist-1.2.0.tgz - autoclosed
security vulnerability
## CVE-2020-7598 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary> <p> <details><summary><b>minimist-0.0.8.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p> <p>Path to dependency file: /AngOCR/ui/package.json</p> <p>Path to vulnerable library: /ui/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - karma-3.1.1.tgz (Root Library) - optimist-0.6.1.tgz - :x: **minimist-0.0.8.tgz** (Vulnerable Library) </details> <details><summary><b>minimist-1.2.0.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p> <p>Path to dependency file: /AngOCR/ui/package.json</p> <p>Path to vulnerable library: /ui/node_modules/@angular/compiler-cli/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - ts-node-4.1.0.tgz (Root Library) - :x: **minimist-1.2.0.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload. <p>Publish Date: 2020-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-03-11</p> <p>Fix Resolution (minimist): 0.2.1</p> <p>Direct dependency fix Resolution (karma): 5.0.0</p><p>Fix Resolution (minimist): 1.2.3</p> <p>Direct dependency fix Resolution (ts-node): 5.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7598 (Medium) detected in minimist-0.0.8.tgz, minimist-1.2.0.tgz - autoclosed - ## CVE-2020-7598 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary> <p> <details><summary><b>minimist-0.0.8.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p> <p>Path to dependency file: /AngOCR/ui/package.json</p> <p>Path to vulnerable library: /ui/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - karma-3.1.1.tgz (Root Library) - optimist-0.6.1.tgz - :x: **minimist-0.0.8.tgz** (Vulnerable Library) </details> <details><summary><b>minimist-1.2.0.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p> <p>Path to dependency file: /AngOCR/ui/package.json</p> <p>Path to vulnerable library: /ui/node_modules/@angular/compiler-cli/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - ts-node-4.1.0.tgz (Root Library) - :x: **minimist-1.2.0.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload. <p>Publish Date: 2020-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-03-11</p> <p>Fix Resolution (minimist): 0.2.1</p> <p>Direct dependency fix Resolution (karma): 5.0.0</p><p>Fix Resolution (minimist): 1.2.3</p> <p>Direct dependency fix Resolution (ts-node): 5.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in minimist tgz minimist tgz autoclosed cve medium severity vulnerability vulnerable libraries minimist tgz minimist tgz minimist tgz parse argument options library home page a href path to dependency file angocr ui package json path to vulnerable library ui node modules minimist package json dependency hierarchy karma tgz root library optimist tgz x minimist tgz vulnerable library minimist tgz parse argument options library home page a href path to dependency file angocr ui package json path to vulnerable library ui node modules angular compiler cli node modules minimist package json dependency hierarchy ts node tgz root library x minimist tgz vulnerable library vulnerability details minimist before could be tricked into adding or modifying properties of object prototype using a constructor or proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution minimist direct dependency fix resolution karma fix resolution minimist direct dependency fix resolution ts node step up your open source security game with mend
0
780,747
27,406,713,012
IssuesEvent
2023-03-01 07:31:14
prgrms-web-devcourse/Team-Kkini-Mukvengers-FE
https://api.github.com/repos/prgrms-web-devcourse/Team-Kkini-Mukvengers-FE
opened
google maps의 remote image를 사용하는 img 태그 최적화
Priority: Medium Perf
## 📕 작업 설명 > google maps의 remote image를 사용하는 img 태그 최적화 ## 📖 To-Do list - [ ] Pre Loading - [ ] webp 변환 - remote image라서 가능할지 안할지 확실치 않음. - [ ] srcSet, sizes 속성 적용
1.0
google maps의 remote image를 사용하는 img 태그 최적화 - ## 📕 작업 설명 > google maps의 remote image를 사용하는 img 태그 최적화 ## 📖 To-Do list - [ ] Pre Loading - [ ] webp 변환 - remote image라서 가능할지 안할지 확실치 않음. - [ ] srcSet, sizes 속성 적용
non_defect
google maps의 remote image를 사용하는 img 태그 최적화 📕 작업 설명 google maps의 remote image를 사용하는 img 태그 최적화 📖 to do list pre loading webp 변환 remote image라서 가능할지 안할지 확실치 않음 srcset sizes 속성 적용
0
25,570
4,386,598,378
IssuesEvent
2016-08-08 13:23:26
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
closed
Carousel page links redirect to #
defect
In showcase, carousel page links are not working. Currently they redirect to http://www.primefaces.org/primeng/# Instead of changing carousel page. I tested showcase http://www.primefaces.org/primeng/#/carousel with Firefox and Chrome and both browsers suffered from the same issue.
1.0
Carousel page links redirect to # - In showcase, carousel page links are not working. Currently they redirect to http://www.primefaces.org/primeng/# Instead of changing carousel page. I tested showcase http://www.primefaces.org/primeng/#/carousel with Firefox and Chrome and both browsers suffered from the same issue.
defect
carousel page links redirect to in showcase carousel page links are not working currently they redirect to instead of changing carousel page i tested showcase with firefox and chrome and both browsers suffered from the same issue
1
23,464
3,829,548,670
IssuesEvent
2016-03-31 11:11:18
pearu/pylibnidaqmx
https://api.github.com/repos/pearu/pylibnidaqmx
closed
Add AnalogOutput status calls
auto-migrated Priority-Medium Type-Defect
``` I'm attaching a patch that adds the GetWriteCurrWritePos() and GetWriteTotalSampPerChanGenerated() calls. These are useful for getting the status of an AnalogOutput task while it's running. I put them in the Task base object. Perhaps these should be split into AnalogOutput and the corresponding GetRead calls should be in AnalogInput. Thanks, Mark diff --git a/nidaqmx/libnidaqmx.py b/nidaqmx/libnidaqmx.py index a349b72..d7e86d4 100644 --- a/nidaqmx/libnidaqmx.py +++ b/nidaqmx/libnidaqmx.py @@ -2556,6 +2556,26 @@ class Task(uInt32): CALL('GetReadTotalSampPerChanAcquired', self, ctypes.byref(d)) return d.value + def get_write_current_position(self): + """ + Indicates the position in the buffer of the next sample to generate. + This value is the same for all channels in the task. + """ + d = uInt64(0) + CALL('GetWriteCurrWritePos', self, ctypes.byref(d)) + return d.value + + def get_write_samples_per_channel_generated(self): + """ + Indicates the total number of samples generated by each + channel. NI-DAQmx returns a single value because this value is + the same for all channels. + """ + d = uInt64(0) + CALL('GetWriteTotalSampPerChanGenerated', self, ctypes.byref(d)) + return d.value + + def wait_until_done(self, timeout=-1): """ Waits for the measurement or generation to complete. Use this ``` Original issue reported on code.google.com by `mhis...@gmail.com` on 29 Apr 2014 at 9:51
1.0
Add AnalogOutput status calls - ``` I'm attaching a patch that adds the GetWriteCurrWritePos() and GetWriteTotalSampPerChanGenerated() calls. These are useful for getting the status of an AnalogOutput task while it's running. I put them in the Task base object. Perhaps these should be split into AnalogOutput and the corresponding GetRead calls should be in AnalogInput. Thanks, Mark diff --git a/nidaqmx/libnidaqmx.py b/nidaqmx/libnidaqmx.py index a349b72..d7e86d4 100644 --- a/nidaqmx/libnidaqmx.py +++ b/nidaqmx/libnidaqmx.py @@ -2556,6 +2556,26 @@ class Task(uInt32): CALL('GetReadTotalSampPerChanAcquired', self, ctypes.byref(d)) return d.value + def get_write_current_position(self): + """ + Indicates the position in the buffer of the next sample to generate. + This value is the same for all channels in the task. + """ + d = uInt64(0) + CALL('GetWriteCurrWritePos', self, ctypes.byref(d)) + return d.value + + def get_write_samples_per_channel_generated(self): + """ + Indicates the total number of samples generated by each + channel. NI-DAQmx returns a single value because this value is + the same for all channels. + """ + d = uInt64(0) + CALL('GetWriteTotalSampPerChanGenerated', self, ctypes.byref(d)) + return d.value + + def wait_until_done(self, timeout=-1): """ Waits for the measurement or generation to complete. Use this ``` Original issue reported on code.google.com by `mhis...@gmail.com` on 29 Apr 2014 at 9:51
defect
add analogoutput status calls i m attaching a patch that adds the getwritecurrwritepos and getwritetotalsampperchangenerated calls these are useful for getting the status of an analogoutput task while it s running i put them in the task base object perhaps these should be split into analogoutput and the corresponding getread calls should be in analoginput thanks mark diff git a nidaqmx libnidaqmx py b nidaqmx libnidaqmx py index a nidaqmx libnidaqmx py b nidaqmx libnidaqmx py class task call getreadtotalsampperchanacquired self ctypes byref d return d value def get write current position self indicates the position in the buffer of the next sample to generate this value is the same for all channels in the task d call getwritecurrwritepos self ctypes byref d return d value def get write samples per channel generated self indicates the total number of samples generated by each channel ni daqmx returns a single value because this value is the same for all channels d call getwritetotalsampperchangenerated self ctypes byref d return d value def wait until done self timeout waits for the measurement or generation to complete use this original issue reported on code google com by mhis gmail com on apr at
1
16,430
2,894,559,830
IssuesEvent
2015-06-16 00:50:19
obophenotype/cell-ontology
https://api.github.com/repos/obophenotype/cell-ontology
closed
Move repo to github
auto-migrated Priority-Medium Type-Defect
``` google code will probably go away at some point in the future. It won't be missed by me. We should probably think about moving in the next 12 months. Issues: * tracker migration (good incentive to go through and close everything?) * ensuring no barriers of entry to non git-experts (todo: test github's svn bridge) ``` Original issue reported on code.google.com by `cmung...@gmail.com` on 31 Mar 2014 at 2:29
1.0
Move repo to github - ``` google code will probably go away at some point in the future. It won't be missed by me. We should probably think about moving in the next 12 months. Issues: * tracker migration (good incentive to go through and close everything?) * ensuring no barriers of entry to non git-experts (todo: test github's svn bridge) ``` Original issue reported on code.google.com by `cmung...@gmail.com` on 31 Mar 2014 at 2:29
defect
move repo to github google code will probably go away at some point in the future it won t be missed by me we should probably think about moving in the next months issues tracker migration good incentive to go through and close everything ensuring no barriers of entry to non git experts todo test github s svn bridge original issue reported on code google com by cmung gmail com on mar at
1
4,283
2,610,090,817
IssuesEvent
2015-02-26 18:27:28
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
深圳痘痘治疗
auto-migrated Priority-Medium Type-Defect
``` 深圳痘痘治疗【深圳韩方科颜全国热线400-869-1818,24小时QQ4008 691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方—�� �韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科颜� ��业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康祛 痘技术并结合先进“先进豪华彩光”仪,开创国内专业治疗�� �刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:44
1.0
深圳痘痘治疗 - ``` 深圳痘痘治疗【深圳韩方科颜全国热线400-869-1818,24小时QQ4008 691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方—�� �韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科颜� ��业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康祛 痘技术并结合先进“先进豪华彩光”仪,开创国内专业治疗�� �刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:44
defect
深圳痘痘治疗 深圳痘痘治疗【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方—�� �韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科颜� ��业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康祛 痘技术并结合先进“先进豪华彩光”仪,开创国内专业治疗�� �刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘。 original issue reported on code google com by szft com on may at
1
821,269
30,813,811,136
IssuesEvent
2023-08-01 12:16:18
wp-media/wp-rocket
https://api.github.com/repos/wp-media/wp-rocket
opened
Used CSS isnot automatically generated for home page when switch theme while preload is off
type: bug priority: low severity: moderate module: remove unused css
**Before submitting an issue please check that you’ve completed the following steps:** - Made sure you’re on the latest version => 3.14.2.1 - Used the search feature to ensure that the bug hasn’t been reported before **Describe the bug** When we switch theme while preload is off, we display a notice that used CSS was created for the home page although it was not **To Reproduce** Steps to reproduce the behavior: 1. Disable preload 2. Enable RUCSS 3. Visit the home page => used CSS is there 4. switch the theme and check the used CSS table **Expected behavior** Entry is added for the home page **Additional context** Add any other context about the problem here. - It was working fine on 3.12.6.1 **Backlog Grooming (for WP Media dev team use only)** - [ ] Reproduce the problem - [ ] Identify the root cause - [ ] Scope a solution - [ ] Estimate the effort
1.0
Used CSS isnot automatically generated for home page when switch theme while preload is off - **Before submitting an issue please check that you’ve completed the following steps:** - Made sure you’re on the latest version => 3.14.2.1 - Used the search feature to ensure that the bug hasn’t been reported before **Describe the bug** When we switch theme while preload is off, we display a notice that used CSS was created for the home page although it was not **To Reproduce** Steps to reproduce the behavior: 1. Disable preload 2. Enable RUCSS 3. Visit the home page => used CSS is there 4. switch the theme and check the used CSS table **Expected behavior** Entry is added for the home page **Additional context** Add any other context about the problem here. - It was working fine on 3.12.6.1 **Backlog Grooming (for WP Media dev team use only)** - [ ] Reproduce the problem - [ ] Identify the root cause - [ ] Scope a solution - [ ] Estimate the effort
non_defect
used css isnot automatically generated for home page when switch theme while preload is off before submitting an issue please check that you’ve completed the following steps made sure you’re on the latest version used the search feature to ensure that the bug hasn’t been reported before describe the bug when we switch theme while preload is off we display a notice that used css was created for the home page although it was not to reproduce steps to reproduce the behavior disable preload enable rucss visit the home page used css is there switch the theme and check the used css table expected behavior entry is added for the home page additional context add any other context about the problem here it was working fine on backlog grooming for wp media dev team use only reproduce the problem identify the root cause scope a solution estimate the effort
0
94,533
27,226,327,869
IssuesEvent
2023-02-21 09:58:26
tensorflow/tfjs
https://api.github.com/repos/tensorflow/tfjs
opened
Docker M2 Macbook ARM invalid ELF header
type:build/install
I use a M2 MacBook. We integrate tensorflow/tfjs into our nestJS backend. When I start the nestJS enviorment for itself, everthing is fine, and tensorflow works. Wenn I use Docker to deploy it on a DigitalOcean Dropet, everything is fine and tensorflow works. **Describe the problem** When i use the same Dockerfile (which i use to deploy on digital ocean) to start it local on my M2, it crashs. `NODE_OPTIONS='--max-old-space-size=3000' nest start 2023-02-20 18:21:41 Error: /usr/src/app/node_modules/@tensorflow/tfjs-node/lib/napi-v8/tfjs_binding.node: invalid ELF header 2023-02-20 18:21:41 at Object.Module._extensions..node (node:internal/modules/cjs/loader:1302:18) 2023-02-20 18:21:41 at Module.load (node:internal/modules/cjs/loader:1081:32) 2023-02-20 18:21:41 at Function.Module._load (node:internal/modules/cjs/loader:922:12) 2023-02-20 18:21:41 at Module.require (node:internal/modules/cjs/loader:1105:19) 2023-02-20 18:21:41 at require (node:internal/modules/cjs/helpers:103:18) 2023-02-20 18:21:41 at Object.<anonymous> (/usr/src/app/node_modules/@tensorflow/tfjs-node/dist/index.js:72:16) 2023-02-20 18:21:41 at Module._compile (node:internal/modules/cjs/loader:1218:14) 2023-02-20 18:21:41 at Object.Module._extensions..js (node:internal/modules/cjs/loader:1272:10) 2023-02-20 18:21:41 at Module.load (node:internal/modules/cjs/loader:1081:32) 2023-02-20 18:21:41 at Function.Module._load (node:internal/modules/cjs/loader:922:12) 2023-02-20 18:21:41 error Command failed with exit code 1.` ### Dockerfile `FROM node:18.12.1 EXPOSE 8080 WORKDIR /usr/src/app COPY package*.json ./ COPY yarn.lock ./ RUN yarn COPY . . CMD [ "yarn","start"] ` ### These were my previous solutions that did not work 1. Add ` platform: linux/amd64` to the correct place in the docker-compose file, whit this result <img width="118" alt="image" src="https://user-images.githubusercontent.com/125969421/220311961-6415f84c-2628-4408-85af-08befaa5e789.png"> Thanks for helping me :)
1.0
Docker M2 Macbook ARM invalid ELF header - I use a M2 MacBook. We integrate tensorflow/tfjs into our nestJS backend. When I start the nestJS enviorment for itself, everthing is fine, and tensorflow works. Wenn I use Docker to deploy it on a DigitalOcean Dropet, everything is fine and tensorflow works. **Describe the problem** When i use the same Dockerfile (which i use to deploy on digital ocean) to start it local on my M2, it crashs. `NODE_OPTIONS='--max-old-space-size=3000' nest start 2023-02-20 18:21:41 Error: /usr/src/app/node_modules/@tensorflow/tfjs-node/lib/napi-v8/tfjs_binding.node: invalid ELF header 2023-02-20 18:21:41 at Object.Module._extensions..node (node:internal/modules/cjs/loader:1302:18) 2023-02-20 18:21:41 at Module.load (node:internal/modules/cjs/loader:1081:32) 2023-02-20 18:21:41 at Function.Module._load (node:internal/modules/cjs/loader:922:12) 2023-02-20 18:21:41 at Module.require (node:internal/modules/cjs/loader:1105:19) 2023-02-20 18:21:41 at require (node:internal/modules/cjs/helpers:103:18) 2023-02-20 18:21:41 at Object.<anonymous> (/usr/src/app/node_modules/@tensorflow/tfjs-node/dist/index.js:72:16) 2023-02-20 18:21:41 at Module._compile (node:internal/modules/cjs/loader:1218:14) 2023-02-20 18:21:41 at Object.Module._extensions..js (node:internal/modules/cjs/loader:1272:10) 2023-02-20 18:21:41 at Module.load (node:internal/modules/cjs/loader:1081:32) 2023-02-20 18:21:41 at Function.Module._load (node:internal/modules/cjs/loader:922:12) 2023-02-20 18:21:41 error Command failed with exit code 1.` ### Dockerfile `FROM node:18.12.1 EXPOSE 8080 WORKDIR /usr/src/app COPY package*.json ./ COPY yarn.lock ./ RUN yarn COPY . . CMD [ "yarn","start"] ` ### These were my previous solutions that did not work 1. Add ` platform: linux/amd64` to the correct place in the docker-compose file, whit this result <img width="118" alt="image" src="https://user-images.githubusercontent.com/125969421/220311961-6415f84c-2628-4408-85af-08befaa5e789.png"> Thanks for helping me :)
non_defect
docker macbook arm invalid elf header i use a macbook we integrate tensorflow tfjs into our nestjs backend when i start the nestjs enviorment for itself everthing is fine and tensorflow works wenn i use docker to deploy it on a digitalocean dropet everything is fine and tensorflow works describe the problem when i use the same dockerfile which i use to deploy on digital ocean to start it local on my it crashs node options max old space size nest start error usr src app node modules tensorflow tfjs node lib napi tfjs binding node invalid elf header at object module extensions node node internal modules cjs loader at module load node internal modules cjs loader at function module load node internal modules cjs loader at module require node internal modules cjs loader at require node internal modules cjs helpers at object usr src app node modules tensorflow tfjs node dist index js at module compile node internal modules cjs loader at object module extensions js node internal modules cjs loader at module load node internal modules cjs loader at function module load node internal modules cjs loader error command failed with exit code dockerfile from node expose workdir usr src app copy package json copy yarn lock run yarn copy cmd these were my previous solutions that did not work add platform linux to the correct place in the docker compose file whit this result img width alt image src thanks for helping me
0
53,442
13,261,617,680
IssuesEvent
2020-08-20 20:13:36
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[DomTools] Bring back I3IsolatedHitsCutModule (Trac #1397)
Migrated from Trac combo core defect
Or at least provide an alternative for filter scripts IceCube_BaseProc.py. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1397">https://code.icecube.wisc.edu/projects/icecube/ticket/1397</a>, reported by olivasand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:14:07", "_ts": "1458335647931556", "description": "Or at least provide an alternative for filter scripts IceCube_BaseProc.py.", "reporter": "olivas", "cc": "", "resolution": "fixed", "time": "2015-10-14T07:06:54", "component": "combo core", "summary": "[DomTools] Bring back I3IsolatedHitsCutModule", "priority": "blocker", "keywords": "", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
[DomTools] Bring back I3IsolatedHitsCutModule (Trac #1397) - Or at least provide an alternative for filter scripts IceCube_BaseProc.py. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1397">https://code.icecube.wisc.edu/projects/icecube/ticket/1397</a>, reported by olivasand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:14:07", "_ts": "1458335647931556", "description": "Or at least provide an alternative for filter scripts IceCube_BaseProc.py.", "reporter": "olivas", "cc": "", "resolution": "fixed", "time": "2015-10-14T07:06:54", "component": "combo core", "summary": "[DomTools] Bring back I3IsolatedHitsCutModule", "priority": "blocker", "keywords": "", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
bring back trac or at least provide an alternative for filter scripts icecube baseproc py migrated from json status closed changetime ts description or at least provide an alternative for filter scripts icecube baseproc py reporter olivas cc resolution fixed time component combo core summary bring back priority blocker keywords milestone owner olivas type defect
1
959
2,594,375,625
IssuesEvent
2015-02-20 02:32:23
BALL-Project/ball
https://api.github.com/repos/BALL-Project/ball
closed
Changing RTfact->OpenGL in Continuous Loop mode results in crash
C: VIEW P: major R: fixed T: defect
**Reported by nicste on 12 Jun 41056064 04:55 UTC** Changing from RTfact rendering mode to OpenGL rendering mode while having the "Continuous Loop" enabled results in a crash of BALLView
1.0
Changing RTfact->OpenGL in Continuous Loop mode results in crash - **Reported by nicste on 12 Jun 41056064 04:55 UTC** Changing from RTfact rendering mode to OpenGL rendering mode while having the "Continuous Loop" enabled results in a crash of BALLView
defect
changing rtfact opengl in continuous loop mode results in crash reported by nicste on jun utc changing from rtfact rendering mode to opengl rendering mode while having the continuous loop enabled results in a crash of ballview
1
75,781
9,320,813,996
IssuesEvent
2019-03-27 00:58:41
JordanMartinez/purescript-jordans-reference
https://api.github.com/repos/JordanMartinez/purescript-jordans-reference
opened
Wrap a 'Show' type class in a newtype
Design-Patterns enhancement
Probably a horrible name for this issue, but that's essentially what's going on here. From Liam on the Slack channel: ```purescript newtype Showable a = Showable (∀ r . (Show a ⇒ a → r) → r) makeShowable ∷ ∀ a . Show a ⇒ a → Showable a makeShowable a = Showable \f → f a runShowable ∷ ∀ a . Showable a → String runShowable (Showable s) = s show ```
1.0
Wrap a 'Show' type class in a newtype - Probably a horrible name for this issue, but that's essentially what's going on here. From Liam on the Slack channel: ```purescript newtype Showable a = Showable (∀ r . (Show a ⇒ a → r) → r) makeShowable ∷ ∀ a . Show a ⇒ a → Showable a makeShowable a = Showable \f → f a runShowable ∷ ∀ a . Showable a → String runShowable (Showable s) = s show ```
non_defect
wrap a show type class in a newtype probably a horrible name for this issue but that s essentially what s going on here from liam on the slack channel purescript newtype showable a showable ∀ r show a ⇒ a → r → r makeshowable ∷ ∀ a show a ⇒ a → showable a makeshowable a showable f → f a runshowable ∷ ∀ a showable a → string runshowable showable s s show
0
11,428
2,651,459,035
IssuesEvent
2015-03-16 11:40:27
douglasdrumond/cloaked-computing-machine
https://api.github.com/repos/douglasdrumond/cloaked-computing-machine
opened
[CLOSED] Apple Key + Key Customization Doesn't Work
auto-migrated Priority-Medium Type-Defect
<a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Monday Mar 16, 2015 at 09:15 GMT_ _Originally opened as https://github.com/douglasdrumond/macvim/issues/14_ ---- ``` For example, customizing <D-2>, which works in other versions of vim, does not work in this application. ``` Original issue reported on code.google.com by `JasonCRo...@gmail.com` on 12 Sep 2007 at 4:02
1.0
[CLOSED] Apple Key + Key Customization Doesn't Work - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Monday Mar 16, 2015 at 09:15 GMT_ _Originally opened as https://github.com/douglasdrumond/macvim/issues/14_ ---- ``` For example, customizing <D-2>, which works in other versions of vim, does not work in this application. ``` Original issue reported on code.google.com by `JasonCRo...@gmail.com` on 12 Sep 2007 at 4:02
defect
apple key key customization doesn t work issue by monday mar at gmt originally opened as for example customizing which works in other versions of vim does not work in this application original issue reported on code google com by jasoncro gmail com on sep at
1
45,261
12,691,232,362
IssuesEvent
2020-06-21 16:01:22
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
Add support to retrieve data from associations of _joinData (a.k.a ternary association)
ORM defect pinned
This is a (multiple allowed): * [ ] bug * [x] enhancement * [x] feature-discussion (RFC) * CakePHP Version: 3.7. * Platform and Target: PHP 7.3 Mysql 5. ### What you did I have created a simple database with those models: _Doctors_ **belongsToMany** _Patients_ **through** _Appointments_ _Appointments_ **hasMany** _Examinations_ Full SQL: https://pastebin.com/9fY8Vibx As you can see, I have a classic "ternary association". So, I have tried to create a kind-of report showing _all doctors with theirs appointments date including patient name and a list of requested exams_ **Controller**: ```php $table = TableRegistry::getTableLocator()->get('Doctors'); $data = $table->find() ->contain('Patients') ->contain('Appointments.Examinations'); //Error ! $this->set(compact('data')); ``` **Template:** ```php <?php echo "<table>"; echo $this->Html->tableHeaders(['Doctor','Patient','Appointment Date', 'Exams']); foreach ($data as $doctor){ echo "<tr>"; echo "<td rowspan='".count($doctor->patients)."'>".$doctor->name."</td>"; foreach ($doctor->patients as $key => $patient) { if($key == !array_key_first($doctor->patients)){ echo "<tr>"; } echo "<td>".$patient->name."</td>"; echo "<td>".$patient->_joinData->date."</td>"; //$exams = array_column($patient->_joinData->examinations, 'name'); //Error ! //echo "<td>".implode('; ', $exams); //Error ! echo "</tr>"; } } ``` ### What happened The "pivot" table it is not associated with doctors so it is no possible to reach by the "normal" way the associated data of this _joinData ### What you expected to happen Be able to get Associated data from a model defined using "setThrough". Some changes could be applied on **findAssociation()** to "infere" (like a OWL Reasoner) that Doctors (or Patients) has a association with Appointments thanks to the association belongsToMany This data would be accessed on _joinData like regular entities when you choose to use contain() or maybe a new Query method like "containJoinData"
1.0
Add support to retrieve data from associations of _joinData (a.k.a ternary association) - This is a (multiple allowed): * [ ] bug * [x] enhancement * [x] feature-discussion (RFC) * CakePHP Version: 3.7. * Platform and Target: PHP 7.3 Mysql 5. ### What you did I have created a simple database with those models: _Doctors_ **belongsToMany** _Patients_ **through** _Appointments_ _Appointments_ **hasMany** _Examinations_ Full SQL: https://pastebin.com/9fY8Vibx As you can see, I have a classic "ternary association". So, I have tried to create a kind-of report showing _all doctors with theirs appointments date including patient name and a list of requested exams_ **Controller**: ```php $table = TableRegistry::getTableLocator()->get('Doctors'); $data = $table->find() ->contain('Patients') ->contain('Appointments.Examinations'); //Error ! $this->set(compact('data')); ``` **Template:** ```php <?php echo "<table>"; echo $this->Html->tableHeaders(['Doctor','Patient','Appointment Date', 'Exams']); foreach ($data as $doctor){ echo "<tr>"; echo "<td rowspan='".count($doctor->patients)."'>".$doctor->name."</td>"; foreach ($doctor->patients as $key => $patient) { if($key == !array_key_first($doctor->patients)){ echo "<tr>"; } echo "<td>".$patient->name."</td>"; echo "<td>".$patient->_joinData->date."</td>"; //$exams = array_column($patient->_joinData->examinations, 'name'); //Error ! //echo "<td>".implode('; ', $exams); //Error ! echo "</tr>"; } } ``` ### What happened The "pivot" table it is not associated with doctors so it is no possible to reach by the "normal" way the associated data of this _joinData ### What you expected to happen Be able to get Associated data from a model defined using "setThrough". Some changes could be applied on **findAssociation()** to "infere" (like a OWL Reasoner) that Doctors (or Patients) has a association with Appointments thanks to the association belongsToMany This data would be accessed on _joinData like regular entities when you choose to use contain() or maybe a new Query method like "containJoinData"
defect
add support to retrieve data from associations of joindata a k a ternary association this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target php mysql what you did i have created a simple database with those models doctors belongstomany patients through appointments appointments hasmany examinations full sql as you can see i have a classic ternary association so i have tried to create a kind of report showing all doctors with theirs appointments date including patient name and a list of requested exams controller php table tableregistry gettablelocator get doctors data table find contain patients contain appointments examinations error this set compact data template php php echo echo this html tableheaders foreach data as doctor echo echo patients doctor name foreach doctor patients as key patient if key array key first doctor patients echo echo patient name echo patient joindata date exams array column patient joindata examinations name error echo implode exams error echo what happened the pivot table it is not associated with doctors so it is no possible to reach by the normal way the associated data of this joindata what you expected to happen be able to get associated data from a model defined using setthrough some changes could be applied on findassociation to infere like a owl reasoner that doctors or patients has a association with appointments thanks to the association belongstomany this data would be accessed on joindata like regular entities when you choose to use contain or maybe a new query method like containjoindata
1
187,997
6,767,176,586
IssuesEvent
2017-10-26 01:36:40
CS2103AUG2017-T13-B1/main
https://api.github.com/repos/CS2103AUG2017-T13-B1/main
closed
Implementing links from insurance profile directly to person profile page
Priority: Medium Status: Completed
Some sort of hyperlink, buttons, or alike to jump from insurance page to profile
1.0
Implementing links from insurance profile directly to person profile page - Some sort of hyperlink, buttons, or alike to jump from insurance page to profile
non_defect
implementing links from insurance profile directly to person profile page some sort of hyperlink buttons or alike to jump from insurance page to profile
0
45,954
13,055,827,704
IssuesEvent
2020-07-30 02:51:18
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
BadDomList service does not close DB connection on exit (Trac #319)
Incomplete Migration Migrated from Trac combo reconstruction defect
Migrated from https://code.icecube.wisc.edu/ticket/319 ```json { "status": "closed", "changetime": "2012-03-18T17:52:31", "description": "The DB version of BadDomList service does not appear to properly\nclose the DB handle when it exits. This results in high numbers of what\nappear as \"dropped\" DB connections.\n\nThe DB interface should be updated to properly close all open DB connections. ", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "_ts": "1332093151000000", "component": "combo reconstruction", "summary": "BadDomList service does not close DB connection on exit", "priority": "normal", "keywords": "", "time": "2011-11-03T14:23:08", "milestone": "", "owner": "jacobi", "type": "defect" } ```
1.0
BadDomList service does not close DB connection on exit (Trac #319) - Migrated from https://code.icecube.wisc.edu/ticket/319 ```json { "status": "closed", "changetime": "2012-03-18T17:52:31", "description": "The DB version of BadDomList service does not appear to properly\nclose the DB handle when it exits. This results in high numbers of what\nappear as \"dropped\" DB connections.\n\nThe DB interface should be updated to properly close all open DB connections. ", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "_ts": "1332093151000000", "component": "combo reconstruction", "summary": "BadDomList service does not close DB connection on exit", "priority": "normal", "keywords": "", "time": "2011-11-03T14:23:08", "milestone": "", "owner": "jacobi", "type": "defect" } ```
defect
baddomlist service does not close db connection on exit trac migrated from json status closed changetime description the db version of baddomlist service does not appear to properly nclose the db handle when it exits this results in high numbers of what nappear as dropped db connections n nthe db interface should be updated to properly close all open db connections reporter blaufuss cc resolution fixed ts component combo reconstruction summary baddomlist service does not close db connection on exit priority normal keywords time milestone owner jacobi type defect
1
44,095
11,965,672,714
IssuesEvent
2020-04-06 00:22:27
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
csgraph.minimum_spanning_tree loses precision
Documentation defect scipy.sparse.csgraph
csgraph.minimum_spanning_tree seems to lose double precision somewhere along the way. This is surprising since all the input and output is formally float64 but still small elements < 1E-8 are rounded to zero. #### Reproducing code example: ``` >>> import numpy as np >>> import scipy.sparse >>> M=np.zeros((3,3), dtype=np.float64) >>> scipy.sparse.csgraph.minimum_spanning_tree(M + 1E-8).toarray() array([[0., 0., 0.], [0., 0., 0.], [0., 0., 0.]]) >>> scipy.sparse.csgraph.minimum_spanning_tree(M + 1E-7).toarray() array([[0.e+00, 1.e-07, 1.e-07], [0.e+00, 0.e+00, 0.e+00], [0.e+00, 0.e+00, 0.e+00]]) >>> scipy.sparse.csgraph.minimum_spanning_tree(M + 1E-8).dtype dtype('float64') ``` #### Scipy/Numpy/Python version information: ``` >>> print(scipy.__version__, np.__version__, sys.version_info) 1.3.1 1.17.2 sys.version_info(major=3, minor=7, micro=4, releaselevel='final', serial=0) ```
1.0
csgraph.minimum_spanning_tree loses precision - csgraph.minimum_spanning_tree seems to lose double precision somewhere along the way. This is surprising since all the input and output is formally float64 but still small elements < 1E-8 are rounded to zero. #### Reproducing code example: ``` >>> import numpy as np >>> import scipy.sparse >>> M=np.zeros((3,3), dtype=np.float64) >>> scipy.sparse.csgraph.minimum_spanning_tree(M + 1E-8).toarray() array([[0., 0., 0.], [0., 0., 0.], [0., 0., 0.]]) >>> scipy.sparse.csgraph.minimum_spanning_tree(M + 1E-7).toarray() array([[0.e+00, 1.e-07, 1.e-07], [0.e+00, 0.e+00, 0.e+00], [0.e+00, 0.e+00, 0.e+00]]) >>> scipy.sparse.csgraph.minimum_spanning_tree(M + 1E-8).dtype dtype('float64') ``` #### Scipy/Numpy/Python version information: ``` >>> print(scipy.__version__, np.__version__, sys.version_info) 1.3.1 1.17.2 sys.version_info(major=3, minor=7, micro=4, releaselevel='final', serial=0) ```
defect
csgraph minimum spanning tree loses precision csgraph minimum spanning tree seems to lose double precision somewhere along the way this is surprising since all the input and output is formally but still small elements are rounded to zero reproducing code example import numpy as np import scipy sparse m np zeros dtype np scipy sparse csgraph minimum spanning tree m toarray array scipy sparse csgraph minimum spanning tree m toarray array scipy sparse csgraph minimum spanning tree m dtype dtype scipy numpy python version information print scipy version np version sys version info sys version info major minor micro releaselevel final serial
1
17,540
3,012,746,958
IssuesEvent
2015-07-29 02:09:08
yawlfoundation/yawl
https://api.github.com/repos/yawlfoundation/yawl
closed
[CLOSED] Can't change the initial value of a net variable
auto-migrated Category-Editor Priority-Critical Type-Defect
<a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Monday Jul 27, 2015 at 03:21 GMT_ _Originally opened as https://github.com/adamsmj/yawl/issues/81_ ---- ``` In the net of issue 78 there is a local net variable that has received the initial value "hello" (including the quotes). This was accepted by the editor but it isn't possible to change this anymore. ``` Original issue reported on code.google.com by `arthurte...@gmail.com` on 8 Aug 2008 at 2:13
1.0
[CLOSED] Can't change the initial value of a net variable - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)** _Monday Jul 27, 2015 at 03:21 GMT_ _Originally opened as https://github.com/adamsmj/yawl/issues/81_ ---- ``` In the net of issue 78 there is a local net variable that has received the initial value "hello" (including the quotes). This was accepted by the editor but it isn't possible to change this anymore. ``` Original issue reported on code.google.com by `arthurte...@gmail.com` on 8 Aug 2008 at 2:13
defect
can t change the initial value of a net variable issue by monday jul at gmt originally opened as in the net of issue there is a local net variable that has received the initial value hello including the quotes this was accepted by the editor but it isn t possible to change this anymore original issue reported on code google com by arthurte gmail com on aug at
1
368,740
25,807,074,113
IssuesEvent
2022-12-11 14:04:29
getsentry/sentry-capacitor
https://api.github.com/repos/getsentry/sentry-capacitor
closed
Make clear in installation docs that the sibling SDK version has to match.
documentation Platform: Capacitor update-sentry-js-sdk
We will need to document in the installation steps to ensure the version of the sibling SDK matches the version of the sentry-javascript dependencies that @sentry/capacitor depends on.
1.0
Make clear in installation docs that the sibling SDK version has to match. - We will need to document in the installation steps to ensure the version of the sibling SDK matches the version of the sentry-javascript dependencies that @sentry/capacitor depends on.
non_defect
make clear in installation docs that the sibling sdk version has to match we will need to document in the installation steps to ensure the version of the sibling sdk matches the version of the sentry javascript dependencies that sentry capacitor depends on
0
8,356
3,163,929,487
IssuesEvent
2015-09-20 18:58:03
acemod/language-arma-atom
https://api.github.com/repos/acemod/language-arma-atom
opened
Update README.md screenshots and information for autocomplete-plus
documentation
- [ ] Screenshots with autocomplete-plus - [ ] Build Dev and Build Release functions shortly presented
1.0
Update README.md screenshots and information for autocomplete-plus - - [ ] Screenshots with autocomplete-plus - [ ] Build Dev and Build Release functions shortly presented
non_defect
update readme md screenshots and information for autocomplete plus screenshots with autocomplete plus build dev and build release functions shortly presented
0
749,764
26,179,015,117
IssuesEvent
2023-01-02 13:21:47
conan-io/conan
https://api.github.com/repos/conan-io/conan
closed
[feature] Virtualenvs should create a sandbox
type: feature stage: queue priority: medium complex: low
Taken from this comment: https://github.com/conan-io/conan/pull/5989#pullrequestreview-326075695 For Conan v2.0 we want the virtualenvs to restore the environment to the state it was before activating it (same variables, same values). So, anything modified inside the virtualenv will be discarded after deactivating it.
1.0
[feature] Virtualenvs should create a sandbox - Taken from this comment: https://github.com/conan-io/conan/pull/5989#pullrequestreview-326075695 For Conan v2.0 we want the virtualenvs to restore the environment to the state it was before activating it (same variables, same values). So, anything modified inside the virtualenv will be discarded after deactivating it.
non_defect
virtualenvs should create a sandbox taken from this comment for conan we want the virtualenvs to restore the environment to the state it was before activating it same variables same values so anything modified inside the virtualenv will be discarded after deactivating it
0
66,309
20,147,310,062
IssuesEvent
2022-02-09 08:57:55
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Record::formatJSON does not respect JSONFormat::wrapSingleColumnRecords
T: Defect C: Functionality P: Low E: All Editions
Calling `Record::formatJSON` with `jsonFormat.wrapSingleColumnRecords(false)` doesn't have any effect. It is only used when formatting a `Result`.
1.0
Record::formatJSON does not respect JSONFormat::wrapSingleColumnRecords - Calling `Record::formatJSON` with `jsonFormat.wrapSingleColumnRecords(false)` doesn't have any effect. It is only used when formatting a `Result`.
defect
record formatjson does not respect jsonformat wrapsinglecolumnrecords calling record formatjson with jsonformat wrapsinglecolumnrecords false doesn t have any effect it is only used when formatting a result
1
14,067
24,346,594,090
IssuesEvent
2022-10-02 11:57:04
PharmaLedger-IMI/epi-workspace
https://api.github.com/repos/PharmaLedger-IMI/epi-workspace
closed
Scalable Solution
user story Non-functional requirement
The productive pilot solution should be able to support at minimum - Organizations - 5/6 companies Products - 80-100 (10-15 per company) Batches - 100-200/product Leaflets - 80-100 for 3 langs (Dutch, German, French, English) Serial nos - 800K - 1M
1.0
Scalable Solution - The productive pilot solution should be able to support at minimum - Organizations - 5/6 companies Products - 80-100 (10-15 per company) Batches - 100-200/product Leaflets - 80-100 for 3 langs (Dutch, German, French, English) Serial nos - 800K - 1M
non_defect
scalable solution the productive pilot solution should be able to support at minimum organizations companies products per company batches product leaflets for langs dutch german french english serial nos
0
61,548
17,023,722,801
IssuesEvent
2021-07-03 03:29:47
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
[roads] Secondary bridges rendered way larger than primary bridges in mapnik on z15-13
Component: mapnik Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 2.34am, Saturday, 18th June 2011]** http://www.openstreetmap.org/?lat=36.107&lon=-115.1832&zoom=14&layers=M Just noticed that secondary bridges are being rendered on z15-z13 a lot bigger than primary bridges that are nearby, which shouldn't be the case. So, I went into the osm.xml file to see if I could spot where the problem was and I 'think' I found it. The problem (I think) lies in the "<Style name="bridges_fill">" section. Now, I don't fully know all that much on how the style sheet works, so, I will not go suggesting exactly what to fix because for all I know, what I suggest could make it worse. But this secondary bridge is rendered bigger than the twinned SR 592 right above it which should be wider since there are two ways there that are combined when rendered at z15-z13. So, any rendering fix would be appreciated for this.
1.0
[roads] Secondary bridges rendered way larger than primary bridges in mapnik on z15-13 - **[Submitted to the original trac issue database at 2.34am, Saturday, 18th June 2011]** http://www.openstreetmap.org/?lat=36.107&lon=-115.1832&zoom=14&layers=M Just noticed that secondary bridges are being rendered on z15-z13 a lot bigger than primary bridges that are nearby, which shouldn't be the case. So, I went into the osm.xml file to see if I could spot where the problem was and I 'think' I found it. The problem (I think) lies in the "<Style name="bridges_fill">" section. Now, I don't fully know all that much on how the style sheet works, so, I will not go suggesting exactly what to fix because for all I know, what I suggest could make it worse. But this secondary bridge is rendered bigger than the twinned SR 592 right above it which should be wider since there are two ways there that are combined when rendered at z15-z13. So, any rendering fix would be appreciated for this.
defect
secondary bridges rendered way larger than primary bridges in mapnik on just noticed that secondary bridges are being rendered on a lot bigger than primary bridges that are nearby which shouldn t be the case so i went into the osm xml file to see if i could spot where the problem was and i think i found it the problem i think lies in the section now i don t fully know all that much on how the style sheet works so i will not go suggesting exactly what to fix because for all i know what i suggest could make it worse but this secondary bridge is rendered bigger than the twinned sr right above it which should be wider since there are two ways there that are combined when rendered at so any rendering fix would be appreciated for this
1
146,839
23,130,784,246
IssuesEvent
2022-07-28 10:09:47
flix/flix
https://api.github.com/repos/flix/flix
closed
Interoperability: Add an expression that allows the (anon.) implementation of a Java interface/class extension
language-design
# Overall Goal: Add new anonymous class/object expression For example: ``` object ##java.lang.Iterator { def hasNext(): Bool = true def next(): String = "Hello" } ``` which will generate a new class called say `Anon$1` which implements the `Iterator` interface and has two methods `hasNext` and `next`. The `object` expression is a Java value of type `##java.lang.Iterator`. (We ignore generics for now-- they are erased anyway.) I propose to work towards this goal through several milestones: # Milestone 1 - [ ] Add a new ` NewObject` expression (right after PutStaticField) that takes a list of `1` class and `n` interfaces. For example, we should be able to compile and run these expressions (trivial as they may be): Ex1: ``` object ##java.lang.Object { } ``` Ex2: ``` object ##java.lang.Serializable { } ``` Ex3: ``` object ##java.lang.Object, ##java.lang.Serializable, ##java.util.RandomAccess { } ``` # Milestone 2 TBD
1.0
Interoperability: Add an expression that allows the (anon.) implementation of a Java interface/class extension - # Overall Goal: Add new anonymous class/object expression For example: ``` object ##java.lang.Iterator { def hasNext(): Bool = true def next(): String = "Hello" } ``` which will generate a new class called say `Anon$1` which implements the `Iterator` interface and has two methods `hasNext` and `next`. The `object` expression is a Java value of type `##java.lang.Iterator`. (We ignore generics for now-- they are erased anyway.) I propose to work towards this goal through several milestones: # Milestone 1 - [ ] Add a new ` NewObject` expression (right after PutStaticField) that takes a list of `1` class and `n` interfaces. For example, we should be able to compile and run these expressions (trivial as they may be): Ex1: ``` object ##java.lang.Object { } ``` Ex2: ``` object ##java.lang.Serializable { } ``` Ex3: ``` object ##java.lang.Object, ##java.lang.Serializable, ##java.util.RandomAccess { } ``` # Milestone 2 TBD
non_defect
interoperability add an expression that allows the anon implementation of a java interface class extension overall goal add new anonymous class object expression for example object java lang iterator def hasnext bool true def next string hello which will generate a new class called say anon which implements the iterator interface and has two methods hasnext and next the object expression is a java value of type java lang iterator we ignore generics for now they are erased anyway i propose to work towards this goal through several milestones milestone add a new newobject expression right after putstaticfield that takes a list of class and n interfaces for example we should be able to compile and run these expressions trivial as they may be object java lang object object java lang serializable object java lang object java lang serializable java util randomaccess milestone tbd
0
3,750
2,610,068,166
IssuesEvent
2015-02-26 18:20:01
chrsmith/jsjsj122
https://api.github.com/repos/chrsmith/jsjsj122
opened
路桥看前列腺炎哪里权威
auto-migrated Priority-Medium Type-Defect
``` 路桥看前列腺炎哪里权威【台州五洲生殖医院】24小时健康咨 询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州 市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108� ��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109 、112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 9:00
1.0
路桥看前列腺炎哪里权威 - ``` 路桥看前列腺炎哪里权威【台州五洲生殖医院】24小时健康咨 询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州 市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108� ��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109 、112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 9:00
defect
路桥看前列腺炎哪里权威 路桥看前列腺炎哪里权威【台州五洲生殖医院】 询热线 微信号tzwzszyy 医院地址 台州 (枫南大转盘旁)乘车线路 、 � �� 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
1
323,120
27,696,559,324
IssuesEvent
2023-03-14 03:00:18
pombase/pombase-chado
https://api.github.com/repos/pombase/pombase-chado
closed
missing data
bug needs testing
Some annotations that were there last week have disappeared form the query builder. No results for: genes annotated with "LYR family" ([PBO:0091211](https://www.pombase.org/term/PBO:0091211)) There were 8 annotations. They are annotated in the contig files, and usually appear in the protein feature section of the page.
1.0
missing data - Some annotations that were there last week have disappeared form the query builder. No results for: genes annotated with "LYR family" ([PBO:0091211](https://www.pombase.org/term/PBO:0091211)) There were 8 annotations. They are annotated in the contig files, and usually appear in the protein feature section of the page.
non_defect
missing data some annotations that were there last week have disappeared form the query builder no results for genes annotated with lyr family there were annotations they are annotated in the contig files and usually appear in the protein feature section of the page
0
66,169
20,025,853,408
IssuesEvent
2022-02-01 21:13:17
idaholab/moose
https://api.github.com/repos/idaholab/moose
closed
Allow PiecewiseConstant functions to control Timestepper
T: defect P: normal
## Bug Description Somewhere along the way, support for `PiecewiseConstant` functions to be able to limit the `IterationAdaptiveDT` timestepper dt via `timestep_limiting_function` was lost...or was never there and never tested. ## Steps to Reproduce `moose_test-opt -i test/tests/time_steppers/piecewise_linear.i Functions/temp_spike/type=PiecewiseConstant` ## Impact Expected functionality will return.
1.0
Allow PiecewiseConstant functions to control Timestepper - ## Bug Description Somewhere along the way, support for `PiecewiseConstant` functions to be able to limit the `IterationAdaptiveDT` timestepper dt via `timestep_limiting_function` was lost...or was never there and never tested. ## Steps to Reproduce `moose_test-opt -i test/tests/time_steppers/piecewise_linear.i Functions/temp_spike/type=PiecewiseConstant` ## Impact Expected functionality will return.
defect
allow piecewiseconstant functions to control timestepper bug description somewhere along the way support for piecewiseconstant functions to be able to limit the iterationadaptivedt timestepper dt via timestep limiting function was lost or was never there and never tested steps to reproduce moose test opt i test tests time steppers piecewise linear i functions temp spike type piecewiseconstant impact expected functionality will return
1
225,462
7,481,906,153
IssuesEvent
2018-04-04 22:24:57
Planteome/plant-experimental-conditions-ontology
https://api.github.com/repos/Planteome/plant-experimental-conditions-ontology
closed
antibiotic treatment
Plant Treatment high priority sourceforge
current name and definition: antibiotic regimen ([EO:0007041](http://purl.obolibrary.org/obo/EO_0007041)): Treatment involving the use of antibiotic for selection purposes. Proposed new name and definition: antibiotic treatment: A chemical treatment ([EO:0007189](http://purl.obolibrary.org/obo/EO_0007189)) involving the use of antibiotic for selection purposes. Reported by: lauramoore1859 Original Ticket: [obo/plant-environment-ontology-eo/70](https://sourceforge.net/p/obo/plant-environment-ontology-eo/70)
1.0
antibiotic treatment - current name and definition: antibiotic regimen ([EO:0007041](http://purl.obolibrary.org/obo/EO_0007041)): Treatment involving the use of antibiotic for selection purposes. Proposed new name and definition: antibiotic treatment: A chemical treatment ([EO:0007189](http://purl.obolibrary.org/obo/EO_0007189)) involving the use of antibiotic for selection purposes. Reported by: lauramoore1859 Original Ticket: [obo/plant-environment-ontology-eo/70](https://sourceforge.net/p/obo/plant-environment-ontology-eo/70)
non_defect
antibiotic treatment current name and definition antibiotic regimen treatment involving the use of antibiotic for selection purposes proposed new name and definition antibiotic treatment a chemical treatment involving the use of antibiotic for selection purposes reported by original ticket
0
60,150
12,059,669,335
IssuesEvent
2020-04-15 19:42:25
ikostan/codewars
https://api.github.com/repos/ikostan/codewars
opened
Fix "method_complexity" issue in kyu_6/help_the_bookseller/stock_list.py
codeclimate codewars issues
Function `stock_list` has a Cognitive Complexity of 10 (exceeds 5 allowed). Consider refactoring. https://codeclimate.com/github/ikostan/codewars/kyu_6/help_the_bookseller/stock_list.py#issue_5e96c3368a97c800010001bb
2.0
Fix "method_complexity" issue in kyu_6/help_the_bookseller/stock_list.py - Function `stock_list` has a Cognitive Complexity of 10 (exceeds 5 allowed). Consider refactoring. https://codeclimate.com/github/ikostan/codewars/kyu_6/help_the_bookseller/stock_list.py#issue_5e96c3368a97c800010001bb
non_defect
fix method complexity issue in kyu help the bookseller stock list py function stock list has a cognitive complexity of exceeds allowed consider refactoring
0
754,324
26,381,856,886
IssuesEvent
2023-01-12 09:17:43
sourceplusplus/sourceplusplus
https://api.github.com/repos/sourceplusplus/sourceplusplus
opened
Probe bridge TCP timeouts
a:bug in:live-platform of:small-effort with:moderate-priority
Probes connected via TCP don't send ping/pong messages and never timeout and therefore are not properly removed from active probe stats. A client ping/pong thread (every 5 seconds) and server timeout (10 seconds) need to be added to match SockJS impl.
1.0
Probe bridge TCP timeouts - Probes connected via TCP don't send ping/pong messages and never timeout and therefore are not properly removed from active probe stats. A client ping/pong thread (every 5 seconds) and server timeout (10 seconds) need to be added to match SockJS impl.
non_defect
probe bridge tcp timeouts probes connected via tcp don t send ping pong messages and never timeout and therefore are not properly removed from active probe stats a client ping pong thread every seconds and server timeout seconds need to be added to match sockjs impl
0
23,082
4,869,026,199
IssuesEvent
2016-11-15 11:17:38
ChrisCummins/clgen
https://api.github.com/repos/ChrisCummins/clgen
closed
Use remote model and sampler in example
Documentation Feature Packaging
Something like: ```sh export model=https://raw.githubusercontent.com/ChrisCummins/clgen/blob/master/tests/data/tiny/corpus.tar.bz2 export sampler=https://raw.githubusercontent.com/ChrisCummins/clgen/blob/master/tests/data/tiny/sampler.json $ clgen $model $sampler ```
1.0
Use remote model and sampler in example - Something like: ```sh export model=https://raw.githubusercontent.com/ChrisCummins/clgen/blob/master/tests/data/tiny/corpus.tar.bz2 export sampler=https://raw.githubusercontent.com/ChrisCummins/clgen/blob/master/tests/data/tiny/sampler.json $ clgen $model $sampler ```
non_defect
use remote model and sampler in example something like sh export model export sampler clgen model sampler
0
8,191
2,965,596,803
IssuesEvent
2015-07-11 00:54:22
gheber/kenzo
https://api.github.com/repos/gheber/kenzo
closed
BAR test fails
bug test
The BAR test in bar.lisp fails with: Failure Details: -------------------------------- BAR []: Unexpected Error: #<TYPE-ERROR expected-type: (SATISFIES CAT::ICMBN-P) datum: ----------------------------------------------------------------------{CMBN 6} <-1 * (-7 -1 1 2 4 -2)> <-1 * (-7 -1 1 4 -2 2)> <1 * (-7 -1 1 4 2 -2)> <-1 * (-7 -1 4 -2 1 2)> <1 * (-7 -1 4 1 -2 2)> <-1 * (-7 -1 4 1 2 -2)> <1 * (-7 1 -1 2 4 -2)> <1 * (-7 1 -1 4 -2 2)> <-1 * (-7 1 -1 4 2 -2)> <-1 * (-7 1 2 -1 4 -2)> ... ... ------------------------------------------------------------------------------ > The value ----------------------------------------------------------------------{CMBN 6} <-1 * (-7 -1 1 2 4 -2)> <-1 * (-7 -1 1 4 -2 2)> <1 * (-7 -1 1 4 2 -2)> <-1 * (-7 -1 4 -2 1 2)> <1 * (-7 -1 4 1 -2 2)> <-1 * (-7 -1 4 1 2 -2)> <1 * (-7 1 -1 2 4 -2)> <1 * (-7 1 -1 4 -2 2)> <-1 * (-7 1 -1 4 2 -2)> <-1 * (-7 1 2 -1 4 -2)> ... ... ------------------------------------------------------------------------------ is not of type (SATISFIES CAT::ICMBN-P)... -------------------------------- This is probably the same issue as #24 and #25.
1.0
BAR test fails - The BAR test in bar.lisp fails with: Failure Details: -------------------------------- BAR []: Unexpected Error: #<TYPE-ERROR expected-type: (SATISFIES CAT::ICMBN-P) datum: ----------------------------------------------------------------------{CMBN 6} <-1 * (-7 -1 1 2 4 -2)> <-1 * (-7 -1 1 4 -2 2)> <1 * (-7 -1 1 4 2 -2)> <-1 * (-7 -1 4 -2 1 2)> <1 * (-7 -1 4 1 -2 2)> <-1 * (-7 -1 4 1 2 -2)> <1 * (-7 1 -1 2 4 -2)> <1 * (-7 1 -1 4 -2 2)> <-1 * (-7 1 -1 4 2 -2)> <-1 * (-7 1 2 -1 4 -2)> ... ... ------------------------------------------------------------------------------ > The value ----------------------------------------------------------------------{CMBN 6} <-1 * (-7 -1 1 2 4 -2)> <-1 * (-7 -1 1 4 -2 2)> <1 * (-7 -1 1 4 2 -2)> <-1 * (-7 -1 4 -2 1 2)> <1 * (-7 -1 4 1 -2 2)> <-1 * (-7 -1 4 1 2 -2)> <1 * (-7 1 -1 2 4 -2)> <1 * (-7 1 -1 4 -2 2)> <-1 * (-7 1 -1 4 2 -2)> <-1 * (-7 1 2 -1 4 -2)> ... ... ------------------------------------------------------------------------------ is not of type (SATISFIES CAT::ICMBN-P)... -------------------------------- This is probably the same issue as #24 and #25.
non_defect
bar test fails the bar test in bar lisp fails with failure details bar unexpected error type error expected type satisfies cat icmbn p datum cmbn the value cmbn is not of type satisfies cat icmbn p this is probably the same issue as and
0
252,672
21,610,232,825
IssuesEvent
2022-05-04 09:20:34
opensrp/opensrp-client-anc
https://api.github.com/repos/opensrp/opensrp-client-anc
closed
[Ona Support Request]: App crashing when one tries to start a contact
high priority Tech Partner (SID Team) client testing
### Affected App or Server Version v1.6.8 ### What kind of support do you need? SID local app crashing when one tries to start a contact ### What is the acceptance criteria for your support request? Able to start and complete a contact ### Relevant Information _No response_
1.0
[Ona Support Request]: App crashing when one tries to start a contact - ### Affected App or Server Version v1.6.8 ### What kind of support do you need? SID local app crashing when one tries to start a contact ### What is the acceptance criteria for your support request? Able to start and complete a contact ### Relevant Information _No response_
non_defect
app crashing when one tries to start a contact affected app or server version what kind of support do you need sid local app crashing when one tries to start a contact what is the acceptance criteria for your support request able to start and complete a contact relevant information no response
0
18,131
10,010,930,434
IssuesEvent
2019-07-15 09:17:56
hlissner/doom-emacs
https://api.github.com/repos/hlissner/doom-emacs
closed
Debugging emacs problems
is:question module:lang:cc re:performance status:unknown
I have been noticing that while editing my C++ files, the speed degrades after a few days of usage (I rarely restart my emacs). I suspect that there's some packages that are being loaded which is slowing down my emacs but it is difficult to debug this since I don't know when these packages get loaded. Is there a way to force doom to load all packages at startup. I know this will slow the startup time significantly but I would like to be able to troubleshoot these speed problems.
True
Debugging emacs problems - I have been noticing that while editing my C++ files, the speed degrades after a few days of usage (I rarely restart my emacs). I suspect that there's some packages that are being loaded which is slowing down my emacs but it is difficult to debug this since I don't know when these packages get loaded. Is there a way to force doom to load all packages at startup. I know this will slow the startup time significantly but I would like to be able to troubleshoot these speed problems.
non_defect
debugging emacs problems i have been noticing that while editing my c files the speed degrades after a few days of usage i rarely restart my emacs i suspect that there s some packages that are being loaded which is slowing down my emacs but it is difficult to debug this since i don t know when these packages get loaded is there a way to force doom to load all packages at startup i know this will slow the startup time significantly but i would like to be able to troubleshoot these speed problems
0
63,262
17,499,849,150
IssuesEvent
2021-08-10 08:04:46
milvus-io/milvus-insight
https://api.github.com/repos/milvus-io/milvus-insight
opened
Server log undefined error
defect
**Describe the bug:** ![image](https://user-images.githubusercontent.com/20559208/128830949-90260531-60dc-4111-b708-6af3c1b96359.png) **Steps to reproduce:** 1. yarn start:dev in server folder **Milvus-insight version:** latest **Milvus version:**
1.0
Server log undefined error - **Describe the bug:** ![image](https://user-images.githubusercontent.com/20559208/128830949-90260531-60dc-4111-b708-6af3c1b96359.png) **Steps to reproduce:** 1. yarn start:dev in server folder **Milvus-insight version:** latest **Milvus version:**
defect
server log undefined error describe the bug steps to reproduce yarn start dev in server folder milvus insight version latest milvus version
1
4,949
2,610,162,147
IssuesEvent
2015-02-26 18:51:26
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Text
auto-migrated Priority-Medium Type-Defect
``` Weapons upgrade text name missing ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 22 Feb 2011 at 2:58
1.0
Text - ``` Weapons upgrade text name missing ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 22 Feb 2011 at 2:58
defect
text weapons upgrade text name missing original issue reported on code google com by gmail com on feb at
1
281,814
24,421,793,313
IssuesEvent
2022-10-05 21:03:12
NCAR/DART
https://api.github.com/repos/NCAR/DART
closed
obs_sequence_tool developer tests
Test
Currently there are no tests to run to check changes to obs_sequence_tool (there is the start of some testing for reading little and big endian obs sequences) These tests should: * go in developer_tests/obs_sequence * be run as part of test_dart * be able to run outside of test_dart for people developing the obs_sequence_tool assigned to @hkershaw-brown
1.0
obs_sequence_tool developer tests - Currently there are no tests to run to check changes to obs_sequence_tool (there is the start of some testing for reading little and big endian obs sequences) These tests should: * go in developer_tests/obs_sequence * be run as part of test_dart * be able to run outside of test_dart for people developing the obs_sequence_tool assigned to @hkershaw-brown
non_defect
obs sequence tool developer tests currently there are no tests to run to check changes to obs sequence tool there is the start of some testing for reading little and big endian obs sequences these tests should go in developer tests obs sequence be run as part of test dart be able to run outside of test dart for people developing the obs sequence tool assigned to hkershaw brown
0
17,028
2,966,760,451
IssuesEvent
2015-07-12 07:15:15
ElvisChiang/welly
https://api.github.com/repos/ElvisChiang/welly
closed
Maximize welly window size
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Mistakenly press unknown keys, the welly window was maximize, but never get back to default size. This is really annoying. 2. 3. What is the expected output? What do you see instead? I expected user could reset the window's size. The maximum size is annoying. What version of the product are you using? On what operating system? OS: Lion (10.7) Welly 2.64 Please provide any additional information below. If welly disables welly window resizing, it should also disable the system hotkey/shortcut key for window maximize on welly. Otherwise, please enable window resize. Attached a screenshot for reference. Thanks. ``` Original issue reported on code.google.com by `kyle.w....@gmail.com` on 1 Jun 2012 at 4:01 Attachments: * [Welly_On_Maximum_Size.png](https://storage.googleapis.com/google-code-attachments/welly/issue-65/comment-0/Welly_On_Maximum_Size.png)
1.0
Maximize welly window size - ``` What steps will reproduce the problem? 1. Mistakenly press unknown keys, the welly window was maximize, but never get back to default size. This is really annoying. 2. 3. What is the expected output? What do you see instead? I expected user could reset the window's size. The maximum size is annoying. What version of the product are you using? On what operating system? OS: Lion (10.7) Welly 2.64 Please provide any additional information below. If welly disables welly window resizing, it should also disable the system hotkey/shortcut key for window maximize on welly. Otherwise, please enable window resize. Attached a screenshot for reference. Thanks. ``` Original issue reported on code.google.com by `kyle.w....@gmail.com` on 1 Jun 2012 at 4:01 Attachments: * [Welly_On_Maximum_Size.png](https://storage.googleapis.com/google-code-attachments/welly/issue-65/comment-0/Welly_On_Maximum_Size.png)
defect
maximize welly window size what steps will reproduce the problem mistakenly press unknown keys the welly window was maximize but never get back to default size this is really annoying what is the expected output what do you see instead i expected user could reset the window s size the maximum size is annoying what version of the product are you using on what operating system os lion welly please provide any additional information below if welly disables welly window resizing it should also disable the system hotkey shortcut key for window maximize on welly otherwise please enable window resize attached a screenshot for reference thanks original issue reported on code google com by kyle w gmail com on jun at attachments
1
16,759
2,941,839,420
IssuesEvent
2015-07-02 10:35:18
google/google-api-dotnet-client
https://api.github.com/repos/google/google-api-dotnet-client
closed
.NET C# Blogger v3 API Bug, Service.Blogs.ListByUser
auto-migrated Component-Api Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Instantiate Blogger Service 2. Call ListByUser using an actual userId from gplus instead of "self" BlogsResource.ListByUserRequest blogListByUserAction = service.Blogs.ListByUser("105499867526169935261"); 3. Run Fetch BlogList list = blogListByUserAction.Fetch(); What is the expected output? What do you see instead? I should get a list of my blogs; however, I receive the following exception (works fine w/ "self"): "Google.GoogleApiRequestException was unhandled by user code Message=Google.Apis.Requests.RequestError [500] No individual errors" What version of the product are you using? On what operating system? .net c# blogger api v3 on windows 7 64 bit. Please provide any additional information below. This also is reproduceable in the sandbox using the userId: https://developers.google.com/blogger/docs/3.0/reference/blogs/listByUser 500 Internal Server Error - Hide headers - cache-control: private, max-age=0 content-encoding: gzip content-length: 66 content-type: application/json; charset=UTF-8 date: Wed, 22 May 2013 21:34:15 GMT expires: Wed, 22 May 2013 21:34:15 GMT server: GSE { "error": { "code": 500, "message": null } } ``` Original issue reported on code.google.com by `cbarthol...@gmail.com` on 22 May 2013 at 9:35
1.0
.NET C# Blogger v3 API Bug, Service.Blogs.ListByUser - ``` What steps will reproduce the problem? 1. Instantiate Blogger Service 2. Call ListByUser using an actual userId from gplus instead of "self" BlogsResource.ListByUserRequest blogListByUserAction = service.Blogs.ListByUser("105499867526169935261"); 3. Run Fetch BlogList list = blogListByUserAction.Fetch(); What is the expected output? What do you see instead? I should get a list of my blogs; however, I receive the following exception (works fine w/ "self"): "Google.GoogleApiRequestException was unhandled by user code Message=Google.Apis.Requests.RequestError [500] No individual errors" What version of the product are you using? On what operating system? .net c# blogger api v3 on windows 7 64 bit. Please provide any additional information below. This also is reproduceable in the sandbox using the userId: https://developers.google.com/blogger/docs/3.0/reference/blogs/listByUser 500 Internal Server Error - Hide headers - cache-control: private, max-age=0 content-encoding: gzip content-length: 66 content-type: application/json; charset=UTF-8 date: Wed, 22 May 2013 21:34:15 GMT expires: Wed, 22 May 2013 21:34:15 GMT server: GSE { "error": { "code": 500, "message": null } } ``` Original issue reported on code.google.com by `cbarthol...@gmail.com` on 22 May 2013 at 9:35
defect
net c blogger api bug service blogs listbyuser what steps will reproduce the problem instantiate blogger service call listbyuser using an actual userid from gplus instead of self blogsresource listbyuserrequest bloglistbyuseraction service blogs listbyuser run fetch bloglist list bloglistbyuseraction fetch what is the expected output what do you see instead i should get a list of my blogs however i receive the following exception works fine w self google googleapirequestexception was unhandled by user code message google apis requests requesterror no individual errors what version of the product are you using on what operating system net c blogger api on windows bit please provide any additional information below this also is reproduceable in the sandbox using the userid internal server error hide headers cache control private max age content encoding gzip content length content type application json charset utf date wed may gmt expires wed may gmt server gse error code message null original issue reported on code google com by cbarthol gmail com on may at
1
45,992
13,055,834,284
IssuesEvent
2020-07-30 02:52:28
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
trigger config parameters are treated as int, not double (Trac #389)
Incomplete Migration Migrated from Trac dataclasses defect
Migrated from https://code.icecube.wisc.edu/ticket/389 ```json { "status": "closed", "changetime": "2014-11-22T18:26:26", "description": "I3TriggerStatus treats trigger settings as integers. But the SLOP trigger has a setting (rel_v) which is set to 0.5. When creating a GCD this parameter is therefore not set correctly and the trigger simulation is not using the correct settings.\n\nIn the database the parameters are saved as string with the correct values.\nI3TriggerStatus should be able to treat a trigger parameter as double.\n\nI've changed this on a local working copy and it works for me, but as a consequence my GCD files are not compatible with an unchanged software.\nHere's what I've done: http://www.pastie.org/private/urikfrsoiwejvi3sg5lxa\n\nA support of double while keeping backwards compatibility is desirable.\n\nEmanuel\n", "reporter": "jacobi", "cc": "", "resolution": "fixed", "_ts": "1416680786877026", "component": "dataclasses", "summary": "trigger config parameters are treated as int, not double", "priority": "normal", "keywords": "", "time": "2012-04-16T16:57:50", "milestone": "", "owner": "olivas", "type": "defect" } ```
1.0
trigger config parameters are treated as int, not double (Trac #389) - Migrated from https://code.icecube.wisc.edu/ticket/389 ```json { "status": "closed", "changetime": "2014-11-22T18:26:26", "description": "I3TriggerStatus treats trigger settings as integers. But the SLOP trigger has a setting (rel_v) which is set to 0.5. When creating a GCD this parameter is therefore not set correctly and the trigger simulation is not using the correct settings.\n\nIn the database the parameters are saved as string with the correct values.\nI3TriggerStatus should be able to treat a trigger parameter as double.\n\nI've changed this on a local working copy and it works for me, but as a consequence my GCD files are not compatible with an unchanged software.\nHere's what I've done: http://www.pastie.org/private/urikfrsoiwejvi3sg5lxa\n\nA support of double while keeping backwards compatibility is desirable.\n\nEmanuel\n", "reporter": "jacobi", "cc": "", "resolution": "fixed", "_ts": "1416680786877026", "component": "dataclasses", "summary": "trigger config parameters are treated as int, not double", "priority": "normal", "keywords": "", "time": "2012-04-16T16:57:50", "milestone": "", "owner": "olivas", "type": "defect" } ```
defect
trigger config parameters are treated as int not double trac migrated from json status closed changetime description treats trigger settings as integers but the slop trigger has a setting rel v which is set to when creating a gcd this parameter is therefore not set correctly and the trigger simulation is not using the correct settings n nin the database the parameters are saved as string with the correct values should be able to treat a trigger parameter as double n ni ve changed this on a local working copy and it works for me but as a consequence my gcd files are not compatible with an unchanged software nhere s what i ve done support of double while keeping backwards compatibility is desirable n nemanuel n reporter jacobi cc resolution fixed ts component dataclasses summary trigger config parameters are treated as int not double priority normal keywords time milestone owner olivas type defect
1
32,041
6,691,016,164
IssuesEvent
2017-10-09 11:30:32
primefaces/primereact
https://api.github.com/repos/primefaces/primereact
closed
Dropdown menu sits behind grid
defect
The dropdown menu in DataTable sites behind the grid when resizableColumns or scrollable is enabled. Caused by the overflow hidden; https://screenshots.firefox.com/49XyjS9g09rpWvUd/localhost
1.0
Dropdown menu sits behind grid - The dropdown menu in DataTable sites behind the grid when resizableColumns or scrollable is enabled. Caused by the overflow hidden; https://screenshots.firefox.com/49XyjS9g09rpWvUd/localhost
defect
dropdown menu sits behind grid the dropdown menu in datatable sites behind the grid when resizablecolumns or scrollable is enabled caused by the overflow hidden
1
3,467
2,610,063,296
IssuesEvent
2015-02-26 18:18:36
chrsmith/jsjsj122
https://api.github.com/repos/chrsmith/jsjsj122
opened
黄岩治不育哪家效果好
auto-migrated Priority-Medium Type-Defect
``` 黄岩治不育哪家效果好【台州五洲生殖医院】24小时健康咨询 热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市 椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、1 18、198及椒江一金清公交车直达枫南小区,乘坐107、105、109、 112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:49
1.0
黄岩治不育哪家效果好 - ``` 黄岩治不育哪家效果好【台州五洲生殖医院】24小时健康咨询 热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市 椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、1 18、198及椒江一金清公交车直达枫南小区,乘坐107、105、109、 112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:49
defect
黄岩治不育哪家效果好 黄岩治不育哪家效果好【台州五洲生殖医院】 热线 微信号tzwzszyy 医院地址 台州市 (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
1
79,151
28,017,776,592
IssuesEvent
2023-03-28 01:06:53
salsadigitalauorg/civictheme_source
https://api.github.com/repos/salsadigitalauorg/civictheme_source
closed
[DEFECT] Example 1
Type: Defect
### Summary S 1 ### Steps to reproduce Sr 1 ### Observed outcome observed outcome 1 ### Expected outcome Expected outcome 1 <br/> --- JIRA: CIVIC-1400
1.0
[DEFECT] Example 1 - ### Summary S 1 ### Steps to reproduce Sr 1 ### Observed outcome observed outcome 1 ### Expected outcome Expected outcome 1 <br/> --- JIRA: CIVIC-1400
defect
example summary s steps to reproduce sr observed outcome observed outcome expected outcome expected outcome jira civic
1
33,302
7,086,008,705
IssuesEvent
2018-01-11 13:09:51
HewlettPackard/yoda-demo
https://api.github.com/repos/HewlettPackard/yoda-demo
closed
Issue editing text in Chrome browser.
S4 - Low T1 - Defect
<p>Lorem ipsum erat semper cursus mauris phasellus hendrerit lacus, metus habitant neque in mattis et.</p> <p>Mi scelerisque mauris hac ut est odio curae inceptos purus sapien, eget tristique donec sollicitudin ut per tincidunt magna laoreet.</p> <p>Iaculis porta sed ipsum enim placerat nisi, pellentesque viverra donec et mollis, libero arcu ullamcorper mattis curae.</p> <p>Nulla etiam curabitur nec integer aptent, non tincidunt volutpat netus aliquet, rutrum dictumst posuere malesuada.</p> > estimate 4
1.0
Issue editing text in Chrome browser. - <p>Lorem ipsum erat semper cursus mauris phasellus hendrerit lacus, metus habitant neque in mattis et.</p> <p>Mi scelerisque mauris hac ut est odio curae inceptos purus sapien, eget tristique donec sollicitudin ut per tincidunt magna laoreet.</p> <p>Iaculis porta sed ipsum enim placerat nisi, pellentesque viverra donec et mollis, libero arcu ullamcorper mattis curae.</p> <p>Nulla etiam curabitur nec integer aptent, non tincidunt volutpat netus aliquet, rutrum dictumst posuere malesuada.</p> > estimate 4
defect
issue editing text in chrome browser lorem ipsum erat semper cursus mauris phasellus hendrerit lacus metus habitant neque in mattis et mi scelerisque mauris hac ut est odio curae inceptos purus sapien eget tristique donec sollicitudin ut per tincidunt magna laoreet iaculis porta sed ipsum enim placerat nisi pellentesque viverra donec et mollis libero arcu ullamcorper mattis curae nulla etiam curabitur nec integer aptent non tincidunt volutpat netus aliquet rutrum dictumst posuere malesuada estimate
1
72,023
23,899,583,499
IssuesEvent
2022-09-08 17:29:59
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
[Markup and meta data] HTML markup isn't valid. (09.01.2)
508/Accessibility authenticated-experience profile 508-defect-2 collab-cycle-feedback Staging CCIssue09.01 CC-Dashboard bad-address-indicator
### General Information #### VFS team name Profile Team #### VFS product name Profile #### VFS feature name Bad Address Indicator #### Point of Contact/Reviewers Brian DeConinck (@briandeconinck) - Accessibility *For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website. --- ### Platform Issue HTML markup isn't valid. ### Issue Details In the "Review your mailing address" alerts, the H2 heading has a `role="alert"` added to it. This isn't a role that's typically allowed to be added to a heading, and it may interfere with the findability of the heading for screen reader users. The `role="alert"` should be placed on a different element. ### Link, screenshot or steps to recreate The "Review your mailing address" alerts are currently coded as: ``` <h2 slot="headline" tabindex="0" role="alert" aria-live="polite" aria-label="The address we have on file for you may not be correct.">Review your mailing address</h2> <p>The mailing address we have on file for you may not be correct.</p> <p><a href="...">...</a></p> ``` ### VA.gov Experience Standard [Category Number 09, Issue Number 01](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html) ### Other References WCAG SC 1.3.1_A --- ### Platform Recommendation Headings (H1 through H6) have an implicit [heading role](https://developer.mozilla.org/en-US/docs/Web/Accessibility/ARIA/Roles/heading_role). That role is used by screen readers to identify what the headings are and announce them to users. When you add a different role attribute to a heading... I'm not totally sure what will happen! Testing with NVDA, these H2s are still announced as headings when you encounter them in the page content. But when you tell NVDA to pull up a list of headings on the page, these H2s aren't included on the list. Something weird is happening with the semantics. I think the intention with the `role="alert"` is to make sure the warning about the bad mailing address is announced to screen reader users on page load, and that does happen. But it might be worth testing it with the `role="alert"` added to a different element. If you're not able to add it directly to the `va-alert` component, maybe try adding it to a `div` containing the `va-alert` component. ### VFS Team Tasks to Complete - [ ] Comment on the ticket if there are questions or concerns - [ ] Close the ticket when the issue has been resolved or validated by your Product Owner. If a team has additional questions or needs Platform help validating the issue, please comment in the ticket.
1.0
[Markup and meta data] HTML markup isn't valid. (09.01.2) - ### General Information #### VFS team name Profile Team #### VFS product name Profile #### VFS feature name Bad Address Indicator #### Point of Contact/Reviewers Brian DeConinck (@briandeconinck) - Accessibility *For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website. --- ### Platform Issue HTML markup isn't valid. ### Issue Details In the "Review your mailing address" alerts, the H2 heading has a `role="alert"` added to it. This isn't a role that's typically allowed to be added to a heading, and it may interfere with the findability of the heading for screen reader users. The `role="alert"` should be placed on a different element. ### Link, screenshot or steps to recreate The "Review your mailing address" alerts are currently coded as: ``` <h2 slot="headline" tabindex="0" role="alert" aria-live="polite" aria-label="The address we have on file for you may not be correct.">Review your mailing address</h2> <p>The mailing address we have on file for you may not be correct.</p> <p><a href="...">...</a></p> ``` ### VA.gov Experience Standard [Category Number 09, Issue Number 01](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html) ### Other References WCAG SC 1.3.1_A --- ### Platform Recommendation Headings (H1 through H6) have an implicit [heading role](https://developer.mozilla.org/en-US/docs/Web/Accessibility/ARIA/Roles/heading_role). That role is used by screen readers to identify what the headings are and announce them to users. When you add a different role attribute to a heading... I'm not totally sure what will happen! Testing with NVDA, these H2s are still announced as headings when you encounter them in the page content. But when you tell NVDA to pull up a list of headings on the page, these H2s aren't included on the list. Something weird is happening with the semantics. I think the intention with the `role="alert"` is to make sure the warning about the bad mailing address is announced to screen reader users on page load, and that does happen. But it might be worth testing it with the `role="alert"` added to a different element. If you're not able to add it directly to the `va-alert` component, maybe try adding it to a `div` containing the `va-alert` component. ### VFS Team Tasks to Complete - [ ] Comment on the ticket if there are questions or concerns - [ ] Close the ticket when the issue has been resolved or validated by your Product Owner. If a team has additional questions or needs Platform help validating the issue, please comment in the ticket.
defect
html markup isn t valid general information vfs team name profile team vfs product name profile vfs feature name bad address indicator point of contact reviewers brian deconinck briandeconinck accessibility for more information on how to interpret this ticket please refer to the guidance on platform website platform issue html markup isn t valid issue details in the review your mailing address alerts the heading has a role alert added to it this isn t a role that s typically allowed to be added to a heading and it may interfere with the findability of the heading for screen reader users the role alert should be placed on a different element link screenshot or steps to recreate the review your mailing address alerts are currently coded as review your mailing address the mailing address we have on file for you may not be correct va gov experience standard other references wcag sc a platform recommendation headings through have an implicit that role is used by screen readers to identify what the headings are and announce them to users when you add a different role attribute to a heading i m not totally sure what will happen testing with nvda these are still announced as headings when you encounter them in the page content but when you tell nvda to pull up a list of headings on the page these aren t included on the list something weird is happening with the semantics i think the intention with the role alert is to make sure the warning about the bad mailing address is announced to screen reader users on page load and that does happen but it might be worth testing it with the role alert added to a different element if you re not able to add it directly to the va alert component maybe try adding it to a div containing the va alert component vfs team tasks to complete comment on the ticket if there are questions or concerns close the ticket when the issue has been resolved or validated by your product owner if a team has additional questions or needs platform help validating the issue please comment in the ticket
1
60,657
17,023,484,491
IssuesEvent
2021-07-03 02:16:06
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
index out of range error (abort and close the application)
Component: merkaartor Priority: major Resolution: worksforme Type: defect
**[Submitted to the original trac issue database at 7.04am, Sunday, 27th September 2009]** i cant reproduce it again. i do several mods in the map and press upload button: ``` -58.3594 , -34.6784 , -58.3539 , -34.6829 getting: "qrc:/Html/ymap.html?WIDTH=512&HEIGHT=512&BBOX=-58.353882,-34.669359,-58.348389,-34.673876" 0 , 0 , 512 , 511 512 , 511 -58.3539 , -34.6694 , -58.3484 , -34.6739 getting: "qrc:/Html/ymap.html?WIDTH=512&HEIGHT=512&BBOX=-58.353882,-34.673876,-58.348389,-34.678394" 0 , 0 , 512 , 512 512 , 512 -58.3539 , -34.6739 , -58.3484 , -34.6784 getting: "qrc:/Html/ymap.html?WIDTH=512&HEIGHT=512&BBOX=-58.353882,-34.678394,-58.348389,-34.682911" 0 , 0 , 512 , 512 512 , 512 -58.3539 , -34.6784 , -58.3484 , -34.6829 ASSERT failure in QList<T>::operator[]: "index out of range", file /usr/include/qt4/QtCore/qlist.h, line 403 Abortado ``` Running version 0.13.2 on Debian testing amd64
1.0
index out of range error (abort and close the application) - **[Submitted to the original trac issue database at 7.04am, Sunday, 27th September 2009]** i cant reproduce it again. i do several mods in the map and press upload button: ``` -58.3594 , -34.6784 , -58.3539 , -34.6829 getting: "qrc:/Html/ymap.html?WIDTH=512&HEIGHT=512&BBOX=-58.353882,-34.669359,-58.348389,-34.673876" 0 , 0 , 512 , 511 512 , 511 -58.3539 , -34.6694 , -58.3484 , -34.6739 getting: "qrc:/Html/ymap.html?WIDTH=512&HEIGHT=512&BBOX=-58.353882,-34.673876,-58.348389,-34.678394" 0 , 0 , 512 , 512 512 , 512 -58.3539 , -34.6739 , -58.3484 , -34.6784 getting: "qrc:/Html/ymap.html?WIDTH=512&HEIGHT=512&BBOX=-58.353882,-34.678394,-58.348389,-34.682911" 0 , 0 , 512 , 512 512 , 512 -58.3539 , -34.6784 , -58.3484 , -34.6829 ASSERT failure in QList<T>::operator[]: "index out of range", file /usr/include/qt4/QtCore/qlist.h, line 403 Abortado ``` Running version 0.13.2 on Debian testing amd64
defect
index out of range error abort and close the application i cant reproduce it again i do several mods in the map and press upload button getting qrc html ymap html width height bbox getting qrc html ymap html width height bbox getting qrc html ymap html width height bbox assert failure in qlist operator index out of range file usr include qtcore qlist h line abortado running version on debian testing
1
661,366
22,051,469,595
IssuesEvent
2022-05-30 09:02:52
googleapis/google-api-nodejs-client
https://api.github.com/repos/googleapis/google-api-nodejs-client
opened
Add support for `androidpublisher.subscriptionsv2`
type: feature request priority: p3
The `androidpublisher` API now includes a `subscriptionsv2` resource that HAS to be used to validate purchases made with Android Billing v5 API released this month. https://developers.google.com/android-publisher/api-ref/rest/v3/purchases.subscriptionsv2/get For information, as a workaround I use a "quick&dirty" polyfill inspired by the existing `subscriptions.get`, that I set as `subscriptions.getv2`... Obviously this should be part of this library. Here's the code. ```js import * as googleapis_common_1 from 'googleapis-common'; androidpublisher.purchases.subscriptions.getv2 = function(paramsOrCallback, optionsOrCallback, callback) { let params = (paramsOrCallback || {}); let options = (optionsOrCallback || {}); if (typeof paramsOrCallback === 'function') { callback = paramsOrCallback; params = {}; options = {}; } if (typeof optionsOrCallback === 'function') { callback = optionsOrCallback; options = {}; } if (params.subscriptionId) delete params.subscriptionId; const rootUrl = options.rootUrl || 'https://androidpublisher.googleapis.com/'; const parameters = { options: Object.assign({ url: (rootUrl + '/androidpublisher/v3/applications/{packageName}/purchases/subscriptionsv2/tokens/{token}').replace(/([^:]\/)\/+/g, '$1'), method: 'GET', }, options), params, requiredParams: ['packageName', 'token'], pathParams: ['packageName', 'token'], context: this.context, }; if (callback) { googleapis_common_1.createAPIRequest(parameters, callback); } else { return googleapis_common_1.createAPIRequest(parameters); } } ```
1.0
Add support for `androidpublisher.subscriptionsv2` - The `androidpublisher` API now includes a `subscriptionsv2` resource that HAS to be used to validate purchases made with Android Billing v5 API released this month. https://developers.google.com/android-publisher/api-ref/rest/v3/purchases.subscriptionsv2/get For information, as a workaround I use a "quick&dirty" polyfill inspired by the existing `subscriptions.get`, that I set as `subscriptions.getv2`... Obviously this should be part of this library. Here's the code. ```js import * as googleapis_common_1 from 'googleapis-common'; androidpublisher.purchases.subscriptions.getv2 = function(paramsOrCallback, optionsOrCallback, callback) { let params = (paramsOrCallback || {}); let options = (optionsOrCallback || {}); if (typeof paramsOrCallback === 'function') { callback = paramsOrCallback; params = {}; options = {}; } if (typeof optionsOrCallback === 'function') { callback = optionsOrCallback; options = {}; } if (params.subscriptionId) delete params.subscriptionId; const rootUrl = options.rootUrl || 'https://androidpublisher.googleapis.com/'; const parameters = { options: Object.assign({ url: (rootUrl + '/androidpublisher/v3/applications/{packageName}/purchases/subscriptionsv2/tokens/{token}').replace(/([^:]\/)\/+/g, '$1'), method: 'GET', }, options), params, requiredParams: ['packageName', 'token'], pathParams: ['packageName', 'token'], context: this.context, }; if (callback) { googleapis_common_1.createAPIRequest(parameters, callback); } else { return googleapis_common_1.createAPIRequest(parameters); } } ```
non_defect
add support for androidpublisher the androidpublisher api now includes a resource that has to be used to validate purchases made with android billing api released this month for information as a workaround i use a quick dirty polyfill inspired by the existing subscriptions get that i set as subscriptions obviously this should be part of this library here s the code js import as googleapis common from googleapis common androidpublisher purchases subscriptions function paramsorcallback optionsorcallback callback let params paramsorcallback let options optionsorcallback if typeof paramsorcallback function callback paramsorcallback params options if typeof optionsorcallback function callback optionsorcallback options if params subscriptionid delete params subscriptionid const rooturl options rooturl const parameters options object assign url rooturl androidpublisher applications packagename purchases tokens token replace g method get options params requiredparams pathparams context this context if callback googleapis common createapirequest parameters callback else return googleapis common createapirequest parameters
0
293,726
25,318,793,132
IssuesEvent
2022-11-18 00:47:24
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
It is better to remove the checkbox "Don't ask again about enabling soft delete" from the 'Delete' dialog if the soft delete is already enabled
🧪 testing :gear: blobs :gear: adls gen2
**Storage Explorer Version**: 1.27.0-dev **Build Number**: 20221117.5 **Branch**: main **Platform/OS**: Windows 10/Linux Ubuntu 18.04/MacOS Ventura 13.0 (Apple M1 Pro) **Architecture**: ia32/x64 **How Found**: Ad-hoc testing **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand one storage account -> Blob Containers. (Soft delete is enabled) 2. Create a blob container -> Upload one blob. 3. Right click the blob -> Click 'Delete'. 4. Check there is no checkbox "Don't ask again about enabling soft delete" on the dialog. ## Expected Experience ## It is better to remove the checkbox "Don't ask again about enabling soft delete" from the 'Delete' dialog. ## Actual Experience ## There is a checkbox "Don't ask again about enabling soft delete" on the 'Delete' dialog. ![image](https://user-images.githubusercontent.com/41351993/202403273-08f1c39f-1118-47ba-809f-bb2f18a7af92.png) ## Addition Context ## This issue also reproduces for ADLS Gen2 blobs when the setting "Enable ADLS Gen2 soft delete (Preview)" is disabled.
1.0
It is better to remove the checkbox "Don't ask again about enabling soft delete" from the 'Delete' dialog if the soft delete is already enabled - **Storage Explorer Version**: 1.27.0-dev **Build Number**: 20221117.5 **Branch**: main **Platform/OS**: Windows 10/Linux Ubuntu 18.04/MacOS Ventura 13.0 (Apple M1 Pro) **Architecture**: ia32/x64 **How Found**: Ad-hoc testing **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand one storage account -> Blob Containers. (Soft delete is enabled) 2. Create a blob container -> Upload one blob. 3. Right click the blob -> Click 'Delete'. 4. Check there is no checkbox "Don't ask again about enabling soft delete" on the dialog. ## Expected Experience ## It is better to remove the checkbox "Don't ask again about enabling soft delete" from the 'Delete' dialog. ## Actual Experience ## There is a checkbox "Don't ask again about enabling soft delete" on the 'Delete' dialog. ![image](https://user-images.githubusercontent.com/41351993/202403273-08f1c39f-1118-47ba-809f-bb2f18a7af92.png) ## Addition Context ## This issue also reproduces for ADLS Gen2 blobs when the setting "Enable ADLS Gen2 soft delete (Preview)" is disabled.
non_defect
it is better to remove the checkbox don t ask again about enabling soft delete from the delete dialog if the soft delete is already enabled storage explorer version dev build number branch main platform os windows linux ubuntu macos ventura apple pro architecture how found ad hoc testing regression from not a regression steps to reproduce expand one storage account blob containers soft delete is enabled create a blob container upload one blob right click the blob click delete check there is no checkbox don t ask again about enabling soft delete on the dialog expected experience it is better to remove the checkbox don t ask again about enabling soft delete from the delete dialog actual experience there is a checkbox don t ask again about enabling soft delete on the delete dialog addition context this issue also reproduces for adls blobs when the setting enable adls soft delete preview is disabled
0
80,795
10,210,545,326
IssuesEvent
2019-08-14 15:00:00
kazukinakamuraKN/realtime_chat_nuxt
https://api.github.com/repos/kazukinakamuraKN/realtime_chat_nuxt
opened
計画
documentation
### 機能概要 リアルタイムチャット ### 機能詳細 - リアルタイムにチャットを行う - チャットの文末に自動で絵文字を追加 - 末尾を認識して、句読点orドットなら削除 - 末尾に絵文字を追加する ### 作業概要 1. チャット機能を追加 1. 絵文字機能を追加 ### タスク - [ ] チャット機能を追加 - https://socket.io/get-started/chat - チャット機能追加 - [ ] 絵文字機能を追加 - https://www.npmjs.com/package/rm-emoji-picker - 絵文字pickerを追加する ### チャレンジ - [ ] チャット機能に追加要素を加える - https://socket.io/get-started/chat
1.0
計画 - ### 機能概要 リアルタイムチャット ### 機能詳細 - リアルタイムにチャットを行う - チャットの文末に自動で絵文字を追加 - 末尾を認識して、句読点orドットなら削除 - 末尾に絵文字を追加する ### 作業概要 1. チャット機能を追加 1. 絵文字機能を追加 ### タスク - [ ] チャット機能を追加 - https://socket.io/get-started/chat - チャット機能追加 - [ ] 絵文字機能を追加 - https://www.npmjs.com/package/rm-emoji-picker - 絵文字pickerを追加する ### チャレンジ - [ ] チャット機能に追加要素を加える - https://socket.io/get-started/chat
non_defect
計画 機能概要 リアルタイムチャット 機能詳細 リアルタイムにチャットを行う チャットの文末に自動で絵文字を追加 末尾を認識して、句読点orドットなら削除 末尾に絵文字を追加する 作業概要 チャット機能を追加 絵文字機能を追加 タスク チャット機能を追加 チャット機能追加 絵文字機能を追加 絵文字pickerを追加する チャレンジ チャット機能に追加要素を加える
0
27,754
5,095,493,914
IssuesEvent
2017-01-03 15:23:37
NewSpring/Holtzman
https://api.github.com/repos/NewSpring/Holtzman
closed
Audio player modal doesn't load
Defect
## Steps to Reproduce 1. start playing audio in the app 2. Go anywhere 3. Tap the audio banner at the bottom of the screen ### Buggy Behavior The modal doesn't actually load. It hides the nav and the audio banner appropriately, but nothing else. ### Expected Behavior It should open the now playing modal
1.0
Audio player modal doesn't load - ## Steps to Reproduce 1. start playing audio in the app 2. Go anywhere 3. Tap the audio banner at the bottom of the screen ### Buggy Behavior The modal doesn't actually load. It hides the nav and the audio banner appropriately, but nothing else. ### Expected Behavior It should open the now playing modal
defect
audio player modal doesn t load steps to reproduce start playing audio in the app go anywhere tap the audio banner at the bottom of the screen buggy behavior the modal doesn t actually load it hides the nav and the audio banner appropriately but nothing else expected behavior it should open the now playing modal
1
86,675
8,042,452,150
IssuesEvent
2018-07-31 08:12:03
ClassicWoW/Nefarian_1.12.1_Bugtracker
https://api.github.com/repos/ClassicWoW/Nefarian_1.12.1_Bugtracker
closed
Natalia Mar'alith benutzt ihre Priester-Fähigkeiten nicht.
Mehr Input/Recherche/Tests nötig
**Welches Verhalten wird beobachtet?** Der NPC "Natalia Mar'alith" verhält sich wie ein normaler Nahkampf-NPC ohne spezielle Fähigkeiten. **Wie sollte es sich verhalten?** Dieser NPC sollte zusätzlich Schattenpriester-Fähigkeiten haben plus die Fähigkeit sich selbst zu heilen. **Schritte zur Reproduzierung** Natalia Mar'alith angreifen. **Zusätzliche Informationen (Screenshots, Videos, Klasse, Rasse, Level, etc.)** Hier sind die Fähigkeiten des NPCs beschrieben: http://wowwiki.wikia.com/wiki/Natalia_Mar%27alith Hier gibt es einen Kommentar von 2006, dass sich dieser NPC selbst heilt: https://wow.freierbund.de/map_and_guide/163325--1/Der_Schlund_des_Wahnsinns-Kommentare.html#add Das ist in den Fähigkeiten des NPCs in der classic-wow-Datenbank nämlich nicht eingetragen. Leider steht nicht drin wie hoch die Heilung ausfällt. Und Videos habe ich keine gefunden von 2005-2006. **Alle Kreaturen, Items, Objekte, Quests, Zauber etc. müssen aus unserer Datenbank verlinkt sein.** Der NPC: https://datenbank.classic-wow.org/?npc=15215
1.0
Natalia Mar'alith benutzt ihre Priester-Fähigkeiten nicht. - **Welches Verhalten wird beobachtet?** Der NPC "Natalia Mar'alith" verhält sich wie ein normaler Nahkampf-NPC ohne spezielle Fähigkeiten. **Wie sollte es sich verhalten?** Dieser NPC sollte zusätzlich Schattenpriester-Fähigkeiten haben plus die Fähigkeit sich selbst zu heilen. **Schritte zur Reproduzierung** Natalia Mar'alith angreifen. **Zusätzliche Informationen (Screenshots, Videos, Klasse, Rasse, Level, etc.)** Hier sind die Fähigkeiten des NPCs beschrieben: http://wowwiki.wikia.com/wiki/Natalia_Mar%27alith Hier gibt es einen Kommentar von 2006, dass sich dieser NPC selbst heilt: https://wow.freierbund.de/map_and_guide/163325--1/Der_Schlund_des_Wahnsinns-Kommentare.html#add Das ist in den Fähigkeiten des NPCs in der classic-wow-Datenbank nämlich nicht eingetragen. Leider steht nicht drin wie hoch die Heilung ausfällt. Und Videos habe ich keine gefunden von 2005-2006. **Alle Kreaturen, Items, Objekte, Quests, Zauber etc. müssen aus unserer Datenbank verlinkt sein.** Der NPC: https://datenbank.classic-wow.org/?npc=15215
non_defect
natalia mar alith benutzt ihre priester fähigkeiten nicht welches verhalten wird beobachtet der npc natalia mar alith verhält sich wie ein normaler nahkampf npc ohne spezielle fähigkeiten wie sollte es sich verhalten dieser npc sollte zusätzlich schattenpriester fähigkeiten haben plus die fähigkeit sich selbst zu heilen schritte zur reproduzierung natalia mar alith angreifen zusätzliche informationen screenshots videos klasse rasse level etc hier sind die fähigkeiten des npcs beschrieben hier gibt es einen kommentar von dass sich dieser npc selbst heilt das ist in den fähigkeiten des npcs in der classic wow datenbank nämlich nicht eingetragen leider steht nicht drin wie hoch die heilung ausfällt und videos habe ich keine gefunden von alle kreaturen items objekte quests zauber etc müssen aus unserer datenbank verlinkt sein der npc
0
62,286
17,023,889,531
IssuesEvent
2021-07-03 04:23:30
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Text above map page overlaps
Component: website Priority: minor Resolution: duplicate Type: defect
**[Submitted to the original trac issue database at 7.41am, Wednesday, 4th December 2013]** The text (traces, diaries, etc.) and boxes (edit, history, etc.) above of the map overlap when the browserscreen is too narrow to support them side by side. When the browserscreen is narrowed even more, the links to traces, diaries etc disappear to be replaced by a "more" dropdown box. This dropdown box needs to be shown earlier to prevent the overlap. See attached screenshots.
1.0
Text above map page overlaps - **[Submitted to the original trac issue database at 7.41am, Wednesday, 4th December 2013]** The text (traces, diaries, etc.) and boxes (edit, history, etc.) above of the map overlap when the browserscreen is too narrow to support them side by side. When the browserscreen is narrowed even more, the links to traces, diaries etc disappear to be replaced by a "more" dropdown box. This dropdown box needs to be shown earlier to prevent the overlap. See attached screenshots.
defect
text above map page overlaps the text traces diaries etc and boxes edit history etc above of the map overlap when the browserscreen is too narrow to support them side by side when the browserscreen is narrowed even more the links to traces diaries etc disappear to be replaced by a more dropdown box this dropdown box needs to be shown earlier to prevent the overlap see attached screenshots
1
35,136
7,595,689,500
IssuesEvent
2018-04-27 06:51:44
STEllAR-GROUP/hpx
https://api.github.com/repos/STEllAR-GROUP/hpx
opened
Setting pu step or offset results in no threads in default pool
category: resource_partitioner type: defect
I expect the following to not fail: ``` ./hello_world --hpx:pu-offset=2 terminate called after throwing an instance of 'std::runtime_error' what(): partitioner::setup_pools: Default pool default has no threads assigned. Please rerun with --hpx:threads=X and check the pool thread assignment zsh: abort (core dumped) ./solution6 --hpx:pu-offset=2 ``` Same goes for `pu-step`.
1.0
Setting pu step or offset results in no threads in default pool - I expect the following to not fail: ``` ./hello_world --hpx:pu-offset=2 terminate called after throwing an instance of 'std::runtime_error' what(): partitioner::setup_pools: Default pool default has no threads assigned. Please rerun with --hpx:threads=X and check the pool thread assignment zsh: abort (core dumped) ./solution6 --hpx:pu-offset=2 ``` Same goes for `pu-step`.
defect
setting pu step or offset results in no threads in default pool i expect the following to not fail hello world hpx pu offset terminate called after throwing an instance of std runtime error what partitioner setup pools default pool default has no threads assigned please rerun with hpx threads x and check the pool thread assignment zsh abort core dumped hpx pu offset same goes for pu step
1
20,068
3,294,338,448
IssuesEvent
2015-10-31 00:30:50
dkfans/keeperfx
https://api.github.com/repos/dkfans/keeperfx
opened
Max efficiency on single treasure room tile (stable)
Priority-Medium Type-Defect
See the attached save game for some treasure room tiles with max efficiency. Being a single room tile they should have very low efficiency. When you sell the bridge tiles around them they get reduced efficiency, but still more than they should. ![fx1g0003.7z](https://cloud.githubusercontent.com/assets/13840686/10860809/9a0fc91c-7f6e-11e5-85a9-87ee50774442.jpg) From 0.4.6.1723(stable branch)
1.0
Max efficiency on single treasure room tile (stable) - See the attached save game for some treasure room tiles with max efficiency. Being a single room tile they should have very low efficiency. When you sell the bridge tiles around them they get reduced efficiency, but still more than they should. ![fx1g0003.7z](https://cloud.githubusercontent.com/assets/13840686/10860809/9a0fc91c-7f6e-11e5-85a9-87ee50774442.jpg) From 0.4.6.1723(stable branch)
defect
max efficiency on single treasure room tile stable see the attached save game for some treasure room tiles with max efficiency being a single room tile they should have very low efficiency when you sell the bridge tiles around them they get reduced efficiency but still more than they should from stable branch
1
53,507
13,172,029,961
IssuesEvent
2020-08-11 17:42:34
netlify/build
https://api.github.com/repos/netlify/build
closed
Add `--statsd.host` and `--statsd.port` CLI flags to `@netlify/config`
feat/speedy-builds type: feature
Background at https://github.com/netlify/buildbot/issues/891 and https://github.com/netlify/buildbot/issues/900 See also https://github.com/netlify/build/issues/1741 We should add a `--statsd.host` and `--statsd.port` CLI flags to `@netlify/build`. Until the buildbot is passing those values to `@netlify/build`, we can hardcode those values.
1.0
Add `--statsd.host` and `--statsd.port` CLI flags to `@netlify/config` - Background at https://github.com/netlify/buildbot/issues/891 and https://github.com/netlify/buildbot/issues/900 See also https://github.com/netlify/build/issues/1741 We should add a `--statsd.host` and `--statsd.port` CLI flags to `@netlify/build`. Until the buildbot is passing those values to `@netlify/build`, we can hardcode those values.
non_defect
add statsd host and statsd port cli flags to netlify config background at and see also we should add a statsd host and statsd port cli flags to netlify build until the buildbot is passing those values to netlify build we can hardcode those values
0
81,615
31,148,136,701
IssuesEvent
2023-08-16 08:05:48
MarcusWolschon/osmeditor4android
https://api.github.com/repos/MarcusWolschon/osmeditor4android
closed
Pending changes count background uncoloured on Android 5 and 6
Defect Minor
On an old (2017) tablet, the Transfer icon's pending changes count lacks a coloured background, making the number difficult to read when the Theme is Light due to the low contrast. It's legible enough in Dark Theme but text isn't coloured. ## Vespucci Version 19.1 BETA 1 ## Download source Google Play Store ## Device (Manufacturer and Model) Amazon Kindle Fire HD 10 Gen 7 [2017] ## Android Version Fire OS 5.7.0.0 ## Behaviour/Symptoms The pending changes count's text is white, but without a coloured background, its visibility on Light Theme has contrast low enough to be barely visible. ## Expected Behaviour A coloured background or at least bold/thick digits with text coloured appropriately for themed contrast. ## How to recreate 1. Launch Vespucci on an old Kindle Fire tablet. 2. Set theme within Advanced Preferences to Light (if not already set). 2. Download data. 3. Move a node or way. 4. Observe the pending changes count on the transfer (↑↓) icon. ## Crash dump submitted (no or yes + date) no ## Any other potentially relevant information Not an issue on an Android 10 phone (LG V50) or Android 13 phone (Samsung Galaxy Z Fold3).
1.0
Pending changes count background uncoloured on Android 5 and 6 - On an old (2017) tablet, the Transfer icon's pending changes count lacks a coloured background, making the number difficult to read when the Theme is Light due to the low contrast. It's legible enough in Dark Theme but text isn't coloured. ## Vespucci Version 19.1 BETA 1 ## Download source Google Play Store ## Device (Manufacturer and Model) Amazon Kindle Fire HD 10 Gen 7 [2017] ## Android Version Fire OS 5.7.0.0 ## Behaviour/Symptoms The pending changes count's text is white, but without a coloured background, its visibility on Light Theme has contrast low enough to be barely visible. ## Expected Behaviour A coloured background or at least bold/thick digits with text coloured appropriately for themed contrast. ## How to recreate 1. Launch Vespucci on an old Kindle Fire tablet. 2. Set theme within Advanced Preferences to Light (if not already set). 2. Download data. 3. Move a node or way. 4. Observe the pending changes count on the transfer (↑↓) icon. ## Crash dump submitted (no or yes + date) no ## Any other potentially relevant information Not an issue on an Android 10 phone (LG V50) or Android 13 phone (Samsung Galaxy Z Fold3).
defect
pending changes count background uncoloured on android and on an old tablet the transfer icon s pending changes count lacks a coloured background making the number difficult to read when the theme is light due to the low contrast it s legible enough in dark theme but text isn t coloured vespucci version beta download source google play store device manufacturer and model amazon kindle fire hd gen android version fire os behaviour symptoms the pending changes count s text is white but without a coloured background its visibility on light theme has contrast low enough to be barely visible expected behaviour a coloured background or at least bold thick digits with text coloured appropriately for themed contrast how to recreate launch vespucci on an old kindle fire tablet set theme within advanced preferences to light if not already set download data move a node or way observe the pending changes count on the transfer ↑↓ icon crash dump submitted no or yes date no any other potentially relevant information not an issue on an android phone lg or android phone samsung galaxy z
1
230,235
17,606,905,961
IssuesEvent
2021-08-17 18:20:10
OHDSI/CommonDataModel
https://api.github.com/repos/OHDSI/CommonDataModel
closed
Article on how to handle drug infusions in the CDM
Documentation
We need an article for the CDM website on how to represent drug infusions in a standard way
1.0
Article on how to handle drug infusions in the CDM - We need an article for the CDM website on how to represent drug infusions in a standard way
non_defect
article on how to handle drug infusions in the cdm we need an article for the cdm website on how to represent drug infusions in a standard way
0
3,895
4,700,624,250
IssuesEvent
2016-10-12 19:04:57
UPDDI/mps-database-server
https://api.github.com/repos/UPDDI/mps-database-server
opened
Refactor Forms to be less Verbose
infrastructure refactor
Currently, some fields are written in a very verbose way in the templates. Ideally, we should have a DRY (Don't Repeat Yourself) solution for this. Initial work has been made to have a file for "generic fields," but it is not totally robust and is not employed everywhere in can be used. Here are some considerations: - We should be able to group fields together - We should be able to change the column sizes for the label and the input - It should make use of as much from the form/view as possible - Ideally it should not have labels etc. that diverge from the models
1.0
Refactor Forms to be less Verbose - Currently, some fields are written in a very verbose way in the templates. Ideally, we should have a DRY (Don't Repeat Yourself) solution for this. Initial work has been made to have a file for "generic fields," but it is not totally robust and is not employed everywhere in can be used. Here are some considerations: - We should be able to group fields together - We should be able to change the column sizes for the label and the input - It should make use of as much from the form/view as possible - Ideally it should not have labels etc. that diverge from the models
non_defect
refactor forms to be less verbose currently some fields are written in a very verbose way in the templates ideally we should have a dry don t repeat yourself solution for this initial work has been made to have a file for generic fields but it is not totally robust and is not employed everywhere in can be used here are some considerations we should be able to group fields together we should be able to change the column sizes for the label and the input it should make use of as much from the form view as possible ideally it should not have labels etc that diverge from the models
0
84,951
15,728,371,772
IssuesEvent
2021-03-29 13:45:08
ssobue/oauth2-provider
https://api.github.com/repos/ssobue/oauth2-provider
closed
CVE-2020-10672 (High) detected in jackson-databind-2.9.9.jar
security vulnerability
## CVE-2020-10672 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /oauth2-provider/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library) - spring-boot-starter-json-2.1.5.RELEASE.jar - :x: **jackson-databind-2.9.9.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.aries.transaction.jms.internal.XaPooledConnectionFactory (aka aries.transaction.jms). <p>Publish Date: 2020-03-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10672>CVE-2020-10672</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-10672">https://nvd.nist.gov/vuln/detail/CVE-2020-10672</a></p> <p>Release Date: 2020-03-18</p> <p>Fix Resolution: jackson-databind-2.9.10.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-10672 (High) detected in jackson-databind-2.9.9.jar - ## CVE-2020-10672 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /oauth2-provider/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library) - spring-boot-starter-json-2.1.5.RELEASE.jar - :x: **jackson-databind-2.9.9.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.aries.transaction.jms.internal.XaPooledConnectionFactory (aka aries.transaction.jms). <p>Publish Date: 2020-03-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10672>CVE-2020-10672</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-10672">https://nvd.nist.gov/vuln/detail/CVE-2020-10672</a></p> <p>Release Date: 2020-03-18</p> <p>Fix Resolution: jackson-databind-2.9.10.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file provider pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache aries transaction jms internal xapooledconnectionfactory aka aries transaction jms publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jackson databind step up your open source security game with whitesource
0
43,111
11,487,846,780
IssuesEvent
2020-02-11 12:49:18
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
closed
PDBWriter incompatible with StringIO if atom coordinates go out of range
Component-Writers Format-PDB defect
**Expected behavior** Assuming that PDBWriter is supposed to support a StringIO object as a valid destination, I expect that attempting to write a PDB with atom coordinates out of range will raise the "PDB files must have coordinate values between [...]" exception. **Actual behavior** The check attempts to `os.remove` the StringIO object and does not catch the resulting exception, so the coordinate exception never gets raised. ``` ~/narupa-protocols/narupa-protocol/python-libraries/narupa-ase/examples/nglclient.py in mda_to_pdb_str(universe) 54 def mda_to_pdb_str(universe: mda.Universe): 55 with StringIO() as str_io, mda.coordinates.PDB.PDBWriter(str_io) as writer: ---> 56 writer.write(universe.atoms) 57 pdb = str_io.getvalue() 58 return pdb ~/anaconda3/envs/narupa-dev/lib/python3.7/site-packages/MDAnalysis/coordinates/PDB.py in write(self, obj) 765 # write_all_timesteps() to dump everything in one go, or do the 766 # traditional loop over frames --> 767 self.write_next_timestep(self.ts, multiframe=self._multiframe) 768 self._write_pdb_bonds() 769 # END record is written when file is being close()d ~/anaconda3/envs/narupa-dev/lib/python3.7/site-packages/MDAnalysis/coordinates/PDB.py in write_next_timestep(self, ts, **kwargs) 839 raise NoDataError("PBDWriter: no coordinate data to write to " 840 "trajectory file") --> 841 self._check_pdb_coordinates() 842 self._write_timestep(ts, **kwargs) 843 ~/anaconda3/envs/narupa-dev/lib/python3.7/site-packages/MDAnalysis/coordinates/PDB.py in _check_pdb_coordinates(self) 670 self.close() 671 try: --> 672 os.remove(self.filename) 673 except OSError as err: 674 if err.errno == errno.ENOENT: TypeError: remove: path should be string, bytes or os.PathLike, not _io.StringIO ``` **Code to reproduce the behavior** Show us how to reproduce the failiure. If you can, use trajectory files from the test data. ``` import io import MDAnalysis as mda from MDAnalysis.tests.datafiles import PDB u = mda.Universe(PDB) u.atoms.translate([-9999, -9999, -9999]) with io.StringIO() as str_io, mda.coordinates.PDB.PDBWriter(str_io) as writer: #writer.filename = "" # uncomment to avoid issue writer.write(u.atoms) ``` **Currently version of MDAnalysis** - Which version are you using? (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) 0.20.1 - Which version of Python (`python -V`)? Python 3.7.3 - Which operating system? Windows 10
1.0
PDBWriter incompatible with StringIO if atom coordinates go out of range - **Expected behavior** Assuming that PDBWriter is supposed to support a StringIO object as a valid destination, I expect that attempting to write a PDB with atom coordinates out of range will raise the "PDB files must have coordinate values between [...]" exception. **Actual behavior** The check attempts to `os.remove` the StringIO object and does not catch the resulting exception, so the coordinate exception never gets raised. ``` ~/narupa-protocols/narupa-protocol/python-libraries/narupa-ase/examples/nglclient.py in mda_to_pdb_str(universe) 54 def mda_to_pdb_str(universe: mda.Universe): 55 with StringIO() as str_io, mda.coordinates.PDB.PDBWriter(str_io) as writer: ---> 56 writer.write(universe.atoms) 57 pdb = str_io.getvalue() 58 return pdb ~/anaconda3/envs/narupa-dev/lib/python3.7/site-packages/MDAnalysis/coordinates/PDB.py in write(self, obj) 765 # write_all_timesteps() to dump everything in one go, or do the 766 # traditional loop over frames --> 767 self.write_next_timestep(self.ts, multiframe=self._multiframe) 768 self._write_pdb_bonds() 769 # END record is written when file is being close()d ~/anaconda3/envs/narupa-dev/lib/python3.7/site-packages/MDAnalysis/coordinates/PDB.py in write_next_timestep(self, ts, **kwargs) 839 raise NoDataError("PBDWriter: no coordinate data to write to " 840 "trajectory file") --> 841 self._check_pdb_coordinates() 842 self._write_timestep(ts, **kwargs) 843 ~/anaconda3/envs/narupa-dev/lib/python3.7/site-packages/MDAnalysis/coordinates/PDB.py in _check_pdb_coordinates(self) 670 self.close() 671 try: --> 672 os.remove(self.filename) 673 except OSError as err: 674 if err.errno == errno.ENOENT: TypeError: remove: path should be string, bytes or os.PathLike, not _io.StringIO ``` **Code to reproduce the behavior** Show us how to reproduce the failiure. If you can, use trajectory files from the test data. ``` import io import MDAnalysis as mda from MDAnalysis.tests.datafiles import PDB u = mda.Universe(PDB) u.atoms.translate([-9999, -9999, -9999]) with io.StringIO() as str_io, mda.coordinates.PDB.PDBWriter(str_io) as writer: #writer.filename = "" # uncomment to avoid issue writer.write(u.atoms) ``` **Currently version of MDAnalysis** - Which version are you using? (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) 0.20.1 - Which version of Python (`python -V`)? Python 3.7.3 - Which operating system? Windows 10
defect
pdbwriter incompatible with stringio if atom coordinates go out of range expected behavior assuming that pdbwriter is supposed to support a stringio object as a valid destination i expect that attempting to write a pdb with atom coordinates out of range will raise the pdb files must have coordinate values between exception actual behavior the check attempts to os remove the stringio object and does not catch the resulting exception so the coordinate exception never gets raised narupa protocols narupa protocol python libraries narupa ase examples nglclient py in mda to pdb str universe def mda to pdb str universe mda universe with stringio as str io mda coordinates pdb pdbwriter str io as writer writer write universe atoms pdb str io getvalue return pdb envs narupa dev lib site packages mdanalysis coordinates pdb py in write self obj write all timesteps to dump everything in one go or do the traditional loop over frames self write next timestep self ts multiframe self multiframe self write pdb bonds end record is written when file is being close d envs narupa dev lib site packages mdanalysis coordinates pdb py in write next timestep self ts kwargs raise nodataerror pbdwriter no coordinate data to write to trajectory file self check pdb coordinates self write timestep ts kwargs envs narupa dev lib site packages mdanalysis coordinates pdb py in check pdb coordinates self self close try os remove self filename except oserror as err if err errno errno enoent typeerror remove path should be string bytes or os pathlike not io stringio code to reproduce the behavior show us how to reproduce the failiure if you can use trajectory files from the test data import io import mdanalysis as mda from mdanalysis tests datafiles import pdb u mda universe pdb u atoms translate with io stringio as str io mda coordinates pdb pdbwriter str io as writer writer filename uncomment to avoid issue writer write u atoms currently version of mdanalysis which version are you using run python c import mdanalysis as mda print mda version which version of python python v python which operating system windows
1
465,993
13,395,784,188
IssuesEvent
2020-09-03 08:57:13
googleapis/java-redis
https://api.github.com/repos/googleapis/java-redis
closed
Synthesis failed for java-redis
api: redis autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate java-redis. :broken_heart: Here's the output from running `synth.py`: ``` hl Saved ./Jinja2-2.11.2-py2.py3-none-any.whl Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5)) Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6)) Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7)) Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz Complete output from command python setup.py egg_info: Traceback (most recent call last): File "<string>", line 1, in <module> File "/tmpfs/tmp/tmp6m3w5les/setuptools-tmp/setuptools/__init__.py", line 6, in <module> import distutils.core File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 82, in create_module return importlib.import_module('._distutils', 'setuptools') File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) ModuleNotFoundError: No module named 'setuptools._distutils' ---------------------------------------- (Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-rc4r6uhv/pypandoc/ ) ERROR: no such package '@gapic_generator_python_pip_deps//': pip_import failed: Collecting click==7.1.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 1)) Using cached https://files.pythonhosted.org/packages/d2/3d/fa76db83bf75c4f8d338c2fd15c8d33fdd7ad23a9b5e57eb6c5de26b430e/click-7.1.2-py2.py3-none-any.whl Saved ./click-7.1.2-py2.py3-none-any.whl Collecting google-api-core==1.22.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 2)) Using cached https://files.pythonhosted.org/packages/e0/2d/7c6c75013105e1d2b6eaa1bf18a56995be1dbc673c38885aea31136e9918/google_api_core-1.22.1-py2.py3-none-any.whl Saved ./google_api_core-1.22.1-py2.py3-none-any.whl Collecting googleapis-common-protos==1.52.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 3)) Using cached https://files.pythonhosted.org/packages/03/74/3956721ea1eb4bcf7502a311fdaa60b85bd751de4e57d1943afe9b334141/googleapis_common_protos-1.52.0-py2.py3-none-any.whl Saved ./googleapis_common_protos-1.52.0-py2.py3-none-any.whl Collecting jinja2==2.11.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 4)) Using cached https://files.pythonhosted.org/packages/30/9e/f663a2aa66a09d838042ae1a2c5659828bb9b41ea3a6efa20a20fd92b121/Jinja2-2.11.2-py2.py3-none-any.whl Saved ./Jinja2-2.11.2-py2.py3-none-any.whl Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5)) Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6)) Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7)) Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz Complete output from command python setup.py egg_info: Traceback (most recent call last): File "<string>", line 1, in <module> File "/tmpfs/tmp/tmp6m3w5les/setuptools-tmp/setuptools/__init__.py", line 6, in <module> import distutils.core File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 82, in create_module return importlib.import_module('._distutils', 'setuptools') File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) ModuleNotFoundError: No module named 'setuptools._distutils' ---------------------------------------- (Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-rc4r6uhv/pypandoc/ ) INFO: Elapsed time: 2.315s INFO: 0 processes. FAILED: Build did NOT complete successfully (0 packages loaded) FAILED: Build did NOT complete successfully (0 packages loaded) Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/java-redis/synth.py", line 28, in <module> bazel_target=f'//google/cloud/{service}/{version}:google-cloud-{service}-{version}-java', File "/tmpfs/src/github/synthtool/synthtool/languages/java.py", line 298, in bazel_library library = gapic.java_library(service=service, version=version, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 62, in java_library service, version, "java", tar_strip_components=0, **kwargs File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 183, in _generate_code shell.run(bazel_run_args) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bazel', '--max_idle_secs=240', 'build', '//google/cloud/redis/v1:google-cloud-redis-v1-java']' returned non-zero exit status 1. 2020-08-30 14:23:48,202 autosynth [ERROR] > Synthesis failed 2020-08-30 14:23:48,203 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at f3c275b chore: regenerate README (#177) 2020-08-30 14:23:48,208 autosynth [DEBUG] > Running: git checkout autosynth Switched to branch 'autosynth' 2020-08-30 14:23:48,213 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 690, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 539, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 670, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 375, in synthesize_loop has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 273, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/29b10bde-06f7-44fc-9da8-8806922323b8/targets/github%2Fsynthtool;config=default/tests;query=java-redis;failed=false).
1.0
Synthesis failed for java-redis - Hello! Autosynth couldn't regenerate java-redis. :broken_heart: Here's the output from running `synth.py`: ``` hl Saved ./Jinja2-2.11.2-py2.py3-none-any.whl Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5)) Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6)) Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7)) Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz Complete output from command python setup.py egg_info: Traceback (most recent call last): File "<string>", line 1, in <module> File "/tmpfs/tmp/tmp6m3w5les/setuptools-tmp/setuptools/__init__.py", line 6, in <module> import distutils.core File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 82, in create_module return importlib.import_module('._distutils', 'setuptools') File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) ModuleNotFoundError: No module named 'setuptools._distutils' ---------------------------------------- (Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-rc4r6uhv/pypandoc/ ) ERROR: no such package '@gapic_generator_python_pip_deps//': pip_import failed: Collecting click==7.1.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 1)) Using cached https://files.pythonhosted.org/packages/d2/3d/fa76db83bf75c4f8d338c2fd15c8d33fdd7ad23a9b5e57eb6c5de26b430e/click-7.1.2-py2.py3-none-any.whl Saved ./click-7.1.2-py2.py3-none-any.whl Collecting google-api-core==1.22.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 2)) Using cached https://files.pythonhosted.org/packages/e0/2d/7c6c75013105e1d2b6eaa1bf18a56995be1dbc673c38885aea31136e9918/google_api_core-1.22.1-py2.py3-none-any.whl Saved ./google_api_core-1.22.1-py2.py3-none-any.whl Collecting googleapis-common-protos==1.52.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 3)) Using cached https://files.pythonhosted.org/packages/03/74/3956721ea1eb4bcf7502a311fdaa60b85bd751de4e57d1943afe9b334141/googleapis_common_protos-1.52.0-py2.py3-none-any.whl Saved ./googleapis_common_protos-1.52.0-py2.py3-none-any.whl Collecting jinja2==2.11.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 4)) Using cached https://files.pythonhosted.org/packages/30/9e/f663a2aa66a09d838042ae1a2c5659828bb9b41ea3a6efa20a20fd92b121/Jinja2-2.11.2-py2.py3-none-any.whl Saved ./Jinja2-2.11.2-py2.py3-none-any.whl Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5)) Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6)) Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7)) Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz Complete output from command python setup.py egg_info: Traceback (most recent call last): File "<string>", line 1, in <module> File "/tmpfs/tmp/tmp6m3w5les/setuptools-tmp/setuptools/__init__.py", line 6, in <module> import distutils.core File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 82, in create_module return importlib.import_module('._distutils', 'setuptools') File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) ModuleNotFoundError: No module named 'setuptools._distutils' ---------------------------------------- (Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-rc4r6uhv/pypandoc/ ) INFO: Elapsed time: 2.315s INFO: 0 processes. FAILED: Build did NOT complete successfully (0 packages loaded) FAILED: Build did NOT complete successfully (0 packages loaded) Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/java-redis/synth.py", line 28, in <module> bazel_target=f'//google/cloud/{service}/{version}:google-cloud-{service}-{version}-java', File "/tmpfs/src/github/synthtool/synthtool/languages/java.py", line 298, in bazel_library library = gapic.java_library(service=service, version=version, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 62, in java_library service, version, "java", tar_strip_components=0, **kwargs File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 183, in _generate_code shell.run(bazel_run_args) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bazel', '--max_idle_secs=240', 'build', '//google/cloud/redis/v1:google-cloud-redis-v1-java']' returned non-zero exit status 1. 2020-08-30 14:23:48,202 autosynth [ERROR] > Synthesis failed 2020-08-30 14:23:48,203 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at f3c275b chore: regenerate README (#177) 2020-08-30 14:23:48,208 autosynth [DEBUG] > Running: git checkout autosynth Switched to branch 'autosynth' 2020-08-30 14:23:48,213 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 690, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 539, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 670, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 375, in synthesize_loop has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 273, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/29b10bde-06f7-44fc-9da8-8806922323b8/targets/github%2Fsynthtool;config=default/tests;query=java-redis;failed=false).
non_defect
synthesis failed for java redis hello autosynth couldn t regenerate java redis broken heart here s the output from running synth py hl saved none any whl collecting markupsafe from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved markupsafe whl collecting protobuf from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved protobuf whl collecting pypandoc from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached complete output from command python setup py egg info traceback most recent call last file line in file tmpfs tmp setuptools tmp setuptools init py line in import distutils core file tmpfs src github synthtool env lib site packages distutils hack init py line in create module return importlib import module distutils setuptools file home kbuilder pyenv versions lib importlib init py line in import module return bootstrap gcd import name package level modulenotfounderror no module named setuptools distutils command python setup py egg info failed with error code in tmpfs tmp pip build pypandoc error no such package gapic generator python pip deps pip import failed collecting click from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved click none any whl collecting google api core from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved google api core none any whl collecting googleapis common protos from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved googleapis common protos none any whl collecting from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved none any whl collecting markupsafe from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved markupsafe whl collecting protobuf from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved protobuf whl collecting pypandoc from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached complete output from command python setup py egg info traceback most recent call last file line in file tmpfs tmp setuptools tmp setuptools init py line in import distutils core file tmpfs src github synthtool env lib site packages distutils hack init py line in create module return importlib import module distutils setuptools file home kbuilder pyenv versions lib importlib init py line in import module return bootstrap gcd import name package level modulenotfounderror no module named setuptools distutils command python setup py egg info failed with error code in tmpfs tmp pip build pypandoc info elapsed time info processes failed build did not complete successfully packages loaded failed build did not complete successfully packages loaded traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool java redis synth py line in bazel target f google cloud service version google cloud service version java file tmpfs src github synthtool synthtool languages java py line in bazel library library gapic java library service service version version kwargs file tmpfs src github synthtool synthtool gcp gapic bazel py line in java library service version java tar strip components kwargs file tmpfs src github synthtool synthtool gcp gapic bazel py line in generate code shell run bazel run args file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git reset hard head head is now at chore regenerate readme autosynth running git checkout autosynth switched to branch autosynth autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop has changes toolbox synthesize version in new branch synthesizer youngest file tmpfs src github synthtool autosynth synth py line in synthesize version in new branch synthesizer synthesize synth log path self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0