Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
29,521 | 5,640,841,258 | IssuesEvent | 2017-04-06 17:20:16 | DanwareCreations/DotKEGG | https://api.github.com/repos/DanwareCreations/DotKEGG | closed | Add a "Release Process" Document | documentation enhancement | Add a simple MarkDown file to the root of the repo, explaining the steps to take on a release branch. These should include the steps that were taken during the v0.2.0-alpha release, as well as updating documentation in the S3 Bucket, and updating the latest version numbers on the main README.
Then, it will be easy to see what should be included in a future publish script! | 1.0 | Add a "Release Process" Document - Add a simple MarkDown file to the root of the repo, explaining the steps to take on a release branch. These should include the steps that were taken during the v0.2.0-alpha release, as well as updating documentation in the S3 Bucket, and updating the latest version numbers on the main README.
Then, it will be easy to see what should be included in a future publish script! | non_defect | add a release process document add a simple markdown file to the root of the repo explaining the steps to take on a release branch these should include the steps that were taken during the alpha release as well as updating documentation in the bucket and updating the latest version numbers on the main readme then it will be easy to see what should be included in a future publish script | 0 |
53,903 | 13,262,479,614 | IssuesEvent | 2020-08-20 21:53:04 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | [icetray] svn revision info required (Trac #2284) | Migrated from Trac combo core defect | If cmake can't detect the svn revision, the build fails:
```text
In file included from <command-line>:0:0:
/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx: In member function 'I3TrayInfo I3TrayInfoService::GetConfig()':
/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:22: error: 'SVN_REVISION' was not declared in this scope
#define SVN_REVISION SVN_REVISION-NOTFOUND
^
/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION'
the_config.svn_revision = SVN_REVISION;
^
/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:35: error: 'NOTFOUND' was not declared in this scope
#define SVN_REVISION SVN_REVISION-NOTFOUND
^
/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION'
the_config.svn_revision = SVN_REVISION;
^
```
This is blocking the cvmfs build for combo.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2284">https://code.icecube.wisc.edu/projects/icecube/ticket/2284</a>, reported by david.schultzand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-05-16T22:09:18",
"_ts": "1558044558386076",
"description": "If cmake can't detect the svn revision, the build fails:\n\n{{{\nIn file included from <command-line>:0:0:\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx: In member function 'I3TrayInfo I3TrayInfoService::GetConfig()':\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:22: error: 'SVN_REVISION' was not declared in this scope\n#define SVN_REVISION SVN_REVISION-NOTFOUND\n ^\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION'\n the_config.svn_revision = SVN_REVISION;\n ^\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:35: error: 'NOTFOUND' was not declared in this scope\n#define SVN_REVISION SVN_REVISION-NOTFOUND\n ^\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION'\n the_config.svn_revision = SVN_REVISION;\n ^\n}}}\n\nThis is blocking the cvmfs build for combo.",
"reporter": "david.schultz",
"cc": "olivas",
"resolution": "fixed",
"time": "2019-05-16T20:06:50",
"component": "combo core",
"summary": "[icetray] svn revision info required",
"priority": "critical",
"keywords": "",
"milestone": "Vernal Equinox 2019",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [icetray] svn revision info required (Trac #2284) - If cmake can't detect the svn revision, the build fails:
```text
In file included from <command-line>:0:0:
/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx: In member function 'I3TrayInfo I3TrayInfoService::GetConfig()':
/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:22: error: 'SVN_REVISION' was not declared in this scope
#define SVN_REVISION SVN_REVISION-NOTFOUND
^
/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION'
the_config.svn_revision = SVN_REVISION;
^
/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:35: error: 'NOTFOUND' was not declared in this scope
#define SVN_REVISION SVN_REVISION-NOTFOUND
^
/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION'
the_config.svn_revision = SVN_REVISION;
^
```
This is blocking the cvmfs build for combo.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2284">https://code.icecube.wisc.edu/projects/icecube/ticket/2284</a>, reported by david.schultzand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-05-16T22:09:18",
"_ts": "1558044558386076",
"description": "If cmake can't detect the svn revision, the build fails:\n\n{{{\nIn file included from <command-line>:0:0:\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx: In member function 'I3TrayInfo I3TrayInfoService::GetConfig()':\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:22: error: 'SVN_REVISION' was not declared in this scope\n#define SVN_REVISION SVN_REVISION-NOTFOUND\n ^\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION'\n the_config.svn_revision = SVN_REVISION;\n ^\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC1/icetray/CMakeFiles/workspace_config.h:29:35: error: 'NOTFOUND' was not declared in this scope\n#define SVN_REVISION SVN_REVISION-NOTFOUND\n ^\n/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/metaprojects/combo/V00-00-00-RC1/icetray/private/icetray/I3TrayInfoService.cxx:63:29: note: in expansion of macro 'SVN_REVISION'\n the_config.svn_revision = SVN_REVISION;\n ^\n}}}\n\nThis is blocking the cvmfs build for combo.",
"reporter": "david.schultz",
"cc": "olivas",
"resolution": "fixed",
"time": "2019-05-16T20:06:50",
"component": "combo core",
"summary": "[icetray] svn revision info required",
"priority": "critical",
"keywords": "",
"milestone": "Vernal Equinox 2019",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | svn revision info required trac if cmake can t detect the svn revision the build fails text in file included from cvmfs icecube opensciencegrid org metaprojects combo icetray private icetray cxx in member function getconfig cvmfs icecube opensciencegrid org rhel metaprojects combo icetray cmakefiles workspace config h error svn revision was not declared in this scope define svn revision svn revision notfound cvmfs icecube opensciencegrid org metaprojects combo icetray private icetray cxx note in expansion of macro svn revision the config svn revision svn revision cvmfs icecube opensciencegrid org rhel metaprojects combo icetray cmakefiles workspace config h error notfound was not declared in this scope define svn revision svn revision notfound cvmfs icecube opensciencegrid org metaprojects combo icetray private icetray cxx note in expansion of macro svn revision the config svn revision svn revision this is blocking the cvmfs build for combo migrated from json status closed changetime ts description if cmake can t detect the svn revision the build fails n n nin file included from n cvmfs icecube opensciencegrid org metaprojects combo icetray private icetray cxx in member function getconfig n cvmfs icecube opensciencegrid org rhel metaprojects combo icetray cmakefiles workspace config h error svn revision was not declared in this scope n define svn revision svn revision notfound n n cvmfs icecube opensciencegrid org metaprojects combo icetray private icetray cxx note in expansion of macro svn revision n the config svn revision svn revision n n cvmfs icecube opensciencegrid org rhel metaprojects combo icetray cmakefiles workspace config h error notfound was not declared in this scope n define svn revision svn revision notfound n n cvmfs icecube opensciencegrid org metaprojects combo icetray private icetray cxx note in expansion of macro svn revision n the config svn revision svn revision n n n nthis is blocking the cvmfs build for combo reporter david schultz cc olivas resolution fixed time component combo core summary svn revision info required priority critical keywords milestone vernal equinox owner nega type defect | 1 |
269,658 | 23,456,864,875 | IssuesEvent | 2022-08-16 09:37:49 | oracle/helidon | https://api.github.com/repos/oracle/helidon | opened | JPA integration tests should run on every build | jpa/jta testing | This should be fixed for:
- helidon-2.x
- helidon-3.x
- main
The `tests/integration/jpa` module is only executed when profile `jpa` is enabled. They do not run in pipeline or during local builds.
Currently the tests seem to fail.
The test should be changed to be runnable in Java without any changes (e.g. if I do `mvn install -Ptests`, the test should run against an in-memory database.
When run with additional profiles, they can be used to validate other databases and/or native image (this should be pipeline specific and done in the appropriate scripts).
We should not have code that is not compiled and executed in the repository. | 1.0 | JPA integration tests should run on every build - This should be fixed for:
- helidon-2.x
- helidon-3.x
- main
The `tests/integration/jpa` module is only executed when profile `jpa` is enabled. They do not run in pipeline or during local builds.
Currently the tests seem to fail.
The test should be changed to be runnable in Java without any changes (e.g. if I do `mvn install -Ptests`, the test should run against an in-memory database.
When run with additional profiles, they can be used to validate other databases and/or native image (this should be pipeline specific and done in the appropriate scripts).
We should not have code that is not compiled and executed in the repository. | non_defect | jpa integration tests should run on every build this should be fixed for helidon x helidon x main the tests integration jpa module is only executed when profile jpa is enabled they do not run in pipeline or during local builds currently the tests seem to fail the test should be changed to be runnable in java without any changes e g if i do mvn install ptests the test should run against an in memory database when run with additional profiles they can be used to validate other databases and or native image this should be pipeline specific and done in the appropriate scripts we should not have code that is not compiled and executed in the repository | 0 |
66,781 | 20,672,094,028 | IssuesEvent | 2022-03-10 04:07:56 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | /rainbow breaks compound emojis | T-Defect S-Minor A-Emoji A-Slash-Commands O-Uncommon | ### Steps to reproduce
Send `/rainbow 🐻❄️`
### Outcome
#### What did you expect?

#### What happened instead?

### Operating system
NixOS unstable
### Browser information
Firefox 97.0.1
### URL for webapp
develop.element.io
### Application version
Element version: 2d3757c37f20-react-7b71a9fb6543-js-35a375e3d290 Olm version: 3.2.8
### Homeserver
Synapse 1.54.0
### Will you send logs?
No | 1.0 | /rainbow breaks compound emojis - ### Steps to reproduce
Send `/rainbow 🐻❄️`
### Outcome
#### What did you expect?

#### What happened instead?

### Operating system
NixOS unstable
### Browser information
Firefox 97.0.1
### URL for webapp
develop.element.io
### Application version
Element version: 2d3757c37f20-react-7b71a9fb6543-js-35a375e3d290 Olm version: 3.2.8
### Homeserver
Synapse 1.54.0
### Will you send logs?
No | defect | rainbow breaks compound emojis steps to reproduce send rainbow 🐻❄️ outcome what did you expect what happened instead operating system nixos unstable browser information firefox url for webapp develop element io application version element version react js olm version homeserver synapse will you send logs no | 1 |
60,305 | 17,023,392,216 | IssuesEvent | 2021-07-03 01:47:27 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | The apidb tasks are not documented. | Component: osmosis Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 12.47am, Sunday, 26th April 2009]**
To support the new PostgreSQL database schema, a new set of tasks with an apidb prefix have been created which obsolete the existing mysql tasks. These new tasks have not been documented on the osmosis wiki page(s) yet. | 1.0 | The apidb tasks are not documented. - **[Submitted to the original trac issue database at 12.47am, Sunday, 26th April 2009]**
To support the new PostgreSQL database schema, a new set of tasks with an apidb prefix have been created which obsolete the existing mysql tasks. These new tasks have not been documented on the osmosis wiki page(s) yet. | defect | the apidb tasks are not documented to support the new postgresql database schema a new set of tasks with an apidb prefix have been created which obsolete the existing mysql tasks these new tasks have not been documented on the osmosis wiki page s yet | 1 |
625,621 | 19,758,790,491 | IssuesEvent | 2022-01-16 03:19:28 | input-output-hk/cardano-db-sync | https://api.github.com/repos/input-output-hk/cardano-db-sync | closed | Empty TX inputs/outputs | bug priority high | **OS**
Debian
**Versions**
The `db-sync` version (eg `cardano-db-sync --version`): cardano-db-sync-extended 12.0.0 - linux-x86_64 - ghc-8.10
git revision 9d0180571482ee4c6acb6fbc6bf55b5a4e2ee833
PostgreSQL version: 14.1
**Build/Install Method**
The method you use to build or install `cardano-db-sync`: Hydra binaries
**Run method**
The method you used to run `cardano-db-sync` (eg Nix/Docker/systemd/none): systemd
**Additional context**
The database restored from snapshot db-sync-snapshot-schema-12-block-6647999-x86_64.tgz
**Problem Report**
There are no rows for tx f9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234 in the tx_out table in DB Sync 12
```
SELECT id FROM tx WHERE hash = '\xf9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234';
id
----------
25318047
(1 row)
```
```
SELECT id
FROM tx_out
WHERE tx_id = (SELECT id FROM tx WHERE hash = '\xf9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234');
id
----
(0 rows)
```
DB Sync 11 (cardano-db-sync-extended 11.0.4 - linux-x86_64 - ghc-8.10
git revision b5cd53d2e86f45127d98de429e8a2b6879fd8a56) feels good
```
SELECT id
FROM tx_out
WHERE tx_id = (SELECT id FROM tx WHERE hash = '\xf9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234');
id
----------
67617947
(1 row)
``` | 1.0 | Empty TX inputs/outputs - **OS**
Debian
**Versions**
The `db-sync` version (eg `cardano-db-sync --version`): cardano-db-sync-extended 12.0.0 - linux-x86_64 - ghc-8.10
git revision 9d0180571482ee4c6acb6fbc6bf55b5a4e2ee833
PostgreSQL version: 14.1
**Build/Install Method**
The method you use to build or install `cardano-db-sync`: Hydra binaries
**Run method**
The method you used to run `cardano-db-sync` (eg Nix/Docker/systemd/none): systemd
**Additional context**
The database restored from snapshot db-sync-snapshot-schema-12-block-6647999-x86_64.tgz
**Problem Report**
There are no rows for tx f9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234 in the tx_out table in DB Sync 12
```
SELECT id FROM tx WHERE hash = '\xf9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234';
id
----------
25318047
(1 row)
```
```
SELECT id
FROM tx_out
WHERE tx_id = (SELECT id FROM tx WHERE hash = '\xf9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234');
id
----
(0 rows)
```
DB Sync 11 (cardano-db-sync-extended 11.0.4 - linux-x86_64 - ghc-8.10
git revision b5cd53d2e86f45127d98de429e8a2b6879fd8a56) feels good
```
SELECT id
FROM tx_out
WHERE tx_id = (SELECT id FROM tx WHERE hash = '\xf9ed2fef27cdcf60c863ba03f27d0e38f39c5047cf73ffdf2428b48edbe83234');
id
----------
67617947
(1 row)
``` | non_defect | empty tx inputs outputs os debian versions the db sync version eg cardano db sync version cardano db sync extended linux ghc git revision postgresql version build install method the method you use to build or install cardano db sync hydra binaries run method the method you used to run cardano db sync eg nix docker systemd none systemd additional context the database restored from snapshot db sync snapshot schema block tgz problem report there are no rows for tx in the tx out table in db sync select id from tx where hash id row select id from tx out where tx id select id from tx where hash id rows db sync cardano db sync extended linux ghc git revision feels good select id from tx out where tx id select id from tx where hash id row | 0 |
34,598 | 12,294,008,724 | IssuesEvent | 2020-05-10 21:40:24 | universalturtles/Lisibilite | https://api.github.com/repos/universalturtles/Lisibilite | opened | [STORY]: Authorize Rest calls using API | Security | **A single line user story**
Ability to authorize rest calls using API
**Details about the user story**
Ability to authorize using the below header
Authorization: api-key base64(api-id:api-key)
**Definition of Done**
If token is valid used should be able to user the API
If token in invalid, user should not be able to use the API
If token is revoked, user should not be able to use the API
Log the user usage into a DB with timestamp (For future rate limiting)
**Additional Comments/Suggestions/Considerations if any**
| True | [STORY]: Authorize Rest calls using API - **A single line user story**
Ability to authorize rest calls using API
**Details about the user story**
Ability to authorize using the below header
Authorization: api-key base64(api-id:api-key)
**Definition of Done**
If token is valid used should be able to user the API
If token in invalid, user should not be able to use the API
If token is revoked, user should not be able to use the API
Log the user usage into a DB with timestamp (For future rate limiting)
**Additional Comments/Suggestions/Considerations if any**
| non_defect | authorize rest calls using api a single line user story ability to authorize rest calls using api details about the user story ability to authorize using the below header authorization api key api id api key definition of done if token is valid used should be able to user the api if token in invalid user should not be able to use the api if token is revoked user should not be able to use the api log the user usage into a db with timestamp for future rate limiting additional comments suggestions considerations if any | 0 |
9,049 | 2,615,125,354 | IssuesEvent | 2015-03-01 05:53:26 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | closed | NullPointerException in GoogleJsonResponseExecption | auto-migrated Milestone-Version1.8.0 Priority-Critical Type-Defect | ```
Version of google-api-java-client (e.g. 1.5.0-beta)?
1.7.0-beta
Java environment (e.g. Java 6, Android 2.3, App Engine)?
All
Describe the problem.
It appears that Google APIs sometimes return an HTTP error response with
Content-Type "application/json; charset=UTF-8" but with no actual content. It
results in this stack trace:
Thread [main] (Suspended (exception NullPointerException))
InputStreamReader(Reader).<init>(Object) line: 78
InputStreamReader.<init>(InputStream, Charset) line: 113
GsonFactory.createJsonParser(InputStream) line: 48
JsonHttpParser.parserForResponse(JsonFactory, HttpResponse) line: 117
GoogleJsonResponseException.from(JsonFactory, HttpResponse) line: 121
GoogleJsonResponseException.execute(JsonFactory, HttpRequest) line: 187
How would you expect it to be fixed?
We should be robust to this kind of problem.
That said, this is incorrect behavior from the Google servers. I'll bring that
up in a separate bug list.
```
Original issue reported on code.google.com by `yan...@google.com` on 27 Mar 2012 at 3:47 | 1.0 | NullPointerException in GoogleJsonResponseExecption - ```
Version of google-api-java-client (e.g. 1.5.0-beta)?
1.7.0-beta
Java environment (e.g. Java 6, Android 2.3, App Engine)?
All
Describe the problem.
It appears that Google APIs sometimes return an HTTP error response with
Content-Type "application/json; charset=UTF-8" but with no actual content. It
results in this stack trace:
Thread [main] (Suspended (exception NullPointerException))
InputStreamReader(Reader).<init>(Object) line: 78
InputStreamReader.<init>(InputStream, Charset) line: 113
GsonFactory.createJsonParser(InputStream) line: 48
JsonHttpParser.parserForResponse(JsonFactory, HttpResponse) line: 117
GoogleJsonResponseException.from(JsonFactory, HttpResponse) line: 121
GoogleJsonResponseException.execute(JsonFactory, HttpRequest) line: 187
How would you expect it to be fixed?
We should be robust to this kind of problem.
That said, this is incorrect behavior from the Google servers. I'll bring that
up in a separate bug list.
```
Original issue reported on code.google.com by `yan...@google.com` on 27 Mar 2012 at 3:47 | defect | nullpointerexception in googlejsonresponseexecption version of google api java client e g beta beta java environment e g java android app engine all describe the problem it appears that google apis sometimes return an http error response with content type application json charset utf but with no actual content it results in this stack trace thread suspended exception nullpointerexception inputstreamreader reader object line inputstreamreader inputstream charset line gsonfactory createjsonparser inputstream line jsonhttpparser parserforresponse jsonfactory httpresponse line googlejsonresponseexception from jsonfactory httpresponse line googlejsonresponseexception execute jsonfactory httprequest line how would you expect it to be fixed we should be robust to this kind of problem that said this is incorrect behavior from the google servers i ll bring that up in a separate bug list original issue reported on code google com by yan google com on mar at | 1 |
39,045 | 15,861,534,008 | IssuesEvent | 2021-04-08 10:26:25 | terraform-providers/terraform-provider-azurerm | https://api.github.com/repos/terraform-providers/terraform-provider-azurerm | closed | Support for Service Bus Topic and Queue Data Sources | enhancement service/servicebus | <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
I would like a way to refer to a Service Bus Topic or Queue as a Data Source
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* azurerm_servicebus_topic
* azurerm_servicebus_queue
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
data "azurerm_servicebus_topic" "example" {
name = "tfex_servicebus_topic"
namespace_name = "tfex-servicebus-namespace"
resource_group_name = "terraform-servicebus"
}
data "azurerm_servicebus_queue" "example" {
name = "tfex_servicebus_queue"
namespace_name = "tfex-servicebus-namespace"
resource_group_name = "terraform-servicebus"
}
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/
--->
| 2.0 | Support for Service Bus Topic and Queue Data Sources - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
I would like a way to refer to a Service Bus Topic or Queue as a Data Source
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* azurerm_servicebus_topic
* azurerm_servicebus_queue
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
data "azurerm_servicebus_topic" "example" {
name = "tfex_servicebus_topic"
namespace_name = "tfex-servicebus-namespace"
resource_group_name = "terraform-servicebus"
}
data "azurerm_servicebus_queue" "example" {
name = "tfex_servicebus_queue"
namespace_name = "tfex-servicebus-namespace"
resource_group_name = "terraform-servicebus"
}
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/
--->
| non_defect | support for service bus topic and queue data sources community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description i would like a way to refer to a service bus topic or queue as a data source new or affected resource s azurerm servicebus topic azurerm servicebus queue potential terraform configuration hcl data azurerm servicebus topic example name tfex servicebus topic namespace name tfex servicebus namespace resource group name terraform servicebus data azurerm servicebus queue example name tfex servicebus queue namespace name tfex servicebus namespace resource group name terraform servicebus references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example | 0 |
66,037 | 19,906,569,005 | IssuesEvent | 2022-01-25 13:24:45 | data4development/iati-workbench | https://api.github.com/repos/data4development/iati-workbench | opened | port narratives merge logic from iati-activities to iati-organisations | Defect Spreadsheets2IATI | The current merge logic for organisation files still produces for instance <description> without <narratives>.
Port the fixes in iati-activities to iati-organisations for merging. | 1.0 | port narratives merge logic from iati-activities to iati-organisations - The current merge logic for organisation files still produces for instance <description> without <narratives>.
Port the fixes in iati-activities to iati-organisations for merging. | defect | port narratives merge logic from iati activities to iati organisations the current merge logic for organisation files still produces for instance without port the fixes in iati activities to iati organisations for merging | 1 |
7,192 | 2,610,357,154 | IssuesEvent | 2015-02-26 19:55:40 | chrsmith/scribefire-chrome | https://api.github.com/repos/chrsmith/scribefire-chrome | closed | UNUSABLE | auto-migrated Priority-Medium Type-Defect | ```
What's the problem?
The side bar is on the wrong side
What browser are you using?
Firefox 7.01
What version of ScribeFire are you running?
I dunno. It crashes FF when I try to open it, but when I manage to get it
started, the sidebar will be on the WRONG SIDE.
PUT IT THE F*CK BACK.
```
-----
Original issue reported on code.google.com by `twstdgen...@gmail.com` on 5 Nov 2011 at 5:59 | 1.0 | UNUSABLE - ```
What's the problem?
The side bar is on the wrong side
What browser are you using?
Firefox 7.01
What version of ScribeFire are you running?
I dunno. It crashes FF when I try to open it, but when I manage to get it
started, the sidebar will be on the WRONG SIDE.
PUT IT THE F*CK BACK.
```
-----
Original issue reported on code.google.com by `twstdgen...@gmail.com` on 5 Nov 2011 at 5:59 | defect | unusable what s the problem the side bar is on the wrong side what browser are you using firefox what version of scribefire are you running i dunno it crashes ff when i try to open it but when i manage to get it started the sidebar will be on the wrong side put it the f ck back original issue reported on code google com by twstdgen gmail com on nov at | 1 |
84,293 | 10,368,983,206 | IssuesEvent | 2019-09-07 21:45:19 | SuperCowPowers/bat | https://api.github.com/repos/SuperCowPowers/bat | closed | Switch docs to markdown instead of RST | documentation usability | Going to just use Github's documentation system instead of readthedocs... | 1.0 | Switch docs to markdown instead of RST - Going to just use Github's documentation system instead of readthedocs... | non_defect | switch docs to markdown instead of rst going to just use github s documentation system instead of readthedocs | 0 |
18,035 | 3,021,549,513 | IssuesEvent | 2015-07-31 15:21:14 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | fetchMany() doesn't work with CUBRID | C: DB: CUBRID P: Medium R: Fixed T: Defect | There is an exception when calling `fetchMany()` on CUBRID:
```
java.lang.UnsupportedOperationException
at cubrid.jdbc.driver.CUBRIDStatement.getMoreResults(CUBRIDStatement.java:649)
at org.jooq.tools.jdbc.DefaultStatement.getMoreResults(DefaultStatement.java:257)
at org.jooq.tools.jdbc.DefaultStatement.getMoreResults(DefaultStatement.java:257)
at org.jooq.impl.Utils.consumeResultSets(Utils.java:2548)
at org.jooq.impl.AbstractResultQuery.execute(AbstractResultQuery.java:282)
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:342)
at org.jooq.impl.AbstractResultQuery.fetchMany(AbstractResultQuery.java:353)
at org.jooq.impl.DefaultDSLContext.fetchMany(DefaultDSLContext.java:540)
at org.jooq.test.all.testcases.FetchTests.testFetchMany(FetchTests.java:544)
at org.jooq.test.jOOQAbstractTest.testFetchMany(jOOQAbstractTest.java:2106)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86)
at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192)
```
The problem is in jOOQ's calling `Statement.getMoreResults(int)`, which isn't implemented by CUBRID's JDBC driver:
```java
// Call this only when there was at least one ResultSet.
// Otherwise, this call is not supported by ojdbc...
if (anyResults)
ctx.statement().getMoreResults(Statement.CLOSE_ALL_RESULTS);
``` | 1.0 | fetchMany() doesn't work with CUBRID - There is an exception when calling `fetchMany()` on CUBRID:
```
java.lang.UnsupportedOperationException
at cubrid.jdbc.driver.CUBRIDStatement.getMoreResults(CUBRIDStatement.java:649)
at org.jooq.tools.jdbc.DefaultStatement.getMoreResults(DefaultStatement.java:257)
at org.jooq.tools.jdbc.DefaultStatement.getMoreResults(DefaultStatement.java:257)
at org.jooq.impl.Utils.consumeResultSets(Utils.java:2548)
at org.jooq.impl.AbstractResultQuery.execute(AbstractResultQuery.java:282)
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:342)
at org.jooq.impl.AbstractResultQuery.fetchMany(AbstractResultQuery.java:353)
at org.jooq.impl.DefaultDSLContext.fetchMany(DefaultDSLContext.java:540)
at org.jooq.test.all.testcases.FetchTests.testFetchMany(FetchTests.java:544)
at org.jooq.test.jOOQAbstractTest.testFetchMany(jOOQAbstractTest.java:2106)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86)
at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192)
```
The problem is in jOOQ's calling `Statement.getMoreResults(int)`, which isn't implemented by CUBRID's JDBC driver:
```java
// Call this only when there was at least one ResultSet.
// Otherwise, this call is not supported by ojdbc...
if (anyResults)
ctx.statement().getMoreResults(Statement.CLOSE_ALL_RESULTS);
``` | defect | fetchmany doesn t work with cubrid there is an exception when calling fetchmany on cubrid java lang unsupportedoperationexception at cubrid jdbc driver cubridstatement getmoreresults cubridstatement java at org jooq tools jdbc defaultstatement getmoreresults defaultstatement java at org jooq tools jdbc defaultstatement getmoreresults defaultstatement java at org jooq impl utils consumeresultsets utils java at org jooq impl abstractresultquery execute abstractresultquery java at org jooq impl abstractquery execute abstractquery java at org jooq impl abstractresultquery fetchmany abstractresultquery java at org jooq impl defaultdslcontext fetchmany defaultdslcontext java at org jooq test all testcases fetchtests testfetchmany fetchtests java at org jooq test jooqabstracttest testfetchmany jooqabstracttest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner run parentrunner java at org eclipse jdt internal runner run java at org eclipse jdt internal junit runner testexecution run testexecution java at org eclipse jdt internal junit runner remotetestrunner runtests remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner runtests remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner run remotetestrunner java at org eclipse jdt internal junit runner remotetestrunner main remotetestrunner java the problem is in jooq s calling statement getmoreresults int which isn t implemented by cubrid s jdbc driver java call this only when there was at least one resultset otherwise this call is not supported by ojdbc if anyresults ctx statement getmoreresults statement close all results | 1 |
80,529 | 30,316,309,354 | IssuesEvent | 2023-07-10 15:48:08 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | make-icons.sh doesn't make the right icons | T-Defect P3 S-Minor A-Packaging | It's unclear where the icons are coming from if not the script | 1.0 | make-icons.sh doesn't make the right icons - It's unclear where the icons are coming from if not the script | defect | make icons sh doesn t make the right icons it s unclear where the icons are coming from if not the script | 1 |
54,930 | 14,046,874,913 | IssuesEvent | 2020-11-02 05:55:55 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | Unboxing may produce NPE warning in intellij when using count(*) | T: Defect | ### Expected behavior
No warning
### Actual behavior
Unboxing may produce NPE warning in intellij
### Steps to reproduce the problem
Code like in stackoverflow [question](https://stackoverflow.com/a/19676187/411965):
```
int count =
DSL.using(configuration)
.select(DSL.count())
.from(Table)
.fetchOne(0, int.class);
```
When the generated code has nullable annotations intellij yields the above warning.
The relevant gradle configuration is:
```
generate {
generatedAnnotation = true
nullableAnnotation = true
nullableAnnotationType = 'javax.annotation.Nullable'
}
```
### Versions
- nu.studer.jooq: 4.2
- jOOQ: 3.14
| 1.0 | Unboxing may produce NPE warning in intellij when using count(*) - ### Expected behavior
No warning
### Actual behavior
Unboxing may produce NPE warning in intellij
### Steps to reproduce the problem
Code like in stackoverflow [question](https://stackoverflow.com/a/19676187/411965):
```
int count =
DSL.using(configuration)
.select(DSL.count())
.from(Table)
.fetchOne(0, int.class);
```
When the generated code has nullable annotations intellij yields the above warning.
The relevant gradle configuration is:
```
generate {
generatedAnnotation = true
nullableAnnotation = true
nullableAnnotationType = 'javax.annotation.Nullable'
}
```
### Versions
- nu.studer.jooq: 4.2
- jOOQ: 3.14
| defect | unboxing may produce npe warning in intellij when using count expected behavior no warning actual behavior unboxing may produce npe warning in intellij steps to reproduce the problem code like in stackoverflow int count dsl using configuration select dsl count from table fetchone int class when the generated code has nullable annotations intellij yields the above warning the relevant gradle configuration is generate generatedannotation true nullableannotation true nullableannotationtype javax annotation nullable versions nu studer jooq jooq | 1 |
38,409 | 8,799,194,185 | IssuesEvent | 2018-12-24 12:36:40 | vim/vim | https://api.github.com/repos/vim/vim | closed | terminal resize during file recovery | Priority-Medium auto-migrated defect | ```
What steps will reproduce the problem?
1. open a file with vim
2. in a new terminal, open the same file with another vim instance
3. resize your terminal
What is the expected output? What do you see instead?
I expect the text to reformat while the terminal resizes. Instead, the text
gets garbled and never formats properly again. In some cases, the output
remains affected after dismissing the prompt, during editing. In either case,
the problem can sometimes be fixed by resizing the terminal window once again
while editing.
What version of the product are you using? On what operating system?
$;vim --version
VIM - Vi IMproved 7.3 (2010 Aug 15, compiled Nov 16 2010 17:05:25)
Included patches: 1-56
Modified by <bugzilla@redhat.com>
Compiled by <bugzilla@redhat.com>
Huge version without GUI. Features included (+) or not (-):
+arabic +autocmd -balloon_eval -browse ++builtin_terms +byte_offset +cindent
-clientserver -clipboard +cmdline_compl +cmdline_hist +cmdline_info +comments
+conceal +cryptv +cscope +cursorbind +cursorshape +dialog_con +diff +digraphs
-dnd -ebcdic +emacs_tags +eval +ex_extra +extra_search +farsi +file_in_path
+find_in_path +float +folding -footer +fork() +gettext -hangul_input +iconv
+insert_expand +jumplist +keymap +langmap +libcall +linebreak +lispindent
+listcmds +localmap -lua +menu +mksession +modify_fname +mouse -mouseshape
+mouse_dec +mouse_gpm -mouse_jsbterm +mouse_netterm -mouse_sysmouse
+mouse_xterm +multi_byte +multi_lang -mzscheme +netbeans_intg -osfiletype
+path_extra +perl +persistent_undo +postscript +printer +profile +python
-python3 +quickfix +reltime +rightleft +ruby +scrollbind +signs +smartindent
-sniff +startuptime +statusline -sun_workshop +syntax +tag_binary
+tag_old_static -tag_any_white -tcl +terminfo +termresponse +textobjects +title
-toolbar +user_commands +vertsplit +virtualedit +visual +visualextra +viminfo
+vreplace +wildignore +wildmenu +windows +writebackup -X11 -xfontset -xim -xsmp
-xterm_clipboard -xterm_save
system vimrc file: "/etc/vimrc"
user vimrc file: "$HOME/.vimrc"
user exrc file: "$HOME/.exrc"
fall-back for $VIM: "/etc"
f-b for $VIMRUNTIME: "/usr/share/vim/vim73"
Compilation: gcc -c -I. -Iproto -DHAVE_CONFIG_H -O2 -g -pipe -Wall
-fexceptions -fstack-protector --param=ssp-buffer-size=4 -m64 -mtune=generic
-D_GNU_SOURCE -D_FILE_OFFSET_BITS=64 -D_FORTIFY_SOURCE=1
Linking: gcc -L. -rdynamic -Wl,-export-dynamic -Wl,--enable-new-dtags
-Wl,-rpath,/usr/lib64/perl5/CORE -L/usr/local/lib -Wl,--as-needed -o vim
-lm -lnsl -lselinux -lncurses -lacl -lattr -lgpm -ldl
-Wl,--enable-new-dtags -Wl,-rpath,/usr/lib64/perl5/CORE -fstack-protector
-L/usr/lib64/perl5/CORE -lperl -lresolv -lnsl -ldl -lm -lcrypt -lutil -lpthread
-lc -L/usr/lib64/python2.7/config -lpython2.7 -lpthread -ldl -lutil -lm
-Xlinker -export-dynamic -lruby -lpthread -lrt -ldl -lcrypt -lm
Fedora 14, XFCE window manager, and xfce4-terminal
Please provide any additional information below.
```
Original issue reported on code.google.com by `michaelj...@gmail.com` on 7 Jun 2011 at 7:35
| 1.0 | terminal resize during file recovery - ```
What steps will reproduce the problem?
1. open a file with vim
2. in a new terminal, open the same file with another vim instance
3. resize your terminal
What is the expected output? What do you see instead?
I expect the text to reformat while the terminal resizes. Instead, the text
gets garbled and never formats properly again. In some cases, the output
remains affected after dismissing the prompt, during editing. In either case,
the problem can sometimes be fixed by resizing the terminal window once again
while editing.
What version of the product are you using? On what operating system?
$;vim --version
VIM - Vi IMproved 7.3 (2010 Aug 15, compiled Nov 16 2010 17:05:25)
Included patches: 1-56
Modified by <bugzilla@redhat.com>
Compiled by <bugzilla@redhat.com>
Huge version without GUI. Features included (+) or not (-):
+arabic +autocmd -balloon_eval -browse ++builtin_terms +byte_offset +cindent
-clientserver -clipboard +cmdline_compl +cmdline_hist +cmdline_info +comments
+conceal +cryptv +cscope +cursorbind +cursorshape +dialog_con +diff +digraphs
-dnd -ebcdic +emacs_tags +eval +ex_extra +extra_search +farsi +file_in_path
+find_in_path +float +folding -footer +fork() +gettext -hangul_input +iconv
+insert_expand +jumplist +keymap +langmap +libcall +linebreak +lispindent
+listcmds +localmap -lua +menu +mksession +modify_fname +mouse -mouseshape
+mouse_dec +mouse_gpm -mouse_jsbterm +mouse_netterm -mouse_sysmouse
+mouse_xterm +multi_byte +multi_lang -mzscheme +netbeans_intg -osfiletype
+path_extra +perl +persistent_undo +postscript +printer +profile +python
-python3 +quickfix +reltime +rightleft +ruby +scrollbind +signs +smartindent
-sniff +startuptime +statusline -sun_workshop +syntax +tag_binary
+tag_old_static -tag_any_white -tcl +terminfo +termresponse +textobjects +title
-toolbar +user_commands +vertsplit +virtualedit +visual +visualextra +viminfo
+vreplace +wildignore +wildmenu +windows +writebackup -X11 -xfontset -xim -xsmp
-xterm_clipboard -xterm_save
system vimrc file: "/etc/vimrc"
user vimrc file: "$HOME/.vimrc"
user exrc file: "$HOME/.exrc"
fall-back for $VIM: "/etc"
f-b for $VIMRUNTIME: "/usr/share/vim/vim73"
Compilation: gcc -c -I. -Iproto -DHAVE_CONFIG_H -O2 -g -pipe -Wall
-fexceptions -fstack-protector --param=ssp-buffer-size=4 -m64 -mtune=generic
-D_GNU_SOURCE -D_FILE_OFFSET_BITS=64 -D_FORTIFY_SOURCE=1
Linking: gcc -L. -rdynamic -Wl,-export-dynamic -Wl,--enable-new-dtags
-Wl,-rpath,/usr/lib64/perl5/CORE -L/usr/local/lib -Wl,--as-needed -o vim
-lm -lnsl -lselinux -lncurses -lacl -lattr -lgpm -ldl
-Wl,--enable-new-dtags -Wl,-rpath,/usr/lib64/perl5/CORE -fstack-protector
-L/usr/lib64/perl5/CORE -lperl -lresolv -lnsl -ldl -lm -lcrypt -lutil -lpthread
-lc -L/usr/lib64/python2.7/config -lpython2.7 -lpthread -ldl -lutil -lm
-Xlinker -export-dynamic -lruby -lpthread -lrt -ldl -lcrypt -lm
Fedora 14, XFCE window manager, and xfce4-terminal
Please provide any additional information below.
```
Original issue reported on code.google.com by `michaelj...@gmail.com` on 7 Jun 2011 at 7:35
| defect | terminal resize during file recovery what steps will reproduce the problem open a file with vim in a new terminal open the same file with another vim instance resize your terminal what is the expected output what do you see instead i expect the text to reformat while the terminal resizes instead the text gets garbled and never formats properly again in some cases the output remains affected after dismissing the prompt during editing in either case the problem can sometimes be fixed by resizing the terminal window once again while editing what version of the product are you using on what operating system vim version vim vi improved aug compiled nov included patches modified by compiled by huge version without gui features included or not arabic autocmd balloon eval browse builtin terms byte offset cindent clientserver clipboard cmdline compl cmdline hist cmdline info comments conceal cryptv cscope cursorbind cursorshape dialog con diff digraphs dnd ebcdic emacs tags eval ex extra extra search farsi file in path find in path float folding footer fork gettext hangul input iconv insert expand jumplist keymap langmap libcall linebreak lispindent listcmds localmap lua menu mksession modify fname mouse mouseshape mouse dec mouse gpm mouse jsbterm mouse netterm mouse sysmouse mouse xterm multi byte multi lang mzscheme netbeans intg osfiletype path extra perl persistent undo postscript printer profile python quickfix reltime rightleft ruby scrollbind signs smartindent sniff startuptime statusline sun workshop syntax tag binary tag old static tag any white tcl terminfo termresponse textobjects title toolbar user commands vertsplit virtualedit visual visualextra viminfo vreplace wildignore wildmenu windows writebackup xfontset xim xsmp xterm clipboard xterm save system vimrc file etc vimrc user vimrc file home vimrc user exrc file home exrc fall back for vim etc f b for vimruntime usr share vim compilation gcc c i iproto dhave config h g pipe wall fexceptions fstack protector param ssp buffer size mtune generic d gnu source d file offset bits d fortify source linking gcc l rdynamic wl export dynamic wl enable new dtags wl rpath usr core l usr local lib wl as needed o vim lm lnsl lselinux lncurses lacl lattr lgpm ldl wl enable new dtags wl rpath usr core fstack protector l usr core lperl lresolv lnsl ldl lm lcrypt lutil lpthread lc l usr config lpthread ldl lutil lm xlinker export dynamic lruby lpthread lrt ldl lcrypt lm fedora xfce window manager and terminal please provide any additional information below original issue reported on code google com by michaelj gmail com on jun at | 1 |
62,192 | 7,551,047,845 | IssuesEvent | 2018-04-18 18:50:32 | phetsims/area-model-common | https://api.github.com/repos/phetsims/area-model-common | opened | Keypad has no way to be dismissed | design:general | @EthanWJohnson brought up the concern that there is no way to dismiss a keypad without pressing "Enter". This means that the user would be forced to overwrite an entry with a new value whenever they wish to dismiss the keypad. Here are 4 quick brainstorms to avoid this:
(1) Make the keypad modal. (Means you couldn't jump between edit boxes on the fly).
(2) Add a "Cancel" button to the keypad.
(3) Dismiss the keypad by clicking on the box/pencil button you are currently editing.
(4) Have blank entries NOT actually overwrite the current value with a blank value. Instead, submitting blank entries does nothing. | 1.0 | Keypad has no way to be dismissed - @EthanWJohnson brought up the concern that there is no way to dismiss a keypad without pressing "Enter". This means that the user would be forced to overwrite an entry with a new value whenever they wish to dismiss the keypad. Here are 4 quick brainstorms to avoid this:
(1) Make the keypad modal. (Means you couldn't jump between edit boxes on the fly).
(2) Add a "Cancel" button to the keypad.
(3) Dismiss the keypad by clicking on the box/pencil button you are currently editing.
(4) Have blank entries NOT actually overwrite the current value with a blank value. Instead, submitting blank entries does nothing. | non_defect | keypad has no way to be dismissed ethanwjohnson brought up the concern that there is no way to dismiss a keypad without pressing enter this means that the user would be forced to overwrite an entry with a new value whenever they wish to dismiss the keypad here are quick brainstorms to avoid this make the keypad modal means you couldn t jump between edit boxes on the fly add a cancel button to the keypad dismiss the keypad by clicking on the box pencil button you are currently editing have blank entries not actually overwrite the current value with a blank value instead submitting blank entries does nothing | 0 |
38,978 | 9,103,965,110 | IssuesEvent | 2019-02-20 17:00:31 | idaholab/moose | https://api.github.com/repos/idaholab/moose | opened | Error check in FieldSplit Preconditioner | C: MOOSE P: minor T: defect | ## Rationale
<!--What is the reason for this enhancement or what error are you reporting?-->
No error is reported when you use invalid variable names with the FSP system.
## Description
<!--Provide details of the enhancement or instructions for reproducing the error.-->
See the discussion here:
https://groups.google.com/forum/?utm_medium=email&utm_source=footer#!msg/moose-users/RuMWaoi14Ho/9fjoneVrBQAJ
## Impact
<!--How will the changes impact the code, developers, and users? Discuss changes to the
internal interfaces and public API.-->
Minor: Bug but not a widely used system. | 1.0 | Error check in FieldSplit Preconditioner - ## Rationale
<!--What is the reason for this enhancement or what error are you reporting?-->
No error is reported when you use invalid variable names with the FSP system.
## Description
<!--Provide details of the enhancement or instructions for reproducing the error.-->
See the discussion here:
https://groups.google.com/forum/?utm_medium=email&utm_source=footer#!msg/moose-users/RuMWaoi14Ho/9fjoneVrBQAJ
## Impact
<!--How will the changes impact the code, developers, and users? Discuss changes to the
internal interfaces and public API.-->
Minor: Bug but not a widely used system. | defect | error check in fieldsplit preconditioner rationale no error is reported when you use invalid variable names with the fsp system description see the discussion here impact how will the changes impact the code developers and users discuss changes to the internal interfaces and public api minor bug but not a widely used system | 1 |
41,116 | 5,308,759,160 | IssuesEvent | 2017-02-12 01:07:03 | RestComm/Restcomm-Connect | https://api.github.com/repos/RestComm/Restcomm-Connect | opened | Assuming Subaccount identity to help a newbie | 2. Enhancement UI UX Visual App Designer | Its common to have to help a new user who is learning how to use Restcomm. It would be practical to be able to assume a subaccount identity without the subaccount user disclosing their password. Restcomm should log for audit purpose who performed actions for any given account - the account holder themselves or a superaccount holder.
<img width="1124" alt="screen shot 2017-02-11 at 5 03 25 pm" src="https://cloud.githubusercontent.com/assets/2234901/22858789/5a8dfb9e-f07c-11e6-9dc0-33426fef40e6.png">
| 1.0 | Assuming Subaccount identity to help a newbie - Its common to have to help a new user who is learning how to use Restcomm. It would be practical to be able to assume a subaccount identity without the subaccount user disclosing their password. Restcomm should log for audit purpose who performed actions for any given account - the account holder themselves or a superaccount holder.
<img width="1124" alt="screen shot 2017-02-11 at 5 03 25 pm" src="https://cloud.githubusercontent.com/assets/2234901/22858789/5a8dfb9e-f07c-11e6-9dc0-33426fef40e6.png">
| non_defect | assuming subaccount identity to help a newbie its common to have to help a new user who is learning how to use restcomm it would be practical to be able to assume a subaccount identity without the subaccount user disclosing their password restcomm should log for audit purpose who performed actions for any given account the account holder themselves or a superaccount holder img width alt screen shot at pm src | 0 |
75,195 | 7,463,475,959 | IssuesEvent | 2018-04-01 05:56:46 | slackertracker/slackertracker | https://api.github.com/repos/slackertracker/slackertracker | closed | Test Models | Models Tests | Create some simple tests for the models to make sure they save correctly to the database.
- User
- Comment
- Reaction | 1.0 | Test Models - Create some simple tests for the models to make sure they save correctly to the database.
- User
- Comment
- Reaction | non_defect | test models create some simple tests for the models to make sure they save correctly to the database user comment reaction | 0 |
128,493 | 10,540,656,001 | IssuesEvent | 2019-10-02 08:55:45 | ckeditor/ckeditor4 | https://api.github.com/repos/ckeditor/ckeditor4 | closed | Unstable Promise AMD test | status:confirmed target:minor type:failingtest | ## Type of report
Failing test
## Provide detailed reproduction steps (if any)
Sometimes Promise AMD loading test fails randomly, for example:

It sometimes fails on Chrome sometimes on Firefox so it's not any specific browser fault. | 1.0 | Unstable Promise AMD test - ## Type of report
Failing test
## Provide detailed reproduction steps (if any)
Sometimes Promise AMD loading test fails randomly, for example:

It sometimes fails on Chrome sometimes on Firefox so it's not any specific browser fault. | non_defect | unstable promise amd test type of report failing test provide detailed reproduction steps if any sometimes promise amd loading test fails randomly for example it sometimes fails on chrome sometimes on firefox so it s not any specific browser fault | 0 |
78,824 | 27,773,922,282 | IssuesEvent | 2023-03-16 16:01:10 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Desktop client loses keys when disk is full | T-Defect S-Critical A-E2EE A-Storage O-Occasional | <!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
### Description
I ran out of disk space and I believe I rebooted in a disk full condition in order to lose swap files. After reboot, I get vector-im/element-web#2996. I'm not sure if vector-im/element-web#3660 is related or if desktop uses a real file API where you can avoid writing to a file when there is no disk space.
### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- OSX Desktop 0.9.9
| 1.0 | Desktop client loses keys when disk is full - <!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
### Description
I ran out of disk space and I believe I rebooted in a disk full condition in order to lose swap files. After reboot, I get vector-im/element-web#2996. I'm not sure if vector-im/element-web#3660 is related or if desktop uses a real file API where you can avoid writing to a file when there is no disk space.
### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- OSX Desktop 0.9.9
| defect | desktop client loses keys when disk is full this is a bug report template by following the instructions below and filling out the sections with your information you will help the us to get all the necessary data to fix your issue you can also preview your report before submitting it you may remove sections that aren t relevant to your particular case text between marks will be invisible in the report description i ran out of disk space and i believe i rebooted in a disk full condition in order to lose swap files after reboot i get vector im element web i m not sure if vector im element web is related or if desktop uses a real file api where you can avoid writing to a file when there is no disk space version information osx desktop | 1 |
18,399 | 3,054,490,588 | IssuesEvent | 2015-08-13 03:11:00 | eczarny/spectacle | https://api.github.com/repos/eczarny/spectacle | closed | Alternate keyboard settings not working | defect ★★ | I nearly always have the "Dvorak – Qwerty ⌘" keyboard setting enabled, which allows me to type in Dvorak (all the vowels on the left side of the home row, all the common consonants on the right side of the home row—google Dvorak if you haven't encountered this before) but still lets me use all the standard shortcuts I am used to. For example, the "V" key types a "J" when Dvorak is enabled, but with "Dvorak – Qwerty ⌘", typing ⌘ - "V key" will still do "Paste". (As opposed to just straight up Dvorak layout, wherein I would have to type "⌘ - Period key", since the period key types a "V" in Dvorak.)
Not so with Spectacle. For example, to enter the shortcut "⌥⌘J" to center a window, I have to type "⌥⌘C", since Spectacle doesn't recognize the "– Qwerty ⌘" aspect of the keyboard layout. (The "C" key types "J" in standard Dvorak layout.)
I know this is somewhat esoteric, but all other apps deal with this keyboard setting okay, with very few exceptions. (LibreOffice mostly works with it, with a few quirks; Acrobat Reader is the only major app I've found that flatly refuses to recognize the shortcuts correctly.) | 1.0 | Alternate keyboard settings not working - I nearly always have the "Dvorak – Qwerty ⌘" keyboard setting enabled, which allows me to type in Dvorak (all the vowels on the left side of the home row, all the common consonants on the right side of the home row—google Dvorak if you haven't encountered this before) but still lets me use all the standard shortcuts I am used to. For example, the "V" key types a "J" when Dvorak is enabled, but with "Dvorak – Qwerty ⌘", typing ⌘ - "V key" will still do "Paste". (As opposed to just straight up Dvorak layout, wherein I would have to type "⌘ - Period key", since the period key types a "V" in Dvorak.)
Not so with Spectacle. For example, to enter the shortcut "⌥⌘J" to center a window, I have to type "⌥⌘C", since Spectacle doesn't recognize the "– Qwerty ⌘" aspect of the keyboard layout. (The "C" key types "J" in standard Dvorak layout.)
I know this is somewhat esoteric, but all other apps deal with this keyboard setting okay, with very few exceptions. (LibreOffice mostly works with it, with a few quirks; Acrobat Reader is the only major app I've found that flatly refuses to recognize the shortcuts correctly.) | defect | alternate keyboard settings not working i nearly always have the dvorak – qwerty ⌘ keyboard setting enabled which allows me to type in dvorak all the vowels on the left side of the home row all the common consonants on the right side of the home row—google dvorak if you haven t encountered this before but still lets me use all the standard shortcuts i am used to for example the v key types a j when dvorak is enabled but with dvorak – qwerty ⌘ typing ⌘ v key will still do paste as opposed to just straight up dvorak layout wherein i would have to type ⌘ period key since the period key types a v in dvorak not so with spectacle for example to enter the shortcut ⌥⌘j to center a window i have to type ⌥⌘c since spectacle doesn t recognize the – qwerty ⌘ aspect of the keyboard layout the c key types j in standard dvorak layout i know this is somewhat esoteric but all other apps deal with this keyboard setting okay with very few exceptions libreoffice mostly works with it with a few quirks acrobat reader is the only major app i ve found that flatly refuses to recognize the shortcuts correctly | 1 |
81,448 | 30,853,449,970 | IssuesEvent | 2023-08-02 18:34:30 | dotCMS/core | https://api.github.com/repos/dotCMS/core | closed | We are not cleaning the cache Page correctly | Type : Defect Merged QA : Passed Internal Team : Falcon dotCMS : Experiments | ### Parent Issue
https://github.com/dotCMS/core/issues/22222
### Problem Statement
Right now we are cleaning the Page Cache when any change come over the Page, but we are cleaning this Cache using the MultiTree of the page but a Page can be render on any Variant even when it does not have any MultiTree specific for that Variant (it is render by fallback but the Cache is full with the result).
Maybe we need to clean up the Page's Cache for all the existing no archived Variants
### Steps to Reproduce
Pending
### Acceptance Criteria
The cache must be clean for all the Variant where the Page was render no matter the existing MultiTree
### dotCMS Version
Lates in master
### Proposed Objective
Core Features
### Proposed Priority
Priority 2 - Important
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
_No response_
### Assumptions & Initiation Needs
_No response_
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | 1.0 | We are not cleaning the cache Page correctly - ### Parent Issue
https://github.com/dotCMS/core/issues/22222
### Problem Statement
Right now we are cleaning the Page Cache when any change come over the Page, but we are cleaning this Cache using the MultiTree of the page but a Page can be render on any Variant even when it does not have any MultiTree specific for that Variant (it is render by fallback but the Cache is full with the result).
Maybe we need to clean up the Page's Cache for all the existing no archived Variants
### Steps to Reproduce
Pending
### Acceptance Criteria
The cache must be clean for all the Variant where the Page was render no matter the existing MultiTree
### dotCMS Version
Lates in master
### Proposed Objective
Core Features
### Proposed Priority
Priority 2 - Important
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
_No response_
### Assumptions & Initiation Needs
_No response_
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | defect | we are not cleaning the cache page correctly parent issue problem statement right now we are cleaning the page cache when any change come over the page but we are cleaning this cache using the multitree of the page but a page can be render on any variant even when it does not have any multitree specific for that variant it is render by fallback but the cache is full with the result maybe we need to clean up the page s cache for all the existing no archived variants steps to reproduce pending acceptance criteria the cache must be clean for all the variant where the page was render no matter the existing multitree dotcms version lates in master proposed objective core features proposed priority priority important external links slack conversations support tickets figma designs etc no response assumptions initiation needs no response quality assurance notes workarounds no response sub tasks estimates no response | 1 |
13,122 | 9,882,899,989 | IssuesEvent | 2019-06-24 18:02:22 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | https for image computer-vision | cognitive-services/svc computer-vision/subsvc cxp doc-enhancement triaged | Hello,
The documentation says that we can access our local endpoint with https://localhost:5000 - it appears that only http is supported. Is there a configuration setting to force https?
Thanks
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 439c70ec-4484-25de-4942-939112188d58
* Version Independent ID: 08f0bb9d-c6a0-d35b-8b3e-e46cca613487
* Content: [How to install and run containers - Computer Vision](https://docs.microsoft.com/en-us/azure/cognitive-services/computer-vision/computer-vision-how-to-install-containers#feedback)
* Content Source: [articles/cognitive-services/Computer-vision/computer-vision-how-to-install-containers.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Computer-vision/computer-vision-how-to-install-containers.md)
* Service: **cognitive-services**
* Sub-service: **computer-vision**
* GitHub Login: @IEvangelist
* Microsoft Alias: **IEvangelist** | 1.0 | https for image computer-vision - Hello,
The documentation says that we can access our local endpoint with https://localhost:5000 - it appears that only http is supported. Is there a configuration setting to force https?
Thanks
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 439c70ec-4484-25de-4942-939112188d58
* Version Independent ID: 08f0bb9d-c6a0-d35b-8b3e-e46cca613487
* Content: [How to install and run containers - Computer Vision](https://docs.microsoft.com/en-us/azure/cognitive-services/computer-vision/computer-vision-how-to-install-containers#feedback)
* Content Source: [articles/cognitive-services/Computer-vision/computer-vision-how-to-install-containers.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Computer-vision/computer-vision-how-to-install-containers.md)
* Service: **cognitive-services**
* Sub-service: **computer-vision**
* GitHub Login: @IEvangelist
* Microsoft Alias: **IEvangelist** | non_defect | https for image computer vision hello the documentation says that we can access our local endpoint with it appears that only http is supported is there a configuration setting to force https thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service cognitive services sub service computer vision github login ievangelist microsoft alias ievangelist | 0 |
22,621 | 3,670,923,412 | IssuesEvent | 2016-02-22 02:39:24 | gperftools/gperftools | https://api.github.com/repos/gperftools/gperftools | closed | tcmalloc crash at CheckCachedSizeClass | Priority-Medium Status-New Type-Defect | Originally reported on Google Code with ID 519
```
#3 0x0000000000dc00a8 in google::FlushLogFilesUnsafe (min_severity=0) at thirdparty/glog-0.3.2/src/logging.cc:1511
#4 0x0000000000dc618c in google::(anonymous namespace)::FailureSignalHandler (signal_number=11,
signal_info=0x63712270,
ucontext=0x63712140) at thirdparty/glog-0.3.2/src/signalhandler.cc:324
#5 <signal handler called>
#6 0x00000000009d1bd4 in CheckCachedSizeClass (ptr=0x7fbdf92f28a0) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:946
#7 0x00000000009d1c0e in CheckedMallocResult (result=0x7fbdf92f28a0) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:950
#8 0x00000000009d225b in (anonymous namespace)::do_malloc (size=96) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:1084
#9 0x00000000009d2ea6 in (anonymous namespace)::cpp_alloc (size=96, nothrow=false)
at thirdparty/google-perftools-1.7/src/tcmalloc.cc:1360
#10 0x00000000009d5ff7 in MallocBlock::Allocate (size=48, type=-21308287)
at thirdparty/google-perftools-1.7/src/debugallocation.cc:537
#11 0x00000000009d3942 in DebugAllocate (size=48, type=-21308287) at thirdparty/google-perftools-1.7/src/debugallocation.cc:970
#12 0x00000000009d6b26 in cpp_debug_alloc (size=48, new_type=-21308287, nothrow=false)
at thirdparty/google-perftools-1.7/src/debugallocation.cc:1203
#13 0x0000000000e3d0e7 in operator new (size=48) at thirdparty/google-perftools-1.7/src/debugallocation.cc:1250
#14 0x000000000095a001 in __gnu_cxx::new_allocator<std::_Rb_tree_node<std::pair<torca::master::Job
const* const, unsigned int> > >::allocate (this=0x7fbe1b61d020, __n=1)
at /usr/local/lib/gcc/x86_64-unknown-linux-gnu/4.5.1/../../../../include/c++/4.5.1/ext/new_allocator.h:89
#15 0x0000000000959be0 in std::_Rb_tree<torca::master::Job const*, std::pair<torca::master::Job
const* const, unsigned int>, std::_Select1st<std::pair<torca::master::Job const* const,
unsigned int> >, torca::master::JobCompare, std::allocator<std::pair<torca::master::Job
const* const, unsigned int> > >::_M_get_node (this=0x7fbe1b61d020)
at /usr/local/lib/gcc/x86_64-unknown-linux-gnu/4.5.1/../../../../include/c++/4.5.1/bits/stl_tree.h:359
#16 0x00000000009590ef in std::_Rb_tree<torca::master::Job const*, std::pair<torca::master::Job
const* const, unsigned int>, std::_Select1st<std::pair<torca::master::Job const* const,
unsigned int> >, torca::master::JobCompare, std::allocator<std::pair<torca::master::Job
const* const, unsigned int> > >::_M_create_node (this=0x7fbe1b61d020, __x=...)
What steps will reproduce the problem?
I don't know how to reproduce it, it happened just once.
What is the expected output? What do you see instead?
My program crash at
return reinterpret_cast<Leaf*>(root_->ptrs[i1]->ptrs[i2])->values[i3];
values[i3] = 0xFFFFFFFFDEADBEAF。
What version of the product are you using? On what operating system?
I use google-perftools-1.7 on SUSE.
Please provide any additional information below.
```
Reported by `tyrionqq` on 2013-04-10 06:17:18
| 1.0 | tcmalloc crash at CheckCachedSizeClass - Originally reported on Google Code with ID 519
```
#3 0x0000000000dc00a8 in google::FlushLogFilesUnsafe (min_severity=0) at thirdparty/glog-0.3.2/src/logging.cc:1511
#4 0x0000000000dc618c in google::(anonymous namespace)::FailureSignalHandler (signal_number=11,
signal_info=0x63712270,
ucontext=0x63712140) at thirdparty/glog-0.3.2/src/signalhandler.cc:324
#5 <signal handler called>
#6 0x00000000009d1bd4 in CheckCachedSizeClass (ptr=0x7fbdf92f28a0) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:946
#7 0x00000000009d1c0e in CheckedMallocResult (result=0x7fbdf92f28a0) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:950
#8 0x00000000009d225b in (anonymous namespace)::do_malloc (size=96) at thirdparty/google-perftools-1.7/src/tcmalloc.cc:1084
#9 0x00000000009d2ea6 in (anonymous namespace)::cpp_alloc (size=96, nothrow=false)
at thirdparty/google-perftools-1.7/src/tcmalloc.cc:1360
#10 0x00000000009d5ff7 in MallocBlock::Allocate (size=48, type=-21308287)
at thirdparty/google-perftools-1.7/src/debugallocation.cc:537
#11 0x00000000009d3942 in DebugAllocate (size=48, type=-21308287) at thirdparty/google-perftools-1.7/src/debugallocation.cc:970
#12 0x00000000009d6b26 in cpp_debug_alloc (size=48, new_type=-21308287, nothrow=false)
at thirdparty/google-perftools-1.7/src/debugallocation.cc:1203
#13 0x0000000000e3d0e7 in operator new (size=48) at thirdparty/google-perftools-1.7/src/debugallocation.cc:1250
#14 0x000000000095a001 in __gnu_cxx::new_allocator<std::_Rb_tree_node<std::pair<torca::master::Job
const* const, unsigned int> > >::allocate (this=0x7fbe1b61d020, __n=1)
at /usr/local/lib/gcc/x86_64-unknown-linux-gnu/4.5.1/../../../../include/c++/4.5.1/ext/new_allocator.h:89
#15 0x0000000000959be0 in std::_Rb_tree<torca::master::Job const*, std::pair<torca::master::Job
const* const, unsigned int>, std::_Select1st<std::pair<torca::master::Job const* const,
unsigned int> >, torca::master::JobCompare, std::allocator<std::pair<torca::master::Job
const* const, unsigned int> > >::_M_get_node (this=0x7fbe1b61d020)
at /usr/local/lib/gcc/x86_64-unknown-linux-gnu/4.5.1/../../../../include/c++/4.5.1/bits/stl_tree.h:359
#16 0x00000000009590ef in std::_Rb_tree<torca::master::Job const*, std::pair<torca::master::Job
const* const, unsigned int>, std::_Select1st<std::pair<torca::master::Job const* const,
unsigned int> >, torca::master::JobCompare, std::allocator<std::pair<torca::master::Job
const* const, unsigned int> > >::_M_create_node (this=0x7fbe1b61d020, __x=...)
What steps will reproduce the problem?
I don't know how to reproduce it, it happened just once.
What is the expected output? What do you see instead?
My program crash at
return reinterpret_cast<Leaf*>(root_->ptrs[i1]->ptrs[i2])->values[i3];
values[i3] = 0xFFFFFFFFDEADBEAF。
What version of the product are you using? On what operating system?
I use google-perftools-1.7 on SUSE.
Please provide any additional information below.
```
Reported by `tyrionqq` on 2013-04-10 06:17:18
| defect | tcmalloc crash at checkcachedsizeclass originally reported on google code with id in google flushlogfilesunsafe min severity at thirdparty glog src logging cc in google anonymous namespace failuresignalhandler signal number signal info ucontext at thirdparty glog src signalhandler cc in checkcachedsizeclass ptr at thirdparty google perftools src tcmalloc cc in checkedmallocresult result at thirdparty google perftools src tcmalloc cc in anonymous namespace do malloc size at thirdparty google perftools src tcmalloc cc in anonymous namespace cpp alloc size nothrow false at thirdparty google perftools src tcmalloc cc in mallocblock allocate size type at thirdparty google perftools src debugallocation cc in debugallocate size type at thirdparty google perftools src debugallocation cc in cpp debug alloc size new type nothrow false at thirdparty google perftools src debugallocation cc in operator new size at thirdparty google perftools src debugallocation cc in gnu cxx new allocator std rb tree node std pair torca master job const const unsigned int allocate this n at usr local lib gcc unknown linux gnu include c ext new allocator h in std rb tree torca master job const std pair torca master job const const unsigned int std std pair torca master job const const unsigned int torca master jobcompare std allocator std pair torca master job const const unsigned int m get node this at usr local lib gcc unknown linux gnu include c bits stl tree h in std rb tree torca master job const std pair torca master job const const unsigned int std std pair torca master job const const unsigned int torca master jobcompare std allocator std pair torca master job const const unsigned int m create node this x what steps will reproduce the problem i don t know how to reproduce it it happened just once what is the expected output what do you see instead my program crash at return reinterpret cast root ptrs ptrs values values 。 what version of the product are you using on what operating system i use google perftools on suse please provide any additional information below reported by tyrionqq on | 1 |
329,446 | 10,019,946,478 | IssuesEvent | 2019-07-16 11:23:27 | aiidateam/aiida-core | https://api.github.com/repos/aiidateam/aiida-core | closed | Update requirements for `plumpy` dependency | priority/important topic/dependencies | New version has just been released `plumpy==0.14.2` | 1.0 | Update requirements for `plumpy` dependency - New version has just been released `plumpy==0.14.2` | non_defect | update requirements for plumpy dependency new version has just been released plumpy | 0 |
102,165 | 31,847,506,646 | IssuesEvent | 2023-09-14 21:13:34 | orbeon/orbeon-forms | https://api.github.com/repos/orbeon/orbeon-forms | closed | JavaScript error in Form Builder when copying a Formatted Text Area | Module: Form Builder | This happens when doing the copy while the menu is displayed. Somehow I don't get the error when I just move it. Below, I first try to move it from left to right, which works fine, but then I get the error when copying from right to left. This is with the latest code on the `2022.1-pe` branch.

| 1.0 | JavaScript error in Form Builder when copying a Formatted Text Area - This happens when doing the copy while the menu is displayed. Somehow I don't get the error when I just move it. Below, I first try to move it from left to right, which works fine, but then I get the error when copying from right to left. This is with the latest code on the `2022.1-pe` branch.

| non_defect | javascript error in form builder when copying a formatted text area this happens when doing the copy while the menu is displayed somehow i don t get the error when i just move it below i first try to move it from left to right which works fine but then i get the error when copying from right to left this is with the latest code on the pe branch | 0 |
37,605 | 8,468,380,240 | IssuesEvent | 2018-10-23 19:34:33 | idaholab/raven | https://api.github.com/repos/idaholab/raven | closed | MOOSE Input Drops Text After Second '=' | defect external interfaces priority_critical | --------
Issue Description
--------
##### What did you expect to see happen?
I expected `function = 'alpha := total_volume * current_power_density * (1 - tau); beta := tau * current_power_density; alpha * beta'` in my moose input file to be parsed to the same thing when it spawned all the jobs.
##### What did you see instead?
Instead it dropped all the text after the second `=` putting in the new input file `function = 'alpha :` causing the run to error.
##### Do you have a suggested fix for the development team?
I do! Expect a merge request soon. :smiley: And it is running now with the fix.
##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or improvement?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [x] 1. If the issue is a defect, is the defect fixed?
- [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [x] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [x] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
| 1.0 | MOOSE Input Drops Text After Second '=' - --------
Issue Description
--------
##### What did you expect to see happen?
I expected `function = 'alpha := total_volume * current_power_density * (1 - tau); beta := tau * current_power_density; alpha * beta'` in my moose input file to be parsed to the same thing when it spawned all the jobs.
##### What did you see instead?
Instead it dropped all the text after the second `=` putting in the new input file `function = 'alpha :` causing the run to error.
##### Do you have a suggested fix for the development team?
I do! Expect a merge request soon. :smiley: And it is running now with the fix.
##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or improvement?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [x] 1. If the issue is a defect, is the defect fixed?
- [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [x] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [x] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
| defect | moose input drops text after second issue description what did you expect to see happen i expected function alpha total volume current power density tau beta tau current power density alpha beta in my moose input file to be parsed to the same thing when it spawned all the jobs what did you see instead instead it dropped all the text after the second putting in the new input file function alpha causing the run to error do you have a suggested fix for the development team i do expect a merge request soon smiley and it is running now with the fix please attach the input file s that generate this error the simpler the input the faster we can find the issue for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided | 1 |
73,482 | 24,652,420,724 | IssuesEvent | 2022-10-17 19:51:44 | shedskin/shedskin | https://api.github.com/repos/shedskin/shedskin | closed | Crash when using string.split() | auto-migrated Priority-Medium Type-Defect platform-osx | **Issue by inn...@gmail.com**
_13 Sep 2011 at 3:27 GMT_
_Originally opened on Google Code_
---
```
What steps will reproduce the problem?
1. Compile and run the following code:
a =
'7,0,1,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,1,0,2,1,0,0,0,0,0,0,0,0,0,0
,0,0,0,0,0,0,0,0,2,2,1,0,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,3,1,0,0
,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'
b =
'7,0,1,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,1,0,2,1,0,0,0,0,0,0,0,0,0,0
,0,0,0,0,0,0,0,0,2,2,1,0,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,3,1,0,0
,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'
c =
'7,0,1,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,1,0,2,1,0,0,0,0,0,0,0,0,0,0
,0,0,0,0,0,0,0,0,2,2,1,0,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,3,1,0,0
,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'
while True:
d = a.split(',')
d = b.split(',')
d = c.split(',')
What is the expected output? What do you see instead?
Segfaults for some reason. Doesn't happen under gdb, which makes the problem
even trickier.
What version of the product are you using? On what operating system?
0.9, OS X 10.7.0, gcc 4.2.1
```
| 1.0 | Crash when using string.split() - **Issue by inn...@gmail.com**
_13 Sep 2011 at 3:27 GMT_
_Originally opened on Google Code_
---
```
What steps will reproduce the problem?
1. Compile and run the following code:
a =
'7,0,1,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,1,0,2,1,0,0,0,0,0,0,0,0,0,0
,0,0,0,0,0,0,0,0,2,2,1,0,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,3,1,0,0
,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'
b =
'7,0,1,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,1,0,2,1,0,0,0,0,0,0,0,0,0,0
,0,0,0,0,0,0,0,0,2,2,1,0,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,3,1,0,0
,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'
c =
'7,0,1,0,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,1,0,2,1,0,0,0,0,0,0,0,0,0,0
,0,0,0,0,0,0,0,0,2,2,1,0,3,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,3,1,0,0
,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'
while True:
d = a.split(',')
d = b.split(',')
d = c.split(',')
What is the expected output? What do you see instead?
Segfaults for some reason. Doesn't happen under gdb, which makes the problem
even trickier.
What version of the product are you using? On what operating system?
0.9, OS X 10.7.0, gcc 4.2.1
```
| defect | crash when using string split issue by inn gmail com sep at gmt originally opened on google code what steps will reproduce the problem compile and run the following code a b c while true d a split d b split d c split what is the expected output what do you see instead segfaults for some reason doesn t happen under gdb which makes the problem even trickier what version of the product are you using on what operating system os x gcc | 1 |
19,951 | 3,283,913,298 | IssuesEvent | 2015-10-28 14:46:34 | ARMmbed/uvisor | https://api.github.com/repos/ARMmbed/uvisor | opened | Hardware floating point support | defect | - add hardware floating point support for the main box as a first step
- extend floating point support to other secure boxes later
| 1.0 | Hardware floating point support - - add hardware floating point support for the main box as a first step
- extend floating point support to other secure boxes later
| defect | hardware floating point support add hardware floating point support for the main box as a first step extend floating point support to other secure boxes later | 1 |
10,165 | 2,618,939,869 | IssuesEvent | 2015-03-03 00:03:40 | marmarek/test | https://api.github.com/repos/marmarek/test | closed | unbind_all_network_devices: fix regexp to cover all pci devs | C: core P: major R: fixed T: defect | **Reported by joanna on 13 May 40287457 01:20 UTC**
The regexp used by unbind_all_network_devices script ignores devices that have hex digits (a-f) in their BDF addresses. As reported here:
http://groups.google.com/group/qubes-devel/browse_thread/thread/66d517a6c67085f9
| 1.0 | unbind_all_network_devices: fix regexp to cover all pci devs - **Reported by joanna on 13 May 40287457 01:20 UTC**
The regexp used by unbind_all_network_devices script ignores devices that have hex digits (a-f) in their BDF addresses. As reported here:
http://groups.google.com/group/qubes-devel/browse_thread/thread/66d517a6c67085f9
| defect | unbind all network devices fix regexp to cover all pci devs reported by joanna on may utc the regexp used by unbind all network devices script ignores devices that have hex digits a f in their bdf addresses as reported here | 1 |
110,172 | 13,906,623,893 | IssuesEvent | 2020-10-20 11:32:02 | httpwg/httpbis-issues | https://api.github.com/repos/httpwg/httpbis-issues | opened | List-type header fields vs Set-Cookie (Trac #129) | Incomplete Migration Migrated from Trac design p1-messaging | Migrated from https://trac.ietf.org/ticket/129
```json
{
"status": "closed",
"changetime": "2012-07-14T13:35:46",
"_ts": "1342272946296005",
"description": "Part 1, Section 4.2 ([http://tools.ietf.org/html/draft-ietf-httpbis-p1-messaging-03#section-4.2]) states:\n\n\"''Multiple message-header fields with the same field-name MAY be present in a message if and only if the entire field-value for that header field is defined as a comma-separated list [i.e., #(values)]. It MUST be possible to combine the multiple header fields into one \"field-name: field-value\" pair, without changing the semantics of the message, by appending each subsequent field-value to the first, each separated by a comma. The order in which header fields with the same field-name are received is therefore significant to the interpretation of the combined field value, and thus a proxy MUST NOT change the order of these field values when a message is forwarded.''\"\n\nThis is known to be incompatible with Set-Cookie as implemented in practice, as opposed as defined in RFC 2109. \n\nIn particular, Jamie Lokier points out:\n\n> RFC2109 is not implemented by anybody as far as I know.\n> \n> Firstly, cookie _values_ in Set-Cookie may contain a comma which\n> _mustn't_ be quoted because quotes are considered part of the value.\n> When a value is unquoted, RFC2109 says it must match token syntax, but\n> even today that's not conformed to. And RFC2109 doesn't describe an\n> \"expires=\" attribute, but of course nearly all cookies have one, and\n> they don't have the \"max-age=\" attribute with RFC2109 recommands.\n> Finally, as you note, unquoted comma in expires attributes - in fact\n> quoting is not allowed historically for that either.\n> \n> See how many RFC2109 non-compliances you can find in this header I got\n> today from Google, for example.\n> \n>\n{{{\nSet-Cookie: PREF=ID=823cb075fecf6437:TM=1195776675:LM=1195776675:S=WADqk8jBntt5y3gk; expires=Sun, 22-Nov-2009 00:11:15\n}}}\n> \n> (That nobody implements RFC2109 is implied in RFC2965, which obsoletes\n> RFC2109 and in section 9 talks about using Set-Cookie2 alongside\n> Netscape style Set-Cookies, not mentioning RFC2109 style Set-Cookiess.\n> I think this reflects the observation at the time that the change of\n> Set-Cookie syntax promoted in RFC2109 wasn't taken up, probably\n> because it's not backward compatible.)\n\n\nIt seems to me that it would be a service to implementors to minimally add a Note pointing out this special case.\n",
"reporter": "julian.reschke@gmx.de",
"cc": "",
"resolution": "fixed",
"time": "2008-08-13T20:05:12",
"component": "p1-messaging",
"summary": "List-type header fields vs Set-Cookie",
"priority": "normal",
"keywords": "",
"milestone": "unassigned",
"owner": "",
"type": "design",
"severity": "Active WG Document"
}
```
| 1.0 | List-type header fields vs Set-Cookie (Trac #129) - Migrated from https://trac.ietf.org/ticket/129
```json
{
"status": "closed",
"changetime": "2012-07-14T13:35:46",
"_ts": "1342272946296005",
"description": "Part 1, Section 4.2 ([http://tools.ietf.org/html/draft-ietf-httpbis-p1-messaging-03#section-4.2]) states:\n\n\"''Multiple message-header fields with the same field-name MAY be present in a message if and only if the entire field-value for that header field is defined as a comma-separated list [i.e., #(values)]. It MUST be possible to combine the multiple header fields into one \"field-name: field-value\" pair, without changing the semantics of the message, by appending each subsequent field-value to the first, each separated by a comma. The order in which header fields with the same field-name are received is therefore significant to the interpretation of the combined field value, and thus a proxy MUST NOT change the order of these field values when a message is forwarded.''\"\n\nThis is known to be incompatible with Set-Cookie as implemented in practice, as opposed as defined in RFC 2109. \n\nIn particular, Jamie Lokier points out:\n\n> RFC2109 is not implemented by anybody as far as I know.\n> \n> Firstly, cookie _values_ in Set-Cookie may contain a comma which\n> _mustn't_ be quoted because quotes are considered part of the value.\n> When a value is unquoted, RFC2109 says it must match token syntax, but\n> even today that's not conformed to. And RFC2109 doesn't describe an\n> \"expires=\" attribute, but of course nearly all cookies have one, and\n> they don't have the \"max-age=\" attribute with RFC2109 recommands.\n> Finally, as you note, unquoted comma in expires attributes - in fact\n> quoting is not allowed historically for that either.\n> \n> See how many RFC2109 non-compliances you can find in this header I got\n> today from Google, for example.\n> \n>\n{{{\nSet-Cookie: PREF=ID=823cb075fecf6437:TM=1195776675:LM=1195776675:S=WADqk8jBntt5y3gk; expires=Sun, 22-Nov-2009 00:11:15\n}}}\n> \n> (That nobody implements RFC2109 is implied in RFC2965, which obsoletes\n> RFC2109 and in section 9 talks about using Set-Cookie2 alongside\n> Netscape style Set-Cookies, not mentioning RFC2109 style Set-Cookiess.\n> I think this reflects the observation at the time that the change of\n> Set-Cookie syntax promoted in RFC2109 wasn't taken up, probably\n> because it's not backward compatible.)\n\n\nIt seems to me that it would be a service to implementors to minimally add a Note pointing out this special case.\n",
"reporter": "julian.reschke@gmx.de",
"cc": "",
"resolution": "fixed",
"time": "2008-08-13T20:05:12",
"component": "p1-messaging",
"summary": "List-type header fields vs Set-Cookie",
"priority": "normal",
"keywords": "",
"milestone": "unassigned",
"owner": "",
"type": "design",
"severity": "Active WG Document"
}
```
| non_defect | list type header fields vs set cookie trac migrated from json status closed changetime ts description part section states n n multiple message header fields with the same field name may be present in a message if and only if the entire field value for that header field is defined as a comma separated list it must be possible to combine the multiple header fields into one field name field value pair without changing the semantics of the message by appending each subsequent field value to the first each separated by a comma the order in which header fields with the same field name are received is therefore significant to the interpretation of the combined field value and thus a proxy must not change the order of these field values when a message is forwarded n nthis is known to be incompatible with set cookie as implemented in practice as opposed as defined in rfc n nin particular jamie lokier points out n n is not implemented by anybody as far as i know n n firstly cookie values in set cookie may contain a comma which n mustn t be quoted because quotes are considered part of the value n when a value is unquoted says it must match token syntax but n even today that s not conformed to and doesn t describe an n expires attribute but of course nearly all cookies have one and n they don t have the max age attribute with recommands n finally as you note unquoted comma in expires attributes in fact n quoting is not allowed historically for that either n n see how many non compliances you can find in this header i got n today from google for example n n n nset cookie pref id tm lm s expires sun nov n n n that nobody implements is implied in which obsoletes n and in section talks about using set alongside n netscape style set cookies not mentioning style set cookiess n i think this reflects the observation at the time that the change of n set cookie syntax promoted in wasn t taken up probably n because it s not backward compatible n n nit seems to me that it would be a service to implementors to minimally add a note pointing out this special case n reporter julian reschke gmx de cc resolution fixed time component messaging summary list type header fields vs set cookie priority normal keywords milestone unassigned owner type design severity active wg document | 0 |
28,355 | 2,701,067,469 | IssuesEvent | 2015-04-04 21:38:50 | TrinityCore/TrinityCore | https://api.github.com/repos/TrinityCore/TrinityCore | closed | [DB|NPC] Vhel'kur - Missing Waypoints | Comp-Database Feedback-PatchFix Priority-Cosmetic | http://de.wowhead.com/npc=21801
As you can see, this npc should have some waypoints and 'flys' in a circle.
If someoone profide some sniffs for this creature, i would copypaste it.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/4246515-db-npc-vhel-kur-missing-waypoints?utm_campaign=plugin&utm_content=tracker%2F1310&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F1310&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | [DB|NPC] Vhel'kur - Missing Waypoints - http://de.wowhead.com/npc=21801
As you can see, this npc should have some waypoints and 'flys' in a circle.
If someoone profide some sniffs for this creature, i would copypaste it.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/4246515-db-npc-vhel-kur-missing-waypoints?utm_campaign=plugin&utm_content=tracker%2F1310&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F1310&utm_medium=issues&utm_source=github).
</bountysource-plugin> | non_defect | vhel kur missing waypoints as you can see this npc should have some waypoints and flys in a circle if someoone profide some sniffs for this creature i would copypaste it want to back this issue we accept bounties via | 0 |
17,023 | 2,966,754,139 | IssuesEvent | 2015-07-12 07:03:58 | ElvisChiang/welly | https://api.github.com/repos/ElvisChiang/welly | closed | 站台無法編輯 | auto-migrated Priority-Medium Type-Defect | ```
如圖所示
原因不明的無法編輯站台
圖片中可以看出目前有四個儲存的站台
但是站台列表當中卻無列出任何東西,也無法選取
這個問題從OSX 10.8~10.9都一樣(我從10.8開始加入MAC家族的)
希望能解決這個BUG
謝謝
```
Original issue reported on code.google.com by `CHCOO...@gmail.com` on 6 Dec 2013 at 8:58
Attachments:
* [螢幕快照 2013-12-06 下午3.57.05.png](https://storage.googleapis.com/google-code-attachments/welly/issue-69/comment-0/螢幕快照 2013-12-06 下午3.57.05.png)
| 1.0 | 站台無法編輯 - ```
如圖所示
原因不明的無法編輯站台
圖片中可以看出目前有四個儲存的站台
但是站台列表當中卻無列出任何東西,也無法選取
這個問題從OSX 10.8~10.9都一樣(我從10.8開始加入MAC家族的)
希望能解決這個BUG
謝謝
```
Original issue reported on code.google.com by `CHCOO...@gmail.com` on 6 Dec 2013 at 8:58
Attachments:
* [螢幕快照 2013-12-06 下午3.57.05.png](https://storage.googleapis.com/google-code-attachments/welly/issue-69/comment-0/螢幕快照 2013-12-06 下午3.57.05.png)
| defect | 站台無法編輯 如圖所示 原因不明的無法編輯站台 圖片中可以看出目前有四個儲存的站台 但是站台列表當中卻無列出任何東西,也無法選取 這個問題從osx ( ) 希望能解決這個bug 謝謝 original issue reported on code google com by chcoo gmail com on dec at attachments png | 1 |
46,097 | 13,055,851,745 | IssuesEvent | 2020-07-30 02:55:35 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | reported issues with return values of tray.Execute(). (Trac #579) | IceTray Incomplete Migration Migrated from Trac defect | Migrated from https://code.icecube.wisc.edu/ticket/579
```json
{
"status": "closed",
"changetime": "2009-11-24T14:58:23",
"description": "As reported by David and Martin:\n\nHi Erik,\n\nNo, Martin's script ran without fatals. And without fatals, he should get normal, complete output files, and that is not what was happening, because flat-ntuple (the module he added to his script in order to study the results) calls TTree::Write() during the Finish() method, not in its destructor. So that is why it is important to call tray.Finish() after tray.Execute() terminated normally.\n\nDavid\n\nOn Tue, 24 Nov 2009, Erik Blaufuss wrote:\n\nHi David, Martin\n\nThat's a good suggestion, but why he's only seeing\nincomplete files, etc when tray.Execute() has a fatal call?\n\nWhy should there be any fatal calls during normal processing and\nwhy would you expect sensible content in a file after a log_fatal()\nwas generated somewhere?\n\nI didn't put too much thought into the bottom of that\nscript (copied from Robert F in reality), since I\nwas more concerned about the settings for the modules...\n\nLater\n-Erik\nOn Nov 24, 2009, at 8:13 AM, David Boersma wrote:\n\nHi Erik,\n\nMartin Bissok, PhD student here in Aachen, ran into some weird issues with incomplete/empty ROOT files, which we eventually traced down to this construction at the end of his python script:\n\ntry:\n tray.Execute()\nexcept Exception, ex:\n print ex\n tray.Finish()\n sys.exit(1)\n\nThis was copied from the official IC77 script:\nhttp://code.icecube.wisc.edu/svn/sandbox/blaufuss/ic77-retrig/base_filter_modules.py\nwhich Martin had used as a starting point to develop his GC filter work.\n\nThe error here is that tray.Finish() *should* also be run when tray.Execute() returns normally. Otherwise the Finish() method of modules does not get called, potentially leading to unfinished business like not properly closed files, unflushed buffers and what not. When we copied a tray.Finish() immediately after the Execute() line, Martin's ROOT files got the expected content. So, like this:\n\ntry:\n tray.Execute()\n tray.Finish()\nexcept Exception, ex:\n print ex\n tray.Finish()\n sys.exit(1)\n\nThough actually I'm not entirely certain if tray.Finish() is really useful after e.g. a log_fatal() was thrown or a segfault occurred. Maybe the tray.Finish() really wasn't meant to be in the except: leg, but intended to have been copied into the try: leg all along?\n\nDavid\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "invalid",
"_ts": "1259074703000000",
"component": "IceTray",
"summary": "reported issues with return values of tray.Execute().",
"priority": "normal",
"keywords": "",
"time": "2009-11-24T14:55:13",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| 1.0 | reported issues with return values of tray.Execute(). (Trac #579) - Migrated from https://code.icecube.wisc.edu/ticket/579
```json
{
"status": "closed",
"changetime": "2009-11-24T14:58:23",
"description": "As reported by David and Martin:\n\nHi Erik,\n\nNo, Martin's script ran without fatals. And without fatals, he should get normal, complete output files, and that is not what was happening, because flat-ntuple (the module he added to his script in order to study the results) calls TTree::Write() during the Finish() method, not in its destructor. So that is why it is important to call tray.Finish() after tray.Execute() terminated normally.\n\nDavid\n\nOn Tue, 24 Nov 2009, Erik Blaufuss wrote:\n\nHi David, Martin\n\nThat's a good suggestion, but why he's only seeing\nincomplete files, etc when tray.Execute() has a fatal call?\n\nWhy should there be any fatal calls during normal processing and\nwhy would you expect sensible content in a file after a log_fatal()\nwas generated somewhere?\n\nI didn't put too much thought into the bottom of that\nscript (copied from Robert F in reality), since I\nwas more concerned about the settings for the modules...\n\nLater\n-Erik\nOn Nov 24, 2009, at 8:13 AM, David Boersma wrote:\n\nHi Erik,\n\nMartin Bissok, PhD student here in Aachen, ran into some weird issues with incomplete/empty ROOT files, which we eventually traced down to this construction at the end of his python script:\n\ntry:\n tray.Execute()\nexcept Exception, ex:\n print ex\n tray.Finish()\n sys.exit(1)\n\nThis was copied from the official IC77 script:\nhttp://code.icecube.wisc.edu/svn/sandbox/blaufuss/ic77-retrig/base_filter_modules.py\nwhich Martin had used as a starting point to develop his GC filter work.\n\nThe error here is that tray.Finish() *should* also be run when tray.Execute() returns normally. Otherwise the Finish() method of modules does not get called, potentially leading to unfinished business like not properly closed files, unflushed buffers and what not. When we copied a tray.Finish() immediately after the Execute() line, Martin's ROOT files got the expected content. So, like this:\n\ntry:\n tray.Execute()\n tray.Finish()\nexcept Exception, ex:\n print ex\n tray.Finish()\n sys.exit(1)\n\nThough actually I'm not entirely certain if tray.Finish() is really useful after e.g. a log_fatal() was thrown or a segfault occurred. Maybe the tray.Finish() really wasn't meant to be in the except: leg, but intended to have been copied into the try: leg all along?\n\nDavid\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "invalid",
"_ts": "1259074703000000",
"component": "IceTray",
"summary": "reported issues with return values of tray.Execute().",
"priority": "normal",
"keywords": "",
"time": "2009-11-24T14:55:13",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| defect | reported issues with return values of tray execute trac migrated from json status closed changetime description as reported by david and martin n nhi erik n nno martin s script ran without fatals and without fatals he should get normal complete output files and that is not what was happening because flat ntuple the module he added to his script in order to study the results calls ttree write during the finish method not in its destructor so that is why it is important to call tray finish after tray execute terminated normally n ndavid n non tue nov erik blaufuss wrote n nhi david martin n nthat s a good suggestion but why he s only seeing nincomplete files etc when tray execute has a fatal call n nwhy should there be any fatal calls during normal processing and nwhy would you expect sensible content in a file after a log fatal nwas generated somewhere n ni didn t put too much thought into the bottom of that nscript copied from robert f in reality since i nwas more concerned about the settings for the modules n nlater n erik non nov at am david boersma wrote n nhi erik n nmartin bissok phd student here in aachen ran into some weird issues with incomplete empty root files which we eventually traced down to this construction at the end of his python script n ntry n tray execute nexcept exception ex n print ex n tray finish n sys exit n nthis was copied from the official script n martin had used as a starting point to develop his gc filter work n nthe error here is that tray finish should also be run when tray execute returns normally otherwise the finish method of modules does not get called potentially leading to unfinished business like not properly closed files unflushed buffers and what not when we copied a tray finish immediately after the execute line martin s root files got the expected content so like this n ntry n tray execute n tray finish nexcept exception ex n print ex n tray finish n sys exit n nthough actually i m not entirely certain if tray finish is really useful after e g a log fatal was thrown or a segfault occurred maybe the tray finish really wasn t meant to be in the except leg but intended to have been copied into the try leg all along n ndavid n reporter blaufuss cc resolution invalid ts component icetray summary reported issues with return values of tray execute priority normal keywords time milestone owner troy type defect | 1 |
23,406 | 3,813,104,726 | IssuesEvent | 2016-03-28 01:52:41 | huboard/huboard | https://api.github.com/repos/huboard/huboard | closed | Important display problems since we installed HB:E v1.6.107 | 3 - In Review Defect Enterprise ★★★ | We are experiencing the following problems:
- [ ] Tasks columns display is not fitting the screen
- [ ] Columns width are not always the same (this cause the problem above)
- [ ] The X that was visible on a filter when selected is not anymore displayed
We'll have to roll-out to the previous release as these display problems are too much disrupting unless this can be fixed asap.
<!---
@huboard:{"order":3.161309661618673e-10,"milestone_order":631,"custom_state":""}
-->
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/30589749-important-display-problems-since-we-installed-hb-e-v1-6-107?utm_campaign=plugin&utm_content=tracker%2F282362&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F282362&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | Important display problems since we installed HB:E v1.6.107 - We are experiencing the following problems:
- [ ] Tasks columns display is not fitting the screen
- [ ] Columns width are not always the same (this cause the problem above)
- [ ] The X that was visible on a filter when selected is not anymore displayed
We'll have to roll-out to the previous release as these display problems are too much disrupting unless this can be fixed asap.
<!---
@huboard:{"order":3.161309661618673e-10,"milestone_order":631,"custom_state":""}
-->
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/30589749-important-display-problems-since-we-installed-hb-e-v1-6-107?utm_campaign=plugin&utm_content=tracker%2F282362&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F282362&utm_medium=issues&utm_source=github).
</bountysource-plugin> | defect | important display problems since we installed hb e we are experiencing the following problems tasks columns display is not fitting the screen columns width are not always the same this cause the problem above the x that was visible on a filter when selected is not anymore displayed we ll have to roll out to the previous release as these display problems are too much disrupting unless this can be fixed asap huboard order milestone order custom state want to back this issue we accept bounties via | 1 |
1,108 | 2,856,001,064 | IssuesEvent | 2015-06-02 13:00:32 | zaphoyd/websocketpp | https://api.github.com/repos/zaphoyd/websocketpp | closed | Two SSL records for one message. | iostream Transport Performance Issue | Hello,
I wrote a programm using Websocket++, what exchange mesages between client and server.
I didn't use any embeded transports, but use library in iostream manner, handling send and receive by myself.
Whenewer I call client.send(), I got two writes in my write handle, set by set_write_handler().
First for hrader, and second - for message body.
As a result, I got two SSL records in channel.
What produces unnecessary overhead. Two bytes of header results in may bytes of SSL record.
I, as a library user, would prefer to get in write handle complete message, with header, body and footer (if any), so I can wrap it into single SSL record. | True | Two SSL records for one message. - Hello,
I wrote a programm using Websocket++, what exchange mesages between client and server.
I didn't use any embeded transports, but use library in iostream manner, handling send and receive by myself.
Whenewer I call client.send(), I got two writes in my write handle, set by set_write_handler().
First for hrader, and second - for message body.
As a result, I got two SSL records in channel.
What produces unnecessary overhead. Two bytes of header results in may bytes of SSL record.
I, as a library user, would prefer to get in write handle complete message, with header, body and footer (if any), so I can wrap it into single SSL record. | non_defect | two ssl records for one message hello i wrote a programm using websocket what exchange mesages between client and server i didn t use any embeded transports but use library in iostream manner handling send and receive by myself whenewer i call client send i got two writes in my write handle set by set write handler first for hrader and second for message body as a result i got two ssl records in channel what produces unnecessary overhead two bytes of header results in may bytes of ssl record i as a library user would prefer to get in write handle complete message with header body and footer if any so i can wrap it into single ssl record | 0 |
100,562 | 30,724,406,268 | IssuesEvent | 2023-07-27 18:25:14 | apache/beam | https://api.github.com/repos/apache/beam | closed | ./gradlew :beam-runners-google-cloud-dataflow-java:validatesRunner fails due to invalid GCS path | build P3 bug | The default root is gs://temp-storage-for-end-to-end-tests/ and the code adds a slash and then the test case name, so it results in two slashes in a row, rejected by the filesystem implementation.
Imported from Jira [BEAM-6715](https://issues.apache.org/jira/browse/BEAM-6715). Original Jira may contain additional context.
Reported by: kenn. | 1.0 | ./gradlew :beam-runners-google-cloud-dataflow-java:validatesRunner fails due to invalid GCS path - The default root is gs://temp-storage-for-end-to-end-tests/ and the code adds a slash and then the test case name, so it results in two slashes in a row, rejected by the filesystem implementation.
Imported from Jira [BEAM-6715](https://issues.apache.org/jira/browse/BEAM-6715). Original Jira may contain additional context.
Reported by: kenn. | non_defect | gradlew beam runners google cloud dataflow java validatesrunner fails due to invalid gcs path the default root is gs temp storage for end to end tests and the code adds a slash and then the test case name so it results in two slashes in a row rejected by the filesystem implementation imported from jira original jira may contain additional context reported by kenn | 0 |
217,689 | 24,348,934,982 | IssuesEvent | 2022-10-02 17:48:47 | venkateshreddypala/AngOCR | https://api.github.com/repos/venkateshreddypala/AngOCR | closed | CVE-2020-7598 (Medium) detected in minimist-0.0.8.tgz, minimist-1.2.0.tgz - autoclosed | security vulnerability | ## CVE-2020-7598 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary>
<p>
<details><summary><b>minimist-0.0.8.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p>
<p>Path to dependency file: /AngOCR/ui/package.json</p>
<p>Path to vulnerable library: /ui/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- karma-3.1.1.tgz (Root Library)
- optimist-0.6.1.tgz
- :x: **minimist-0.0.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimist-1.2.0.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p>
<p>Path to dependency file: /AngOCR/ui/package.json</p>
<p>Path to vulnerable library: /ui/node_modules/@angular/compiler-cli/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- ts-node-4.1.0.tgz (Root Library)
- :x: **minimist-1.2.0.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload.
<p>Publish Date: 2020-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-03-11</p>
<p>Fix Resolution (minimist): 0.2.1</p>
<p>Direct dependency fix Resolution (karma): 5.0.0</p><p>Fix Resolution (minimist): 1.2.3</p>
<p>Direct dependency fix Resolution (ts-node): 5.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7598 (Medium) detected in minimist-0.0.8.tgz, minimist-1.2.0.tgz - autoclosed - ## CVE-2020-7598 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary>
<p>
<details><summary><b>minimist-0.0.8.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p>
<p>Path to dependency file: /AngOCR/ui/package.json</p>
<p>Path to vulnerable library: /ui/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- karma-3.1.1.tgz (Root Library)
- optimist-0.6.1.tgz
- :x: **minimist-0.0.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimist-1.2.0.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p>
<p>Path to dependency file: /AngOCR/ui/package.json</p>
<p>Path to vulnerable library: /ui/node_modules/@angular/compiler-cli/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- ts-node-4.1.0.tgz (Root Library)
- :x: **minimist-1.2.0.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload.
<p>Publish Date: 2020-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-03-11</p>
<p>Fix Resolution (minimist): 0.2.1</p>
<p>Direct dependency fix Resolution (karma): 5.0.0</p><p>Fix Resolution (minimist): 1.2.3</p>
<p>Direct dependency fix Resolution (ts-node): 5.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in minimist tgz minimist tgz autoclosed cve medium severity vulnerability vulnerable libraries minimist tgz minimist tgz minimist tgz parse argument options library home page a href path to dependency file angocr ui package json path to vulnerable library ui node modules minimist package json dependency hierarchy karma tgz root library optimist tgz x minimist tgz vulnerable library minimist tgz parse argument options library home page a href path to dependency file angocr ui package json path to vulnerable library ui node modules angular compiler cli node modules minimist package json dependency hierarchy ts node tgz root library x minimist tgz vulnerable library vulnerability details minimist before could be tricked into adding or modifying properties of object prototype using a constructor or proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution minimist direct dependency fix resolution karma fix resolution minimist direct dependency fix resolution ts node step up your open source security game with mend | 0 |
780,747 | 27,406,713,012 | IssuesEvent | 2023-03-01 07:31:14 | prgrms-web-devcourse/Team-Kkini-Mukvengers-FE | https://api.github.com/repos/prgrms-web-devcourse/Team-Kkini-Mukvengers-FE | opened | google maps의 remote image를 사용하는 img 태그 최적화 | Priority: Medium Perf | ## 📕 작업 설명
> google maps의 remote image를 사용하는 img 태그 최적화
## 📖 To-Do list
- [ ] Pre Loading
- [ ] webp 변환
- remote image라서 가능할지 안할지 확실치 않음.
- [ ] srcSet, sizes 속성 적용
| 1.0 | google maps의 remote image를 사용하는 img 태그 최적화 - ## 📕 작업 설명
> google maps의 remote image를 사용하는 img 태그 최적화
## 📖 To-Do list
- [ ] Pre Loading
- [ ] webp 변환
- remote image라서 가능할지 안할지 확실치 않음.
- [ ] srcSet, sizes 속성 적용
| non_defect | google maps의 remote image를 사용하는 img 태그 최적화 📕 작업 설명 google maps의 remote image를 사용하는 img 태그 최적화 📖 to do list pre loading webp 변환 remote image라서 가능할지 안할지 확실치 않음 srcset sizes 속성 적용 | 0 |
25,570 | 4,386,598,378 | IssuesEvent | 2016-08-08 13:23:26 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | closed | Carousel page links redirect to # | defect | In showcase, carousel page links are not working. Currently they redirect to http://www.primefaces.org/primeng/# Instead of changing carousel page. I tested showcase http://www.primefaces.org/primeng/#/carousel with Firefox and Chrome and both browsers suffered from the same issue. | 1.0 | Carousel page links redirect to # - In showcase, carousel page links are not working. Currently they redirect to http://www.primefaces.org/primeng/# Instead of changing carousel page. I tested showcase http://www.primefaces.org/primeng/#/carousel with Firefox and Chrome and both browsers suffered from the same issue. | defect | carousel page links redirect to in showcase carousel page links are not working currently they redirect to instead of changing carousel page i tested showcase with firefox and chrome and both browsers suffered from the same issue | 1 |
23,464 | 3,829,548,670 | IssuesEvent | 2016-03-31 11:11:18 | pearu/pylibnidaqmx | https://api.github.com/repos/pearu/pylibnidaqmx | closed | Add AnalogOutput status calls | auto-migrated Priority-Medium Type-Defect | ```
I'm attaching a patch that adds the GetWriteCurrWritePos() and
GetWriteTotalSampPerChanGenerated() calls.
These are useful for getting the status of an AnalogOutput task while it's
running.
I put them in the Task base object. Perhaps these should be split into
AnalogOutput and the corresponding GetRead calls should be in AnalogInput.
Thanks,
Mark
diff --git a/nidaqmx/libnidaqmx.py b/nidaqmx/libnidaqmx.py
index a349b72..d7e86d4 100644
--- a/nidaqmx/libnidaqmx.py
+++ b/nidaqmx/libnidaqmx.py
@@ -2556,6 +2556,26 @@ class Task(uInt32):
CALL('GetReadTotalSampPerChanAcquired', self, ctypes.byref(d))
return d.value
+ def get_write_current_position(self):
+ """
+ Indicates the position in the buffer of the next sample to generate.
+ This value is the same for all channels in the task.
+ """
+ d = uInt64(0)
+ CALL('GetWriteCurrWritePos', self, ctypes.byref(d))
+ return d.value
+
+ def get_write_samples_per_channel_generated(self):
+ """
+ Indicates the total number of samples generated by each
+ channel. NI-DAQmx returns a single value because this value is
+ the same for all channels.
+ """
+ d = uInt64(0)
+ CALL('GetWriteTotalSampPerChanGenerated', self, ctypes.byref(d))
+ return d.value
+
+
def wait_until_done(self, timeout=-1):
"""
Waits for the measurement or generation to complete. Use this
```
Original issue reported on code.google.com by `mhis...@gmail.com` on 29 Apr 2014 at 9:51 | 1.0 | Add AnalogOutput status calls - ```
I'm attaching a patch that adds the GetWriteCurrWritePos() and
GetWriteTotalSampPerChanGenerated() calls.
These are useful for getting the status of an AnalogOutput task while it's
running.
I put them in the Task base object. Perhaps these should be split into
AnalogOutput and the corresponding GetRead calls should be in AnalogInput.
Thanks,
Mark
diff --git a/nidaqmx/libnidaqmx.py b/nidaqmx/libnidaqmx.py
index a349b72..d7e86d4 100644
--- a/nidaqmx/libnidaqmx.py
+++ b/nidaqmx/libnidaqmx.py
@@ -2556,6 +2556,26 @@ class Task(uInt32):
CALL('GetReadTotalSampPerChanAcquired', self, ctypes.byref(d))
return d.value
+ def get_write_current_position(self):
+ """
+ Indicates the position in the buffer of the next sample to generate.
+ This value is the same for all channels in the task.
+ """
+ d = uInt64(0)
+ CALL('GetWriteCurrWritePos', self, ctypes.byref(d))
+ return d.value
+
+ def get_write_samples_per_channel_generated(self):
+ """
+ Indicates the total number of samples generated by each
+ channel. NI-DAQmx returns a single value because this value is
+ the same for all channels.
+ """
+ d = uInt64(0)
+ CALL('GetWriteTotalSampPerChanGenerated', self, ctypes.byref(d))
+ return d.value
+
+
def wait_until_done(self, timeout=-1):
"""
Waits for the measurement or generation to complete. Use this
```
Original issue reported on code.google.com by `mhis...@gmail.com` on 29 Apr 2014 at 9:51 | defect | add analogoutput status calls i m attaching a patch that adds the getwritecurrwritepos and getwritetotalsampperchangenerated calls these are useful for getting the status of an analogoutput task while it s running i put them in the task base object perhaps these should be split into analogoutput and the corresponding getread calls should be in analoginput thanks mark diff git a nidaqmx libnidaqmx py b nidaqmx libnidaqmx py index a nidaqmx libnidaqmx py b nidaqmx libnidaqmx py class task call getreadtotalsampperchanacquired self ctypes byref d return d value def get write current position self indicates the position in the buffer of the next sample to generate this value is the same for all channels in the task d call getwritecurrwritepos self ctypes byref d return d value def get write samples per channel generated self indicates the total number of samples generated by each channel ni daqmx returns a single value because this value is the same for all channels d call getwritetotalsampperchangenerated self ctypes byref d return d value def wait until done self timeout waits for the measurement or generation to complete use this original issue reported on code google com by mhis gmail com on apr at | 1 |
16,430 | 2,894,559,830 | IssuesEvent | 2015-06-16 00:50:19 | obophenotype/cell-ontology | https://api.github.com/repos/obophenotype/cell-ontology | closed | Move repo to github | auto-migrated Priority-Medium Type-Defect | ```
google code will probably go away at some point in the future. It won't be
missed by me.
We should probably think about moving in the next 12 months. Issues:
* tracker migration (good incentive to go through and close everything?)
* ensuring no barriers of entry to non git-experts (todo: test github's svn bridge)
```
Original issue reported on code.google.com by `cmung...@gmail.com` on 31 Mar 2014 at 2:29 | 1.0 | Move repo to github - ```
google code will probably go away at some point in the future. It won't be
missed by me.
We should probably think about moving in the next 12 months. Issues:
* tracker migration (good incentive to go through and close everything?)
* ensuring no barriers of entry to non git-experts (todo: test github's svn bridge)
```
Original issue reported on code.google.com by `cmung...@gmail.com` on 31 Mar 2014 at 2:29 | defect | move repo to github google code will probably go away at some point in the future it won t be missed by me we should probably think about moving in the next months issues tracker migration good incentive to go through and close everything ensuring no barriers of entry to non git experts todo test github s svn bridge original issue reported on code google com by cmung gmail com on mar at | 1 |
4,283 | 2,610,090,817 | IssuesEvent | 2015-02-26 18:27:28 | chrsmith/dsdsdaadf | https://api.github.com/repos/chrsmith/dsdsdaadf | opened | 深圳痘痘治疗 | auto-migrated Priority-Medium Type-Defect | ```
深圳痘痘治疗【深圳韩方科颜全国热线400-869-1818,24小时QQ4008
691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方—��
�韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科颜�
��业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康祛
痘技术并结合先进“先进豪华彩光”仪,开创国内专业治疗��
�刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:44 | 1.0 | 深圳痘痘治疗 - ```
深圳痘痘治疗【深圳韩方科颜全国热线400-869-1818,24小时QQ4008
691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方—��
�韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科颜�
��业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康祛
痘技术并结合先进“先进豪华彩光”仪,开创国内专业治疗��
�刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:44 | defect | 深圳痘痘治疗 深圳痘痘治疗【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方—�� �韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科颜� ��业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康祛 痘技术并结合先进“先进豪华彩光”仪,开创国内专业治疗�� �刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘。 original issue reported on code google com by szft com on may at | 1 |
821,269 | 30,813,811,136 | IssuesEvent | 2023-08-01 12:16:18 | wp-media/wp-rocket | https://api.github.com/repos/wp-media/wp-rocket | opened | Used CSS isnot automatically generated for home page when switch theme while preload is off | type: bug priority: low severity: moderate module: remove unused css | **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version => 3.14.2.1
- Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
When we switch theme while preload is off, we display a notice that used CSS was created for the home page although it was not
**To Reproduce**
Steps to reproduce the behavior:
1. Disable preload
2. Enable RUCSS
3. Visit the home page => used CSS is there
4. switch the theme and check the used CSS table
**Expected behavior**
Entry is added for the home page
**Additional context**
Add any other context about the problem here.
- It was working fine on 3.12.6.1
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| 1.0 | Used CSS isnot automatically generated for home page when switch theme while preload is off - **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version => 3.14.2.1
- Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
When we switch theme while preload is off, we display a notice that used CSS was created for the home page although it was not
**To Reproduce**
Steps to reproduce the behavior:
1. Disable preload
2. Enable RUCSS
3. Visit the home page => used CSS is there
4. switch the theme and check the used CSS table
**Expected behavior**
Entry is added for the home page
**Additional context**
Add any other context about the problem here.
- It was working fine on 3.12.6.1
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| non_defect | used css isnot automatically generated for home page when switch theme while preload is off before submitting an issue please check that you’ve completed the following steps made sure you’re on the latest version used the search feature to ensure that the bug hasn’t been reported before describe the bug when we switch theme while preload is off we display a notice that used css was created for the home page although it was not to reproduce steps to reproduce the behavior disable preload enable rucss visit the home page used css is there switch the theme and check the used css table expected behavior entry is added for the home page additional context add any other context about the problem here it was working fine on backlog grooming for wp media dev team use only reproduce the problem identify the root cause scope a solution estimate the effort | 0 |
94,533 | 27,226,327,869 | IssuesEvent | 2023-02-21 09:58:26 | tensorflow/tfjs | https://api.github.com/repos/tensorflow/tfjs | opened | Docker M2 Macbook ARM invalid ELF header | type:build/install | I use a M2 MacBook.
We integrate tensorflow/tfjs into our nestJS backend.
When I start the nestJS enviorment for itself, everthing is fine, and tensorflow works.
Wenn I use Docker to deploy it on a DigitalOcean Dropet, everything is fine and tensorflow works.
**Describe the problem**
When i use the same Dockerfile (which i use to deploy on digital ocean) to start it local on my M2, it crashs.
`NODE_OPTIONS='--max-old-space-size=3000' nest start
2023-02-20 18:21:41 Error: /usr/src/app/node_modules/@tensorflow/tfjs-node/lib/napi-v8/tfjs_binding.node: invalid ELF header
2023-02-20 18:21:41 at Object.Module._extensions..node (node:internal/modules/cjs/loader:1302:18)
2023-02-20 18:21:41 at Module.load (node:internal/modules/cjs/loader:1081:32)
2023-02-20 18:21:41 at Function.Module._load (node:internal/modules/cjs/loader:922:12)
2023-02-20 18:21:41 at Module.require (node:internal/modules/cjs/loader:1105:19)
2023-02-20 18:21:41 at require (node:internal/modules/cjs/helpers:103:18)
2023-02-20 18:21:41 at Object.<anonymous> (/usr/src/app/node_modules/@tensorflow/tfjs-node/dist/index.js:72:16)
2023-02-20 18:21:41 at Module._compile (node:internal/modules/cjs/loader:1218:14)
2023-02-20 18:21:41 at Object.Module._extensions..js (node:internal/modules/cjs/loader:1272:10)
2023-02-20 18:21:41 at Module.load (node:internal/modules/cjs/loader:1081:32)
2023-02-20 18:21:41 at Function.Module._load (node:internal/modules/cjs/loader:922:12)
2023-02-20 18:21:41 error Command failed with exit code 1.`
### Dockerfile
`FROM node:18.12.1
EXPOSE 8080
WORKDIR /usr/src/app
COPY package*.json ./
COPY yarn.lock ./
RUN yarn
COPY . .
CMD [ "yarn","start"]
`
### These were my previous solutions that did not work
1. Add ` platform: linux/amd64` to the correct place in the docker-compose file, whit this result
<img width="118" alt="image" src="https://user-images.githubusercontent.com/125969421/220311961-6415f84c-2628-4408-85af-08befaa5e789.png">
Thanks for helping me :)
| 1.0 | Docker M2 Macbook ARM invalid ELF header - I use a M2 MacBook.
We integrate tensorflow/tfjs into our nestJS backend.
When I start the nestJS enviorment for itself, everthing is fine, and tensorflow works.
Wenn I use Docker to deploy it on a DigitalOcean Dropet, everything is fine and tensorflow works.
**Describe the problem**
When i use the same Dockerfile (which i use to deploy on digital ocean) to start it local on my M2, it crashs.
`NODE_OPTIONS='--max-old-space-size=3000' nest start
2023-02-20 18:21:41 Error: /usr/src/app/node_modules/@tensorflow/tfjs-node/lib/napi-v8/tfjs_binding.node: invalid ELF header
2023-02-20 18:21:41 at Object.Module._extensions..node (node:internal/modules/cjs/loader:1302:18)
2023-02-20 18:21:41 at Module.load (node:internal/modules/cjs/loader:1081:32)
2023-02-20 18:21:41 at Function.Module._load (node:internal/modules/cjs/loader:922:12)
2023-02-20 18:21:41 at Module.require (node:internal/modules/cjs/loader:1105:19)
2023-02-20 18:21:41 at require (node:internal/modules/cjs/helpers:103:18)
2023-02-20 18:21:41 at Object.<anonymous> (/usr/src/app/node_modules/@tensorflow/tfjs-node/dist/index.js:72:16)
2023-02-20 18:21:41 at Module._compile (node:internal/modules/cjs/loader:1218:14)
2023-02-20 18:21:41 at Object.Module._extensions..js (node:internal/modules/cjs/loader:1272:10)
2023-02-20 18:21:41 at Module.load (node:internal/modules/cjs/loader:1081:32)
2023-02-20 18:21:41 at Function.Module._load (node:internal/modules/cjs/loader:922:12)
2023-02-20 18:21:41 error Command failed with exit code 1.`
### Dockerfile
`FROM node:18.12.1
EXPOSE 8080
WORKDIR /usr/src/app
COPY package*.json ./
COPY yarn.lock ./
RUN yarn
COPY . .
CMD [ "yarn","start"]
`
### These were my previous solutions that did not work
1. Add ` platform: linux/amd64` to the correct place in the docker-compose file, whit this result
<img width="118" alt="image" src="https://user-images.githubusercontent.com/125969421/220311961-6415f84c-2628-4408-85af-08befaa5e789.png">
Thanks for helping me :)
| non_defect | docker macbook arm invalid elf header i use a macbook we integrate tensorflow tfjs into our nestjs backend when i start the nestjs enviorment for itself everthing is fine and tensorflow works wenn i use docker to deploy it on a digitalocean dropet everything is fine and tensorflow works describe the problem when i use the same dockerfile which i use to deploy on digital ocean to start it local on my it crashs node options max old space size nest start error usr src app node modules tensorflow tfjs node lib napi tfjs binding node invalid elf header at object module extensions node node internal modules cjs loader at module load node internal modules cjs loader at function module load node internal modules cjs loader at module require node internal modules cjs loader at require node internal modules cjs helpers at object usr src app node modules tensorflow tfjs node dist index js at module compile node internal modules cjs loader at object module extensions js node internal modules cjs loader at module load node internal modules cjs loader at function module load node internal modules cjs loader error command failed with exit code dockerfile from node expose workdir usr src app copy package json copy yarn lock run yarn copy cmd these were my previous solutions that did not work add platform linux to the correct place in the docker compose file whit this result img width alt image src thanks for helping me | 0 |
53,442 | 13,261,617,680 | IssuesEvent | 2020-08-20 20:13:36 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | [DomTools] Bring back I3IsolatedHitsCutModule (Trac #1397) | Migrated from Trac combo core defect | Or at least provide an alternative for filter scripts IceCube_BaseProc.py.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1397">https://code.icecube.wisc.edu/projects/icecube/ticket/1397</a>, reported by olivasand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:07",
"_ts": "1458335647931556",
"description": "Or at least provide an alternative for filter scripts IceCube_BaseProc.py.",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"time": "2015-10-14T07:06:54",
"component": "combo core",
"summary": "[DomTools] Bring back I3IsolatedHitsCutModule",
"priority": "blocker",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [DomTools] Bring back I3IsolatedHitsCutModule (Trac #1397) - Or at least provide an alternative for filter scripts IceCube_BaseProc.py.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1397">https://code.icecube.wisc.edu/projects/icecube/ticket/1397</a>, reported by olivasand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:07",
"_ts": "1458335647931556",
"description": "Or at least provide an alternative for filter scripts IceCube_BaseProc.py.",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"time": "2015-10-14T07:06:54",
"component": "combo core",
"summary": "[DomTools] Bring back I3IsolatedHitsCutModule",
"priority": "blocker",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| defect | bring back trac or at least provide an alternative for filter scripts icecube baseproc py migrated from json status closed changetime ts description or at least provide an alternative for filter scripts icecube baseproc py reporter olivas cc resolution fixed time component combo core summary bring back priority blocker keywords milestone owner olivas type defect | 1 |
959 | 2,594,375,625 | IssuesEvent | 2015-02-20 02:32:23 | BALL-Project/ball | https://api.github.com/repos/BALL-Project/ball | closed | Changing RTfact->OpenGL in Continuous Loop mode results in crash | C: VIEW P: major R: fixed T: defect | **Reported by nicste on 12 Jun 41056064 04:55 UTC**
Changing from RTfact rendering mode to OpenGL rendering mode while having the "Continuous Loop" enabled results in a crash of BALLView | 1.0 | Changing RTfact->OpenGL in Continuous Loop mode results in crash - **Reported by nicste on 12 Jun 41056064 04:55 UTC**
Changing from RTfact rendering mode to OpenGL rendering mode while having the "Continuous Loop" enabled results in a crash of BALLView | defect | changing rtfact opengl in continuous loop mode results in crash reported by nicste on jun utc changing from rtfact rendering mode to opengl rendering mode while having the continuous loop enabled results in a crash of ballview | 1 |
75,781 | 9,320,813,996 | IssuesEvent | 2019-03-27 00:58:41 | JordanMartinez/purescript-jordans-reference | https://api.github.com/repos/JordanMartinez/purescript-jordans-reference | opened | Wrap a 'Show' type class in a newtype | Design-Patterns enhancement | Probably a horrible name for this issue, but that's essentially what's going on here.
From Liam on the Slack channel:
```purescript
newtype Showable a = Showable (∀ r . (Show a ⇒ a → r) → r)
makeShowable ∷ ∀ a . Show a ⇒ a → Showable a
makeShowable a = Showable \f → f a
runShowable ∷ ∀ a . Showable a → String
runShowable (Showable s) = s show
``` | 1.0 | Wrap a 'Show' type class in a newtype - Probably a horrible name for this issue, but that's essentially what's going on here.
From Liam on the Slack channel:
```purescript
newtype Showable a = Showable (∀ r . (Show a ⇒ a → r) → r)
makeShowable ∷ ∀ a . Show a ⇒ a → Showable a
makeShowable a = Showable \f → f a
runShowable ∷ ∀ a . Showable a → String
runShowable (Showable s) = s show
``` | non_defect | wrap a show type class in a newtype probably a horrible name for this issue but that s essentially what s going on here from liam on the slack channel purescript newtype showable a showable ∀ r show a ⇒ a → r → r makeshowable ∷ ∀ a show a ⇒ a → showable a makeshowable a showable f → f a runshowable ∷ ∀ a showable a → string runshowable showable s s show | 0 |
11,428 | 2,651,459,035 | IssuesEvent | 2015-03-16 11:40:27 | douglasdrumond/cloaked-computing-machine | https://api.github.com/repos/douglasdrumond/cloaked-computing-machine | opened | [CLOSED] Apple Key + Key Customization Doesn't Work | auto-migrated Priority-Medium Type-Defect | <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday Mar 16, 2015 at 09:15 GMT_
_Originally opened as https://github.com/douglasdrumond/macvim/issues/14_
----
```
For example, customizing <D-2>, which works in other versions of vim, does
not work in this application.
```
Original issue reported on code.google.com by `JasonCRo...@gmail.com` on 12 Sep 2007 at 4:02
| 1.0 | [CLOSED] Apple Key + Key Customization Doesn't Work - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday Mar 16, 2015 at 09:15 GMT_
_Originally opened as https://github.com/douglasdrumond/macvim/issues/14_
----
```
For example, customizing <D-2>, which works in other versions of vim, does
not work in this application.
```
Original issue reported on code.google.com by `JasonCRo...@gmail.com` on 12 Sep 2007 at 4:02
| defect | apple key key customization doesn t work issue by monday mar at gmt originally opened as for example customizing which works in other versions of vim does not work in this application original issue reported on code google com by jasoncro gmail com on sep at | 1 |
45,261 | 12,691,232,362 | IssuesEvent | 2020-06-21 16:01:22 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | Add support to retrieve data from associations of _joinData (a.k.a ternary association) | ORM defect pinned | This is a (multiple allowed):
* [ ] bug
* [x] enhancement
* [x] feature-discussion (RFC)
* CakePHP Version: 3.7.
* Platform and Target: PHP 7.3 Mysql 5.
### What you did
I have created a simple database with those models:
_Doctors_ **belongsToMany** _Patients_ **through** _Appointments_
_Appointments_ **hasMany** _Examinations_
Full SQL: https://pastebin.com/9fY8Vibx
As you can see, I have a classic "ternary association".
So, I have tried to create a kind-of report showing _all doctors with theirs appointments date including patient name and a list of requested exams_
**Controller**:
```php
$table = TableRegistry::getTableLocator()->get('Doctors');
$data = $table->find()
->contain('Patients')
->contain('Appointments.Examinations'); //Error !
$this->set(compact('data'));
```
**Template:**
```php
<?php
echo "<table>";
echo $this->Html->tableHeaders(['Doctor','Patient','Appointment Date', 'Exams']);
foreach ($data as $doctor){
echo "<tr>";
echo "<td rowspan='".count($doctor->patients)."'>".$doctor->name."</td>";
foreach ($doctor->patients as $key => $patient) {
if($key == !array_key_first($doctor->patients)){
echo "<tr>";
}
echo "<td>".$patient->name."</td>";
echo "<td>".$patient->_joinData->date."</td>";
//$exams = array_column($patient->_joinData->examinations, 'name'); //Error !
//echo "<td>".implode('; ', $exams); //Error !
echo "</tr>";
}
}
```
### What happened
The "pivot" table it is not associated with doctors so it is no possible to reach by the "normal" way the associated data of this _joinData
### What you expected to happen
Be able to get Associated data from a model defined using "setThrough".
Some changes could be applied on **findAssociation()** to "infere" (like a OWL Reasoner) that Doctors (or Patients) has a association with Appointments thanks to the association belongsToMany
This data would be accessed on _joinData like regular entities when you choose to use contain() or maybe a new Query method like "containJoinData"
| 1.0 | Add support to retrieve data from associations of _joinData (a.k.a ternary association) - This is a (multiple allowed):
* [ ] bug
* [x] enhancement
* [x] feature-discussion (RFC)
* CakePHP Version: 3.7.
* Platform and Target: PHP 7.3 Mysql 5.
### What you did
I have created a simple database with those models:
_Doctors_ **belongsToMany** _Patients_ **through** _Appointments_
_Appointments_ **hasMany** _Examinations_
Full SQL: https://pastebin.com/9fY8Vibx
As you can see, I have a classic "ternary association".
So, I have tried to create a kind-of report showing _all doctors with theirs appointments date including patient name and a list of requested exams_
**Controller**:
```php
$table = TableRegistry::getTableLocator()->get('Doctors');
$data = $table->find()
->contain('Patients')
->contain('Appointments.Examinations'); //Error !
$this->set(compact('data'));
```
**Template:**
```php
<?php
echo "<table>";
echo $this->Html->tableHeaders(['Doctor','Patient','Appointment Date', 'Exams']);
foreach ($data as $doctor){
echo "<tr>";
echo "<td rowspan='".count($doctor->patients)."'>".$doctor->name."</td>";
foreach ($doctor->patients as $key => $patient) {
if($key == !array_key_first($doctor->patients)){
echo "<tr>";
}
echo "<td>".$patient->name."</td>";
echo "<td>".$patient->_joinData->date."</td>";
//$exams = array_column($patient->_joinData->examinations, 'name'); //Error !
//echo "<td>".implode('; ', $exams); //Error !
echo "</tr>";
}
}
```
### What happened
The "pivot" table it is not associated with doctors so it is no possible to reach by the "normal" way the associated data of this _joinData
### What you expected to happen
Be able to get Associated data from a model defined using "setThrough".
Some changes could be applied on **findAssociation()** to "infere" (like a OWL Reasoner) that Doctors (or Patients) has a association with Appointments thanks to the association belongsToMany
This data would be accessed on _joinData like regular entities when you choose to use contain() or maybe a new Query method like "containJoinData"
| defect | add support to retrieve data from associations of joindata a k a ternary association this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target php mysql what you did i have created a simple database with those models doctors belongstomany patients through appointments appointments hasmany examinations full sql as you can see i have a classic ternary association so i have tried to create a kind of report showing all doctors with theirs appointments date including patient name and a list of requested exams controller php table tableregistry gettablelocator get doctors data table find contain patients contain appointments examinations error this set compact data template php php echo echo this html tableheaders foreach data as doctor echo echo patients doctor name foreach doctor patients as key patient if key array key first doctor patients echo echo patient name echo patient joindata date exams array column patient joindata examinations name error echo implode exams error echo what happened the pivot table it is not associated with doctors so it is no possible to reach by the normal way the associated data of this joindata what you expected to happen be able to get associated data from a model defined using setthrough some changes could be applied on findassociation to infere like a owl reasoner that doctors or patients has a association with appointments thanks to the association belongstomany this data would be accessed on joindata like regular entities when you choose to use contain or maybe a new query method like containjoindata | 1 |
187,997 | 6,767,176,586 | IssuesEvent | 2017-10-26 01:36:40 | CS2103AUG2017-T13-B1/main | https://api.github.com/repos/CS2103AUG2017-T13-B1/main | closed | Implementing links from insurance profile directly to person profile page | Priority: Medium Status: Completed | Some sort of hyperlink, buttons, or alike to jump from insurance page to profile | 1.0 | Implementing links from insurance profile directly to person profile page - Some sort of hyperlink, buttons, or alike to jump from insurance page to profile | non_defect | implementing links from insurance profile directly to person profile page some sort of hyperlink buttons or alike to jump from insurance page to profile | 0 |
45,954 | 13,055,827,704 | IssuesEvent | 2020-07-30 02:51:18 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | BadDomList service does not close DB connection on exit (Trac #319) | Incomplete Migration Migrated from Trac combo reconstruction defect | Migrated from https://code.icecube.wisc.edu/ticket/319
```json
{
"status": "closed",
"changetime": "2012-03-18T17:52:31",
"description": "The DB version of BadDomList service does not appear to properly\nclose the DB handle when it exits. This results in high numbers of what\nappear as \"dropped\" DB connections.\n\nThe DB interface should be updated to properly close all open DB connections. ",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1332093151000000",
"component": "combo reconstruction",
"summary": "BadDomList service does not close DB connection on exit",
"priority": "normal",
"keywords": "",
"time": "2011-11-03T14:23:08",
"milestone": "",
"owner": "jacobi",
"type": "defect"
}
```
| 1.0 | BadDomList service does not close DB connection on exit (Trac #319) - Migrated from https://code.icecube.wisc.edu/ticket/319
```json
{
"status": "closed",
"changetime": "2012-03-18T17:52:31",
"description": "The DB version of BadDomList service does not appear to properly\nclose the DB handle when it exits. This results in high numbers of what\nappear as \"dropped\" DB connections.\n\nThe DB interface should be updated to properly close all open DB connections. ",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1332093151000000",
"component": "combo reconstruction",
"summary": "BadDomList service does not close DB connection on exit",
"priority": "normal",
"keywords": "",
"time": "2011-11-03T14:23:08",
"milestone": "",
"owner": "jacobi",
"type": "defect"
}
```
| defect | baddomlist service does not close db connection on exit trac migrated from json status closed changetime description the db version of baddomlist service does not appear to properly nclose the db handle when it exits this results in high numbers of what nappear as dropped db connections n nthe db interface should be updated to properly close all open db connections reporter blaufuss cc resolution fixed ts component combo reconstruction summary baddomlist service does not close db connection on exit priority normal keywords time milestone owner jacobi type defect | 1 |
44,095 | 11,965,672,714 | IssuesEvent | 2020-04-06 00:22:27 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | csgraph.minimum_spanning_tree loses precision | Documentation defect scipy.sparse.csgraph | csgraph.minimum_spanning_tree seems to lose double precision somewhere along the way.
This is surprising since all the input and output is formally float64 but still small elements < 1E-8 are rounded to zero.
#### Reproducing code example:
```
>>> import numpy as np
>>> import scipy.sparse
>>> M=np.zeros((3,3), dtype=np.float64)
>>> scipy.sparse.csgraph.minimum_spanning_tree(M + 1E-8).toarray()
array([[0., 0., 0.],
[0., 0., 0.],
[0., 0., 0.]])
>>> scipy.sparse.csgraph.minimum_spanning_tree(M + 1E-7).toarray()
array([[0.e+00, 1.e-07, 1.e-07],
[0.e+00, 0.e+00, 0.e+00],
[0.e+00, 0.e+00, 0.e+00]])
>>> scipy.sparse.csgraph.minimum_spanning_tree(M + 1E-8).dtype
dtype('float64')
```
#### Scipy/Numpy/Python version information:
```
>>> print(scipy.__version__, np.__version__, sys.version_info)
1.3.1 1.17.2 sys.version_info(major=3, minor=7, micro=4, releaselevel='final', serial=0)
``` | 1.0 | csgraph.minimum_spanning_tree loses precision - csgraph.minimum_spanning_tree seems to lose double precision somewhere along the way.
This is surprising since all the input and output is formally float64 but still small elements < 1E-8 are rounded to zero.
#### Reproducing code example:
```
>>> import numpy as np
>>> import scipy.sparse
>>> M=np.zeros((3,3), dtype=np.float64)
>>> scipy.sparse.csgraph.minimum_spanning_tree(M + 1E-8).toarray()
array([[0., 0., 0.],
[0., 0., 0.],
[0., 0., 0.]])
>>> scipy.sparse.csgraph.minimum_spanning_tree(M + 1E-7).toarray()
array([[0.e+00, 1.e-07, 1.e-07],
[0.e+00, 0.e+00, 0.e+00],
[0.e+00, 0.e+00, 0.e+00]])
>>> scipy.sparse.csgraph.minimum_spanning_tree(M + 1E-8).dtype
dtype('float64')
```
#### Scipy/Numpy/Python version information:
```
>>> print(scipy.__version__, np.__version__, sys.version_info)
1.3.1 1.17.2 sys.version_info(major=3, minor=7, micro=4, releaselevel='final', serial=0)
``` | defect | csgraph minimum spanning tree loses precision csgraph minimum spanning tree seems to lose double precision somewhere along the way this is surprising since all the input and output is formally but still small elements are rounded to zero reproducing code example import numpy as np import scipy sparse m np zeros dtype np scipy sparse csgraph minimum spanning tree m toarray array scipy sparse csgraph minimum spanning tree m toarray array scipy sparse csgraph minimum spanning tree m dtype dtype scipy numpy python version information print scipy version np version sys version info sys version info major minor micro releaselevel final serial | 1 |
17,540 | 3,012,746,958 | IssuesEvent | 2015-07-29 02:09:08 | yawlfoundation/yawl | https://api.github.com/repos/yawlfoundation/yawl | closed | [CLOSED] Can't change the initial value of a net variable | auto-migrated Category-Editor Priority-Critical Type-Defect | <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday Jul 27, 2015 at 03:21 GMT_
_Originally opened as https://github.com/adamsmj/yawl/issues/81_
----
```
In the net of issue 78 there is a local net variable that has received the
initial value "hello" (including the quotes). This was accepted by the
editor but it isn't possible to change this anymore.
```
Original issue reported on code.google.com by `arthurte...@gmail.com` on 8 Aug 2008 at 2:13
| 1.0 | [CLOSED] Can't change the initial value of a net variable - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday Jul 27, 2015 at 03:21 GMT_
_Originally opened as https://github.com/adamsmj/yawl/issues/81_
----
```
In the net of issue 78 there is a local net variable that has received the
initial value "hello" (including the quotes). This was accepted by the
editor but it isn't possible to change this anymore.
```
Original issue reported on code.google.com by `arthurte...@gmail.com` on 8 Aug 2008 at 2:13
| defect | can t change the initial value of a net variable issue by monday jul at gmt originally opened as in the net of issue there is a local net variable that has received the initial value hello including the quotes this was accepted by the editor but it isn t possible to change this anymore original issue reported on code google com by arthurte gmail com on aug at | 1 |
368,740 | 25,807,074,113 | IssuesEvent | 2022-12-11 14:04:29 | getsentry/sentry-capacitor | https://api.github.com/repos/getsentry/sentry-capacitor | closed | Make clear in installation docs that the sibling SDK version has to match. | documentation Platform: Capacitor update-sentry-js-sdk | We will need to document in the installation steps to ensure the version of the sibling SDK matches the version of the sentry-javascript dependencies that @sentry/capacitor depends on. | 1.0 | Make clear in installation docs that the sibling SDK version has to match. - We will need to document in the installation steps to ensure the version of the sibling SDK matches the version of the sentry-javascript dependencies that @sentry/capacitor depends on. | non_defect | make clear in installation docs that the sibling sdk version has to match we will need to document in the installation steps to ensure the version of the sibling sdk matches the version of the sentry javascript dependencies that sentry capacitor depends on | 0 |
8,356 | 3,163,929,487 | IssuesEvent | 2015-09-20 18:58:03 | acemod/language-arma-atom | https://api.github.com/repos/acemod/language-arma-atom | opened | Update README.md screenshots and information for autocomplete-plus | documentation | - [ ] Screenshots with autocomplete-plus
- [ ] Build Dev and Build Release functions shortly presented | 1.0 | Update README.md screenshots and information for autocomplete-plus - - [ ] Screenshots with autocomplete-plus
- [ ] Build Dev and Build Release functions shortly presented | non_defect | update readme md screenshots and information for autocomplete plus screenshots with autocomplete plus build dev and build release functions shortly presented | 0 |
749,764 | 26,179,015,117 | IssuesEvent | 2023-01-02 13:21:47 | conan-io/conan | https://api.github.com/repos/conan-io/conan | closed | [feature] Virtualenvs should create a sandbox | type: feature stage: queue priority: medium complex: low | Taken from this comment: https://github.com/conan-io/conan/pull/5989#pullrequestreview-326075695
For Conan v2.0 we want the virtualenvs to restore the environment to the state it was before activating it (same variables, same values). So, anything modified inside the virtualenv will be discarded after deactivating it.
| 1.0 | [feature] Virtualenvs should create a sandbox - Taken from this comment: https://github.com/conan-io/conan/pull/5989#pullrequestreview-326075695
For Conan v2.0 we want the virtualenvs to restore the environment to the state it was before activating it (same variables, same values). So, anything modified inside the virtualenv will be discarded after deactivating it.
| non_defect | virtualenvs should create a sandbox taken from this comment for conan we want the virtualenvs to restore the environment to the state it was before activating it same variables same values so anything modified inside the virtualenv will be discarded after deactivating it | 0 |
66,309 | 20,147,310,062 | IssuesEvent | 2022-02-09 08:57:55 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | Record::formatJSON does not respect JSONFormat::wrapSingleColumnRecords | T: Defect C: Functionality P: Low E: All Editions | Calling `Record::formatJSON` with `jsonFormat.wrapSingleColumnRecords(false)` doesn't have any effect. It is only used when formatting a `Result`. | 1.0 | Record::formatJSON does not respect JSONFormat::wrapSingleColumnRecords - Calling `Record::formatJSON` with `jsonFormat.wrapSingleColumnRecords(false)` doesn't have any effect. It is only used when formatting a `Result`. | defect | record formatjson does not respect jsonformat wrapsinglecolumnrecords calling record formatjson with jsonformat wrapsinglecolumnrecords false doesn t have any effect it is only used when formatting a result | 1 |
14,067 | 24,346,594,090 | IssuesEvent | 2022-10-02 11:57:04 | PharmaLedger-IMI/epi-workspace | https://api.github.com/repos/PharmaLedger-IMI/epi-workspace | closed | Scalable Solution | user story Non-functional requirement | The productive pilot solution should be able to support at minimum -
Organizations - 5/6 companies
Products - 80-100 (10-15 per company)
Batches - 100-200/product
Leaflets - 80-100 for 3 langs (Dutch, German, French, English)
Serial nos - 800K - 1M
| 1.0 | Scalable Solution - The productive pilot solution should be able to support at minimum -
Organizations - 5/6 companies
Products - 80-100 (10-15 per company)
Batches - 100-200/product
Leaflets - 80-100 for 3 langs (Dutch, German, French, English)
Serial nos - 800K - 1M
| non_defect | scalable solution the productive pilot solution should be able to support at minimum organizations companies products per company batches product leaflets for langs dutch german french english serial nos | 0 |
61,548 | 17,023,722,801 | IssuesEvent | 2021-07-03 03:29:47 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | [roads] Secondary bridges rendered way larger than primary bridges in mapnik on z15-13 | Component: mapnik Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 2.34am, Saturday, 18th June 2011]**
http://www.openstreetmap.org/?lat=36.107&lon=-115.1832&zoom=14&layers=M
Just noticed that secondary bridges are being rendered on z15-z13 a lot bigger than primary bridges that are nearby, which shouldn't be the case.
So, I went into the osm.xml file to see if I could spot where the problem was and I 'think' I found it.
The problem (I think) lies in the "<Style name="bridges_fill">" section. Now, I don't fully know all that much on how the style sheet works, so, I will not go suggesting exactly what to fix because for all I know, what I suggest could make it worse.
But this secondary bridge is rendered bigger than the twinned SR 592 right above it which should be wider since there are two ways there that are combined when rendered at z15-z13. So, any rendering fix would be appreciated for this. | 1.0 | [roads] Secondary bridges rendered way larger than primary bridges in mapnik on z15-13 - **[Submitted to the original trac issue database at 2.34am, Saturday, 18th June 2011]**
http://www.openstreetmap.org/?lat=36.107&lon=-115.1832&zoom=14&layers=M
Just noticed that secondary bridges are being rendered on z15-z13 a lot bigger than primary bridges that are nearby, which shouldn't be the case.
So, I went into the osm.xml file to see if I could spot where the problem was and I 'think' I found it.
The problem (I think) lies in the "<Style name="bridges_fill">" section. Now, I don't fully know all that much on how the style sheet works, so, I will not go suggesting exactly what to fix because for all I know, what I suggest could make it worse.
But this secondary bridge is rendered bigger than the twinned SR 592 right above it which should be wider since there are two ways there that are combined when rendered at z15-z13. So, any rendering fix would be appreciated for this. | defect | secondary bridges rendered way larger than primary bridges in mapnik on just noticed that secondary bridges are being rendered on a lot bigger than primary bridges that are nearby which shouldn t be the case so i went into the osm xml file to see if i could spot where the problem was and i think i found it the problem i think lies in the section now i don t fully know all that much on how the style sheet works so i will not go suggesting exactly what to fix because for all i know what i suggest could make it worse but this secondary bridge is rendered bigger than the twinned sr right above it which should be wider since there are two ways there that are combined when rendered at so any rendering fix would be appreciated for this | 1 |
146,839 | 23,130,784,246 | IssuesEvent | 2022-07-28 10:09:47 | flix/flix | https://api.github.com/repos/flix/flix | closed | Interoperability: Add an expression that allows the (anon.) implementation of a Java interface/class extension | language-design | # Overall Goal: Add new anonymous class/object expression
For example:
```
object ##java.lang.Iterator {
def hasNext(): Bool = true
def next(): String = "Hello"
}
```
which will generate a new class called say `Anon$1` which implements the `Iterator` interface and has two methods `hasNext` and `next`. The `object` expression is a Java value of type `##java.lang.Iterator`. (We ignore generics for now-- they are erased anyway.)
I propose to work towards this goal through several milestones:
# Milestone 1
- [ ] Add a new ` NewObject` expression (right after PutStaticField) that takes a list of `1` class and `n` interfaces.
For example, we should be able to compile and run these expressions (trivial as they may be):
Ex1:
```
object ##java.lang.Object { }
```
Ex2:
```
object ##java.lang.Serializable { }
```
Ex3:
```
object ##java.lang.Object, ##java.lang.Serializable, ##java.util.RandomAccess { }
```
# Milestone 2
TBD
| 1.0 | Interoperability: Add an expression that allows the (anon.) implementation of a Java interface/class extension - # Overall Goal: Add new anonymous class/object expression
For example:
```
object ##java.lang.Iterator {
def hasNext(): Bool = true
def next(): String = "Hello"
}
```
which will generate a new class called say `Anon$1` which implements the `Iterator` interface and has two methods `hasNext` and `next`. The `object` expression is a Java value of type `##java.lang.Iterator`. (We ignore generics for now-- they are erased anyway.)
I propose to work towards this goal through several milestones:
# Milestone 1
- [ ] Add a new ` NewObject` expression (right after PutStaticField) that takes a list of `1` class and `n` interfaces.
For example, we should be able to compile and run these expressions (trivial as they may be):
Ex1:
```
object ##java.lang.Object { }
```
Ex2:
```
object ##java.lang.Serializable { }
```
Ex3:
```
object ##java.lang.Object, ##java.lang.Serializable, ##java.util.RandomAccess { }
```
# Milestone 2
TBD
| non_defect | interoperability add an expression that allows the anon implementation of a java interface class extension overall goal add new anonymous class object expression for example object java lang iterator def hasnext bool true def next string hello which will generate a new class called say anon which implements the iterator interface and has two methods hasnext and next the object expression is a java value of type java lang iterator we ignore generics for now they are erased anyway i propose to work towards this goal through several milestones milestone add a new newobject expression right after putstaticfield that takes a list of class and n interfaces for example we should be able to compile and run these expressions trivial as they may be object java lang object object java lang serializable object java lang object java lang serializable java util randomaccess milestone tbd | 0 |
3,750 | 2,610,068,166 | IssuesEvent | 2015-02-26 18:20:01 | chrsmith/jsjsj122 | https://api.github.com/repos/chrsmith/jsjsj122 | opened | 路桥看前列腺炎哪里权威 | auto-migrated Priority-Medium Type-Defect | ```
路桥看前列腺炎哪里权威【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 9:00 | 1.0 | 路桥看前列腺炎哪里权威 - ```
路桥看前列腺炎哪里权威【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 9:00 | defect | 路桥看前列腺炎哪里权威 路桥看前列腺炎哪里权威【台州五洲生殖医院】 询热线 微信号tzwzszyy 医院地址 台州 (枫南大转盘旁)乘车线路 、 � �� 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at | 1 |
323,120 | 27,696,559,324 | IssuesEvent | 2023-03-14 03:00:18 | pombase/pombase-chado | https://api.github.com/repos/pombase/pombase-chado | closed | missing data | bug needs testing |
Some annotations that were there last week have disappeared form the query builder.
No results for: genes annotated with "LYR family" ([PBO:0091211](https://www.pombase.org/term/PBO:0091211))
There were 8 annotations.
They are annotated in the contig files, and usually appear in the protein feature section of the page. | 1.0 | missing data -
Some annotations that were there last week have disappeared form the query builder.
No results for: genes annotated with "LYR family" ([PBO:0091211](https://www.pombase.org/term/PBO:0091211))
There were 8 annotations.
They are annotated in the contig files, and usually appear in the protein feature section of the page. | non_defect | missing data some annotations that were there last week have disappeared form the query builder no results for genes annotated with lyr family there were annotations they are annotated in the contig files and usually appear in the protein feature section of the page | 0 |
66,169 | 20,025,853,408 | IssuesEvent | 2022-02-01 21:13:17 | idaholab/moose | https://api.github.com/repos/idaholab/moose | closed | Allow PiecewiseConstant functions to control Timestepper | T: defect P: normal | ## Bug Description
Somewhere along the way, support for `PiecewiseConstant` functions to be able to limit the `IterationAdaptiveDT` timestepper dt via `timestep_limiting_function` was lost...or was never there and never tested.
## Steps to Reproduce
`moose_test-opt -i test/tests/time_steppers/piecewise_linear.i Functions/temp_spike/type=PiecewiseConstant`
## Impact
Expected functionality will return.
| 1.0 | Allow PiecewiseConstant functions to control Timestepper - ## Bug Description
Somewhere along the way, support for `PiecewiseConstant` functions to be able to limit the `IterationAdaptiveDT` timestepper dt via `timestep_limiting_function` was lost...or was never there and never tested.
## Steps to Reproduce
`moose_test-opt -i test/tests/time_steppers/piecewise_linear.i Functions/temp_spike/type=PiecewiseConstant`
## Impact
Expected functionality will return.
| defect | allow piecewiseconstant functions to control timestepper bug description somewhere along the way support for piecewiseconstant functions to be able to limit the iterationadaptivedt timestepper dt via timestep limiting function was lost or was never there and never tested steps to reproduce moose test opt i test tests time steppers piecewise linear i functions temp spike type piecewiseconstant impact expected functionality will return | 1 |
225,462 | 7,481,906,153 | IssuesEvent | 2018-04-04 22:24:57 | Planteome/plant-experimental-conditions-ontology | https://api.github.com/repos/Planteome/plant-experimental-conditions-ontology | closed | antibiotic treatment | Plant Treatment high priority sourceforge | current name and definition:
antibiotic regimen ([EO:0007041](http://purl.obolibrary.org/obo/EO_0007041)): Treatment involving the use of antibiotic for selection purposes.
Proposed new name and definition:
antibiotic treatment: A chemical treatment ([EO:0007189](http://purl.obolibrary.org/obo/EO_0007189)) involving the use of antibiotic for selection purposes.
Reported by: lauramoore1859
Original Ticket: [obo/plant-environment-ontology-eo/70](https://sourceforge.net/p/obo/plant-environment-ontology-eo/70)
| 1.0 | antibiotic treatment - current name and definition:
antibiotic regimen ([EO:0007041](http://purl.obolibrary.org/obo/EO_0007041)): Treatment involving the use of antibiotic for selection purposes.
Proposed new name and definition:
antibiotic treatment: A chemical treatment ([EO:0007189](http://purl.obolibrary.org/obo/EO_0007189)) involving the use of antibiotic for selection purposes.
Reported by: lauramoore1859
Original Ticket: [obo/plant-environment-ontology-eo/70](https://sourceforge.net/p/obo/plant-environment-ontology-eo/70)
| non_defect | antibiotic treatment current name and definition antibiotic regimen treatment involving the use of antibiotic for selection purposes proposed new name and definition antibiotic treatment a chemical treatment involving the use of antibiotic for selection purposes reported by original ticket | 0 |
60,150 | 12,059,669,335 | IssuesEvent | 2020-04-15 19:42:25 | ikostan/codewars | https://api.github.com/repos/ikostan/codewars | opened | Fix "method_complexity" issue in kyu_6/help_the_bookseller/stock_list.py | codeclimate codewars issues | Function `stock_list` has a Cognitive Complexity of 10 (exceeds 5 allowed). Consider refactoring.
https://codeclimate.com/github/ikostan/codewars/kyu_6/help_the_bookseller/stock_list.py#issue_5e96c3368a97c800010001bb | 2.0 | Fix "method_complexity" issue in kyu_6/help_the_bookseller/stock_list.py - Function `stock_list` has a Cognitive Complexity of 10 (exceeds 5 allowed). Consider refactoring.
https://codeclimate.com/github/ikostan/codewars/kyu_6/help_the_bookseller/stock_list.py#issue_5e96c3368a97c800010001bb | non_defect | fix method complexity issue in kyu help the bookseller stock list py function stock list has a cognitive complexity of exceeds allowed consider refactoring | 0 |
754,324 | 26,381,856,886 | IssuesEvent | 2023-01-12 09:17:43 | sourceplusplus/sourceplusplus | https://api.github.com/repos/sourceplusplus/sourceplusplus | opened | Probe bridge TCP timeouts | a:bug in:live-platform of:small-effort with:moderate-priority | Probes connected via TCP don't send ping/pong messages and never timeout and therefore are not properly removed from active probe stats. A client ping/pong thread (every 5 seconds) and server timeout (10 seconds) need to be added to match SockJS impl. | 1.0 | Probe bridge TCP timeouts - Probes connected via TCP don't send ping/pong messages and never timeout and therefore are not properly removed from active probe stats. A client ping/pong thread (every 5 seconds) and server timeout (10 seconds) need to be added to match SockJS impl. | non_defect | probe bridge tcp timeouts probes connected via tcp don t send ping pong messages and never timeout and therefore are not properly removed from active probe stats a client ping pong thread every seconds and server timeout seconds need to be added to match sockjs impl | 0 |
23,082 | 4,869,026,199 | IssuesEvent | 2016-11-15 11:17:38 | ChrisCummins/clgen | https://api.github.com/repos/ChrisCummins/clgen | closed | Use remote model and sampler in example | Documentation Feature Packaging | Something like:
```sh
export model=https://raw.githubusercontent.com/ChrisCummins/clgen/blob/master/tests/data/tiny/corpus.tar.bz2
export sampler=https://raw.githubusercontent.com/ChrisCummins/clgen/blob/master/tests/data/tiny/sampler.json
$ clgen $model $sampler
``` | 1.0 | Use remote model and sampler in example - Something like:
```sh
export model=https://raw.githubusercontent.com/ChrisCummins/clgen/blob/master/tests/data/tiny/corpus.tar.bz2
export sampler=https://raw.githubusercontent.com/ChrisCummins/clgen/blob/master/tests/data/tiny/sampler.json
$ clgen $model $sampler
``` | non_defect | use remote model and sampler in example something like sh export model export sampler clgen model sampler | 0 |
8,191 | 2,965,596,803 | IssuesEvent | 2015-07-11 00:54:22 | gheber/kenzo | https://api.github.com/repos/gheber/kenzo | closed | BAR test fails | bug test | The BAR test in bar.lisp fails with:
Failure Details:
--------------------------------
BAR []:
Unexpected Error: #<TYPE-ERROR expected-type: (SATISFIES CAT::ICMBN-P)
datum:
----------------------------------------------------------------------{CMBN 6}
<-1 * (-7 -1 1 2 4 -2)>
<-1 * (-7 -1 1 4 -2 2)>
<1 * (-7 -1 1 4 2 -2)>
<-1 * (-7 -1 4 -2 1 2)>
<1 * (-7 -1 4 1 -2 2)>
<-1 * (-7 -1 4 1 2 -2)>
<1 * (-7 1 -1 2 4 -2)>
<1 * (-7 1 -1 4 -2 2)>
<-1 * (-7 1 -1 4 2 -2)>
<-1 * (-7 1 2 -1 4 -2)>
... ...
------------------------------------------------------------------------------
>
The value
----------------------------------------------------------------------{CMBN 6}
<-1 * (-7 -1 1 2 4 -2)>
<-1 * (-7 -1 1 4 -2 2)>
<1 * (-7 -1 1 4 2 -2)>
<-1 * (-7 -1 4 -2 1 2)>
<1 * (-7 -1 4 1 -2 2)>
<-1 * (-7 -1 4 1 2 -2)>
<1 * (-7 1 -1 2 4 -2)>
<1 * (-7 1 -1 4 -2 2)>
<-1 * (-7 1 -1 4 2 -2)>
<-1 * (-7 1 2 -1 4 -2)>
... ...
------------------------------------------------------------------------------
is not of type
(SATISFIES CAT::ICMBN-P)...
--------------------------------
This is probably the same issue as #24 and #25. | 1.0 | BAR test fails - The BAR test in bar.lisp fails with:
Failure Details:
--------------------------------
BAR []:
Unexpected Error: #<TYPE-ERROR expected-type: (SATISFIES CAT::ICMBN-P)
datum:
----------------------------------------------------------------------{CMBN 6}
<-1 * (-7 -1 1 2 4 -2)>
<-1 * (-7 -1 1 4 -2 2)>
<1 * (-7 -1 1 4 2 -2)>
<-1 * (-7 -1 4 -2 1 2)>
<1 * (-7 -1 4 1 -2 2)>
<-1 * (-7 -1 4 1 2 -2)>
<1 * (-7 1 -1 2 4 -2)>
<1 * (-7 1 -1 4 -2 2)>
<-1 * (-7 1 -1 4 2 -2)>
<-1 * (-7 1 2 -1 4 -2)>
... ...
------------------------------------------------------------------------------
>
The value
----------------------------------------------------------------------{CMBN 6}
<-1 * (-7 -1 1 2 4 -2)>
<-1 * (-7 -1 1 4 -2 2)>
<1 * (-7 -1 1 4 2 -2)>
<-1 * (-7 -1 4 -2 1 2)>
<1 * (-7 -1 4 1 -2 2)>
<-1 * (-7 -1 4 1 2 -2)>
<1 * (-7 1 -1 2 4 -2)>
<1 * (-7 1 -1 4 -2 2)>
<-1 * (-7 1 -1 4 2 -2)>
<-1 * (-7 1 2 -1 4 -2)>
... ...
------------------------------------------------------------------------------
is not of type
(SATISFIES CAT::ICMBN-P)...
--------------------------------
This is probably the same issue as #24 and #25. | non_defect | bar test fails the bar test in bar lisp fails with failure details bar unexpected error type error expected type satisfies cat icmbn p datum cmbn the value cmbn is not of type satisfies cat icmbn p this is probably the same issue as and | 0 |
252,672 | 21,610,232,825 | IssuesEvent | 2022-05-04 09:20:34 | opensrp/opensrp-client-anc | https://api.github.com/repos/opensrp/opensrp-client-anc | closed | [Ona Support Request]: App crashing when one tries to start a contact | high priority Tech Partner (SID Team) client testing | ### Affected App or Server Version
v1.6.8
### What kind of support do you need?
SID local app crashing when one tries to start a contact
### What is the acceptance criteria for your support request?
Able to start and complete a contact
### Relevant Information
_No response_ | 1.0 | [Ona Support Request]: App crashing when one tries to start a contact - ### Affected App or Server Version
v1.6.8
### What kind of support do you need?
SID local app crashing when one tries to start a contact
### What is the acceptance criteria for your support request?
Able to start and complete a contact
### Relevant Information
_No response_ | non_defect | app crashing when one tries to start a contact affected app or server version what kind of support do you need sid local app crashing when one tries to start a contact what is the acceptance criteria for your support request able to start and complete a contact relevant information no response | 0 |
18,131 | 10,010,930,434 | IssuesEvent | 2019-07-15 09:17:56 | hlissner/doom-emacs | https://api.github.com/repos/hlissner/doom-emacs | closed | Debugging emacs problems | is:question module:lang:cc re:performance status:unknown | I have been noticing that while editing my C++ files, the speed degrades after a few days of usage (I rarely restart my emacs). I suspect that there's some packages that are being loaded which is slowing down my emacs but it is difficult to debug this since I don't know when these packages get loaded.
Is there a way to force doom to load all packages at startup. I know this will slow the startup time significantly but I would like to be able to troubleshoot these speed problems. | True | Debugging emacs problems - I have been noticing that while editing my C++ files, the speed degrades after a few days of usage (I rarely restart my emacs). I suspect that there's some packages that are being loaded which is slowing down my emacs but it is difficult to debug this since I don't know when these packages get loaded.
Is there a way to force doom to load all packages at startup. I know this will slow the startup time significantly but I would like to be able to troubleshoot these speed problems. | non_defect | debugging emacs problems i have been noticing that while editing my c files the speed degrades after a few days of usage i rarely restart my emacs i suspect that there s some packages that are being loaded which is slowing down my emacs but it is difficult to debug this since i don t know when these packages get loaded is there a way to force doom to load all packages at startup i know this will slow the startup time significantly but i would like to be able to troubleshoot these speed problems | 0 |
63,262 | 17,499,849,150 | IssuesEvent | 2021-08-10 08:04:46 | milvus-io/milvus-insight | https://api.github.com/repos/milvus-io/milvus-insight | opened | Server log undefined error | defect | **Describe the bug:**

**Steps to reproduce:**
1. yarn start:dev in server folder
**Milvus-insight version:**
latest
**Milvus version:**
| 1.0 | Server log undefined error - **Describe the bug:**

**Steps to reproduce:**
1. yarn start:dev in server folder
**Milvus-insight version:**
latest
**Milvus version:**
| defect | server log undefined error describe the bug steps to reproduce yarn start dev in server folder milvus insight version latest milvus version | 1 |
4,949 | 2,610,162,147 | IssuesEvent | 2015-02-26 18:51:26 | chrsmith/republic-at-war | https://api.github.com/repos/chrsmith/republic-at-war | closed | Text | auto-migrated Priority-Medium Type-Defect | ```
Weapons upgrade text name missing
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 22 Feb 2011 at 2:58 | 1.0 | Text - ```
Weapons upgrade text name missing
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 22 Feb 2011 at 2:58 | defect | text weapons upgrade text name missing original issue reported on code google com by gmail com on feb at | 1 |
281,814 | 24,421,793,313 | IssuesEvent | 2022-10-05 21:03:12 | NCAR/DART | https://api.github.com/repos/NCAR/DART | closed | obs_sequence_tool developer tests | Test | Currently there are no tests to run to check changes to obs_sequence_tool (there is the start of some testing for reading little and big endian obs sequences)
These tests should:
* go in developer_tests/obs_sequence
* be run as part of test_dart
* be able to run outside of test_dart for people developing the obs_sequence_tool
assigned to @hkershaw-brown | 1.0 | obs_sequence_tool developer tests - Currently there are no tests to run to check changes to obs_sequence_tool (there is the start of some testing for reading little and big endian obs sequences)
These tests should:
* go in developer_tests/obs_sequence
* be run as part of test_dart
* be able to run outside of test_dart for people developing the obs_sequence_tool
assigned to @hkershaw-brown | non_defect | obs sequence tool developer tests currently there are no tests to run to check changes to obs sequence tool there is the start of some testing for reading little and big endian obs sequences these tests should go in developer tests obs sequence be run as part of test dart be able to run outside of test dart for people developing the obs sequence tool assigned to hkershaw brown | 0 |
17,028 | 2,966,760,451 | IssuesEvent | 2015-07-12 07:15:15 | ElvisChiang/welly | https://api.github.com/repos/ElvisChiang/welly | closed | Maximize welly window size | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Mistakenly press unknown keys, the welly window was maximize, but never get
back to default size. This is really annoying.
2.
3.
What is the expected output? What do you see instead?
I expected user could reset the window's size. The maximum size is annoying.
What version of the product are you using? On what operating system?
OS: Lion (10.7)
Welly 2.64
Please provide any additional information below.
If welly disables welly window resizing, it should also disable the system
hotkey/shortcut key for window maximize on welly. Otherwise, please enable
window resize.
Attached a screenshot for reference. Thanks.
```
Original issue reported on code.google.com by `kyle.w....@gmail.com` on 1 Jun 2012 at 4:01
Attachments:
* [Welly_On_Maximum_Size.png](https://storage.googleapis.com/google-code-attachments/welly/issue-65/comment-0/Welly_On_Maximum_Size.png)
| 1.0 | Maximize welly window size - ```
What steps will reproduce the problem?
1. Mistakenly press unknown keys, the welly window was maximize, but never get
back to default size. This is really annoying.
2.
3.
What is the expected output? What do you see instead?
I expected user could reset the window's size. The maximum size is annoying.
What version of the product are you using? On what operating system?
OS: Lion (10.7)
Welly 2.64
Please provide any additional information below.
If welly disables welly window resizing, it should also disable the system
hotkey/shortcut key for window maximize on welly. Otherwise, please enable
window resize.
Attached a screenshot for reference. Thanks.
```
Original issue reported on code.google.com by `kyle.w....@gmail.com` on 1 Jun 2012 at 4:01
Attachments:
* [Welly_On_Maximum_Size.png](https://storage.googleapis.com/google-code-attachments/welly/issue-65/comment-0/Welly_On_Maximum_Size.png)
| defect | maximize welly window size what steps will reproduce the problem mistakenly press unknown keys the welly window was maximize but never get back to default size this is really annoying what is the expected output what do you see instead i expected user could reset the window s size the maximum size is annoying what version of the product are you using on what operating system os lion welly please provide any additional information below if welly disables welly window resizing it should also disable the system hotkey shortcut key for window maximize on welly otherwise please enable window resize attached a screenshot for reference thanks original issue reported on code google com by kyle w gmail com on jun at attachments | 1 |
16,759 | 2,941,839,420 | IssuesEvent | 2015-07-02 10:35:18 | google/google-api-dotnet-client | https://api.github.com/repos/google/google-api-dotnet-client | closed | .NET C# Blogger v3 API Bug, Service.Blogs.ListByUser | auto-migrated Component-Api Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Instantiate Blogger Service
2. Call ListByUser using an actual userId from gplus instead of "self"
BlogsResource.ListByUserRequest blogListByUserAction =
service.Blogs.ListByUser("105499867526169935261");
3. Run Fetch
BlogList list = blogListByUserAction.Fetch();
What is the expected output? What do you see instead?
I should get a list of my blogs; however, I receive the following exception
(works fine w/ "self"):
"Google.GoogleApiRequestException was unhandled by user code
Message=Google.Apis.Requests.RequestError
[500]
No individual errors"
What version of the product are you using? On what operating system?
.net c# blogger api v3 on windows 7 64 bit.
Please provide any additional information below.
This also is reproduceable in the sandbox using the userId:
https://developers.google.com/blogger/docs/3.0/reference/blogs/listByUser
500 Internal Server Error
- Hide headers -
cache-control: private, max-age=0
content-encoding: gzip
content-length: 66
content-type: application/json; charset=UTF-8
date: Wed, 22 May 2013 21:34:15 GMT
expires: Wed, 22 May 2013 21:34:15 GMT
server: GSE
{
"error": {
"code": 500,
"message": null
}
}
```
Original issue reported on code.google.com by `cbarthol...@gmail.com` on 22 May 2013 at 9:35 | 1.0 | .NET C# Blogger v3 API Bug, Service.Blogs.ListByUser - ```
What steps will reproduce the problem?
1. Instantiate Blogger Service
2. Call ListByUser using an actual userId from gplus instead of "self"
BlogsResource.ListByUserRequest blogListByUserAction =
service.Blogs.ListByUser("105499867526169935261");
3. Run Fetch
BlogList list = blogListByUserAction.Fetch();
What is the expected output? What do you see instead?
I should get a list of my blogs; however, I receive the following exception
(works fine w/ "self"):
"Google.GoogleApiRequestException was unhandled by user code
Message=Google.Apis.Requests.RequestError
[500]
No individual errors"
What version of the product are you using? On what operating system?
.net c# blogger api v3 on windows 7 64 bit.
Please provide any additional information below.
This also is reproduceable in the sandbox using the userId:
https://developers.google.com/blogger/docs/3.0/reference/blogs/listByUser
500 Internal Server Error
- Hide headers -
cache-control: private, max-age=0
content-encoding: gzip
content-length: 66
content-type: application/json; charset=UTF-8
date: Wed, 22 May 2013 21:34:15 GMT
expires: Wed, 22 May 2013 21:34:15 GMT
server: GSE
{
"error": {
"code": 500,
"message": null
}
}
```
Original issue reported on code.google.com by `cbarthol...@gmail.com` on 22 May 2013 at 9:35 | defect | net c blogger api bug service blogs listbyuser what steps will reproduce the problem instantiate blogger service call listbyuser using an actual userid from gplus instead of self blogsresource listbyuserrequest bloglistbyuseraction service blogs listbyuser run fetch bloglist list bloglistbyuseraction fetch what is the expected output what do you see instead i should get a list of my blogs however i receive the following exception works fine w self google googleapirequestexception was unhandled by user code message google apis requests requesterror no individual errors what version of the product are you using on what operating system net c blogger api on windows bit please provide any additional information below this also is reproduceable in the sandbox using the userid internal server error hide headers cache control private max age content encoding gzip content length content type application json charset utf date wed may gmt expires wed may gmt server gse error code message null original issue reported on code google com by cbarthol gmail com on may at | 1 |
45,992 | 13,055,834,284 | IssuesEvent | 2020-07-30 02:52:28 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | trigger config parameters are treated as int, not double (Trac #389) | Incomplete Migration Migrated from Trac dataclasses defect | Migrated from https://code.icecube.wisc.edu/ticket/389
```json
{
"status": "closed",
"changetime": "2014-11-22T18:26:26",
"description": "I3TriggerStatus treats trigger settings as integers. But the SLOP trigger has a setting (rel_v) which is set to 0.5. When creating a GCD this parameter is therefore not set correctly and the trigger simulation is not using the correct settings.\n\nIn the database the parameters are saved as string with the correct values.\nI3TriggerStatus should be able to treat a trigger parameter as double.\n\nI've changed this on a local working copy and it works for me, but as a consequence my GCD files are not compatible with an unchanged software.\nHere's what I've done: http://www.pastie.org/private/urikfrsoiwejvi3sg5lxa\n\nA support of double while keeping backwards compatibility is desirable.\n\nEmanuel\n",
"reporter": "jacobi",
"cc": "",
"resolution": "fixed",
"_ts": "1416680786877026",
"component": "dataclasses",
"summary": "trigger config parameters are treated as int, not double",
"priority": "normal",
"keywords": "",
"time": "2012-04-16T16:57:50",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
| 1.0 | trigger config parameters are treated as int, not double (Trac #389) - Migrated from https://code.icecube.wisc.edu/ticket/389
```json
{
"status": "closed",
"changetime": "2014-11-22T18:26:26",
"description": "I3TriggerStatus treats trigger settings as integers. But the SLOP trigger has a setting (rel_v) which is set to 0.5. When creating a GCD this parameter is therefore not set correctly and the trigger simulation is not using the correct settings.\n\nIn the database the parameters are saved as string with the correct values.\nI3TriggerStatus should be able to treat a trigger parameter as double.\n\nI've changed this on a local working copy and it works for me, but as a consequence my GCD files are not compatible with an unchanged software.\nHere's what I've done: http://www.pastie.org/private/urikfrsoiwejvi3sg5lxa\n\nA support of double while keeping backwards compatibility is desirable.\n\nEmanuel\n",
"reporter": "jacobi",
"cc": "",
"resolution": "fixed",
"_ts": "1416680786877026",
"component": "dataclasses",
"summary": "trigger config parameters are treated as int, not double",
"priority": "normal",
"keywords": "",
"time": "2012-04-16T16:57:50",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
| defect | trigger config parameters are treated as int not double trac migrated from json status closed changetime description treats trigger settings as integers but the slop trigger has a setting rel v which is set to when creating a gcd this parameter is therefore not set correctly and the trigger simulation is not using the correct settings n nin the database the parameters are saved as string with the correct values should be able to treat a trigger parameter as double n ni ve changed this on a local working copy and it works for me but as a consequence my gcd files are not compatible with an unchanged software nhere s what i ve done support of double while keeping backwards compatibility is desirable n nemanuel n reporter jacobi cc resolution fixed ts component dataclasses summary trigger config parameters are treated as int not double priority normal keywords time milestone owner olivas type defect | 1 |
32,041 | 6,691,016,164 | IssuesEvent | 2017-10-09 11:30:32 | primefaces/primereact | https://api.github.com/repos/primefaces/primereact | closed | Dropdown menu sits behind grid | defect | The dropdown menu in DataTable sites behind the grid when resizableColumns or scrollable is enabled.
Caused by the overflow hidden;
https://screenshots.firefox.com/49XyjS9g09rpWvUd/localhost | 1.0 | Dropdown menu sits behind grid - The dropdown menu in DataTable sites behind the grid when resizableColumns or scrollable is enabled.
Caused by the overflow hidden;
https://screenshots.firefox.com/49XyjS9g09rpWvUd/localhost | defect | dropdown menu sits behind grid the dropdown menu in datatable sites behind the grid when resizablecolumns or scrollable is enabled caused by the overflow hidden | 1 |
3,467 | 2,610,063,296 | IssuesEvent | 2015-02-26 18:18:36 | chrsmith/jsjsj122 | https://api.github.com/repos/chrsmith/jsjsj122 | opened | 黄岩治不育哪家效果好 | auto-migrated Priority-Medium Type-Defect | ```
黄岩治不育哪家效果好【台州五洲生殖医院】24小时健康咨询
热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市
椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、1
18、198及椒江一金清公交车直达枫南小区,乘坐107、105、109、
112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:49 | 1.0 | 黄岩治不育哪家效果好 - ```
黄岩治不育哪家效果好【台州五洲生殖医院】24小时健康咨询
热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市
椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、1
18、198及椒江一金清公交车直达枫南小区,乘坐107、105、109、
112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:49 | defect | 黄岩治不育哪家效果好 黄岩治不育哪家效果好【台州五洲生殖医院】 热线 微信号tzwzszyy 医院地址 台州市 (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at | 1 |
79,151 | 28,017,776,592 | IssuesEvent | 2023-03-28 01:06:53 | salsadigitalauorg/civictheme_source | https://api.github.com/repos/salsadigitalauorg/civictheme_source | closed | [DEFECT] Example 1 | Type: Defect | ### Summary
S 1
### Steps to reproduce
Sr 1
### Observed outcome
observed outcome 1
### Expected outcome
Expected outcome 1
<br/>
---
JIRA: CIVIC-1400 | 1.0 | [DEFECT] Example 1 - ### Summary
S 1
### Steps to reproduce
Sr 1
### Observed outcome
observed outcome 1
### Expected outcome
Expected outcome 1
<br/>
---
JIRA: CIVIC-1400 | defect | example summary s steps to reproduce sr observed outcome observed outcome expected outcome expected outcome jira civic | 1 |
33,302 | 7,086,008,705 | IssuesEvent | 2018-01-11 13:09:51 | HewlettPackard/yoda-demo | https://api.github.com/repos/HewlettPackard/yoda-demo | closed | Issue editing text in Chrome browser. | S4 - Low T1 - Defect | <p>Lorem ipsum erat semper cursus mauris phasellus hendrerit lacus, metus habitant neque in mattis et.</p>
<p>Mi scelerisque mauris hac ut est odio curae inceptos purus sapien, eget tristique donec sollicitudin ut per tincidunt magna laoreet.</p>
<p>Iaculis porta sed ipsum enim placerat nisi, pellentesque viverra donec et mollis, libero arcu ullamcorper mattis curae.</p>
<p>Nulla etiam curabitur nec integer aptent, non tincidunt volutpat netus aliquet, rutrum dictumst posuere malesuada.</p>
> estimate 4
| 1.0 | Issue editing text in Chrome browser. - <p>Lorem ipsum erat semper cursus mauris phasellus hendrerit lacus, metus habitant neque in mattis et.</p>
<p>Mi scelerisque mauris hac ut est odio curae inceptos purus sapien, eget tristique donec sollicitudin ut per tincidunt magna laoreet.</p>
<p>Iaculis porta sed ipsum enim placerat nisi, pellentesque viverra donec et mollis, libero arcu ullamcorper mattis curae.</p>
<p>Nulla etiam curabitur nec integer aptent, non tincidunt volutpat netus aliquet, rutrum dictumst posuere malesuada.</p>
> estimate 4
| defect | issue editing text in chrome browser lorem ipsum erat semper cursus mauris phasellus hendrerit lacus metus habitant neque in mattis et mi scelerisque mauris hac ut est odio curae inceptos purus sapien eget tristique donec sollicitudin ut per tincidunt magna laoreet iaculis porta sed ipsum enim placerat nisi pellentesque viverra donec et mollis libero arcu ullamcorper mattis curae nulla etiam curabitur nec integer aptent non tincidunt volutpat netus aliquet rutrum dictumst posuere malesuada estimate | 1 |
72,023 | 23,899,583,499 | IssuesEvent | 2022-09-08 17:29:59 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | opened | [Markup and meta data] HTML markup isn't valid. (09.01.2) | 508/Accessibility authenticated-experience profile 508-defect-2 collab-cycle-feedback Staging CCIssue09.01 CC-Dashboard bad-address-indicator | ### General Information
#### VFS team name
Profile Team
#### VFS product name
Profile
#### VFS feature name
Bad Address Indicator
#### Point of Contact/Reviewers
Brian DeConinck (@briandeconinck) - Accessibility
*For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website.
---
### Platform Issue
HTML markup isn't valid.
### Issue Details
In the "Review your mailing address" alerts, the H2 heading has a `role="alert"` added to it. This isn't a role that's typically allowed to be added to a heading, and it may interfere with the findability of the heading for screen reader users. The `role="alert"` should be placed on a different element.
### Link, screenshot or steps to recreate
The "Review your mailing address" alerts are currently coded as:
```
<h2 slot="headline" tabindex="0" role="alert" aria-live="polite" aria-label="The address we have on file for you may not be correct.">Review your mailing address</h2>
<p>The mailing address we have on file for you may not be correct.</p>
<p><a href="...">...</a></p>
```
### VA.gov Experience Standard
[Category Number 09, Issue Number 01](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html)
### Other References
WCAG SC 1.3.1_A
---
### Platform Recommendation
Headings (H1 through H6) have an implicit [heading role](https://developer.mozilla.org/en-US/docs/Web/Accessibility/ARIA/Roles/heading_role). That role is used by screen readers to identify what the headings are and announce them to users.
When you add a different role attribute to a heading... I'm not totally sure what will happen! Testing with NVDA, these H2s are still announced as headings when you encounter them in the page content. But when you tell NVDA to pull up a list of headings on the page, these H2s aren't included on the list. Something weird is happening with the semantics.
I think the intention with the `role="alert"` is to make sure the warning about the bad mailing address is announced to screen reader users on page load, and that does happen. But it might be worth testing it with the `role="alert"` added to a different element. If you're not able to add it directly to the `va-alert` component, maybe try adding it to a `div` containing the `va-alert` component.
### VFS Team Tasks to Complete
- [ ] Comment on the ticket if there are questions or concerns
- [ ] Close the ticket when the issue has been resolved or validated by your Product Owner. If a team has additional questions or needs Platform help validating the issue, please comment in the ticket. | 1.0 | [Markup and meta data] HTML markup isn't valid. (09.01.2) - ### General Information
#### VFS team name
Profile Team
#### VFS product name
Profile
#### VFS feature name
Bad Address Indicator
#### Point of Contact/Reviewers
Brian DeConinck (@briandeconinck) - Accessibility
*For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website.
---
### Platform Issue
HTML markup isn't valid.
### Issue Details
In the "Review your mailing address" alerts, the H2 heading has a `role="alert"` added to it. This isn't a role that's typically allowed to be added to a heading, and it may interfere with the findability of the heading for screen reader users. The `role="alert"` should be placed on a different element.
### Link, screenshot or steps to recreate
The "Review your mailing address" alerts are currently coded as:
```
<h2 slot="headline" tabindex="0" role="alert" aria-live="polite" aria-label="The address we have on file for you may not be correct.">Review your mailing address</h2>
<p>The mailing address we have on file for you may not be correct.</p>
<p><a href="...">...</a></p>
```
### VA.gov Experience Standard
[Category Number 09, Issue Number 01](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html)
### Other References
WCAG SC 1.3.1_A
---
### Platform Recommendation
Headings (H1 through H6) have an implicit [heading role](https://developer.mozilla.org/en-US/docs/Web/Accessibility/ARIA/Roles/heading_role). That role is used by screen readers to identify what the headings are and announce them to users.
When you add a different role attribute to a heading... I'm not totally sure what will happen! Testing with NVDA, these H2s are still announced as headings when you encounter them in the page content. But when you tell NVDA to pull up a list of headings on the page, these H2s aren't included on the list. Something weird is happening with the semantics.
I think the intention with the `role="alert"` is to make sure the warning about the bad mailing address is announced to screen reader users on page load, and that does happen. But it might be worth testing it with the `role="alert"` added to a different element. If you're not able to add it directly to the `va-alert` component, maybe try adding it to a `div` containing the `va-alert` component.
### VFS Team Tasks to Complete
- [ ] Comment on the ticket if there are questions or concerns
- [ ] Close the ticket when the issue has been resolved or validated by your Product Owner. If a team has additional questions or needs Platform help validating the issue, please comment in the ticket. | defect | html markup isn t valid general information vfs team name profile team vfs product name profile vfs feature name bad address indicator point of contact reviewers brian deconinck briandeconinck accessibility for more information on how to interpret this ticket please refer to the guidance on platform website platform issue html markup isn t valid issue details in the review your mailing address alerts the heading has a role alert added to it this isn t a role that s typically allowed to be added to a heading and it may interfere with the findability of the heading for screen reader users the role alert should be placed on a different element link screenshot or steps to recreate the review your mailing address alerts are currently coded as review your mailing address the mailing address we have on file for you may not be correct va gov experience standard other references wcag sc a platform recommendation headings through have an implicit that role is used by screen readers to identify what the headings are and announce them to users when you add a different role attribute to a heading i m not totally sure what will happen testing with nvda these are still announced as headings when you encounter them in the page content but when you tell nvda to pull up a list of headings on the page these aren t included on the list something weird is happening with the semantics i think the intention with the role alert is to make sure the warning about the bad mailing address is announced to screen reader users on page load and that does happen but it might be worth testing it with the role alert added to a different element if you re not able to add it directly to the va alert component maybe try adding it to a div containing the va alert component vfs team tasks to complete comment on the ticket if there are questions or concerns close the ticket when the issue has been resolved or validated by your product owner if a team has additional questions or needs platform help validating the issue please comment in the ticket | 1 |
60,657 | 17,023,484,491 | IssuesEvent | 2021-07-03 02:16:06 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | index out of range error (abort and close the application) | Component: merkaartor Priority: major Resolution: worksforme Type: defect | **[Submitted to the original trac issue database at 7.04am, Sunday, 27th September 2009]**
i cant reproduce it again.
i do several mods in the map and press upload button:
```
-58.3594 , -34.6784 , -58.3539 , -34.6829
getting: "qrc:/Html/ymap.html?WIDTH=512&HEIGHT=512&BBOX=-58.353882,-34.669359,-58.348389,-34.673876"
0 , 0 , 512 , 511
512 , 511
-58.3539 , -34.6694 , -58.3484 , -34.6739
getting: "qrc:/Html/ymap.html?WIDTH=512&HEIGHT=512&BBOX=-58.353882,-34.673876,-58.348389,-34.678394"
0 , 0 , 512 , 512
512 , 512
-58.3539 , -34.6739 , -58.3484 , -34.6784
getting: "qrc:/Html/ymap.html?WIDTH=512&HEIGHT=512&BBOX=-58.353882,-34.678394,-58.348389,-34.682911"
0 , 0 , 512 , 512
512 , 512
-58.3539 , -34.6784 , -58.3484 , -34.6829
ASSERT failure in QList<T>::operator[]: "index out of range", file /usr/include/qt4/QtCore/qlist.h, line 403
Abortado
```
Running version 0.13.2 on Debian testing amd64
| 1.0 | index out of range error (abort and close the application) - **[Submitted to the original trac issue database at 7.04am, Sunday, 27th September 2009]**
i cant reproduce it again.
i do several mods in the map and press upload button:
```
-58.3594 , -34.6784 , -58.3539 , -34.6829
getting: "qrc:/Html/ymap.html?WIDTH=512&HEIGHT=512&BBOX=-58.353882,-34.669359,-58.348389,-34.673876"
0 , 0 , 512 , 511
512 , 511
-58.3539 , -34.6694 , -58.3484 , -34.6739
getting: "qrc:/Html/ymap.html?WIDTH=512&HEIGHT=512&BBOX=-58.353882,-34.673876,-58.348389,-34.678394"
0 , 0 , 512 , 512
512 , 512
-58.3539 , -34.6739 , -58.3484 , -34.6784
getting: "qrc:/Html/ymap.html?WIDTH=512&HEIGHT=512&BBOX=-58.353882,-34.678394,-58.348389,-34.682911"
0 , 0 , 512 , 512
512 , 512
-58.3539 , -34.6784 , -58.3484 , -34.6829
ASSERT failure in QList<T>::operator[]: "index out of range", file /usr/include/qt4/QtCore/qlist.h, line 403
Abortado
```
Running version 0.13.2 on Debian testing amd64
| defect | index out of range error abort and close the application i cant reproduce it again i do several mods in the map and press upload button getting qrc html ymap html width height bbox getting qrc html ymap html width height bbox getting qrc html ymap html width height bbox assert failure in qlist operator index out of range file usr include qtcore qlist h line abortado running version on debian testing | 1 |
661,366 | 22,051,469,595 | IssuesEvent | 2022-05-30 09:02:52 | googleapis/google-api-nodejs-client | https://api.github.com/repos/googleapis/google-api-nodejs-client | opened | Add support for `androidpublisher.subscriptionsv2` | type: feature request priority: p3 | The `androidpublisher` API now includes a `subscriptionsv2` resource that HAS to be used to validate purchases made with Android Billing v5 API released this month.
https://developers.google.com/android-publisher/api-ref/rest/v3/purchases.subscriptionsv2/get
For information, as a workaround I use a "quick&dirty" polyfill inspired by the existing `subscriptions.get`, that I set as `subscriptions.getv2`... Obviously this should be part of this library. Here's the code.
```js
import * as googleapis_common_1 from 'googleapis-common';
androidpublisher.purchases.subscriptions.getv2 = function(paramsOrCallback, optionsOrCallback, callback) {
let params = (paramsOrCallback ||
{});
let options = (optionsOrCallback || {});
if (typeof paramsOrCallback === 'function') {
callback = paramsOrCallback;
params = {};
options = {};
}
if (typeof optionsOrCallback === 'function') {
callback = optionsOrCallback;
options = {};
}
if (params.subscriptionId) delete params.subscriptionId;
const rootUrl = options.rootUrl || 'https://androidpublisher.googleapis.com/';
const parameters = {
options: Object.assign({
url: (rootUrl +
'/androidpublisher/v3/applications/{packageName}/purchases/subscriptionsv2/tokens/{token}').replace(/([^:]\/)\/+/g, '$1'),
method: 'GET',
}, options),
params,
requiredParams: ['packageName', 'token'],
pathParams: ['packageName', 'token'],
context: this.context,
};
if (callback) {
googleapis_common_1.createAPIRequest(parameters, callback);
}
else {
return googleapis_common_1.createAPIRequest(parameters);
}
}
``` | 1.0 | Add support for `androidpublisher.subscriptionsv2` - The `androidpublisher` API now includes a `subscriptionsv2` resource that HAS to be used to validate purchases made with Android Billing v5 API released this month.
https://developers.google.com/android-publisher/api-ref/rest/v3/purchases.subscriptionsv2/get
For information, as a workaround I use a "quick&dirty" polyfill inspired by the existing `subscriptions.get`, that I set as `subscriptions.getv2`... Obviously this should be part of this library. Here's the code.
```js
import * as googleapis_common_1 from 'googleapis-common';
androidpublisher.purchases.subscriptions.getv2 = function(paramsOrCallback, optionsOrCallback, callback) {
let params = (paramsOrCallback ||
{});
let options = (optionsOrCallback || {});
if (typeof paramsOrCallback === 'function') {
callback = paramsOrCallback;
params = {};
options = {};
}
if (typeof optionsOrCallback === 'function') {
callback = optionsOrCallback;
options = {};
}
if (params.subscriptionId) delete params.subscriptionId;
const rootUrl = options.rootUrl || 'https://androidpublisher.googleapis.com/';
const parameters = {
options: Object.assign({
url: (rootUrl +
'/androidpublisher/v3/applications/{packageName}/purchases/subscriptionsv2/tokens/{token}').replace(/([^:]\/)\/+/g, '$1'),
method: 'GET',
}, options),
params,
requiredParams: ['packageName', 'token'],
pathParams: ['packageName', 'token'],
context: this.context,
};
if (callback) {
googleapis_common_1.createAPIRequest(parameters, callback);
}
else {
return googleapis_common_1.createAPIRequest(parameters);
}
}
``` | non_defect | add support for androidpublisher the androidpublisher api now includes a resource that has to be used to validate purchases made with android billing api released this month for information as a workaround i use a quick dirty polyfill inspired by the existing subscriptions get that i set as subscriptions obviously this should be part of this library here s the code js import as googleapis common from googleapis common androidpublisher purchases subscriptions function paramsorcallback optionsorcallback callback let params paramsorcallback let options optionsorcallback if typeof paramsorcallback function callback paramsorcallback params options if typeof optionsorcallback function callback optionsorcallback options if params subscriptionid delete params subscriptionid const rooturl options rooturl const parameters options object assign url rooturl androidpublisher applications packagename purchases tokens token replace g method get options params requiredparams pathparams context this context if callback googleapis common createapirequest parameters callback else return googleapis common createapirequest parameters | 0 |
293,726 | 25,318,793,132 | IssuesEvent | 2022-11-18 00:47:24 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | closed | It is better to remove the checkbox "Don't ask again about enabling soft delete" from the 'Delete' dialog if the soft delete is already enabled | 🧪 testing :gear: blobs :gear: adls gen2 | **Storage Explorer Version**: 1.27.0-dev
**Build Number**: 20221117.5
**Branch**: main
**Platform/OS**: Windows 10/Linux Ubuntu 18.04/MacOS Ventura 13.0 (Apple M1 Pro)
**Architecture**: ia32/x64
**How Found**: Ad-hoc testing
**Regression From**: Not a regression
## Steps to Reproduce ##
1. Expand one storage account -> Blob Containers. (Soft delete is enabled)
2. Create a blob container -> Upload one blob.
3. Right click the blob -> Click 'Delete'.
4. Check there is no checkbox "Don't ask again about enabling soft delete" on the dialog.
## Expected Experience ##
It is better to remove the checkbox "Don't ask again about enabling soft delete" from the 'Delete' dialog.
## Actual Experience ##
There is a checkbox "Don't ask again about enabling soft delete" on the 'Delete' dialog.

## Addition Context ##
This issue also reproduces for ADLS Gen2 blobs when the setting "Enable ADLS Gen2 soft delete (Preview)" is disabled. | 1.0 | It is better to remove the checkbox "Don't ask again about enabling soft delete" from the 'Delete' dialog if the soft delete is already enabled - **Storage Explorer Version**: 1.27.0-dev
**Build Number**: 20221117.5
**Branch**: main
**Platform/OS**: Windows 10/Linux Ubuntu 18.04/MacOS Ventura 13.0 (Apple M1 Pro)
**Architecture**: ia32/x64
**How Found**: Ad-hoc testing
**Regression From**: Not a regression
## Steps to Reproduce ##
1. Expand one storage account -> Blob Containers. (Soft delete is enabled)
2. Create a blob container -> Upload one blob.
3. Right click the blob -> Click 'Delete'.
4. Check there is no checkbox "Don't ask again about enabling soft delete" on the dialog.
## Expected Experience ##
It is better to remove the checkbox "Don't ask again about enabling soft delete" from the 'Delete' dialog.
## Actual Experience ##
There is a checkbox "Don't ask again about enabling soft delete" on the 'Delete' dialog.

## Addition Context ##
This issue also reproduces for ADLS Gen2 blobs when the setting "Enable ADLS Gen2 soft delete (Preview)" is disabled. | non_defect | it is better to remove the checkbox don t ask again about enabling soft delete from the delete dialog if the soft delete is already enabled storage explorer version dev build number branch main platform os windows linux ubuntu macos ventura apple pro architecture how found ad hoc testing regression from not a regression steps to reproduce expand one storage account blob containers soft delete is enabled create a blob container upload one blob right click the blob click delete check there is no checkbox don t ask again about enabling soft delete on the dialog expected experience it is better to remove the checkbox don t ask again about enabling soft delete from the delete dialog actual experience there is a checkbox don t ask again about enabling soft delete on the delete dialog addition context this issue also reproduces for adls blobs when the setting enable adls soft delete preview is disabled | 0 |
80,795 | 10,210,545,326 | IssuesEvent | 2019-08-14 15:00:00 | kazukinakamuraKN/realtime_chat_nuxt | https://api.github.com/repos/kazukinakamuraKN/realtime_chat_nuxt | opened | 計画 | documentation | ### 機能概要
リアルタイムチャット
### 機能詳細
- リアルタイムにチャットを行う
- チャットの文末に自動で絵文字を追加
- 末尾を認識して、句読点orドットなら削除
- 末尾に絵文字を追加する
### 作業概要
1. チャット機能を追加
1. 絵文字機能を追加
### タスク
- [ ] チャット機能を追加
- https://socket.io/get-started/chat
- チャット機能追加
- [ ] 絵文字機能を追加
- https://www.npmjs.com/package/rm-emoji-picker
- 絵文字pickerを追加する
### チャレンジ
- [ ] チャット機能に追加要素を加える
- https://socket.io/get-started/chat | 1.0 | 計画 - ### 機能概要
リアルタイムチャット
### 機能詳細
- リアルタイムにチャットを行う
- チャットの文末に自動で絵文字を追加
- 末尾を認識して、句読点orドットなら削除
- 末尾に絵文字を追加する
### 作業概要
1. チャット機能を追加
1. 絵文字機能を追加
### タスク
- [ ] チャット機能を追加
- https://socket.io/get-started/chat
- チャット機能追加
- [ ] 絵文字機能を追加
- https://www.npmjs.com/package/rm-emoji-picker
- 絵文字pickerを追加する
### チャレンジ
- [ ] チャット機能に追加要素を加える
- https://socket.io/get-started/chat | non_defect | 計画 機能概要 リアルタイムチャット 機能詳細 リアルタイムにチャットを行う チャットの文末に自動で絵文字を追加 末尾を認識して、句読点orドットなら削除 末尾に絵文字を追加する 作業概要 チャット機能を追加 絵文字機能を追加 タスク チャット機能を追加 チャット機能追加 絵文字機能を追加 絵文字pickerを追加する チャレンジ チャット機能に追加要素を加える | 0 |
27,754 | 5,095,493,914 | IssuesEvent | 2017-01-03 15:23:37 | NewSpring/Holtzman | https://api.github.com/repos/NewSpring/Holtzman | closed | Audio player modal doesn't load | Defect | ## Steps to Reproduce
1. start playing audio in the app
2. Go anywhere
3. Tap the audio banner at the bottom of the screen
### Buggy Behavior
The modal doesn't actually load. It hides the nav and the audio banner appropriately, but nothing else.
### Expected Behavior
It should open the now playing modal
| 1.0 | Audio player modal doesn't load - ## Steps to Reproduce
1. start playing audio in the app
2. Go anywhere
3. Tap the audio banner at the bottom of the screen
### Buggy Behavior
The modal doesn't actually load. It hides the nav and the audio banner appropriately, but nothing else.
### Expected Behavior
It should open the now playing modal
| defect | audio player modal doesn t load steps to reproduce start playing audio in the app go anywhere tap the audio banner at the bottom of the screen buggy behavior the modal doesn t actually load it hides the nav and the audio banner appropriately but nothing else expected behavior it should open the now playing modal | 1 |
86,675 | 8,042,452,150 | IssuesEvent | 2018-07-31 08:12:03 | ClassicWoW/Nefarian_1.12.1_Bugtracker | https://api.github.com/repos/ClassicWoW/Nefarian_1.12.1_Bugtracker | closed | Natalia Mar'alith benutzt ihre Priester-Fähigkeiten nicht. | Mehr Input/Recherche/Tests nötig | **Welches Verhalten wird beobachtet?**
Der NPC "Natalia Mar'alith" verhält sich wie ein normaler Nahkampf-NPC ohne spezielle Fähigkeiten.
**Wie sollte es sich verhalten?**
Dieser NPC sollte zusätzlich Schattenpriester-Fähigkeiten haben plus die Fähigkeit sich selbst zu heilen.
**Schritte zur Reproduzierung**
Natalia Mar'alith angreifen.
**Zusätzliche Informationen (Screenshots, Videos, Klasse, Rasse, Level, etc.)**
Hier sind die Fähigkeiten des NPCs beschrieben:
http://wowwiki.wikia.com/wiki/Natalia_Mar%27alith
Hier gibt es einen Kommentar von 2006, dass sich dieser NPC selbst heilt:
https://wow.freierbund.de/map_and_guide/163325--1/Der_Schlund_des_Wahnsinns-Kommentare.html#add
Das ist in den Fähigkeiten des NPCs in der classic-wow-Datenbank nämlich nicht eingetragen. Leider steht nicht drin wie hoch die Heilung ausfällt. Und Videos habe ich keine gefunden von 2005-2006.
**Alle Kreaturen, Items, Objekte, Quests, Zauber etc. müssen aus unserer Datenbank verlinkt sein.**
Der NPC:
https://datenbank.classic-wow.org/?npc=15215
| 1.0 | Natalia Mar'alith benutzt ihre Priester-Fähigkeiten nicht. - **Welches Verhalten wird beobachtet?**
Der NPC "Natalia Mar'alith" verhält sich wie ein normaler Nahkampf-NPC ohne spezielle Fähigkeiten.
**Wie sollte es sich verhalten?**
Dieser NPC sollte zusätzlich Schattenpriester-Fähigkeiten haben plus die Fähigkeit sich selbst zu heilen.
**Schritte zur Reproduzierung**
Natalia Mar'alith angreifen.
**Zusätzliche Informationen (Screenshots, Videos, Klasse, Rasse, Level, etc.)**
Hier sind die Fähigkeiten des NPCs beschrieben:
http://wowwiki.wikia.com/wiki/Natalia_Mar%27alith
Hier gibt es einen Kommentar von 2006, dass sich dieser NPC selbst heilt:
https://wow.freierbund.de/map_and_guide/163325--1/Der_Schlund_des_Wahnsinns-Kommentare.html#add
Das ist in den Fähigkeiten des NPCs in der classic-wow-Datenbank nämlich nicht eingetragen. Leider steht nicht drin wie hoch die Heilung ausfällt. Und Videos habe ich keine gefunden von 2005-2006.
**Alle Kreaturen, Items, Objekte, Quests, Zauber etc. müssen aus unserer Datenbank verlinkt sein.**
Der NPC:
https://datenbank.classic-wow.org/?npc=15215
| non_defect | natalia mar alith benutzt ihre priester fähigkeiten nicht welches verhalten wird beobachtet der npc natalia mar alith verhält sich wie ein normaler nahkampf npc ohne spezielle fähigkeiten wie sollte es sich verhalten dieser npc sollte zusätzlich schattenpriester fähigkeiten haben plus die fähigkeit sich selbst zu heilen schritte zur reproduzierung natalia mar alith angreifen zusätzliche informationen screenshots videos klasse rasse level etc hier sind die fähigkeiten des npcs beschrieben hier gibt es einen kommentar von dass sich dieser npc selbst heilt das ist in den fähigkeiten des npcs in der classic wow datenbank nämlich nicht eingetragen leider steht nicht drin wie hoch die heilung ausfällt und videos habe ich keine gefunden von alle kreaturen items objekte quests zauber etc müssen aus unserer datenbank verlinkt sein der npc | 0 |
62,286 | 17,023,889,531 | IssuesEvent | 2021-07-03 04:23:30 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Text above map page overlaps | Component: website Priority: minor Resolution: duplicate Type: defect | **[Submitted to the original trac issue database at 7.41am, Wednesday, 4th December 2013]**
The text (traces, diaries, etc.) and boxes (edit, history, etc.) above of the map overlap when the browserscreen is too narrow to support them side by side. When the browserscreen is narrowed even more, the links to traces, diaries etc disappear to be replaced by a "more" dropdown box.
This dropdown box needs to be shown earlier to prevent the overlap.
See attached screenshots. | 1.0 | Text above map page overlaps - **[Submitted to the original trac issue database at 7.41am, Wednesday, 4th December 2013]**
The text (traces, diaries, etc.) and boxes (edit, history, etc.) above of the map overlap when the browserscreen is too narrow to support them side by side. When the browserscreen is narrowed even more, the links to traces, diaries etc disappear to be replaced by a "more" dropdown box.
This dropdown box needs to be shown earlier to prevent the overlap.
See attached screenshots. | defect | text above map page overlaps the text traces diaries etc and boxes edit history etc above of the map overlap when the browserscreen is too narrow to support them side by side when the browserscreen is narrowed even more the links to traces diaries etc disappear to be replaced by a more dropdown box this dropdown box needs to be shown earlier to prevent the overlap see attached screenshots | 1 |
35,136 | 7,595,689,500 | IssuesEvent | 2018-04-27 06:51:44 | STEllAR-GROUP/hpx | https://api.github.com/repos/STEllAR-GROUP/hpx | opened | Setting pu step or offset results in no threads in default pool | category: resource_partitioner type: defect | I expect the following to not fail:
```
./hello_world --hpx:pu-offset=2
terminate called after throwing an instance of 'std::runtime_error'
what(): partitioner::setup_pools: Default pool default has no threads assigned. Please rerun with --hpx:threads=X and check the pool thread assignment
zsh: abort (core dumped) ./solution6 --hpx:pu-offset=2
```
Same goes for `pu-step`. | 1.0 | Setting pu step or offset results in no threads in default pool - I expect the following to not fail:
```
./hello_world --hpx:pu-offset=2
terminate called after throwing an instance of 'std::runtime_error'
what(): partitioner::setup_pools: Default pool default has no threads assigned. Please rerun with --hpx:threads=X and check the pool thread assignment
zsh: abort (core dumped) ./solution6 --hpx:pu-offset=2
```
Same goes for `pu-step`. | defect | setting pu step or offset results in no threads in default pool i expect the following to not fail hello world hpx pu offset terminate called after throwing an instance of std runtime error what partitioner setup pools default pool default has no threads assigned please rerun with hpx threads x and check the pool thread assignment zsh abort core dumped hpx pu offset same goes for pu step | 1 |
20,068 | 3,294,338,448 | IssuesEvent | 2015-10-31 00:30:50 | dkfans/keeperfx | https://api.github.com/repos/dkfans/keeperfx | opened | Max efficiency on single treasure room tile (stable) | Priority-Medium Type-Defect | See the attached save game for some treasure room tiles with max efficiency. Being a single room tile they should have very low efficiency. When you sell the bridge tiles around them they get reduced efficiency, but still more than they should.

From 0.4.6.1723(stable branch) | 1.0 | Max efficiency on single treasure room tile (stable) - See the attached save game for some treasure room tiles with max efficiency. Being a single room tile they should have very low efficiency. When you sell the bridge tiles around them they get reduced efficiency, but still more than they should.

From 0.4.6.1723(stable branch) | defect | max efficiency on single treasure room tile stable see the attached save game for some treasure room tiles with max efficiency being a single room tile they should have very low efficiency when you sell the bridge tiles around them they get reduced efficiency but still more than they should from stable branch | 1 |
53,507 | 13,172,029,961 | IssuesEvent | 2020-08-11 17:42:34 | netlify/build | https://api.github.com/repos/netlify/build | closed | Add `--statsd.host` and `--statsd.port` CLI flags to `@netlify/config` | feat/speedy-builds type: feature | Background at https://github.com/netlify/buildbot/issues/891 and https://github.com/netlify/buildbot/issues/900
See also https://github.com/netlify/build/issues/1741
We should add a `--statsd.host` and `--statsd.port` CLI flags to `@netlify/build`.
Until the buildbot is passing those values to `@netlify/build`, we can hardcode those values. | 1.0 | Add `--statsd.host` and `--statsd.port` CLI flags to `@netlify/config` - Background at https://github.com/netlify/buildbot/issues/891 and https://github.com/netlify/buildbot/issues/900
See also https://github.com/netlify/build/issues/1741
We should add a `--statsd.host` and `--statsd.port` CLI flags to `@netlify/build`.
Until the buildbot is passing those values to `@netlify/build`, we can hardcode those values. | non_defect | add statsd host and statsd port cli flags to netlify config background at and see also we should add a statsd host and statsd port cli flags to netlify build until the buildbot is passing those values to netlify build we can hardcode those values | 0 |
81,615 | 31,148,136,701 | IssuesEvent | 2023-08-16 08:05:48 | MarcusWolschon/osmeditor4android | https://api.github.com/repos/MarcusWolschon/osmeditor4android | closed | Pending changes count background uncoloured on Android 5 and 6 | Defect Minor | On an old (2017) tablet, the Transfer icon's pending changes count lacks a coloured background, making the number difficult to read when the Theme is Light due to the low contrast. It's legible enough in Dark Theme but text isn't coloured.
## Vespucci Version
19.1 BETA 1
## Download source
Google Play Store
## Device (Manufacturer and Model)
Amazon Kindle Fire HD 10 Gen 7 [2017]
## Android Version
Fire OS 5.7.0.0
## Behaviour/Symptoms
The pending changes count's text is white, but without a coloured background, its visibility on Light Theme has contrast low enough to be barely visible.
## Expected Behaviour
A coloured background or at least bold/thick digits with text coloured appropriately for themed contrast.
## How to recreate
1. Launch Vespucci on an old Kindle Fire tablet.
2. Set theme within Advanced Preferences to Light (if not already set).
2. Download data.
3. Move a node or way.
4. Observe the pending changes count on the transfer (↑↓) icon.
## Crash dump submitted (no or yes + date)
no
## Any other potentially relevant information
Not an issue on an Android 10 phone (LG V50) or Android 13 phone (Samsung Galaxy Z Fold3).
| 1.0 | Pending changes count background uncoloured on Android 5 and 6 - On an old (2017) tablet, the Transfer icon's pending changes count lacks a coloured background, making the number difficult to read when the Theme is Light due to the low contrast. It's legible enough in Dark Theme but text isn't coloured.
## Vespucci Version
19.1 BETA 1
## Download source
Google Play Store
## Device (Manufacturer and Model)
Amazon Kindle Fire HD 10 Gen 7 [2017]
## Android Version
Fire OS 5.7.0.0
## Behaviour/Symptoms
The pending changes count's text is white, but without a coloured background, its visibility on Light Theme has contrast low enough to be barely visible.
## Expected Behaviour
A coloured background or at least bold/thick digits with text coloured appropriately for themed contrast.
## How to recreate
1. Launch Vespucci on an old Kindle Fire tablet.
2. Set theme within Advanced Preferences to Light (if not already set).
2. Download data.
3. Move a node or way.
4. Observe the pending changes count on the transfer (↑↓) icon.
## Crash dump submitted (no or yes + date)
no
## Any other potentially relevant information
Not an issue on an Android 10 phone (LG V50) or Android 13 phone (Samsung Galaxy Z Fold3).
| defect | pending changes count background uncoloured on android and on an old tablet the transfer icon s pending changes count lacks a coloured background making the number difficult to read when the theme is light due to the low contrast it s legible enough in dark theme but text isn t coloured vespucci version beta download source google play store device manufacturer and model amazon kindle fire hd gen android version fire os behaviour symptoms the pending changes count s text is white but without a coloured background its visibility on light theme has contrast low enough to be barely visible expected behaviour a coloured background or at least bold thick digits with text coloured appropriately for themed contrast how to recreate launch vespucci on an old kindle fire tablet set theme within advanced preferences to light if not already set download data move a node or way observe the pending changes count on the transfer ↑↓ icon crash dump submitted no or yes date no any other potentially relevant information not an issue on an android phone lg or android phone samsung galaxy z | 1 |
230,235 | 17,606,905,961 | IssuesEvent | 2021-08-17 18:20:10 | OHDSI/CommonDataModel | https://api.github.com/repos/OHDSI/CommonDataModel | closed | Article on how to handle drug infusions in the CDM | Documentation | We need an article for the CDM website on how to represent drug infusions in a standard way | 1.0 | Article on how to handle drug infusions in the CDM - We need an article for the CDM website on how to represent drug infusions in a standard way | non_defect | article on how to handle drug infusions in the cdm we need an article for the cdm website on how to represent drug infusions in a standard way | 0 |
3,895 | 4,700,624,250 | IssuesEvent | 2016-10-12 19:04:57 | UPDDI/mps-database-server | https://api.github.com/repos/UPDDI/mps-database-server | opened | Refactor Forms to be less Verbose | infrastructure refactor | Currently, some fields are written in a very verbose way in the templates. Ideally, we should have a DRY (Don't Repeat Yourself) solution for this. Initial work has been made to have a file for "generic fields," but it is not totally robust and is not employed everywhere in can be used.
Here are some considerations:
- We should be able to group fields together
- We should be able to change the column sizes for the label and the input
- It should make use of as much from the form/view as possible
- Ideally it should not have labels etc. that diverge from the models | 1.0 | Refactor Forms to be less Verbose - Currently, some fields are written in a very verbose way in the templates. Ideally, we should have a DRY (Don't Repeat Yourself) solution for this. Initial work has been made to have a file for "generic fields," but it is not totally robust and is not employed everywhere in can be used.
Here are some considerations:
- We should be able to group fields together
- We should be able to change the column sizes for the label and the input
- It should make use of as much from the form/view as possible
- Ideally it should not have labels etc. that diverge from the models | non_defect | refactor forms to be less verbose currently some fields are written in a very verbose way in the templates ideally we should have a dry don t repeat yourself solution for this initial work has been made to have a file for generic fields but it is not totally robust and is not employed everywhere in can be used here are some considerations we should be able to group fields together we should be able to change the column sizes for the label and the input it should make use of as much from the form view as possible ideally it should not have labels etc that diverge from the models | 0 |
84,951 | 15,728,371,772 | IssuesEvent | 2021-03-29 13:45:08 | ssobue/oauth2-provider | https://api.github.com/repos/ssobue/oauth2-provider | closed | CVE-2020-10672 (High) detected in jackson-databind-2.9.9.jar | security vulnerability | ## CVE-2020-10672 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /oauth2-provider/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.1.5.RELEASE.jar
- :x: **jackson-databind-2.9.9.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.aries.transaction.jms.internal.XaPooledConnectionFactory (aka aries.transaction.jms).
<p>Publish Date: 2020-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10672>CVE-2020-10672</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-10672">https://nvd.nist.gov/vuln/detail/CVE-2020-10672</a></p>
<p>Release Date: 2020-03-18</p>
<p>Fix Resolution: jackson-databind-2.9.10.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-10672 (High) detected in jackson-databind-2.9.9.jar - ## CVE-2020-10672 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /oauth2-provider/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.1.5.RELEASE.jar
- :x: **jackson-databind-2.9.9.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.aries.transaction.jms.internal.XaPooledConnectionFactory (aka aries.transaction.jms).
<p>Publish Date: 2020-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10672>CVE-2020-10672</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-10672">https://nvd.nist.gov/vuln/detail/CVE-2020-10672</a></p>
<p>Release Date: 2020-03-18</p>
<p>Fix Resolution: jackson-databind-2.9.10.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file provider pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache aries transaction jms internal xapooledconnectionfactory aka aries transaction jms publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jackson databind step up your open source security game with whitesource | 0 |
43,111 | 11,487,846,780 | IssuesEvent | 2020-02-11 12:49:18 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | PDBWriter incompatible with StringIO if atom coordinates go out of range | Component-Writers Format-PDB defect | **Expected behavior**
Assuming that PDBWriter is supposed to support a StringIO object as a valid destination, I expect that attempting to write a PDB with atom coordinates out of range will raise the "PDB files must have coordinate values between [...]" exception.
**Actual behavior**
The check attempts to `os.remove` the StringIO object and does not catch the resulting exception, so the coordinate exception never gets raised.
```
~/narupa-protocols/narupa-protocol/python-libraries/narupa-ase/examples/nglclient.py in mda_to_pdb_str(universe)
54 def mda_to_pdb_str(universe: mda.Universe):
55 with StringIO() as str_io, mda.coordinates.PDB.PDBWriter(str_io) as writer:
---> 56 writer.write(universe.atoms)
57 pdb = str_io.getvalue()
58 return pdb
~/anaconda3/envs/narupa-dev/lib/python3.7/site-packages/MDAnalysis/coordinates/PDB.py in write(self, obj)
765 # write_all_timesteps() to dump everything in one go, or do the
766 # traditional loop over frames
--> 767 self.write_next_timestep(self.ts, multiframe=self._multiframe)
768 self._write_pdb_bonds()
769 # END record is written when file is being close()d
~/anaconda3/envs/narupa-dev/lib/python3.7/site-packages/MDAnalysis/coordinates/PDB.py in write_next_timestep(self, ts, **kwargs)
839 raise NoDataError("PBDWriter: no coordinate data to write to "
840 "trajectory file")
--> 841 self._check_pdb_coordinates()
842 self._write_timestep(ts, **kwargs)
843
~/anaconda3/envs/narupa-dev/lib/python3.7/site-packages/MDAnalysis/coordinates/PDB.py in _check_pdb_coordinates(self)
670 self.close()
671 try:
--> 672 os.remove(self.filename)
673 except OSError as err:
674 if err.errno == errno.ENOENT:
TypeError: remove: path should be string, bytes or os.PathLike, not _io.StringIO
```
**Code to reproduce the behavior**
Show us how to reproduce the failiure. If you can, use trajectory files from the test data.
```
import io
import MDAnalysis as mda
from MDAnalysis.tests.datafiles import PDB
u = mda.Universe(PDB)
u.atoms.translate([-9999, -9999, -9999])
with io.StringIO() as str_io, mda.coordinates.PDB.PDBWriter(str_io) as writer:
#writer.filename = "" # uncomment to avoid issue
writer.write(u.atoms)
```
**Currently version of MDAnalysis**
- Which version are you using? (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`)
0.20.1
- Which version of Python (`python -V`)?
Python 3.7.3
- Which operating system?
Windows 10
| 1.0 | PDBWriter incompatible with StringIO if atom coordinates go out of range - **Expected behavior**
Assuming that PDBWriter is supposed to support a StringIO object as a valid destination, I expect that attempting to write a PDB with atom coordinates out of range will raise the "PDB files must have coordinate values between [...]" exception.
**Actual behavior**
The check attempts to `os.remove` the StringIO object and does not catch the resulting exception, so the coordinate exception never gets raised.
```
~/narupa-protocols/narupa-protocol/python-libraries/narupa-ase/examples/nglclient.py in mda_to_pdb_str(universe)
54 def mda_to_pdb_str(universe: mda.Universe):
55 with StringIO() as str_io, mda.coordinates.PDB.PDBWriter(str_io) as writer:
---> 56 writer.write(universe.atoms)
57 pdb = str_io.getvalue()
58 return pdb
~/anaconda3/envs/narupa-dev/lib/python3.7/site-packages/MDAnalysis/coordinates/PDB.py in write(self, obj)
765 # write_all_timesteps() to dump everything in one go, or do the
766 # traditional loop over frames
--> 767 self.write_next_timestep(self.ts, multiframe=self._multiframe)
768 self._write_pdb_bonds()
769 # END record is written when file is being close()d
~/anaconda3/envs/narupa-dev/lib/python3.7/site-packages/MDAnalysis/coordinates/PDB.py in write_next_timestep(self, ts, **kwargs)
839 raise NoDataError("PBDWriter: no coordinate data to write to "
840 "trajectory file")
--> 841 self._check_pdb_coordinates()
842 self._write_timestep(ts, **kwargs)
843
~/anaconda3/envs/narupa-dev/lib/python3.7/site-packages/MDAnalysis/coordinates/PDB.py in _check_pdb_coordinates(self)
670 self.close()
671 try:
--> 672 os.remove(self.filename)
673 except OSError as err:
674 if err.errno == errno.ENOENT:
TypeError: remove: path should be string, bytes or os.PathLike, not _io.StringIO
```
**Code to reproduce the behavior**
Show us how to reproduce the failiure. If you can, use trajectory files from the test data.
```
import io
import MDAnalysis as mda
from MDAnalysis.tests.datafiles import PDB
u = mda.Universe(PDB)
u.atoms.translate([-9999, -9999, -9999])
with io.StringIO() as str_io, mda.coordinates.PDB.PDBWriter(str_io) as writer:
#writer.filename = "" # uncomment to avoid issue
writer.write(u.atoms)
```
**Currently version of MDAnalysis**
- Which version are you using? (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`)
0.20.1
- Which version of Python (`python -V`)?
Python 3.7.3
- Which operating system?
Windows 10
| defect | pdbwriter incompatible with stringio if atom coordinates go out of range expected behavior assuming that pdbwriter is supposed to support a stringio object as a valid destination i expect that attempting to write a pdb with atom coordinates out of range will raise the pdb files must have coordinate values between exception actual behavior the check attempts to os remove the stringio object and does not catch the resulting exception so the coordinate exception never gets raised narupa protocols narupa protocol python libraries narupa ase examples nglclient py in mda to pdb str universe def mda to pdb str universe mda universe with stringio as str io mda coordinates pdb pdbwriter str io as writer writer write universe atoms pdb str io getvalue return pdb envs narupa dev lib site packages mdanalysis coordinates pdb py in write self obj write all timesteps to dump everything in one go or do the traditional loop over frames self write next timestep self ts multiframe self multiframe self write pdb bonds end record is written when file is being close d envs narupa dev lib site packages mdanalysis coordinates pdb py in write next timestep self ts kwargs raise nodataerror pbdwriter no coordinate data to write to trajectory file self check pdb coordinates self write timestep ts kwargs envs narupa dev lib site packages mdanalysis coordinates pdb py in check pdb coordinates self self close try os remove self filename except oserror as err if err errno errno enoent typeerror remove path should be string bytes or os pathlike not io stringio code to reproduce the behavior show us how to reproduce the failiure if you can use trajectory files from the test data import io import mdanalysis as mda from mdanalysis tests datafiles import pdb u mda universe pdb u atoms translate with io stringio as str io mda coordinates pdb pdbwriter str io as writer writer filename uncomment to avoid issue writer write u atoms currently version of mdanalysis which version are you using run python c import mdanalysis as mda print mda version which version of python python v python which operating system windows | 1 |
465,993 | 13,395,784,188 | IssuesEvent | 2020-09-03 08:57:13 | googleapis/java-redis | https://api.github.com/repos/googleapis/java-redis | closed | Synthesis failed for java-redis | api: redis autosynth failure priority: p1 type: bug | Hello! Autosynth couldn't regenerate java-redis. :broken_heart:
Here's the output from running `synth.py`:
```
hl
Saved ./Jinja2-2.11.2-py2.py3-none-any.whl
Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5))
Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6))
Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7))
Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz
Complete output from command python setup.py egg_info:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmpfs/tmp/tmp6m3w5les/setuptools-tmp/setuptools/__init__.py", line 6, in <module>
import distutils.core
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 82, in create_module
return importlib.import_module('._distutils', 'setuptools')
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
ModuleNotFoundError: No module named 'setuptools._distutils'
----------------------------------------
(Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-rc4r6uhv/pypandoc/
)
ERROR: no such package '@gapic_generator_python_pip_deps//': pip_import failed: Collecting click==7.1.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 1))
Using cached https://files.pythonhosted.org/packages/d2/3d/fa76db83bf75c4f8d338c2fd15c8d33fdd7ad23a9b5e57eb6c5de26b430e/click-7.1.2-py2.py3-none-any.whl
Saved ./click-7.1.2-py2.py3-none-any.whl
Collecting google-api-core==1.22.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 2))
Using cached https://files.pythonhosted.org/packages/e0/2d/7c6c75013105e1d2b6eaa1bf18a56995be1dbc673c38885aea31136e9918/google_api_core-1.22.1-py2.py3-none-any.whl
Saved ./google_api_core-1.22.1-py2.py3-none-any.whl
Collecting googleapis-common-protos==1.52.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 3))
Using cached https://files.pythonhosted.org/packages/03/74/3956721ea1eb4bcf7502a311fdaa60b85bd751de4e57d1943afe9b334141/googleapis_common_protos-1.52.0-py2.py3-none-any.whl
Saved ./googleapis_common_protos-1.52.0-py2.py3-none-any.whl
Collecting jinja2==2.11.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 4))
Using cached https://files.pythonhosted.org/packages/30/9e/f663a2aa66a09d838042ae1a2c5659828bb9b41ea3a6efa20a20fd92b121/Jinja2-2.11.2-py2.py3-none-any.whl
Saved ./Jinja2-2.11.2-py2.py3-none-any.whl
Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5))
Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6))
Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7))
Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz
Complete output from command python setup.py egg_info:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmpfs/tmp/tmp6m3w5les/setuptools-tmp/setuptools/__init__.py", line 6, in <module>
import distutils.core
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 82, in create_module
return importlib.import_module('._distutils', 'setuptools')
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
ModuleNotFoundError: No module named 'setuptools._distutils'
----------------------------------------
(Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-rc4r6uhv/pypandoc/
)
INFO: Elapsed time: 2.315s
INFO: 0 processes.
FAILED: Build did NOT complete successfully (0 packages loaded)
FAILED: Build did NOT complete successfully (0 packages loaded)
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/java-redis/synth.py", line 28, in <module>
bazel_target=f'//google/cloud/{service}/{version}:google-cloud-{service}-{version}-java',
File "/tmpfs/src/github/synthtool/synthtool/languages/java.py", line 298, in bazel_library
library = gapic.java_library(service=service, version=version, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 62, in java_library
service, version, "java", tar_strip_components=0, **kwargs
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 183, in _generate_code
shell.run(bazel_run_args)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['bazel', '--max_idle_secs=240', 'build', '//google/cloud/redis/v1:google-cloud-redis-v1-java']' returned non-zero exit status 1.
2020-08-30 14:23:48,202 autosynth [ERROR] > Synthesis failed
2020-08-30 14:23:48,203 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at f3c275b chore: regenerate README (#177)
2020-08-30 14:23:48,208 autosynth [DEBUG] > Running: git checkout autosynth
Switched to branch 'autosynth'
2020-08-30 14:23:48,213 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 690, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 539, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 670, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 375, in synthesize_loop
has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 273, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/29b10bde-06f7-44fc-9da8-8806922323b8/targets/github%2Fsynthtool;config=default/tests;query=java-redis;failed=false).
| 1.0 | Synthesis failed for java-redis - Hello! Autosynth couldn't regenerate java-redis. :broken_heart:
Here's the output from running `synth.py`:
```
hl
Saved ./Jinja2-2.11.2-py2.py3-none-any.whl
Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5))
Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6))
Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7))
Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz
Complete output from command python setup.py egg_info:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmpfs/tmp/tmp6m3w5les/setuptools-tmp/setuptools/__init__.py", line 6, in <module>
import distutils.core
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 82, in create_module
return importlib.import_module('._distutils', 'setuptools')
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
ModuleNotFoundError: No module named 'setuptools._distutils'
----------------------------------------
(Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-rc4r6uhv/pypandoc/
)
ERROR: no such package '@gapic_generator_python_pip_deps//': pip_import failed: Collecting click==7.1.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 1))
Using cached https://files.pythonhosted.org/packages/d2/3d/fa76db83bf75c4f8d338c2fd15c8d33fdd7ad23a9b5e57eb6c5de26b430e/click-7.1.2-py2.py3-none-any.whl
Saved ./click-7.1.2-py2.py3-none-any.whl
Collecting google-api-core==1.22.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 2))
Using cached https://files.pythonhosted.org/packages/e0/2d/7c6c75013105e1d2b6eaa1bf18a56995be1dbc673c38885aea31136e9918/google_api_core-1.22.1-py2.py3-none-any.whl
Saved ./google_api_core-1.22.1-py2.py3-none-any.whl
Collecting googleapis-common-protos==1.52.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 3))
Using cached https://files.pythonhosted.org/packages/03/74/3956721ea1eb4bcf7502a311fdaa60b85bd751de4e57d1943afe9b334141/googleapis_common_protos-1.52.0-py2.py3-none-any.whl
Saved ./googleapis_common_protos-1.52.0-py2.py3-none-any.whl
Collecting jinja2==2.11.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 4))
Using cached https://files.pythonhosted.org/packages/30/9e/f663a2aa66a09d838042ae1a2c5659828bb9b41ea3a6efa20a20fd92b121/Jinja2-2.11.2-py2.py3-none-any.whl
Saved ./Jinja2-2.11.2-py2.py3-none-any.whl
Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5))
Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6))
Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7))
Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz
Complete output from command python setup.py egg_info:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmpfs/tmp/tmp6m3w5les/setuptools-tmp/setuptools/__init__.py", line 6, in <module>
import distutils.core
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 82, in create_module
return importlib.import_module('._distutils', 'setuptools')
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
ModuleNotFoundError: No module named 'setuptools._distutils'
----------------------------------------
(Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-rc4r6uhv/pypandoc/
)
INFO: Elapsed time: 2.315s
INFO: 0 processes.
FAILED: Build did NOT complete successfully (0 packages loaded)
FAILED: Build did NOT complete successfully (0 packages loaded)
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/java-redis/synth.py", line 28, in <module>
bazel_target=f'//google/cloud/{service}/{version}:google-cloud-{service}-{version}-java',
File "/tmpfs/src/github/synthtool/synthtool/languages/java.py", line 298, in bazel_library
library = gapic.java_library(service=service, version=version, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 62, in java_library
service, version, "java", tar_strip_components=0, **kwargs
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 183, in _generate_code
shell.run(bazel_run_args)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['bazel', '--max_idle_secs=240', 'build', '//google/cloud/redis/v1:google-cloud-redis-v1-java']' returned non-zero exit status 1.
2020-08-30 14:23:48,202 autosynth [ERROR] > Synthesis failed
2020-08-30 14:23:48,203 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at f3c275b chore: regenerate README (#177)
2020-08-30 14:23:48,208 autosynth [DEBUG] > Running: git checkout autosynth
Switched to branch 'autosynth'
2020-08-30 14:23:48,213 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 690, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 539, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 670, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 375, in synthesize_loop
has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 273, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/29b10bde-06f7-44fc-9da8-8806922323b8/targets/github%2Fsynthtool;config=default/tests;query=java-redis;failed=false).
| non_defect | synthesis failed for java redis hello autosynth couldn t regenerate java redis broken heart here s the output from running synth py hl saved none any whl collecting markupsafe from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved markupsafe whl collecting protobuf from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved protobuf whl collecting pypandoc from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached complete output from command python setup py egg info traceback most recent call last file line in file tmpfs tmp setuptools tmp setuptools init py line in import distutils core file tmpfs src github synthtool env lib site packages distutils hack init py line in create module return importlib import module distutils setuptools file home kbuilder pyenv versions lib importlib init py line in import module return bootstrap gcd import name package level modulenotfounderror no module named setuptools distutils command python setup py egg info failed with error code in tmpfs tmp pip build pypandoc error no such package gapic generator python pip deps pip import failed collecting click from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved click none any whl collecting google api core from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved google api core none any whl collecting googleapis common protos from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved googleapis common protos none any whl collecting from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved none any whl collecting markupsafe from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved markupsafe whl collecting protobuf from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved protobuf whl collecting pypandoc from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached complete output from command python setup py egg info traceback most recent call last file line in file tmpfs tmp setuptools tmp setuptools init py line in import distutils core file tmpfs src github synthtool env lib site packages distutils hack init py line in create module return importlib import module distutils setuptools file home kbuilder pyenv versions lib importlib init py line in import module return bootstrap gcd import name package level modulenotfounderror no module named setuptools distutils command python setup py egg info failed with error code in tmpfs tmp pip build pypandoc info elapsed time info processes failed build did not complete successfully packages loaded failed build did not complete successfully packages loaded traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool java redis synth py line in bazel target f google cloud service version google cloud service version java file tmpfs src github synthtool synthtool languages java py line in bazel library library gapic java library service service version version kwargs file tmpfs src github synthtool synthtool gcp gapic bazel py line in java library service version java tar strip components kwargs file tmpfs src github synthtool synthtool gcp gapic bazel py line in generate code shell run bazel run args file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git reset hard head head is now at chore regenerate readme autosynth running git checkout autosynth switched to branch autosynth autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop has changes toolbox synthesize version in new branch synthesizer youngest file tmpfs src github synthtool autosynth synth py line in synthesize version in new branch synthesizer synthesize synth log path self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.