Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 1 744 | labels stringlengths 4 574 | body stringlengths 9 211k | index stringclasses 10 values | text_combine stringlengths 96 211k | label stringclasses 2 values | text stringlengths 96 188k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
17,609 | 23,428,329,647 | IssuesEvent | 2022-08-14 18:34:49 | ForNeVeR/Cesium | https://api.github.com/repos/ForNeVeR/Cesium | opened | Nested include support | kind:feature area:preprocessor | Look for the number `xxx` in the code to find clues to resolve this issue. | 1.0 | Nested include support - Look for the number `xxx` in the code to find clues to resolve this issue. | process | nested include support look for the number xxx in the code to find clues to resolve this issue | 1 |
19,842 | 26,244,352,652 | IssuesEvent | 2023-01-05 14:08:56 | metabase/metabase | https://api.github.com/repos/metabase/metabase | closed | Pivot tables won't respect the order set up in the query builder | Type:Bug Priority:P2 Querying/Processor Visualization/Tables | **Describe the bug**
If you try to do a pivot table with an order clause, it won't get respected, since the pivot table sends several queries to the DB and the last one (the one that generates the final table) does not bring the ORDER BY from the initial query (it does something like ORDER BY "source"."pivot-grouping" ASC, no matter the order you have).
Seems to have been designed this way
https://github.com/metabase/metabase/blob/3fe17270f2fbdcdb104d5ccdcebf612913364413/src/metabase/query_processor/pivot.clj#L104
**Logs**
E.g. this is with DESC order
```
2022-12-28 22:47:57.524 UTC [85] LOG: execute <unnamed>: SET SESSION CHARACTERISTICS AS TRANSACTION ISOLATION LEVEL READ UNCOMMITTED
2022-12-28 22:47:57.525 UTC [85] LOG: execute <unnamed>: BEGIN READ ONLY
2022-12-28 22:47:57.525 UTC [85] LOG: execute <unnamed>/C_37: -- Metabase:: userID: 1 queryType: MBQL queryHash: b3f10d2564045d5b6c5006c90ca16ae5e70f22701cbaf0bacbcb67ac11e753f2
SELECT "source"."products__via__product_id__vendor" AS "products__via__product_id__vendor", "source"."pivot-grouping" AS "pivot-grouping", sum("source"."quantity") AS "sum", sum("source"."total") AS "sum_2" FROM (SELECT "public"."orders"."id" AS "id", "public"."orders"."user_id" AS "user_id", "public"."orders"."product_id" AS "product_id", "public"."orders"."subtotal" AS "subtotal", "public"."orders"."tax" AS "tax", "public"."orders"."total" AS "total", "public"."orders"."discount" AS "discount", "public"."orders"."created_at" AS "created_at", "public"."orders"."quantity" AS "quantity", abs(0) AS "pivot-grouping", "products__via__product_id"."vendor" AS "products__via__product_id__vendor", "products__via__product_id"."id" AS "products__via__product_id__id" FROM "public"."orders" LEFT JOIN "public"."products" "products__via__product_id" ON "public"."orders"."product_id" = "products__via__product_id"."id") "source" GROUP BY "source"."products__via__product_id__vendor", "source"."pivot-grouping" ORDER BY "source"."products__via__product_id__vendor" ASC, "source"."pivot-grouping" ASC
2022-12-28 22:47:57.547 UTC [85] LOG: execute S_2: ROLLBACK
2022-12-28 22:47:57.548 UTC [85] LOG: execute <unnamed>: SET SESSION CHARACTERISTICS AS TRANSACTION ISOLATION LEVEL READ COMMITTED
2022-12-28 22:47:57.562 UTC [85] LOG: execute <unnamed>: SET SESSION CHARACTERISTICS AS TRANSACTION ISOLATION LEVEL READ UNCOMMITTED
2022-12-28 22:47:57.562 UTC [85] LOG: execute <unnamed>: BEGIN READ ONLY
2022-12-28 22:47:57.562 UTC [85] LOG: execute <unnamed>/C_38: -- Metabase:: userID: 1 queryType: MBQL queryHash: cc962af2784c4fd2a9c2dbb673ec525d446a3be48e38dd513c8a82526c66bbe7
SELECT "source"."pivot-grouping" AS "pivot-grouping", sum("source"."quantity") AS "sum", sum("source"."total") AS "sum_2" FROM (SELECT "public"."orders"."id" AS "id", "public"."orders"."user_id" AS "user_id", "public"."orders"."product_id" AS "product_id", "public"."orders"."subtotal" AS "subtotal", "public"."orders"."tax" AS "tax", "public"."orders"."total" AS "total", "public"."orders"."discount" AS "discount", "public"."orders"."created_at" AS "created_at", "public"."orders"."quantity" AS "quantity", abs(1) AS "pivot-grouping" FROM "public"."orders") "source" GROUP BY "source"."pivot-grouping" ORDER BY "source"."pivot-grouping" ASC
```
This is with the ASC order
```
2022-12-28 22:51:28.834 UTC [85] LOG: execute <unnamed>: SET SESSION CHARACTERISTICS AS TRANSACTION ISOLATION LEVEL READ UNCOMMITTED
2022-12-28 22:51:28.834 UTC [85] LOG: execute <unnamed>: BEGIN READ ONLY
2022-12-28 22:51:28.834 UTC [85] LOG: execute <unnamed>/C_39: -- Metabase:: userID: 1 queryType: MBQL queryHash: b3f10d2564045d5b6c5006c90ca16ae5e70f22701cbaf0bacbcb67ac11e753f2
SELECT "source"."products__via__product_id__vendor" AS "products__via__product_id__vendor", "source"."pivot-grouping" AS "pivot-grouping", sum("source"."quantity") AS "sum", sum("source"."total") AS "sum_2" FROM (SELECT "public"."orders"."id" AS "id", "public"."orders"."user_id" AS "user_id", "public"."orders"."product_id" AS "product_id", "public"."orders"."subtotal" AS "subtotal", "public"."orders"."tax" AS "tax", "public"."orders"."total" AS "total", "public"."orders"."discount" AS "discount", "public"."orders"."created_at" AS "created_at", "public"."orders"."quantity" AS "quantity", abs(0) AS "pivot-grouping", "products__via__product_id"."vendor" AS "products__via__product_id__vendor", "products__via__product_id"."id" AS "products__via__product_id__id" FROM "public"."orders" LEFT JOIN "public"."products" "products__via__product_id" ON "public"."orders"."product_id" = "products__via__product_id"."id") "source" GROUP BY "source"."products__via__product_id__vendor", "source"."pivot-grouping" ORDER BY "source"."products__via__product_id__vendor" ASC, "source"."pivot-grouping" ASC
2022-12-28 22:51:28.856 UTC [85] LOG: execute S_2: ROLLBACK
2022-12-28 22:51:28.856 UTC [85] LOG: execute <unnamed>: SET SESSION CHARACTERISTICS AS TRANSACTION ISOLATION LEVEL READ COMMITTED
2022-12-28 22:51:28.876 UTC [85] LOG: execute <unnamed>: SET SESSION CHARACTERISTICS AS TRANSACTION ISOLATION LEVEL READ UNCOMMITTED
2022-12-28 22:51:28.876 UTC [85] LOG: execute <unnamed>: BEGIN READ ONLY
2022-12-28 22:51:28.876 UTC [85] LOG: execute <unnamed>/C_40: -- Metabase:: userID: 1 queryType: MBQL queryHash: cc962af2784c4fd2a9c2dbb673ec525d446a3be48e38dd513c8a82526c66bbe7
SELECT "source"."pivot-grouping" AS "pivot-grouping", sum("source"."quantity") AS "sum", sum("source"."total") AS "sum_2" FROM (SELECT "public"."orders"."id" AS "id", "public"."orders"."user_id" AS "user_id", "public"."orders"."product_id" AS "product_id", "public"."orders"."subtotal" AS "subtotal", "public"."orders"."tax" AS "tax", "public"."orders"."total" AS "total", "public"."orders"."discount" AS "discount", "public"."orders"."created_at" AS "created_at", "public"."orders"."quantity" AS "quantity", abs(1) AS "pivot-grouping" FROM "public"."orders") "source" GROUP BY "source"."pivot-grouping" ORDER BY "source"."pivot-grouping" ASC
```
**To Reproduce**
1) Do a sum of total and sum of quantity by product vendor with the orders table, visualize it as a pivot table.
2) Then change the order in the query builder and visualize it again, see that the order didn't change
**Expected behavior**
Order BY should be done in the first query and then passed to the last query. This is important as many customers have more than 10K series and the queries don't return with all data so they can't sort on the frontend
**Screenshots**
NA
**Information about your Metabase Installation:**
- Metabase version: 45.1
- Metabase hosting environment: Docker
- Metabase internal database: Postgres
**Severity**
P2
**Additional context**
Reported by a customer | 1.0 | Pivot tables won't respect the order set up in the query builder - **Describe the bug**
If you try to do a pivot table with an order clause, it won't get respected, since the pivot table sends several queries to the DB and the last one (the one that generates the final table) does not bring the ORDER BY from the initial query (it does something like ORDER BY "source"."pivot-grouping" ASC, no matter the order you have).
Seems to have been designed this way
https://github.com/metabase/metabase/blob/3fe17270f2fbdcdb104d5ccdcebf612913364413/src/metabase/query_processor/pivot.clj#L104
**Logs**
E.g. this is with DESC order
```
2022-12-28 22:47:57.524 UTC [85] LOG: execute <unnamed>: SET SESSION CHARACTERISTICS AS TRANSACTION ISOLATION LEVEL READ UNCOMMITTED
2022-12-28 22:47:57.525 UTC [85] LOG: execute <unnamed>: BEGIN READ ONLY
2022-12-28 22:47:57.525 UTC [85] LOG: execute <unnamed>/C_37: -- Metabase:: userID: 1 queryType: MBQL queryHash: b3f10d2564045d5b6c5006c90ca16ae5e70f22701cbaf0bacbcb67ac11e753f2
SELECT "source"."products__via__product_id__vendor" AS "products__via__product_id__vendor", "source"."pivot-grouping" AS "pivot-grouping", sum("source"."quantity") AS "sum", sum("source"."total") AS "sum_2" FROM (SELECT "public"."orders"."id" AS "id", "public"."orders"."user_id" AS "user_id", "public"."orders"."product_id" AS "product_id", "public"."orders"."subtotal" AS "subtotal", "public"."orders"."tax" AS "tax", "public"."orders"."total" AS "total", "public"."orders"."discount" AS "discount", "public"."orders"."created_at" AS "created_at", "public"."orders"."quantity" AS "quantity", abs(0) AS "pivot-grouping", "products__via__product_id"."vendor" AS "products__via__product_id__vendor", "products__via__product_id"."id" AS "products__via__product_id__id" FROM "public"."orders" LEFT JOIN "public"."products" "products__via__product_id" ON "public"."orders"."product_id" = "products__via__product_id"."id") "source" GROUP BY "source"."products__via__product_id__vendor", "source"."pivot-grouping" ORDER BY "source"."products__via__product_id__vendor" ASC, "source"."pivot-grouping" ASC
2022-12-28 22:47:57.547 UTC [85] LOG: execute S_2: ROLLBACK
2022-12-28 22:47:57.548 UTC [85] LOG: execute <unnamed>: SET SESSION CHARACTERISTICS AS TRANSACTION ISOLATION LEVEL READ COMMITTED
2022-12-28 22:47:57.562 UTC [85] LOG: execute <unnamed>: SET SESSION CHARACTERISTICS AS TRANSACTION ISOLATION LEVEL READ UNCOMMITTED
2022-12-28 22:47:57.562 UTC [85] LOG: execute <unnamed>: BEGIN READ ONLY
2022-12-28 22:47:57.562 UTC [85] LOG: execute <unnamed>/C_38: -- Metabase:: userID: 1 queryType: MBQL queryHash: cc962af2784c4fd2a9c2dbb673ec525d446a3be48e38dd513c8a82526c66bbe7
SELECT "source"."pivot-grouping" AS "pivot-grouping", sum("source"."quantity") AS "sum", sum("source"."total") AS "sum_2" FROM (SELECT "public"."orders"."id" AS "id", "public"."orders"."user_id" AS "user_id", "public"."orders"."product_id" AS "product_id", "public"."orders"."subtotal" AS "subtotal", "public"."orders"."tax" AS "tax", "public"."orders"."total" AS "total", "public"."orders"."discount" AS "discount", "public"."orders"."created_at" AS "created_at", "public"."orders"."quantity" AS "quantity", abs(1) AS "pivot-grouping" FROM "public"."orders") "source" GROUP BY "source"."pivot-grouping" ORDER BY "source"."pivot-grouping" ASC
```
This is with the ASC order
```
2022-12-28 22:51:28.834 UTC [85] LOG: execute <unnamed>: SET SESSION CHARACTERISTICS AS TRANSACTION ISOLATION LEVEL READ UNCOMMITTED
2022-12-28 22:51:28.834 UTC [85] LOG: execute <unnamed>: BEGIN READ ONLY
2022-12-28 22:51:28.834 UTC [85] LOG: execute <unnamed>/C_39: -- Metabase:: userID: 1 queryType: MBQL queryHash: b3f10d2564045d5b6c5006c90ca16ae5e70f22701cbaf0bacbcb67ac11e753f2
SELECT "source"."products__via__product_id__vendor" AS "products__via__product_id__vendor", "source"."pivot-grouping" AS "pivot-grouping", sum("source"."quantity") AS "sum", sum("source"."total") AS "sum_2" FROM (SELECT "public"."orders"."id" AS "id", "public"."orders"."user_id" AS "user_id", "public"."orders"."product_id" AS "product_id", "public"."orders"."subtotal" AS "subtotal", "public"."orders"."tax" AS "tax", "public"."orders"."total" AS "total", "public"."orders"."discount" AS "discount", "public"."orders"."created_at" AS "created_at", "public"."orders"."quantity" AS "quantity", abs(0) AS "pivot-grouping", "products__via__product_id"."vendor" AS "products__via__product_id__vendor", "products__via__product_id"."id" AS "products__via__product_id__id" FROM "public"."orders" LEFT JOIN "public"."products" "products__via__product_id" ON "public"."orders"."product_id" = "products__via__product_id"."id") "source" GROUP BY "source"."products__via__product_id__vendor", "source"."pivot-grouping" ORDER BY "source"."products__via__product_id__vendor" ASC, "source"."pivot-grouping" ASC
2022-12-28 22:51:28.856 UTC [85] LOG: execute S_2: ROLLBACK
2022-12-28 22:51:28.856 UTC [85] LOG: execute <unnamed>: SET SESSION CHARACTERISTICS AS TRANSACTION ISOLATION LEVEL READ COMMITTED
2022-12-28 22:51:28.876 UTC [85] LOG: execute <unnamed>: SET SESSION CHARACTERISTICS AS TRANSACTION ISOLATION LEVEL READ UNCOMMITTED
2022-12-28 22:51:28.876 UTC [85] LOG: execute <unnamed>: BEGIN READ ONLY
2022-12-28 22:51:28.876 UTC [85] LOG: execute <unnamed>/C_40: -- Metabase:: userID: 1 queryType: MBQL queryHash: cc962af2784c4fd2a9c2dbb673ec525d446a3be48e38dd513c8a82526c66bbe7
SELECT "source"."pivot-grouping" AS "pivot-grouping", sum("source"."quantity") AS "sum", sum("source"."total") AS "sum_2" FROM (SELECT "public"."orders"."id" AS "id", "public"."orders"."user_id" AS "user_id", "public"."orders"."product_id" AS "product_id", "public"."orders"."subtotal" AS "subtotal", "public"."orders"."tax" AS "tax", "public"."orders"."total" AS "total", "public"."orders"."discount" AS "discount", "public"."orders"."created_at" AS "created_at", "public"."orders"."quantity" AS "quantity", abs(1) AS "pivot-grouping" FROM "public"."orders") "source" GROUP BY "source"."pivot-grouping" ORDER BY "source"."pivot-grouping" ASC
```
**To Reproduce**
1) Do a sum of total and sum of quantity by product vendor with the orders table, visualize it as a pivot table.
2) Then change the order in the query builder and visualize it again, see that the order didn't change
**Expected behavior**
Order BY should be done in the first query and then passed to the last query. This is important as many customers have more than 10K series and the queries don't return with all data so they can't sort on the frontend
**Screenshots**
NA
**Information about your Metabase Installation:**
- Metabase version: 45.1
- Metabase hosting environment: Docker
- Metabase internal database: Postgres
**Severity**
P2
**Additional context**
Reported by a customer | process | pivot tables won t respect the order set up in the query builder describe the bug if you try to do a pivot table with an order clause it won t get respected since the pivot table sends several queries to the db and the last one the one that generates the final table does not bring the order by from the initial query it does something like order by source pivot grouping asc no matter the order you have seems to have been designed this way logs e g this is with desc order utc log execute set session characteristics as transaction isolation level read uncommitted utc log execute begin read only utc log execute c metabase userid querytype mbql queryhash select source products via product id vendor as products via product id vendor source pivot grouping as pivot grouping sum source quantity as sum sum source total as sum from select public orders id as id public orders user id as user id public orders product id as product id public orders subtotal as subtotal public orders tax as tax public orders total as total public orders discount as discount public orders created at as created at public orders quantity as quantity abs as pivot grouping products via product id vendor as products via product id vendor products via product id id as products via product id id from public orders left join public products products via product id on public orders product id products via product id id source group by source products via product id vendor source pivot grouping order by source products via product id vendor asc source pivot grouping asc utc log execute s rollback utc log execute set session characteristics as transaction isolation level read committed utc log execute set session characteristics as transaction isolation level read uncommitted utc log execute begin read only utc log execute c metabase userid querytype mbql queryhash select source pivot grouping as pivot grouping sum source quantity as sum sum source total as sum from select public orders id as id public orders user id as user id public orders product id as product id public orders subtotal as subtotal public orders tax as tax public orders total as total public orders discount as discount public orders created at as created at public orders quantity as quantity abs as pivot grouping from public orders source group by source pivot grouping order by source pivot grouping asc this is with the asc order utc log execute set session characteristics as transaction isolation level read uncommitted utc log execute begin read only utc log execute c metabase userid querytype mbql queryhash select source products via product id vendor as products via product id vendor source pivot grouping as pivot grouping sum source quantity as sum sum source total as sum from select public orders id as id public orders user id as user id public orders product id as product id public orders subtotal as subtotal public orders tax as tax public orders total as total public orders discount as discount public orders created at as created at public orders quantity as quantity abs as pivot grouping products via product id vendor as products via product id vendor products via product id id as products via product id id from public orders left join public products products via product id on public orders product id products via product id id source group by source products via product id vendor source pivot grouping order by source products via product id vendor asc source pivot grouping asc utc log execute s rollback utc log execute set session characteristics as transaction isolation level read committed utc log execute set session characteristics as transaction isolation level read uncommitted utc log execute begin read only utc log execute c metabase userid querytype mbql queryhash select source pivot grouping as pivot grouping sum source quantity as sum sum source total as sum from select public orders id as id public orders user id as user id public orders product id as product id public orders subtotal as subtotal public orders tax as tax public orders total as total public orders discount as discount public orders created at as created at public orders quantity as quantity abs as pivot grouping from public orders source group by source pivot grouping order by source pivot grouping asc to reproduce do a sum of total and sum of quantity by product vendor with the orders table visualize it as a pivot table then change the order in the query builder and visualize it again see that the order didn t change expected behavior order by should be done in the first query and then passed to the last query this is important as many customers have more than series and the queries don t return with all data so they can t sort on the frontend screenshots na information about your metabase installation metabase version metabase hosting environment docker metabase internal database postgres severity additional context reported by a customer | 1 |
694,643 | 23,822,634,892 | IssuesEvent | 2022-09-05 12:35:53 | Together-Java/TJ-Bot | https://api.github.com/repos/Together-Java/TJ-Bot | opened | Use slash-command mentions | enhancement good first issue priority: normal | ## Mentions
Discord recently released a new feature called [slash command mentions](https://netcord.site/mention-slash-commands-in-chat/).

The syntax for this is simple, its pure text in a specific format like `</ask:987249686383591447>` where the number is the ID of the command. The ID can be get in developer mode by right clicking the command when about to use it:

## Proposal
The bot should make use of this when writing messages that contain such mentions, for example when explaining the help system:


and possibly more. Its likely easy to identify with a text search of `/` in the code base.
## Details
When implementing, we should add utility to not have to write this syntax all the time. Similar to how we can do `author.getAsMention()`, we should have some sort of `DiscordUtils.mentionSlashCommand("ask", 987249686383591447)`.
We already have such an utility class somewhere, which can be reused for this. | 1.0 | Use slash-command mentions - ## Mentions
Discord recently released a new feature called [slash command mentions](https://netcord.site/mention-slash-commands-in-chat/).

The syntax for this is simple, its pure text in a specific format like `</ask:987249686383591447>` where the number is the ID of the command. The ID can be get in developer mode by right clicking the command when about to use it:

## Proposal
The bot should make use of this when writing messages that contain such mentions, for example when explaining the help system:


and possibly more. Its likely easy to identify with a text search of `/` in the code base.
## Details
When implementing, we should add utility to not have to write this syntax all the time. Similar to how we can do `author.getAsMention()`, we should have some sort of `DiscordUtils.mentionSlashCommand("ask", 987249686383591447)`.
We already have such an utility class somewhere, which can be reused for this. | non_process | use slash command mentions mentions discord recently released a new feature called the syntax for this is simple its pure text in a specific format like where the number is the id of the command the id can be get in developer mode by right clicking the command when about to use it proposal the bot should make use of this when writing messages that contain such mentions for example when explaining the help system and possibly more its likely easy to identify with a text search of in the code base details when implementing we should add utility to not have to write this syntax all the time similar to how we can do author getasmention we should have some sort of discordutils mentionslashcommand ask we already have such an utility class somewhere which can be reused for this | 0 |
215,231 | 16,656,890,069 | IssuesEvent | 2021-06-05 17:41:17 | Realm667/WolfenDoom | https://api.github.com/repos/Realm667/WolfenDoom | closed | Commander Keen actor problems | actor bug playtesting | - [x] CKSmirky can teleport into a wall and get stuck.

- [x] The slug poison flatsprite does not align to slopes.

Both could have been fixed already. | 1.0 | Commander Keen actor problems - - [x] CKSmirky can teleport into a wall and get stuck.

- [x] The slug poison flatsprite does not align to slopes.

Both could have been fixed already. | non_process | commander keen actor problems cksmirky can teleport into a wall and get stuck the slug poison flatsprite does not align to slopes both could have been fixed already | 0 |
6,353 | 9,413,531,986 | IssuesEvent | 2019-04-10 08:03:01 | tomwrobel/ora_ocfl | https://api.github.com/repos/tomwrobel/ora_ocfl | opened | Generate valid inventory.json.sha512 | !birkland/ocfl AS import/export process ocfl_spec | Currently, Aaron Birkland's code does not generate a valid inventory.json sha512 hash: the hash is present, but the filename is missing, e.g.
```
9cc7b2c8958ed9a2d7ec7e8571e2265046f9ce4993295fa42b1fe1d71489e5f97ea21cc36daeabbd43d5c08140f42464ee5c9aba8b0918bc537c9f8d49773107
```
not
```
inventory.json
``` | 1.0 | Generate valid inventory.json.sha512 - Currently, Aaron Birkland's code does not generate a valid inventory.json sha512 hash: the hash is present, but the filename is missing, e.g.
```
9cc7b2c8958ed9a2d7ec7e8571e2265046f9ce4993295fa42b1fe1d71489e5f97ea21cc36daeabbd43d5c08140f42464ee5c9aba8b0918bc537c9f8d49773107
```
not
```
inventory.json
``` | process | generate valid inventory json currently aaron birkland s code does not generate a valid inventory json hash the hash is present but the filename is missing e g not inventory json | 1 |
11,000 | 13,789,085,475 | IssuesEvent | 2020-10-09 08:19:43 | tikv/tikv | https://api.github.com/repos/tikv/tikv | reopened | copr: optimize ChunkedVecSized performance | sig/coprocessor type/enhancement | ## Development Task
https://github.com/tikv/tikv/pull/8503 adds benchmark for structures modified in Chunk-based Computing RFC. The result shows that `ChunkedVecSized<T>` is 2x slower than original `Vec<Option<T>>`. We should figure out what caused such regression, and optimize performance of `ChunkedVecSized`.
<img width="715" alt="image" src="https://user-images.githubusercontent.com/4198311/91149691-f88f4d00-e6ed-11ea-9ccc-e99b548d2247.png">
| 1.0 | copr: optimize ChunkedVecSized performance - ## Development Task
https://github.com/tikv/tikv/pull/8503 adds benchmark for structures modified in Chunk-based Computing RFC. The result shows that `ChunkedVecSized<T>` is 2x slower than original `Vec<Option<T>>`. We should figure out what caused such regression, and optimize performance of `ChunkedVecSized`.
<img width="715" alt="image" src="https://user-images.githubusercontent.com/4198311/91149691-f88f4d00-e6ed-11ea-9ccc-e99b548d2247.png">
| process | copr optimize chunkedvecsized performance development task adds benchmark for structures modified in chunk based computing rfc the result shows that chunkedvecsized is slower than original vec we should figure out what caused such regression and optimize performance of chunkedvecsized img width alt image src | 1 |
10,981 | 13,783,156,325 | IssuesEvent | 2020-10-08 18:45:44 | zaimoni/Cataclysm | https://api.github.com/repos/zaimoni/Cataclysm | opened | LLVM and GCC have intentionally broken std::variant -- no fix planned | formal correctness process | This theoretically affects the stability of the subtraction operator for global positioning locations. Further uses of std::variant should require a reasonably high bar. MSVC++ has no such problem.
We'll just have to very carefully test any use of this C++17 STL type, unfortunately.
https://bugs.llvm.org/show_bug.cgi?id=34632#c29 indexed by https://old.reddit.com/r/cpp/comments/j7gn2d/stdvariant_is_broken_in_clang_and_it_exposes_how/ . Breakage appears to be any optimization level above O0. Example code triggering incorrect compiling is provided.
Rationale for won't-fix is:
> Daniel Berlin 2017-10-23 09:03:44 PDT
> So, as a preliminary, i can break every test case GCC has fixed, because they are pretty simple workarounds that will still break. Richard knows this as well :)
>
> Instead of responding to your specific test cases, I'm going to give you two general rules of interpretation all the compilers i know of follow:
>
> 1. It must be possible to determine the effect of TBAA using only the information available in the current translation unit
> 2. Adding more translation units to the scope of what the compiler sees should not change the result.
>
> Compilers consider anything that violate this to be a bug in the standard or a bug in interpretation.
>
> Anything else creates either a terrible user experience (things work depending on how many files they are split across or how they are refactored), would not allow the use of TBAA in general, or requires additional, non-TBAA info to get anywhere by default (Your malloc example would require IPA points-to to determine whether a thing came from a malloc or not, for example).
>
> Rule #1 is in fact, the reason for the "visible union" rule that compilers follow - so that the result does not change when you split it into multiple files at function boundaries.
>
> ....
>
> However, TL;DR, if your memory model is that types can be changed at any time for allocated memory, in ways that are invisible to other translation units, TBAA is impossible to apply to anything by default without breaking some cases.
>
> I don't think such an interpretation (or standard) makes a lot of sense. If that's what C/C++ really want, they should remove effective type rules entirely, as they just create a mess.
(Above, TBAA := Type Based Alias Analysis.) | 1.0 | LLVM and GCC have intentionally broken std::variant -- no fix planned - This theoretically affects the stability of the subtraction operator for global positioning locations. Further uses of std::variant should require a reasonably high bar. MSVC++ has no such problem.
We'll just have to very carefully test any use of this C++17 STL type, unfortunately.
https://bugs.llvm.org/show_bug.cgi?id=34632#c29 indexed by https://old.reddit.com/r/cpp/comments/j7gn2d/stdvariant_is_broken_in_clang_and_it_exposes_how/ . Breakage appears to be any optimization level above O0. Example code triggering incorrect compiling is provided.
Rationale for won't-fix is:
> Daniel Berlin 2017-10-23 09:03:44 PDT
> So, as a preliminary, i can break every test case GCC has fixed, because they are pretty simple workarounds that will still break. Richard knows this as well :)
>
> Instead of responding to your specific test cases, I'm going to give you two general rules of interpretation all the compilers i know of follow:
>
> 1. It must be possible to determine the effect of TBAA using only the information available in the current translation unit
> 2. Adding more translation units to the scope of what the compiler sees should not change the result.
>
> Compilers consider anything that violate this to be a bug in the standard or a bug in interpretation.
>
> Anything else creates either a terrible user experience (things work depending on how many files they are split across or how they are refactored), would not allow the use of TBAA in general, or requires additional, non-TBAA info to get anywhere by default (Your malloc example would require IPA points-to to determine whether a thing came from a malloc or not, for example).
>
> Rule #1 is in fact, the reason for the "visible union" rule that compilers follow - so that the result does not change when you split it into multiple files at function boundaries.
>
> ....
>
> However, TL;DR, if your memory model is that types can be changed at any time for allocated memory, in ways that are invisible to other translation units, TBAA is impossible to apply to anything by default without breaking some cases.
>
> I don't think such an interpretation (or standard) makes a lot of sense. If that's what C/C++ really want, they should remove effective type rules entirely, as they just create a mess.
(Above, TBAA := Type Based Alias Analysis.) | process | llvm and gcc have intentionally broken std variant no fix planned this theoretically affects the stability of the subtraction operator for global positioning locations further uses of std variant should require a reasonably high bar msvc has no such problem we ll just have to very carefully test any use of this c stl type unfortunately indexed by breakage appears to be any optimization level above example code triggering incorrect compiling is provided rationale for won t fix is daniel berlin pdt so as a preliminary i can break every test case gcc has fixed because they are pretty simple workarounds that will still break richard knows this as well instead of responding to your specific test cases i m going to give you two general rules of interpretation all the compilers i know of follow it must be possible to determine the effect of tbaa using only the information available in the current translation unit adding more translation units to the scope of what the compiler sees should not change the result compilers consider anything that violate this to be a bug in the standard or a bug in interpretation anything else creates either a terrible user experience things work depending on how many files they are split across or how they are refactored would not allow the use of tbaa in general or requires additional non tbaa info to get anywhere by default your malloc example would require ipa points to to determine whether a thing came from a malloc or not for example rule is in fact the reason for the visible union rule that compilers follow so that the result does not change when you split it into multiple files at function boundaries however tl dr if your memory model is that types can be changed at any time for allocated memory in ways that are invisible to other translation units tbaa is impossible to apply to anything by default without breaking some cases i don t think such an interpretation or standard makes a lot of sense if that s what c c really want they should remove effective type rules entirely as they just create a mess above tbaa type based alias analysis | 1 |
10,790 | 13,608,997,645 | IssuesEvent | 2020-09-23 03:58:35 | googleapis/java-language | https://api.github.com/repos/googleapis/java-language | closed | Dependency Dashboard | api: language type: process | This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-project-info-reports-plugin-3.x -->build(deps): update dependency org.apache.maven.plugins:maven-project-info-reports-plugin to v3.1.1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-language-1.x -->chore(deps): update dependency com.google.cloud:google-cloud-language to v1.101.0
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
| 1.0 | Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-project-info-reports-plugin-3.x -->build(deps): update dependency org.apache.maven.plugins:maven-project-info-reports-plugin to v3.1.1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-language-1.x -->chore(deps): update dependency com.google.cloud:google-cloud-language to v1.101.0
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
| process | dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any build deps update dependency org apache maven plugins maven project info reports plugin to chore deps update dependency com google cloud google cloud language to check this box to trigger a request for renovate to run again on this repository | 1 |
62,026 | 15,144,672,513 | IssuesEvent | 2021-02-11 01:59:24 | Vyvy-vi/TearDrops | https://api.github.com/repos/Vyvy-vi/TearDrops | closed | [DOCS/IDEA]: Command listing automation | build v1.1+ documentation enhancement ideas-needed no-issue-activity status: on-hold | ## This might be a crazy idea
We can pick up the content of our `commands` and perhaps assign parameters to the command(using the `discord.py` commands class methods or by typing it out in the doc-strings with identifiers).
This would be linked with a GitHub Action, that reads that content and generates a .json file and uploads that to mongo perhaps and updates a .md file with the parsed content of that json file.
### But Why?
- #137 raises an issue about the help command becoming more advanced.
If we are making specific structures for making help command more detailed, it would be best if we also work on the documentation inside the repo at the same time. Just fetching the docs-data from the code-documentation after better help command is made, might make it easier to make detailed docs about the bot.
- If we sync it with a github action, keeping the repo docs up-to-date might be an easier task as making long detailed documentation for commands, isn't something we have time for and at present, there are not many new contributors who arrive for making docs to this repo.
- This fetched data stored on mongo, might help us in logging when a command was added and what was the description structure etc.
- The fetched data in json might be useful if later on, we decide to put detailed documentation on a site. | 1.0 | [DOCS/IDEA]: Command listing automation - ## This might be a crazy idea
We can pick up the content of our `commands` and perhaps assign parameters to the command(using the `discord.py` commands class methods or by typing it out in the doc-strings with identifiers).
This would be linked with a GitHub Action, that reads that content and generates a .json file and uploads that to mongo perhaps and updates a .md file with the parsed content of that json file.
### But Why?
- #137 raises an issue about the help command becoming more advanced.
If we are making specific structures for making help command more detailed, it would be best if we also work on the documentation inside the repo at the same time. Just fetching the docs-data from the code-documentation after better help command is made, might make it easier to make detailed docs about the bot.
- If we sync it with a github action, keeping the repo docs up-to-date might be an easier task as making long detailed documentation for commands, isn't something we have time for and at present, there are not many new contributors who arrive for making docs to this repo.
- This fetched data stored on mongo, might help us in logging when a command was added and what was the description structure etc.
- The fetched data in json might be useful if later on, we decide to put detailed documentation on a site. | non_process | command listing automation this might be a crazy idea we can pick up the content of our commands and perhaps assign parameters to the command using the discord py commands class methods or by typing it out in the doc strings with identifiers this would be linked with a github action that reads that content and generates a json file and uploads that to mongo perhaps and updates a md file with the parsed content of that json file but why raises an issue about the help command becoming more advanced if we are making specific structures for making help command more detailed it would be best if we also work on the documentation inside the repo at the same time just fetching the docs data from the code documentation after better help command is made might make it easier to make detailed docs about the bot if we sync it with a github action keeping the repo docs up to date might be an easier task as making long detailed documentation for commands isn t something we have time for and at present there are not many new contributors who arrive for making docs to this repo this fetched data stored on mongo might help us in logging when a command was added and what was the description structure etc the fetched data in json might be useful if later on we decide to put detailed documentation on a site | 0 |
19,476 | 25,787,797,691 | IssuesEvent | 2022-12-09 22:37:13 | dtcenter/MET | https://api.github.com/repos/dtcenter/MET | closed | Add support for EPA AirNow ASCII data in ASCII2NC | type: new feature requestor: NOAA/other reporting: DTC NOAA R2O required: FOR DEVELOPMENT RELEASE MET: PreProcessing Tools (Point) priority: high | ## Describe the New Feature ##
Per dtcenter/METplus#1515 and @PerryShafran-NOAA, the EPA will switch from providing BUFR to providing ASCII data to NOAA. The new feature is to add support for this new dataset to ASCII2NC.
### Acceptance Testing ###
Sample files exist on seneca here: /home/dadriaan/projects/airnow/shafran_data/
From dtcenter/METplus#1515:
> There are several files here to look at, but they both provide the same data. The file with all the data is the HourlyAQObs data. The HourlyData is a neater looking file with the same data, but the lat/lon of each station is not provided there, so you'll have to use the HourlyAQObs file, which is a CSV file.
>
> It would be great to be able to also read in the daily file, which I think we can use to validate daily model items. That is the daily_data.dat file, and there are two of them. Unfortunately, they are in the HourlyData format, but you should be able to use the lat/lon of the HourlyAQObs file to find these stations. Here's the fact sheet on the daily file:
>
> As you can see, they don't make this easy. However, we could just stick with hourly data and just sum/average them the way we are doing now in PB2NC.
There are four file types:
1) "HourlyAQObs" (Docs: [HourlyAQObsFactSheet.pdf](https://github.com/dtcenter/MET/files/8514707/HourlyAQObsFactSheet.1.pdf))
2) "HourlyData" (Docs: [HourlyDataFactSheet.pdf](https://github.com/dtcenter/MET/files/8514896/HourlyDataFactSheet.1.pdf))
3) "daily_data" (Docs: unknown)
4) "daily_data_v2" (Docs: [DailyDataFactSheet.pdf](https://github.com/dtcenter/MET/files/8514899/DailyDataFactSheet.1.pdf))
The documentation for file type 4) only describes "daily_data_v2", and no information was provided about the "daily_data" file from the user. @JohnHalleyGotway notes:
> ...looking at a diff of the two, I recommend using daily_data_v2.dat. The former does NOT contain lat/lon info for each site and would require a lookup table for that. The v2 version DOES contain the lat/lon info (it would seems) and would therefore be simpler to process.
### Time Estimate ###
1 day of work
### Sub-Issues ###
- [ ] Document different file type metadata
- [ ] Add support for reading these files in ASCII2NC
- [ ] Test support
### Relevant Deadlines ###
NONE.
### Funding Source ###
2792541
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose): need to update ASCII2NC wrapper
## New Feature Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [x] Complete the issue definition above, including the **Time Estimate** and **Funding source**.
- [x] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| 1.0 | Add support for EPA AirNow ASCII data in ASCII2NC - ## Describe the New Feature ##
Per dtcenter/METplus#1515 and @PerryShafran-NOAA, the EPA will switch from providing BUFR to providing ASCII data to NOAA. The new feature is to add support for this new dataset to ASCII2NC.
### Acceptance Testing ###
Sample files exist on seneca here: /home/dadriaan/projects/airnow/shafran_data/
From dtcenter/METplus#1515:
> There are several files here to look at, but they both provide the same data. The file with all the data is the HourlyAQObs data. The HourlyData is a neater looking file with the same data, but the lat/lon of each station is not provided there, so you'll have to use the HourlyAQObs file, which is a CSV file.
>
> It would be great to be able to also read in the daily file, which I think we can use to validate daily model items. That is the daily_data.dat file, and there are two of them. Unfortunately, they are in the HourlyData format, but you should be able to use the lat/lon of the HourlyAQObs file to find these stations. Here's the fact sheet on the daily file:
>
> As you can see, they don't make this easy. However, we could just stick with hourly data and just sum/average them the way we are doing now in PB2NC.
There are four file types:
1) "HourlyAQObs" (Docs: [HourlyAQObsFactSheet.pdf](https://github.com/dtcenter/MET/files/8514707/HourlyAQObsFactSheet.1.pdf))
2) "HourlyData" (Docs: [HourlyDataFactSheet.pdf](https://github.com/dtcenter/MET/files/8514896/HourlyDataFactSheet.1.pdf))
3) "daily_data" (Docs: unknown)
4) "daily_data_v2" (Docs: [DailyDataFactSheet.pdf](https://github.com/dtcenter/MET/files/8514899/DailyDataFactSheet.1.pdf))
The documentation for file type 4) only describes "daily_data_v2", and no information was provided about the "daily_data" file from the user. @JohnHalleyGotway notes:
> ...looking at a diff of the two, I recommend using daily_data_v2.dat. The former does NOT contain lat/lon info for each site and would require a lookup table for that. The v2 version DOES contain the lat/lon info (it would seems) and would therefore be simpler to process.
### Time Estimate ###
1 day of work
### Sub-Issues ###
- [ ] Document different file type metadata
- [ ] Add support for reading these files in ASCII2NC
- [ ] Test support
### Relevant Deadlines ###
NONE.
### Funding Source ###
2792541
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose): need to update ASCII2NC wrapper
## New Feature Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [x] Complete the issue definition above, including the **Time Estimate** and **Funding source**.
- [x] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| process | add support for epa airnow ascii data in describe the new feature per dtcenter metplus and perryshafran noaa the epa will switch from providing bufr to providing ascii data to noaa the new feature is to add support for this new dataset to acceptance testing sample files exist on seneca here home dadriaan projects airnow shafran data from dtcenter metplus there are several files here to look at but they both provide the same data the file with all the data is the hourlyaqobs data the hourlydata is a neater looking file with the same data but the lat lon of each station is not provided there so you ll have to use the hourlyaqobs file which is a csv file it would be great to be able to also read in the daily file which i think we can use to validate daily model items that is the daily data dat file and there are two of them unfortunately they are in the hourlydata format but you should be able to use the lat lon of the hourlyaqobs file to find these stations here s the fact sheet on the daily file as you can see they don t make this easy however we could just stick with hourly data and just sum average them the way we are doing now in there are four file types hourlyaqobs docs hourlydata docs daily data docs unknown daily data docs the documentation for file type only describes daily data and no information was provided about the daily data file from the user johnhalleygotway notes looking at a diff of the two i recommend using daily data dat the former does not contain lat lon info for each site and would require a lookup table for that the version does contain the lat lon info it would seems and would therefore be simpler to process time estimate day of work sub issues document different file type metadata add support for reading these files in test support relevant deadlines none funding source define related issue s consider the impact to the other metplus components need to update wrapper new feature checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue | 1 |
240,681 | 20,070,191,981 | IssuesEvent | 2022-02-04 05:12:13 | storj/gateway-mt | https://api.github.com/repos/storj/gateway-mt | opened | missing content-length header produces different error response to S3 | s3-tests | One of the [s3-tests produces a failure](https://github.com/storj/splunk-s3-tests/blob/main/s3tests/functional/test_headers.py#L346). The test expected 411 Length Required when Content-Length header was omitted from the request, but it returns 400 MD5 mismatch error, a bit of a red herring.
After looking at the code, minio `PutObjectHandler` checks `r.ContentLength` from `http.Request`, but this is determined by the content body, even if the `Content-Length` header is missing from the request. It is set as `-1` if the request was HTTP/2, according to the code in `readRequest` in the go http package. In the s3 test, it was observed `r.ContentLength` was `0` with the missing header.
A suggested fix is to check Content-Length header was presented in the request in minio handlers that are checking for `size == -1`, or change the size check to `<= 0`. I'm not entirely sure of the implications of the latter change though. | 1.0 | missing content-length header produces different error response to S3 - One of the [s3-tests produces a failure](https://github.com/storj/splunk-s3-tests/blob/main/s3tests/functional/test_headers.py#L346). The test expected 411 Length Required when Content-Length header was omitted from the request, but it returns 400 MD5 mismatch error, a bit of a red herring.
After looking at the code, minio `PutObjectHandler` checks `r.ContentLength` from `http.Request`, but this is determined by the content body, even if the `Content-Length` header is missing from the request. It is set as `-1` if the request was HTTP/2, according to the code in `readRequest` in the go http package. In the s3 test, it was observed `r.ContentLength` was `0` with the missing header.
A suggested fix is to check Content-Length header was presented in the request in minio handlers that are checking for `size == -1`, or change the size check to `<= 0`. I'm not entirely sure of the implications of the latter change though. | non_process | missing content length header produces different error response to one of the the test expected length required when content length header was omitted from the request but it returns mismatch error a bit of a red herring after looking at the code minio putobjecthandler checks r contentlength from http request but this is determined by the content body even if the content length header is missing from the request it is set as if the request was http according to the code in readrequest in the go http package in the test it was observed r contentlength was with the missing header a suggested fix is to check content length header was presented in the request in minio handlers that are checking for size or change the size check to i m not entirely sure of the implications of the latter change though | 0 |
216,522 | 7,309,036,238 | IssuesEvent | 2018-02-28 10:23:27 | wso2/product-is | https://api.github.com/repos/wso2/product-is | opened | Location given for -d parameter in forget me tool doc is not valid | Affected/5.5.0-Alpha2 Priority/High Type/Docs | Location given for -d parameter in forget me tool doc [1] is not valid
[1] https://docs.wso2.com/display/IS550/Removing+References+to+Deleted+User+Identities
It says below :
d | The configuration directory to use when the tool is run. If you do not specify a value for this option, thedefault conf directory will be used. | No | -d /users/john/forgetme/config
/users/john/forgetme/config is not a valid location
| 1.0 | Location given for -d parameter in forget me tool doc is not valid - Location given for -d parameter in forget me tool doc [1] is not valid
[1] https://docs.wso2.com/display/IS550/Removing+References+to+Deleted+User+Identities
It says below :
d | The configuration directory to use when the tool is run. If you do not specify a value for this option, thedefault conf directory will be used. | No | -d /users/john/forgetme/config
/users/john/forgetme/config is not a valid location
| non_process | location given for d parameter in forget me tool doc is not valid location given for d parameter in forget me tool doc is not valid it says below d the configuration directory to use when the tool is run if you do not specify a value for this option thedefault conf directory will be used no d users john forgetme config users john forgetme config is not a valid location | 0 |
77,412 | 26,979,160,010 | IssuesEvent | 2023-02-09 11:49:09 | vector-im/element-call | https://api.github.com/repos/vector-im/element-call | opened | NPE during call hangup | T-Defect | ### Steps to reproduce
Unknown, but this was when a call to a device failed and we tried to dispose the call.
### Outcome
#### What did you expect?
#### What happened instead?
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Will you send logs?
Yes | 1.0 | NPE during call hangup - ### Steps to reproduce
Unknown, but this was when a call to a device failed and we tried to dispose the call.
### Outcome
#### What did you expect?
#### What happened instead?
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Will you send logs?
Yes | non_process | npe during call hangup steps to reproduce unknown but this was when a call to a device failed and we tried to dispose the call outcome what did you expect what happened instead operating system no response browser information no response url for webapp no response will you send logs yes | 0 |
76,892 | 15,496,226,610 | IssuesEvent | 2021-03-11 02:17:24 | n-devs/example-scripts-react | https://api.github.com/repos/n-devs/example-scripts-react | opened | CVE-2019-20920 (High) detected in handlebars-4.1.2.tgz | security vulnerability | ## CVE-2019-20920 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.2.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.2.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.2.tgz</a></p>
<p>Path to dependency file: /example-scripts-react/package.json</p>
<p>Path to vulnerable library: example-scripts-react/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- jest-24.7.1.tgz (Root Library)
- jest-cli-24.8.0.tgz
- core-24.8.0.tgz
- reporters-24.8.0.tgz
- istanbul-reports-2.2.6.tgz
- :x: **handlebars-4.1.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Handlebars before 3.0.8 and 4.x before 4.5.3 is vulnerable to Arbitrary Code Execution. The lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript. This can be used to run arbitrary code on a server processing Handlebars templates or in a victim's browser (effectively serving as XSS).
<p>Publish Date: 2020-09-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20920>CVE-2019-20920</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1324">https://www.npmjs.com/advisories/1324</a></p>
<p>Release Date: 2020-10-15</p>
<p>Fix Resolution: handlebars - 4.5.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-20920 (High) detected in handlebars-4.1.2.tgz - ## CVE-2019-20920 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.2.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.2.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.2.tgz</a></p>
<p>Path to dependency file: /example-scripts-react/package.json</p>
<p>Path to vulnerable library: example-scripts-react/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- jest-24.7.1.tgz (Root Library)
- jest-cli-24.8.0.tgz
- core-24.8.0.tgz
- reporters-24.8.0.tgz
- istanbul-reports-2.2.6.tgz
- :x: **handlebars-4.1.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Handlebars before 3.0.8 and 4.x before 4.5.3 is vulnerable to Arbitrary Code Execution. The lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript. This can be used to run arbitrary code on a server processing Handlebars templates or in a victim's browser (effectively serving as XSS).
<p>Publish Date: 2020-09-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20920>CVE-2019-20920</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1324">https://www.npmjs.com/advisories/1324</a></p>
<p>Release Date: 2020-10-15</p>
<p>Fix Resolution: handlebars - 4.5.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in handlebars tgz cve high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file example scripts react package json path to vulnerable library example scripts react node modules handlebars package json dependency hierarchy jest tgz root library jest cli tgz core tgz reporters tgz istanbul reports tgz x handlebars tgz vulnerable library vulnerability details handlebars before and x before is vulnerable to arbitrary code execution the lookup helper fails to properly validate templates allowing attackers to submit templates that execute arbitrary javascript this can be used to run arbitrary code on a server processing handlebars templates or in a victim s browser effectively serving as xss publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars step up your open source security game with whitesource | 0 |
262,949 | 23,026,358,660 | IssuesEvent | 2022-07-22 09:34:52 | redpanda-data/redpanda | https://api.github.com/repos/redpanda-data/redpanda | opened | v22.2.1-rc1 manual cloud testing | kind/enhance area/redpanda area/tests | Initial test plan:
- [ ] deploy in cloud (BYOC and/or FMC) - see https://vectorizedio.atlassian.net/wiki/spaces/CORE/pages/195199013/Testing+in+BYOC
- [ ] run basic client workload
- [ ] stress testing | 1.0 | v22.2.1-rc1 manual cloud testing - Initial test plan:
- [ ] deploy in cloud (BYOC and/or FMC) - see https://vectorizedio.atlassian.net/wiki/spaces/CORE/pages/195199013/Testing+in+BYOC
- [ ] run basic client workload
- [ ] stress testing | non_process | manual cloud testing initial test plan deploy in cloud byoc and or fmc see run basic client workload stress testing | 0 |
17,771 | 23,698,777,353 | IssuesEvent | 2022-08-29 16:56:06 | googleapis/cloud-trace-nodejs | https://api.github.com/repos/googleapis/cloud-trace-nodejs | closed | Tests for PostgreSQL and some versions of Hapi should be re-enabled or deleted | type: process api: cloudtrace | The tests for PostgreSQL and for some versions of Hapi have been turned off, because they were not running on Node 14 and Node 15.
They should either be removed completely, or fixed. | 1.0 | Tests for PostgreSQL and some versions of Hapi should be re-enabled or deleted - The tests for PostgreSQL and for some versions of Hapi have been turned off, because they were not running on Node 14 and Node 15.
They should either be removed completely, or fixed. | process | tests for postgresql and some versions of hapi should be re enabled or deleted the tests for postgresql and for some versions of hapi have been turned off because they were not running on node and node they should either be removed completely or fixed | 1 |
4,690 | 7,526,143,257 | IssuesEvent | 2018-04-13 13:09:12 | Activiti/Activiti | https://api.github.com/repos/Activiti/Activiti | opened | Process Instances Model is not working with the Accept Image/svg | blocking process | When I call the api
http://{{domain}}/myapp-rb/v1/process-instances/4159a15f-3f1b-11e8-aa83-0a586460031e/model
Accept image/svg+xml
I'm getting a 406 error.
The API is working fine if I have in the header
Content-Type image/svg+xm
| 1.0 | Process Instances Model is not working with the Accept Image/svg - When I call the api
http://{{domain}}/myapp-rb/v1/process-instances/4159a15f-3f1b-11e8-aa83-0a586460031e/model
Accept image/svg+xml
I'm getting a 406 error.
The API is working fine if I have in the header
Content-Type image/svg+xm
| process | process instances model is not working with the accept image svg when i call the api accept image svg xml i m getting a error the api is working fine if i have in the header content type image svg xm | 1 |
246,078 | 20,821,182,636 | IssuesEvent | 2022-03-18 15:31:27 | phetsims/chains | https://api.github.com/repos/phetsims/chains | opened | CT protocol error Target closed | type:automated-testing | ```
chains : build
Build failed with status code 3:
Running "report-media" task
Running "clean" task
Running "build" task
>> TypeScript compilation complete: 985ms
10:01:27 PM, 452 ms: ../chains/js/chains/ChainsScreen.js
10:01:27 PM, 63 ms: ../chains/js/chains/view/ChainsView.js
10:01:27 PM, 34 ms: ../chains/js/chains-main.js
10:01:27 PM, 24 ms: ../chains/js/chains-phet-io-overrides.js
10:01:27 PM, 23 ms: ../chains/js/chains.js
10:01:27 PM, 62 ms: ../chains/js/chainsStrings.ts
Building runnable repository (chains, brands: phet, phet-io)
Building brand: phet
>> Webpack build complete: 7795ms
>> Production minification complete: 44033ms (1656901 bytes)
>> Debug minification complete: 0ms (18125771 bytes)
Building brand: phet-io
>> Webpack build complete: 6368ms
>> Production minification complete: 38518ms (1688812 bytes)
>> Debug minification complete: 39620ms (1992105 bytes)
>> No client guides found at ../phet-io-client-guides/chains/, no guides being built.
>> TypeScript compilation complete: 1010ms
Fatal error: Protocol error (Target.closeTarget): Target closed.
Snapshot from 3/17/2022, 8:34:38 PM
``` | 1.0 | CT protocol error Target closed - ```
chains : build
Build failed with status code 3:
Running "report-media" task
Running "clean" task
Running "build" task
>> TypeScript compilation complete: 985ms
10:01:27 PM, 452 ms: ../chains/js/chains/ChainsScreen.js
10:01:27 PM, 63 ms: ../chains/js/chains/view/ChainsView.js
10:01:27 PM, 34 ms: ../chains/js/chains-main.js
10:01:27 PM, 24 ms: ../chains/js/chains-phet-io-overrides.js
10:01:27 PM, 23 ms: ../chains/js/chains.js
10:01:27 PM, 62 ms: ../chains/js/chainsStrings.ts
Building runnable repository (chains, brands: phet, phet-io)
Building brand: phet
>> Webpack build complete: 7795ms
>> Production minification complete: 44033ms (1656901 bytes)
>> Debug minification complete: 0ms (18125771 bytes)
Building brand: phet-io
>> Webpack build complete: 6368ms
>> Production minification complete: 38518ms (1688812 bytes)
>> Debug minification complete: 39620ms (1992105 bytes)
>> No client guides found at ../phet-io-client-guides/chains/, no guides being built.
>> TypeScript compilation complete: 1010ms
Fatal error: Protocol error (Target.closeTarget): Target closed.
Snapshot from 3/17/2022, 8:34:38 PM
``` | non_process | ct protocol error target closed chains build build failed with status code running report media task running clean task running build task typescript compilation complete pm ms chains js chains chainsscreen js pm ms chains js chains view chainsview js pm ms chains js chains main js pm ms chains js chains phet io overrides js pm ms chains js chains js pm ms chains js chainsstrings ts building runnable repository chains brands phet phet io building brand phet webpack build complete production minification complete bytes debug minification complete bytes building brand phet io webpack build complete production minification complete bytes debug minification complete bytes no client guides found at phet io client guides chains no guides being built typescript compilation complete fatal error protocol error target closetarget target closed snapshot from pm | 0 |
546,854 | 16,020,342,279 | IssuesEvent | 2021-04-20 21:55:32 | Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2 | https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2 | opened | Сartography Inaccuracies | :exclamation: priority high :question: suggestion :question: cartography | <!--
DO NOT REMOVE PRE-EXISTING LINES
IF YOU WANT TO SUGGEST A FEW THINGS, OPEN A NEW ISSUE PER EVERY SUGGESTION
----------------------------------------------------------------------------------------------------------
-->
**Describe your suggestion in full detail below:**
So it's time to list all inaccuracies that should be fixed.
- [ ] This mountain range. If it's inspired by WoW, its position isn't right.
<details>
<summary>Click to expand</summary>


</details>
- [ ] No trees in Eastern Lordaeron and Quel'thalas.
- [ ] Gilneas county has a weird texture.
<details>
<summary>Click to expand</summary>

</details>
- [ ] Blackwald has rock texture.
<details>
<summary>Click to expand</summary>

</details>
- [ ] Not existing mountain range in Azshara.
<details>
<summary>Click to expand</summary>

</details>
- [ ] No trees in Ashenvale.
- [ ] No trees in Hyjal.
- [ ] No trees in Northrend.
- [ ] Northrend has weird textures in many places. Like rock texture here.
<details>
<summary>Click to expand</summary>

</details>
- [ ] Not enough jungle trees in Krasarang. | 1.0 | Сartography Inaccuracies - <!--
DO NOT REMOVE PRE-EXISTING LINES
IF YOU WANT TO SUGGEST A FEW THINGS, OPEN A NEW ISSUE PER EVERY SUGGESTION
----------------------------------------------------------------------------------------------------------
-->
**Describe your suggestion in full detail below:**
So it's time to list all inaccuracies that should be fixed.
- [ ] This mountain range. If it's inspired by WoW, its position isn't right.
<details>
<summary>Click to expand</summary>


</details>
- [ ] No trees in Eastern Lordaeron and Quel'thalas.
- [ ] Gilneas county has a weird texture.
<details>
<summary>Click to expand</summary>

</details>
- [ ] Blackwald has rock texture.
<details>
<summary>Click to expand</summary>

</details>
- [ ] Not existing mountain range in Azshara.
<details>
<summary>Click to expand</summary>

</details>
- [ ] No trees in Ashenvale.
- [ ] No trees in Hyjal.
- [ ] No trees in Northrend.
- [ ] Northrend has weird textures in many places. Like rock texture here.
<details>
<summary>Click to expand</summary>

</details>
- [ ] Not enough jungle trees in Krasarang. | non_process | сartography inaccuracies do not remove pre existing lines if you want to suggest a few things open a new issue per every suggestion describe your suggestion in full detail below so it s time to list all inaccuracies that should be fixed this mountain range if it s inspired by wow its position isn t right click to expand no trees in eastern lordaeron and quel thalas gilneas county has a weird texture click to expand blackwald has rock texture click to expand not existing mountain range in azshara click to expand no trees in ashenvale no trees in hyjal no trees in northrend northrend has weird textures in many places like rock texture here click to expand not enough jungle trees in krasarang | 0 |
19,208 | 25,340,923,742 | IssuesEvent | 2022-11-18 21:31:56 | opensearch-project/data-prepper | https://api.github.com/repos/opensearch-project/data-prepper | closed | Aggregate only some events | enhancement plugin - processor | **Is your feature request related to a problem? Please describe.**
Support aggregating only some events meeting a certain value.
**Describe the solution you'd like**
Provide an `aggregate_when` configuration to the `aggregate` processor.
**Describe alternatives you've considered (Optional)**
* Support `when` on actions, but this requires any action needing this to support a `when` configuration.
* Processor-based when statements. This is not currently available.
**Additional context**
This is implemented by #2018
| 1.0 | Aggregate only some events - **Is your feature request related to a problem? Please describe.**
Support aggregating only some events meeting a certain value.
**Describe the solution you'd like**
Provide an `aggregate_when` configuration to the `aggregate` processor.
**Describe alternatives you've considered (Optional)**
* Support `when` on actions, but this requires any action needing this to support a `when` configuration.
* Processor-based when statements. This is not currently available.
**Additional context**
This is implemented by #2018
| process | aggregate only some events is your feature request related to a problem please describe support aggregating only some events meeting a certain value describe the solution you d like provide an aggregate when configuration to the aggregate processor describe alternatives you ve considered optional support when on actions but this requires any action needing this to support a when configuration processor based when statements this is not currently available additional context this is implemented by | 1 |
4,354 | 7,260,288,103 | IssuesEvent | 2018-02-18 07:41:43 | MobileOrg/mobileorg | https://api.github.com/repos/MobileOrg/mobileorg | closed | Is the use of a crash reporter worth the effort? | development process question | The last bugs reported by users were somehow hard to track down. We can't ask the user for their org files. Therefore to have access to crash reports might help. As it seems there were no crash reports provided within the Apple world (TestFlight, Xcode, ...).
The use of an external service might help here. I stumbled across [Hockey App](https://hockeyapp.net) which is free for up to two Apps. There might be others. | 1.0 | Is the use of a crash reporter worth the effort? - The last bugs reported by users were somehow hard to track down. We can't ask the user for their org files. Therefore to have access to crash reports might help. As it seems there were no crash reports provided within the Apple world (TestFlight, Xcode, ...).
The use of an external service might help here. I stumbled across [Hockey App](https://hockeyapp.net) which is free for up to two Apps. There might be others. | process | is the use of a crash reporter worth the effort the last bugs reported by users were somehow hard to track down we can t ask the user for their org files therefore to have access to crash reports might help as it seems there were no crash reports provided within the apple world testflight xcode the use of an external service might help here i stumbled across which is free for up to two apps there might be others | 1 |
645,197 | 20,997,707,676 | IssuesEvent | 2022-03-29 14:45:25 | thoth-station/user-api | https://api.github.com/repos/thoth-station/user-api | closed | Remove `count` and `limit` from advise endpoint | kind/feature priority/important-longterm sig/user-experience triage/accepted | **Is your feature request related to a problem? Please describe.**
As per discussion in the tech talk, we could drop these two parameters on the advise endpoint. They should be controlled based on the backend configuration rather than allowing users to supply them to the resolution process.
**Describe the solution you'd like**
Remove `count` and `limit` from the advise endpoint. Also, it would be great to make sure that removing these values does not affect scheduling adviser in a deployment negatively (defaults are properly used).
| 1.0 | Remove `count` and `limit` from advise endpoint - **Is your feature request related to a problem? Please describe.**
As per discussion in the tech talk, we could drop these two parameters on the advise endpoint. They should be controlled based on the backend configuration rather than allowing users to supply them to the resolution process.
**Describe the solution you'd like**
Remove `count` and `limit` from the advise endpoint. Also, it would be great to make sure that removing these values does not affect scheduling adviser in a deployment negatively (defaults are properly used).
| non_process | remove count and limit from advise endpoint is your feature request related to a problem please describe as per discussion in the tech talk we could drop these two parameters on the advise endpoint they should be controlled based on the backend configuration rather than allowing users to supply them to the resolution process describe the solution you d like remove count and limit from the advise endpoint also it would be great to make sure that removing these values does not affect scheduling adviser in a deployment negatively defaults are properly used | 0 |
153,342 | 5,890,005,933 | IssuesEvent | 2017-05-17 14:08:13 | nim-lang/Nim | https://api.github.com/repos/nim-lang/Nim | closed | LibreSSL isn't recognized as legit SSL library | High Priority Stdlib | When running Nimble on a distro with LibreSSL installed in lieu of OpenSSL (like on Void Linux), Nim doesn't seem to recognize it and asks for another versions.
```
$ nimble
could not load: libcrypto.so(|.10|.1.0.1|.1.0.0|.0.9.9|.0.9.8)
```
```
$ nimble
could not load: libssl.so(|.10|.1.0.1|.1.0.0|.0.9.9|.0.9.8)
```
Current and dirty fixes are the creation of symbolic links:
```
$ ln -s /usr/lib/libcrypto.so.38 /usr/lib/libcrypto.so.10
$ ln -s /usr/lib/libssl.so.39 /usr/lib/libssl.so.10
```
| 1.0 | LibreSSL isn't recognized as legit SSL library - When running Nimble on a distro with LibreSSL installed in lieu of OpenSSL (like on Void Linux), Nim doesn't seem to recognize it and asks for another versions.
```
$ nimble
could not load: libcrypto.so(|.10|.1.0.1|.1.0.0|.0.9.9|.0.9.8)
```
```
$ nimble
could not load: libssl.so(|.10|.1.0.1|.1.0.0|.0.9.9|.0.9.8)
```
Current and dirty fixes are the creation of symbolic links:
```
$ ln -s /usr/lib/libcrypto.so.38 /usr/lib/libcrypto.so.10
$ ln -s /usr/lib/libssl.so.39 /usr/lib/libssl.so.10
```
| non_process | libressl isn t recognized as legit ssl library when running nimble on a distro with libressl installed in lieu of openssl like on void linux nim doesn t seem to recognize it and asks for another versions nimble could not load libcrypto so nimble could not load libssl so current and dirty fixes are the creation of symbolic links ln s usr lib libcrypto so usr lib libcrypto so ln s usr lib libssl so usr lib libssl so | 0 |
2,218 | 2,603,991,278 | IssuesEvent | 2015-02-24 19:06:41 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳疱疹的治疗办法 | auto-migrated Priority-Medium Type-Defect | ```
沈阳疱疹的治疗办法〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:31 | 1.0 | 沈阳疱疹的治疗办法 - ```
沈阳疱疹的治疗办法〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:31 | non_process | 沈阳疱疹的治疗办法 沈阳疱疹的治疗办法〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位� �� 。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 original issue reported on code google com by gmail com on jun at | 0 |
3,118 | 6,149,509,212 | IssuesEvent | 2017-06-27 20:17:38 | phpDocumentor/phpDocumentor2 | https://api.github.com/repos/phpDocumentor/phpDocumentor2 | closed | Fatal error: Failed opening required "dompdf_config.inc.php" in Bootstrap.php on line 178 | Bug Release process | Hi
I've installed phpDocumnetor2 using pearl I a got a success message "install ok: channel://pear.phpdoc.org/phpDocumentor-2.9.0" but when I put in the console phpdoc I got the following error:
PHP Warning: require_once(C:\xampp\htdocs\siselc-thomas\vendor/dompdf/dompdf/dompdf_config.inc.php): failed to open stream: No such file or directory in C:\xampp\php\pear\phpDocumentor\src\phpDocumentor\Bootstrap.php on line 178
Warning: require_once(C:\xampp\htdocs\siselc-thomas\vendor/dompdf/dompdf/dompdf_config.inc.php): failed to open stream: No such file or directory in C:\xampp\php\pear\phpDocumentor\src\phpDocumentor\Bootstrap.php on line 178
PHP Fatal error: require_once(): Failed opening required 'C:\xampp\htdocs\siselc-thomas\vendor/dompdf/dompdf/dompdf_config.inc.php' (include_path='C:\xampp\php\PEAR') in C:\xampp\php\pear\phpDocumentor\src\phpDocumentor\Bootstrap.php on line 178
Fatal error: require_once(): Failed opening required 'C:\xampp\htdocs\siselc-thomas\vendor/dompdf/dompdf/dompdf_config.inc.php' (include_path='C:\xampp\php\PEAR') in C:\xampp\php\pear\phpDocumentor\src\phpDocumentor\Bootstrap.php on line 178
I am not an expert in this son I tried to install DOMPDF, but I am still getting the same error.
Thanks a lot | 1.0 | Fatal error: Failed opening required "dompdf_config.inc.php" in Bootstrap.php on line 178 - Hi
I've installed phpDocumnetor2 using pearl I a got a success message "install ok: channel://pear.phpdoc.org/phpDocumentor-2.9.0" but when I put in the console phpdoc I got the following error:
PHP Warning: require_once(C:\xampp\htdocs\siselc-thomas\vendor/dompdf/dompdf/dompdf_config.inc.php): failed to open stream: No such file or directory in C:\xampp\php\pear\phpDocumentor\src\phpDocumentor\Bootstrap.php on line 178
Warning: require_once(C:\xampp\htdocs\siselc-thomas\vendor/dompdf/dompdf/dompdf_config.inc.php): failed to open stream: No such file or directory in C:\xampp\php\pear\phpDocumentor\src\phpDocumentor\Bootstrap.php on line 178
PHP Fatal error: require_once(): Failed opening required 'C:\xampp\htdocs\siselc-thomas\vendor/dompdf/dompdf/dompdf_config.inc.php' (include_path='C:\xampp\php\PEAR') in C:\xampp\php\pear\phpDocumentor\src\phpDocumentor\Bootstrap.php on line 178
Fatal error: require_once(): Failed opening required 'C:\xampp\htdocs\siselc-thomas\vendor/dompdf/dompdf/dompdf_config.inc.php' (include_path='C:\xampp\php\PEAR') in C:\xampp\php\pear\phpDocumentor\src\phpDocumentor\Bootstrap.php on line 178
I am not an expert in this son I tried to install DOMPDF, but I am still getting the same error.
Thanks a lot | process | fatal error failed opening required dompdf config inc php in bootstrap php on line hi i ve installed using pearl i a got a success message install ok channel pear phpdoc org phpdocumentor but when i put in the console phpdoc i got the following error php warning require once c xampp htdocs siselc thomas vendor dompdf dompdf dompdf config inc php failed to open stream no such file or directory in c xampp php pear phpdocumentor src phpdocumentor bootstrap php on line warning require once c xampp htdocs siselc thomas vendor dompdf dompdf dompdf config inc php failed to open stream no such file or directory in c xampp php pear phpdocumentor src phpdocumentor bootstrap php on line php fatal error require once failed opening required c xampp htdocs siselc thomas vendor dompdf dompdf dompdf config inc php include path c xampp php pear in c xampp php pear phpdocumentor src phpdocumentor bootstrap php on line fatal error require once failed opening required c xampp htdocs siselc thomas vendor dompdf dompdf dompdf config inc php include path c xampp php pear in c xampp php pear phpdocumentor src phpdocumentor bootstrap php on line i am not an expert in this son i tried to install dompdf but i am still getting the same error thanks a lot | 1 |
19,048 | 25,050,303,271 | IssuesEvent | 2022-11-05 20:02:03 | nodejs/node | https://api.github.com/repos/nodejs/node | closed | child_process.spawn is considerably slower in Node 12+ | child_process performance | ### Version
Node v12+
### Platform
MacOS (Big Sur)
### Summary
We've discovered an issue with Electron 12 and above (Node 14 and above), where spawning a child_process takes considerably longer than Electron 11 and lower (Node 12 and lower).
The issue appears to be related to the Node version - Here's an example app, with instructions, to demonstrate:
https://github.com/dev-manager-uk/spawn-delay
After running the app initially, you'll notice the spawn function takes ~10ms.
After upgrading Electron to 12 or higher, you'll notice the spawn function takes ~500ms!
What's changed to cause this and can anything be done about it?
Thanks!
-- UPDATE --
This bug only occurs on macOS running Big Sur AND only with Electron 12+. I tested on Windows 10 and macOS running Mojave and could not replicate. Electron 11 and lower are also fine.
**To summarise, this delay only occurs on macOS running Big Sur with Electron 12 and greater.** | 1.0 | child_process.spawn is considerably slower in Node 12+ - ### Version
Node v12+
### Platform
MacOS (Big Sur)
### Summary
We've discovered an issue with Electron 12 and above (Node 14 and above), where spawning a child_process takes considerably longer than Electron 11 and lower (Node 12 and lower).
The issue appears to be related to the Node version - Here's an example app, with instructions, to demonstrate:
https://github.com/dev-manager-uk/spawn-delay
After running the app initially, you'll notice the spawn function takes ~10ms.
After upgrading Electron to 12 or higher, you'll notice the spawn function takes ~500ms!
What's changed to cause this and can anything be done about it?
Thanks!
-- UPDATE --
This bug only occurs on macOS running Big Sur AND only with Electron 12+. I tested on Windows 10 and macOS running Mojave and could not replicate. Electron 11 and lower are also fine.
**To summarise, this delay only occurs on macOS running Big Sur with Electron 12 and greater.** | process | child process spawn is considerably slower in node version node platform macos big sur summary we ve discovered an issue with electron and above node and above where spawning a child process takes considerably longer than electron and lower node and lower the issue appears to be related to the node version here s an example app with instructions to demonstrate after running the app initially you ll notice the spawn function takes after upgrading electron to or higher you ll notice the spawn function takes what s changed to cause this and can anything be done about it thanks update this bug only occurs on macos running big sur and only with electron i tested on windows and macos running mojave and could not replicate electron and lower are also fine to summarise this delay only occurs on macos running big sur with electron and greater | 1 |
9,259 | 12,294,552,717 | IssuesEvent | 2020-05-11 00:30:05 | allinurl/goaccess | https://api.github.com/repos/allinurl/goaccess | closed | Allow %h (or a new format specifier) to accept looked-up hostnames | add enhancement log-processing | First off, thank you for writing GoAccess and releasing it!
Could you please implement the ability to have `%h` (or perhaps a new format specifier like `%H`) accept hostnames that have already been looked up. In other words, to be able to accept log files where IP addresses are not recorded but the looked-up names are:
```
crawl-66-249-65-222.googlebot.com - - [21/Sep/2019:09:00:06 +1000] "GET /robots.txt HTTP/1.1" 200 571 "-" "Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)"
```
When I run `goaccess` with such log files, it complains with something like:
```
Parsed 10 lines producing the following errors:
Token 'crawl32.bl.semrush.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token 'msnbot-207-46-13-178.search.msn.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Format Errors - Verify your log/date/time format
```
(I tried to search whether someone had already filed this issue, but could not find anything to match).
Thanks! | 1.0 | Allow %h (or a new format specifier) to accept looked-up hostnames - First off, thank you for writing GoAccess and releasing it!
Could you please implement the ability to have `%h` (or perhaps a new format specifier like `%H`) accept hostnames that have already been looked up. In other words, to be able to accept log files where IP addresses are not recorded but the looked-up names are:
```
crawl-66-249-65-222.googlebot.com - - [21/Sep/2019:09:00:06 +1000] "GET /robots.txt HTTP/1.1" 200 571 "-" "Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)"
```
When I run `goaccess` with such log files, it complains with something like:
```
Parsed 10 lines producing the following errors:
Token 'crawl32.bl.semrush.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token 'msnbot-207-46-13-178.search.msn.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Token '141-8-143-196.spider.yandex.com' doesn't match specifier '%h'
Format Errors - Verify your log/date/time format
```
(I tried to search whether someone had already filed this issue, but could not find anything to match).
Thanks! | process | allow h or a new format specifier to accept looked up hostnames first off thank you for writing goaccess and releasing it could you please implement the ability to have h or perhaps a new format specifier like h accept hostnames that have already been looked up in other words to be able to accept log files where ip addresses are not recorded but the looked up names are crawl googlebot com get robots txt http mozilla compatible googlebot when i run goaccess with such log files it complains with something like parsed lines producing the following errors token bl semrush com doesn t match specifier h token spider yandex com doesn t match specifier h token msnbot search msn com doesn t match specifier h token spider yandex com doesn t match specifier h token spider yandex com doesn t match specifier h token spider yandex com doesn t match specifier h token spider yandex com doesn t match specifier h token spider yandex com doesn t match specifier h token spider yandex com doesn t match specifier h token spider yandex com doesn t match specifier h format errors verify your log date time format i tried to search whether someone had already filed this issue but could not find anything to match thanks | 1 |
11,263 | 14,048,899,735 | IssuesEvent | 2020-11-02 09:31:47 | pystatgen/sgkit | https://api.github.com/repos/pystatgen/sgkit | closed | Stop using __all__ for top-level namespace | process + tools | We seem to have a lot of "git conflict potential" in the current setup. The two main places seem to be in the ``known_third_party`` in setup.cfg and manually listing all API functions in ``__init__.py``. I can see how ``known_third_party`` will settle down after a while, but manually updating ``__init__.py`` for every new function is going to lead to lots of conflicts and be pretty tedious for maintainers.
Is it worth keeping such a strict hold on the top-level namespace? People can still use undocumented methods by accessing the individual packages, so I've never really seen the point in manually managing the ``__all__`` variable. | 1.0 | Stop using __all__ for top-level namespace - We seem to have a lot of "git conflict potential" in the current setup. The two main places seem to be in the ``known_third_party`` in setup.cfg and manually listing all API functions in ``__init__.py``. I can see how ``known_third_party`` will settle down after a while, but manually updating ``__init__.py`` for every new function is going to lead to lots of conflicts and be pretty tedious for maintainers.
Is it worth keeping such a strict hold on the top-level namespace? People can still use undocumented methods by accessing the individual packages, so I've never really seen the point in manually managing the ``__all__`` variable. | process | stop using all for top level namespace we seem to have a lot of git conflict potential in the current setup the two main places seem to be in the known third party in setup cfg and manually listing all api functions in init py i can see how known third party will settle down after a while but manually updating init py for every new function is going to lead to lots of conflicts and be pretty tedious for maintainers is it worth keeping such a strict hold on the top level namespace people can still use undocumented methods by accessing the individual packages so i ve never really seen the point in manually managing the all variable | 1 |
10,184 | 13,044,162,860 | IssuesEvent | 2020-07-29 03:47:37 | tikv/tikv | https://api.github.com/repos/tikv/tikv | closed | UCP: Migrate scalar function `TruncateReal` from TiDB | challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor |
## Description
Port the scalar function `TruncateReal` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @iosmanthus
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
| 2.0 | UCP: Migrate scalar function `TruncateReal` from TiDB -
## Description
Port the scalar function `TruncateReal` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @iosmanthus
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
| process | ucp migrate scalar function truncatereal from tidb description port the scalar function truncatereal from tidb to coprocessor score mentor s iosmanthus recommended skills rust programming learning materials already implemented expressions ported from tidb | 1 |
75,508 | 3,463,442,516 | IssuesEvent | 2015-12-21 10:01:32 | phusion/passenger | https://api.github.com/repos/phusion/passenger | closed | Graceful shutdown signal when using websockets | Enhancement EnterpriseCustomer Priority/High | When the new `passenger_abort_websockets_on_process_shutdown` is set to `false`, Passenger will wait with sending a termination signal until the application has closed its websockets. Passenger should have a way of signaling the app that it should commence websocket termination. | 1.0 | Graceful shutdown signal when using websockets - When the new `passenger_abort_websockets_on_process_shutdown` is set to `false`, Passenger will wait with sending a termination signal until the application has closed its websockets. Passenger should have a way of signaling the app that it should commence websocket termination. | non_process | graceful shutdown signal when using websockets when the new passenger abort websockets on process shutdown is set to false passenger will wait with sending a termination signal until the application has closed its websockets passenger should have a way of signaling the app that it should commence websocket termination | 0 |
7,530 | 10,606,153,451 | IssuesEvent | 2019-10-10 22:20:40 | googleapis/google-cloud-python | https://api.github.com/repos/googleapis/google-cloud-python | closed | Storage: add systests for V4 signed URLs with CSEK headers | api: storage backend testing type: process | /cc @frankyn Follow-on from PR #7460.
Currently, the GCS back-end has a bug for signed URLs using customer-supplied encryption key headers (`x-goog-encryption-key` and `x-goog-encryption-key-sha256`). Once a fix for that bug is rolled out, add system tests which show using the CSEK headers to build V4 signed upload / download URLs, and exercise them. | 1.0 | Storage: add systests for V4 signed URLs with CSEK headers - /cc @frankyn Follow-on from PR #7460.
Currently, the GCS back-end has a bug for signed URLs using customer-supplied encryption key headers (`x-goog-encryption-key` and `x-goog-encryption-key-sha256`). Once a fix for that bug is rolled out, add system tests which show using the CSEK headers to build V4 signed upload / download URLs, and exercise them. | process | storage add systests for signed urls with csek headers cc frankyn follow on from pr currently the gcs back end has a bug for signed urls using customer supplied encryption key headers x goog encryption key and x goog encryption key once a fix for that bug is rolled out add system tests which show using the csek headers to build signed upload download urls and exercise them | 1 |
20,922 | 27,758,991,130 | IssuesEvent | 2023-03-16 06:26:31 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | The backward call hangs on Torch for CPU | high priority oncall: distributed module: multiprocessing module: convolution module: deadlock | ### 🐛 Describe the bug
Very tiny model with two convolutions hangs on loss.backward() call.
```python
UNet(
(down_path): ModuleList(
(0): UNetConvBlock(
(block): Conv2d(3, 2, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
)
)
(last): Conv2d(2, 13, kernel_size=(1, 1), stride=(1, 1))
)
```
Snippet from stacktrace in GDB:
```
#0 0x00007fca0afed99f in __GI___poll (fds=0x7fc9ebbd0040, nfds=1, timeout=5000) at ../sysdeps/unix/sysv/linux/poll.c:29
```
It’s reproduced on torch with CPU only, GPU one works fine.
The batch size is also important. No hang with batch size = 1, but freezes with batch size = 4.
Profiling of the script for different batches shows different implementation called:
No hang with `_slow_conv2d_forward` and `aten::_slow_conv2d_backward`.
Hang happens with `aten::mkldnn_convolution` and `aten::convolution_backward`.
Snapshot of reproducing script:
```python
def main():
train_set = MockDataset()
train_loader = DataLoader(train_set, batch_size=4, num_workers=1, drop_last=True)
model = UNet(n_classes=13)
print(model)
device = 'cpu'
model.to(device)
optimizer = SGD(model.parameters(), lr=1e-3)
for epoch in range(1):
model.train()
for step, batch_data in enumerate(train_loader):
inputs = batch_data[0].to(device)
labels = batch_data[1].to(device)
optimizer.zero_grad()
outputs = model(inputs)
print(outputs.shape, labels.shape)
loss = nn.CrossEntropyLoss()(outputs, labels)
print(f"\nBefore HANG {loss}\n")
loss.backward()
print("\nAFTER HANG\n")
```
CPU: Intel(R) Core™ i9-10980XE CPU @ 3.00GHz (can’t reproduce on Intel(R) Core™ i9-10920X CPU @ 3.50GHz)
The issue is reproduced in the docker when the python script is launched in this way:
```bash
sh -c "(script.sh > log.txt 2>&1; wait)" >&- 2>&- &
```
Content of the script.sh:
```bash
#!/bin/bash -xe
. venv/bin/activate
python3 main.py
```
[The full reproducer, GDB stack trace and profile log for different batches](https://gist.github.com/ljaljushkin/c6daf65ea0c6e119b9336c5edba09a0c)
Original issue from the forum: https://discuss.pytorch.org/t/backward-hangs-on-torch-for-cpu/169282
### Versions
Collecting environment information...
PyTorch version: 1.13.1+cpu
Is debug build: False
CUDA used to build PyTorch: Could not collect
ROCM used to build PyTorch: N/A
OS: Ubuntu 20.04.4 LTS (x86_64)
GCC version: (Ubuntu 9.4.0-1ubuntu1~20.04.1) 9.4.0
Clang version: Could not collect
CMake version: Could not collect
Libc version: glibc-2.31
Python version: 3.8.10 (default, Mar 15 2022, 12:22:08) [GCC 9.4.0] (64-bit runtime)
Python platform: Linux-5.15.0-46-generic-x86_64-with-glibc2.29
Is CUDA available: False
CUDA runtime version: 11.1.105
CUDA_MODULE_LOADING set to: N/A
GPU models and configuration:
GPU 0: NVIDIA GeForce RTX 3090
GPU 1: NVIDIA GeForce RTX 3090
Nvidia driver version: 470.161.03
cuDNN version: Probably one of the following:
/usr/lib/x86_64-linux-gnu/libcudnn.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.0.5
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
Versions of relevant libraries:
[pip3] numpy==1.23.5
[pip3] torch==1.13.1+cpu <-------- also reproduced with torch==1.12.1+cpu and torch==1.9.1+cpu
[pip3] torchvision==0.14.1+cpu <-------- also reproduced with torchvision==0.13.1+cpu and torch==0.10.1+cpu
[conda] Could not collect
cc @ezyang @gchanan @zou3519 @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @osalpekar @jiayisuse @H-Huang @kwen2501 @awgu @VitalyFedyunin @ejguan | 1.0 | The backward call hangs on Torch for CPU - ### 🐛 Describe the bug
Very tiny model with two convolutions hangs on loss.backward() call.
```python
UNet(
(down_path): ModuleList(
(0): UNetConvBlock(
(block): Conv2d(3, 2, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
)
)
(last): Conv2d(2, 13, kernel_size=(1, 1), stride=(1, 1))
)
```
Snippet from stacktrace in GDB:
```
#0 0x00007fca0afed99f in __GI___poll (fds=0x7fc9ebbd0040, nfds=1, timeout=5000) at ../sysdeps/unix/sysv/linux/poll.c:29
```
It’s reproduced on torch with CPU only, GPU one works fine.
The batch size is also important. No hang with batch size = 1, but freezes with batch size = 4.
Profiling of the script for different batches shows different implementation called:
No hang with `_slow_conv2d_forward` and `aten::_slow_conv2d_backward`.
Hang happens with `aten::mkldnn_convolution` and `aten::convolution_backward`.
Snapshot of reproducing script:
```python
def main():
train_set = MockDataset()
train_loader = DataLoader(train_set, batch_size=4, num_workers=1, drop_last=True)
model = UNet(n_classes=13)
print(model)
device = 'cpu'
model.to(device)
optimizer = SGD(model.parameters(), lr=1e-3)
for epoch in range(1):
model.train()
for step, batch_data in enumerate(train_loader):
inputs = batch_data[0].to(device)
labels = batch_data[1].to(device)
optimizer.zero_grad()
outputs = model(inputs)
print(outputs.shape, labels.shape)
loss = nn.CrossEntropyLoss()(outputs, labels)
print(f"\nBefore HANG {loss}\n")
loss.backward()
print("\nAFTER HANG\n")
```
CPU: Intel(R) Core™ i9-10980XE CPU @ 3.00GHz (can’t reproduce on Intel(R) Core™ i9-10920X CPU @ 3.50GHz)
The issue is reproduced in the docker when the python script is launched in this way:
```bash
sh -c "(script.sh > log.txt 2>&1; wait)" >&- 2>&- &
```
Content of the script.sh:
```bash
#!/bin/bash -xe
. venv/bin/activate
python3 main.py
```
[The full reproducer, GDB stack trace and profile log for different batches](https://gist.github.com/ljaljushkin/c6daf65ea0c6e119b9336c5edba09a0c)
Original issue from the forum: https://discuss.pytorch.org/t/backward-hangs-on-torch-for-cpu/169282
### Versions
Collecting environment information...
PyTorch version: 1.13.1+cpu
Is debug build: False
CUDA used to build PyTorch: Could not collect
ROCM used to build PyTorch: N/A
OS: Ubuntu 20.04.4 LTS (x86_64)
GCC version: (Ubuntu 9.4.0-1ubuntu1~20.04.1) 9.4.0
Clang version: Could not collect
CMake version: Could not collect
Libc version: glibc-2.31
Python version: 3.8.10 (default, Mar 15 2022, 12:22:08) [GCC 9.4.0] (64-bit runtime)
Python platform: Linux-5.15.0-46-generic-x86_64-with-glibc2.29
Is CUDA available: False
CUDA runtime version: 11.1.105
CUDA_MODULE_LOADING set to: N/A
GPU models and configuration:
GPU 0: NVIDIA GeForce RTX 3090
GPU 1: NVIDIA GeForce RTX 3090
Nvidia driver version: 470.161.03
cuDNN version: Probably one of the following:
/usr/lib/x86_64-linux-gnu/libcudnn.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.0.5
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
Versions of relevant libraries:
[pip3] numpy==1.23.5
[pip3] torch==1.13.1+cpu <-------- also reproduced with torch==1.12.1+cpu and torch==1.9.1+cpu
[pip3] torchvision==0.14.1+cpu <-------- also reproduced with torchvision==0.13.1+cpu and torch==0.10.1+cpu
[conda] Could not collect
cc @ezyang @gchanan @zou3519 @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @osalpekar @jiayisuse @H-Huang @kwen2501 @awgu @VitalyFedyunin @ejguan | process | the backward call hangs on torch for cpu 🐛 describe the bug very tiny model with two convolutions hangs on loss backward call python unet down path modulelist unetconvblock block kernel size stride padding last kernel size stride snippet from stacktrace in gdb in gi poll fds nfds timeout at sysdeps unix sysv linux poll c it’s reproduced on torch with cpu only gpu one works fine the batch size is also important no hang with batch size but freezes with batch size profiling of the script for different batches shows different implementation called no hang with slow forward and aten slow backward hang happens with aten mkldnn convolution and aten convolution backward snapshot of reproducing script python def main train set mockdataset train loader dataloader train set batch size num workers drop last true model unet n classes print model device cpu model to device optimizer sgd model parameters lr for epoch in range model train for step batch data in enumerate train loader inputs batch data to device labels batch data to device optimizer zero grad outputs model inputs print outputs shape labels shape loss nn crossentropyloss outputs labels print f nbefore hang loss n loss backward print nafter hang n cpu intel r core™ cpu can’t reproduce on intel r core™ cpu the issue is reproduced in the docker when the python script is launched in this way bash sh c script sh log txt wait content of the script sh bash bin bash xe venv bin activate main py original issue from the forum versions collecting environment information pytorch version cpu is debug build false cuda used to build pytorch could not collect rocm used to build pytorch n a os ubuntu lts gcc version ubuntu clang version could not collect cmake version could not collect libc version glibc python version default mar bit runtime python platform linux generic with is cuda available false cuda runtime version cuda module loading set to n a gpu models and configuration gpu nvidia geforce rtx gpu nvidia geforce rtx nvidia driver version cudnn version probably one of the following usr lib linux gnu libcudnn so usr lib linux gnu libcudnn adv infer so usr lib linux gnu libcudnn adv train so usr lib linux gnu libcudnn cnn infer so usr lib linux gnu libcudnn cnn train so usr lib linux gnu libcudnn ops infer so usr lib linux gnu libcudnn ops train so hip runtime version n a miopen runtime version n a is xnnpack available true versions of relevant libraries numpy torch cpu also reproduced with torch cpu and torch cpu torchvision cpu also reproduced with torchvision cpu and torch cpu could not collect cc ezyang gchanan mrshenli zhaojuanmao satgera rohan varma gqchen aazzolini osalpekar jiayisuse h huang awgu vitalyfedyunin ejguan | 1 |
11,262 | 14,048,642,220 | IssuesEvent | 2020-11-02 09:09:46 | tikv/tikv | https://api.github.com/repos/tikv/tikv | closed | Dust seal the old coprocessor executors' code | difficulty/medium sig/coprocessor type/feature-request | ## Feature Request
### Is your feature request related to a problem? Please describe:
The migration of the coprocessor from the old executor framework toward the batch executor framework is almost done due to our contributor's [effort](https://github.com/tikv/tikv/pull/8322). However, after our batch execution framework support server-side streaming requests, some of the builtin functions are not yet migrated from the old execution framework. We need to port them to the new framework codebase and try to open its push down switch from TiDB. Finally, we could test them via [copr-test](https://github.com/tikv/copr-test).
Here is the function list that not port to batch execution framework but implemented in the old execution framework.
- [ ] `GreatestDecimal`
- [ ] `GreatestString`
- [ ] `LeastInt`
- [ ] `LeastReal`
- [ ] `LeastDecimal`
- [ ] `LeastString`
- [ ] `LeastTime`
- [ ] `IntervalInt`
- [ ] `IntervalReal`
- [ ] `TruncateDecimal`
- [ ] `TruncateUint`
- [ ] `Compress`
- [ ] `Uncompress`
- [ ] `RegexpSig`
- [ ] `RegexpUTF8Sig`
- [ ] `DateDiff`
- [ ] `Date`
- [ ] `MonthName`
- [ ] `WeekWithoutMode`
- [ ] `YearWeekWithMode`
- [ ] `YearWeekWithoutMode`
- [ ] `AddDatetimeAndDuration`
- [ ] `AddDatetimeAndString`
- [ ] `AddTimeDateTimeNull`
- [ ] `AddTimeStringNull`
- [ ] `AddDurationAndDuration`
- [ ] `AddDurationAndString`
- [ ] `AddTimeDurationNull`
- [ ] `SubDatetimeAndDuration`
- [ ] `SubDatetimeAndString`
- [ ] `SubTimeDateTimeNull`
- [ ] `SubDurationAndDuration`
- [ ] `SubDurationAndString`
- [ ] `SubTimeDurationNull`
- [ ] `ToSeconds`
- [ ] `Instr`
- [ ] `Locate2ArgsUTF8`
- [ ] `Locate3ArgsUTF8`
- [ ] `Lower`
- [ ] `Quote`
- [ ] `RpadUTF8`
- [ ] `Substring2ArgsUTF8`
- [ ] `Substring3ArgsUTF8`
- [ ] `Substring2Args`
- [ ] `Substring3Args`
- [ ] `Trim2Args` | 1.0 | Dust seal the old coprocessor executors' code - ## Feature Request
### Is your feature request related to a problem? Please describe:
The migration of the coprocessor from the old executor framework toward the batch executor framework is almost done due to our contributor's [effort](https://github.com/tikv/tikv/pull/8322). However, after our batch execution framework support server-side streaming requests, some of the builtin functions are not yet migrated from the old execution framework. We need to port them to the new framework codebase and try to open its push down switch from TiDB. Finally, we could test them via [copr-test](https://github.com/tikv/copr-test).
Here is the function list that not port to batch execution framework but implemented in the old execution framework.
- [ ] `GreatestDecimal`
- [ ] `GreatestString`
- [ ] `LeastInt`
- [ ] `LeastReal`
- [ ] `LeastDecimal`
- [ ] `LeastString`
- [ ] `LeastTime`
- [ ] `IntervalInt`
- [ ] `IntervalReal`
- [ ] `TruncateDecimal`
- [ ] `TruncateUint`
- [ ] `Compress`
- [ ] `Uncompress`
- [ ] `RegexpSig`
- [ ] `RegexpUTF8Sig`
- [ ] `DateDiff`
- [ ] `Date`
- [ ] `MonthName`
- [ ] `WeekWithoutMode`
- [ ] `YearWeekWithMode`
- [ ] `YearWeekWithoutMode`
- [ ] `AddDatetimeAndDuration`
- [ ] `AddDatetimeAndString`
- [ ] `AddTimeDateTimeNull`
- [ ] `AddTimeStringNull`
- [ ] `AddDurationAndDuration`
- [ ] `AddDurationAndString`
- [ ] `AddTimeDurationNull`
- [ ] `SubDatetimeAndDuration`
- [ ] `SubDatetimeAndString`
- [ ] `SubTimeDateTimeNull`
- [ ] `SubDurationAndDuration`
- [ ] `SubDurationAndString`
- [ ] `SubTimeDurationNull`
- [ ] `ToSeconds`
- [ ] `Instr`
- [ ] `Locate2ArgsUTF8`
- [ ] `Locate3ArgsUTF8`
- [ ] `Lower`
- [ ] `Quote`
- [ ] `RpadUTF8`
- [ ] `Substring2ArgsUTF8`
- [ ] `Substring3ArgsUTF8`
- [ ] `Substring2Args`
- [ ] `Substring3Args`
- [ ] `Trim2Args` | process | dust seal the old coprocessor executors code feature request is your feature request related to a problem please describe the migration of the coprocessor from the old executor framework toward the batch executor framework is almost done due to our contributor s however after our batch execution framework support server side streaming requests some of the builtin functions are not yet migrated from the old execution framework we need to port them to the new framework codebase and try to open its push down switch from tidb finally we could test them via here is the function list that not port to batch execution framework but implemented in the old execution framework greatestdecimal greateststring leastint leastreal leastdecimal leaststring leasttime intervalint intervalreal truncatedecimal truncateuint compress uncompress regexpsig datediff date monthname weekwithoutmode yearweekwithmode yearweekwithoutmode adddatetimeandduration adddatetimeandstring addtimedatetimenull addtimestringnull adddurationandduration adddurationandstring addtimedurationnull subdatetimeandduration subdatetimeandstring subtimedatetimenull subdurationandduration subdurationandstring subtimedurationnull toseconds instr lower quote | 1 |
6,241 | 9,199,358,117 | IssuesEvent | 2019-03-07 14:47:18 | googlegenomics/gcp-variant-transforms | https://api.github.com/repos/googlegenomics/gcp-variant-transforms | closed | Automate release process as much as possible | P3 enhancement process | The existing release process requires a few manual steps. These can be easily automated through a bash script. | 1.0 | Automate release process as much as possible - The existing release process requires a few manual steps. These can be easily automated through a bash script. | process | automate release process as much as possible the existing release process requires a few manual steps these can be easily automated through a bash script | 1 |
507 | 2,963,137,720 | IssuesEvent | 2015-07-10 08:15:29 | Philpax/athena | https://api.github.com/repos/Philpax/athena | closed | Creation of binary expressions from instructions | ast-process | `add(a, b)` should be mapped to `a + b`, etc.
Incomplete list of potential expression conversion targets:
`add(a, b)` -> `a + b`
`mul(a, b)` -> `a * b`
`ishl(a, b)` -> `a << b`
`ushr(a, b)` -> `a >> b`
`ge(a, b)` -> `a >= b` | 1.0 | Creation of binary expressions from instructions - `add(a, b)` should be mapped to `a + b`, etc.
Incomplete list of potential expression conversion targets:
`add(a, b)` -> `a + b`
`mul(a, b)` -> `a * b`
`ishl(a, b)` -> `a << b`
`ushr(a, b)` -> `a >> b`
`ge(a, b)` -> `a >= b` | process | creation of binary expressions from instructions add a b should be mapped to a b etc incomplete list of potential expression conversion targets add a b a b mul a b a b ishl a b a b ushr a b a b ge a b a b | 1 |
77,828 | 27,183,862,877 | IssuesEvent | 2023-02-19 00:31:25 | amyjko/bookish | https://api.github.com/repos/amyjko/bookish | closed | Image doesn't appear after upload | defect writing | Repro:
* Drag an image to book or chapter cover
* Image chooser shows uploaded image but the cover image says it was unable to get the URL
There are some wonky things about the Google Cloud feedback about file upload success. Need to work around some of their defects or see if the APIs are fixed. Part of the problem is that our current approach is triggered by a new file change in Google Cloud, and then the client polls until it finds the resized image. This is not only unreliable, but also slower, since it means we check on every image change. We should trigger it on the client. | 1.0 | Image doesn't appear after upload - Repro:
* Drag an image to book or chapter cover
* Image chooser shows uploaded image but the cover image says it was unable to get the URL
There are some wonky things about the Google Cloud feedback about file upload success. Need to work around some of their defects or see if the APIs are fixed. Part of the problem is that our current approach is triggered by a new file change in Google Cloud, and then the client polls until it finds the resized image. This is not only unreliable, but also slower, since it means we check on every image change. We should trigger it on the client. | non_process | image doesn t appear after upload repro drag an image to book or chapter cover image chooser shows uploaded image but the cover image says it was unable to get the url there are some wonky things about the google cloud feedback about file upload success need to work around some of their defects or see if the apis are fixed part of the problem is that our current approach is triggered by a new file change in google cloud and then the client polls until it finds the resized image this is not only unreliable but also slower since it means we check on every image change we should trigger it on the client | 0 |
127,413 | 18,010,455,240 | IssuesEvent | 2021-09-16 07:59:57 | maddyCode23/linux-4.1.15 | https://api.github.com/repos/maddyCode23/linux-4.1.15 | opened | CVE-2016-4557 (High) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2016-4557 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The replace_map_fd_with_map_ptr function in kernel/bpf/verifier.c in the Linux kernel before 4.5.5 does not properly maintain an fd data structure, which allows local users to gain privileges or cause a denial of service (use-after-free) via crafted BPF instructions that reference an incorrect file descriptor.
<p>Publish Date: 2016-05-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-4557>CVE-2016-4557</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-4557">https://nvd.nist.gov/vuln/detail/CVE-2016-4557</a></p>
<p>Release Date: 2016-05-23</p>
<p>Fix Resolution: 4.5.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2016-4557 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2016-4557 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The replace_map_fd_with_map_ptr function in kernel/bpf/verifier.c in the Linux kernel before 4.5.5 does not properly maintain an fd data structure, which allows local users to gain privileges or cause a denial of service (use-after-free) via crafted BPF instructions that reference an incorrect file descriptor.
<p>Publish Date: 2016-05-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-4557>CVE-2016-4557</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-4557">https://nvd.nist.gov/vuln/detail/CVE-2016-4557</a></p>
<p>Release Date: 2016-05-23</p>
<p>Fix Resolution: 4.5.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href vulnerable source files kernel bpf verifier c kernel bpf verifier c vulnerability details the replace map fd with map ptr function in kernel bpf verifier c in the linux kernel before does not properly maintain an fd data structure which allows local users to gain privileges or cause a denial of service use after free via crafted bpf instructions that reference an incorrect file descriptor publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
17,499 | 23,311,733,457 | IssuesEvent | 2022-08-08 08:52:40 | pystatgen/sgkit | https://api.github.com/repos/pystatgen/sgkit | closed | Fix windows wheels tests | process + tools | Don't test on Python 3.10, and make sure Python 3.8 and 3.9 use cbgen 1.0.1 not 1.0.2 (like #878, but in setup.cfg). | 1.0 | Fix windows wheels tests - Don't test on Python 3.10, and make sure Python 3.8 and 3.9 use cbgen 1.0.1 not 1.0.2 (like #878, but in setup.cfg). | process | fix windows wheels tests don t test on python and make sure python and use cbgen not like but in setup cfg | 1 |
3,915 | 17,500,657,036 | IssuesEvent | 2021-08-10 09:01:30 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | closed | MiKo_3205 should be aware of 'yield return' | bug Area: analyzer Area: maintainability | Multiple `yield return` calls without empy lines in between should not trigger a violation. | True | MiKo_3205 should be aware of 'yield return' - Multiple `yield return` calls without empy lines in between should not trigger a violation. | non_process | miko should be aware of yield return multiple yield return calls without empy lines in between should not trigger a violation | 0 |
94,802 | 8,518,046,284 | IssuesEvent | 2018-11-01 10:13:21 | actnforchildren/mental_health_app | https://api.github.com/repos/actnforchildren/mental_health_app | closed | Emotion logging - Emojis | please-test priority-2 | When logging my daily emotions,
I would like to choose from cartoon characters with generic facial expressions
So I am able to attribute my emotions to how they are represented visually.
### Acceptance Criteria
- [x] Users can choose from a pre-determined list of emoji which can reflect how they're feeling
- [x] Emojis will show the emotions: Happy, Excited, Angry, Sad, Worried, I don't know, Something Else
For the MVP, we will use the generic emojis everyone is familiar with.
Relates #10 #8 | 1.0 | Emotion logging - Emojis - When logging my daily emotions,
I would like to choose from cartoon characters with generic facial expressions
So I am able to attribute my emotions to how they are represented visually.
### Acceptance Criteria
- [x] Users can choose from a pre-determined list of emoji which can reflect how they're feeling
- [x] Emojis will show the emotions: Happy, Excited, Angry, Sad, Worried, I don't know, Something Else
For the MVP, we will use the generic emojis everyone is familiar with.
Relates #10 #8 | non_process | emotion logging emojis when logging my daily emotions i would like to choose from cartoon characters with generic facial expressions so i am able to attribute my emotions to how they are represented visually acceptance criteria users can choose from a pre determined list of emoji which can reflect how they re feeling emojis will show the emotions happy excited angry sad worried i don t know something else for the mvp we will use the generic emojis everyone is familiar with relates | 0 |
246,362 | 20,842,899,094 | IssuesEvent | 2022-03-21 04:03:13 | dapr/java-sdk | https://api.github.com/repos/dapr/java-sdk | closed | Fix Service invocation path segments automatic encode of `/` into %2f | kind/bug triaged/resolved area/client area/test/integration-tests P1 size/XS | ## Expected Behavior
Given a method name containing "/" in service invocation, properly split path segments and call dapr.
## Actual Behavior
Currently in PubSub IT,
https://github.com/dapr/java-sdk/blob/master/sdk-tests/src/test/java/io/dapr/it/pubsub/http/PubSubIT.java#L396-L400
The method being called is `messages/ttlTopic` which gets encoded as `messages%2fttlTopic` which was working prior to fix https://github.com/dapr/dapr/issues/4008.
With the fix merged to Dapr, on Feb 8th, any build of Dapr after that fails the PubSub IT since it tries to invoke `http://...../messages%2fttltopic` instead of `http://..../messages/ttlTopic`.
See https://github.com/dapr/java-sdk/runs/5533656018?check_suite_focus=true#step:20:6305
## Steps to Reproduce the Problem
Change Dapr Ref in build.yaml to any latest master ref from dapr/dapr and run the ITs.
## Release Note
<!-- How should the fix for this issue be communicated in our release notes? It can be populated later. -->
<!-- Keep it as a single line. Examples: -->
<!-- RELEASE NOTE: **ADD** New feature in Dapr. -->
<!-- RELEASE NOTE: **FIX** Bug in runtime. -->
<!-- RELEASE NOTE: **UPDATE** Runtime dependency. -->
RELEASE NOTE: Fix service invocation path segments to avoid automatic encoding of `/` to `%2f` | 2.0 | Fix Service invocation path segments automatic encode of `/` into %2f - ## Expected Behavior
Given a method name containing "/" in service invocation, properly split path segments and call dapr.
## Actual Behavior
Currently in PubSub IT,
https://github.com/dapr/java-sdk/blob/master/sdk-tests/src/test/java/io/dapr/it/pubsub/http/PubSubIT.java#L396-L400
The method being called is `messages/ttlTopic` which gets encoded as `messages%2fttlTopic` which was working prior to fix https://github.com/dapr/dapr/issues/4008.
With the fix merged to Dapr, on Feb 8th, any build of Dapr after that fails the PubSub IT since it tries to invoke `http://...../messages%2fttltopic` instead of `http://..../messages/ttlTopic`.
See https://github.com/dapr/java-sdk/runs/5533656018?check_suite_focus=true#step:20:6305
## Steps to Reproduce the Problem
Change Dapr Ref in build.yaml to any latest master ref from dapr/dapr and run the ITs.
## Release Note
<!-- How should the fix for this issue be communicated in our release notes? It can be populated later. -->
<!-- Keep it as a single line. Examples: -->
<!-- RELEASE NOTE: **ADD** New feature in Dapr. -->
<!-- RELEASE NOTE: **FIX** Bug in runtime. -->
<!-- RELEASE NOTE: **UPDATE** Runtime dependency. -->
RELEASE NOTE: Fix service invocation path segments to avoid automatic encoding of `/` to `%2f` | non_process | fix service invocation path segments automatic encode of into expected behavior given a method name containing in service invocation properly split path segments and call dapr actual behavior currently in pubsub it the method being called is messages ttltopic which gets encoded as messages which was working prior to fix with the fix merged to dapr on feb any build of dapr after that fails the pubsub it since it tries to invoke instead of see steps to reproduce the problem change dapr ref in build yaml to any latest master ref from dapr dapr and run the its release note release note fix service invocation path segments to avoid automatic encoding of to | 0 |
93,163 | 3,886,470,099 | IssuesEvent | 2016-04-14 01:11:17 | Solinea/goldstone-server | https://api.github.com/repos/Solinea/goldstone-server | reopened | Remove plugin install from goldstone-search image | priority 4: low type: enhancement | The following plugins should be installed as part of solinea/logstash image, and not goldstone-search.
logstash-filter-translate
logstash-input-http
Will increase the efficiency of the image build, and only focus on the goldstone-specific build portion. | 1.0 | Remove plugin install from goldstone-search image - The following plugins should be installed as part of solinea/logstash image, and not goldstone-search.
logstash-filter-translate
logstash-input-http
Will increase the efficiency of the image build, and only focus on the goldstone-specific build portion. | non_process | remove plugin install from goldstone search image the following plugins should be installed as part of solinea logstash image and not goldstone search logstash filter translate logstash input http will increase the efficiency of the image build and only focus on the goldstone specific build portion | 0 |
3,711 | 6,732,531,559 | IssuesEvent | 2017-10-18 11:53:41 | lockedata/rcms | https://api.github.com/repos/lockedata/rcms | opened | Manage speaker submission | conference team odoo processes | ## Detailed task
- Review submissions
- Accept/reject sessions
- Email speakers
## Assessing the task
Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks.
Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback.
## Extra Info
- Site: [odoo](//http://188.166.159.192:8069)
- System documentation: [odoo docs](https://www.odoo.com/page/docs)
- Role: Conference team
- Area: Processes
| 1.0 | Manage speaker submission - ## Detailed task
- Review submissions
- Accept/reject sessions
- Email speakers
## Assessing the task
Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks.
Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback.
## Extra Info
- Site: [odoo](//http://188.166.159.192:8069)
- System documentation: [odoo docs](https://www.odoo.com/page/docs)
- Role: Conference team
- Area: Processes
| process | manage speaker submission detailed task review submissions accept reject sessions email speakers assessing the task try to perform the task use google and the system documentation to help part of what we re trying to assess how easy it is for people to work out how to do tasks use a 👍 reaction to this task if you were able to perform the task use a 👎 reaction to the task if you could not complete it add a reply with any comments or feedback extra info site system documentation role conference team area processes | 1 |
16,106 | 20,358,060,079 | IssuesEvent | 2022-02-20 08:59:57 | fmnas/fmnas-site | https://api.github.com/repos/fmnas/fmnas-site | closed | Email the applicant a copy of their application as a PDF. | public backend form processor medium (3-8h) +1 from Sean |
---
_This issue has been automatically created by [todo-actions](https://github.com/apps/todo-actions) based on a TODO comment found in [public/application/index.php:91](https://github.com/fmnas/fmnas-site/blob/main/public/application/index.php#L91). It will automatically be closed when the TODO comment is removed from the default branch (main)._ | 1.0 | Email the applicant a copy of their application as a PDF. -
---
_This issue has been automatically created by [todo-actions](https://github.com/apps/todo-actions) based on a TODO comment found in [public/application/index.php:91](https://github.com/fmnas/fmnas-site/blob/main/public/application/index.php#L91). It will automatically be closed when the TODO comment is removed from the default branch (main)._ | process | email the applicant a copy of their application as a pdf this issue has been automatically created by based on a todo comment found in it will automatically be closed when the todo comment is removed from the default branch main | 1 |
19,683 | 26,033,581,876 | IssuesEvent | 2022-12-22 00:57:26 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | DISABLED test_terminate_signal (__main__.SpawnTest) | module: multiprocessing triaged module: flaky-tests skipped | Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](http://torch-ci.com/failure/test_terminate_signal%2C%20SpawnTest) and the most recent [workflow logs](https://github.com/pytorch/pytorch/actions/runs/1890592409).
Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 red and 3 green.
cc @VitalyFedyunin | 1.0 | DISABLED test_terminate_signal (__main__.SpawnTest) - Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](http://torch-ci.com/failure/test_terminate_signal%2C%20SpawnTest) and the most recent [workflow logs](https://github.com/pytorch/pytorch/actions/runs/1890592409).
Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 red and 3 green.
cc @VitalyFedyunin | process | disabled test terminate signal main spawntest platforms linux this test was disabled because it is failing in ci see and the most recent over the past hours it has been determined flaky in workflow s with red and green cc vitalyfedyunin | 1 |
491,391 | 14,163,236,934 | IssuesEvent | 2020-11-12 01:51:39 | Baystation12/Baystation12 | https://api.github.com/repos/Baystation12/Baystation12 | closed | Clicking 'Reconnect' on the arctic planet shuttle helm causes the MC to crash. | Bug :bug: Could Reproduce :bug: Priority: High ⚠ | <!--
Anything inside tags like these is a comment and will not be displayed in the final issue.
Be careful not to write inside them!
Every field other than 'specific information for locating' is required.
If you do not fill out the 'specific information' field, please delete the header.
/!\ Omitting or not answering a required field will result in your issue being closed. /!\
Repeated violation of this rule, or joke or spam issues, will result in punishment.
PUT YOUR ANSWERS ON THE BLANK LINES BELOW THE HEADERS
(The lines with four #'s)
Don't edit them or delete them - it's part of the formatting
-->
#### Description of issue
Clicking 'reconnect' on the helm console of the arctic planet shuttle (the one that spawns there) will crash the MC.
#### Difference between expected and actual behavior
It should definitely not crash the MC. It should, at most, do nothing, probably (unless it should be a working shuttle?)
#### Steps to reproduce
1. Power the area if needed.
2. Go to the helm console (or sensor console, I think), and click "reconnect" in the interface.
#### Specific information for locating
Site is the Bluespace River.
#### Length of time in which bug has been known to occur
Unsure, maybe forever? Not sure if this has been found or even tried before.
#### Client version, Server revision & Game ID
Client Version: 513
Server Revision: 2bca524e7623248c4050373d84e9212ad5882af2 - dev - 2020-11-05
#### Issue bingo
<!-- Check these by writing an x inside the [ ] (like this: [x])-->
<!-- Don't forget to remove the space between the brackets, or it won't work! -->
- [x] Issue could be reproduced at least once
- [ ] Issue could be reproduced by different players
- [x] Issue could be reproduced in multiple rounds
- [x] Issue happened in a recent (less than 7 days ago) round
- [x] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
| 1.0 | Clicking 'Reconnect' on the arctic planet shuttle helm causes the MC to crash. - <!--
Anything inside tags like these is a comment and will not be displayed in the final issue.
Be careful not to write inside them!
Every field other than 'specific information for locating' is required.
If you do not fill out the 'specific information' field, please delete the header.
/!\ Omitting or not answering a required field will result in your issue being closed. /!\
Repeated violation of this rule, or joke or spam issues, will result in punishment.
PUT YOUR ANSWERS ON THE BLANK LINES BELOW THE HEADERS
(The lines with four #'s)
Don't edit them or delete them - it's part of the formatting
-->
#### Description of issue
Clicking 'reconnect' on the helm console of the arctic planet shuttle (the one that spawns there) will crash the MC.
#### Difference between expected and actual behavior
It should definitely not crash the MC. It should, at most, do nothing, probably (unless it should be a working shuttle?)
#### Steps to reproduce
1. Power the area if needed.
2. Go to the helm console (or sensor console, I think), and click "reconnect" in the interface.
#### Specific information for locating
Site is the Bluespace River.
#### Length of time in which bug has been known to occur
Unsure, maybe forever? Not sure if this has been found or even tried before.
#### Client version, Server revision & Game ID
Client Version: 513
Server Revision: 2bca524e7623248c4050373d84e9212ad5882af2 - dev - 2020-11-05
#### Issue bingo
<!-- Check these by writing an x inside the [ ] (like this: [x])-->
<!-- Don't forget to remove the space between the brackets, or it won't work! -->
- [x] Issue could be reproduced at least once
- [ ] Issue could be reproduced by different players
- [x] Issue could be reproduced in multiple rounds
- [x] Issue happened in a recent (less than 7 days ago) round
- [x] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
| non_process | clicking reconnect on the arctic planet shuttle helm causes the mc to crash anything inside tags like these is a comment and will not be displayed in the final issue be careful not to write inside them every field other than specific information for locating is required if you do not fill out the specific information field please delete the header omitting or not answering a required field will result in your issue being closed repeated violation of this rule or joke or spam issues will result in punishment put your answers on the blank lines below the headers the lines with four s don t edit them or delete them it s part of the formatting description of issue clicking reconnect on the helm console of the arctic planet shuttle the one that spawns there will crash the mc difference between expected and actual behavior it should definitely not crash the mc it should at most do nothing probably unless it should be a working shuttle steps to reproduce power the area if needed go to the helm console or sensor console i think and click reconnect in the interface specific information for locating site is the bluespace river length of time in which bug has been known to occur unsure maybe forever not sure if this has been found or even tried before client version server revision game id client version server revision dev issue bingo issue could be reproduced at least once issue could be reproduced by different players issue could be reproduced in multiple rounds issue happened in a recent less than days ago round | 0 |
46,072 | 13,055,847,926 | IssuesEvent | 2020-07-30 02:54:55 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | PYTHON_LOGGING eanabled causes builds to fail. (Trac #539) | Incomplete Migration Migrated from Trac cmake defect | Migrated from https://code.icecube.wisc.edu/ticket/539
```json
{
"status": "closed",
"changetime": "2009-03-01T00:40:18",
"description": "When PYTHON_LOGGING in CMakeCache.txt is set to True, dataio fails to compile:\n\n/disk02/home/blaufuss/icework/offline-software/trunk/src/dataio/public/dataio/I3File.h:48:\n error: ISO C++ forbids declaration of \u2018map\u2019 with no type\n/disk02/home/blaufuss/icework/offline-software/trunk/src/dataio/public/dataio/I3File.h:48:\n error: typedef name may not be a nested-name-specifier\n\nSeems some include files are getting munged.\n\n",
"reporter": "anonymous",
"cc": "",
"resolution": "fixed",
"_ts": "1235868018000000",
"component": "cmake",
"summary": "PYTHON_LOGGING eanabled causes builds to fail.",
"priority": "normal",
"keywords": "",
"time": "2009-02-28T19:41:04",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| 1.0 | PYTHON_LOGGING eanabled causes builds to fail. (Trac #539) - Migrated from https://code.icecube.wisc.edu/ticket/539
```json
{
"status": "closed",
"changetime": "2009-03-01T00:40:18",
"description": "When PYTHON_LOGGING in CMakeCache.txt is set to True, dataio fails to compile:\n\n/disk02/home/blaufuss/icework/offline-software/trunk/src/dataio/public/dataio/I3File.h:48:\n error: ISO C++ forbids declaration of \u2018map\u2019 with no type\n/disk02/home/blaufuss/icework/offline-software/trunk/src/dataio/public/dataio/I3File.h:48:\n error: typedef name may not be a nested-name-specifier\n\nSeems some include files are getting munged.\n\n",
"reporter": "anonymous",
"cc": "",
"resolution": "fixed",
"_ts": "1235868018000000",
"component": "cmake",
"summary": "PYTHON_LOGGING eanabled causes builds to fail.",
"priority": "normal",
"keywords": "",
"time": "2009-02-28T19:41:04",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| non_process | python logging eanabled causes builds to fail trac migrated from json status closed changetime description when python logging in cmakecache txt is set to true dataio fails to compile n n home blaufuss icework offline software trunk src dataio public dataio h n error iso c forbids declaration of with no type n home blaufuss icework offline software trunk src dataio public dataio h n error typedef name may not be a nested name specifier n nseems some include files are getting munged n n reporter anonymous cc resolution fixed ts component cmake summary python logging eanabled causes builds to fail priority normal keywords time milestone owner troy type defect | 0 |
22,530 | 31,653,710,061 | IssuesEvent | 2023-09-07 02:00:10 | lizhihao6/get-daily-arxiv-noti | https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti | opened | New submissions for Thu, 7 Sep 23 | event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB | ## Keyword: events
### Domain Adaptation for Efficiently Fine-tuning Vision Transformer with Encrypted Images
- **Authors:** Teru Nagamori, Sayaka Shiota, Hitoshi Kiya
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Cryptography and Security (cs.CR); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2309.02556
- **Pdf link:** https://arxiv.org/pdf/2309.02556
- **Abstract**
In recent years, deep neural networks (DNNs) trained with transformed data have been applied to various applications such as privacy-preserving learning, access control, and adversarial defenses. However, the use of transformed data decreases the performance of models. Accordingly, in this paper, we propose a novel method for fine-tuning models with transformed images under the use of the vision transformer (ViT). The proposed domain adaptation method does not cause the accuracy degradation of models, and it is carried out on the basis of the embedding structure of ViT. In experiments, we confirmed that the proposed method prevents accuracy degradation even when using encrypted images with the CIFAR-10 and CIFAR-100 datasets.
### SlAction: Non-intrusive, Lightweight Obstructive Sleep Apnea Detection using Infrared Video
- **Authors:** You Rim Choi, Gyeongseon Eo, Wonhyuck Youn, Hyojin Lee, Haemin Jang, Dongyoon Kim, Hyunwoo Shin, Hyung-Sin Kim
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2309.02713
- **Pdf link:** https://arxiv.org/pdf/2309.02713
- **Abstract**
Obstructive sleep apnea (OSA) is a prevalent sleep disorder affecting approximately one billion people world-wide. The current gold standard for diagnosing OSA, Polysomnography (PSG), involves an overnight hospital stay with multiple attached sensors, leading to potential inaccuracies due to the first-night effect. To address this, we present SlAction, a non-intrusive OSA detection system for daily sleep environments using infrared videos. Recognizing that sleep videos exhibit minimal motion, this work investigates the fundamental question: "Are respiratory events adequately reflected in human motions during sleep?" Analyzing the largest sleep video dataset of 5,098 hours, we establish correlations between OSA events and human motions during sleep. Our approach uses a low frame rate (2.5 FPS), a large size (60 seconds) and step (30 seconds) for sliding window analysis to capture slow and long-term motions related to OSA. Furthermore, we utilize a lightweight deep neural network for resource-constrained devices, ensuring all video streams are processed locally without compromising privacy. Evaluations show that SlAction achieves an average F1 score of 87.6% in detecting OSA across various environments. Implementing SlAction on NVIDIA Jetson Nano enables real-time inference (~3 seconds for a 60-second video clip), highlighting its potential for early detection and personalized treatment of OSA.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### Image Aesthetics Assessment via Learnable Queries
- **Authors:** Zhiwei Xiong, Yunfan Zhang, Zhiqi Shen, Peiran Ren, Han Yu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.02861
- **Pdf link:** https://arxiv.org/pdf/2309.02861
- **Abstract**
Image aesthetics assessment (IAA) aims to estimate the aesthetics of images. Depending on the content of an image, diverse criteria need to be selected to assess its aesthetics. Existing works utilize pre-trained vision backbones based on content knowledge to learn image aesthetics. However, training those backbones is time-consuming and suffers from attention dispersion. Inspired by learnable queries in vision-language alignment, we propose the Image Aesthetics Assessment via Learnable Queries (IAA-LQ) approach. It adapts learnable queries to extract aesthetic features from pre-trained image features obtained from a frozen image encoder. Extensive experiments on real-world data demonstrate the advantages of IAA-LQ, beating the best state-of-the-art method by 2.2% and 2.1% in terms of SRCC and PLCC, respectively.
### Do We Still Need Non-Maximum Suppression? Accurate Confidence Estimates and Implicit Duplication Modeling with IoU-Aware Calibration
- **Authors:** Johannes Gilg, Torben Teepe, Fabian Herzog, Philipp Wolters, Gerhard Rigoll
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.03110
- **Pdf link:** https://arxiv.org/pdf/2309.03110
- **Abstract**
Object detectors are at the heart of many semi- and fully autonomous decision systems and are poised to become even more indispensable. They are, however, still lacking in accessibility and can sometimes produce unreliable predictions. Especially concerning in this regard are the -- essentially hand-crafted -- non-maximum suppression algorithms that lead to an obfuscated prediction process and biased confidence estimates. We show that we can eliminate classic NMS-style post-processing by using IoU-aware calibration. IoU-aware calibration is a conditional Beta calibration; this makes it parallelizable with no hyper-parameters. Instead of arbitrary cutoffs or discounts, it implicitly accounts for the likelihood of each detection being a duplicate and adjusts the confidence score accordingly, resulting in empirically based precision estimates for each detection. Our extensive experiments on diverse detection architectures show that the proposed IoU-aware calibration can successfully model duplicate detections and improve calibration. Compared to the standard sequential NMS and calibration approach, our joint modeling can deliver performance gains over the best NMS-based alternative while producing consistently better-calibrated confidence predictions with less complexity. The \hyperlink{https://github.com/Blueblue4/IoU-AwareCalibration}{code} for all our experiments is publicly available.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### Compressing Vision Transformers for Low-Resource Visual Learning
- **Authors:** Eric Youn, Sai Mitheran J, Sanjana Prabhu, Siyuan Chen
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2309.02617
- **Pdf link:** https://arxiv.org/pdf/2309.02617
- **Abstract**
Vision transformer (ViT) and its variants have swept through visual learning leaderboards and offer state-of-the-art accuracy in tasks such as image classification, object detection, and semantic segmentation by attending to different parts of the visual input and capturing long-range spatial dependencies. However, these models are large and computation-heavy. For instance, the recently proposed ViT-B model has 86M parameters making it impractical for deployment on resource-constrained devices. As a result, their deployment on mobile and edge scenarios is limited. In our work, we aim to take a step toward bringing vision transformers to the edge by utilizing popular model compression techniques such as distillation, pruning, and quantization. Our chosen application environment is an unmanned aerial vehicle (UAV) that is battery-powered and memory-constrained, carrying a single-board computer on the scale of an NVIDIA Jetson Nano with 4GB of RAM. On the other hand, the UAV requires high accuracy close to that of state-of-the-art ViTs to ensure safe object avoidance in autonomous navigation, or correct localization of humans in search-and-rescue. Inference latency should also be minimized given the application requirements. Hence, our target is to enable rapid inference of a vision transformer on an NVIDIA Jetson Nano (4GB) with minimal accuracy loss. This allows us to deploy ViTs on resource-constrained devices, opening up new possibilities in surveillance, environmental monitoring, etc. Our implementation is made available at https://github.com/chensy7/efficient-vit.
### Bandwidth-efficient Inference for Neural Image Compression
- **Authors:** Shanzhi Yin, Tongda Xu, Yongsheng Liang, Yuanyuan Wang, Yanghao Li, Yan Wang, Jingjing Liu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2309.02855
- **Pdf link:** https://arxiv.org/pdf/2309.02855
- **Abstract**
With neural networks growing deeper and feature maps growing larger, limited communication bandwidth with external memory (or DRAM) and power constraints become a bottleneck in implementing network inference on mobile and edge devices. In this paper, we propose an end-to-end differentiable bandwidth efficient neural inference method with the activation compressed by neural data compression method. Specifically, we propose a transform-quantization-entropy coding pipeline for activation compression with symmetric exponential Golomb coding and a data-dependent Gaussian entropy model for arithmetic coding. Optimized with existing model quantization methods, low-level task of image compression can achieve up to 19x bandwidth reduction with 6.21x energy saving.
## Keyword: RAW
### FArMARe: a Furniture-Aware Multi-task methodology for Recommending Apartments based on the user interests
- **Authors:** Ali Abdari, Alex Falcon, Giuseppe Serra
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Multimedia (cs.MM)
- **Arxiv link:** https://arxiv.org/abs/2309.03100
- **Pdf link:** https://arxiv.org/pdf/2309.03100
- **Abstract**
Nowadays, many people frequently have to search for new accommodation options. Searching for a suitable apartment is a time-consuming process, especially because visiting them is often mandatory to assess the truthfulness of the advertisements found on the Web. While this process could be alleviated by visiting the apartments in the metaverse, the Web-based recommendation platforms are not suitable for the task. To address this shortcoming, in this paper, we define a new problem called text-to-apartment recommendation, which requires ranking the apartments based on their relevance to a textual query expressing the user's interests. To tackle this problem, we introduce FArMARe, a multi-task approach that supports cross-modal contrastive training with a furniture-aware objective. Since public datasets related to indoor scenes do not contain detailed descriptions of the furniture, we collect and annotate a dataset comprising more than 6000 apartments. A thorough experimentation with three different methods and two raw feature extraction procedures reveals the effectiveness of FArMARe in dealing with the problem at hand.
## Keyword: raw image
There is no result
| 2.0 | New submissions for Thu, 7 Sep 23 - ## Keyword: events
### Domain Adaptation for Efficiently Fine-tuning Vision Transformer with Encrypted Images
- **Authors:** Teru Nagamori, Sayaka Shiota, Hitoshi Kiya
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Cryptography and Security (cs.CR); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2309.02556
- **Pdf link:** https://arxiv.org/pdf/2309.02556
- **Abstract**
In recent years, deep neural networks (DNNs) trained with transformed data have been applied to various applications such as privacy-preserving learning, access control, and adversarial defenses. However, the use of transformed data decreases the performance of models. Accordingly, in this paper, we propose a novel method for fine-tuning models with transformed images under the use of the vision transformer (ViT). The proposed domain adaptation method does not cause the accuracy degradation of models, and it is carried out on the basis of the embedding structure of ViT. In experiments, we confirmed that the proposed method prevents accuracy degradation even when using encrypted images with the CIFAR-10 and CIFAR-100 datasets.
### SlAction: Non-intrusive, Lightweight Obstructive Sleep Apnea Detection using Infrared Video
- **Authors:** You Rim Choi, Gyeongseon Eo, Wonhyuck Youn, Hyojin Lee, Haemin Jang, Dongyoon Kim, Hyunwoo Shin, Hyung-Sin Kim
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2309.02713
- **Pdf link:** https://arxiv.org/pdf/2309.02713
- **Abstract**
Obstructive sleep apnea (OSA) is a prevalent sleep disorder affecting approximately one billion people world-wide. The current gold standard for diagnosing OSA, Polysomnography (PSG), involves an overnight hospital stay with multiple attached sensors, leading to potential inaccuracies due to the first-night effect. To address this, we present SlAction, a non-intrusive OSA detection system for daily sleep environments using infrared videos. Recognizing that sleep videos exhibit minimal motion, this work investigates the fundamental question: "Are respiratory events adequately reflected in human motions during sleep?" Analyzing the largest sleep video dataset of 5,098 hours, we establish correlations between OSA events and human motions during sleep. Our approach uses a low frame rate (2.5 FPS), a large size (60 seconds) and step (30 seconds) for sliding window analysis to capture slow and long-term motions related to OSA. Furthermore, we utilize a lightweight deep neural network for resource-constrained devices, ensuring all video streams are processed locally without compromising privacy. Evaluations show that SlAction achieves an average F1 score of 87.6% in detecting OSA across various environments. Implementing SlAction on NVIDIA Jetson Nano enables real-time inference (~3 seconds for a 60-second video clip), highlighting its potential for early detection and personalized treatment of OSA.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### Image Aesthetics Assessment via Learnable Queries
- **Authors:** Zhiwei Xiong, Yunfan Zhang, Zhiqi Shen, Peiran Ren, Han Yu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.02861
- **Pdf link:** https://arxiv.org/pdf/2309.02861
- **Abstract**
Image aesthetics assessment (IAA) aims to estimate the aesthetics of images. Depending on the content of an image, diverse criteria need to be selected to assess its aesthetics. Existing works utilize pre-trained vision backbones based on content knowledge to learn image aesthetics. However, training those backbones is time-consuming and suffers from attention dispersion. Inspired by learnable queries in vision-language alignment, we propose the Image Aesthetics Assessment via Learnable Queries (IAA-LQ) approach. It adapts learnable queries to extract aesthetic features from pre-trained image features obtained from a frozen image encoder. Extensive experiments on real-world data demonstrate the advantages of IAA-LQ, beating the best state-of-the-art method by 2.2% and 2.1% in terms of SRCC and PLCC, respectively.
### Do We Still Need Non-Maximum Suppression? Accurate Confidence Estimates and Implicit Duplication Modeling with IoU-Aware Calibration
- **Authors:** Johannes Gilg, Torben Teepe, Fabian Herzog, Philipp Wolters, Gerhard Rigoll
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2309.03110
- **Pdf link:** https://arxiv.org/pdf/2309.03110
- **Abstract**
Object detectors are at the heart of many semi- and fully autonomous decision systems and are poised to become even more indispensable. They are, however, still lacking in accessibility and can sometimes produce unreliable predictions. Especially concerning in this regard are the -- essentially hand-crafted -- non-maximum suppression algorithms that lead to an obfuscated prediction process and biased confidence estimates. We show that we can eliminate classic NMS-style post-processing by using IoU-aware calibration. IoU-aware calibration is a conditional Beta calibration; this makes it parallelizable with no hyper-parameters. Instead of arbitrary cutoffs or discounts, it implicitly accounts for the likelihood of each detection being a duplicate and adjusts the confidence score accordingly, resulting in empirically based precision estimates for each detection. Our extensive experiments on diverse detection architectures show that the proposed IoU-aware calibration can successfully model duplicate detections and improve calibration. Compared to the standard sequential NMS and calibration approach, our joint modeling can deliver performance gains over the best NMS-based alternative while producing consistently better-calibrated confidence predictions with less complexity. The \hyperlink{https://github.com/Blueblue4/IoU-AwareCalibration}{code} for all our experiments is publicly available.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### Compressing Vision Transformers for Low-Resource Visual Learning
- **Authors:** Eric Youn, Sai Mitheran J, Sanjana Prabhu, Siyuan Chen
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2309.02617
- **Pdf link:** https://arxiv.org/pdf/2309.02617
- **Abstract**
Vision transformer (ViT) and its variants have swept through visual learning leaderboards and offer state-of-the-art accuracy in tasks such as image classification, object detection, and semantic segmentation by attending to different parts of the visual input and capturing long-range spatial dependencies. However, these models are large and computation-heavy. For instance, the recently proposed ViT-B model has 86M parameters making it impractical for deployment on resource-constrained devices. As a result, their deployment on mobile and edge scenarios is limited. In our work, we aim to take a step toward bringing vision transformers to the edge by utilizing popular model compression techniques such as distillation, pruning, and quantization. Our chosen application environment is an unmanned aerial vehicle (UAV) that is battery-powered and memory-constrained, carrying a single-board computer on the scale of an NVIDIA Jetson Nano with 4GB of RAM. On the other hand, the UAV requires high accuracy close to that of state-of-the-art ViTs to ensure safe object avoidance in autonomous navigation, or correct localization of humans in search-and-rescue. Inference latency should also be minimized given the application requirements. Hence, our target is to enable rapid inference of a vision transformer on an NVIDIA Jetson Nano (4GB) with minimal accuracy loss. This allows us to deploy ViTs on resource-constrained devices, opening up new possibilities in surveillance, environmental monitoring, etc. Our implementation is made available at https://github.com/chensy7/efficient-vit.
### Bandwidth-efficient Inference for Neural Image Compression
- **Authors:** Shanzhi Yin, Tongda Xu, Yongsheng Liang, Yuanyuan Wang, Yanghao Li, Yan Wang, Jingjing Liu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2309.02855
- **Pdf link:** https://arxiv.org/pdf/2309.02855
- **Abstract**
With neural networks growing deeper and feature maps growing larger, limited communication bandwidth with external memory (or DRAM) and power constraints become a bottleneck in implementing network inference on mobile and edge devices. In this paper, we propose an end-to-end differentiable bandwidth efficient neural inference method with the activation compressed by neural data compression method. Specifically, we propose a transform-quantization-entropy coding pipeline for activation compression with symmetric exponential Golomb coding and a data-dependent Gaussian entropy model for arithmetic coding. Optimized with existing model quantization methods, low-level task of image compression can achieve up to 19x bandwidth reduction with 6.21x energy saving.
## Keyword: RAW
### FArMARe: a Furniture-Aware Multi-task methodology for Recommending Apartments based on the user interests
- **Authors:** Ali Abdari, Alex Falcon, Giuseppe Serra
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Multimedia (cs.MM)
- **Arxiv link:** https://arxiv.org/abs/2309.03100
- **Pdf link:** https://arxiv.org/pdf/2309.03100
- **Abstract**
Nowadays, many people frequently have to search for new accommodation options. Searching for a suitable apartment is a time-consuming process, especially because visiting them is often mandatory to assess the truthfulness of the advertisements found on the Web. While this process could be alleviated by visiting the apartments in the metaverse, the Web-based recommendation platforms are not suitable for the task. To address this shortcoming, in this paper, we define a new problem called text-to-apartment recommendation, which requires ranking the apartments based on their relevance to a textual query expressing the user's interests. To tackle this problem, we introduce FArMARe, a multi-task approach that supports cross-modal contrastive training with a furniture-aware objective. Since public datasets related to indoor scenes do not contain detailed descriptions of the furniture, we collect and annotate a dataset comprising more than 6000 apartments. A thorough experimentation with three different methods and two raw feature extraction procedures reveals the effectiveness of FArMARe in dealing with the problem at hand.
## Keyword: raw image
There is no result
| process | new submissions for thu sep keyword events domain adaptation for efficiently fine tuning vision transformer with encrypted images authors teru nagamori sayaka shiota hitoshi kiya subjects computer vision and pattern recognition cs cv cryptography and security cs cr machine learning cs lg arxiv link pdf link abstract in recent years deep neural networks dnns trained with transformed data have been applied to various applications such as privacy preserving learning access control and adversarial defenses however the use of transformed data decreases the performance of models accordingly in this paper we propose a novel method for fine tuning models with transformed images under the use of the vision transformer vit the proposed domain adaptation method does not cause the accuracy degradation of models and it is carried out on the basis of the embedding structure of vit in experiments we confirmed that the proposed method prevents accuracy degradation even when using encrypted images with the cifar and cifar datasets slaction non intrusive lightweight obstructive sleep apnea detection using infrared video authors you rim choi gyeongseon eo wonhyuck youn hyojin lee haemin jang dongyoon kim hyunwoo shin hyung sin kim subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract obstructive sleep apnea osa is a prevalent sleep disorder affecting approximately one billion people world wide the current gold standard for diagnosing osa polysomnography psg involves an overnight hospital stay with multiple attached sensors leading to potential inaccuracies due to the first night effect to address this we present slaction a non intrusive osa detection system for daily sleep environments using infrared videos recognizing that sleep videos exhibit minimal motion this work investigates the fundamental question are respiratory events adequately reflected in human motions during sleep analyzing the largest sleep video dataset of hours we establish correlations between osa events and human motions during sleep our approach uses a low frame rate fps a large size seconds and step seconds for sliding window analysis to capture slow and long term motions related to osa furthermore we utilize a lightweight deep neural network for resource constrained devices ensuring all video streams are processed locally without compromising privacy evaluations show that slaction achieves an average score of in detecting osa across various environments implementing slaction on nvidia jetson nano enables real time inference seconds for a second video clip highlighting its potential for early detection and personalized treatment of osa keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp image aesthetics assessment via learnable queries authors zhiwei xiong yunfan zhang zhiqi shen peiran ren han yu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract image aesthetics assessment iaa aims to estimate the aesthetics of images depending on the content of an image diverse criteria need to be selected to assess its aesthetics existing works utilize pre trained vision backbones based on content knowledge to learn image aesthetics however training those backbones is time consuming and suffers from attention dispersion inspired by learnable queries in vision language alignment we propose the image aesthetics assessment via learnable queries iaa lq approach it adapts learnable queries to extract aesthetic features from pre trained image features obtained from a frozen image encoder extensive experiments on real world data demonstrate the advantages of iaa lq beating the best state of the art method by and in terms of srcc and plcc respectively do we still need non maximum suppression accurate confidence estimates and implicit duplication modeling with iou aware calibration authors johannes gilg torben teepe fabian herzog philipp wolters gerhard rigoll subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract object detectors are at the heart of many semi and fully autonomous decision systems and are poised to become even more indispensable they are however still lacking in accessibility and can sometimes produce unreliable predictions especially concerning in this regard are the essentially hand crafted non maximum suppression algorithms that lead to an obfuscated prediction process and biased confidence estimates we show that we can eliminate classic nms style post processing by using iou aware calibration iou aware calibration is a conditional beta calibration this makes it parallelizable with no hyper parameters instead of arbitrary cutoffs or discounts it implicitly accounts for the likelihood of each detection being a duplicate and adjusts the confidence score accordingly resulting in empirically based precision estimates for each detection our extensive experiments on diverse detection architectures show that the proposed iou aware calibration can successfully model duplicate detections and improve calibration compared to the standard sequential nms and calibration approach our joint modeling can deliver performance gains over the best nms based alternative while producing consistently better calibrated confidence predictions with less complexity the hyperlink for all our experiments is publicly available keyword image signal processing there is no result keyword image signal process there is no result keyword compression compressing vision transformers for low resource visual learning authors eric youn sai mitheran j sanjana prabhu siyuan chen subjects computer vision and pattern recognition cs cv machine learning cs lg arxiv link pdf link abstract vision transformer vit and its variants have swept through visual learning leaderboards and offer state of the art accuracy in tasks such as image classification object detection and semantic segmentation by attending to different parts of the visual input and capturing long range spatial dependencies however these models are large and computation heavy for instance the recently proposed vit b model has parameters making it impractical for deployment on resource constrained devices as a result their deployment on mobile and edge scenarios is limited in our work we aim to take a step toward bringing vision transformers to the edge by utilizing popular model compression techniques such as distillation pruning and quantization our chosen application environment is an unmanned aerial vehicle uav that is battery powered and memory constrained carrying a single board computer on the scale of an nvidia jetson nano with of ram on the other hand the uav requires high accuracy close to that of state of the art vits to ensure safe object avoidance in autonomous navigation or correct localization of humans in search and rescue inference latency should also be minimized given the application requirements hence our target is to enable rapid inference of a vision transformer on an nvidia jetson nano with minimal accuracy loss this allows us to deploy vits on resource constrained devices opening up new possibilities in surveillance environmental monitoring etc our implementation is made available at bandwidth efficient inference for neural image compression authors shanzhi yin tongda xu yongsheng liang yuanyuan wang yanghao li yan wang jingjing liu subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract with neural networks growing deeper and feature maps growing larger limited communication bandwidth with external memory or dram and power constraints become a bottleneck in implementing network inference on mobile and edge devices in this paper we propose an end to end differentiable bandwidth efficient neural inference method with the activation compressed by neural data compression method specifically we propose a transform quantization entropy coding pipeline for activation compression with symmetric exponential golomb coding and a data dependent gaussian entropy model for arithmetic coding optimized with existing model quantization methods low level task of image compression can achieve up to bandwidth reduction with energy saving keyword raw farmare a furniture aware multi task methodology for recommending apartments based on the user interests authors ali abdari alex falcon giuseppe serra subjects computer vision and pattern recognition cs cv multimedia cs mm arxiv link pdf link abstract nowadays many people frequently have to search for new accommodation options searching for a suitable apartment is a time consuming process especially because visiting them is often mandatory to assess the truthfulness of the advertisements found on the web while this process could be alleviated by visiting the apartments in the metaverse the web based recommendation platforms are not suitable for the task to address this shortcoming in this paper we define a new problem called text to apartment recommendation which requires ranking the apartments based on their relevance to a textual query expressing the user s interests to tackle this problem we introduce farmare a multi task approach that supports cross modal contrastive training with a furniture aware objective since public datasets related to indoor scenes do not contain detailed descriptions of the furniture we collect and annotate a dataset comprising more than apartments a thorough experimentation with three different methods and two raw feature extraction procedures reveals the effectiveness of farmare in dealing with the problem at hand keyword raw image there is no result | 1 |
14,233 | 17,154,427,825 | IssuesEvent | 2021-07-14 03:49:57 | amor71/LiuAlgoTrader | https://api.github.com/repos/amor71/LiuAlgoTrader | closed | ML support | enhancement in-process no-issue-activity | **Is your feature request related to a problem? Please describe.**
some strategies seem to work well on certain stocks and certain setups, and less for others.
**Describe the solution you'd like**
extend market miner for off-market calculation of a NN that will predict the affinity of stock to a specific strategy
**Describe alternatives you've considered**
direct caluclatuon
| 1.0 | ML support - **Is your feature request related to a problem? Please describe.**
some strategies seem to work well on certain stocks and certain setups, and less for others.
**Describe the solution you'd like**
extend market miner for off-market calculation of a NN that will predict the affinity of stock to a specific strategy
**Describe alternatives you've considered**
direct caluclatuon
| process | ml support is your feature request related to a problem please describe some strategies seem to work well on certain stocks and certain setups and less for others describe the solution you d like extend market miner for off market calculation of a nn that will predict the affinity of stock to a specific strategy describe alternatives you ve considered direct caluclatuon | 1 |
477,301 | 13,759,486,474 | IssuesEvent | 2020-10-07 03:09:59 | AY2021S1-CS2113T-T12-4/tp | https://api.github.com/repos/AY2021S1-CS2113T-T12-4/tp | closed | Implement a standard DateTime parser | priority.High type.Task | The DateTime parser will allow scheduler--; to accept both 24hr and 12hr format. | 1.0 | Implement a standard DateTime parser - The DateTime parser will allow scheduler--; to accept both 24hr and 12hr format. | non_process | implement a standard datetime parser the datetime parser will allow scheduler to accept both and format | 0 |
131,372 | 18,276,432,786 | IssuesEvent | 2021-10-04 19:25:26 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Last item of the menu is obscured by the hovered URL and it's not possible to scroll a bit past that | bug Team:Kibana-Design design-only responsive design | **Kibana version:**
7.14
**Elasticsearch version:**
**Server OS version:**
**Browser version:**
Chrome: 91.0.4472.114
**Browser OS version:**
OS X
**Original install method (e.g. download page, yum, from source, etc.):**
**Describe the bug:**
Needlessly obscured menu item
**Steps to reproduce:**
1. Scroll to the bottom of the long side menu
2. Hover over an item
3. The last menu item is no longer visible, due to the tooltip obscuring it
**Expected behavior:**
Retaining visibility, eg. adding a more substantial bottom padding so the user scrolls the last item out of harm's way if they scroll to the very bottom
**Screenshots (if relevant):**

**Errors in browser console (if relevant):**
**Provide logs and/or server output (if relevant):**
**Any additional context:**
cc @yuliacech | 3.0 | Last item of the menu is obscured by the hovered URL and it's not possible to scroll a bit past that - **Kibana version:**
7.14
**Elasticsearch version:**
**Server OS version:**
**Browser version:**
Chrome: 91.0.4472.114
**Browser OS version:**
OS X
**Original install method (e.g. download page, yum, from source, etc.):**
**Describe the bug:**
Needlessly obscured menu item
**Steps to reproduce:**
1. Scroll to the bottom of the long side menu
2. Hover over an item
3. The last menu item is no longer visible, due to the tooltip obscuring it
**Expected behavior:**
Retaining visibility, eg. adding a more substantial bottom padding so the user scrolls the last item out of harm's way if they scroll to the very bottom
**Screenshots (if relevant):**

**Errors in browser console (if relevant):**
**Provide logs and/or server output (if relevant):**
**Any additional context:**
cc @yuliacech | non_process | last item of the menu is obscured by the hovered url and it s not possible to scroll a bit past that kibana version elasticsearch version server os version browser version chrome browser os version os x original install method e g download page yum from source etc describe the bug needlessly obscured menu item steps to reproduce scroll to the bottom of the long side menu hover over an item the last menu item is no longer visible due to the tooltip obscuring it expected behavior retaining visibility eg adding a more substantial bottom padding so the user scrolls the last item out of harm s way if they scroll to the very bottom screenshots if relevant errors in browser console if relevant provide logs and or server output if relevant any additional context cc yuliacech | 0 |
248,011 | 20,988,903,591 | IssuesEvent | 2022-03-29 07:27:23 | opencurve/curve | https://api.github.com/repos/opencurve/curve | closed | curvefs_tool query-fs return wrong fs | bug need test | **Describe the bug (描述bug)**

**To Reproduce (复现方法)**
**Expected behavior (期望行为)**
不要默认参数了,要么传fsid,要么传fsname,两个都不传或者两个都传就报错。如果命令写错了,展示example。
Don't use default parameters. Either pass fsid or fsname. If you pass neither or both, it will report an error. If the command is wrong, show example.
**Versions (各种版本)**
curvefs_tool: 50fe52bb+debug
**Additional context/screenshots (更多上下文/截图)**
| 1.0 | curvefs_tool query-fs return wrong fs - **Describe the bug (描述bug)**

**To Reproduce (复现方法)**
**Expected behavior (期望行为)**
不要默认参数了,要么传fsid,要么传fsname,两个都不传或者两个都传就报错。如果命令写错了,展示example。
Don't use default parameters. Either pass fsid or fsname. If you pass neither or both, it will report an error. If the command is wrong, show example.
**Versions (各种版本)**
curvefs_tool: 50fe52bb+debug
**Additional context/screenshots (更多上下文/截图)**
| non_process | curvefs tool query fs return wrong fs describe the bug 描述bug to reproduce 复现方法 expected behavior 期望行为 不要默认参数了,要么传fsid,要么传fsname,两个都不传或者两个都传就报错。如果命令写错了,展示example。 don t use default parameters either pass fsid or fsname if you pass neither or both it will report an error if the command is wrong show example versions 各种版本 curvefs tool debug additional context screenshots 更多上下文 截图 | 0 |
5,992 | 8,805,374,888 | IssuesEvent | 2018-12-26 19:14:06 | dita-ot/dita-ot | https://api.github.com/repos/dita-ot/dita-ot | closed | Improve error message for keyref that includes topic ID | enhancement preprocess/keyref priority/medium stale | GIven a keyref like this:
```
<xref keyref="getting-started/concept_qd2_gt4_t5"/>.
```
Where "concept_qd2_gt4_t5" is the ID of a topic (not an element within a topic), the OT fails to resolve the reference and does not produce a working link.
The keyref is incorrect as authored: you are only supposed to include non-topic element IDs, so this case is author error.
However, the OT could detect this error and report it as "You specified a topic ID as part of a key ref, that won't work."
In addition, it could also go ahead and treat the keyref as a keyref to the topic since that's what the author thought they were requesting.
The current behavior is to simply report the reference as unresolvable. But using oXygen, for example, the link is resolved (because oXygen is clearly not distinguishing between keyrefs to topics by ID an other keyrefs--they should also report this issue or simply not allow it to occur). So it's confusing to a typical user as to why the link didn't work when clearly the reference is correct.
It took me a while to figure out what I had done wrong, and that was only in the context of making this issue report. And I helped design the keyref mechanism.
| 1.0 | Improve error message for keyref that includes topic ID - GIven a keyref like this:
```
<xref keyref="getting-started/concept_qd2_gt4_t5"/>.
```
Where "concept_qd2_gt4_t5" is the ID of a topic (not an element within a topic), the OT fails to resolve the reference and does not produce a working link.
The keyref is incorrect as authored: you are only supposed to include non-topic element IDs, so this case is author error.
However, the OT could detect this error and report it as "You specified a topic ID as part of a key ref, that won't work."
In addition, it could also go ahead and treat the keyref as a keyref to the topic since that's what the author thought they were requesting.
The current behavior is to simply report the reference as unresolvable. But using oXygen, for example, the link is resolved (because oXygen is clearly not distinguishing between keyrefs to topics by ID an other keyrefs--they should also report this issue or simply not allow it to occur). So it's confusing to a typical user as to why the link didn't work when clearly the reference is correct.
It took me a while to figure out what I had done wrong, and that was only in the context of making this issue report. And I helped design the keyref mechanism.
| process | improve error message for keyref that includes topic id given a keyref like this where concept is the id of a topic not an element within a topic the ot fails to resolve the reference and does not produce a working link the keyref is incorrect as authored you are only supposed to include non topic element ids so this case is author error however the ot could detect this error and report it as you specified a topic id as part of a key ref that won t work in addition it could also go ahead and treat the keyref as a keyref to the topic since that s what the author thought they were requesting the current behavior is to simply report the reference as unresolvable but using oxygen for example the link is resolved because oxygen is clearly not distinguishing between keyrefs to topics by id an other keyrefs they should also report this issue or simply not allow it to occur so it s confusing to a typical user as to why the link didn t work when clearly the reference is correct it took me a while to figure out what i had done wrong and that was only in the context of making this issue report and i helped design the keyref mechanism | 1 |
9,483 | 2,906,017,775 | IssuesEvent | 2015-06-19 06:51:14 | XVincentX/pollsApiClient | https://api.github.com/repos/XVincentX/pollsApiClient | opened | Pre announcement bugs | Blocking Design Engineering | ## To do List
- [ ] CSS are not prefixed; thus it does not work on **Safari**
- [ ] Flex layout is not aligned in _phone_ mode [Image](https://cloud.githubusercontent.com/assets/1416224/8248440/6256f00e-165f-11e5-8fa3-a2d2707da759.png)
- [ ] Fork on Github ribbon covers the title in _phone_ mode
- [ ] Missing validation and messages from **formly**
- [ ] Missing error messages from poll | 1.0 | Pre announcement bugs - ## To do List
- [ ] CSS are not prefixed; thus it does not work on **Safari**
- [ ] Flex layout is not aligned in _phone_ mode [Image](https://cloud.githubusercontent.com/assets/1416224/8248440/6256f00e-165f-11e5-8fa3-a2d2707da759.png)
- [ ] Fork on Github ribbon covers the title in _phone_ mode
- [ ] Missing validation and messages from **formly**
- [ ] Missing error messages from poll | non_process | pre announcement bugs to do list css are not prefixed thus it does not work on safari flex layout is not aligned in phone mode fork on github ribbon covers the title in phone mode missing validation and messages from formly missing error messages from poll | 0 |
17,898 | 23,873,234,943 | IssuesEvent | 2022-09-07 16:29:57 | DSpace/dspace-angular | https://api.github.com/repos/DSpace/dspace-angular | closed | Allow for Processes to be deleted from the User Interface | bug improvement help wanted high priority e/8 tools:processes | **Is your feature request related to a problem? Please describe.**
Processes currently cannot be deleted from the User Interface, even though `DELETE` is possible from the REST API: https://github.com/DSpace/RestContract/blob/main/processes-endpoint.md#execution-deletion
Without this feature, the list of processes will just grow and grow, even though they can be deleted via the REST API.
**Describe the solution you'd like**
At a minimum, we should allow individual processes to be deleted (add a delete button when viewing a process).
It also might be nice to provide a way to bulk delete processes, either via a series of checkboxes, or via a new `processes_cleanup` script which can delete all processes which are greater than ___ days old. | 1.0 | Allow for Processes to be deleted from the User Interface - **Is your feature request related to a problem? Please describe.**
Processes currently cannot be deleted from the User Interface, even though `DELETE` is possible from the REST API: https://github.com/DSpace/RestContract/blob/main/processes-endpoint.md#execution-deletion
Without this feature, the list of processes will just grow and grow, even though they can be deleted via the REST API.
**Describe the solution you'd like**
At a minimum, we should allow individual processes to be deleted (add a delete button when viewing a process).
It also might be nice to provide a way to bulk delete processes, either via a series of checkboxes, or via a new `processes_cleanup` script which can delete all processes which are greater than ___ days old. | process | allow for processes to be deleted from the user interface is your feature request related to a problem please describe processes currently cannot be deleted from the user interface even though delete is possible from the rest api without this feature the list of processes will just grow and grow even though they can be deleted via the rest api describe the solution you d like at a minimum we should allow individual processes to be deleted add a delete button when viewing a process it also might be nice to provide a way to bulk delete processes either via a series of checkboxes or via a new processes cleanup script which can delete all processes which are greater than days old | 1 |
58,454 | 3,089,494,541 | IssuesEvent | 2015-08-25 21:48:03 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | Persistent services, replication controllers and pods | kind/support priority/P3 team/mesosphere | I am currently running kubernetes on top of mesos in a development cluster using some VMs. One of the things I am interested in is persistent services, replication controllers and pods.
Currently, if I deploy a service, replication controller or pod and reboot the whole cluster, all my deployments are lost. I would need to run `kubectl create ...` to recreate them.
That means, if a kubenetes cluster loses power, there needs to be some mechanism to recreate those pods, services and rcs when it comes back up.
Are there any plans to put those things into long term persistence so that when the whole cluster is restarted, the pods, rcs and services can automatically come back up?
I think this would also be quite useful to allow upgrade the scheduler and the various components easily.
xref https://github.com/mesosphere/kubernetes-mesos/issues/446 | 1.0 | Persistent services, replication controllers and pods - I am currently running kubernetes on top of mesos in a development cluster using some VMs. One of the things I am interested in is persistent services, replication controllers and pods.
Currently, if I deploy a service, replication controller or pod and reboot the whole cluster, all my deployments are lost. I would need to run `kubectl create ...` to recreate them.
That means, if a kubenetes cluster loses power, there needs to be some mechanism to recreate those pods, services and rcs when it comes back up.
Are there any plans to put those things into long term persistence so that when the whole cluster is restarted, the pods, rcs and services can automatically come back up?
I think this would also be quite useful to allow upgrade the scheduler and the various components easily.
xref https://github.com/mesosphere/kubernetes-mesos/issues/446 | non_process | persistent services replication controllers and pods i am currently running kubernetes on top of mesos in a development cluster using some vms one of the things i am interested in is persistent services replication controllers and pods currently if i deploy a service replication controller or pod and reboot the whole cluster all my deployments are lost i would need to run kubectl create to recreate them that means if a kubenetes cluster loses power there needs to be some mechanism to recreate those pods services and rcs when it comes back up are there any plans to put those things into long term persistence so that when the whole cluster is restarted the pods rcs and services can automatically come back up i think this would also be quite useful to allow upgrade the scheduler and the various components easily xref | 0 |
5,793 | 8,638,873,353 | IssuesEvent | 2018-11-23 16:14:23 | Dweepa/DataAnalytics- | https://api.github.com/repos/Dweepa/DataAnalytics- | closed | Dimensionality reduction | Cleaning and pre-processing | Check if dimension reduction is required and make the necessary reductions. | 1.0 | Dimensionality reduction - Check if dimension reduction is required and make the necessary reductions. | process | dimensionality reduction check if dimension reduction is required and make the necessary reductions | 1 |
17,710 | 23,606,647,131 | IssuesEvent | 2022-08-24 08:50:01 | Data-Product-Business/open-data-product-spec | https://api.github.com/repos/Data-Product-Business/open-data-product-spec | opened | Add product content Schema to the Product level | enhancement unprocessed | **Idea Description**
Data Product consumers want to see the content schema and access it to validate stream from the product in case it is for example API driven.
The Schema could be inline, internal reference or as link. The example sketches below
**Examples**
**As inline**
`Schema
type= "inline"
value ="schema elements here inside string..."
`
**As ref**
The Schema would be defined separately in the spec. This option is not applicable as is now. This would require more changes is the specification.
`Schema
type= "ref"
value ="$stream"
`
**As url**
`Schema
type= "url"
value ="https://someurl.com/schema.json" (must either JSON or XML)
`
| 1.0 | Add product content Schema to the Product level - **Idea Description**
Data Product consumers want to see the content schema and access it to validate stream from the product in case it is for example API driven.
The Schema could be inline, internal reference or as link. The example sketches below
**Examples**
**As inline**
`Schema
type= "inline"
value ="schema elements here inside string..."
`
**As ref**
The Schema would be defined separately in the spec. This option is not applicable as is now. This would require more changes is the specification.
`Schema
type= "ref"
value ="$stream"
`
**As url**
`Schema
type= "url"
value ="https://someurl.com/schema.json" (must either JSON or XML)
`
| process | add product content schema to the product level idea description data product consumers want to see the content schema and access it to validate stream from the product in case it is for example api driven the schema could be inline internal reference or as link the example sketches below examples as inline schema type inline value schema elements here inside string as ref the schema would be defined separately in the spec this option is not applicable as is now this would require more changes is the specification schema type ref value stream as url schema type url value must either json or xml | 1 |
11,154 | 13,957,693,511 | IssuesEvent | 2020-10-24 08:11:00 | alexanderkotsev/geoportal | https://api.github.com/repos/alexanderkotsev/geoportal | opened | NL: Harvesting request | Geoportal Harvesting process NL - The Netherlands | From: Sanders, Mark [Mark.Sanders@kadaster.nl]
Sent: 21 January 2019 10:32
To: JRC INSPIRE SUPPORT
Subject: Geoportal Helpdesk
Dear Helpdesk,
The Geoportal harvester which is configured to harvest metadata weekly for the NL inspire datasets did not run last Friday.
We are frequently updating the metadata for the NL Inspire dataset to comply with the Inspire regulations. Therefore we would benefit from a more frequently running harvester. Could you please change the harvest frequency from weekly to daily.
Kind regards,
Mark Sanders
Softwareontwikkelaar PDOK | 1.0 | NL: Harvesting request - From: Sanders, Mark [Mark.Sanders@kadaster.nl]
Sent: 21 January 2019 10:32
To: JRC INSPIRE SUPPORT
Subject: Geoportal Helpdesk
Dear Helpdesk,
The Geoportal harvester which is configured to harvest metadata weekly for the NL inspire datasets did not run last Friday.
We are frequently updating the metadata for the NL Inspire dataset to comply with the Inspire regulations. Therefore we would benefit from a more frequently running harvester. Could you please change the harvest frequency from weekly to daily.
Kind regards,
Mark Sanders
Softwareontwikkelaar PDOK | process | nl harvesting request from sanders mark sent january to jrc inspire support subject geoportal helpdesk dear helpdesk the geoportal harvester which is configured to harvest metadata weekly for the nl inspire datasets did not run last friday we are frequently updating the metadata for the nl inspire dataset to comply with the inspire regulations therefore we would benefit from a more frequently running harvester could you please change the harvest frequency from weekly to daily kind regards mark sanders softwareontwikkelaar pdok | 1 |
5,302 | 8,121,861,602 | IssuesEvent | 2018-08-16 09:34:44 | openvstorage/framework | https://api.github.com/repos/openvstorage/framework | closed | template's ownership | process_wontfix | Once a vdisk is set as a template, it can't be used as a vdisk any more.
In the vpool's directory, there'are bunch of vdisk.raws. It's very nice that I can manipulate vdisk.raw and install a VM on it. But I have to remember which one is a template that I shouldn't use for this operation. Is a template really special or different from a vdisk? | 1.0 | template's ownership - Once a vdisk is set as a template, it can't be used as a vdisk any more.
In the vpool's directory, there'are bunch of vdisk.raws. It's very nice that I can manipulate vdisk.raw and install a VM on it. But I have to remember which one is a template that I shouldn't use for this operation. Is a template really special or different from a vdisk? | process | template s ownership once a vdisk is set as a template it can t be used as a vdisk any more in the vpool s directory there are bunch of vdisk raws it s very nice that i can manipulate vdisk raw and install a vm on it but i have to remember which one is a template that i shouldn t use for this operation is a template really special or different from a vdisk | 1 |
98,857 | 16,389,517,882 | IssuesEvent | 2021-05-17 14:31:22 | Thanraj/linux-1 | https://api.github.com/repos/Thanraj/linux-1 | opened | CVE-2019-19071 (High) detected in linuxv5.0 | security vulnerability | ## CVE-2019-19071 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.0</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-1/commits/9738d89d33cb0f3ac708908509b82eafc007d557">9738d89d33cb0f3ac708908509b82eafc007d557</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-1/drivers/net/wireless/rsi/rsi_91x_mgmt.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-1/drivers/net/wireless/rsi/rsi_91x_mgmt.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A memory leak in the rsi_send_beacon() function in drivers/net/wireless/rsi/rsi_91x_mgmt.c in the Linux kernel through 5.3.11 allows attackers to cause a denial of service (memory consumption) by triggering rsi_prepare_beacon() failures, aka CID-d563131ef23c.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19071>CVE-2019-19071</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2019-19071">https://www.linuxkernelcves.com/cves/CVE-2019-19071</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: v5.5-rc1,v4.14.159,v4.19.89,v5.3.16,v5.4.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-19071 (High) detected in linuxv5.0 - ## CVE-2019-19071 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.0</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-1/commits/9738d89d33cb0f3ac708908509b82eafc007d557">9738d89d33cb0f3ac708908509b82eafc007d557</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-1/drivers/net/wireless/rsi/rsi_91x_mgmt.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-1/drivers/net/wireless/rsi/rsi_91x_mgmt.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A memory leak in the rsi_send_beacon() function in drivers/net/wireless/rsi/rsi_91x_mgmt.c in the Linux kernel through 5.3.11 allows attackers to cause a denial of service (memory consumption) by triggering rsi_prepare_beacon() failures, aka CID-d563131ef23c.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19071>CVE-2019-19071</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2019-19071">https://www.linuxkernelcves.com/cves/CVE-2019-19071</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: v5.5-rc1,v4.14.159,v4.19.89,v5.3.16,v5.4.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch master vulnerable source files linux drivers net wireless rsi rsi mgmt c linux drivers net wireless rsi rsi mgmt c vulnerability details a memory leak in the rsi send beacon function in drivers net wireless rsi rsi mgmt c in the linux kernel through allows attackers to cause a denial of service memory consumption by triggering rsi prepare beacon failures aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
15,733 | 19,908,359,588 | IssuesEvent | 2022-01-25 14:56:43 | hashicorp/terraform-cdk | https://api.github.com/repos/hashicorp/terraform-cdk | closed | Build a Vagrant Box for Development | enhancement waiting-on-answer dev-process | <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
It would be great to have a [Vagrant](https://www.vagrantup.com/) setup which contains everything we need for development and testing both on Linux and Windows.
The current Docker image works, but it's not a great experience (at least on a Mac).
<!--- Please leave a helpful description of the feature request here. --->
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation?
--->
| 1.0 | Build a Vagrant Box for Development - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
It would be great to have a [Vagrant](https://www.vagrantup.com/) setup which contains everything we need for development and testing both on Linux and Windows.
The current Docker image works, but it's not a great experience (at least on a Mac).
<!--- Please leave a helpful description of the feature request here. --->
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation?
--->
| process | build a vagrant box for development community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description it would be great to have a setup which contains everything we need for development and testing both on linux and windows the current docker image works but it s not a great experience at least on a mac references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation | 1 |
795,388 | 28,071,238,071 | IssuesEvent | 2023-03-29 19:14:44 | svthalia/Reaxit | https://api.github.com/repos/svthalia/Reaxit | closed | App closes when going back from menu drawer screen | bug priority: medium | ### Describe the bug
When going back from a screen opened from the menu drawer the app closes instead of going back to the welcome screen.
### How to reproduce
Steps to reproduce the behavior:
1. Go to screen from the menu drawer.
2. Click on android back button.
3. The app closes.
### Expected behaviour
The welcome screen opens.
| 1.0 | App closes when going back from menu drawer screen - ### Describe the bug
When going back from a screen opened from the menu drawer the app closes instead of going back to the welcome screen.
### How to reproduce
Steps to reproduce the behavior:
1. Go to screen from the menu drawer.
2. Click on android back button.
3. The app closes.
### Expected behaviour
The welcome screen opens.
| non_process | app closes when going back from menu drawer screen describe the bug when going back from a screen opened from the menu drawer the app closes instead of going back to the welcome screen how to reproduce steps to reproduce the behavior go to screen from the menu drawer click on android back button the app closes expected behaviour the welcome screen opens | 0 |
63,120 | 6,826,605,651 | IssuesEvent | 2017-11-08 14:39:33 | sympy/sympy | https://api.github.com/repos/sympy/sympy | closed | Add unit test for 'bosonic' operator | Easy to Fix physics Testing | Look at pull request https://github.com/sympy/sympy/pull/10677 . It adds latex string for "Bosonic operator", though no unit test has been added.
The task is to add write unit test case for those changes (i.e for 'Bosonic operator').
**NOTE**: The submitted PR should use the changes submitted PR original author, i.e. ownership of commits should retain with original author of PR #10677. | 1.0 | Add unit test for 'bosonic' operator - Look at pull request https://github.com/sympy/sympy/pull/10677 . It adds latex string for "Bosonic operator", though no unit test has been added.
The task is to add write unit test case for those changes (i.e for 'Bosonic operator').
**NOTE**: The submitted PR should use the changes submitted PR original author, i.e. ownership of commits should retain with original author of PR #10677. | non_process | add unit test for bosonic operator look at pull request it adds latex string for bosonic operator though no unit test has been added the task is to add write unit test case for those changes i e for bosonic operator note the submitted pr should use the changes submitted pr original author i e ownership of commits should retain with original author of pr | 0 |
218,571 | 16,996,886,974 | IssuesEvent | 2021-07-01 07:43:31 | dzhw/zofar | https://api.github.com/repos/dzhw/zofar | closed | exclusive category should be visually demarcated | 3 inSprint prio: 1 status: testing | Especially for multiple choice questions an answer option should be clearly distanced if there is the attribute 'exclusive="true"' in qml.
Preferred would be an automated solution against a css snippet in additional css within the template. | 1.0 | exclusive category should be visually demarcated - Especially for multiple choice questions an answer option should be clearly distanced if there is the attribute 'exclusive="true"' in qml.
Preferred would be an automated solution against a css snippet in additional css within the template. | non_process | exclusive category should be visually demarcated especially for multiple choice questions an answer option should be clearly distanced if there is the attribute exclusive true in qml preferred would be an automated solution against a css snippet in additional css within the template | 0 |
453,000 | 13,063,092,589 | IssuesEvent | 2020-07-30 16:02:20 | kubesphere/kubesphere | https://api.github.com/repos/kubesphere/kubesphere | closed | Lack of configuration, custom monitoring, grayscale publishing and build image permission settings in one cluster project under muti-clustor namespace | area/multicluster kind/bug priority/high | Describe the Bug
Lack of configuration, custom monitoring, grayscale publishing and build image permission settings in one cluster project under muti-clustor namespace
Versions Used
KubeSphere:3.0.0
Environment
testing env
http://139.198.12.26:30880/
How To Reproduce
Steps to reproduce the behavior:
1.Go to 'Access control' from 'platform management' of home page
2.Click on 'multi-cluster-ws' namespace
3.Click 'project management'
4.Click 'one-cluster-project' and then click 'project role' under 'project setting '
5.Create one role and select permission
6.Lack of configuration, custom monitoring, grayscale publishing and build image permission settings



Expected behavior
advise add these permissions
/kind bug
/area multicluster
/assign @zryfish
/milestone 3.0.0 | 1.0 | Lack of configuration, custom monitoring, grayscale publishing and build image permission settings in one cluster project under muti-clustor namespace - Describe the Bug
Lack of configuration, custom monitoring, grayscale publishing and build image permission settings in one cluster project under muti-clustor namespace
Versions Used
KubeSphere:3.0.0
Environment
testing env
http://139.198.12.26:30880/
How To Reproduce
Steps to reproduce the behavior:
1.Go to 'Access control' from 'platform management' of home page
2.Click on 'multi-cluster-ws' namespace
3.Click 'project management'
4.Click 'one-cluster-project' and then click 'project role' under 'project setting '
5.Create one role and select permission
6.Lack of configuration, custom monitoring, grayscale publishing and build image permission settings



Expected behavior
advise add these permissions
/kind bug
/area multicluster
/assign @zryfish
/milestone 3.0.0 | non_process | lack of configuration custom monitoring grayscale publishing and build image permission settings in one cluster project under muti clustor namespace describe the bug lack of configuration custom monitoring grayscale publishing and build image permission settings in one cluster project under muti clustor namespace versions used kubesphere environment testing env how to reproduce steps to reproduce the behavior go to access control from platform management of home page click on multi cluster ws namespace click project management click one cluster project and then click project role under project setting create one role and select permission lack of configuration custom monitoring grayscale publishing and build image permission settings expected behavior advise add these permissions kind bug area multicluster assign zryfish milestone | 0 |
13,095 | 15,443,344,035 | IssuesEvent | 2021-03-08 09:01:12 | pystatgen/sgkit | https://api.github.com/repos/pystatgen/sgkit | closed | Fix mypy errors with Xarray 0.17.0 | process + tools | Xarray 0.17.0 declares types so some mypy ignore statements are no longer needed. | 1.0 | Fix mypy errors with Xarray 0.17.0 - Xarray 0.17.0 declares types so some mypy ignore statements are no longer needed. | process | fix mypy errors with xarray xarray declares types so some mypy ignore statements are no longer needed | 1 |
13,040 | 15,384,951,608 | IssuesEvent | 2021-03-03 05:39:38 | GoogleCloudPlatform/fda-mystudies | https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies | closed | [iOS] Resources > HTTP hyperlinks are not loading in resources section | Bug P1 Process: Fixed Process: Tested QA Process: Tested dev iOS | **Steps:**
1. Edit study
2. Navigate to resources
3. Add a resource
4. Configure HTTP hyperlink (eg. http://google.com)
5. Publish updates
6. Open the link from iOS Mobile
**Actual:** HTTP hyperlinks are not loading in the resources section
**Expected:** HTTP hyperlinks should load | 3.0 | [iOS] Resources > HTTP hyperlinks are not loading in resources section - **Steps:**
1. Edit study
2. Navigate to resources
3. Add a resource
4. Configure HTTP hyperlink (eg. http://google.com)
5. Publish updates
6. Open the link from iOS Mobile
**Actual:** HTTP hyperlinks are not loading in the resources section
**Expected:** HTTP hyperlinks should load | process | resources http hyperlinks are not loading in resources section steps edit study navigate to resources add a resource configure http hyperlink eg publish updates open the link from ios mobile actual http hyperlinks are not loading in the resources section expected http hyperlinks should load | 1 |
59,225 | 11,951,237,274 | IssuesEvent | 2020-04-03 16:31:34 | pookage/atomic-bomberpook | https://api.github.com/repos/pookage/atomic-bomberpook | opened | Create a test scene | 🎨 Concept 🤖 Code | - [ ] Lay out a series of planes in a 15x11 grid
- [ ] Place grey 'indestructable' cubes on every other tile with the exception of the outermost ring
- [ ] Fill all of the grid with green 'destructable' cubes except in a 3x3 grid in the top-left corner
- [ ] Add a pyramid to the top-left corner of the scene to represent the player | 1.0 | Create a test scene - - [ ] Lay out a series of planes in a 15x11 grid
- [ ] Place grey 'indestructable' cubes on every other tile with the exception of the outermost ring
- [ ] Fill all of the grid with green 'destructable' cubes except in a 3x3 grid in the top-left corner
- [ ] Add a pyramid to the top-left corner of the scene to represent the player | non_process | create a test scene lay out a series of planes in a grid place grey indestructable cubes on every other tile with the exception of the outermost ring fill all of the grid with green destructable cubes except in a grid in the top left corner add a pyramid to the top left corner of the scene to represent the player | 0 |
23,060 | 3,755,995,433 | IssuesEvent | 2016-03-13 01:40:27 | libarchive/libarchive | https://api.github.com/repos/libarchive/libarchive | closed | Failed tests on Solaris 10 Sparc | OpSys-All Priority-Medium Type-Defect | Original [issue 313](https://code.google.com/p/libarchive/issues/detail?id=313) created by Google Code user `honkman42` on 2013-04-10T12:52:19.000Z:
```
<b>What steps will reproduce the problem?</b>
1. Compile and run the testsuite
<b>What is the expected output? What do you see instead?</b>
gmake[2]: Entering directory `/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2'
If tests fail or crash, details will be in:
/tmp/libarchive_test.2013-04-10T14.16.07-000
Reference files will be read from: /home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2/libarchive/test
Exercising: libarchive 3.1.2
0: test_acl_freebsd_nfs4 ok
1: test_acl_freebsd_posix1e ok
2: test_acl_nfs4 ok
3: test_acl_pax ok
4: test_acl_posix1e ok
5: test_archive_api_feature ok
6: test_archive_clear_error ok
7: test_archive_cmdline ok
8: test_archive_md5 ok
9: test_archive_rmd160 ok
10: test_archive_sha1 ok
11: test_archive_sha256 ok
12: test_archive_sha384 ok
13: test_archive_sha512 ok
14: test_archive_getdate ok
15: test_archive_match_owner ok
16: test_archive_match_path ok
17: test_archive_match_time ok
18: test_archive_pathmatch ok
19: test_archive_read_close_twice ok
20: test_archive_read_close_twice_open_fd ok
21: test_archive_read_close_twice_open_filename ok
22: test_archive_read_multiple_data_objects ok
23: test_archive_read_next_header_empty ok
24: test_archive_read_next_header_raw ok
25: test_archive_read_open2 ok
26: test_archive_read_set_filter_option ok
27: test_archive_read_set_format_option ok
28: test_archive_read_set_option ok
29: test_archive_read_set_options ok
30: test_archive_read_support ok
31: test_archive_set_error ok
32: test_archive_string ok
33: test_archive_string_conversion ok
34: test_archive_write_add_filter_by_name_b64encode ok
35: test_archive_write_add_filter_by_name_bzip2 ok
36: test_archive_write_add_filter_by_name_compress ok
37: test_archive_write_add_filter_by_name_grzip sh: grzip: not found
ok
38: test_archive_write_add_filter_by_name_gzip ok
39: test_archive_write_add_filter_by_name_lrzip sh: lrzip: not found
ok
40: test_archive_write_add_filter_by_name_lzip ok
41: test_archive_write_add_filter_by_name_lzma ok
42: test_archive_write_add_filter_by_name_lzop ok
43: test_archive_write_add_filter_by_name_uuencode ok
44: test_archive_write_add_filter_by_name_xz ok
45: test_archive_write_set_filter_option ok
46: test_archive_write_set_format_by_name_7zip ok
47: test_archive_write_set_format_by_name_ar ok
48: test_archive_write_set_format_by_name_arbsd ok
49: test_archive_write_set_format_by_name_argnu ok
50: test_archive_write_set_format_by_name_arsvr4 ok
51: test_archive_write_set_format_by_name_bsdtar ok
52: test_archive_write_set_format_by_name_cd9660 ok
53: test_archive_write_set_format_by_name_cpio ok
54: test_archive_write_set_format_by_name_gnutar ok
55: test_archive_write_set_format_by_name_iso ok
56: test_archive_write_set_format_by_name_iso9660 ok
57: test_archive_write_set_format_by_name_mtree ok
58: test_archive_write_set_format_by_name_mtree_classicok
59: test_archive_write_set_format_by_name_newc ok
60: test_archive_write_set_format_by_name_odc ok
61: test_archive_write_set_format_by_name_oldtar ok
62: test_archive_write_set_format_by_name_pax ok
63: test_archive_write_set_format_by_name_paxr ok
64: test_archive_write_set_format_by_name_posix ok
65: test_archive_write_set_format_by_name_rpax ok
66: test_archive_write_set_format_by_name_shar ok
67: test_archive_write_set_format_by_name_shardump ok
68: test_archive_write_set_format_by_name_ustar ok
69: test_archive_write_set_format_by_name_v7tar ok
70: test_archive_write_set_format_by_name_v7 ok
71: test_archive_write_set_format_by_name_xar ok
72: test_archive_write_set_format_by_name_zip ok
73: test_archive_write_set_format_option ok
74: test_archive_write_set_option ok
75: test_archive_write_set_options ok
76: test_bad_fd ok
77: test_compat_bzip2 ok
78: test_compat_cpio ok
79: test_compat_gtar ok
80: test_compat_gzip ok
81: test_compat_lzip ok
82: test_compat_lzma ok
83: test_compat_lzop ok
84: test_compat_mac ok
85: test_compat_pax_libarchive_2x ok
86: test_compat_solaris_pax_sparse ok
87: test_compat_solaris_tar_acl ok
88: test_compat_tar_hardlink ok
89: test_compat_uudecode ok
90: test_compat_xz ok
91: test_compat_zip FAIL
92: test_empty_write ok
93: test_entry ok
94: test_entry_strmode ok
95: test_extattr_freebsd ok
96: test_filter_count ok
97: test_fuzz_ar ok
98: test_fuzz_cab ok
99: test_fuzz_cpio
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--format violated
gzip: stdin: invalid compressed data--crc error
gzip: stdin: unexpected end of file
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: unexpected end of file
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
ok
100: test_fuzz_iso9660 ok
101: test_fuzz_lzh ok
102: test_fuzz_mtree ok
103: test_fuzz_rar ok
104: test_fuzz_tar
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--format violated
gzip: stdin: invalid compressed data--format violated
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: decompression OK, trailing garbage ignored
gzip: stdin: invalid compressed data--crc error
gzip: stdin: decompression OK, trailing garbage ignored
gzip: stdin: decompression OK, trailing garbage ignored
gzip: stdin: unexpected end of file
gzip: stdin: invalid compressed data--format violated
gzip: stdin: decompression OK, trailing garbage ignored
gzip: stdin: invalid compressed data--format violated
gzip: stdin: invalid compressed data--format violated
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: decompression OK, trailing garbage ignored
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: decompression OK, trailing garbage ignored
ok
105: test_fuzz_zip ok
106: test_gnutar_filename_encoding ok
107: test_link_resolver ok
108: test_open_failure ok
109: test_open_fd ok
110: test_open_file ok
111: test_open_filename ok
112: test_pax_filename_encoding ok
113: test_read_data_large ok
114: test_read_disk ok
115: test_read_disk_directory_traversals ok
116: test_read_disk_entry_from_file ok
117: test_read_extract ok
118: test_read_file_nonexistent ok
119: test_read_filter_grzip ok
120: test_read_filter_lrzip ok
121: test_read_filter_lzop ok
122: test_read_filter_lzop_multiple_parts ok
123: test_read_filter_program ok
124: test_read_filter_program_signature ok
125: test_read_filter_uudecode ok
126: test_read_filter_uudecode_base64 ok
127: test_read_format_7zip ok
128: test_read_format_7zip_bzip2 ok
129: test_read_format_7zip_copy ok
130: test_read_format_7zip_deflate ok
131: test_read_format_7zip_empty ok
132: test_read_format_7zip_lzma1 ok
133: test_read_format_7zip_lzma2 ok
134: test_read_format_7zip_ppmd ok
135: test_read_format_ar ok
136: test_read_format_cab FAIL
137: test_read_format_cab_filename ok
...
209: test_read_format_xar ok
210: test_read_format_zip FAIL
211: test_read_format_zip_comment_stored ok
212: test_read_format_zip_filename ok
213: test_read_format_zip_mac_metadata FAIL
214: test_read_format_zip_sfx ok
...
227: test_read_truncated_filter_gzip
gzip: stdin: unexpected end of file
ok
228: test_read_truncated_filter_lzip ok
229: test_read_truncated_filter_lzma ok
230: test_read_truncated_filter_lzop sh: lzop: not found
ok
231: test_read_truncated_filter_xz ok
...
257: test_write_filter_lzma ok
258: test_write_filter_lzop FAIL
259: test_write_filter_program ok
...
308: test_write_zip_set_compression_store ok
309: test_zip_filename_encoding ok
Totals:
Tests run: 310
Tests failed: 5
Assertions checked:19467410
Assertions failed: 549
Skips reported: 148
Failing tests:
91: test_compat_zip (2 failures)
136: test_read_format_cab (530 failures)
210: test_read_format_zip (14 failures)
213: test_read_format_zip_mac_metadata (2 failures)
258: test_write_filter_lzop (1 failures)
Details for failing tests: /tmp/libarchive_test.2013-04-10T14.16.07-000
FAIL: libarchive_test
If tests fail or crash, details will be in:
/tmp/bsdtar_test.2013-04-10T14.21.00-000
Reference files will be read from: /home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2/tar/test
Running tests on: "/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2/bsdtar"
Exercising: bsdtar 3.1.2 - libarchive 3.1.2
0: test_0 ok
1: test_basic ok
2: test_copy ok
3: test_empty_mtree ok
4: test_extract_tar_Z ok
5: test_extract_tar_bz2 ok
6: test_extract_tar_grz sh: grzip: not found
ok
7: test_extract_tar_gz ok
8: test_extract_tar_lrz sh: lrzip: not found
ok
9: test_extract_tar_lz ok
10: test_extract_tar_lzma ok
11: test_extract_tar_lzo ok
12: test_extract_tar_xz ok
13: test_format_newc ok
14: test_help ok
15: test_option_C_upper ok
16: test_option_H_upper ok
17: test_option_L_upper ok
18: test_option_O_upper ok
19: test_option_T_upper ok
20: test_option_U_upper ok
21: test_option_X_upper ok
22: test_option_a ok
23: test_option_b ok
24: test_option_b64encode ok
25: test_option_exclude ok
26: test_option_gid_gname ok
27: test_option_grzip ok
28: test_option_j ok
29: test_option_k ok
30: test_option_keep_newer_files ok
31: test_option_lrzip ok
32: test_option_lzma ok
33: test_option_lzop ok
34: test_option_n ok
35: test_option_newer_than ok
36: test_option_nodump ok
37: test_option_older_than ok
38: test_option_q ok
39: test_option_r ok
40: test_option_s ok
41: test_option_uid_uname ok
42: test_option_uuencode ok
43: test_option_xz ok
44: test_option_z ok
45: test_patterns ok
46: test_print_longpath ok
47: test_stdio ok
48: test_strip_components ok
49: test_symlink_dir ok
50: test_version ok
51: test_windows ok
Totals:
Tests run: 52
Tests failed: 0
Assertions checked: 7716
Assertions failed: 0
Skips reported: 6
52 tests passed, no failures
PASS: bsdtar_test
If tests fail or crash, details will be in:
/tmp/bsdcpio_test.2013-04-10T14.21.12-000
Reference files will be read from: /home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2/cpio/test
Running tests on: "/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2/bsdcpio"
Exercising: bsdcpio 3.1.2 -- libarchive 3.1.2
0: test_0 ok
1: test_basic ok
2: test_cmdline ok
3: test_extract_cpio_Z ok
4: test_extract_cpio_bz2 ok
5: test_extract_cpio_grz sh: grzip: not found
ok
6: test_extract_cpio_gz ok
7: test_extract_cpio_lrz sh: lrzip: not found
ok
8: test_extract_cpio_lz ok
9: test_extract_cpio_lzma ok
10: test_extract_cpio_lzo sh: lzop: not found
ok
11: test_extract_cpio_xz ok
12: test_format_newc ok
13: test_gcpio_compat ok
14: test_option_0 ok
15: test_option_B_upper ok
16: test_option_C_upper ok
17: test_option_J_upper ok
18: test_option_L_upper ok
19: test_option_Z_upper ok
20: test_option_a ok
21: test_option_b64encode ok
22: test_option_c ok
23: test_option_d ok
24: test_option_f ok
25: test_option_grzip ok
26: test_option_help ok
27: test_option_l ok
28: test_option_lrzip ok
29: test_option_lzma ok
30: test_option_lzop ok
31: test_option_m ok
32: test_option_t ok
33: test_option_u ok
34: test_option_uuencode ok
35: test_option_version ok
36: test_option_xz ok
37: test_option_y ok
38: test_option_z ok
39: test_owner_parse ok
40: test_passthrough_dotdot ok
41: test_passthrough_reverse ok
Totals:
Tests run: 42
Tests failed: 0
Assertions checked: 936
Assertions failed: 0
Skips reported: 5
42 tests passed, no failures
PASS: bsdcpio_test
====================================================
1 of 3 tests failed
Please report to libarchive-discuss@googlegroups.com
====================================================
<b>What version are you using?</b>
3.1.2
<b>On what operating system?</b>
Solaris 10 Sparc
<b>How did you build? (cmake, configure, or pre-packaged binary)</b>
cd work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2 && /usr/bin/env -i HOME="/home/dam" PATH="/home/dam/mgar/pkg/.buildsys/v2/gar/bin/sos12-wrappers:/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/install-isa-sparcv8plus/opt/csw/bin:/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/install-isa-sparcv8plus/opt/csw/bin:/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/install-isa-sparcv8plus/opt/csw/sbin:/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/install-isa-sparcv8plus/opt/csw/sbin:/opt/csw/bin:/opt/csw/bin:/opt/csw/sbin:/opt/csw/sbin:/opt/SUNWspro/bin:/home/dam/mgar/pkg/.buildsys/v2/gar/bin:/usr/bin:/usr/sbin:/usr/java/bin:/usr/ccs/bin:/usr/openwin/bin" LC_ALL="C" prefix="/opt/csw" exec_prefix="/opt/csw" bindir="/opt/csw/bin" sbindir="/opt/csw/sbin" libexecdir="/opt/csw/libexec" datadir="/opt/csw/share" sysconfdir="/etc/opt/csw" sharedstatedir="/opt/csw/share" localstatedir="/var/opt/csw" libdir="/opt/csw/lib" infodir="/opt/csw/share/info" lispdir="/opt/csw/share/emacs/site-lisp" includedir="/opt/csw/include" mandir="/opt/csw/share/man" docdir="/opt/csw/share/doc" sourcedir="/opt/csw/src" CPPFLAGS="-I/opt/csw/include/cryptopp -I/opt/csw/include" CFLAGS="-xO3 -m32 -xarch=sparc" CXXFLAGS="-xO3 -m32 -xarch=sparc" LDFLAGS="-m32 -xarch=sparc -L/opt/csw/lib" FFLAGS="-xO3 -m32 -xarch=sparc" FCFLAGS="-xO3 -m32 -xarch=sparc" F77="/opt/SUNWspro/bin/f77" FC="/opt/SUNWspro/bin/f95" ASFLAGS="" OPTFLAGS="-xO3 -m32 -xarch=sparc" CC="/opt/SUNWspro/bin/cc" CXX="/opt/SUNWspro/bin/CC" CC_HOME="/opt/SUNWspro" CC_VERSION="Sun C 5.9 SunOS_sparc Patch 124867-16 2010/08/11" CXX_VERSION="Sun C++ 5.9 SunOS_sparc Patch 124863-30 2012/07/11" GARCH="sparc" GAROSREL="5.10" GARPACKAGE="trunk" LD_OPTIONS="-R/opt/csw/lib/\$ISALIST -R/opt/csw/lib -M /home/dam/mgar/pkg/.buildsys/v2/gar/lib/map.solaris10 -B direct -z ignore" gmake COMMON_CFLAGS="" -C . check
<b>What compiler or development environment (please include version)?</b>
Sun Studio 12
<b>Please provide any additional information below.</b>
The logs for the failed tests are available at http://buildfarm.opencsw.org/~dam/libarchive_test.2013-04-10T14.16.07-000/
``` | 1.0 | Failed tests on Solaris 10 Sparc - Original [issue 313](https://code.google.com/p/libarchive/issues/detail?id=313) created by Google Code user `honkman42` on 2013-04-10T12:52:19.000Z:
```
<b>What steps will reproduce the problem?</b>
1. Compile and run the testsuite
<b>What is the expected output? What do you see instead?</b>
gmake[2]: Entering directory `/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2'
If tests fail or crash, details will be in:
/tmp/libarchive_test.2013-04-10T14.16.07-000
Reference files will be read from: /home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2/libarchive/test
Exercising: libarchive 3.1.2
0: test_acl_freebsd_nfs4 ok
1: test_acl_freebsd_posix1e ok
2: test_acl_nfs4 ok
3: test_acl_pax ok
4: test_acl_posix1e ok
5: test_archive_api_feature ok
6: test_archive_clear_error ok
7: test_archive_cmdline ok
8: test_archive_md5 ok
9: test_archive_rmd160 ok
10: test_archive_sha1 ok
11: test_archive_sha256 ok
12: test_archive_sha384 ok
13: test_archive_sha512 ok
14: test_archive_getdate ok
15: test_archive_match_owner ok
16: test_archive_match_path ok
17: test_archive_match_time ok
18: test_archive_pathmatch ok
19: test_archive_read_close_twice ok
20: test_archive_read_close_twice_open_fd ok
21: test_archive_read_close_twice_open_filename ok
22: test_archive_read_multiple_data_objects ok
23: test_archive_read_next_header_empty ok
24: test_archive_read_next_header_raw ok
25: test_archive_read_open2 ok
26: test_archive_read_set_filter_option ok
27: test_archive_read_set_format_option ok
28: test_archive_read_set_option ok
29: test_archive_read_set_options ok
30: test_archive_read_support ok
31: test_archive_set_error ok
32: test_archive_string ok
33: test_archive_string_conversion ok
34: test_archive_write_add_filter_by_name_b64encode ok
35: test_archive_write_add_filter_by_name_bzip2 ok
36: test_archive_write_add_filter_by_name_compress ok
37: test_archive_write_add_filter_by_name_grzip sh: grzip: not found
ok
38: test_archive_write_add_filter_by_name_gzip ok
39: test_archive_write_add_filter_by_name_lrzip sh: lrzip: not found
ok
40: test_archive_write_add_filter_by_name_lzip ok
41: test_archive_write_add_filter_by_name_lzma ok
42: test_archive_write_add_filter_by_name_lzop ok
43: test_archive_write_add_filter_by_name_uuencode ok
44: test_archive_write_add_filter_by_name_xz ok
45: test_archive_write_set_filter_option ok
46: test_archive_write_set_format_by_name_7zip ok
47: test_archive_write_set_format_by_name_ar ok
48: test_archive_write_set_format_by_name_arbsd ok
49: test_archive_write_set_format_by_name_argnu ok
50: test_archive_write_set_format_by_name_arsvr4 ok
51: test_archive_write_set_format_by_name_bsdtar ok
52: test_archive_write_set_format_by_name_cd9660 ok
53: test_archive_write_set_format_by_name_cpio ok
54: test_archive_write_set_format_by_name_gnutar ok
55: test_archive_write_set_format_by_name_iso ok
56: test_archive_write_set_format_by_name_iso9660 ok
57: test_archive_write_set_format_by_name_mtree ok
58: test_archive_write_set_format_by_name_mtree_classicok
59: test_archive_write_set_format_by_name_newc ok
60: test_archive_write_set_format_by_name_odc ok
61: test_archive_write_set_format_by_name_oldtar ok
62: test_archive_write_set_format_by_name_pax ok
63: test_archive_write_set_format_by_name_paxr ok
64: test_archive_write_set_format_by_name_posix ok
65: test_archive_write_set_format_by_name_rpax ok
66: test_archive_write_set_format_by_name_shar ok
67: test_archive_write_set_format_by_name_shardump ok
68: test_archive_write_set_format_by_name_ustar ok
69: test_archive_write_set_format_by_name_v7tar ok
70: test_archive_write_set_format_by_name_v7 ok
71: test_archive_write_set_format_by_name_xar ok
72: test_archive_write_set_format_by_name_zip ok
73: test_archive_write_set_format_option ok
74: test_archive_write_set_option ok
75: test_archive_write_set_options ok
76: test_bad_fd ok
77: test_compat_bzip2 ok
78: test_compat_cpio ok
79: test_compat_gtar ok
80: test_compat_gzip ok
81: test_compat_lzip ok
82: test_compat_lzma ok
83: test_compat_lzop ok
84: test_compat_mac ok
85: test_compat_pax_libarchive_2x ok
86: test_compat_solaris_pax_sparse ok
87: test_compat_solaris_tar_acl ok
88: test_compat_tar_hardlink ok
89: test_compat_uudecode ok
90: test_compat_xz ok
91: test_compat_zip FAIL
92: test_empty_write ok
93: test_entry ok
94: test_entry_strmode ok
95: test_extattr_freebsd ok
96: test_filter_count ok
97: test_fuzz_ar ok
98: test_fuzz_cab ok
99: test_fuzz_cpio
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--format violated
gzip: stdin: invalid compressed data--crc error
gzip: stdin: unexpected end of file
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: unexpected end of file
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
ok
100: test_fuzz_iso9660 ok
101: test_fuzz_lzh ok
102: test_fuzz_mtree ok
103: test_fuzz_rar ok
104: test_fuzz_tar
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--format violated
gzip: stdin: invalid compressed data--format violated
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: decompression OK, trailing garbage ignored
gzip: stdin: invalid compressed data--crc error
gzip: stdin: decompression OK, trailing garbage ignored
gzip: stdin: decompression OK, trailing garbage ignored
gzip: stdin: unexpected end of file
gzip: stdin: invalid compressed data--format violated
gzip: stdin: decompression OK, trailing garbage ignored
gzip: stdin: invalid compressed data--format violated
gzip: stdin: invalid compressed data--format violated
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: decompression OK, trailing garbage ignored
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: invalid compressed data--crc error
gzip: stdin: invalid compressed data--length error
gzip: stdin: decompression OK, trailing garbage ignored
ok
105: test_fuzz_zip ok
106: test_gnutar_filename_encoding ok
107: test_link_resolver ok
108: test_open_failure ok
109: test_open_fd ok
110: test_open_file ok
111: test_open_filename ok
112: test_pax_filename_encoding ok
113: test_read_data_large ok
114: test_read_disk ok
115: test_read_disk_directory_traversals ok
116: test_read_disk_entry_from_file ok
117: test_read_extract ok
118: test_read_file_nonexistent ok
119: test_read_filter_grzip ok
120: test_read_filter_lrzip ok
121: test_read_filter_lzop ok
122: test_read_filter_lzop_multiple_parts ok
123: test_read_filter_program ok
124: test_read_filter_program_signature ok
125: test_read_filter_uudecode ok
126: test_read_filter_uudecode_base64 ok
127: test_read_format_7zip ok
128: test_read_format_7zip_bzip2 ok
129: test_read_format_7zip_copy ok
130: test_read_format_7zip_deflate ok
131: test_read_format_7zip_empty ok
132: test_read_format_7zip_lzma1 ok
133: test_read_format_7zip_lzma2 ok
134: test_read_format_7zip_ppmd ok
135: test_read_format_ar ok
136: test_read_format_cab FAIL
137: test_read_format_cab_filename ok
...
209: test_read_format_xar ok
210: test_read_format_zip FAIL
211: test_read_format_zip_comment_stored ok
212: test_read_format_zip_filename ok
213: test_read_format_zip_mac_metadata FAIL
214: test_read_format_zip_sfx ok
...
227: test_read_truncated_filter_gzip
gzip: stdin: unexpected end of file
ok
228: test_read_truncated_filter_lzip ok
229: test_read_truncated_filter_lzma ok
230: test_read_truncated_filter_lzop sh: lzop: not found
ok
231: test_read_truncated_filter_xz ok
...
257: test_write_filter_lzma ok
258: test_write_filter_lzop FAIL
259: test_write_filter_program ok
...
308: test_write_zip_set_compression_store ok
309: test_zip_filename_encoding ok
Totals:
Tests run: 310
Tests failed: 5
Assertions checked:19467410
Assertions failed: 549
Skips reported: 148
Failing tests:
91: test_compat_zip (2 failures)
136: test_read_format_cab (530 failures)
210: test_read_format_zip (14 failures)
213: test_read_format_zip_mac_metadata (2 failures)
258: test_write_filter_lzop (1 failures)
Details for failing tests: /tmp/libarchive_test.2013-04-10T14.16.07-000
FAIL: libarchive_test
If tests fail or crash, details will be in:
/tmp/bsdtar_test.2013-04-10T14.21.00-000
Reference files will be read from: /home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2/tar/test
Running tests on: "/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2/bsdtar"
Exercising: bsdtar 3.1.2 - libarchive 3.1.2
0: test_0 ok
1: test_basic ok
2: test_copy ok
3: test_empty_mtree ok
4: test_extract_tar_Z ok
5: test_extract_tar_bz2 ok
6: test_extract_tar_grz sh: grzip: not found
ok
7: test_extract_tar_gz ok
8: test_extract_tar_lrz sh: lrzip: not found
ok
9: test_extract_tar_lz ok
10: test_extract_tar_lzma ok
11: test_extract_tar_lzo ok
12: test_extract_tar_xz ok
13: test_format_newc ok
14: test_help ok
15: test_option_C_upper ok
16: test_option_H_upper ok
17: test_option_L_upper ok
18: test_option_O_upper ok
19: test_option_T_upper ok
20: test_option_U_upper ok
21: test_option_X_upper ok
22: test_option_a ok
23: test_option_b ok
24: test_option_b64encode ok
25: test_option_exclude ok
26: test_option_gid_gname ok
27: test_option_grzip ok
28: test_option_j ok
29: test_option_k ok
30: test_option_keep_newer_files ok
31: test_option_lrzip ok
32: test_option_lzma ok
33: test_option_lzop ok
34: test_option_n ok
35: test_option_newer_than ok
36: test_option_nodump ok
37: test_option_older_than ok
38: test_option_q ok
39: test_option_r ok
40: test_option_s ok
41: test_option_uid_uname ok
42: test_option_uuencode ok
43: test_option_xz ok
44: test_option_z ok
45: test_patterns ok
46: test_print_longpath ok
47: test_stdio ok
48: test_strip_components ok
49: test_symlink_dir ok
50: test_version ok
51: test_windows ok
Totals:
Tests run: 52
Tests failed: 0
Assertions checked: 7716
Assertions failed: 0
Skips reported: 6
52 tests passed, no failures
PASS: bsdtar_test
If tests fail or crash, details will be in:
/tmp/bsdcpio_test.2013-04-10T14.21.12-000
Reference files will be read from: /home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2/cpio/test
Running tests on: "/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2/bsdcpio"
Exercising: bsdcpio 3.1.2 -- libarchive 3.1.2
0: test_0 ok
1: test_basic ok
2: test_cmdline ok
3: test_extract_cpio_Z ok
4: test_extract_cpio_bz2 ok
5: test_extract_cpio_grz sh: grzip: not found
ok
6: test_extract_cpio_gz ok
7: test_extract_cpio_lrz sh: lrzip: not found
ok
8: test_extract_cpio_lz ok
9: test_extract_cpio_lzma ok
10: test_extract_cpio_lzo sh: lzop: not found
ok
11: test_extract_cpio_xz ok
12: test_format_newc ok
13: test_gcpio_compat ok
14: test_option_0 ok
15: test_option_B_upper ok
16: test_option_C_upper ok
17: test_option_J_upper ok
18: test_option_L_upper ok
19: test_option_Z_upper ok
20: test_option_a ok
21: test_option_b64encode ok
22: test_option_c ok
23: test_option_d ok
24: test_option_f ok
25: test_option_grzip ok
26: test_option_help ok
27: test_option_l ok
28: test_option_lrzip ok
29: test_option_lzma ok
30: test_option_lzop ok
31: test_option_m ok
32: test_option_t ok
33: test_option_u ok
34: test_option_uuencode ok
35: test_option_version ok
36: test_option_xz ok
37: test_option_y ok
38: test_option_z ok
39: test_owner_parse ok
40: test_passthrough_dotdot ok
41: test_passthrough_reverse ok
Totals:
Tests run: 42
Tests failed: 0
Assertions checked: 936
Assertions failed: 0
Skips reported: 5
42 tests passed, no failures
PASS: bsdcpio_test
====================================================
1 of 3 tests failed
Please report to libarchive-discuss@googlegroups.com
====================================================
<b>What version are you using?</b>
3.1.2
<b>On what operating system?</b>
Solaris 10 Sparc
<b>How did you build? (cmake, configure, or pre-packaged binary)</b>
cd work/solaris10-sparc/build-isa-sparcv8plus/libarchive-3.1.2 && /usr/bin/env -i HOME="/home/dam" PATH="/home/dam/mgar/pkg/.buildsys/v2/gar/bin/sos12-wrappers:/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/install-isa-sparcv8plus/opt/csw/bin:/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/install-isa-sparcv8plus/opt/csw/bin:/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/install-isa-sparcv8plus/opt/csw/sbin:/home/dam/mgar/pkg/libarchive/trunk/work/solaris10-sparc/install-isa-sparcv8plus/opt/csw/sbin:/opt/csw/bin:/opt/csw/bin:/opt/csw/sbin:/opt/csw/sbin:/opt/SUNWspro/bin:/home/dam/mgar/pkg/.buildsys/v2/gar/bin:/usr/bin:/usr/sbin:/usr/java/bin:/usr/ccs/bin:/usr/openwin/bin" LC_ALL="C" prefix="/opt/csw" exec_prefix="/opt/csw" bindir="/opt/csw/bin" sbindir="/opt/csw/sbin" libexecdir="/opt/csw/libexec" datadir="/opt/csw/share" sysconfdir="/etc/opt/csw" sharedstatedir="/opt/csw/share" localstatedir="/var/opt/csw" libdir="/opt/csw/lib" infodir="/opt/csw/share/info" lispdir="/opt/csw/share/emacs/site-lisp" includedir="/opt/csw/include" mandir="/opt/csw/share/man" docdir="/opt/csw/share/doc" sourcedir="/opt/csw/src" CPPFLAGS="-I/opt/csw/include/cryptopp -I/opt/csw/include" CFLAGS="-xO3 -m32 -xarch=sparc" CXXFLAGS="-xO3 -m32 -xarch=sparc" LDFLAGS="-m32 -xarch=sparc -L/opt/csw/lib" FFLAGS="-xO3 -m32 -xarch=sparc" FCFLAGS="-xO3 -m32 -xarch=sparc" F77="/opt/SUNWspro/bin/f77" FC="/opt/SUNWspro/bin/f95" ASFLAGS="" OPTFLAGS="-xO3 -m32 -xarch=sparc" CC="/opt/SUNWspro/bin/cc" CXX="/opt/SUNWspro/bin/CC" CC_HOME="/opt/SUNWspro" CC_VERSION="Sun C 5.9 SunOS_sparc Patch 124867-16 2010/08/11" CXX_VERSION="Sun C++ 5.9 SunOS_sparc Patch 124863-30 2012/07/11" GARCH="sparc" GAROSREL="5.10" GARPACKAGE="trunk" LD_OPTIONS="-R/opt/csw/lib/\$ISALIST -R/opt/csw/lib -M /home/dam/mgar/pkg/.buildsys/v2/gar/lib/map.solaris10 -B direct -z ignore" gmake COMMON_CFLAGS="" -C . check
<b>What compiler or development environment (please include version)?</b>
Sun Studio 12
<b>Please provide any additional information below.</b>
The logs for the failed tests are available at http://buildfarm.opencsw.org/~dam/libarchive_test.2013-04-10T14.16.07-000/
``` | non_process | failed tests on solaris sparc original created by google code user on what steps will reproduce the problem compile and run the testsuite what is the expected output what do you see instead gmake entering directory home dam mgar pkg libarchive trunk work sparc build isa libarchive if tests fail or crash details will be in tmp libarchive test reference files will be read from home dam mgar pkg libarchive trunk work sparc build isa libarchive libarchive test exercising libarchive test acl freebsd ok test acl freebsd ok test acl ok test acl pax ok test acl ok test archive api feature ok test archive clear error ok test archive cmdline ok test archive ok test archive ok test archive ok test archive ok test archive ok test archive ok test archive getdate ok test archive match owner ok test archive match path ok test archive match time ok test archive pathmatch ok test archive read close twice ok test archive read close twice open fd ok test archive read close twice open filename ok test archive read multiple data objects ok test archive read next header empty ok test archive read next header raw ok test archive read ok test archive read set filter option ok test archive read set format option ok test archive read set option ok test archive read set options ok test archive read support ok test archive set error ok test archive string ok test archive string conversion ok test archive write add filter by name ok test archive write add filter by name ok test archive write add filter by name compress ok test archive write add filter by name grzip sh grzip not found ok test archive write add filter by name gzip ok test archive write add filter by name lrzip sh lrzip not found ok test archive write add filter by name lzip ok test archive write add filter by name lzma ok test archive write add filter by name lzop ok test archive write add filter by name uuencode ok test archive write add filter by name xz ok test archive write set filter option ok test archive write set format by name ok test archive write set format by name ar ok test archive write set format by name arbsd ok test archive write set format by name argnu ok test archive write set format by name ok test archive write set format by name bsdtar ok test archive write set format by name ok test archive write set format by name cpio ok test archive write set format by name gnutar ok test archive write set format by name iso ok test archive write set format by name ok test archive write set format by name mtree ok test archive write set format by name mtree classicok test archive write set format by name newc ok test archive write set format by name odc ok test archive write set format by name oldtar ok test archive write set format by name pax ok test archive write set format by name paxr ok test archive write set format by name posix ok test archive write set format by name rpax ok test archive write set format by name shar ok test archive write set format by name shardump ok test archive write set format by name ustar ok test archive write set format by name ok test archive write set format by name ok test archive write set format by name xar ok test archive write set format by name zip ok test archive write set format option ok test archive write set option ok test archive write set options ok test bad fd ok test compat ok test compat cpio ok test compat gtar ok test compat gzip ok test compat lzip ok test compat lzma ok test compat lzop ok test compat mac ok test compat pax libarchive ok test compat solaris pax sparse ok test compat solaris tar acl ok test compat tar hardlink ok test compat uudecode ok test compat xz ok test compat zip fail test empty write ok test entry ok test entry strmode ok test extattr freebsd ok test filter count ok test fuzz ar ok test fuzz cab ok test fuzz cpio gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin invalid compressed data crc error gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin invalid compressed data crc error gzip stdin invalid compressed data format violated gzip stdin invalid compressed data crc error gzip stdin unexpected end of file gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin invalid compressed data crc error gzip stdin unexpected end of file gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error ok test fuzz ok test fuzz lzh ok test fuzz mtree ok test fuzz rar ok test fuzz tar gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin invalid compressed data crc error gzip stdin invalid compressed data format violated gzip stdin invalid compressed data format violated gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin invalid compressed data length error gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin decompression ok trailing garbage ignored gzip stdin invalid compressed data crc error gzip stdin decompression ok trailing garbage ignored gzip stdin decompression ok trailing garbage ignored gzip stdin unexpected end of file gzip stdin invalid compressed data format violated gzip stdin decompression ok trailing garbage ignored gzip stdin invalid compressed data format violated gzip stdin invalid compressed data format violated gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin decompression ok trailing garbage ignored gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin invalid compressed data crc error gzip stdin invalid compressed data length error gzip stdin decompression ok trailing garbage ignored ok test fuzz zip ok test gnutar filename encoding ok test link resolver ok test open failure ok test open fd ok test open file ok test open filename ok test pax filename encoding ok test read data large ok test read disk ok test read disk directory traversals ok test read disk entry from file ok test read extract ok test read file nonexistent ok test read filter grzip ok test read filter lrzip ok test read filter lzop ok test read filter lzop multiple parts ok test read filter program ok test read filter program signature ok test read filter uudecode ok test read filter uudecode ok test read format ok test read format ok test read format copy ok test read format deflate ok test read format empty ok test read format ok test read format ok test read format ppmd ok test read format ar ok test read format cab fail test read format cab filename ok test read format xar ok test read format zip fail test read format zip comment stored ok test read format zip filename ok test read format zip mac metadata fail test read format zip sfx ok test read truncated filter gzip gzip stdin unexpected end of file ok test read truncated filter lzip ok test read truncated filter lzma ok test read truncated filter lzop sh lzop not found ok test read truncated filter xz ok test write filter lzma ok test write filter lzop fail test write filter program ok test write zip set compression store ok test zip filename encoding ok totals tests run tests failed assertions checked assertions failed skips reported failing tests test compat zip failures test read format cab failures test read format zip failures test read format zip mac metadata failures test write filter lzop failures details for failing tests tmp libarchive test fail libarchive test if tests fail or crash details will be in tmp bsdtar test reference files will be read from home dam mgar pkg libarchive trunk work sparc build isa libarchive tar test running tests on quot home dam mgar pkg libarchive trunk work sparc build isa libarchive bsdtar quot exercising bsdtar libarchive test ok test basic ok test copy ok test empty mtree ok test extract tar z ok test extract tar ok test extract tar grz sh grzip not found ok test extract tar gz ok test extract tar lrz sh lrzip not found ok test extract tar lz ok test extract tar lzma ok test extract tar lzo ok test extract tar xz ok test format newc ok test help ok test option c upper ok test option h upper ok test option l upper ok test option o upper ok test option t upper ok test option u upper ok test option x upper ok test option a ok test option b ok test option ok test option exclude ok test option gid gname ok test option grzip ok test option j ok test option k ok test option keep newer files ok test option lrzip ok test option lzma ok test option lzop ok test option n ok test option newer than ok test option nodump ok test option older than ok test option q ok test option r ok test option s ok test option uid uname ok test option uuencode ok test option xz ok test option z ok test patterns ok test print longpath ok test stdio ok test strip components ok test symlink dir ok test version ok test windows ok totals tests run tests failed assertions checked assertions failed skips reported tests passed no failures pass bsdtar test if tests fail or crash details will be in tmp bsdcpio test reference files will be read from home dam mgar pkg libarchive trunk work sparc build isa libarchive cpio test running tests on quot home dam mgar pkg libarchive trunk work sparc build isa libarchive bsdcpio quot exercising bsdcpio libarchive test ok test basic ok test cmdline ok test extract cpio z ok test extract cpio ok test extract cpio grz sh grzip not found ok test extract cpio gz ok test extract cpio lrz sh lrzip not found ok test extract cpio lz ok test extract cpio lzma ok test extract cpio lzo sh lzop not found ok test extract cpio xz ok test format newc ok test gcpio compat ok test option ok test option b upper ok test option c upper ok test option j upper ok test option l upper ok test option z upper ok test option a ok test option ok test option c ok test option d ok test option f ok test option grzip ok test option help ok test option l ok test option lrzip ok test option lzma ok test option lzop ok test option m ok test option t ok test option u ok test option uuencode ok test option version ok test option xz ok test option y ok test option z ok test owner parse ok test passthrough dotdot ok test passthrough reverse ok totals tests run tests failed assertions checked assertions failed skips reported tests passed no failures pass bsdcpio test of tests failed please report to libarchive discuss googlegroups com what version are you using on what operating system solaris sparc how did you build cmake configure or pre packaged binary cd work sparc build isa libarchive amp amp usr bin env i home quot home dam quot path quot home dam mgar pkg buildsys gar bin wrappers home dam mgar pkg libarchive trunk work sparc install isa opt csw bin home dam mgar pkg libarchive trunk work sparc install isa opt csw bin home dam mgar pkg libarchive trunk work sparc install isa opt csw sbin home dam mgar pkg libarchive trunk work sparc install isa opt csw sbin opt csw bin opt csw bin opt csw sbin opt csw sbin opt sunwspro bin home dam mgar pkg buildsys gar bin usr bin usr sbin usr java bin usr ccs bin usr openwin bin quot lc all quot c quot prefix quot opt csw quot exec prefix quot opt csw quot bindir quot opt csw bin quot sbindir quot opt csw sbin quot libexecdir quot opt csw libexec quot datadir quot opt csw share quot sysconfdir quot etc opt csw quot sharedstatedir quot opt csw share quot localstatedir quot var opt csw quot libdir quot opt csw lib quot infodir quot opt csw share info quot lispdir quot opt csw share emacs site lisp quot includedir quot opt csw include quot mandir quot opt csw share man quot docdir quot opt csw share doc quot sourcedir quot opt csw src quot cppflags quot i opt csw include cryptopp i opt csw include quot cflags quot xarch sparc quot cxxflags quot xarch sparc quot ldflags quot xarch sparc l opt csw lib quot fflags quot xarch sparc quot fcflags quot xarch sparc quot quot opt sunwspro bin quot fc quot opt sunwspro bin quot asflags quot quot optflags quot xarch sparc quot cc quot opt sunwspro bin cc quot cxx quot opt sunwspro bin cc quot cc home quot opt sunwspro quot cc version quot sun c sunos sparc patch quot cxx version quot sun c sunos sparc patch quot garch quot sparc quot garosrel quot quot garpackage quot trunk quot ld options quot r opt csw lib isalist r opt csw lib m home dam mgar pkg buildsys gar lib map b direct z ignore quot gmake common cflags quot quot c check what compiler or development environment please include version sun studio please provide any additional information below the logs for the failed tests are available at | 0 |
15,706 | 19,848,440,451 | IssuesEvent | 2022-01-21 09:33:48 | prisma/prisma | https://api.github.com/repos/prisma/prisma | opened | Error: [/root/build/libs/mongodb-schema-describer/src/lib.rs:35:41] called `Option::unwrap()` on a `None` value | bug/1-repro-available kind/bug process/candidate topic: error reporting team/migrations topic: mongodb | <!-- If required, please update the title to be clear and descriptive -->
Command: `prisma db pull`
Version: `3.8.1`
Binary Version: `34df67547cf5598f5a6cd3eb45f14ee70c3fb86f`
Report: https://prisma-errors.netlify.app/report/13648
OS: `x64 linux 4.19.128-microsoft-standard`
JS Stacktrace:
```
Error: [/root/build/libs/mongodb-schema-describer/src/lib.rs:35:41] called `Option::unwrap()` on a `None` value
at ChildProcess.<anonymous> (/home/hosein/projects/cheekara/website/api/node_modules/prisma/build/index.js:46398:30)
at ChildProcess.emit (events.js:315:20)
at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12)
```
Rust Stacktrace:
```
0: user_facing_errors::Error::new_in_panic_hook
1: user_facing_errors::panic_hook::set_panic_hook::{{closure}}
2: std::panicking::rust_panic_with_hook
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:628:17
3: std::panicking::begin_panic_handler::{{closure}}
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:519:13
4: std::sys_common::backtrace::__rust_end_short_backtrace
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/sys_common/backtrace.rs:139:18
5: rust_begin_unwind
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:517:5
6: core::panicking::panic_fmt
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/core/src/panicking.rs:100:14
7: core::panicking::panic
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/core/src/panicking.rs:50:5
8: <mongodb_introspection_connector::MongoDbIntrospectionConnector as introspection_connector::IntrospectionConnector>::introspect::{{closure}}
9: <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll
10: <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll
11: <futures_util::future::either::Either<A,B> as core::future::future::Future>::poll
12: <futures_util::future::future::Then<Fut1,Fut2,F> as core::future::future::Future>::poll
13: <futures_util::future::either::Either<A,B> as core::future::future::Future>::poll
14: json_rpc_stdio::handle_stdin_next_line::{{closure}}
15: <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll
16: introspection_engine::main
17: std::sys_common::backtrace::__rust_begin_short_backtrace
18: std::rt::lang_start::{{closure}}
19: core::ops::function::impls::<impl core::ops::function::FnOnce<A> for &F>::call_once
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/core/src/ops/function.rs:259:13
std::panicking::try::do_call
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:403:40
std::panicking::try
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:367:19
std::panic::catch_unwind
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panic.rs:133:14
std::rt::lang_start_internal::{{closure}}
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/rt.rs:128:48
std::panicking::try::do_call
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:403:40
std::panicking::try
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:367:19
std::panic::catch_unwind
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panic.rs:133:14
std::rt::lang_start_internal
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/rt.rs:128:20
20: std::rt::lang_start
21: __libc_start_main
22: <unknown>
```
| 1.0 | Error: [/root/build/libs/mongodb-schema-describer/src/lib.rs:35:41] called `Option::unwrap()` on a `None` value - <!-- If required, please update the title to be clear and descriptive -->
Command: `prisma db pull`
Version: `3.8.1`
Binary Version: `34df67547cf5598f5a6cd3eb45f14ee70c3fb86f`
Report: https://prisma-errors.netlify.app/report/13648
OS: `x64 linux 4.19.128-microsoft-standard`
JS Stacktrace:
```
Error: [/root/build/libs/mongodb-schema-describer/src/lib.rs:35:41] called `Option::unwrap()` on a `None` value
at ChildProcess.<anonymous> (/home/hosein/projects/cheekara/website/api/node_modules/prisma/build/index.js:46398:30)
at ChildProcess.emit (events.js:315:20)
at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12)
```
Rust Stacktrace:
```
0: user_facing_errors::Error::new_in_panic_hook
1: user_facing_errors::panic_hook::set_panic_hook::{{closure}}
2: std::panicking::rust_panic_with_hook
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:628:17
3: std::panicking::begin_panic_handler::{{closure}}
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:519:13
4: std::sys_common::backtrace::__rust_end_short_backtrace
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/sys_common/backtrace.rs:139:18
5: rust_begin_unwind
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:517:5
6: core::panicking::panic_fmt
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/core/src/panicking.rs:100:14
7: core::panicking::panic
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/core/src/panicking.rs:50:5
8: <mongodb_introspection_connector::MongoDbIntrospectionConnector as introspection_connector::IntrospectionConnector>::introspect::{{closure}}
9: <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll
10: <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll
11: <futures_util::future::either::Either<A,B> as core::future::future::Future>::poll
12: <futures_util::future::future::Then<Fut1,Fut2,F> as core::future::future::Future>::poll
13: <futures_util::future::either::Either<A,B> as core::future::future::Future>::poll
14: json_rpc_stdio::handle_stdin_next_line::{{closure}}
15: <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll
16: introspection_engine::main
17: std::sys_common::backtrace::__rust_begin_short_backtrace
18: std::rt::lang_start::{{closure}}
19: core::ops::function::impls::<impl core::ops::function::FnOnce<A> for &F>::call_once
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/core/src/ops/function.rs:259:13
std::panicking::try::do_call
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:403:40
std::panicking::try
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:367:19
std::panic::catch_unwind
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panic.rs:133:14
std::rt::lang_start_internal::{{closure}}
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/rt.rs:128:48
std::panicking::try::do_call
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:403:40
std::panicking::try
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panicking.rs:367:19
std::panic::catch_unwind
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/panic.rs:133:14
std::rt::lang_start_internal
at /rustc/f1edd0429582dd29cccacaf50fd134b05593bd9c/library/std/src/rt.rs:128:20
20: std::rt::lang_start
21: __libc_start_main
22: <unknown>
```
| process | error called option unwrap on a none value command prisma db pull version binary version report os linux microsoft standard js stacktrace error called option unwrap on a none value at childprocess home hosein projects cheekara website api node modules prisma build index js at childprocess emit events js at process childprocess handle onexit internal child process js rust stacktrace user facing errors error new in panic hook user facing errors panic hook set panic hook closure std panicking rust panic with hook at rustc library std src panicking rs std panicking begin panic handler closure at rustc library std src panicking rs std sys common backtrace rust end short backtrace at rustc library std src sys common backtrace rs rust begin unwind at rustc library std src panicking rs core panicking panic fmt at rustc library core src panicking rs core panicking panic at rustc library core src panicking rs introspect closure as core future future future poll as core future future future poll as core future future future poll as core future future future poll as core future future future poll json rpc stdio handle stdin next line closure as core future future future poll introspection engine main std sys common backtrace rust begin short backtrace std rt lang start closure core ops function impls for f call once at rustc library core src ops function rs std panicking try do call at rustc library std src panicking rs std panicking try at rustc library std src panicking rs std panic catch unwind at rustc library std src panic rs std rt lang start internal closure at rustc library std src rt rs std panicking try do call at rustc library std src panicking rs std panicking try at rustc library std src panicking rs std panic catch unwind at rustc library std src panic rs std rt lang start internal at rustc library std src rt rs std rt lang start libc start main | 1 |
19,966 | 26,444,300,188 | IssuesEvent | 2023-01-16 05:13:33 | bazelbuild/bazel | https://api.github.com/repos/bazelbuild/bazel | closed | [Mirror] Gazelle 0.29.0 | P2 type: process team-OSS mirror request | ### Please list the URLs of the archives you'd like to mirror:
Please mirror https://github.com/bazelbuild/bazel-gazelle/releases/download/v0.29.0/bazel-gazelle-v0.29.0.tar.gz | 1.0 | [Mirror] Gazelle 0.29.0 - ### Please list the URLs of the archives you'd like to mirror:
Please mirror https://github.com/bazelbuild/bazel-gazelle/releases/download/v0.29.0/bazel-gazelle-v0.29.0.tar.gz | process | gazelle please list the urls of the archives you d like to mirror please mirror | 1 |
62,399 | 17,023,914,952 | IssuesEvent | 2021-07-03 04:32:05 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | irregular behaviour of Mapnik | Component: mapnik Priority: minor Resolution: invalid Type: defect | **[Submitted to the original trac issue database at 5.06pm, Wednesday, 10th December 2014]**
Problem:
[https://www.openstreetmap.org/relation/1564971 One multipolygon]
```
type=multipolygon
landuse=farmmland
```
with two inner ways
```
landuse=orchard
```
One orchard gets rendered, the other not.
Why? | 1.0 | irregular behaviour of Mapnik - **[Submitted to the original trac issue database at 5.06pm, Wednesday, 10th December 2014]**
Problem:
[https://www.openstreetmap.org/relation/1564971 One multipolygon]
```
type=multipolygon
landuse=farmmland
```
with two inner ways
```
landuse=orchard
```
One orchard gets rendered, the other not.
Why? | non_process | irregular behaviour of mapnik problem type multipolygon landuse farmmland with two inner ways landuse orchard one orchard gets rendered the other not why | 0 |
11,582 | 2,659,137,098 | IssuesEvent | 2015-03-18 19:14:24 | sbsrouteur/sbsrouteur | https://api.github.com/repos/sbsrouteur/sbsrouteur | closed | Fix incorrect speed estimate | auto-migrated Milestone-V0.34 Priority-Medium Type-Defect | ```
Some angles yield an incorrect speed estimate
```
Original issue reported on code.google.com by `sbsrout...@free.fr` on 16 Feb 2013 at 1:22 | 1.0 | Fix incorrect speed estimate - ```
Some angles yield an incorrect speed estimate
```
Original issue reported on code.google.com by `sbsrout...@free.fr` on 16 Feb 2013 at 1:22 | non_process | fix incorrect speed estimate some angles yield an incorrect speed estimate original issue reported on code google com by sbsrout free fr on feb at | 0 |
31,566 | 13,559,937,978 | IssuesEvent | 2020-09-18 00:20:26 | terraform-providers/terraform-provider-aws | https://api.github.com/repos/terraform-providers/terraform-provider-aws | closed | dynamodb data source global_secondary_index list missing entries | needs-triage service/dynamodb | <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
Terraform v0.12.8
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* aws_dynamodb_table
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
data "aws_dynamodb_table" "orders" {
name = var.orders_dynamodb_table
}
```
### Expected Behavior
Terraform state should include both of the GSIs that are on the referenced table.
```
terraform state show module.cloud_portal.data.aws_dynamodb_table.orders
data "aws_dynamodb_table" "orders" {
...
global_secondary_index = [
{
hash_key = "OrganizationId"
name = "OrganizationIndex"
non_key_attributes = []
projection_type = "ALL"
range_key = ""
read_capacity = 20
write_capacity = 10
},
{
hash_key = "ProductId"
name = "ProductIndex"
non_key_attributes = []
projection_type = "ALL"
range_key = ""
read_capacity = 20
write_capacity = 10
},
]
...
}
```
### Actual Behavior
Terraform state only includes one of the GSIs that are on the referenced table.
```
terraform state show module.cloud_portal.data.aws_dynamodb_table.orders
data "aws_dynamodb_table" "orders" {
...
global_secondary_index = [
{
hash_key = "OrganizationId"
name = "OrganizationIndex"
non_key_attributes = []
projection_type = "ALL"
range_key = ""
read_capacity = 20
write_capacity = 10
},
]
...
}
```
### Steps to Reproduce
1. Create a dynamodb table with two GSIs
2. Reference that table in a `data "aws_dynamodb_table"` block
3. `terraform state show data.aws_dynamodb_table.<my_test_table>`
4. Find the GSI list and see that it has only one index listed.
| 1.0 | dynamodb data source global_secondary_index list missing entries - <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
Terraform v0.12.8
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* aws_dynamodb_table
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
data "aws_dynamodb_table" "orders" {
name = var.orders_dynamodb_table
}
```
### Expected Behavior
Terraform state should include both of the GSIs that are on the referenced table.
```
terraform state show module.cloud_portal.data.aws_dynamodb_table.orders
data "aws_dynamodb_table" "orders" {
...
global_secondary_index = [
{
hash_key = "OrganizationId"
name = "OrganizationIndex"
non_key_attributes = []
projection_type = "ALL"
range_key = ""
read_capacity = 20
write_capacity = 10
},
{
hash_key = "ProductId"
name = "ProductIndex"
non_key_attributes = []
projection_type = "ALL"
range_key = ""
read_capacity = 20
write_capacity = 10
},
]
...
}
```
### Actual Behavior
Terraform state only includes one of the GSIs that are on the referenced table.
```
terraform state show module.cloud_portal.data.aws_dynamodb_table.orders
data "aws_dynamodb_table" "orders" {
...
global_secondary_index = [
{
hash_key = "OrganizationId"
name = "OrganizationIndex"
non_key_attributes = []
projection_type = "ALL"
range_key = ""
read_capacity = 20
write_capacity = 10
},
]
...
}
```
### Steps to Reproduce
1. Create a dynamodb table with two GSIs
2. Reference that table in a `data "aws_dynamodb_table"` block
3. `terraform state show data.aws_dynamodb_table.<my_test_table>`
4. Find the GSI list and see that it has only one index listed.
| non_process | dynamodb data source global secondary index list missing entries please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version terraform affected resource s aws dynamodb table terraform configuration files hcl data aws dynamodb table orders name var orders dynamodb table expected behavior terraform state should include both of the gsis that are on the referenced table terraform state show module cloud portal data aws dynamodb table orders data aws dynamodb table orders global secondary index hash key organizationid name organizationindex non key attributes projection type all range key read capacity write capacity hash key productid name productindex non key attributes projection type all range key read capacity write capacity actual behavior terraform state only includes one of the gsis that are on the referenced table terraform state show module cloud portal data aws dynamodb table orders data aws dynamodb table orders global secondary index hash key organizationid name organizationindex non key attributes projection type all range key read capacity write capacity steps to reproduce create a dynamodb table with two gsis reference that table in a data aws dynamodb table block terraform state show data aws dynamodb table find the gsi list and see that it has only one index listed | 0 |
19,844 | 26,245,464,610 | IssuesEvent | 2023-01-05 14:56:08 | celo-org/celo-monorepo | https://api.github.com/repos/celo-org/celo-monorepo | closed | Shepherd Core Contracts 8 Release | epic release-process Component: Identity ASv2 | Github issue to keep track checklist in [Core Contracts release process](https://docs.celo.org/community/release-process/smart-contracts#promotion-process).
T:
- [x] @0xarthurxyz: Create a Github issue tracking all these checklist items as an audit log
- This is it: #9711
- [x] @0xarthurxyz: Implement the [git management steps](https://docs.celo.org/community/release-process/smart-contracts#When-a-new-release-branch-is-cut) for when a new release branch is cut.
- **Release branch**:
- > A new release branch is created release/core-contracts/${N} with the contracts to be audited.
- > there already exists a release/core-contracts/8 branch, but it's deprecated - can be deleted and a new one of the same name be created at your release branch. [~Source: @m-chrzan]
- Commands used (for review):
```bash
git push origin --delete release/core-contracts/8 # delete existing release/core-contracts/8 branch (advised by Martin)
git checkout origin/ASv2 # moves to development branch
git branch release/core-contracts/8 # creates release branch
git checkout release/core-contracts/8 # moves to release branch
git log # checks last commit on release branch is as expected
git push origin release/core-contracts/8 # publishes release branch
```
- [x] this is the release branch: [`release/core-contracts/8`](https://github.com/celo-org/celo-monorepo/commits/release/core-contracts/8) (cut from our development branch [`ASv2`](https://github.com/celo-org/celo-monorepo/tree/ASv2/packages/protocol/contracts/identity), last commit `c1bd0574`)
- **Tag commit**:
- > The latest commit on the release branch is tagged with `core-contracts.v${N}.pre-audit`.
- Commands used (for review):
```bash
git log # check commits on release branch
git show core-contracts.v8.pre-audit # checks if tag already exists (it does, Eela already tagged it)
```
- [x] this is the tag: [`core-contracts.v8.pre-audit`](https://github.com/celo-org/celo-monorepo/releases/tag/core-contracts.v8.pre-audit) for commit `c1bd0574`
- **Update config.yml**:
- > On master branch, [`.circleci/config.yml`](https://github.com/celo-org/celo-monorepo/blob/c1bd057484c4601436eb35c563d431f80d5f94b6/.circleci/config.yml#L37) should be edited so that the variable `RELEASE_TAG` points to the tag `celo-core-contracts-v${N}.pre-audit` so that all future changes to master are versioned against the new release. ~ Source: [Git release management](https://docs.celo.org/community/release-process/smart-contracts#when-a-new-release-branch-is-cut)
- > On the release branch, it should remain `core-contracts.v7`. [~Source: @m-chrzan]
- Commands used (for review):
```bash
# currently on master branch in monorepo
git branch 0xarthurxyz/modify-circleci-configyml-for-CoreContracts8Release # makes branch for my 1 line change
git checkout 0xarthurxyz/modify-circleci-configyml-for-CoreContracts8Release # moves to my branch
# made 1 line change
# committed to my branch (in VS Code UI)
# pushed my branch to remote (in VS Code UI)
# created PR in Github UI
```
- [x] this is the PR: https://github.com/celo-org/celo-monorepo/pull/9734
- **Pre-release notes**:
- > On Github, a pre-release Github release should be created pointing at the latest tag on the release branch. ~ Source: [Git release management](https://docs.celo.org/community/release-process/smart-contracts#when-a-new-release-branch-is-cut)
- Followed official Github instructions here: [Managing releases in a repository](https://docs.github.com/en/repositories/releasing-projects-on-github/managing-releases-in-a-repository
)
- [x] this is the pre-release: [Core Contracts Release 8](https://github.com/celo-org/celo-monorepo/releases/tag/core-contracts.v8.pre-audit)
- [x] @eelanagaraj: Submit release branch to a reputable third party **auditor for review**.
- Completed with Hacken.io
- **Begin drafting release notes.**
- [x] #9724
T+1w:
- [x] Receive report from auditors.
- [x] @0xarthurxyz: Add audit summary to final draft of the release notes.
- If all issues in the audit report have straightforward fixes:
- **CGP**:
- [ ] @0xarthurxyz: Submit a governance proposal draft
- Started the CGP draft on this branch: [0xarthurxyz/cgp-61-core-contracts-8](https://github.com/celo-org/governance/tree/0xarthurxyz/cgp-61-core-contracts-8)
- Added: [cgp-0061.md](https://github.com/celo-org/governance/blob/0xarthurxyz/cgp-61-core-contracts-8/CGPs/cgp-0061.md) (following example of Core Contract 7 release: [cgp-0055.md](https://github.com/celo-org/governance/blob/main/CGPs/cgp-0055.md))
- Added: [cgp-0061.json](https://github.com/celo-org/governance/blob/0xarthurxyz/cgp-61-core-contracts-8/CGPs/cgp-0061/cgp-0061.json) (following example of Core Contract 7 release: [cgp-0055.json](https://github.com/celo-org/governance/blob/main/CGPs/cgp-0055/cgp-0055.json))
- **`.json` file for CGP**:
- [x] @alecps: Add any initialization data to the CGP that should be included as part of the proposal
- **Forum post**:
- [x] @0xarthurxyz: Announce forthcoming smart contract release on: https://forum.celo.org/c/governance
- Created draft forum post: [Core Contracts Release 8 [DRAFT]](https://forum.celo.org/t/core-contracts-release-8-draft/4050)
- [x] Present "_Core Contract Release 8_" at Celo Governance Call
- https://github.com/celo-org/governance/issues/164
- [x] @eelanagaraj: Commit audit fixes to the release branch
- This is the release branch: https://github.com/celo-org/celo-monorepo/tree/ASv2/packages/protocol/contracts/identity
- [x] Submit audit fixes to auditors for review.
- Completed with: #9678
- [x] @alecps: Tag the first release candidate commit according to the [git release management instructions](https://docs.celo.org/community/release-process/smart-contracts#During-the-release-proposal-stage).
- [x] @0xarthurxyz: Let the community know about the upcoming release proposal by posting details to the Governance category on https://forum.celo.org and cross post in the [Discord #governance channel](https://discord.com/channels/600834479145353243/704805825373274134).See the 'Communication guidelines' section below for information on what your post should contain.
T+2w
- [ ] @alecps: On Tuesday: Run the smart contract release script in order to to deploy the contracts to Baklava as well as submit a governance proposal.
- [ ] @0xarthurxyz: Transition proposal through Baklava governance process.
- [ ] @0xarthurxyz: Update your forum post with the Baklava `PROPOSAL_ID`, updated timings (if any changes), and notify the community in the [Discord #governance channel](https://discord.com/channels/600834479145353243/704805825373274134).
T+3w
- [ ] @alecps: Confirm all contracts working as intended on Baklava.
- [ ] @alecps: Run the [smart contract release script](https://docs.celo.org/community/release-process/smart-contracts#release-process) in order to to deploy the contracts to Alfajores as well as submit a governance proposal.
- [ ] @0xarthurxyz: Update your forum post with the Alfajores `PROPOSAL_ID`, updated timings (if any changes), and notify the community in the [Discord #governance channel](https://discord.com/channels/600834479145353243/704805825373274134).
T+4w
- [ ] @alecps: Confirm all contracts working as intended on Alfajores.
- [ ] @0xarthurxyz: Confirm audit is complete and make the release notes and forum post contain a link to it.
- [ ] @alecps: On Tuesday: Run the [smart contract release script](https://docs.celo.org/community/release-process/smart-contracts#build-process) in order to to deploy the contracts to Mainnet as well as submit a governance proposal.
- [ ] Update the corresponding governance proposal with the updated on-chain `PROPOSAL_ID` and mark CGP status as `"PROPOSED"`.
- [ ] @0xarthurxyz: Update your forum post with the Mainnet `PROPOSAL_ID`, updated timings (if any changes), and notify the community in the [Discord #governance channel](https://discord.com/channels/600834479145353243/704805825373274134).
- At this point all stakeholders are encouraged to [verify](https://docs.celo.org/community/release-process/smart-contracts#verify-release-process) the proposed contracts deployed match the contracts from the release branch.
- [ ] @0xarthurxyz: Monitor the progress of the proposal through the [governance process](https://docs.celo.org/celo-codebase/protocol/governance).
- Currently the governance process should take approximately 1 week: 24 hours for the dequeue process, 24 hours for the approval process, and 5 days for the referendum process. After which, the proposal is either declined or is ready to be executed within 3 days.
- For updated timeframes, use the celocli: `celocli network:parameters`
T+5w
- If the proposal passed:
- [ ] @alecps: Confirm all contracts working as intended on Mainnet.
- [ ] @0xarthurxyz: Update your forum post with the Mainnet governance outcome (`Passed` or `Rejected`) and notify the community in the [Discord #governance channel](https://discord.com/channels/600834479145353243/704805825373274134).
- [ ] @0xarthurxyz: Change corresponding CGP status to `EXCECUTED`.
- [ ] @alecps: Merge the release branch into `master` with a merge commit
- If the proposal failed:
- Change corresponding CGP status to `EXPIRED`. | 1.0 | Shepherd Core Contracts 8 Release - Github issue to keep track checklist in [Core Contracts release process](https://docs.celo.org/community/release-process/smart-contracts#promotion-process).
T:
- [x] @0xarthurxyz: Create a Github issue tracking all these checklist items as an audit log
- This is it: #9711
- [x] @0xarthurxyz: Implement the [git management steps](https://docs.celo.org/community/release-process/smart-contracts#When-a-new-release-branch-is-cut) for when a new release branch is cut.
- **Release branch**:
- > A new release branch is created release/core-contracts/${N} with the contracts to be audited.
- > there already exists a release/core-contracts/8 branch, but it's deprecated - can be deleted and a new one of the same name be created at your release branch. [~Source: @m-chrzan]
- Commands used (for review):
```bash
git push origin --delete release/core-contracts/8 # delete existing release/core-contracts/8 branch (advised by Martin)
git checkout origin/ASv2 # moves to development branch
git branch release/core-contracts/8 # creates release branch
git checkout release/core-contracts/8 # moves to release branch
git log # checks last commit on release branch is as expected
git push origin release/core-contracts/8 # publishes release branch
```
- [x] this is the release branch: [`release/core-contracts/8`](https://github.com/celo-org/celo-monorepo/commits/release/core-contracts/8) (cut from our development branch [`ASv2`](https://github.com/celo-org/celo-monorepo/tree/ASv2/packages/protocol/contracts/identity), last commit `c1bd0574`)
- **Tag commit**:
- > The latest commit on the release branch is tagged with `core-contracts.v${N}.pre-audit`.
- Commands used (for review):
```bash
git log # check commits on release branch
git show core-contracts.v8.pre-audit # checks if tag already exists (it does, Eela already tagged it)
```
- [x] this is the tag: [`core-contracts.v8.pre-audit`](https://github.com/celo-org/celo-monorepo/releases/tag/core-contracts.v8.pre-audit) for commit `c1bd0574`
- **Update config.yml**:
- > On master branch, [`.circleci/config.yml`](https://github.com/celo-org/celo-monorepo/blob/c1bd057484c4601436eb35c563d431f80d5f94b6/.circleci/config.yml#L37) should be edited so that the variable `RELEASE_TAG` points to the tag `celo-core-contracts-v${N}.pre-audit` so that all future changes to master are versioned against the new release. ~ Source: [Git release management](https://docs.celo.org/community/release-process/smart-contracts#when-a-new-release-branch-is-cut)
- > On the release branch, it should remain `core-contracts.v7`. [~Source: @m-chrzan]
- Commands used (for review):
```bash
# currently on master branch in monorepo
git branch 0xarthurxyz/modify-circleci-configyml-for-CoreContracts8Release # makes branch for my 1 line change
git checkout 0xarthurxyz/modify-circleci-configyml-for-CoreContracts8Release # moves to my branch
# made 1 line change
# committed to my branch (in VS Code UI)
# pushed my branch to remote (in VS Code UI)
# created PR in Github UI
```
- [x] this is the PR: https://github.com/celo-org/celo-monorepo/pull/9734
- **Pre-release notes**:
- > On Github, a pre-release Github release should be created pointing at the latest tag on the release branch. ~ Source: [Git release management](https://docs.celo.org/community/release-process/smart-contracts#when-a-new-release-branch-is-cut)
- Followed official Github instructions here: [Managing releases in a repository](https://docs.github.com/en/repositories/releasing-projects-on-github/managing-releases-in-a-repository
)
- [x] this is the pre-release: [Core Contracts Release 8](https://github.com/celo-org/celo-monorepo/releases/tag/core-contracts.v8.pre-audit)
- [x] @eelanagaraj: Submit release branch to a reputable third party **auditor for review**.
- Completed with Hacken.io
- **Begin drafting release notes.**
- [x] #9724
T+1w:
- [x] Receive report from auditors.
- [x] @0xarthurxyz: Add audit summary to final draft of the release notes.
- If all issues in the audit report have straightforward fixes:
- **CGP**:
- [ ] @0xarthurxyz: Submit a governance proposal draft
- Started the CGP draft on this branch: [0xarthurxyz/cgp-61-core-contracts-8](https://github.com/celo-org/governance/tree/0xarthurxyz/cgp-61-core-contracts-8)
- Added: [cgp-0061.md](https://github.com/celo-org/governance/blob/0xarthurxyz/cgp-61-core-contracts-8/CGPs/cgp-0061.md) (following example of Core Contract 7 release: [cgp-0055.md](https://github.com/celo-org/governance/blob/main/CGPs/cgp-0055.md))
- Added: [cgp-0061.json](https://github.com/celo-org/governance/blob/0xarthurxyz/cgp-61-core-contracts-8/CGPs/cgp-0061/cgp-0061.json) (following example of Core Contract 7 release: [cgp-0055.json](https://github.com/celo-org/governance/blob/main/CGPs/cgp-0055/cgp-0055.json))
- **`.json` file for CGP**:
- [x] @alecps: Add any initialization data to the CGP that should be included as part of the proposal
- **Forum post**:
- [x] @0xarthurxyz: Announce forthcoming smart contract release on: https://forum.celo.org/c/governance
- Created draft forum post: [Core Contracts Release 8 [DRAFT]](https://forum.celo.org/t/core-contracts-release-8-draft/4050)
- [x] Present "_Core Contract Release 8_" at Celo Governance Call
- https://github.com/celo-org/governance/issues/164
- [x] @eelanagaraj: Commit audit fixes to the release branch
- This is the release branch: https://github.com/celo-org/celo-monorepo/tree/ASv2/packages/protocol/contracts/identity
- [x] Submit audit fixes to auditors for review.
- Completed with: #9678
- [x] @alecps: Tag the first release candidate commit according to the [git release management instructions](https://docs.celo.org/community/release-process/smart-contracts#During-the-release-proposal-stage).
- [x] @0xarthurxyz: Let the community know about the upcoming release proposal by posting details to the Governance category on https://forum.celo.org and cross post in the [Discord #governance channel](https://discord.com/channels/600834479145353243/704805825373274134).See the 'Communication guidelines' section below for information on what your post should contain.
T+2w
- [ ] @alecps: On Tuesday: Run the smart contract release script in order to to deploy the contracts to Baklava as well as submit a governance proposal.
- [ ] @0xarthurxyz: Transition proposal through Baklava governance process.
- [ ] @0xarthurxyz: Update your forum post with the Baklava `PROPOSAL_ID`, updated timings (if any changes), and notify the community in the [Discord #governance channel](https://discord.com/channels/600834479145353243/704805825373274134).
T+3w
- [ ] @alecps: Confirm all contracts working as intended on Baklava.
- [ ] @alecps: Run the [smart contract release script](https://docs.celo.org/community/release-process/smart-contracts#release-process) in order to to deploy the contracts to Alfajores as well as submit a governance proposal.
- [ ] @0xarthurxyz: Update your forum post with the Alfajores `PROPOSAL_ID`, updated timings (if any changes), and notify the community in the [Discord #governance channel](https://discord.com/channels/600834479145353243/704805825373274134).
T+4w
- [ ] @alecps: Confirm all contracts working as intended on Alfajores.
- [ ] @0xarthurxyz: Confirm audit is complete and make the release notes and forum post contain a link to it.
- [ ] @alecps: On Tuesday: Run the [smart contract release script](https://docs.celo.org/community/release-process/smart-contracts#build-process) in order to to deploy the contracts to Mainnet as well as submit a governance proposal.
- [ ] Update the corresponding governance proposal with the updated on-chain `PROPOSAL_ID` and mark CGP status as `"PROPOSED"`.
- [ ] @0xarthurxyz: Update your forum post with the Mainnet `PROPOSAL_ID`, updated timings (if any changes), and notify the community in the [Discord #governance channel](https://discord.com/channels/600834479145353243/704805825373274134).
- At this point all stakeholders are encouraged to [verify](https://docs.celo.org/community/release-process/smart-contracts#verify-release-process) the proposed contracts deployed match the contracts from the release branch.
- [ ] @0xarthurxyz: Monitor the progress of the proposal through the [governance process](https://docs.celo.org/celo-codebase/protocol/governance).
- Currently the governance process should take approximately 1 week: 24 hours for the dequeue process, 24 hours for the approval process, and 5 days for the referendum process. After which, the proposal is either declined or is ready to be executed within 3 days.
- For updated timeframes, use the celocli: `celocli network:parameters`
T+5w
- If the proposal passed:
- [ ] @alecps: Confirm all contracts working as intended on Mainnet.
- [ ] @0xarthurxyz: Update your forum post with the Mainnet governance outcome (`Passed` or `Rejected`) and notify the community in the [Discord #governance channel](https://discord.com/channels/600834479145353243/704805825373274134).
- [ ] @0xarthurxyz: Change corresponding CGP status to `EXCECUTED`.
- [ ] @alecps: Merge the release branch into `master` with a merge commit
- If the proposal failed:
- Change corresponding CGP status to `EXPIRED`. | process | shepherd core contracts release github issue to keep track checklist in t create a github issue tracking all these checklist items as an audit log this is it implement the for when a new release branch is cut release branch a new release branch is created release core contracts n with the contracts to be audited there already exists a release core contracts branch but it s deprecated can be deleted and a new one of the same name be created at your release branch commands used for review bash git push origin delete release core contracts delete existing release core contracts branch advised by martin git checkout origin moves to development branch git branch release core contracts creates release branch git checkout release core contracts moves to release branch git log checks last commit on release branch is as expected git push origin release core contracts publishes release branch this is the release branch cut from our development branch last commit tag commit the latest commit on the release branch is tagged with core contracts v n pre audit commands used for review bash git log check commits on release branch git show core contracts pre audit checks if tag already exists it does eela already tagged it this is the tag for commit update config yml on master branch should be edited so that the variable release tag points to the tag celo core contracts v n pre audit so that all future changes to master are versioned against the new release source on the release branch it should remain core contracts commands used for review bash currently on master branch in monorepo git branch modify circleci configyml for makes branch for my line change git checkout modify circleci configyml for moves to my branch made line change committed to my branch in vs code ui pushed my branch to remote in vs code ui created pr in github ui this is the pr pre release notes on github a pre release github release should be created pointing at the latest tag on the release branch source followed official github instructions here this is the pre release eelanagaraj submit release branch to a reputable third party auditor for review completed with hacken io begin drafting release notes t receive report from auditors add audit summary to final draft of the release notes if all issues in the audit report have straightforward fixes cgp submit a governance proposal draft started the cgp draft on this branch added following example of core contract release added following example of core contract release json file for cgp alecps add any initialization data to the cgp that should be included as part of the proposal forum post announce forthcoming smart contract release on created draft forum post present core contract release at celo governance call eelanagaraj commit audit fixes to the release branch this is the release branch submit audit fixes to auditors for review completed with alecps tag the first release candidate commit according to the let the community know about the upcoming release proposal by posting details to the governance category on and cross post in the the communication guidelines section below for information on what your post should contain t alecps on tuesday run the smart contract release script in order to to deploy the contracts to baklava as well as submit a governance proposal transition proposal through baklava governance process update your forum post with the baklava proposal id updated timings if any changes and notify the community in the t alecps confirm all contracts working as intended on baklava alecps run the in order to to deploy the contracts to alfajores as well as submit a governance proposal update your forum post with the alfajores proposal id updated timings if any changes and notify the community in the t alecps confirm all contracts working as intended on alfajores confirm audit is complete and make the release notes and forum post contain a link to it alecps on tuesday run the in order to to deploy the contracts to mainnet as well as submit a governance proposal update the corresponding governance proposal with the updated on chain proposal id and mark cgp status as proposed update your forum post with the mainnet proposal id updated timings if any changes and notify the community in the at this point all stakeholders are encouraged to the proposed contracts deployed match the contracts from the release branch monitor the progress of the proposal through the currently the governance process should take approximately week hours for the dequeue process hours for the approval process and days for the referendum process after which the proposal is either declined or is ready to be executed within days for updated timeframes use the celocli celocli network parameters t if the proposal passed alecps confirm all contracts working as intended on mainnet update your forum post with the mainnet governance outcome passed or rejected and notify the community in the change corresponding cgp status to excecuted alecps merge the release branch into master with a merge commit if the proposal failed change corresponding cgp status to expired | 1 |
21,731 | 30,242,940,247 | IssuesEvent | 2023-07-06 14:36:11 | microsoft/vscode | https://api.github.com/repos/microsoft/vscode | closed | Conpty causes pty host on remote to restart immediately | bug terminal-conpty terminal-process | Repro:
1. Open OSS
2. Ensure conpty is enabled
3. Open new test resolver window
4. Close other windows
5. Open terminal
6. Close test resolver window
7. Open OSS, 🐛 pwsh will launch, then flicker and the prompt will be gone. Looking at the logs reveals the pty host almost immediately restarts because it exits, see `Pty Host exiting, code=1` below
Some extra logs showing traces below:
```
2023-07-05 13:00:56.076 [trace] node-pty.IPty#spawn C:\Users\Daniel\AppData\Local\Microsoft\WindowsApps\Microsoft.PowerShell_8wekyb3d8bbwe\pwsh.exe [] ...
2023-07-05 13:00:56.125 [trace] [RPC Event] PtyService#_onProcessReady.fire({"id":2,"event":{"pid":10176,"cwd":"C:\\Users\\Daniel","windowsPty":{"backend":"conpty","buildNumber":22621}}})
2023-07-05 13:00:56.126 [trace] [RPC Response] PtyService#start undefined
2023-07-05 13:00:56.142 [trace] [RPC Event] PtyService#_onDidChangeProperty.fire({"id":2,"property":{"type":"title","value":"pwsh.exe"}})
2023-07-05 13:00:56.142 [trace] [RPC Event] PtyService#_onDidChangeProperty.fire({"id":2,"property":{"type":"shellType"}})
2023-07-05 13:00:56.268 [trace] node-pty.IPty#onData [?25l[2J[m[HPowerShell 7.3.5
]0;C:\Program Files\WindowsApps\Microsoft.PowerShell_7.3.5.0_x64__8wekyb3d8bbwe\pwsh.exe[?25h
2023-07-05 13:00:56.286 [trace] [RPC Event] PtyService#_onProcessData.fire({"id":2,"event":"\u001b[?25l\u001b[2J\u001b[m\u001b[HPowerShell 7.3.5\r\n\u001b]0;C:\\Program Files\\WindowsApps\\Microsoft.PowerShell_7.3.5.0_x64__8wekyb3d8bbwe\\pwsh.exe\u0007\u001b[?25h"})
2023-07-05 13:00:56.333 [trace] node-pty.IPty#kill
2023-07-05 13:00:56.340 [trace] [RPC Event] PtyService#_onProcessExit.fire({"id":1,"event":0})
2023-07-05 13:00:56.414 [trace] Pty Host exiting, code=1
2023-07-05 13:00:56.414 [trace] TerminalProcess#shutdown
at TerminalProcess.shutdown (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\terminalProcess.js:375:64)
at PersistentTerminalProcess.shutdown (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\ptyService.js:762:42)
at c:\Github\microsoft\vscode\out\vs\platform\terminal\node\ptyService.js:89:25
at c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:144:17
at Object.dispose (c:\Github\microsoft\vscode\out\vs\base\common\functional.js:18:25)
at dispose (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:96:27)
at DisposableStore.clear (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:191:17)
at DisposableStore.dispose (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:175:18)
at PtyService.dispose (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:239:25)
at process.<anonymous> (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\ptyHostMain.js:68:24)
at Object.onceWrapper (node:events:628:26)
at process.emit (node:events:513:28)
at process._fatalException (node:internal/process/execution:174:19)
2023-07-05 13:00:56.414 [trace] TerminalProcess#_queueProcessExit
at TerminalProcess._queueProcessExit (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\terminalProcess.js:314:73)
at TerminalProcess.shutdown (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\terminalProcess.js:384:26)
at PersistentTerminalProcess.shutdown (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\ptyService.js:762:42)
at c:\Github\microsoft\vscode\out\vs\platform\terminal\node\ptyService.js:89:25
at c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:144:17
at Object.dispose (c:\Github\microsoft\vscode\out\vs\base\common\functional.js:18:25)
at dispose (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:96:27)
at DisposableStore.clear (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:191:17)
at DisposableStore.dispose (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:175:18)
at PtyService.dispose (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:239:25)
at process.<anonymous> (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\ptyHostMain.js:68:24)
at Object.onceWrapper (node:events:628:26)
at process.emit (node:events:513:28)
at process._fatalException (node:internal/process/execution:174:19)
```
Prompt briefly appears and then goes away:

| 1.0 | Conpty causes pty host on remote to restart immediately - Repro:
1. Open OSS
2. Ensure conpty is enabled
3. Open new test resolver window
4. Close other windows
5. Open terminal
6. Close test resolver window
7. Open OSS, 🐛 pwsh will launch, then flicker and the prompt will be gone. Looking at the logs reveals the pty host almost immediately restarts because it exits, see `Pty Host exiting, code=1` below
Some extra logs showing traces below:
```
2023-07-05 13:00:56.076 [trace] node-pty.IPty#spawn C:\Users\Daniel\AppData\Local\Microsoft\WindowsApps\Microsoft.PowerShell_8wekyb3d8bbwe\pwsh.exe [] ...
2023-07-05 13:00:56.125 [trace] [RPC Event] PtyService#_onProcessReady.fire({"id":2,"event":{"pid":10176,"cwd":"C:\\Users\\Daniel","windowsPty":{"backend":"conpty","buildNumber":22621}}})
2023-07-05 13:00:56.126 [trace] [RPC Response] PtyService#start undefined
2023-07-05 13:00:56.142 [trace] [RPC Event] PtyService#_onDidChangeProperty.fire({"id":2,"property":{"type":"title","value":"pwsh.exe"}})
2023-07-05 13:00:56.142 [trace] [RPC Event] PtyService#_onDidChangeProperty.fire({"id":2,"property":{"type":"shellType"}})
2023-07-05 13:00:56.268 [trace] node-pty.IPty#onData [?25l[2J[m[HPowerShell 7.3.5
]0;C:\Program Files\WindowsApps\Microsoft.PowerShell_7.3.5.0_x64__8wekyb3d8bbwe\pwsh.exe[?25h
2023-07-05 13:00:56.286 [trace] [RPC Event] PtyService#_onProcessData.fire({"id":2,"event":"\u001b[?25l\u001b[2J\u001b[m\u001b[HPowerShell 7.3.5\r\n\u001b]0;C:\\Program Files\\WindowsApps\\Microsoft.PowerShell_7.3.5.0_x64__8wekyb3d8bbwe\\pwsh.exe\u0007\u001b[?25h"})
2023-07-05 13:00:56.333 [trace] node-pty.IPty#kill
2023-07-05 13:00:56.340 [trace] [RPC Event] PtyService#_onProcessExit.fire({"id":1,"event":0})
2023-07-05 13:00:56.414 [trace] Pty Host exiting, code=1
2023-07-05 13:00:56.414 [trace] TerminalProcess#shutdown
at TerminalProcess.shutdown (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\terminalProcess.js:375:64)
at PersistentTerminalProcess.shutdown (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\ptyService.js:762:42)
at c:\Github\microsoft\vscode\out\vs\platform\terminal\node\ptyService.js:89:25
at c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:144:17
at Object.dispose (c:\Github\microsoft\vscode\out\vs\base\common\functional.js:18:25)
at dispose (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:96:27)
at DisposableStore.clear (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:191:17)
at DisposableStore.dispose (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:175:18)
at PtyService.dispose (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:239:25)
at process.<anonymous> (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\ptyHostMain.js:68:24)
at Object.onceWrapper (node:events:628:26)
at process.emit (node:events:513:28)
at process._fatalException (node:internal/process/execution:174:19)
2023-07-05 13:00:56.414 [trace] TerminalProcess#_queueProcessExit
at TerminalProcess._queueProcessExit (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\terminalProcess.js:314:73)
at TerminalProcess.shutdown (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\terminalProcess.js:384:26)
at PersistentTerminalProcess.shutdown (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\ptyService.js:762:42)
at c:\Github\microsoft\vscode\out\vs\platform\terminal\node\ptyService.js:89:25
at c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:144:17
at Object.dispose (c:\Github\microsoft\vscode\out\vs\base\common\functional.js:18:25)
at dispose (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:96:27)
at DisposableStore.clear (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:191:17)
at DisposableStore.dispose (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:175:18)
at PtyService.dispose (c:\Github\microsoft\vscode\out\vs\base\common\lifecycle.js:239:25)
at process.<anonymous> (c:\Github\microsoft\vscode\out\vs\platform\terminal\node\ptyHostMain.js:68:24)
at Object.onceWrapper (node:events:628:26)
at process.emit (node:events:513:28)
at process._fatalException (node:internal/process/execution:174:19)
```
Prompt briefly appears and then goes away:

| process | conpty causes pty host on remote to restart immediately repro open oss ensure conpty is enabled open new test resolver window close other windows open terminal close test resolver window open oss 🐛 pwsh will launch then flicker and the prompt will be gone looking at the logs reveals the pty host almost immediately restarts because it exits see pty host exiting code below some extra logs showing traces below node pty ipty spawn c users daniel appdata local microsoft windowsapps microsoft powershell pwsh exe ptyservice onprocessready fire id event pid cwd c users daniel windowspty backend conpty buildnumber ptyservice start undefined ptyservice ondidchangeproperty fire id property type title value pwsh exe ptyservice ondidchangeproperty fire id property type shelltype node pty ipty ondata m hpowershell c program files windowsapps microsoft powershell pwsh exe ptyservice onprocessdata fire id event c program files windowsapps microsoft powershell pwsh exe node pty ipty kill ptyservice onprocessexit fire id event pty host exiting code terminalprocess shutdown at terminalprocess shutdown c github microsoft vscode out vs platform terminal node terminalprocess js at persistentterminalprocess shutdown c github microsoft vscode out vs platform terminal node ptyservice js at c github microsoft vscode out vs platform terminal node ptyservice js at c github microsoft vscode out vs base common lifecycle js at object dispose c github microsoft vscode out vs base common functional js at dispose c github microsoft vscode out vs base common lifecycle js at disposablestore clear c github microsoft vscode out vs base common lifecycle js at disposablestore dispose c github microsoft vscode out vs base common lifecycle js at ptyservice dispose c github microsoft vscode out vs base common lifecycle js at process c github microsoft vscode out vs platform terminal node ptyhostmain js at object oncewrapper node events at process emit node events at process fatalexception node internal process execution terminalprocess queueprocessexit at terminalprocess queueprocessexit c github microsoft vscode out vs platform terminal node terminalprocess js at terminalprocess shutdown c github microsoft vscode out vs platform terminal node terminalprocess js at persistentterminalprocess shutdown c github microsoft vscode out vs platform terminal node ptyservice js at c github microsoft vscode out vs platform terminal node ptyservice js at c github microsoft vscode out vs base common lifecycle js at object dispose c github microsoft vscode out vs base common functional js at dispose c github microsoft vscode out vs base common lifecycle js at disposablestore clear c github microsoft vscode out vs base common lifecycle js at disposablestore dispose c github microsoft vscode out vs base common lifecycle js at ptyservice dispose c github microsoft vscode out vs base common lifecycle js at process c github microsoft vscode out vs platform terminal node ptyhostmain js at object oncewrapper node events at process emit node events at process fatalexception node internal process execution prompt briefly appears and then goes away | 1 |
5,909 | 8,728,446,530 | IssuesEvent | 2018-12-10 17:24:10 | googleapis/google-cloud-java | https://api.github.com/repos/googleapis/google-cloud-java | closed | Figure out permission settings for requester pays bucket integration tests | api: storage priority: p2 type: process | `testCantCreateWithoutUserProject`, `testCantReadWithoutUserProject`, `testCantCopyWithoutUserProject` and `testFileExistsRequesterPaysNoUserProject` (https://github.com/GoogleCloudPlatform/google-cloud-java/blob/92b87dc9dd28fdaa523ecca0dfce02514a2c7297/google-cloud-clients/google-cloud-contrib/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/it/ITGcsNio.java) are failing because the testing project has `resourcemanager.projects.createBillingAssignment` permission. Currently ignoring these tests. | 1.0 | Figure out permission settings for requester pays bucket integration tests - `testCantCreateWithoutUserProject`, `testCantReadWithoutUserProject`, `testCantCopyWithoutUserProject` and `testFileExistsRequesterPaysNoUserProject` (https://github.com/GoogleCloudPlatform/google-cloud-java/blob/92b87dc9dd28fdaa523ecca0dfce02514a2c7297/google-cloud-clients/google-cloud-contrib/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/it/ITGcsNio.java) are failing because the testing project has `resourcemanager.projects.createBillingAssignment` permission. Currently ignoring these tests. | process | figure out permission settings for requester pays bucket integration tests testcantcreatewithoutuserproject testcantreadwithoutuserproject testcantcopywithoutuserproject and testfileexistsrequesterpaysnouserproject are failing because the testing project has resourcemanager projects createbillingassignment permission currently ignoring these tests | 1 |
147,146 | 13,201,515,835 | IssuesEvent | 2020-08-14 10:18:05 | swat-flights/Master_Flights | https://api.github.com/repos/swat-flights/Master_Flights | opened | Entrega preliminar de lista de actividades | documentation good first issue planning and followup | Se requiere la lista de actividades como documentación indicada en requerimientos del projecto | 1.0 | Entrega preliminar de lista de actividades - Se requiere la lista de actividades como documentación indicada en requerimientos del projecto | non_process | entrega preliminar de lista de actividades se requiere la lista de actividades como documentación indicada en requerimientos del projecto | 0 |
12,524 | 14,967,882,785 | IssuesEvent | 2021-01-27 16:11:26 | panther-labs/panther | https://api.github.com/repos/panther-labs/panther | closed | Alert_Context fails if user passes event | bug p0 team:data processing | ### Describe the bug
If the users decides to return the provided event from `alert_context`, the `alert_context` method will fail
### Steps to reproduce
Steps to reproduce the behavior:
1. Go to Log Analysis and select to create a new rule
2. Create a new rule with body:
```
def rule(event):
return event
def alert_context(event):
return event
```
3. See error
### Expected behavior
Users should be able to return event through alert_context method
### Environment
- Panther version or commit: 1.15.2
| 1.0 | Alert_Context fails if user passes event - ### Describe the bug
If the users decides to return the provided event from `alert_context`, the `alert_context` method will fail
### Steps to reproduce
Steps to reproduce the behavior:
1. Go to Log Analysis and select to create a new rule
2. Create a new rule with body:
```
def rule(event):
return event
def alert_context(event):
return event
```
3. See error
### Expected behavior
Users should be able to return event through alert_context method
### Environment
- Panther version or commit: 1.15.2
| process | alert context fails if user passes event describe the bug if the users decides to return the provided event from alert context the alert context method will fail steps to reproduce steps to reproduce the behavior go to log analysis and select to create a new rule create a new rule with body def rule event return event def alert context event return event see error expected behavior users should be able to return event through alert context method environment panther version or commit | 1 |
14,776 | 18,051,381,452 | IssuesEvent | 2021-09-19 20:07:44 | googlemaps/maps-sdk-for-ios-samples | https://api.github.com/repos/googlemaps/maps-sdk-for-ios-samples | closed | snippet-bot full scan | type: process stale | <!-- probot comment [11337970]-->
## snippet-bot scan result
Life is too short to manually check unmatched region tags.
Here is the result:
- [ ] [snippets/Podfile:14](https://github.com/googlemaps/maps-sdk-for-ios-samples/blob/a0fc38b/snippets/Podfile#L14), tag `maps_ios_get_started_install_podfile` doesn't have a matching start tag
- [ ] [snippets/Podfile:9](https://github.com/googlemaps/maps-sdk-for-ios-samples/blob/a0fc38b/snippets/Podfile#L9), tag `maps_ios_places_get_started_install_podfile` doesn't have a matching end tag
---
Report generated by [snippet-bot](https://github.com/apps/snippet-bot).
If you find problems with this result, please file an issue at:
https://github.com/googleapis/repo-automation-bots/issues.
| 1.0 | snippet-bot full scan - <!-- probot comment [11337970]-->
## snippet-bot scan result
Life is too short to manually check unmatched region tags.
Here is the result:
- [ ] [snippets/Podfile:14](https://github.com/googlemaps/maps-sdk-for-ios-samples/blob/a0fc38b/snippets/Podfile#L14), tag `maps_ios_get_started_install_podfile` doesn't have a matching start tag
- [ ] [snippets/Podfile:9](https://github.com/googlemaps/maps-sdk-for-ios-samples/blob/a0fc38b/snippets/Podfile#L9), tag `maps_ios_places_get_started_install_podfile` doesn't have a matching end tag
---
Report generated by [snippet-bot](https://github.com/apps/snippet-bot).
If you find problems with this result, please file an issue at:
https://github.com/googleapis/repo-automation-bots/issues.
| process | snippet bot full scan snippet bot scan result life is too short to manually check unmatched region tags here is the result tag maps ios get started install podfile doesn t have a matching start tag tag maps ios places get started install podfile doesn t have a matching end tag report generated by if you find problems with this result please file an issue at | 1 |
1,496 | 4,074,094,257 | IssuesEvent | 2016-05-28 06:38:07 | opentrials/opentrials | https://api.github.com/repos/opentrials/opentrials | opened | Implement persons deduplication | enhancement Processors | Entities with dedup based on unique slug:
- condition
- intervention
- location
- organisation
Entities with dedup can't be based on unique slug:
- trial
- person (not implemented for now) | 1.0 | Implement persons deduplication - Entities with dedup based on unique slug:
- condition
- intervention
- location
- organisation
Entities with dedup can't be based on unique slug:
- trial
- person (not implemented for now) | process | implement persons deduplication entities with dedup based on unique slug condition intervention location organisation entities with dedup can t be based on unique slug trial person not implemented for now | 1 |
509,436 | 14,730,988,118 | IssuesEvent | 2021-01-06 14:04:04 | fedora-infra/anitya | https://api.github.com/repos/fedora-infra/anitya | closed | Rewrite projects pages | Medium Priority enhancement groomed hacktoberfest | # Description
The projects menu in Anitya is broken for some time and it will be nice to fix it. There are plenty of small things that needs to be addressed.
## Requirements
- [x] Remove `Odd version found` menu option (Odd version no longer exists, because we are retrieving all new versions, not only latest)
- [x] Add header row (this will add more readability and it will not be less confusing for users)
- [x] Add page turner to bottom of the page (most of the time you want to go to next page, you actually looking for something on the page and scrolling down, it will be good to be able to go to next page on both top and bottom of the table)
- [x] Show backend on projects table (this information could be valuable to see)
- [x] Show correct projects in correct categories (use `check_failed` and `versions`) (currently some categories either doesn't show anything or shows wrong projects, it will be nice to fix this)
- [ ] Add ability to sort by column (this will be very nice, if you are looking for project with specific backend or the most failures)
## How this will make Anitya better
Anitya will be much more user friendly when looking for projects and the projects menu will be much more helpful for users. | 1.0 | Rewrite projects pages - # Description
The projects menu in Anitya is broken for some time and it will be nice to fix it. There are plenty of small things that needs to be addressed.
## Requirements
- [x] Remove `Odd version found` menu option (Odd version no longer exists, because we are retrieving all new versions, not only latest)
- [x] Add header row (this will add more readability and it will not be less confusing for users)
- [x] Add page turner to bottom of the page (most of the time you want to go to next page, you actually looking for something on the page and scrolling down, it will be good to be able to go to next page on both top and bottom of the table)
- [x] Show backend on projects table (this information could be valuable to see)
- [x] Show correct projects in correct categories (use `check_failed` and `versions`) (currently some categories either doesn't show anything or shows wrong projects, it will be nice to fix this)
- [ ] Add ability to sort by column (this will be very nice, if you are looking for project with specific backend or the most failures)
## How this will make Anitya better
Anitya will be much more user friendly when looking for projects and the projects menu will be much more helpful for users. | non_process | rewrite projects pages description the projects menu in anitya is broken for some time and it will be nice to fix it there are plenty of small things that needs to be addressed requirements remove odd version found menu option odd version no longer exists because we are retrieving all new versions not only latest add header row this will add more readability and it will not be less confusing for users add page turner to bottom of the page most of the time you want to go to next page you actually looking for something on the page and scrolling down it will be good to be able to go to next page on both top and bottom of the table show backend on projects table this information could be valuable to see show correct projects in correct categories use check failed and versions currently some categories either doesn t show anything or shows wrong projects it will be nice to fix this add ability to sort by column this will be very nice if you are looking for project with specific backend or the most failures how this will make anitya better anitya will be much more user friendly when looking for projects and the projects menu will be much more helpful for users | 0 |
16,667 | 2,925,506,214 | IssuesEvent | 2015-06-26 06:31:04 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | MKL test failures | defect scipy.linalg | On Ubuntu 14.10 64-bit with system Python 2.7, latest MKL (15.0.0), cython 0.22.x, numpy 1.9.2, scipy 0.16 branch I get about 22 failures/errors with `scipy.test('full')`. `numpy.test('full')` passes (except for one test that runs out of disk space). I was running MKL version 13 last week and had similar failures.
I have tried the following, but they all fail in similar but not identical ways:
- using the default values in `numpy/distuils/*`
- [intel's instructions] (https://software.intel.com/en-us/articles/numpyscipy-with-intel-mkl)
- another [person's instructions] (https://gehrcke.de/2014/02/building-numpy-and-scipy-with-intel-compilers-and-intel-mkl-on-a-64-bit-machine)
I have also rolled back to the `maintenance/0.15.x` branch, and I get about ~~16~~ EDIT: 3 failures there, too -- so I assume there is something wrong with my compilation setup but I'm not sure what it could be...
All failures are [in this gist] (https://gist.github.com/Eric89GXL/cc95600b8dc0d67cce66), but here are a few example failures:
```
======================================================================
ERROR: test_decomp_update.TestQRdelete_f.test_economic_p_row_fat
----------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/lib/python2.7/dist-packages/nose/case.py", line 197, in runTest
self.test(*self.arg)
File "/home/larsoner/.local/lib/python2.7/site-packages/scipy/linalg/tests/test_decomp_update.py", line 230, in test_economic_p_row_fat
self.base_economic_p_row_xxx(7)
File "/home/larsoner/.local/lib/python2.7/site-packages/scipy/linalg/tests/test_decomp_update.py", line 216, in base_economic_p_row_xxx
q1, r1 = qr_delete(q, r, row, ndel, overwrite_qr=False)
File "scipy/linalg/_decomp_update.pyx", line 1652, in scipy.linalg._decomp_update.qr_delete (scipy/linalg/_decomp_update.c:27659)
ValueError: Reorthogonalization Failed, unable to perform row deletion.
======================================================================
FAIL: test_decomp_update.TestQRdelete_D.test_economic_1_row
----------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/lib/python2.7/dist-packages/nose/case.py", line 197, in runTest
self.test(*self.arg)
File "/home/larsoner/.local/lib/python2.7/site-packages/scipy/linalg/tests/test_decomp_update.py", line 207, in test_economic_1_row
check_qr(q1, r1, a1, self.rtol, self.atol, False)
File "/home/larsoner/.local/lib/python2.7/site-packages/scipy/linalg/tests/test_decomp_update.py", line 32, in check_qr
assert_unitary(q, rtol, atol, assert_sqr)
File "/home/larsoner/.local/lib/python2.7/site-packages/scipy/linalg/tests/test_decomp_update.py", line 21, in assert_unitary
assert_allclose(aTa, np.eye(a.shape[1]), rtol=rtol, atol=atol)
File "/home/larsoner/.local/lib/python2.7/site-packages/numpy/testing/utils.py", line 1297, in assert_allclose
verbose=verbose, header=header)
File "/home/larsoner/.local/lib/python2.7/site-packages/numpy/testing/utils.py", line 665, in assert_array_compare
raise AssertionError(msg)
AssertionError:
Not equal to tolerance rtol=1e-13, atol=2.22045e-15
(mismatch 100.0%)
x: array([[ 9.803958e-01+0.j , 2.215354e-02-0.092053j,
2.161659e-02-0.030168j, -4.511673e-04-0.013876j,
-4.492602e-04-0.090115j, 3.500837e-02-0.059986j,...
y: array([[ 1., 0., 0., 0., 0., 0., 0.],
[ 0., 1., 0., 0., 0., 0., 0.],
[ 0., 0., 1., 0., 0., 0., 0.],...
======================================================================
FAIL: test_multi (test_odr.TestODR)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/home/larsoner/.local/lib/python2.7/site-packages/scipy/odr/tests/test_odr.py", line 190, in test_multi
0.5101147161764654, 0.5173902330489161]),
File "/home/larsoner/.local/lib/python2.7/site-packages/numpy/testing/utils.py", line 842, in assert_array_almost_equal
precision=decimal)
File "/home/larsoner/.local/lib/python2.7/site-packages/numpy/testing/utils.py", line 665, in assert_array_compare
raise AssertionError(msg)
AssertionError:
Arrays are not almost equal to 6 decimals
(mismatch 100.0%)
x: array([ 4. , 2. , 7. , 0.4, 0.5])
y: array([ 4.379988, 2.433306, 8.002885, 0.510115, 0.51739 ])
```
| 1.0 | MKL test failures - On Ubuntu 14.10 64-bit with system Python 2.7, latest MKL (15.0.0), cython 0.22.x, numpy 1.9.2, scipy 0.16 branch I get about 22 failures/errors with `scipy.test('full')`. `numpy.test('full')` passes (except for one test that runs out of disk space). I was running MKL version 13 last week and had similar failures.
I have tried the following, but they all fail in similar but not identical ways:
- using the default values in `numpy/distuils/*`
- [intel's instructions] (https://software.intel.com/en-us/articles/numpyscipy-with-intel-mkl)
- another [person's instructions] (https://gehrcke.de/2014/02/building-numpy-and-scipy-with-intel-compilers-and-intel-mkl-on-a-64-bit-machine)
I have also rolled back to the `maintenance/0.15.x` branch, and I get about ~~16~~ EDIT: 3 failures there, too -- so I assume there is something wrong with my compilation setup but I'm not sure what it could be...
All failures are [in this gist] (https://gist.github.com/Eric89GXL/cc95600b8dc0d67cce66), but here are a few example failures:
```
======================================================================
ERROR: test_decomp_update.TestQRdelete_f.test_economic_p_row_fat
----------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/lib/python2.7/dist-packages/nose/case.py", line 197, in runTest
self.test(*self.arg)
File "/home/larsoner/.local/lib/python2.7/site-packages/scipy/linalg/tests/test_decomp_update.py", line 230, in test_economic_p_row_fat
self.base_economic_p_row_xxx(7)
File "/home/larsoner/.local/lib/python2.7/site-packages/scipy/linalg/tests/test_decomp_update.py", line 216, in base_economic_p_row_xxx
q1, r1 = qr_delete(q, r, row, ndel, overwrite_qr=False)
File "scipy/linalg/_decomp_update.pyx", line 1652, in scipy.linalg._decomp_update.qr_delete (scipy/linalg/_decomp_update.c:27659)
ValueError: Reorthogonalization Failed, unable to perform row deletion.
======================================================================
FAIL: test_decomp_update.TestQRdelete_D.test_economic_1_row
----------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/lib/python2.7/dist-packages/nose/case.py", line 197, in runTest
self.test(*self.arg)
File "/home/larsoner/.local/lib/python2.7/site-packages/scipy/linalg/tests/test_decomp_update.py", line 207, in test_economic_1_row
check_qr(q1, r1, a1, self.rtol, self.atol, False)
File "/home/larsoner/.local/lib/python2.7/site-packages/scipy/linalg/tests/test_decomp_update.py", line 32, in check_qr
assert_unitary(q, rtol, atol, assert_sqr)
File "/home/larsoner/.local/lib/python2.7/site-packages/scipy/linalg/tests/test_decomp_update.py", line 21, in assert_unitary
assert_allclose(aTa, np.eye(a.shape[1]), rtol=rtol, atol=atol)
File "/home/larsoner/.local/lib/python2.7/site-packages/numpy/testing/utils.py", line 1297, in assert_allclose
verbose=verbose, header=header)
File "/home/larsoner/.local/lib/python2.7/site-packages/numpy/testing/utils.py", line 665, in assert_array_compare
raise AssertionError(msg)
AssertionError:
Not equal to tolerance rtol=1e-13, atol=2.22045e-15
(mismatch 100.0%)
x: array([[ 9.803958e-01+0.j , 2.215354e-02-0.092053j,
2.161659e-02-0.030168j, -4.511673e-04-0.013876j,
-4.492602e-04-0.090115j, 3.500837e-02-0.059986j,...
y: array([[ 1., 0., 0., 0., 0., 0., 0.],
[ 0., 1., 0., 0., 0., 0., 0.],
[ 0., 0., 1., 0., 0., 0., 0.],...
======================================================================
FAIL: test_multi (test_odr.TestODR)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/home/larsoner/.local/lib/python2.7/site-packages/scipy/odr/tests/test_odr.py", line 190, in test_multi
0.5101147161764654, 0.5173902330489161]),
File "/home/larsoner/.local/lib/python2.7/site-packages/numpy/testing/utils.py", line 842, in assert_array_almost_equal
precision=decimal)
File "/home/larsoner/.local/lib/python2.7/site-packages/numpy/testing/utils.py", line 665, in assert_array_compare
raise AssertionError(msg)
AssertionError:
Arrays are not almost equal to 6 decimals
(mismatch 100.0%)
x: array([ 4. , 2. , 7. , 0.4, 0.5])
y: array([ 4.379988, 2.433306, 8.002885, 0.510115, 0.51739 ])
```
| non_process | mkl test failures on ubuntu bit with system python latest mkl cython x numpy scipy branch i get about failures errors with scipy test full numpy test full passes except for one test that runs out of disk space i was running mkl version last week and had similar failures i have tried the following but they all fail in similar but not identical ways using the default values in numpy distuils another i have also rolled back to the maintenance x branch and i get about edit failures there too so i assume there is something wrong with my compilation setup but i m not sure what it could be all failures are but here are a few example failures error test decomp update testqrdelete f test economic p row fat traceback most recent call last file usr lib dist packages nose case py line in runtest self test self arg file home larsoner local lib site packages scipy linalg tests test decomp update py line in test economic p row fat self base economic p row xxx file home larsoner local lib site packages scipy linalg tests test decomp update py line in base economic p row xxx qr delete q r row ndel overwrite qr false file scipy linalg decomp update pyx line in scipy linalg decomp update qr delete scipy linalg decomp update c valueerror reorthogonalization failed unable to perform row deletion fail test decomp update testqrdelete d test economic row traceback most recent call last file usr lib dist packages nose case py line in runtest self test self arg file home larsoner local lib site packages scipy linalg tests test decomp update py line in test economic row check qr self rtol self atol false file home larsoner local lib site packages scipy linalg tests test decomp update py line in check qr assert unitary q rtol atol assert sqr file home larsoner local lib site packages scipy linalg tests test decomp update py line in assert unitary assert allclose ata np eye a shape rtol rtol atol atol file home larsoner local lib site packages numpy testing utils py line in assert allclose verbose verbose header header file home larsoner local lib site packages numpy testing utils py line in assert array compare raise assertionerror msg assertionerror not equal to tolerance rtol atol mismatch x array j y array fail test multi test odr testodr traceback most recent call last file home larsoner local lib site packages scipy odr tests test odr py line in test multi file home larsoner local lib site packages numpy testing utils py line in assert array almost equal precision decimal file home larsoner local lib site packages numpy testing utils py line in assert array compare raise assertionerror msg assertionerror arrays are not almost equal to decimals mismatch x array y array | 0 |
24,213 | 5,038,683,783 | IssuesEvent | 2016-12-18 11:56:50 | poweradmin/poweradmin | https://api.github.com/repos/poweradmin/poweradmin | closed | Missing update info from older versions (2.1.3 -> 2.1.4) | documentation | Hi
I am missing the update info from 2.1.3 to 2.1.4 (so I can upgrade a installation to the latest). Is it possible to add this to the wiki?
/ Claus | 1.0 | Missing update info from older versions (2.1.3 -> 2.1.4) - Hi
I am missing the update info from 2.1.3 to 2.1.4 (so I can upgrade a installation to the latest). Is it possible to add this to the wiki?
/ Claus | non_process | missing update info from older versions hi i am missing the update info from to so i can upgrade a installation to the latest is it possible to add this to the wiki claus | 0 |
339,239 | 24,614,939,088 | IssuesEvent | 2022-10-15 07:16:34 | alterae/mksite | https://api.github.com/repos/alterae/mksite | opened | Website overhaul | documentation | Something shiny, with good typography and nice gradients and shadows and the like. Should include an html version of the man pages(s) (see #8), as well as a tutorial and any other documentation necessary to explain how to use it. The main page shouldn't duplicate the readme, and the readme should be shortened to account for having proper docs.
We should probably hold off on this until the software is more mature and less subject to change. | 1.0 | Website overhaul - Something shiny, with good typography and nice gradients and shadows and the like. Should include an html version of the man pages(s) (see #8), as well as a tutorial and any other documentation necessary to explain how to use it. The main page shouldn't duplicate the readme, and the readme should be shortened to account for having proper docs.
We should probably hold off on this until the software is more mature and less subject to change. | non_process | website overhaul something shiny with good typography and nice gradients and shadows and the like should include an html version of the man pages s see as well as a tutorial and any other documentation necessary to explain how to use it the main page shouldn t duplicate the readme and the readme should be shortened to account for having proper docs we should probably hold off on this until the software is more mature and less subject to change | 0 |
6,148 | 9,023,605,163 | IssuesEvent | 2019-02-07 07:57:46 | hashicorp/packer | https://api.github.com/repos/hashicorp/packer | closed | Vsphere-template post processor failing with "Artifact type mitchellh.vmware does not fit this requirement" | post-processor/vsphere-template question | Packer v1.3.4
MacOS Mojave
Packer Log output
https://gist.github.com/spstratis/b8abe1795bea6dd45b0d60cba06a8066
Packer Build Scrip
https://gist.github.com/spstratis/097a47cad65f258320ecae7808f75085
I'm trying to build a vmware-iso locally using vmware fusion and then have the vsphere post processeor export it to vSphere and then have the vsphere-template post-processor convert it into template. It's failing with the following error though.
```* Post-processor failed: The Packer vSphere Template post-processor can only take an artifact from the VMware-iso builder, built on ESXi (i.e. remote) or an artifact from the vSphere post-processor. Artifact type mitchellh.vmware does not fit this requirement```
The vsphere post processor works fine, the issue occurs after it tries to run the vsphere-template post-processor. I saw another issue that was closed for this from a few months ago. It looks like a fix was merged into the newest release (1.3.4) for it but I'm still running into it.
I also tried adding the `vmx_data property` for `"bios.hddorder": ""` but that also did not work. Any ideas?
Thanks!
| 1.0 | Vsphere-template post processor failing with "Artifact type mitchellh.vmware does not fit this requirement" - Packer v1.3.4
MacOS Mojave
Packer Log output
https://gist.github.com/spstratis/b8abe1795bea6dd45b0d60cba06a8066
Packer Build Scrip
https://gist.github.com/spstratis/097a47cad65f258320ecae7808f75085
I'm trying to build a vmware-iso locally using vmware fusion and then have the vsphere post processeor export it to vSphere and then have the vsphere-template post-processor convert it into template. It's failing with the following error though.
```* Post-processor failed: The Packer vSphere Template post-processor can only take an artifact from the VMware-iso builder, built on ESXi (i.e. remote) or an artifact from the vSphere post-processor. Artifact type mitchellh.vmware does not fit this requirement```
The vsphere post processor works fine, the issue occurs after it tries to run the vsphere-template post-processor. I saw another issue that was closed for this from a few months ago. It looks like a fix was merged into the newest release (1.3.4) for it but I'm still running into it.
I also tried adding the `vmx_data property` for `"bios.hddorder": ""` but that also did not work. Any ideas?
Thanks!
| process | vsphere template post processor failing with artifact type mitchellh vmware does not fit this requirement packer macos mojave packer log output packer build scrip i m trying to build a vmware iso locally using vmware fusion and then have the vsphere post processeor export it to vsphere and then have the vsphere template post processor convert it into template it s failing with the following error though post processor failed the packer vsphere template post processor can only take an artifact from the vmware iso builder built on esxi i e remote or an artifact from the vsphere post processor artifact type mitchellh vmware does not fit this requirement the vsphere post processor works fine the issue occurs after it tries to run the vsphere template post processor i saw another issue that was closed for this from a few months ago it looks like a fix was merged into the newest release for it but i m still running into it i also tried adding the vmx data property for bios hddorder but that also did not work any ideas thanks | 1 |
173,034 | 13,384,433,657 | IssuesEvent | 2020-09-02 12:00:42 | ZaneDubya/MedievaLandsPublic | https://api.github.com/repos/ZaneDubya/MedievaLandsPublic | closed | Twinion/MTS interaction: are weapons spawning in Twinion? (Jair) | Type-Testing | Jair reports that new weapons might not be spawning in Twinion, or at a much decreased rate. | 1.0 | Twinion/MTS interaction: are weapons spawning in Twinion? (Jair) - Jair reports that new weapons might not be spawning in Twinion, or at a much decreased rate. | non_process | twinion mts interaction are weapons spawning in twinion jair jair reports that new weapons might not be spawning in twinion or at a much decreased rate | 0 |
13,358 | 15,820,062,519 | IssuesEvent | 2021-04-05 18:24:50 | MicrosoftDocs/azure-devops-docs | https://api.github.com/repos/MicrosoftDocs/azure-devops-docs | closed | stageDependencies unavailable in AzDo 2019.1 | Pri1 devops-cicd-process/tech devops/prod doc-bug | Documentation for AzDo 2019 mentions stageDependencies, though those are not available unfortunately:
```
An error occurred while loading the YAML build pipeline. Unrecognized value: 'stageDependencies'
```
Which probably is expected since the feature was only added in sprint [168](https://docs.microsoft.com/en-us/azure/devops/release-notes/2020/sprint-168-update)
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: dd7e0bd3-1f7d-d7b6-cc72-5ef63c31b46a
* Version Independent ID: dae87abd-b73d-9120-bcdb-6097d4b40f2a
* Content: [Define variables - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/variables?view=azure-devops-2019&tabs=yaml%2Cbatch#set-a-multi-job-output-variable)
* Content Source: [docs/pipelines/process/variables.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/variables.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | 1.0 | stageDependencies unavailable in AzDo 2019.1 - Documentation for AzDo 2019 mentions stageDependencies, though those are not available unfortunately:
```
An error occurred while loading the YAML build pipeline. Unrecognized value: 'stageDependencies'
```
Which probably is expected since the feature was only added in sprint [168](https://docs.microsoft.com/en-us/azure/devops/release-notes/2020/sprint-168-update)
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: dd7e0bd3-1f7d-d7b6-cc72-5ef63c31b46a
* Version Independent ID: dae87abd-b73d-9120-bcdb-6097d4b40f2a
* Content: [Define variables - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/variables?view=azure-devops-2019&tabs=yaml%2Cbatch#set-a-multi-job-output-variable)
* Content Source: [docs/pipelines/process/variables.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/variables.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | process | stagedependencies unavailable in azdo documentation for azdo mentions stagedependencies though those are not available unfortunately an error occurred while loading the yaml build pipeline unrecognized value stagedependencies which probably is expected since the feature was only added in sprint document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id bcdb content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam | 1 |
15,387 | 19,571,629,782 | IssuesEvent | 2022-01-04 10:37:29 | prisma/prisma | https://api.github.com/repos/prisma/prisma | closed | Full-text search doesn't work with filter operators/ conditions – `OR` | bug/1-repro-available kind/bug process/candidate team/client topic: full-text search | ### Bug description
Executing the following query throws an error when using the `OR` operator with full-text search:
```typescript
const query = 'Prisma rocks'.split(' ').join('&')
const results = await prisma.book.findMany({
where: {
OR: {
title: {
search: query
},
content: {
search: query
}
}
},
})
```
The following error is thrown by Prisma Client:
```bash
{ query: 'Prisma & rocks' }
{
error: PrismaClientUnknownRequestError:
Invalid `prisma.book.findMany()` invocation:
Error occurred during query execution:
ConnectorError(ConnectorError { user_facing_error: None, kind: QueryError(Error { kind: Db, cause: Some(DbError { severity: "ERROR", parsed_severity: Some(Error), code: SqlState("42601"), message: "syntax error in tsquery: \"david copperfield\"", detail: None, hint: None, position: None, where_: None, schema: None, table: None, column: None, datatype: None, constraint: None, file: Some("tsquery.c"), line: Some(689), routine: Some("makepol") }) }) })
at cb (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/@prisma/client/runtime/index.js:38692:17)
at async handler (webpack-internal:///./pages/api/search.ts:40:29)
at async Object.apiResolver (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/api-utils.js:102:9)
at async DevServer.handleApiRequest (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:1064:9)
at async Object.fn (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:951:37)
at async Router.execute (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/router.js:222:32)
at async DevServer.run (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:1135:29)
at async DevServer.run (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/dev/next-dev-server.js:445:20)
at async DevServer.handleRequest (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:325:20) {
clientVersion: '3.7.0'
}
```
If filter operators are not supposed to work when paired with full-text search, we should provide some level of type-safety around it.
### How to reproduce
<!--
1. Go to '...'
2. Change '....'
3. Run '....'
4. See error
-->
Query
```typescript
const query = 'Prisma rocks'.split(' ').join(' & ')
const results = await prisma.book.findMany({
where: {
OR: {
title: {
search: query
},
content: {
search: query
}
}
},
})
```
### Expected behavior
No error if the query is executed, or provide type safety that filter operators can't be used when using FTS
### Prisma information
<!-- Do not include your database credentials when sharing your Prisma schema! -->
Schema:
```prisma
generator client {
provider = "prisma-client-js"
previewFeatures = ["fullTextSearch"]
}
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
model Book {
id Int @id @default(autoincrement())
title String
content String
url String
cover String
authors String[]
}
```
Migration
### Environment & setup
- OS: Mac OS
- Database: PostgreSQL
- Node.js version: 16.13
### Prisma Version
```
prisma : 3.7.0
@prisma/client : 3.7.0
Current platform : darwin
Query Engine (Node-API) : libquery-engine 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/libquery_engine-darwin.dylib.node)
Migration Engine : migration-engine-cli 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/migration-engine-darwin)
Introspection Engine : introspection-core 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/introspection-engine-darwin)
Format Binary : prisma-fmt 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/prisma-fmt-darwin)
Default Engines Hash : 8746e055198f517658c08a0c426c7eec87f5a85f
Studio : 0.445.0
Preview Features : fullTextSearch
```
| 1.0 | Full-text search doesn't work with filter operators/ conditions – `OR` - ### Bug description
Executing the following query throws an error when using the `OR` operator with full-text search:
```typescript
const query = 'Prisma rocks'.split(' ').join('&')
const results = await prisma.book.findMany({
where: {
OR: {
title: {
search: query
},
content: {
search: query
}
}
},
})
```
The following error is thrown by Prisma Client:
```bash
{ query: 'Prisma & rocks' }
{
error: PrismaClientUnknownRequestError:
Invalid `prisma.book.findMany()` invocation:
Error occurred during query execution:
ConnectorError(ConnectorError { user_facing_error: None, kind: QueryError(Error { kind: Db, cause: Some(DbError { severity: "ERROR", parsed_severity: Some(Error), code: SqlState("42601"), message: "syntax error in tsquery: \"david copperfield\"", detail: None, hint: None, position: None, where_: None, schema: None, table: None, column: None, datatype: None, constraint: None, file: Some("tsquery.c"), line: Some(689), routine: Some("makepol") }) }) })
at cb (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/@prisma/client/runtime/index.js:38692:17)
at async handler (webpack-internal:///./pages/api/search.ts:40:29)
at async Object.apiResolver (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/api-utils.js:102:9)
at async DevServer.handleApiRequest (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:1064:9)
at async Object.fn (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:951:37)
at async Router.execute (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/router.js:222:32)
at async DevServer.run (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:1135:29)
at async DevServer.run (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/dev/next-dev-server.js:445:20)
at async DevServer.handleRequest (/Users/ruheni/Documents/repos/work/prisma/projects/prisma-fulltextsearch/node_modules/next/dist/server/next-server.js:325:20) {
clientVersion: '3.7.0'
}
```
If filter operators are not supposed to work when paired with full-text search, we should provide some level of type-safety around it.
### How to reproduce
<!--
1. Go to '...'
2. Change '....'
3. Run '....'
4. See error
-->
Query
```typescript
const query = 'Prisma rocks'.split(' ').join(' & ')
const results = await prisma.book.findMany({
where: {
OR: {
title: {
search: query
},
content: {
search: query
}
}
},
})
```
### Expected behavior
No error if the query is executed, or provide type safety that filter operators can't be used when using FTS
### Prisma information
<!-- Do not include your database credentials when sharing your Prisma schema! -->
Schema:
```prisma
generator client {
provider = "prisma-client-js"
previewFeatures = ["fullTextSearch"]
}
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
model Book {
id Int @id @default(autoincrement())
title String
content String
url String
cover String
authors String[]
}
```
Migration
### Environment & setup
- OS: Mac OS
- Database: PostgreSQL
- Node.js version: 16.13
### Prisma Version
```
prisma : 3.7.0
@prisma/client : 3.7.0
Current platform : darwin
Query Engine (Node-API) : libquery-engine 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/libquery_engine-darwin.dylib.node)
Migration Engine : migration-engine-cli 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/migration-engine-darwin)
Introspection Engine : introspection-core 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/introspection-engine-darwin)
Format Binary : prisma-fmt 8746e055198f517658c08a0c426c7eec87f5a85f (at node_modules/@prisma/engines/prisma-fmt-darwin)
Default Engines Hash : 8746e055198f517658c08a0c426c7eec87f5a85f
Studio : 0.445.0
Preview Features : fullTextSearch
```
| process | full text search doesn t work with filter operators conditions – or bug description executing the following query throws an error when using the or operator with full text search typescript const query prisma rocks split join const results await prisma book findmany where or title search query content search query the following error is thrown by prisma client bash query prisma rocks error prismaclientunknownrequesterror invalid prisma book findmany invocation error occurred during query execution connectorerror connectorerror user facing error none kind queryerror error kind db cause some dberror severity error parsed severity some error code sqlstate message syntax error in tsquery david copperfield detail none hint none position none where none schema none table none column none datatype none constraint none file some tsquery c line some routine some makepol at cb users ruheni documents repos work prisma projects prisma fulltextsearch node modules prisma client runtime index js at async handler webpack internal pages api search ts at async object apiresolver users ruheni documents repos work prisma projects prisma fulltextsearch node modules next dist server api utils js at async devserver handleapirequest users ruheni documents repos work prisma projects prisma fulltextsearch node modules next dist server next server js at async object fn users ruheni documents repos work prisma projects prisma fulltextsearch node modules next dist server next server js at async router execute users ruheni documents repos work prisma projects prisma fulltextsearch node modules next dist server router js at async devserver run users ruheni documents repos work prisma projects prisma fulltextsearch node modules next dist server next server js at async devserver run users ruheni documents repos work prisma projects prisma fulltextsearch node modules next dist server dev next dev server js at async devserver handlerequest users ruheni documents repos work prisma projects prisma fulltextsearch node modules next dist server next server js clientversion if filter operators are not supposed to work when paired with full text search we should provide some level of type safety around it how to reproduce go to change run see error query typescript const query prisma rocks split join const results await prisma book findmany where or title search query content search query expected behavior no error if the query is executed or provide type safety that filter operators can t be used when using fts prisma information schema prisma generator client provider prisma client js previewfeatures datasource db provider postgresql url env database url model book id int id default autoincrement title string content string url string cover string authors string migration environment setup os mac os database postgresql node js version prisma version prisma prisma client current platform darwin query engine node api libquery engine at node modules prisma engines libquery engine darwin dylib node migration engine migration engine cli at node modules prisma engines migration engine darwin introspection engine introspection core at node modules prisma engines introspection engine darwin format binary prisma fmt at node modules prisma engines prisma fmt darwin default engines hash studio preview features fulltextsearch | 1 |
306,735 | 23,170,792,243 | IssuesEvent | 2022-07-30 17:26:00 | Nicolas-Hermet/ProGGG | https://api.github.com/repos/Nicolas-Hermet/ProGGG | opened | The app should be a PWA | documentation enhancement Workflow Refactoring | ## Description
The app has to respect PWA standards.
## Objectivs
The github page has to be pushed on stores and behave like an app on any device.
## To Do:
- [ ] Some research might have to be done and discuss here before doing some work.
- [ ] It might impact other issues: change them or discuss these changes if needed
- [ ] Turn the app into a PWA. | 1.0 | The app should be a PWA - ## Description
The app has to respect PWA standards.
## Objectivs
The github page has to be pushed on stores and behave like an app on any device.
## To Do:
- [ ] Some research might have to be done and discuss here before doing some work.
- [ ] It might impact other issues: change them or discuss these changes if needed
- [ ] Turn the app into a PWA. | non_process | the app should be a pwa description the app has to respect pwa standards objectivs the github page has to be pushed on stores and behave like an app on any device to do some research might have to be done and discuss here before doing some work it might impact other issues change them or discuss these changes if needed turn the app into a pwa | 0 |
272,538 | 23,679,861,671 | IssuesEvent | 2022-08-28 16:33:00 | raheemadamboev/kichkina-shahzoda | https://api.github.com/repos/raheemadamboev/kichkina-shahzoda | closed | Test `About` | test | - [x] load proper version of screen according to orientation
- [x] proper ui
- [x] back button
- [x] proper version
- [x] gravity button | 1.0 | Test `About` - - [x] load proper version of screen according to orientation
- [x] proper ui
- [x] back button
- [x] proper version
- [x] gravity button | non_process | test about load proper version of screen according to orientation proper ui back button proper version gravity button | 0 |
12,285 | 14,814,961,002 | IssuesEvent | 2021-01-14 06:15:11 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | [Modeler] Missing variables in modeler expression when used in default expression | Bug Modeller Processing | Another modeler Bug or improvement.
The expression builder dialogs display the available variables when initiated from a processing window (IE an expression parameter in a tool in the model or when setting an expression parameter in the model dialog itself).
But the variables of the model are not available when using the expression builder to set a default value for an expression, even though those variables are properly evaluated at runtime.
This can be misleading as it would appear that the default expression does not have access to those elements or force the user to have a separate way to use the variables that are available in the model.
The fix would be to simply provide the model expression context scope to this expression builder for the expression parameter.
| 1.0 | [Modeler] Missing variables in modeler expression when used in default expression - Another modeler Bug or improvement.
The expression builder dialogs display the available variables when initiated from a processing window (IE an expression parameter in a tool in the model or when setting an expression parameter in the model dialog itself).
But the variables of the model are not available when using the expression builder to set a default value for an expression, even though those variables are properly evaluated at runtime.
This can be misleading as it would appear that the default expression does not have access to those elements or force the user to have a separate way to use the variables that are available in the model.
The fix would be to simply provide the model expression context scope to this expression builder for the expression parameter.
| process | missing variables in modeler expression when used in default expression another modeler bug or improvement the expression builder dialogs display the available variables when initiated from a processing window ie an expression parameter in a tool in the model or when setting an expression parameter in the model dialog itself but the variables of the model are not available when using the expression builder to set a default value for an expression even though those variables are properly evaluated at runtime this can be misleading as it would appear that the default expression does not have access to those elements or force the user to have a separate way to use the variables that are available in the model the fix would be to simply provide the model expression context scope to this expression builder for the expression parameter | 1 |
66,235 | 20,086,240,408 | IssuesEvent | 2022-02-05 02:14:31 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Clicking on the jump to unreads button gives a "Failed to load timeline position" and dumps me in a blank timeline | T-Defect A-Timeline A-Read-Marker | ### Steps to reproduce
1. Visit a room
1. Click the "Jump to first unread message." arrow

1. `Failed to load timeline position` error modal is thrown up
> Failed to load timeline position
>
> Tried to load a specific point in this room's timeline, but was unable to find it.

1. After exiting out of the modal, the timeline is blank and no way to scroll
1. Have to switch rooms to get the timeline again
---
Reproduction/rage-shake logs: https://github.com/matrix-org/element-web-rageshakes/issues/10392
I was in the `!ltpmbfSHehPTwnLDNS:jki.re` room and clicked on the "Jump to first unread message." arrow and got the `Failed to load timeline position` error modal. With the following error in the devtools console.
Relevant code for the error: [`src/components/structures/TimelinePanel.tsx#L1219-L1244`](https://github.com/matrix-org/matrix-react-sdk/blob/7f3f18604417d9567c109201673e60a5f24a30ce/src/components/structures/TimelinePanel.tsx#L1219-L1244)
```
Error loading timeline panel at $iR-8DJRGNhmIpDFDHmRy70gUTurU77eCba-Jt5YIXbo: Error: getEventTimeline result didn't include requested event
```
I'm not sure what room the `$iR-8DJRGNhmIpDFDHmRy70gUTurU77eCba-Jt5YIXbo` event pertains to. Is there a way to look it up and tell which room it actually belongs to?
### Outcome
#### What did you expect?
Jump to first unread button jumps me to whatever the unreads are for that room. Or not be shown if it was for another room and this room doesn't have any unread.
#### What happened instead?
Error modal is shown for a situation that probably shouldn't occur in the first place.
### Operating system
Windows 10
### Browser information
Chrome 97.0.4692.99
### URL for webapp
https://develop.element.io/
### Application version
Element version: 64242a004eb7-react-78e78292cb62-js-b07457726bf5 Olm version: 3.2.8
### Homeserver
matrix.org
### Will you send logs?
Yes | 1.0 | Clicking on the jump to unreads button gives a "Failed to load timeline position" and dumps me in a blank timeline - ### Steps to reproduce
1. Visit a room
1. Click the "Jump to first unread message." arrow

1. `Failed to load timeline position` error modal is thrown up
> Failed to load timeline position
>
> Tried to load a specific point in this room's timeline, but was unable to find it.

1. After exiting out of the modal, the timeline is blank and no way to scroll
1. Have to switch rooms to get the timeline again
---
Reproduction/rage-shake logs: https://github.com/matrix-org/element-web-rageshakes/issues/10392
I was in the `!ltpmbfSHehPTwnLDNS:jki.re` room and clicked on the "Jump to first unread message." arrow and got the `Failed to load timeline position` error modal. With the following error in the devtools console.
Relevant code for the error: [`src/components/structures/TimelinePanel.tsx#L1219-L1244`](https://github.com/matrix-org/matrix-react-sdk/blob/7f3f18604417d9567c109201673e60a5f24a30ce/src/components/structures/TimelinePanel.tsx#L1219-L1244)
```
Error loading timeline panel at $iR-8DJRGNhmIpDFDHmRy70gUTurU77eCba-Jt5YIXbo: Error: getEventTimeline result didn't include requested event
```
I'm not sure what room the `$iR-8DJRGNhmIpDFDHmRy70gUTurU77eCba-Jt5YIXbo` event pertains to. Is there a way to look it up and tell which room it actually belongs to?
### Outcome
#### What did you expect?
Jump to first unread button jumps me to whatever the unreads are for that room. Or not be shown if it was for another room and this room doesn't have any unread.
#### What happened instead?
Error modal is shown for a situation that probably shouldn't occur in the first place.
### Operating system
Windows 10
### Browser information
Chrome 97.0.4692.99
### URL for webapp
https://develop.element.io/
### Application version
Element version: 64242a004eb7-react-78e78292cb62-js-b07457726bf5 Olm version: 3.2.8
### Homeserver
matrix.org
### Will you send logs?
Yes | non_process | clicking on the jump to unreads button gives a failed to load timeline position and dumps me in a blank timeline steps to reproduce visit a room click the jump to first unread message arrow failed to load timeline position error modal is thrown up failed to load timeline position tried to load a specific point in this room s timeline but was unable to find it after exiting out of the modal the timeline is blank and no way to scroll have to switch rooms to get the timeline again reproduction rage shake logs i was in the ltpmbfshehptwnldns jki re room and clicked on the jump to first unread message arrow and got the failed to load timeline position error modal with the following error in the devtools console relevant code for the error error loading timeline panel at ir error geteventtimeline result didn t include requested event i m not sure what room the ir event pertains to is there a way to look it up and tell which room it actually belongs to outcome what did you expect jump to first unread button jumps me to whatever the unreads are for that room or not be shown if it was for another room and this room doesn t have any unread what happened instead error modal is shown for a situation that probably shouldn t occur in the first place operating system windows browser information chrome url for webapp application version element version react js olm version homeserver matrix org will you send logs yes | 0 |
681,461 | 23,311,956,585 | IssuesEvent | 2022-08-08 09:03:44 | woocommerce/woocommerce-ios | https://api.github.com/repos/woocommerce/woocommerce-ios | closed | NSGenericException: Unable to activate constraint with anchors <NSLayoutYAxisAnchor:0x281aa2900 "_TtC11WooCommerceP33... | type: crash priority: low | Sentry Issue: [WOOCOMMERCE-IOS-1CWS](https://sentry.io/organizations/a8c/issues/2274916143/?referrer=github_integration)
```
NSGenericException: Unable to activate constraint with anchors <NSLayoutYAxisAnchor:0x281aa2900 "_TtC11WooCommerceP33_FFA5D4594163A533C445FD7D6757038419NoticeContainerView:0x103aff040.bottom"> and <NSLayoutYAxisAnchor:0x281af01c0 "UITabBar:0x103a3b8f0.top"> because they have no common ancestor. Does the constraint or its anchors reference items in different view hierarchies? That's illegal.
File "DefaultNoticePresenter.swift", line 99, in DefaultNoticePresenter.presentNoticeInForeground
File "DefaultNoticePresenter.swift", line 53, in DefaultNoticePresenter.present
File "DefaultNoticePresenter.swift", line 47, in DefaultNoticePresenter.presentNextNoticeIfPossible
File "DefaultNoticePresenter.swift", line 33, in DefaultNoticePresenter.enqueue
File "<compiler-generated>", in DefaultNoticePresenter
...
(28 additional frame(s) were not displayed)
``` | 1.0 | NSGenericException: Unable to activate constraint with anchors <NSLayoutYAxisAnchor:0x281aa2900 "_TtC11WooCommerceP33... - Sentry Issue: [WOOCOMMERCE-IOS-1CWS](https://sentry.io/organizations/a8c/issues/2274916143/?referrer=github_integration)
```
NSGenericException: Unable to activate constraint with anchors <NSLayoutYAxisAnchor:0x281aa2900 "_TtC11WooCommerceP33_FFA5D4594163A533C445FD7D6757038419NoticeContainerView:0x103aff040.bottom"> and <NSLayoutYAxisAnchor:0x281af01c0 "UITabBar:0x103a3b8f0.top"> because they have no common ancestor. Does the constraint or its anchors reference items in different view hierarchies? That's illegal.
File "DefaultNoticePresenter.swift", line 99, in DefaultNoticePresenter.presentNoticeInForeground
File "DefaultNoticePresenter.swift", line 53, in DefaultNoticePresenter.present
File "DefaultNoticePresenter.swift", line 47, in DefaultNoticePresenter.presentNextNoticeIfPossible
File "DefaultNoticePresenter.swift", line 33, in DefaultNoticePresenter.enqueue
File "<compiler-generated>", in DefaultNoticePresenter
...
(28 additional frame(s) were not displayed)
``` | non_process | nsgenericexception unable to activate constraint with anchors nslayoutyaxisanchor sentry issue nsgenericexception unable to activate constraint with anchors and because they have no common ancestor does the constraint or its anchors reference items in different view hierarchies that s illegal file defaultnoticepresenter swift line in defaultnoticepresenter presentnoticeinforeground file defaultnoticepresenter swift line in defaultnoticepresenter present file defaultnoticepresenter swift line in defaultnoticepresenter presentnextnoticeifpossible file defaultnoticepresenter swift line in defaultnoticepresenter enqueue file in defaultnoticepresenter additional frame s were not displayed | 0 |
9,458 | 12,438,784,864 | IssuesEvent | 2020-05-26 09:00:44 | DevExpress/testcafe-hammerhead | https://api.github.com/repos/DevExpress/testcafe-hammerhead | closed | IE breaks with TypeError: Invalid calling object in _processChildren() and getFirstChild() | AREA: client BROWSER: IE11 FREQUENCY: level 2 SYSTEM: client side processing SYSTEM: shadow UI TYPE: bug | ### What is your Scenario?
Run test on IE on Windows 10.
### What is the Current behavior?
IE stops script with 'Invalid Calling Object' error.
### What is your public web site URL?
https://www.johnmuirhealth.com/doctor
<details>
<summary>Your complete app code (or attach your test files):</summary>
<!-- Paste your app code here: -->
```js
import { Selector } from "testcafe";
fixture `Simple Test`;
test('simple test', async t => {
await t.navigateTo('https://www.johnmuirhealth.com/doctor');
await t.expect( Selector('.provider-name').exists ).ok();
});
```
</details>
<details>
<summary>Screenshots:</summary>
<!-- If applicable, add screenshots to help explain the issue. -->
```
testcafe ie ./tests/simple.test.js
Using locally installed version of TestCafe.
Running tests in:
- Internet Explorer 11.0 / Windows 10
Simple Test
× simple test
1) A JavaScript error occurred on "https://www.johnmuirhealth.com/doctor".
Repeat test actions in the browser and check the console for errors.
If you see this error, it means that the tested website caused it. You can fix it or disable tracking JavaScript errors in TestCafe. To do the latter, enable the
"--skip-js-errors" option.
If this error does not occur, please write a new issue at:
"https://github.com/DevExpress/testcafe/issues/new?template=bug-report.md".
JavaScript error details:
TypeError: Invalid calling object
at r.getFirstChild (http://172.29.37.13:54112/hammerhead.js:12:28542)
at getter (http://172.29.37.13:54112/hammerhead.js:12:10460)
at sc.firstChild (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:127:24)
at Bc.firstChild.get (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:133:226)
at pe (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:190:405)
at Q (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:192:98)
at ve (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:196:253)
at S (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:196:931)
at Anonymous function (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:214:527)
at Global code (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:11:2)
Browser: Internet Explorer 11.0 / Windows 10
1/1 failed (3s)
```
</details>
### Steps to Reproduce:
<!-- Describe what we should do to reproduce the behavior you encountered. -->
1. Put the code above into simple.test.js
2. Execute this command: testcafe ie simple.test.js
3. See the error (see detailed error output above): TypeError: Invalid calling object
Making these changes to hammerhead.js makes it work:
```js
_proto._processChildren = function _processChildren(el) {
if (!el.querySelectorAll) return;
//var children = (0, _querySelector.getNativeQuerySelectorAll)(el).call(el, '*');
var children = el.querySelectorAll('*');
//var length = _nativeMethods.default.nodeListLengthGetter.call(children);
var length = children.length;
for (var i = 0; i < length; i++) {
this._processElement(children[i]);
}
};
```
and
```js
_proto.getFirstChild = function getFirstChild(el) {
//var length = _nativeMethods.default.nodeListLengthGetter.call(el.childNodes);
var length = el.childNodes.length;
var filteredNodes = this._filterNodeList(el.childNodes, length);
return filteredNodes[0] || null;
};
```
### Your Environment details:
* node.js version: 10.16.3
* browser name and version: IE 11
* platform and version: Windows 10
* other: TestCafe 1.8.1
| 1.0 | IE breaks with TypeError: Invalid calling object in _processChildren() and getFirstChild() - ### What is your Scenario?
Run test on IE on Windows 10.
### What is the Current behavior?
IE stops script with 'Invalid Calling Object' error.
### What is your public web site URL?
https://www.johnmuirhealth.com/doctor
<details>
<summary>Your complete app code (or attach your test files):</summary>
<!-- Paste your app code here: -->
```js
import { Selector } from "testcafe";
fixture `Simple Test`;
test('simple test', async t => {
await t.navigateTo('https://www.johnmuirhealth.com/doctor');
await t.expect( Selector('.provider-name').exists ).ok();
});
```
</details>
<details>
<summary>Screenshots:</summary>
<!-- If applicable, add screenshots to help explain the issue. -->
```
testcafe ie ./tests/simple.test.js
Using locally installed version of TestCafe.
Running tests in:
- Internet Explorer 11.0 / Windows 10
Simple Test
× simple test
1) A JavaScript error occurred on "https://www.johnmuirhealth.com/doctor".
Repeat test actions in the browser and check the console for errors.
If you see this error, it means that the tested website caused it. You can fix it or disable tracking JavaScript errors in TestCafe. To do the latter, enable the
"--skip-js-errors" option.
If this error does not occur, please write a new issue at:
"https://github.com/DevExpress/testcafe/issues/new?template=bug-report.md".
JavaScript error details:
TypeError: Invalid calling object
at r.getFirstChild (http://172.29.37.13:54112/hammerhead.js:12:28542)
at getter (http://172.29.37.13:54112/hammerhead.js:12:10460)
at sc.firstChild (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:127:24)
at Bc.firstChild.get (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:133:226)
at pe (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:190:405)
at Q (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:192:98)
at ve (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:196:253)
at S (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:196:931)
at Anonymous function (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:214:527)
at Global code (https://www.johnmuirhealth.com/webcommon/lib/webcomponentsjs-2.1.3/webcomponents-bundle.js:11:2)
Browser: Internet Explorer 11.0 / Windows 10
1/1 failed (3s)
```
</details>
### Steps to Reproduce:
<!-- Describe what we should do to reproduce the behavior you encountered. -->
1. Put the code above into simple.test.js
2. Execute this command: testcafe ie simple.test.js
3. See the error (see detailed error output above): TypeError: Invalid calling object
Making these changes to hammerhead.js makes it work:
```js
_proto._processChildren = function _processChildren(el) {
if (!el.querySelectorAll) return;
//var children = (0, _querySelector.getNativeQuerySelectorAll)(el).call(el, '*');
var children = el.querySelectorAll('*');
//var length = _nativeMethods.default.nodeListLengthGetter.call(children);
var length = children.length;
for (var i = 0; i < length; i++) {
this._processElement(children[i]);
}
};
```
and
```js
_proto.getFirstChild = function getFirstChild(el) {
//var length = _nativeMethods.default.nodeListLengthGetter.call(el.childNodes);
var length = el.childNodes.length;
var filteredNodes = this._filterNodeList(el.childNodes, length);
return filteredNodes[0] || null;
};
```
### Your Environment details:
* node.js version: 10.16.3
* browser name and version: IE 11
* platform and version: Windows 10
* other: TestCafe 1.8.1
| process | ie breaks with typeerror invalid calling object in processchildren and getfirstchild what is your scenario run test on ie on windows what is the current behavior ie stops script with invalid calling object error what is your public web site url your complete app code or attach your test files js import selector from testcafe fixture simple test test simple test async t await t navigateto await t expect selector provider name exists ok screenshots testcafe ie tests simple test js using locally installed version of testcafe running tests in internet explorer windows simple test × simple test a javascript error occurred on repeat test actions in the browser and check the console for errors if you see this error it means that the tested website caused it you can fix it or disable tracking javascript errors in testcafe to do the latter enable the skip js errors option if this error does not occur please write a new issue at javascript error details typeerror invalid calling object at r getfirstchild at getter at sc firstchild at bc firstchild get at pe at q at ve at s at anonymous function at global code browser internet explorer windows failed steps to reproduce put the code above into simple test js execute this command testcafe ie simple test js see the error see detailed error output above typeerror invalid calling object making these changes to hammerhead js makes it work js proto processchildren function processchildren el if el queryselectorall return var children queryselector getnativequeryselectorall el call el var children el queryselectorall var length nativemethods default nodelistlengthgetter call children var length children length for var i i length i this processelement children and js proto getfirstchild function getfirstchild el var length nativemethods default nodelistlengthgetter call el childnodes var length el childnodes length var filterednodes this filternodelist el childnodes length return filterednodes null your environment details node js version browser name and version ie platform and version windows other testcafe | 1 |
14,594 | 17,703,558,656 | IssuesEvent | 2021-08-25 03:16:42 | tdwg/dwc | https://api.github.com/repos/tdwg/dwc | closed | Change term - organismQuantity | Term - change Class - Occurrence non-normative Process - complete | ## Change term
* Submitter: John Wieczorek
* Justification (why is this change necessary?): Clarity
* Proponents (who needs this change): Everyone
Current Term definition: https://dwc.tdwg.org/list/#dwc_organismQuantity
Proposed new attributes of the term:
* Term name (in lowerCamelCase): organismQuantity
* Organized in Class (e.g. Location, Taxon): Occurrence
* Definition of the term: A number or enumeration value for the quantity of organisms.
* Usage comments (recommendations regarding content, etc.): An organismQuantity must have a corresponding organismQuantityType.
* Examples: `27` (organismQuantity) with `individuals` (organismQuantityType). `12.5` (organismQuantity) with `%biomass` (organismQuantityType). `r` (organismQuantity) with `BraunBlanquetScale` (organismQuantityType). **`many` (organismQuantity) with `individuals` (organismQuantityType).**
* Refines (identifier of the broader term this term refines, if applicable): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/organismQuantity-2017-10-06
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): not in ABCD
This proposal is to add the example of "many" and an organismQuantity to complement the proposed usage comment under individualCount (see Issue #285).
| 1.0 | Change term - organismQuantity - ## Change term
* Submitter: John Wieczorek
* Justification (why is this change necessary?): Clarity
* Proponents (who needs this change): Everyone
Current Term definition: https://dwc.tdwg.org/list/#dwc_organismQuantity
Proposed new attributes of the term:
* Term name (in lowerCamelCase): organismQuantity
* Organized in Class (e.g. Location, Taxon): Occurrence
* Definition of the term: A number or enumeration value for the quantity of organisms.
* Usage comments (recommendations regarding content, etc.): An organismQuantity must have a corresponding organismQuantityType.
* Examples: `27` (organismQuantity) with `individuals` (organismQuantityType). `12.5` (organismQuantity) with `%biomass` (organismQuantityType). `r` (organismQuantity) with `BraunBlanquetScale` (organismQuantityType). **`many` (organismQuantity) with `individuals` (organismQuantityType).**
* Refines (identifier of the broader term this term refines, if applicable): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/organismQuantity-2017-10-06
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): not in ABCD
This proposal is to add the example of "many" and an organismQuantity to complement the proposed usage comment under individualCount (see Issue #285).
| process | change term organismquantity change term submitter john wieczorek justification why is this change necessary clarity proponents who needs this change everyone current term definition proposed new attributes of the term term name in lowercamelcase organismquantity organized in class e g location taxon occurrence definition of the term a number or enumeration value for the quantity of organisms usage comments recommendations regarding content etc an organismquantity must have a corresponding organismquantitytype examples organismquantity with individuals organismquantitytype organismquantity with biomass organismquantitytype r organismquantity with braunblanquetscale organismquantitytype many organismquantity with individuals organismquantitytype refines identifier of the broader term this term refines if applicable none replaces identifier of the existing term that would be deprecated and replaced by this term if applicable abcd xpath of the equivalent term in abcd or efg if applicable not in abcd this proposal is to add the example of many and an organismquantity to complement the proposed usage comment under individualcount see issue | 1 |
784,656 | 27,580,870,832 | IssuesEvent | 2023-03-08 16:07:56 | kubernetes-sigs/gateway-api | https://api.github.com/repos/kubernetes-sigs/gateway-api | closed | Configure Timeouts in Gateway API | kind/feature priority/important-longterm | Here are some timeouts that would be benefit users
HTTP
* Request Timeout
* Envoy - https://www.envoyproxy.io/docs/envoy/latest/api-v3/config/route/v3/route_components.proto#envoy-v3-api-field-config-route-v3-routeaction-timeout
* NGINX - https://nginx.org/en/docs/http/ngx_http_proxy_module.html#proxy_send_timeout
* Idle Timeout
* NGINX -
* Envoy - https://www.envoyproxy.io/docs/envoy/latest/api-v3/config/core/v3/protocol.proto#envoy-v3-api-field-config-core-v3-httpprotocoloptions-idle-timeout
TCP
* Connect Timeout
* NGINX - https://nginx.org/en/docs/http/ngx_http_proxy_module.html#proxy_connect_timeout
* Envoy - https://www.envoyproxy.io/docs/envoy/latest/api-v3/config/cluster/v3/cluster.proto#envoy-v3-api-field-config-cluster-v3-cluster-connect-timeout
* Keep Alive
* Upstream
* NGINX - https://nginx.org/en/docs/http/ngx_http_upstream_module.html#keepalive_requests
* Envoy - https://www.envoyproxy.io/docs/envoy/latest/api-v3/config/core/v3/address.proto#config-core-v3-tcpkeepalive
* Downstream
* NGINX - TODO
* Envoy - https://www.envoyproxy.io/docs/envoy/latest/api-v3/config/core/v3/socket_option.proto#config-core-v3-socketoption
Relates to https://github.com/kubernetes-sigs/gateway-api/discussions/1715 | 1.0 | Configure Timeouts in Gateway API - Here are some timeouts that would be benefit users
HTTP
* Request Timeout
* Envoy - https://www.envoyproxy.io/docs/envoy/latest/api-v3/config/route/v3/route_components.proto#envoy-v3-api-field-config-route-v3-routeaction-timeout
* NGINX - https://nginx.org/en/docs/http/ngx_http_proxy_module.html#proxy_send_timeout
* Idle Timeout
* NGINX -
* Envoy - https://www.envoyproxy.io/docs/envoy/latest/api-v3/config/core/v3/protocol.proto#envoy-v3-api-field-config-core-v3-httpprotocoloptions-idle-timeout
TCP
* Connect Timeout
* NGINX - https://nginx.org/en/docs/http/ngx_http_proxy_module.html#proxy_connect_timeout
* Envoy - https://www.envoyproxy.io/docs/envoy/latest/api-v3/config/cluster/v3/cluster.proto#envoy-v3-api-field-config-cluster-v3-cluster-connect-timeout
* Keep Alive
* Upstream
* NGINX - https://nginx.org/en/docs/http/ngx_http_upstream_module.html#keepalive_requests
* Envoy - https://www.envoyproxy.io/docs/envoy/latest/api-v3/config/core/v3/address.proto#config-core-v3-tcpkeepalive
* Downstream
* NGINX - TODO
* Envoy - https://www.envoyproxy.io/docs/envoy/latest/api-v3/config/core/v3/socket_option.proto#config-core-v3-socketoption
Relates to https://github.com/kubernetes-sigs/gateway-api/discussions/1715 | non_process | configure timeouts in gateway api here are some timeouts that would be benefit users http request timeout envoy nginx idle timeout nginx envoy tcp connect timeout nginx envoy keep alive upstream nginx envoy downstream nginx todo envoy relates to | 0 |
19,349 | 25,481,239,972 | IssuesEvent | 2022-11-25 21:26:37 | MPMG-DCC-UFMG/C01 | https://api.github.com/repos/MPMG-DCC-UFMG/C01 | closed | Remover o componente Scrapy-Playwright | [2] Alta Prioridade [0] Desenvolvimento [1] Aprimoramento [3] Processamento Dinâmico | ## Comportamento Esperado
O sistema deve utilizar a biblioteca Playwright diretamente para executar as coletas dinâmicas.
## Comportamento Atual
O sistema utiliza a biblioteca Scrapy-Playwright para fazer a integração entre o coletor e o Playwright. Implementamos dessa forma para garantir uma boa integração das ferramentas e evitar "reinventar a roda". Com o tempo, porém, encontramos algumas instabilidades e problemas no coletor que parecem ser causadas por causa do Scrapy-Playwright, e além disso ao usar essa ferramenta perdemos parte do controle do processo. | 1.0 | Remover o componente Scrapy-Playwright - ## Comportamento Esperado
O sistema deve utilizar a biblioteca Playwright diretamente para executar as coletas dinâmicas.
## Comportamento Atual
O sistema utiliza a biblioteca Scrapy-Playwright para fazer a integração entre o coletor e o Playwright. Implementamos dessa forma para garantir uma boa integração das ferramentas e evitar "reinventar a roda". Com o tempo, porém, encontramos algumas instabilidades e problemas no coletor que parecem ser causadas por causa do Scrapy-Playwright, e além disso ao usar essa ferramenta perdemos parte do controle do processo. | process | remover o componente scrapy playwright comportamento esperado o sistema deve utilizar a biblioteca playwright diretamente para executar as coletas dinâmicas comportamento atual o sistema utiliza a biblioteca scrapy playwright para fazer a integração entre o coletor e o playwright implementamos dessa forma para garantir uma boa integração das ferramentas e evitar reinventar a roda com o tempo porém encontramos algumas instabilidades e problemas no coletor que parecem ser causadas por causa do scrapy playwright e além disso ao usar essa ferramenta perdemos parte do controle do processo | 1 |
159,112 | 20,036,644,461 | IssuesEvent | 2022-02-02 12:38:13 | kapseliboi/dapp | https://api.github.com/repos/kapseliboi/dapp | opened | WS-2019-0075 (Low) detected in web3-0.20.7.tgz, web3-1.2.11.tgz | security vulnerability | ## WS-2019-0075 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>web3-0.20.7.tgz</b>, <b>web3-1.2.11.tgz</b></p></summary>
<p>
<details><summary><b>web3-0.20.7.tgz</b></p></summary>
<p>Ethereum JavaScript API, middleware to talk to a ethereum node over RPC</p>
<p>Library home page: <a href="https://registry.npmjs.org/web3/-/web3-0.20.7.tgz">https://registry.npmjs.org/web3/-/web3-0.20.7.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@toruslabs/torus-embed/node_modules/web3/package.json</p>
<p>
Dependency Hierarchy:
- torus-embed-1.8.2.tgz (Root Library)
- :x: **web3-0.20.7.tgz** (Vulnerable Library)
</details>
<details><summary><b>web3-1.2.11.tgz</b></p></summary>
<p>Ethereum JavaScript API</p>
<p>Library home page: <a href="https://registry.npmjs.org/web3/-/web3-1.2.11.tgz">https://registry.npmjs.org/web3/-/web3-1.2.11.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/web3/package.json</p>
<p>
Dependency Hierarchy:
- :x: **web3-1.2.11.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/dapp/commit/79de7acd382466c6348d970d41ce91b47fc3366d">79de7acd382466c6348d970d41ce91b47fc3366d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of web3 are vulnerable to Insecure Credential Storage
<p>Publish Date: 2019-05-15
<p>URL: <a href=https://github.com/ethereum/web3.js/issues/2739>WS-2019-0075</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/WS-2019-0075">https://nvd.nist.gov/vuln/detail/WS-2019-0075</a></p>
<p>Release Date: 2019-05-15</p>
<p>Fix Resolution: 1.5.3-rc.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0075 (Low) detected in web3-0.20.7.tgz, web3-1.2.11.tgz - ## WS-2019-0075 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>web3-0.20.7.tgz</b>, <b>web3-1.2.11.tgz</b></p></summary>
<p>
<details><summary><b>web3-0.20.7.tgz</b></p></summary>
<p>Ethereum JavaScript API, middleware to talk to a ethereum node over RPC</p>
<p>Library home page: <a href="https://registry.npmjs.org/web3/-/web3-0.20.7.tgz">https://registry.npmjs.org/web3/-/web3-0.20.7.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@toruslabs/torus-embed/node_modules/web3/package.json</p>
<p>
Dependency Hierarchy:
- torus-embed-1.8.2.tgz (Root Library)
- :x: **web3-0.20.7.tgz** (Vulnerable Library)
</details>
<details><summary><b>web3-1.2.11.tgz</b></p></summary>
<p>Ethereum JavaScript API</p>
<p>Library home page: <a href="https://registry.npmjs.org/web3/-/web3-1.2.11.tgz">https://registry.npmjs.org/web3/-/web3-1.2.11.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/web3/package.json</p>
<p>
Dependency Hierarchy:
- :x: **web3-1.2.11.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/dapp/commit/79de7acd382466c6348d970d41ce91b47fc3366d">79de7acd382466c6348d970d41ce91b47fc3366d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of web3 are vulnerable to Insecure Credential Storage
<p>Publish Date: 2019-05-15
<p>URL: <a href=https://github.com/ethereum/web3.js/issues/2739>WS-2019-0075</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/WS-2019-0075">https://nvd.nist.gov/vuln/detail/WS-2019-0075</a></p>
<p>Release Date: 2019-05-15</p>
<p>Fix Resolution: 1.5.3-rc.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | ws low detected in tgz tgz ws low severity vulnerability vulnerable libraries tgz tgz tgz ethereum javascript api middleware to talk to a ethereum node over rpc library home page a href path to dependency file package json path to vulnerable library node modules toruslabs torus embed node modules package json dependency hierarchy torus embed tgz root library x tgz vulnerable library tgz ethereum javascript api library home page a href path to dependency file package json path to vulnerable library node modules package json dependency hierarchy x tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of are vulnerable to insecure credential storage publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rc step up your open source security game with whitesource | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.