Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16 values | text_combine stringlengths 96 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
651,449 | 21,478,944,226 | IssuesEvent | 2022-04-26 15:51:49 | ucb-rit/coldfront | https://api.github.com/repos/ucb-rit/coldfront | closed | Support uploading and processing valid PIDs file in UI | enhancement low priority lrc-only | Every month, the set of valid PIDs changes. The update comes in the form of an `.xlsx` containing the now-valid PIDs.
In #354, a CLI tool to process this file was added. It would be convenient for superusers to have a way to upload and process this file directly in the web UI, using the same underlying logic. | 1.0 | Support uploading and processing valid PIDs file in UI - Every month, the set of valid PIDs changes. The update comes in the form of an `.xlsx` containing the now-valid PIDs.
In #354, a CLI tool to process this file was added. It would be convenient for superusers to have a way to upload and process this file directly in the web UI, using the same underlying logic. | priority | support uploading and processing valid pids file in ui every month the set of valid pids changes the update comes in the form of an xlsx containing the now valid pids in a cli tool to process this file was added it would be convenient for superusers to have a way to upload and process this file directly in the web ui using the same underlying logic | 1 |
292,135 | 8,953,489,670 | IssuesEvent | 2019-01-25 19:37:17 | mathematicalthinking/vmt | https://api.github.com/repos/mathematicalthinking/vmt | closed | Desmos replayer | High Priority enhancement | Desmos replayer has completely broken since implementing tabs and re-composing the workspace
This shouldn't be that hard to implement. | 1.0 | Desmos replayer - Desmos replayer has completely broken since implementing tabs and re-composing the workspace
This shouldn't be that hard to implement. | priority | desmos replayer desmos replayer has completely broken since implementing tabs and re composing the workspace this shouldn t be that hard to implement | 1 |
101,299 | 4,112,497,726 | IssuesEvent | 2016-06-07 10:42:58 | RestComm/mediaserver | https://api.github.com/repos/RestComm/mediaserver | closed | Modify debug log statement for free endpoints to include the endpoint type | enhancement help-wanted Low-Priority | Add the endpoint type to the log statement so it says: "Bridge Free endpoints 49" or "IVR Free endpoints 49" | 1.0 | Modify debug log statement for free endpoints to include the endpoint type - Add the endpoint type to the log statement so it says: "Bridge Free endpoints 49" or "IVR Free endpoints 49" | priority | modify debug log statement for free endpoints to include the endpoint type add the endpoint type to the log statement so it says bridge free endpoints or ivr free endpoints | 1 |
52,667 | 6,649,536,161 | IssuesEvent | 2017-09-28 13:34:31 | SecUSo/privacy-friendly-weather | https://api.github.com/repos/SecUSo/privacy-friendly-weather | closed | This location is not already in the list | bug Redesign | Hi,
I already put the city 'Annecy' as a permanent location bu I can't select it from the main screen.
If I try to add it again, I have the following message: 'This location is already in the list'. As you can see in the screenshot below, 'Annecy' is not here.
Thanks.

| 1.0 | This location is not already in the list - Hi,
I already put the city 'Annecy' as a permanent location bu I can't select it from the main screen.
If I try to add it again, I have the following message: 'This location is already in the list'. As you can see in the screenshot below, 'Annecy' is not here.
Thanks.

| non_priority | this location is not already in the list hi i already put the city annecy as a permanent location bu i can t select it from the main screen if i try to add it again i have the following message this location is already in the list as you can see in the screenshot below annecy is not here thanks | 0 |
50,181 | 6,063,898,279 | IssuesEvent | 2017-06-14 13:12:48 | acstech/corkboard | https://api.github.com/repos/acstech/corkboard | opened | Implement Create Post Endpoint | feature MEDIUM testable | Add ability to create a new post, but for signed in users only. | 1.0 | Implement Create Post Endpoint - Add ability to create a new post, but for signed in users only. | non_priority | implement create post endpoint add ability to create a new post but for signed in users only | 0 |
6,859 | 3,061,892,617 | IssuesEvent | 2015-08-16 01:16:44 | san650/ember-cli-page-object | https://api.github.com/repos/san650/ember-cli-page-object | closed | Document folder conventions | documentation question | One of the ideas @juanazam and I have is to have some conventions on where to generate the object pages.
We would like to propose and document these conventions so everyone can take advantage of them.
__Edit: take a look at #27 and #32__
## Folder structure
1. Put your Page Objects in `<root>/tests/pages/` folder
2. Put you custom components in `<root>/tests/pages/components/` folder
An example could be
```
myApp/tests
โโโ pages
ย ย โโโ team.js
ย ย โโโ users.js
ย ย โโโ users-add.js
ย ย โโโ components
ย ย ย ย โโโ search-box.js
ย ย ย ย โโโ gear-menu.js
```
## Naming conventions (?)
This is something we didn't discuss yet and we would like to hear ideas.
For simple routes like `users.index` it's easy to pick a name for the page object `<root>/tests/pages/user.js`, but when you have complex nested routes it's no so trivial to pick a name. It would be great to have some guidelines. | 1.0 | Document folder conventions - One of the ideas @juanazam and I have is to have some conventions on where to generate the object pages.
We would like to propose and document these conventions so everyone can take advantage of them.
__Edit: take a look at #27 and #32__
## Folder structure
1. Put your Page Objects in `<root>/tests/pages/` folder
2. Put you custom components in `<root>/tests/pages/components/` folder
An example could be
```
myApp/tests
โโโ pages
ย ย โโโ team.js
ย ย โโโ users.js
ย ย โโโ users-add.js
ย ย โโโ components
ย ย ย ย โโโ search-box.js
ย ย ย ย โโโ gear-menu.js
```
## Naming conventions (?)
This is something we didn't discuss yet and we would like to hear ideas.
For simple routes like `users.index` it's easy to pick a name for the page object `<root>/tests/pages/user.js`, but when you have complex nested routes it's no so trivial to pick a name. It would be great to have some guidelines. | non_priority | document folder conventions one of the ideas juanazam and i have is to have some conventions on where to generate the object pages we would like to propose and document these conventions so everyone can take advantage of them edit take a look at and folder structure put your page objects in tests pages folder put you custom components in tests pages components folder an example could be myapp tests โโโ pages ย ย โโโ team js ย ย โโโ users js ย ย โโโ users add js ย ย โโโ components ย ย ย ย โโโ search box js ย ย ย ย โโโ gear menu js naming conventions this is something we didn t discuss yet and we would like to hear ideas for simple routes like users index it s easy to pick a name for the page object tests pages user js but when you have complex nested routes it s no so trivial to pick a name it would be great to have some guidelines | 0 |
579,185 | 17,185,350,763 | IssuesEvent | 2021-07-16 00:24:18 | apcountryman/picolibrary-microchip-megaavr0 | https://api.github.com/repos/apcountryman/picolibrary-microchip-megaavr0 | closed | Add TWI multiplexed signals configuration and lookup | priority-normal status-awaiting_review type-enhancement | Add TWI multiplexed signals configuration and lookup.
- [x] The ATmega4809 TWI multiplexed signals configuration and lookup functions should be defined in the `::picolibrary::Microchip::megaAVR0::Multiplexed_Signals::ATmega4809` namespace in the `include/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809/twi.h`/`source/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809/twi.cc` header/source file pair
- [x] `include/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809/twi.h` should be included by `include/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809.h`
The following TWI multiplexed signals configuration and lookup type aliases should be defined:
- [x] `TWI_Route = Peripheral::PORTMUX::TWI_Route`: TWI routing configuration
The following TWI multiplexed signals configuration and lookup functions should be defined:
- [x] `void set_twi_route( Peripheral::TWI const & twi, TWI_Route route ) noexcept`: Set a TWI peripheral's routing configuration
- [x] `constexpr auto scl_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's SCL pin number
- [x] `constexpr auto scl_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's SCL pin mask
- [x] `constexpr auto sda_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's SDA pin number
- [x] `constexpr auto sda_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's SDA pin mask
- [x] `atuo twi_controller_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's controller pins port
- [x] `auto twi_controller_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's controller pins virtual port
- [x] `auto controller_scl_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's controller SCL pin port
- [x] `auto controller_scl_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's controller SCL pin virtual port
- [x] `constexpr auto controller_scl_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's controller SCL pin number
- [x] `constexpr auto controller_scl_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's controller SCL pin mask
- [x] `auto controller_sda_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's controller SDA pin port
- [x] `auto controller_sda_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's controller SDA pin virtual port
- [x] `constexpr auto controller_sda_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's controller SDA pin number
- [x] `constexpr auto controller_sda_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's controller SDA pin mask
- [x] `auto twi_device_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's device pins port
- [x] `auto twi_device_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's device pins virtual port
- [x] `auto device_scl_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's device SCL pin port
- [x] `auto device_scl_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's device SCL pin virtual port
- [x] `constexpr auto device_scl_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's device SCL pin number
- [x] `constexpr auto device_scl_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's device SCL pin mask
- [x] `auto device_sda_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's device SDA pin port
- [x] `auto device_sda_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's device SDA pin virtual port
- [x] `constexpr auto device_sda_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's device SDA pin number
- [x] `constexpr auto device_sda_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's device SDA pin mask | 1.0 | Add TWI multiplexed signals configuration and lookup - Add TWI multiplexed signals configuration and lookup.
- [x] The ATmega4809 TWI multiplexed signals configuration and lookup functions should be defined in the `::picolibrary::Microchip::megaAVR0::Multiplexed_Signals::ATmega4809` namespace in the `include/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809/twi.h`/`source/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809/twi.cc` header/source file pair
- [x] `include/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809/twi.h` should be included by `include/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809.h`
The following TWI multiplexed signals configuration and lookup type aliases should be defined:
- [x] `TWI_Route = Peripheral::PORTMUX::TWI_Route`: TWI routing configuration
The following TWI multiplexed signals configuration and lookup functions should be defined:
- [x] `void set_twi_route( Peripheral::TWI const & twi, TWI_Route route ) noexcept`: Set a TWI peripheral's routing configuration
- [x] `constexpr auto scl_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's SCL pin number
- [x] `constexpr auto scl_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's SCL pin mask
- [x] `constexpr auto sda_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's SDA pin number
- [x] `constexpr auto sda_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's SDA pin mask
- [x] `atuo twi_controller_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's controller pins port
- [x] `auto twi_controller_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's controller pins virtual port
- [x] `auto controller_scl_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's controller SCL pin port
- [x] `auto controller_scl_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's controller SCL pin virtual port
- [x] `constexpr auto controller_scl_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's controller SCL pin number
- [x] `constexpr auto controller_scl_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's controller SCL pin mask
- [x] `auto controller_sda_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's controller SDA pin port
- [x] `auto controller_sda_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's controller SDA pin virtual port
- [x] `constexpr auto controller_sda_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's controller SDA pin number
- [x] `constexpr auto controller_sda_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's controller SDA pin mask
- [x] `auto twi_device_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's device pins port
- [x] `auto twi_device_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's device pins virtual port
- [x] `auto device_scl_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's device SCL pin port
- [x] `auto device_scl_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's device SCL pin virtual port
- [x] `constexpr auto device_scl_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's device SCL pin number
- [x] `constexpr auto device_scl_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's device SCL pin mask
- [x] `auto device_sda_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's device SDA pin port
- [x] `auto device_sda_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's device SDA pin virtual port
- [x] `constexpr auto device_sda_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's device SDA pin number
- [x] `constexpr auto device_sda_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's device SDA pin mask | priority | add twi multiplexed signals configuration and lookup add twi multiplexed signals configuration and lookup the twi multiplexed signals configuration and lookup functions should be defined in the picolibrary microchip multiplexed signals namespace in the include picolibrary microchip multiplexed signals twi h source picolibrary microchip multiplexed signals twi cc header source file pair include picolibrary microchip multiplexed signals twi h should be included by include picolibrary microchip multiplexed signals h the following twi multiplexed signals configuration and lookup type aliases should be defined twi route peripheral portmux twi route twi routing configuration the following twi multiplexed signals configuration and lookup functions should be defined void set twi route peripheral twi const twi twi route route noexcept set a twi peripheral s routing configuration constexpr auto scl number peripheral twi const twi noexcept std uint t lookup a twi peripheral s scl pin number constexpr auto scl mask peripheral twi const twi noexcept std t lookup a twi peripheral s scl pin mask constexpr auto sda number peripheral twi const twi noexcept std uint t lookup a twi peripheral s sda pin number constexpr auto sda mask peripheral twi const twi noexcept std t lookup a twi peripheral s sda pin mask atuo twi controller port peripheral twi const twi noexcept peripheral port lookup a twi peripheral s controller pins port auto twi controller vport peripheral twi const twi noexcept peripheral vport lookup a twi peripheral s controller pins virtual port auto controller scl port peripheral twi const twi noexcept peripheral port lookup a twi peripheral s controller scl pin port auto controller scl vport peripheral twi const twi noexcept peripheral vport lookup a twi peripheral s controller scl pin virtual port constexpr auto controller scl number peripheral twi const twi noexcept std uint t lookup a twi peripheral s controller scl pin number constexpr auto controller scl mask peripheral twi const twi noexcept std t lookup a twi peripheral s controller scl pin mask auto controller sda port peripheral twi const twi noexcept peripheral port lookup a twi peripheral s controller sda pin port auto controller sda vport peripheral twi const twi noexcept peripheral vport lookup a twi peripheral s controller sda pin virtual port constexpr auto controller sda number peripheral twi const twi noexcept std uint t lookup a twi peripheral s controller sda pin number constexpr auto controller sda mask peripheral twi const twi noexcept std t lookup a twi peripheral s controller sda pin mask auto twi device port peripheral twi const twi noexcept peripheral port lookup a twi peripheral s device pins port auto twi device vport peripheral twi const twi noexcept peripheral vport lookup a twi peripheral s device pins virtual port auto device scl port peripheral twi const twi noexcept peripheral port lookup a twi peripheral s device scl pin port auto device scl vport peripheral twi const twi noexcept peripheral vport lookup a twi peripheral s device scl pin virtual port constexpr auto device scl number peripheral twi const twi noexcept std uint t lookup a twi peripheral s device scl pin number constexpr auto device scl mask peripheral twi const twi noexcept std t lookup a twi peripheral s device scl pin mask auto device sda port peripheral twi const twi noexcept peripheral port lookup a twi peripheral s device sda pin port auto device sda vport peripheral twi const twi noexcept peripheral vport lookup a twi peripheral s device sda pin virtual port constexpr auto device sda number peripheral twi const twi noexcept std uint t lookup a twi peripheral s device sda pin number constexpr auto device sda mask peripheral twi const twi noexcept std t lookup a twi peripheral s device sda pin mask | 1 |
58,274 | 16,458,674,674 | IssuesEvent | 2021-05-21 15:43:31 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Room Specific functionality VS General functionality | T-Defect | 
### Description
It's not logical to have both of these in the same place - entangled in the same area. This causes confusion and users do not know where to look for things.
The video and audio call buttons are room specific just like the info button and partially the search button (also the hide widget and message pinning) this gives the user the feeling that also the notifications bell button is room specific.
The most important fact though is the search all messages, this is very well hidden and has to be navigated to by going to by going to a room (yes I know users are 99% of the time in a room, but there are some cases where they are not e.g. #/home) and then switching from "This Room" to "All Rooms" which basically renders this useless for normal users.

Suggestions:
1. Replace the filter room with a general search which will also filter rooms.
2. add a search all messages button in #/home

and replace the "Explore rooms" button with a home button (this will also make the web logic more similar to mobile)

3. The general search could be placed above the other room specific icons, which will also make room for other future non room specific functionality (for example a calendar), this will also solve where to put the Notifications bell button.
| 1.0 | Room Specific functionality VS General functionality - 
### Description
It's not logical to have both of these in the same place - entangled in the same area. This causes confusion and users do not know where to look for things.
The video and audio call buttons are room specific just like the info button and partially the search button (also the hide widget and message pinning) this gives the user the feeling that also the notifications bell button is room specific.
The most important fact though is the search all messages, this is very well hidden and has to be navigated to by going to by going to a room (yes I know users are 99% of the time in a room, but there are some cases where they are not e.g. #/home) and then switching from "This Room" to "All Rooms" which basically renders this useless for normal users.

Suggestions:
1. Replace the filter room with a general search which will also filter rooms.
2. add a search all messages button in #/home

and replace the "Explore rooms" button with a home button (this will also make the web logic more similar to mobile)

3. The general search could be placed above the other room specific icons, which will also make room for other future non room specific functionality (for example a calendar), this will also solve where to put the Notifications bell button.
| non_priority | room specific functionality vs general functionality description it s not logical to have both of these in the same place entangled in the same area this causes confusion and users do not know where to look for things the video and audio call buttons are room specific just like the info button and partially the search button also the hide widget and message pinning this gives the user the feeling that also the notifications bell button is room specific the most important fact though is the search all messages this is very well hidden and has to be navigated to by going to by going to a room yes i know users are of the time in a room but there are some cases where they are not e g home and then switching from this room to all rooms which basically renders this useless for normal users suggestions replace the filter room with a general search which will also filter rooms add a search all messages button in home and replace the explore rooms button with a home button this will also make the web logic more similar to mobile the general search could be placed above the other room specific icons which will also make room for other future non room specific functionality for example a calendar this will also solve where to put the notifications bell button | 0 |
99,311 | 8,696,946,334 | IssuesEvent | 2018-12-04 18:54:31 | nebula-plugins/gradle-nebula-integration | https://api.github.com/repos/nebula-plugins/gradle-nebula-integration | closed | Dependency insight with preferred dependency | priority testable | Tracking gradle/gradle#7334
Related to dependency insight not showing full information on how/ where a preferred dependency contributes to the selection reason. | 1.0 | Dependency insight with preferred dependency - Tracking gradle/gradle#7334
Related to dependency insight not showing full information on how/ where a preferred dependency contributes to the selection reason. | non_priority | dependency insight with preferred dependency tracking gradle gradle related to dependency insight not showing full information on how where a preferred dependency contributes to the selection reason | 0 |
201,905 | 15,816,491,071 | IssuesEvent | 2021-04-05 13:12:00 | microsoft/code-with-engineering-playbook | https://api.github.com/repos/microsoft/code-with-engineering-playbook | opened | Add "Keeping documentation updated" best practices to Documentation section | documentation | After the new Documentation section is created (#533), add a new "Keeping documentation updated" document under "Best practices". | 1.0 | Add "Keeping documentation updated" best practices to Documentation section - After the new Documentation section is created (#533), add a new "Keeping documentation updated" document under "Best practices". | non_priority | add keeping documentation updated best practices to documentation section after the new documentation section is created add a new keeping documentation updated document under best practices | 0 |
657,721 | 21,802,468,679 | IssuesEvent | 2022-05-16 07:12:38 | teogor/teogor.github.io | https://api.github.com/repos/teogor/teogor.github.io | opened | Sitemap is not updated | @bug @priority-critical | The sitemap is not updated based on the latest values - [sitemap](teogor.github.io/sitemap) | 1.0 | Sitemap is not updated - The sitemap is not updated based on the latest values - [sitemap](teogor.github.io/sitemap) | priority | sitemap is not updated the sitemap is not updated based on the latest values teogor github io sitemap | 1 |
203,266 | 23,131,775,238 | IssuesEvent | 2022-07-28 11:02:30 | Gal-Doron/-gradle-main-app-gal2 | https://api.github.com/repos/Gal-Doron/-gradle-main-app-gal2 | closed | guava-28.1-jre.jar: 1 vulnerabilities (highest severity is: 3.3) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>guava-28.1-jre.jar</b></p></summary>
<p>Guava is a suite of core and expanded libraries that include
utility classes, google's collections, io classes, and much
much more.</p>
<p>Library home page: <a href="https://github.com/google/guava">https://github.com/google/guava</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /modules-2/files-2.1/com.google.guava/guava/28.1-jre/b0e91dcb6a44ffb6221b5027e12a5cb34b841145/guava-28.1-jre.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/-gradle-main-app-gal2/commit/13111bc5f66b724f62da31e0515a2c958de72ac5">13111bc5f66b724f62da31e0515a2c958de72ac5</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-8908](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.3 | guava-28.1-jre.jar | Direct | 30.0-android | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2020-8908</summary>
### Vulnerable Library - <b>guava-28.1-jre.jar</b></p>
<p>Guava is a suite of core and expanded libraries that include
utility classes, google's collections, io classes, and much
much more.</p>
<p>Library home page: <a href="https://github.com/google/guava">https://github.com/google/guava</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /modules-2/files-2.1/com.google.guava/guava/28.1-jre/b0e91dcb6a44ffb6221b5027e12a5cb34b841145/guava-28.1-jre.jar</p>
<p>
Dependency Hierarchy:
- :x: **guava-28.1-jre.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/-gradle-main-app-gal2/commit/13111bc5f66b724f62da31e0515a2c958de72ac5">13111bc5f66b724f62da31e0515a2c958de72ac5</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A temp directory creation vulnerability exists in all versions of Guava, allowing an attacker with access to the machine to potentially access data in a temporary directory created by the Guava API com.google.common.io.Files.createTempDir(). By default, on unix-like systems, the created directory is world-readable (readable by an attacker with access to the system). The method in question has been marked @Deprecated in versions 30.0 and later and should not be used. For Android developers, we recommend choosing a temporary directory API provided by Android, such as context.getCacheDir(). For other Java developers, we recommend migrating to the Java 7 API java.nio.file.Files.createTempDirectory() which explicitly configures permissions of 700, or configuring the Java runtime's java.io.tmpdir system property to point to a location whose permissions are appropriately configured.
<p>Publish Date: 2020-12-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908>CVE-2020-8908</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>3.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908</a></p>
<p>Release Date: 2020-12-10</p>
<p>Fix Resolution: 30.0-android</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | True | guava-28.1-jre.jar: 1 vulnerabilities (highest severity is: 3.3) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>guava-28.1-jre.jar</b></p></summary>
<p>Guava is a suite of core and expanded libraries that include
utility classes, google's collections, io classes, and much
much more.</p>
<p>Library home page: <a href="https://github.com/google/guava">https://github.com/google/guava</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /modules-2/files-2.1/com.google.guava/guava/28.1-jre/b0e91dcb6a44ffb6221b5027e12a5cb34b841145/guava-28.1-jre.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/-gradle-main-app-gal2/commit/13111bc5f66b724f62da31e0515a2c958de72ac5">13111bc5f66b724f62da31e0515a2c958de72ac5</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-8908](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.3 | guava-28.1-jre.jar | Direct | 30.0-android | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2020-8908</summary>
### Vulnerable Library - <b>guava-28.1-jre.jar</b></p>
<p>Guava is a suite of core and expanded libraries that include
utility classes, google's collections, io classes, and much
much more.</p>
<p>Library home page: <a href="https://github.com/google/guava">https://github.com/google/guava</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /modules-2/files-2.1/com.google.guava/guava/28.1-jre/b0e91dcb6a44ffb6221b5027e12a5cb34b841145/guava-28.1-jre.jar</p>
<p>
Dependency Hierarchy:
- :x: **guava-28.1-jre.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/-gradle-main-app-gal2/commit/13111bc5f66b724f62da31e0515a2c958de72ac5">13111bc5f66b724f62da31e0515a2c958de72ac5</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A temp directory creation vulnerability exists in all versions of Guava, allowing an attacker with access to the machine to potentially access data in a temporary directory created by the Guava API com.google.common.io.Files.createTempDir(). By default, on unix-like systems, the created directory is world-readable (readable by an attacker with access to the system). The method in question has been marked @Deprecated in versions 30.0 and later and should not be used. For Android developers, we recommend choosing a temporary directory API provided by Android, such as context.getCacheDir(). For other Java developers, we recommend migrating to the Java 7 API java.nio.file.Files.createTempDirectory() which explicitly configures permissions of 700, or configuring the Java runtime's java.io.tmpdir system property to point to a location whose permissions are appropriately configured.
<p>Publish Date: 2020-12-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908>CVE-2020-8908</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>3.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908</a></p>
<p>Release Date: 2020-12-10</p>
<p>Fix Resolution: 30.0-android</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | non_priority | guava jre jar vulnerabilities highest severity is autoclosed vulnerable library guava jre jar guava is a suite of core and expanded libraries that include utility classes google s collections io classes and much much more library home page a href path to dependency file build gradle path to vulnerable library modules files com google guava guava jre guava jre jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available low guava jre jar direct android details cve vulnerable library guava jre jar guava is a suite of core and expanded libraries that include utility classes google s collections io classes and much much more library home page a href path to dependency file build gradle path to vulnerable library modules files com google guava guava jre guava jre jar dependency hierarchy x guava jre jar vulnerable library found in head commit a href found in base branch main vulnerability details a temp directory creation vulnerability exists in all versions of guava allowing an attacker with access to the machine to potentially access data in a temporary directory created by the guava api com google common io files createtempdir by default on unix like systems the created directory is world readable readable by an attacker with access to the system the method in question has been marked deprecated in versions and later and should not be used for android developers we recommend choosing a temporary directory api provided by android such as context getcachedir for other java developers we recommend migrating to the java api java nio file files createtempdirectory which explicitly configures permissions of or configuring the java runtime s java io tmpdir system property to point to a location whose permissions are appropriately configured publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue | 0 |
11,977 | 3,244,706,531 | IssuesEvent | 2015-10-16 05:10:33 | commitizen/cz-cli | https://api.github.com/repos/commitizen/cz-cli | closed | Need a test suite | enhancement in progress tests | There is currently no test suite for commitizen or a way for adapters to smoke test. | 1.0 | Need a test suite - There is currently no test suite for commitizen or a way for adapters to smoke test. | non_priority | need a test suite there is currently no test suite for commitizen or a way for adapters to smoke test | 0 |
259,799 | 27,725,073,192 | IssuesEvent | 2023-03-15 01:06:25 | safat/conductor | https://api.github.com/repos/safat/conductor | opened | CVE-2019-20920 (High) detected in handlebars-4.0.5.js | Mend: dependency security vulnerability | ## CVE-2019-20920 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.5.js</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.js</a></p>
<p>Path to dependency file: /server/build/resources/main/swagger-ui/index.html</p>
<p>Path to vulnerable library: /d/resources/main/swagger-ui/lib/handlebars-4.0.5.js,/server/src/main/resources/swagger-ui/lib/handlebars-4.0.5.js,/main/resources/swagger-ui/lib/handlebars-4.0.5.js,/server/build/resources/main/swagger-ui/lib/handlebars-4.0.5.js</p>
<p>
Dependency Hierarchy:
- :x: **handlebars-4.0.5.js** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Handlebars before 3.0.8 and 4.x before 4.5.3 is vulnerable to Arbitrary Code Execution. The lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript. This can be used to run arbitrary code on a server processing Handlebars templates or in a victim's browser (effectively serving as XSS).
<p>Publish Date: 2020-09-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-20920>CVE-2019-20920</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1316">https://www.npmjs.com/advisories/1316</a></p>
<p>Release Date: 2020-10-15</p>
<p>Fix Resolution: handlebars - 4.5.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-20920 (High) detected in handlebars-4.0.5.js - ## CVE-2019-20920 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.5.js</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.js</a></p>
<p>Path to dependency file: /server/build/resources/main/swagger-ui/index.html</p>
<p>Path to vulnerable library: /d/resources/main/swagger-ui/lib/handlebars-4.0.5.js,/server/src/main/resources/swagger-ui/lib/handlebars-4.0.5.js,/main/resources/swagger-ui/lib/handlebars-4.0.5.js,/server/build/resources/main/swagger-ui/lib/handlebars-4.0.5.js</p>
<p>
Dependency Hierarchy:
- :x: **handlebars-4.0.5.js** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Handlebars before 3.0.8 and 4.x before 4.5.3 is vulnerable to Arbitrary Code Execution. The lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript. This can be used to run arbitrary code on a server processing Handlebars templates or in a victim's browser (effectively serving as XSS).
<p>Publish Date: 2020-09-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-20920>CVE-2019-20920</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1316">https://www.npmjs.com/advisories/1316</a></p>
<p>Release Date: 2020-10-15</p>
<p>Fix Resolution: handlebars - 4.5.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in handlebars js cve high severity vulnerability vulnerable library handlebars js handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file server build resources main swagger ui index html path to vulnerable library d resources main swagger ui lib handlebars js server src main resources swagger ui lib handlebars js main resources swagger ui lib handlebars js server build resources main swagger ui lib handlebars js dependency hierarchy x handlebars js vulnerable library found in base branch master vulnerability details handlebars before and x before is vulnerable to arbitrary code execution the lookup helper fails to properly validate templates allowing attackers to submit templates that execute arbitrary javascript this can be used to run arbitrary code on a server processing handlebars templates or in a victim s browser effectively serving as xss publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars step up your open source security game with mend | 0 |
261,556 | 19,716,482,317 | IssuesEvent | 2022-01-13 11:29:58 | obss/sahi | https://api.github.com/repos/obss/sahi | reopened | Evaluation & Error Analysis | documentation | โ๏ธ Create COCO formatted prediction results using COCO formatted dataset:

```bash
sahi predict --source image_dir/ --dataset_json_path dataset.json --model_type detectron2 --model_path weight.pt --config_path config.yaml --no_sliced_prediction
```
โ๏ธ Calculate COCO evaluation metrics using the created result.json:

```bash
sahi coco evaluate--dataset_json_path dataset.json --result_json_path result.json
```
โ๏ธ Create error analysis plots using the created result.json:

```bash
sahi coco analyse --dataset_json_path dataset.json --result_json_path result.json
```
๐ฏ **Meaning of the metrics:**
**C75:** Results at 0.75 IOU threshod
**C50:** Results at 0.75 IOU threshold
**Loc:** Results after ignoring localization errors
**Sim:** Results after ignoring supercategory false positives
**Oth:** Results after ignoring all category confusions
**BG:** Results after ignoring all false positives
**FN:** Results after ignoring all false negatives
๐ **Possible model improvements:**
**C75-C50 and C50-Loc**=Potential gain with more accurate bounding box prediction
**Loc-Sim**=Potential gain after fixing supercategory confusions
**Loc-Oth**=Potential gain after fixing category confusions
**Oth-BG**=Potential gain after fixing all false positives
**BG-FN**=Potential gain after fixing all false negatives
| 1.0 | Evaluation & Error Analysis - โ๏ธ Create COCO formatted prediction results using COCO formatted dataset:

```bash
sahi predict --source image_dir/ --dataset_json_path dataset.json --model_type detectron2 --model_path weight.pt --config_path config.yaml --no_sliced_prediction
```
โ๏ธ Calculate COCO evaluation metrics using the created result.json:

```bash
sahi coco evaluate--dataset_json_path dataset.json --result_json_path result.json
```
โ๏ธ Create error analysis plots using the created result.json:

```bash
sahi coco analyse --dataset_json_path dataset.json --result_json_path result.json
```
๐ฏ **Meaning of the metrics:**
**C75:** Results at 0.75 IOU threshod
**C50:** Results at 0.75 IOU threshold
**Loc:** Results after ignoring localization errors
**Sim:** Results after ignoring supercategory false positives
**Oth:** Results after ignoring all category confusions
**BG:** Results after ignoring all false positives
**FN:** Results after ignoring all false negatives
๐ **Possible model improvements:**
**C75-C50 and C50-Loc**=Potential gain with more accurate bounding box prediction
**Loc-Sim**=Potential gain after fixing supercategory confusions
**Loc-Oth**=Potential gain after fixing category confusions
**Oth-BG**=Potential gain after fixing all false positives
**BG-FN**=Potential gain after fixing all false negatives
| non_priority | evaluation error analysis โ๏ธ create coco formatted prediction results using coco formatted dataset bash sahi predict source image dir dataset json path dataset json model type model path weight pt config path config yaml no sliced prediction โ๏ธ calculate coco evaluation metrics using the created result json bash sahi coco evaluate dataset json path dataset json result json path result json โ๏ธ create error analysis plots using the created result json bash sahi coco analyse dataset json path dataset json result json path result json ๐ฏ meaning of the metrics results at iou threshod results at iou threshold loc results after ignoring localization errors sim results after ignoring supercategory false positives oth results after ignoring all category confusions bg results after ignoring all false positives fn results after ignoring all false negatives ๐ possible model improvements and loc potential gain with more accurate bounding box prediction loc sim potential gain after fixing supercategory confusions loc oth potential gain after fixing category confusions oth bg potential gain after fixing all false positives bg fn potential gain after fixing all false negatives | 0 |
390,335 | 26,858,692,689 | IssuesEvent | 2023-02-03 16:31:47 | tmforum-oda/oda-ca | https://api.github.com/repos/tmforum-oda/oda-ca | closed | Docs: Add Prometheus to API Operator documentation | documentation | ## Description
The API apiOperatorIstio includes handling Prometheus APIs and integrating to the Prometheus operator. Need to add documentation describing how this integration works. | 1.0 | Docs: Add Prometheus to API Operator documentation - ## Description
The API apiOperatorIstio includes handling Prometheus APIs and integrating to the Prometheus operator. Need to add documentation describing how this integration works. | non_priority | docs add prometheus to api operator documentation description the api apioperatoristio includes handling prometheus apis and integrating to the prometheus operator need to add documentation describing how this integration works | 0 |
93,668 | 27,012,207,630 | IssuesEvent | 2023-02-10 16:15:00 | apache/camel-k | https://api.github.com/repos/apache/camel-k | closed | Integration Kit image build runs into an endless loop when a lower level image can't be pulled from Registry | kind/bug status/waiting-for-feedback area/build-system | Hi all.
We're still working with Camel-K 1.8.0. We did experience that an Integration Kit build runs into an endless loop when a lower level image can't be pulled from Registry. The build will be tried again and again.
We must delete the Operator, afterwards the Integration Kit build was running properly again.
I'd expect in such a case (lower level image not found), that the Operator does have a mitigation strategy built in (e.g. build the Integration Kit based on an other one existing or from scratch).
So far the question, does the Integration Kit build process be re-worked in 1.9.x. And .. we'd expect a bit more detailed log entries in the Operator logs.
Best
Gerald | 1.0 | Integration Kit image build runs into an endless loop when a lower level image can't be pulled from Registry - Hi all.
We're still working with Camel-K 1.8.0. We did experience that an Integration Kit build runs into an endless loop when a lower level image can't be pulled from Registry. The build will be tried again and again.
We must delete the Operator, afterwards the Integration Kit build was running properly again.
I'd expect in such a case (lower level image not found), that the Operator does have a mitigation strategy built in (e.g. build the Integration Kit based on an other one existing or from scratch).
So far the question, does the Integration Kit build process be re-worked in 1.9.x. And .. we'd expect a bit more detailed log entries in the Operator logs.
Best
Gerald | non_priority | integration kit image build runs into an endless loop when a lower level image can t be pulled from registry hi all we re still working with camel k we did experience that an integration kit build runs into an endless loop when a lower level image can t be pulled from registry the build will be tried again and again we must delete the operator afterwards the integration kit build was running properly again i d expect in such a case lower level image not found that the operator does have a mitigation strategy built in e g build the integration kit based on an other one existing or from scratch so far the question does the integration kit build process be re worked in x and we d expect a bit more detailed log entries in the operator logs best gerald | 0 |
162,538 | 6,154,982,119 | IssuesEvent | 2017-06-28 13:54:06 | k0shk0sh/FastHub | https://api.github.com/repos/k0shk0sh/FastHub | closed | Won't mark notification as read | Priority: Medium Status: Accepted Status: Completed | **App Version: 3.2.0**
**OS Version: 25**
**Model: LGE-Nexus 5X**
I have the settings to __not__ mark as read notifications when I click on them. That way I can check them out and still have them when I reach a desktop version of GitHub.
The problem is that sometimes I want to mark a notification as read because I don't have to see it when I get to the desktop, so I press the โ๏ธ to mark it and it still won't get marked.
I'm guessing it happens because I have the settings set to not mark as read.
_Sent from my LGE Nexus 5X using [FastHub](https://play.google.com/store/apps/details?id=com.fastaccess.github)_ | 1.0 | Won't mark notification as read - **App Version: 3.2.0**
**OS Version: 25**
**Model: LGE-Nexus 5X**
I have the settings to __not__ mark as read notifications when I click on them. That way I can check them out and still have them when I reach a desktop version of GitHub.
The problem is that sometimes I want to mark a notification as read because I don't have to see it when I get to the desktop, so I press the โ๏ธ to mark it and it still won't get marked.
I'm guessing it happens because I have the settings set to not mark as read.
_Sent from my LGE Nexus 5X using [FastHub](https://play.google.com/store/apps/details?id=com.fastaccess.github)_ | priority | won t mark notification as read app version os version model lge nexus i have the settings to not mark as read notifications when i click on them that way i can check them out and still have them when i reach a desktop version of github the problem is that sometimes i want to mark a notification as read because i don t have to see it when i get to the desktop so i press the โ๏ธ to mark it and it still won t get marked i m guessing it happens because i have the settings set to not mark as read sent from my lge nexus using | 1 |
801,886 | 28,506,236,316 | IssuesEvent | 2023-04-18 21:45:20 | envoyproxy/gateway | https://api.github.com/repos/envoyproxy/gateway | closed | Release 0.4.0-rc.1 install.yaml installs EG in default namespace | bug help wanted priority/high | *Description*:
The install yaml in the 0.4.0-rc.1 release https://github.com/envoyproxy/gateway/releases/tag/v0.4.0-rc.1 does not have the metadata.Namespace field set to `envoy-gateway-system` installing all EG resources in the `default` namespace
*Repro steps*:
> Include sample requests, environment, etc. All data and inputs
required to reproduce the bug.
`kubectl apply -f https://github.com/envoyproxy/gateway/releases/download/v0.4.0-rc.1/install.yaml`
>**Note**: If there are privacy concerns, sanitize the data prior to
sharing.
*Environment*:
>Include the environment like gateway version, envoy version and so on.
*Logs*:
>Include the access logs and the Envoy logs.
| 1.0 | Release 0.4.0-rc.1 install.yaml installs EG in default namespace - *Description*:
The install yaml in the 0.4.0-rc.1 release https://github.com/envoyproxy/gateway/releases/tag/v0.4.0-rc.1 does not have the metadata.Namespace field set to `envoy-gateway-system` installing all EG resources in the `default` namespace
*Repro steps*:
> Include sample requests, environment, etc. All data and inputs
required to reproduce the bug.
`kubectl apply -f https://github.com/envoyproxy/gateway/releases/download/v0.4.0-rc.1/install.yaml`
>**Note**: If there are privacy concerns, sanitize the data prior to
sharing.
*Environment*:
>Include the environment like gateway version, envoy version and so on.
*Logs*:
>Include the access logs and the Envoy logs.
| priority | release rc install yaml installs eg in default namespace description the install yaml in the rc release does not have the metadata namespace field set to envoy gateway system installing all eg resources in the default namespace repro steps include sample requests environment etc all data and inputs required to reproduce the bug kubectl apply f note if there are privacy concerns sanitize the data prior to sharing environment include the environment like gateway version envoy version and so on logs include the access logs and the envoy logs | 1 |
447,459 | 31,711,290,553 | IssuesEvent | 2023-09-09 09:57:27 | scrapfly/scrapfly-scrapers | https://api.github.com/repos/scrapfly/scrapfly-scrapers | closed | Trouble following instructions | documentation | Hi there,
I'm having trouble following instructions from https://github.com/scrapfly/scrapfly-scrapers/tree/main/bookingcom-scraper
At point 2 when trying `git clone git@github.com:scrapfly/scrapfly-scrapers.git`
I got:
```git@github.com: Permission denied (publickey).
fatal: Could not read from remote repository.
Please make sure you have the correct access rights
and the repository exists.
```
As a workaround I did `git clone https://github.com/scrapfly/scrapfly-scrapers.git`
Next step is `poetry install .`
I got: `-bash: poetry: command not found`
I go to https://scrapfly.io/blog/how-to-scrape-bookingcom/ and as stated, I try `pip install "httpx[http2,brotli]" parsel`
I got: `-bash: pip: command not found`
As a workaround I did `sudo apt install python3-pip` and then `pip install "httpx[http2,brotli]" parsel`
I got:
```
error: externally-managed-environment
ร This environment is externally managed
โฐโ> To install Python packages system-wide, try apt install
python3-xyz, where xyz is the package you are trying to
install.
If you wish to install a non-Debian-packaged Python package,
create a virtual environment using python3 -m venv path/to/venv.
Then use path/to/venv/bin/python and path/to/venv/bin/pip. Make
sure you have python3-full installed.
If you wish to install a non-Debian packaged Python application,
it may be easiest to use pipx install xyz, which will manage a
virtual environment for you. Make sure you have pipx installed.
See /usr/share/doc/python3.11/README.venv for more information.
note: If you believe this is a mistake, please contact your Python installation or OS distribution provider. You can override this, at the risk of breaking your Python installation or OS, by passing --break-system-packages.
hint: See PEP 668 for the detailed specification.
```
After searching how to install poetry I did: `curl -sSL https://install.python-poetry.org | python3 -`, added Poetry's bin directory in my PATH environment and tried again `poetry install .`
I got:
```
Creating virtualenv scrapfly-booking-tyZw0pBk-py3.11 in /home/xxxx/.cache/pypoetry/virtualenvs
No arguments expected for "install" command, got "."
```
I decided to go for step 3: `poetry run python run.py`
I got:
```
Traceback (most recent call last):
File "/home/esaumell/scrapfly-scrapers/bookingcom-scraper/run.py", line 12, in <module>
import bookingcom
File "/home/esaumell/scrapfly-scrapers/bookingcom-scraper/bookingcom.py", line 19, in <module>
from loguru import logger as log
ModuleNotFoundError: No module named 'loguru'
```
So brave and I went for step 4: `poetry install --with dev`
I got:
```
Installing dependencies from lock file
Package operations: 67 installs, 1 update, 0 removals
โข Installing certifi (2023.5.7)
โข Installing charset-normalizer (3.1.0)
โข Installing idna (3.4)
โข Installing pycparser (2.21)
โข Installing six (1.16.0)
โข Installing urllib3 (2.0.2)
โข Installing attrs (23.1.0)
โข Installing cffi (1.15.1)
โข Installing cssselect (1.2.0)
โข Installing jmespath (1.0.1)
โข Installing isodate (0.6.1)
โข Installing lxml (4.9.2)
โข Installing packaging (23.1)
โข Installing pyasn1 (0.5.0)
โข Installing pyparsing (3.0.9)
โข Installing requests (2.31.0)
โข Downgrading setuptools (68.1.2 -> 67.8.0)
โข Installing soupsieve (2.4.1)
โข Installing w3lib (2.1.1)
โข Installing webencodings (0.5.1)
โข Installing automat (22.10.0)
โข Installing beautifulsoup4 (4.12.2)
โข Installing constantly (15.1.0)
โข Installing cryptography (40.0.2)
โข Installing filelock (3.12.0)
โข Installing html5lib (1.1)
โข Installing hyperlink (21.0.0)
โข Installing incremental (22.10.0)
โข Installing itemadapter (0.8.0)
โข Installing parsel (1.8.1)
โข Installing pyasn1-modules (0.3.0)
โข Installing rdflib (6.3.2)
โข Installing requests-file (1.5.1)
โข Installing typing-extensions (4.6.1)
โข Installing zope-interface (6.0)
โข Installing html-text (0.5.2)
โข Installing iniconfig (2.0.0)
โข Installing itemloaders (1.1.0)
โข Installing jstyleson (0.0.2)
โข Installing mf2py (1.1.2)
โข Installing pluggy (1.0.0)
โข Installing protego (0.2.1)
โข Installing pydispatcher (2.0.7)
โข Installing pyopenssl (23.1.1)
โข Installing pyrdfa3 (3.5.3)
โข Installing queuelib (1.6.2)
โข Installing service-identity (21.1.0)
โข Installing tldextract (3.4.4)
โข Installing twisted (22.10.0)
โข Installing backoff (2.2.1)
โข Installing brotlipy (0.7.0)
โข Installing cchardet (2.1.7): Failed
ChefBuildError
Backend subprocess exited when trying to invoke build_wheel
running bdist_wheel
running build
running build_py
creating build
creating build/lib.linux-aarch64-cpython-311
creating build/lib.linux-aarch64-cpython-311/cchardet
copying src/cchardet/version.py -> build/lib.linux-aarch64-cpython-311/cchardet
copying src/cchardet/__init__.py -> build/lib.linux-aarch64-cpython-311/cchardet
running build_ext
building 'cchardet._cchardet' extension
creating build/temp.linux-aarch64-cpython-311
creating build/temp.linux-aarch64-cpython-311/src
creating build/temp.linux-aarch64-cpython-311/src/cchardet
creating build/temp.linux-aarch64-cpython-311/src/ext
creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet
creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src
creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src/LangModels
aarch64-linux-gnu-gcc -Wsign-compare -DNDEBUG -g -fwrapv -O2 -Wall -g -fstack-protector-strong -Wformat -Werror=format-security -g -fwrapv -O2 -fPIC -Isrc/ext/uchardet/src -I/tmp/tmp2edcn7ai/.venv/include -I/usr/include/python3.11 -c src/cchardet/_cchardet.cpp -o build/temp.linux-aarch64-cpython-311/src/cchardet/_cchardet.o
src/cchardet/_cchardet.cpp:4:10: fatal error: Python.h: No such file or directory
4 | #include "Python.h"
| ^~~~~~~~~~
compilation terminated.
error: command '/usr/bin/aarch64-linux-gnu-gcc' failed with exit code 1
at ~/.local/share/pypoetry/venv/lib/python3.11/site-packages/poetry/installation/chef.py:147 in _prepare
143โ
144โ error = ChefBuildError("\n\n".join(message_parts))
145โ
146โ if error is not None:
โ 147โ raise error from None
148โ
149โ return path
150โ
151โ def _prepare_sdist(self, archive: Path, destination: Path | None = None) -> Path:
Note: This error originates from the build backend, and is likely not a problem with poetry but with cchardet (2.1.7) not supporting PEP 517 builds. You can verify this by running 'pip wheel --use-pep517 "cchardet (==2.1.7)"'.
โข Installing click (8.1.3)
โข Installing decorator (5.1.1)
โข Installing extruct (0.14.0)
โข Installing loguru (0.7.0)
โข Installing msgpack (1.0.5)
โข Installing mypy-extensions (1.0.0)
โข Installing pathspec (0.11.1)
โข Installing platformdirs (3.5.1)
โข Installing pytest (7.3.1)
โข Installing python-dateutil (2.8.2)
โข Installing scrapy (2.9.0)
```
As stated on the output I tried `pip wheel --use-pep517 "cchardet (==2.1.7)"`
I got:
```
Collecting cchardet==2.1.7
Downloading cchardet-2.1.7.tar.gz (653 kB)
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 653.6/653.6 kB 13.5 MB/s eta 0:00:00
Installing build dependencies ... done
Getting requirements to build wheel ... done
Preparing metadata (pyproject.toml) ... done
Building wheels for collected packages: cchardet
Building wheel for cchardet (pyproject.toml) ... error
error: subprocess-exited-with-error
ร Building wheel for cchardet (pyproject.toml) did not run successfully.
โ exit code: 1
โฐโ> [23 lines of output]
running bdist_wheel
running build
running build_py
creating build
creating build/lib.linux-aarch64-cpython-311
creating build/lib.linux-aarch64-cpython-311/cchardet
copying src/cchardet/version.py -> build/lib.linux-aarch64-cpython-311/cchardet
copying src/cchardet/__init__.py -> build/lib.linux-aarch64-cpython-311/cchardet
running build_ext
building 'cchardet._cchardet' extension
creating build/temp.linux-aarch64-cpython-311
creating build/temp.linux-aarch64-cpython-311/src
creating build/temp.linux-aarch64-cpython-311/src/cchardet
creating build/temp.linux-aarch64-cpython-311/src/ext
creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet
creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src
creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src/LangModels
aarch64-linux-gnu-gcc -Wsign-compare -DNDEBUG -g -fwrapv -O2 -Wall -g -fstack-protector-strong -Wformat -Werror=format-security -g -fwrapv -O2 -fPIC -Isrc/ext/uchardet/src -I/usr/include/python3.11 -c src/cchardet/_cchardet.cpp -o build/temp.linux-aarch64-cpython-311/src/cchardet/_cchardet.o
src/cchardet/_cchardet.cpp:4:10: fatal error: Python.h: No such file or directory
4 | #include "Python.h"
| ^~~~~~~~~~
compilation terminated.
error: command '/usr/bin/aarch64-linux-gnu-gcc' failed with exit code 1
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building wheel for cchardet
Failed to build cchardet
ERROR: Failed to build one or more wheels
```
This is running on a fresh Debian 12 setup and I have installed poetry 1.6.1
Searching for the problem I'm suggested to try to downgrade poetry, so I try `poetry self update 1.4` but `poetry install --with dev` fails again on cchardet
I also tried installing `python3-dev` without success. I don't know where to go from here. Any help would be appreciated. | 1.0 | Trouble following instructions - Hi there,
I'm having trouble following instructions from https://github.com/scrapfly/scrapfly-scrapers/tree/main/bookingcom-scraper
At point 2 when trying `git clone git@github.com:scrapfly/scrapfly-scrapers.git`
I got:
```git@github.com: Permission denied (publickey).
fatal: Could not read from remote repository.
Please make sure you have the correct access rights
and the repository exists.
```
As a workaround I did `git clone https://github.com/scrapfly/scrapfly-scrapers.git`
Next step is `poetry install .`
I got: `-bash: poetry: command not found`
I go to https://scrapfly.io/blog/how-to-scrape-bookingcom/ and as stated, I try `pip install "httpx[http2,brotli]" parsel`
I got: `-bash: pip: command not found`
As a workaround I did `sudo apt install python3-pip` and then `pip install "httpx[http2,brotli]" parsel`
I got:
```
error: externally-managed-environment
ร This environment is externally managed
โฐโ> To install Python packages system-wide, try apt install
python3-xyz, where xyz is the package you are trying to
install.
If you wish to install a non-Debian-packaged Python package,
create a virtual environment using python3 -m venv path/to/venv.
Then use path/to/venv/bin/python and path/to/venv/bin/pip. Make
sure you have python3-full installed.
If you wish to install a non-Debian packaged Python application,
it may be easiest to use pipx install xyz, which will manage a
virtual environment for you. Make sure you have pipx installed.
See /usr/share/doc/python3.11/README.venv for more information.
note: If you believe this is a mistake, please contact your Python installation or OS distribution provider. You can override this, at the risk of breaking your Python installation or OS, by passing --break-system-packages.
hint: See PEP 668 for the detailed specification.
```
After searching how to install poetry I did: `curl -sSL https://install.python-poetry.org | python3 -`, added Poetry's bin directory in my PATH environment and tried again `poetry install .`
I got:
```
Creating virtualenv scrapfly-booking-tyZw0pBk-py3.11 in /home/xxxx/.cache/pypoetry/virtualenvs
No arguments expected for "install" command, got "."
```
I decided to go for step 3: `poetry run python run.py`
I got:
```
Traceback (most recent call last):
File "/home/esaumell/scrapfly-scrapers/bookingcom-scraper/run.py", line 12, in <module>
import bookingcom
File "/home/esaumell/scrapfly-scrapers/bookingcom-scraper/bookingcom.py", line 19, in <module>
from loguru import logger as log
ModuleNotFoundError: No module named 'loguru'
```
So brave and I went for step 4: `poetry install --with dev`
I got:
```
Installing dependencies from lock file
Package operations: 67 installs, 1 update, 0 removals
โข Installing certifi (2023.5.7)
โข Installing charset-normalizer (3.1.0)
โข Installing idna (3.4)
โข Installing pycparser (2.21)
โข Installing six (1.16.0)
โข Installing urllib3 (2.0.2)
โข Installing attrs (23.1.0)
โข Installing cffi (1.15.1)
โข Installing cssselect (1.2.0)
โข Installing jmespath (1.0.1)
โข Installing isodate (0.6.1)
โข Installing lxml (4.9.2)
โข Installing packaging (23.1)
โข Installing pyasn1 (0.5.0)
โข Installing pyparsing (3.0.9)
โข Installing requests (2.31.0)
โข Downgrading setuptools (68.1.2 -> 67.8.0)
โข Installing soupsieve (2.4.1)
โข Installing w3lib (2.1.1)
โข Installing webencodings (0.5.1)
โข Installing automat (22.10.0)
โข Installing beautifulsoup4 (4.12.2)
โข Installing constantly (15.1.0)
โข Installing cryptography (40.0.2)
โข Installing filelock (3.12.0)
โข Installing html5lib (1.1)
โข Installing hyperlink (21.0.0)
โข Installing incremental (22.10.0)
โข Installing itemadapter (0.8.0)
โข Installing parsel (1.8.1)
โข Installing pyasn1-modules (0.3.0)
โข Installing rdflib (6.3.2)
โข Installing requests-file (1.5.1)
โข Installing typing-extensions (4.6.1)
โข Installing zope-interface (6.0)
โข Installing html-text (0.5.2)
โข Installing iniconfig (2.0.0)
โข Installing itemloaders (1.1.0)
โข Installing jstyleson (0.0.2)
โข Installing mf2py (1.1.2)
โข Installing pluggy (1.0.0)
โข Installing protego (0.2.1)
โข Installing pydispatcher (2.0.7)
โข Installing pyopenssl (23.1.1)
โข Installing pyrdfa3 (3.5.3)
โข Installing queuelib (1.6.2)
โข Installing service-identity (21.1.0)
โข Installing tldextract (3.4.4)
โข Installing twisted (22.10.0)
โข Installing backoff (2.2.1)
โข Installing brotlipy (0.7.0)
โข Installing cchardet (2.1.7): Failed
ChefBuildError
Backend subprocess exited when trying to invoke build_wheel
running bdist_wheel
running build
running build_py
creating build
creating build/lib.linux-aarch64-cpython-311
creating build/lib.linux-aarch64-cpython-311/cchardet
copying src/cchardet/version.py -> build/lib.linux-aarch64-cpython-311/cchardet
copying src/cchardet/__init__.py -> build/lib.linux-aarch64-cpython-311/cchardet
running build_ext
building 'cchardet._cchardet' extension
creating build/temp.linux-aarch64-cpython-311
creating build/temp.linux-aarch64-cpython-311/src
creating build/temp.linux-aarch64-cpython-311/src/cchardet
creating build/temp.linux-aarch64-cpython-311/src/ext
creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet
creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src
creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src/LangModels
aarch64-linux-gnu-gcc -Wsign-compare -DNDEBUG -g -fwrapv -O2 -Wall -g -fstack-protector-strong -Wformat -Werror=format-security -g -fwrapv -O2 -fPIC -Isrc/ext/uchardet/src -I/tmp/tmp2edcn7ai/.venv/include -I/usr/include/python3.11 -c src/cchardet/_cchardet.cpp -o build/temp.linux-aarch64-cpython-311/src/cchardet/_cchardet.o
src/cchardet/_cchardet.cpp:4:10: fatal error: Python.h: No such file or directory
4 | #include "Python.h"
| ^~~~~~~~~~
compilation terminated.
error: command '/usr/bin/aarch64-linux-gnu-gcc' failed with exit code 1
at ~/.local/share/pypoetry/venv/lib/python3.11/site-packages/poetry/installation/chef.py:147 in _prepare
143โ
144โ error = ChefBuildError("\n\n".join(message_parts))
145โ
146โ if error is not None:
โ 147โ raise error from None
148โ
149โ return path
150โ
151โ def _prepare_sdist(self, archive: Path, destination: Path | None = None) -> Path:
Note: This error originates from the build backend, and is likely not a problem with poetry but with cchardet (2.1.7) not supporting PEP 517 builds. You can verify this by running 'pip wheel --use-pep517 "cchardet (==2.1.7)"'.
โข Installing click (8.1.3)
โข Installing decorator (5.1.1)
โข Installing extruct (0.14.0)
โข Installing loguru (0.7.0)
โข Installing msgpack (1.0.5)
โข Installing mypy-extensions (1.0.0)
โข Installing pathspec (0.11.1)
โข Installing platformdirs (3.5.1)
โข Installing pytest (7.3.1)
โข Installing python-dateutil (2.8.2)
โข Installing scrapy (2.9.0)
```
As stated on the output I tried `pip wheel --use-pep517 "cchardet (==2.1.7)"`
I got:
```
Collecting cchardet==2.1.7
Downloading cchardet-2.1.7.tar.gz (653 kB)
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 653.6/653.6 kB 13.5 MB/s eta 0:00:00
Installing build dependencies ... done
Getting requirements to build wheel ... done
Preparing metadata (pyproject.toml) ... done
Building wheels for collected packages: cchardet
Building wheel for cchardet (pyproject.toml) ... error
error: subprocess-exited-with-error
ร Building wheel for cchardet (pyproject.toml) did not run successfully.
โ exit code: 1
โฐโ> [23 lines of output]
running bdist_wheel
running build
running build_py
creating build
creating build/lib.linux-aarch64-cpython-311
creating build/lib.linux-aarch64-cpython-311/cchardet
copying src/cchardet/version.py -> build/lib.linux-aarch64-cpython-311/cchardet
copying src/cchardet/__init__.py -> build/lib.linux-aarch64-cpython-311/cchardet
running build_ext
building 'cchardet._cchardet' extension
creating build/temp.linux-aarch64-cpython-311
creating build/temp.linux-aarch64-cpython-311/src
creating build/temp.linux-aarch64-cpython-311/src/cchardet
creating build/temp.linux-aarch64-cpython-311/src/ext
creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet
creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src
creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src/LangModels
aarch64-linux-gnu-gcc -Wsign-compare -DNDEBUG -g -fwrapv -O2 -Wall -g -fstack-protector-strong -Wformat -Werror=format-security -g -fwrapv -O2 -fPIC -Isrc/ext/uchardet/src -I/usr/include/python3.11 -c src/cchardet/_cchardet.cpp -o build/temp.linux-aarch64-cpython-311/src/cchardet/_cchardet.o
src/cchardet/_cchardet.cpp:4:10: fatal error: Python.h: No such file or directory
4 | #include "Python.h"
| ^~~~~~~~~~
compilation terminated.
error: command '/usr/bin/aarch64-linux-gnu-gcc' failed with exit code 1
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building wheel for cchardet
Failed to build cchardet
ERROR: Failed to build one or more wheels
```
This is running on a fresh Debian 12 setup and I have installed poetry 1.6.1
Searching for the problem I'm suggested to try to downgrade poetry, so I try `poetry self update 1.4` but `poetry install --with dev` fails again on cchardet
I also tried installing `python3-dev` without success. I don't know where to go from here. Any help would be appreciated. | non_priority | trouble following instructions hi there i m having trouble following instructions from at point when trying git clone git github com scrapfly scrapfly scrapers git i got git github com permission denied publickey fatal could not read from remote repository please make sure you have the correct access rights and the repository exists as a workaround i did git clone next step is poetry install i got bash poetry command not found i go to and as stated i try pip install httpx parsel i got bash pip command not found as a workaround i did sudo apt install pip and then pip install httpx parsel i got error externally managed environment ร this environment is externally managed โฐโ to install python packages system wide try apt install xyz where xyz is the package you are trying to install if you wish to install a non debian packaged python package create a virtual environment using m venv path to venv then use path to venv bin python and path to venv bin pip make sure you have full installed if you wish to install a non debian packaged python application it may be easiest to use pipx install xyz which will manage a virtual environment for you make sure you have pipx installed see usr share doc readme venv for more information note if you believe this is a mistake please contact your python installation or os distribution provider you can override this at the risk of breaking your python installation or os by passing break system packages hint see pep for the detailed specification after searching how to install poetry i did curl ssl added poetry s bin directory in my path environment and tried again poetry install i got creating virtualenv scrapfly booking in home xxxx cache pypoetry virtualenvs no arguments expected for install command got i decided to go for step poetry run python run py i got traceback most recent call last file home esaumell scrapfly scrapers bookingcom scraper run py line in import bookingcom file home esaumell scrapfly scrapers bookingcom scraper bookingcom py line in from loguru import logger as log modulenotfounderror no module named loguru so brave and i went for step poetry install with dev i got installing dependencies from lock file package operations installs update removals โข installing certifi โข installing charset normalizer โข installing idna โข installing pycparser โข installing six โข installing โข installing attrs โข installing cffi โข installing cssselect โข installing jmespath โข installing isodate โข installing lxml โข installing packaging โข installing โข installing pyparsing โข installing requests โข downgrading setuptools โข installing soupsieve โข installing โข installing webencodings โข installing automat โข installing โข installing constantly โข installing cryptography โข installing filelock โข installing โข installing hyperlink โข installing incremental โข installing itemadapter โข installing parsel โข installing modules โข installing rdflib โข installing requests file โข installing typing extensions โข installing zope interface โข installing html text โข installing iniconfig โข installing itemloaders โข installing jstyleson โข installing โข installing pluggy โข installing protego โข installing pydispatcher โข installing pyopenssl โข installing โข installing queuelib โข installing service identity โข installing tldextract โข installing twisted โข installing backoff โข installing brotlipy โข installing cchardet failed chefbuilderror backend subprocess exited when trying to invoke build wheel running bdist wheel running build running build py creating build creating build lib linux cpython creating build lib linux cpython cchardet copying src cchardet version py build lib linux cpython cchardet copying src cchardet init py build lib linux cpython cchardet running build ext building cchardet cchardet extension creating build temp linux cpython creating build temp linux cpython src creating build temp linux cpython src cchardet creating build temp linux cpython src ext creating build temp linux cpython src ext uchardet creating build temp linux cpython src ext uchardet src creating build temp linux cpython src ext uchardet src langmodels linux gnu gcc wsign compare dndebug g fwrapv wall g fstack protector strong wformat werror format security g fwrapv fpic isrc ext uchardet src i tmp venv include i usr include c src cchardet cchardet cpp o build temp linux cpython src cchardet cchardet o src cchardet cchardet cpp fatal error python h no such file or directory include python h compilation terminated error command usr bin linux gnu gcc failed with exit code at local share pypoetry venv lib site packages poetry installation chef py in prepare โ โ error chefbuilderror n n join message parts โ โ if error is not none โ โ raise error from none โ โ return path โ โ def prepare sdist self archive path destination path none none path note this error originates from the build backend and is likely not a problem with poetry but with cchardet not supporting pep builds you can verify this by running pip wheel use cchardet โข installing click โข installing decorator โข installing extruct โข installing loguru โข installing msgpack โข installing mypy extensions โข installing pathspec โข installing platformdirs โข installing pytest โข installing python dateutil โข installing scrapy as stated on the output i tried pip wheel use cchardet i got collecting cchardet downloading cchardet tar gz kb โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ kb mb s eta installing build dependencies done getting requirements to build wheel done preparing metadata pyproject toml done building wheels for collected packages cchardet building wheel for cchardet pyproject toml error error subprocess exited with error ร building wheel for cchardet pyproject toml did not run successfully โ exit code โฐโ running bdist wheel running build running build py creating build creating build lib linux cpython creating build lib linux cpython cchardet copying src cchardet version py build lib linux cpython cchardet copying src cchardet init py build lib linux cpython cchardet running build ext building cchardet cchardet extension creating build temp linux cpython creating build temp linux cpython src creating build temp linux cpython src cchardet creating build temp linux cpython src ext creating build temp linux cpython src ext uchardet creating build temp linux cpython src ext uchardet src creating build temp linux cpython src ext uchardet src langmodels linux gnu gcc wsign compare dndebug g fwrapv wall g fstack protector strong wformat werror format security g fwrapv fpic isrc ext uchardet src i usr include c src cchardet cchardet cpp o build temp linux cpython src cchardet cchardet o src cchardet cchardet cpp fatal error python h no such file or directory include python h compilation terminated error command usr bin linux gnu gcc failed with exit code note this error originates from a subprocess and is likely not a problem with pip error failed building wheel for cchardet failed to build cchardet error failed to build one or more wheels this is running on a fresh debian setup and i have installed poetry searching for the problem i m suggested to try to downgrade poetry so i try poetry self update but poetry install with dev fails again on cchardet i also tried installing dev without success i don t know where to go from here any help would be appreciated | 0 |
155,166 | 13,613,509,672 | IssuesEvent | 2020-09-23 11:59:59 | fluid-project/floeproject.org | https://api.github.com/repos/fluid-project/floeproject.org | closed | Add screen reader or other AT details to the Technical Details section | documentation enhancement | ## Describe the solution you'd like
Add a section to list ATs and their versions in the technical details section of a bug report. | 1.0 | Add screen reader or other AT details to the Technical Details section - ## Describe the solution you'd like
Add a section to list ATs and their versions in the technical details section of a bug report. | non_priority | add screen reader or other at details to the technical details section describe the solution you d like add a section to list ats and their versions in the technical details section of a bug report | 0 |
189,075 | 22,046,977,735 | IssuesEvent | 2022-05-30 03:38:49 | praneethpanasala/linux | https://api.github.com/repos/praneethpanasala/linux | closed | CVE-2019-19534 (Low) detected in linuxlinux-4.19.6 - autoclosed | security vulnerability | ## CVE-2019-19534 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/praneethpanasala/linux/commits/d80c4f847c91020292cb280132b15e2ea147f1a3">d80c4f847c91020292cb280132b15e2ea147f1a3</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/can/usb/peak_usb/pcan_usb_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/can/usb/peak_usb/pcan_usb_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel before 5.3.11, there is an info-leak bug that can be caused by a malicious USB device in the drivers/net/can/usb/peak_usb/pcan_usb_core.c driver, aka CID-f7a1337f0d29.
<p>Publish Date: 2019-12-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19534>CVE-2019-19534</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2019-19534">https://www.linuxkernelcves.com/cves/CVE-2019-19534</a></p>
<p>Release Date: 2019-12-03</p>
<p>Fix Resolution: v5.4-rc7,v3.16.79,v4.14.154,v4.19.84,v4.4.201,v4.9.201,v5.3.11</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-19534 (Low) detected in linuxlinux-4.19.6 - autoclosed - ## CVE-2019-19534 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/praneethpanasala/linux/commits/d80c4f847c91020292cb280132b15e2ea147f1a3">d80c4f847c91020292cb280132b15e2ea147f1a3</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/can/usb/peak_usb/pcan_usb_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/can/usb/peak_usb/pcan_usb_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel before 5.3.11, there is an info-leak bug that can be caused by a malicious USB device in the drivers/net/can/usb/peak_usb/pcan_usb_core.c driver, aka CID-f7a1337f0d29.
<p>Publish Date: 2019-12-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19534>CVE-2019-19534</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2019-19534">https://www.linuxkernelcves.com/cves/CVE-2019-19534</a></p>
<p>Release Date: 2019-12-03</p>
<p>Fix Resolution: v5.4-rc7,v3.16.79,v4.14.154,v4.19.84,v4.4.201,v4.9.201,v5.3.11</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve low detected in linuxlinux autoclosed cve low severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files drivers net can usb peak usb pcan usb core c drivers net can usb peak usb pcan usb core c vulnerability details in the linux kernel before there is an info leak bug that can be caused by a malicious usb device in the drivers net can usb peak usb pcan usb core c driver aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
644,732 | 20,986,104,710 | IssuesEvent | 2022-03-29 03:27:01 | EspressoSystems/cape | https://api.github.com/repos/EspressoSystems/cape | opened | Wallet :name parameter | bug priority: high wallet security wallet API unplanned | The `:name` parameter is used directly as the wallet file name, so this should be changed to something assured to be a valid file name, e.g. no path separators and length supported across platforms Linux, MacOS, and Windows. | 1.0 | Wallet :name parameter - The `:name` parameter is used directly as the wallet file name, so this should be changed to something assured to be a valid file name, e.g. no path separators and length supported across platforms Linux, MacOS, and Windows. | priority | wallet name parameter the name parameter is used directly as the wallet file name so this should be changed to something assured to be a valid file name e g no path separators and length supported across platforms linux macos and windows | 1 |
18,174 | 10,889,362,513 | IssuesEvent | 2019-11-18 18:04:53 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | Preliminary research | Epic Impact: 4-None Project: Warehouse Inventory Service: PM Type: Research Workgroup: AMD Workgroup: SMB | ### [Warehouse Inventory Application comparison matrix](https://docs.google.com/spreadsheets/d/1r2l_KFgjygCCp1E4oaBa5lyK2QzkVcP5qCu2u85UwcY/edit#gid=2118741241)
### Candidates
[Wasp Inventory Control](https://www.waspbarcode.com/inventory-control)
- [x] Form submitted
- [x] Phone call #173
- [x] Evaluate trial account #174
[Finale Inventory](https://www.finaleinventory.com/)
- [x] Email inquiry with basic requirements & demo request
- [x] [Talk with sales rep](https://github.com/cityofaustin/atd-data-tech/issues/156#issuecomment-499175173)
- [x] Evaluate trial account #176
[Asset Panda](https://www.assetpanda.com/asset-tracking-government-2/)
- [x] Form submitted
- [x] [Talk with sales rep](https://github.com/cityofaustin/atd-data-tech/issues/156#issuecomment-499284953)
- [x] Evaluate trial account #178
[Dude Solutions](https://www.dudesolutions.com/solutions/work-asset)
- [x] Call with sales rep
- [x] Demo #233
------
### Ruled out
[Cin7](https://www.cin7.com/) โ Expensive, too tailored to ecommerce
[SKU Vault](https://www.skuvault.com/) โ Expensive, too tailored to ecommerce
[Infoplus](https://www.infopluscommerce.com/) โ Expensive, too tailored to ecommerce
[Grainger](https://www.grainger.com/content/keepstock) โ Only tracks Grainger products
| 1.0 | Preliminary research - ### [Warehouse Inventory Application comparison matrix](https://docs.google.com/spreadsheets/d/1r2l_KFgjygCCp1E4oaBa5lyK2QzkVcP5qCu2u85UwcY/edit#gid=2118741241)
### Candidates
[Wasp Inventory Control](https://www.waspbarcode.com/inventory-control)
- [x] Form submitted
- [x] Phone call #173
- [x] Evaluate trial account #174
[Finale Inventory](https://www.finaleinventory.com/)
- [x] Email inquiry with basic requirements & demo request
- [x] [Talk with sales rep](https://github.com/cityofaustin/atd-data-tech/issues/156#issuecomment-499175173)
- [x] Evaluate trial account #176
[Asset Panda](https://www.assetpanda.com/asset-tracking-government-2/)
- [x] Form submitted
- [x] [Talk with sales rep](https://github.com/cityofaustin/atd-data-tech/issues/156#issuecomment-499284953)
- [x] Evaluate trial account #178
[Dude Solutions](https://www.dudesolutions.com/solutions/work-asset)
- [x] Call with sales rep
- [x] Demo #233
------
### Ruled out
[Cin7](https://www.cin7.com/) โ Expensive, too tailored to ecommerce
[SKU Vault](https://www.skuvault.com/) โ Expensive, too tailored to ecommerce
[Infoplus](https://www.infopluscommerce.com/) โ Expensive, too tailored to ecommerce
[Grainger](https://www.grainger.com/content/keepstock) โ Only tracks Grainger products
| non_priority | preliminary research candidates form submitted phone call evaluate trial account email inquiry with basic requirements demo request evaluate trial account form submitted evaluate trial account call with sales rep demo ruled out โ expensive too tailored to ecommerce โ expensive too tailored to ecommerce โ expensive too tailored to ecommerce โ only tracks grainger products | 0 |
66,124 | 6,989,255,021 | IssuesEvent | 2017-12-14 15:38:44 | khartec/waltz | https://api.github.com/repos/khartec/waltz | closed | Increase Bookmark.URL field length | DDL change fixed (test & close) small change | Currently is it 255, suggestion is to increase to 500 to accommodate longer URLs | 1.0 | Increase Bookmark.URL field length - Currently is it 255, suggestion is to increase to 500 to accommodate longer URLs | non_priority | increase bookmark url field length currently is it suggestion is to increase to to accommodate longer urls | 0 |
17,056 | 2,615,129,869 | IssuesEvent | 2015-03-01 05:59:38 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | closed | setPlusRequestInitializer | auto-migrated Component-Google-APIs Milestone-Version1.12.0 Priority-High Type-Enhancement | ```
External references, such as a standards document, or specification?
Java environments (e.g. Java 6, Android 2.3, App Engine, or All)?
All
Please describe the feature requested.
Here's your typical builder structure:
Plus plus =
new Plus.Builder(httpTransport, jsonFactory, null).setGoogleClientRequestInitializer(
new PlusRequestInitializer(API_KEY)).build();
But it would be easier to use if it were:
Plus plus =
new Plus.Builder(httpTransport, jsonFactory, null).setPlusRequestInitializer(
new PlusRequestInitializer(API_KEY)).build();
In other words, add a setPlusRequestInitializer that looks something like:
public Builder setPlusRequestInitializer(PlusRequestInitializer initializer) {
return (Builder) setGoogleClientRequestInitializer(initializer);
}
Note that we'd still keep setGoogleClientRequestInitializer.
```
Original issue reported on code.google.com by `yan...@google.com` on 31 Oct 2012 at 4:10 | 1.0 | setPlusRequestInitializer - ```
External references, such as a standards document, or specification?
Java environments (e.g. Java 6, Android 2.3, App Engine, or All)?
All
Please describe the feature requested.
Here's your typical builder structure:
Plus plus =
new Plus.Builder(httpTransport, jsonFactory, null).setGoogleClientRequestInitializer(
new PlusRequestInitializer(API_KEY)).build();
But it would be easier to use if it were:
Plus plus =
new Plus.Builder(httpTransport, jsonFactory, null).setPlusRequestInitializer(
new PlusRequestInitializer(API_KEY)).build();
In other words, add a setPlusRequestInitializer that looks something like:
public Builder setPlusRequestInitializer(PlusRequestInitializer initializer) {
return (Builder) setGoogleClientRequestInitializer(initializer);
}
Note that we'd still keep setGoogleClientRequestInitializer.
```
Original issue reported on code.google.com by `yan...@google.com` on 31 Oct 2012 at 4:10 | priority | setplusrequestinitializer external references such as a standards document or specification java environments e g java android app engine or all all please describe the feature requested here s your typical builder structure plus plus new plus builder httptransport jsonfactory null setgoogleclientrequestinitializer new plusrequestinitializer api key build but it would be easier to use if it were plus plus new plus builder httptransport jsonfactory null setplusrequestinitializer new plusrequestinitializer api key build in other words add a setplusrequestinitializer that looks something like public builder setplusrequestinitializer plusrequestinitializer initializer return builder setgoogleclientrequestinitializer initializer note that we d still keep setgoogleclientrequestinitializer original issue reported on code google com by yan google com on oct at | 1 |
81,480 | 3,591,433,657 | IssuesEvent | 2016-02-01 11:45:04 | sa31fh/ims | https://api.github.com/repos/sa31fh/ims | closed | Expected sales keeps moving down | bug High Priority | On category_status.php, if I click on one of the categories e.g. 'Drinks' for three or more times, the 'Expected Sales' field below the table starts moving down one line with every click on a category. It doesn't matter which category/ies are clicked on and in which order. | 1.0 | Expected sales keeps moving down - On category_status.php, if I click on one of the categories e.g. 'Drinks' for three or more times, the 'Expected Sales' field below the table starts moving down one line with every click on a category. It doesn't matter which category/ies are clicked on and in which order. | priority | expected sales keeps moving down on category status php if i click on one of the categories e g drinks for three or more times the expected sales field below the table starts moving down one line with every click on a category it doesn t matter which category ies are clicked on and in which order | 1 |
700,879 | 24,076,569,356 | IssuesEvent | 2022-09-18 21:45:43 | json-schema-org/json-schema-spec | https://api.github.com/repos/json-schema-org/json-schema-spec | closed | Review the concept of how annotations can impact validation | Priority: High core annotation | > I'm wondering if the entire concept of annotations should be revised to something different - @karenetheridge
> This is the direction I'd like to go. - @jdesrosiers
> I'm happy to do that, too, but not in a patch to this version of the spec. Let's open a new issue to discuss that for the next one. - @gregsdennis
_Originally posted in https://github.com/json-schema-org/json-schema-spec/issues/1172#issuecomment-1068698286_
This may be best discussed in a [GH Discussion](https://github.com/json-schema-org/community/discussions). One should be created. | 1.0 | Review the concept of how annotations can impact validation - > I'm wondering if the entire concept of annotations should be revised to something different - @karenetheridge
> This is the direction I'd like to go. - @jdesrosiers
> I'm happy to do that, too, but not in a patch to this version of the spec. Let's open a new issue to discuss that for the next one. - @gregsdennis
_Originally posted in https://github.com/json-schema-org/json-schema-spec/issues/1172#issuecomment-1068698286_
This may be best discussed in a [GH Discussion](https://github.com/json-schema-org/community/discussions). One should be created. | priority | review the concept of how annotations can impact validation i m wondering if the entire concept of annotations should be revised to something different karenetheridge this is the direction i d like to go jdesrosiers i m happy to do that too but not in a patch to this version of the spec let s open a new issue to discuss that for the next one gregsdennis originally posted in this may be best discussed in a one should be created | 1 |
255,551 | 8,125,417,354 | IssuesEvent | 2018-08-16 20:52:14 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | StreetLight shadows missing | Art Medium Priority | Looks like cast shadows got turned off on them for some reason?

| 1.0 | StreetLight shadows missing - Looks like cast shadows got turned off on them for some reason?

| priority | streetlight shadows missing looks like cast shadows got turned off on them for some reason | 1 |
187,742 | 6,760,834,149 | IssuesEvent | 2017-10-24 22:10:11 | b3aver/Automate | https://api.github.com/repos/b3aver/Automate | opened | In viewMode show Actions without the input fields | enhancement priority:minor time:medium topic:ui | Insert span tags with the saved informations.
Substitute also the select fields. | 1.0 | In viewMode show Actions without the input fields - Insert span tags with the saved informations.
Substitute also the select fields. | priority | in viewmode show actions without the input fields insert span tags with the saved informations substitute also the select fields | 1 |
665,748 | 22,327,558,311 | IssuesEvent | 2022-06-14 12:02:30 | laurencedawson/sync-for-reddit | https://api.github.com/repos/laurencedawson/sync-for-reddit | closed | [BUG] When a bottomsheet dialog is shown and the display is off, it doesn't theme | bug low-priority | For example if the app shows the updated dialog while the screen is off it may use the system coloring... | 1.0 | [BUG] When a bottomsheet dialog is shown and the display is off, it doesn't theme - For example if the app shows the updated dialog while the screen is off it may use the system coloring... | priority | when a bottomsheet dialog is shown and the display is off it doesn t theme for example if the app shows the updated dialog while the screen is off it may use the system coloring | 1 |
255,113 | 8,108,882,352 | IssuesEvent | 2018-08-14 04:35:30 | RhoInc/web-codebook | https://api.github.com/repos/RhoInc/web-codebook | closed | Use fixed width font to prevent y-axis label text from getting clipped. | invalid low priority | @brittsikora noticed in testing #62 | 1.0 | Use fixed width font to prevent y-axis label text from getting clipped. - @brittsikora noticed in testing #62 | priority | use fixed width font to prevent y axis label text from getting clipped brittsikora noticed in testing | 1 |
379,060 | 26,351,440,800 | IssuesEvent | 2023-01-11 05:27:12 | tossicat/tossicat-core | https://api.github.com/repos/tossicat/tossicat-core | opened | `README.md`์ ์ด ํ๋ก์ ํธ ์ฅ์ ์ถ๊ฐ | documentation | ํ์ฌ ์ด ํ๋ก์ ํธ๋ง์ ์ฅ์ ์ ์๊ฐํ๋ ๋ถ๋ถ์ด ์์ต๋๋ค. ์ ๊ฐ ์๊ฐํ๊ธฐ์๋ ๋ค์๊ณผ ๊ฐ์ ์ฅ์ ์ด ์๋ค๊ณ ์๊ฐํฉ๋๋ค. ๋ฌผ๋ก ์์ง๊น์ง ์๋ฒฝํ์ง๋ ์์ง๋ง, ๊ณง ์๋ ๋ด์ฉ์ ๋ค ๊ตฌํํ ์ ์๋ค๊ณ ์๊ฐํฉ๋๋ค. ๋ ์ถ๊ฐํ ๋ถ๋ถ์ด๋ ์๋ ๋ด์ฉ์์ ๊ณ ์น ๋ถ๋ถ์ด ์๋ค๋ฉด ์๋์ ์จ์ฃผ์ธ์.
```md
## ์ด ํ๋ก์ ํธ์ ์ฅ์ (์์ง ๋ค ๊ตฌํํ)
1. ๋ฌ์คํธ๋ก ๋ง๋ค์ด์ ๋น ๋ฆ
๋๋ค.
2. ๋ณํํ ๊ฒ๋ค์ด ์ฌ๋ฌ ๊ฐ ํฌํจ๋ ๋ฌธ์ฅ๋ ์ ์ ํ ํ์์ผ๋ก ์
๋ ฅํ์๋ฉด ํ ๋ฒ์ ์ฒ๋ฆฌํ ์ ์์ต๋๋ค.
3. ์ ๋ณํ์ด ๋์ง ์๋ ์ ์ ํ๊ฒ ์๋ฌ ๋ฉ์ธ์ง๋ฅผ ๋ฐํํฉ๋๋ค.
``` | 1.0 | `README.md`์ ์ด ํ๋ก์ ํธ ์ฅ์ ์ถ๊ฐ - ํ์ฌ ์ด ํ๋ก์ ํธ๋ง์ ์ฅ์ ์ ์๊ฐํ๋ ๋ถ๋ถ์ด ์์ต๋๋ค. ์ ๊ฐ ์๊ฐํ๊ธฐ์๋ ๋ค์๊ณผ ๊ฐ์ ์ฅ์ ์ด ์๋ค๊ณ ์๊ฐํฉ๋๋ค. ๋ฌผ๋ก ์์ง๊น์ง ์๋ฒฝํ์ง๋ ์์ง๋ง, ๊ณง ์๋ ๋ด์ฉ์ ๋ค ๊ตฌํํ ์ ์๋ค๊ณ ์๊ฐํฉ๋๋ค. ๋ ์ถ๊ฐํ ๋ถ๋ถ์ด๋ ์๋ ๋ด์ฉ์์ ๊ณ ์น ๋ถ๋ถ์ด ์๋ค๋ฉด ์๋์ ์จ์ฃผ์ธ์.
```md
## ์ด ํ๋ก์ ํธ์ ์ฅ์ (์์ง ๋ค ๊ตฌํํ)
1. ๋ฌ์คํธ๋ก ๋ง๋ค์ด์ ๋น ๋ฆ
๋๋ค.
2. ๋ณํํ ๊ฒ๋ค์ด ์ฌ๋ฌ ๊ฐ ํฌํจ๋ ๋ฌธ์ฅ๋ ์ ์ ํ ํ์์ผ๋ก ์
๋ ฅํ์๋ฉด ํ ๋ฒ์ ์ฒ๋ฆฌํ ์ ์์ต๋๋ค.
3. ์ ๋ณํ์ด ๋์ง ์๋ ์ ์ ํ๊ฒ ์๋ฌ ๋ฉ์ธ์ง๋ฅผ ๋ฐํํฉ๋๋ค.
``` | non_priority | readme md ์ ์ด ํ๋ก์ ํธ ์ฅ์ ์ถ๊ฐ ํ์ฌ ์ด ํ๋ก์ ํธ๋ง์ ์ฅ์ ์ ์๊ฐํ๋ ๋ถ๋ถ์ด ์์ต๋๋ค ์ ๊ฐ ์๊ฐํ๊ธฐ์๋ ๋ค์๊ณผ ๊ฐ์ ์ฅ์ ์ด ์๋ค๊ณ ์๊ฐํฉ๋๋ค ๋ฌผ๋ก ์์ง๊น์ง ์๋ฒฝํ์ง๋ ์์ง๋ง ๊ณง ์๋ ๋ด์ฉ์ ๋ค ๊ตฌํํ ์ ์๋ค๊ณ ์๊ฐํฉ๋๋ค ๋ ์ถ๊ฐํ ๋ถ๋ถ์ด๋ ์๋ ๋ด์ฉ์์ ๊ณ ์น ๋ถ๋ถ์ด ์๋ค๋ฉด ์๋์ ์จ์ฃผ์ธ์ md ์ด ํ๋ก์ ํธ์ ์ฅ์ ์์ง ๋ค ๊ตฌํํ ๋ฌ์คํธ๋ก ๋ง๋ค์ด์ ๋น ๋ฆ
๋๋ค ๋ณํํ ๊ฒ๋ค์ด ์ฌ๋ฌ ๊ฐ ํฌํจ๋ ๋ฌธ์ฅ๋ ์ ์ ํ ํ์์ผ๋ก ์
๋ ฅํ์๋ฉด ํ ๋ฒ์ ์ฒ๋ฆฌํ ์ ์์ต๋๋ค ์ ๋ณํ์ด ๋์ง ์๋ ์ ์ ํ๊ฒ ์๋ฌ ๋ฉ์ธ์ง๋ฅผ ๋ฐํํฉ๋๋ค | 0 |
75,831 | 21,007,491,702 | IssuesEvent | 2022-03-30 01:00:17 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | Build fails to make use of bazelisk, if available. | type:build/install subtype:windows TF 2.8 | <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): WINDOWS
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow installed from (source or binary): https://github.com/tensorflow/tensorflow.git
- TensorFlow version: r2.
- Python version: 3.10
- Installed using virtualenv? pip? conda?: GIT
- Bazel version (if compiling from source): 5.0
- GCC/Compiler version (if compiling from source): 11.2
- CUDA/cuDNN version: NA
- GPU model and memory: NA
**Describe the problem**
When using the [build from source](https://www.tensorflow.org/install/source_windows) https://www.tensorflow.org/install/source_windows , Even when 'bazelisk' is installed, the install run fails from config.py stating
'Cannot find bazel. Please install bazel.',
This could easly be fixed, like
```
if bazel_executable is None:
bazel_executable = which('bazelisk')
print('bazel_executable not found, instead using bazelisk @ ',bazel_executable)
if bazel_executable is None:
print('Cannot find bazel. Please install bazel.')
sys.exit(1)
```
**Provide the exact sequence of commands / steps that you executed before running into the problem**
Remove from path variable /Un Install bazel.
Install bazelisk.
execute ./config or python config.py
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
**Error**: 'Cannot find bazel. Please install bazel.'
| 1.0 | Build fails to make use of bazelisk, if available. - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): WINDOWS
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow installed from (source or binary): https://github.com/tensorflow/tensorflow.git
- TensorFlow version: r2.
- Python version: 3.10
- Installed using virtualenv? pip? conda?: GIT
- Bazel version (if compiling from source): 5.0
- GCC/Compiler version (if compiling from source): 11.2
- CUDA/cuDNN version: NA
- GPU model and memory: NA
**Describe the problem**
When using the [build from source](https://www.tensorflow.org/install/source_windows) https://www.tensorflow.org/install/source_windows , Even when 'bazelisk' is installed, the install run fails from config.py stating
'Cannot find bazel. Please install bazel.',
This could easly be fixed, like
```
if bazel_executable is None:
bazel_executable = which('bazelisk')
print('bazel_executable not found, instead using bazelisk @ ',bazel_executable)
if bazel_executable is None:
print('Cannot find bazel. Please install bazel.')
sys.exit(1)
```
**Provide the exact sequence of commands / steps that you executed before running into the problem**
Remove from path variable /Un Install bazel.
Install bazelisk.
execute ./config or python config.py
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
**Error**: 'Cannot find bazel. Please install bazel.'
| non_priority | build fails to make use of bazelisk if available please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution e g linux ubuntu windows mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device tensorflow installed from source or binary tensorflow version python version installed using virtualenv pip conda git bazel version if compiling from source gcc compiler version if compiling from source cuda cudnn version na gpu model and memory na describe the problem when using the even when bazelisk is installed the install run fails from config py stating cannot find bazel please install bazel this could easly be fixed like if bazel executable is none bazel executable which bazelisk print bazel executable not found instead using bazelisk bazel executable if bazel executable is none print cannot find bazel please install bazel sys exit provide the exact sequence of commands steps that you executed before running into the problem remove from path variable un install bazel install bazelisk execute config or python config py any other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached error cannot find bazel please install bazel | 0 |
150,461 | 5,773,143,350 | IssuesEvent | 2017-04-28 00:40:08 | coreos/flannel | https://api.github.com/repos/coreos/flannel | closed | Add consistent flannel configuration. | area/configuration help wanted priority/Pmaybe | Hi,
it is extremely annoying to configure flannel from cloud-config. When doing flannel configuration, you have to think about YAML, JSON, SHELL syntaxes and catch bugs in different files.
Please add consistent configuration for flannel in single file.
| 1.0 | Add consistent flannel configuration. - Hi,
it is extremely annoying to configure flannel from cloud-config. When doing flannel configuration, you have to think about YAML, JSON, SHELL syntaxes and catch bugs in different files.
Please add consistent configuration for flannel in single file.
| priority | add consistent flannel configuration hi it is extremely annoying to configure flannel from cloud config when doing flannel configuration you have to think about yaml json shell syntaxes and catch bugs in different files please add consistent configuration for flannel in single file | 1 |
174,424 | 14,481,915,642 | IssuesEvent | 2020-12-10 13:18:18 | OpenEnergyPlatform/open-MaStR | https://api.github.com/repos/OpenEnergyPlatform/open-MaStR | closed | streamline docs | :book: documentation | ### Tasks
- [x] Move README content partially to RTD
- [x] Think about including remaining README into index.rst
- [x] Sort content in sections download, post-processing, analyze with content in getting started | 1.0 | streamline docs - ### Tasks
- [x] Move README content partially to RTD
- [x] Think about including remaining README into index.rst
- [x] Sort content in sections download, post-processing, analyze with content in getting started | non_priority | streamline docs tasks move readme content partially to rtd think about including remaining readme into index rst sort content in sections download post processing analyze with content in getting started | 0 |
502,599 | 14,562,721,282 | IssuesEvent | 2020-12-17 00:44:17 | Sage-Bionetworks/sageseqr | https://api.github.com/repos/Sage-Bionetworks/sageseqr | closed | Use githubr to version control markdown | low priority wontfix | The `githubr` package can point to a specific commit of a file in a repository. The package loads the markdown code into the users working directory and is currently hard coded as "sageseqr-report.Rmd". To track the correct version via a commit using `githubr`, the user would need to push the changed file to their repo before running the plan. This is difficult to message to the user and may over complicate the workflow.
On the other hand, it might be possible to either `push` the file as a step in the workflow if the repository argument is provided.
For now - marking this as low priority since the `sageseqr` package version is tracked in the activity provenance and I think the general use case for the workflow will be to run it with minimal modifications to the markdown document. | 1.0 | Use githubr to version control markdown - The `githubr` package can point to a specific commit of a file in a repository. The package loads the markdown code into the users working directory and is currently hard coded as "sageseqr-report.Rmd". To track the correct version via a commit using `githubr`, the user would need to push the changed file to their repo before running the plan. This is difficult to message to the user and may over complicate the workflow.
On the other hand, it might be possible to either `push` the file as a step in the workflow if the repository argument is provided.
For now - marking this as low priority since the `sageseqr` package version is tracked in the activity provenance and I think the general use case for the workflow will be to run it with minimal modifications to the markdown document. | priority | use githubr to version control markdown the githubr package can point to a specific commit of a file in a repository the package loads the markdown code into the users working directory and is currently hard coded as sageseqr report rmd to track the correct version via a commit using githubr the user would need to push the changed file to their repo before running the plan this is difficult to message to the user and may over complicate the workflow on the other hand it might be possible to either push the file as a step in the workflow if the repository argument is provided for now marking this as low priority since the sageseqr package version is tracked in the activity provenance and i think the general use case for the workflow will be to run it with minimal modifications to the markdown document | 1 |
468,315 | 13,465,137,065 | IssuesEvent | 2020-09-09 20:21:32 | googleapis/python-bigquery | https://api.github.com/repos/googleapis/python-bigquery | closed | Using string array with magic block results in whitespace removal from first entry of the array | api: bigquery priority: p2 type: bug | On passing a dict ` { 'cats': ['apple orange', 'pear plum']}` as params in bigquery magic cell, the first value is changed to `appleorange` - the space character is filtered out.
#### Environment details
- OS type and version: colab
- Python version: Python 3.6.9
- pip version: 19.3.1
- `google-cloud-bigquery` version: 1.21.0
#### Steps to reproduce
A colab notebook illustrating the error:
https://colab.research.google.com/gist/amardeep/63ec303ba8bac3db9849f4044cd19ff1/test-bigquery-array-parameter-bug.ipynb
#### Code example
```python
params = {
'cats': ['apple orange', 'pear plum']
}
%%bigquery --params $params
SELECT * FROM UNNEST(@cats)
```
This results in the output:
```
f0_
0 appleorange
1 pear plum
```
| 1.0 | Using string array with magic block results in whitespace removal from first entry of the array - On passing a dict ` { 'cats': ['apple orange', 'pear plum']}` as params in bigquery magic cell, the first value is changed to `appleorange` - the space character is filtered out.
#### Environment details
- OS type and version: colab
- Python version: Python 3.6.9
- pip version: 19.3.1
- `google-cloud-bigquery` version: 1.21.0
#### Steps to reproduce
A colab notebook illustrating the error:
https://colab.research.google.com/gist/amardeep/63ec303ba8bac3db9849f4044cd19ff1/test-bigquery-array-parameter-bug.ipynb
#### Code example
```python
params = {
'cats': ['apple orange', 'pear plum']
}
%%bigquery --params $params
SELECT * FROM UNNEST(@cats)
```
This results in the output:
```
f0_
0 appleorange
1 pear plum
```
| priority | using string array with magic block results in whitespace removal from first entry of the array on passing a dict cats as params in bigquery magic cell the first value is changed to appleorange the space character is filtered out environment details os type and version colab python version python pip version google cloud bigquery version steps to reproduce a colab notebook illustrating the error code example python params cats bigquery params params select from unnest cats this results in the output appleorange pear plum | 1 |
689,006 | 23,603,752,717 | IssuesEvent | 2022-08-24 06:13:49 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Missing information on the Dry run report generation configuration in migration guide | Priority/Low docs Severity/Major Component/Migration 6.0.0-Migration Affected-6.0.0 QA-Reported 5.11-migration 6.0.0-post-ga | **Describe the issue:**
During a Migration testing if you want to generate dry run reports that can be done from the migration-config.yaml. But migration guide doesnt include the correct steps with example on how to do that.
https://is.docs.wso2.com/en/latest/setup/migrating-userstore-managers/#how-to-run Dry Run section
To generate dry run reports we have to pass the reportPath as a string value and this reportPath always needs to be defined under a **parameter** config. If we dont pass it as a string or dont include it under a **parameter** config report generatio will not get success and you will end up in error saying migration process stopped.
So lets include a sample config for the correct way of doing it
<IS_HOME>/migration-resources/migration-config.yaml
```
name: "GroupsAndRolesMigrator"
order: 4
parameters:
# Absolute path for the dry report. This is required in the dry run mode.
reportPath: "/home/ubuntu/WSO2/IS600/MigrationTesting/reports"
```
- [x] Mention you need to pass the reportPath as a string value
- [x] Mention you need to define reportPath under a **parameter** config
- [x] Mention that you need to uncomment and include reportPath in the migration-config.yaml uner each migrator that you wants to get the reports generated ( At the moment there are 5 occurences in the migration-config.yaml to uncomment)
Attaching sample config file for reference
[resources.zip](https://github.com/wso2/product-is/files/8888114/resources.zip)
| 1.0 | Missing information on the Dry run report generation configuration in migration guide - **Describe the issue:**
During a Migration testing if you want to generate dry run reports that can be done from the migration-config.yaml. But migration guide doesnt include the correct steps with example on how to do that.
https://is.docs.wso2.com/en/latest/setup/migrating-userstore-managers/#how-to-run Dry Run section
To generate dry run reports we have to pass the reportPath as a string value and this reportPath always needs to be defined under a **parameter** config. If we dont pass it as a string or dont include it under a **parameter** config report generatio will not get success and you will end up in error saying migration process stopped.
So lets include a sample config for the correct way of doing it
<IS_HOME>/migration-resources/migration-config.yaml
```
name: "GroupsAndRolesMigrator"
order: 4
parameters:
# Absolute path for the dry report. This is required in the dry run mode.
reportPath: "/home/ubuntu/WSO2/IS600/MigrationTesting/reports"
```
- [x] Mention you need to pass the reportPath as a string value
- [x] Mention you need to define reportPath under a **parameter** config
- [x] Mention that you need to uncomment and include reportPath in the migration-config.yaml uner each migrator that you wants to get the reports generated ( At the moment there are 5 occurences in the migration-config.yaml to uncomment)
Attaching sample config file for reference
[resources.zip](https://github.com/wso2/product-is/files/8888114/resources.zip)
| priority | missing information on the dry run report generation configuration in migration guide describe the issue during a migration testing if you want to generate dry run reports that can be done from the migration config yaml but migration guide doesnt include the correct steps with example on how to do that dry run section to generate dry run reports we have to pass the reportpath as a string value and this reportpath always needs to be defined under a parameter config if we dont pass it as a string or dont include it under a parameter config report generatio will not get success and you will end up in error saying migration process stopped so lets include a sample config for the correct way of doing it migration resources migration config yaml name groupsandrolesmigrator order parameters absolute path for the dry report this is required in the dry run mode reportpath home ubuntu migrationtesting reports mention you need to pass the reportpath as a string value mention you need to define reportpath under a parameter config mention that you need to uncomment and include reportpath in the migration config yaml uner each migrator that you wants to get the reports generated at the moment there are occurences in the migration config yaml to uncomment attaching sample config file for reference | 1 |
266,259 | 20,143,479,793 | IssuesEvent | 2022-02-09 03:24:14 | Websoft9/ansible-ruby | https://api.github.com/repos/Websoft9/ansible-ruby | opened | Some pages are not fully translated | documentation | ## Language
- [ ] Chinese
- [x] English
## Content error
* URL: https://support.websoft9.com/docs/ruby/solution-upgrade.html#upgrade-ruby
* 
## Content optimization
Provide us with content optimization, e.g easy to read, simplify the steps, step by step, precise expression.
* URL:
* Description:
## New content
Describe important content what do you think is important in this documentation.
**What content do you want to add?**
(describe here)
**Steps for the content**
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. Modify file '....'
| 1.0 | Some pages are not fully translated - ## Language
- [ ] Chinese
- [x] English
## Content error
* URL: https://support.websoft9.com/docs/ruby/solution-upgrade.html#upgrade-ruby
* 
## Content optimization
Provide us with content optimization, e.g easy to read, simplify the steps, step by step, precise expression.
* URL:
* Description:
## New content
Describe important content what do you think is important in this documentation.
**What content do you want to add?**
(describe here)
**Steps for the content**
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. Modify file '....'
| non_priority | some pages are not fully translated language chinese english content error url content optimization provide us with content optimization e g easy to read simplify the steps step by step precise expression url description new content describe important content what do you think is important in this documentation what content do you want to add describe here steps for the content go to click on scroll down to modify file | 0 |
8,063 | 5,376,205,928 | IssuesEvent | 2017-02-23 08:23:13 | apinf/platform | https://api.github.com/repos/apinf/platform | closed | In Suomi setting, Filtering option in API catalog needs Finnish translations | i18n/l10n in progress Usability Issue | _From @Nazarah on February 1, 2017 7:30_


_Copied from original issue: Digipalvelutehdas/APIKA#337_ | True | In Suomi setting, Filtering option in API catalog needs Finnish translations - _From @Nazarah on February 1, 2017 7:30_


_Copied from original issue: Digipalvelutehdas/APIKA#337_ | non_priority | in suomi setting filtering option in api catalog needs finnish translations from nazarah on february copied from original issue digipalvelutehdas apika | 0 |
264,910 | 8,321,332,077 | IssuesEvent | 2018-09-25 23:35:02 | phenoscape/phenoscape-kb-services | https://api.github.com/repos/phenoscape/phenoscape-kb-services | closed | Apiary "Console" view doesn't work for Term Search in Firefox | bug low priority | In Firefox, the other API methods do work, just [the term search](http://docs.phenoscapekb.apiary.io/#reference/terms/term-search/search-for-terms-in-the-kb-by-text-match-on-a-property-value.-`rdfs:label`-is-searched-by-default.?console=1) doesn't. It does work in Chrome and Safari (cc @xu-hong). Not sure why Firefox is an issue.
Given that this is specific to one browser and is mostly controlled by Apiary, I don't think it qualifies as a priority. But perhaps there is something that you can see as a difference in the Term Search that might be responsible for this.
| 1.0 | Apiary "Console" view doesn't work for Term Search in Firefox - In Firefox, the other API methods do work, just [the term search](http://docs.phenoscapekb.apiary.io/#reference/terms/term-search/search-for-terms-in-the-kb-by-text-match-on-a-property-value.-`rdfs:label`-is-searched-by-default.?console=1) doesn't. It does work in Chrome and Safari (cc @xu-hong). Not sure why Firefox is an issue.
Given that this is specific to one browser and is mostly controlled by Apiary, I don't think it qualifies as a priority. But perhaps there is something that you can see as a difference in the Term Search that might be responsible for this.
| priority | apiary console view doesn t work for term search in firefox in firefox the other api methods do work just doesn t it does work in chrome and safari cc xu hong not sure why firefox is an issue given that this is specific to one browser and is mostly controlled by apiary i don t think it qualifies as a priority but perhaps there is something that you can see as a difference in the term search that might be responsible for this | 1 |
347,767 | 24,898,371,024 | IssuesEvent | 2022-10-28 18:05:55 | SeekyCt/ppcdis | https://api.github.com/repos/SeekyCt/ppcdis | opened | Document .ctors and .dtors weirdness | documentation high priority | There's a bunch of features / methods to work around the MWLD weirdness with the .ctors and .dtors sections that aren't really explained anywhere yet | 1.0 | Document .ctors and .dtors weirdness - There's a bunch of features / methods to work around the MWLD weirdness with the .ctors and .dtors sections that aren't really explained anywhere yet | non_priority | document ctors and dtors weirdness there s a bunch of features methods to work around the mwld weirdness with the ctors and dtors sections that aren t really explained anywhere yet | 0 |
663,782 | 22,206,461,399 | IssuesEvent | 2022-06-07 15:14:49 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | closed | FireEdge onprem HCI errors | Type: Bug Status: Accepted Priority: Normal Category: Provision Category: FireEdge | **Description**
HCI onprem cluster deployment is affected by a couple of minor error/typos:
- The provider filed at `usr/share/one/oneprovision/edge-clusters/metal/provisions/onprem.d/fireedge.yml` should be `onprem` instead of `oneprem`.
- When any of the `Hosts to run ...` fields are empty it will result in the creation of a host named `undefined`.
**Details**
- Affected Component: FireEdge - OneProvision - HCI Deployment
- Version: 6.4.0
**Additional context**
Add any other context about the problem here.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Code committed
- [ ] Testing - QA
- [ ] Documentation (Release notes - resolved issues, compatibility, known issues)
| 1.0 | FireEdge onprem HCI errors - **Description**
HCI onprem cluster deployment is affected by a couple of minor error/typos:
- The provider filed at `usr/share/one/oneprovision/edge-clusters/metal/provisions/onprem.d/fireedge.yml` should be `onprem` instead of `oneprem`.
- When any of the `Hosts to run ...` fields are empty it will result in the creation of a host named `undefined`.
**Details**
- Affected Component: FireEdge - OneProvision - HCI Deployment
- Version: 6.4.0
**Additional context**
Add any other context about the problem here.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Code committed
- [ ] Testing - QA
- [ ] Documentation (Release notes - resolved issues, compatibility, known issues)
| priority | fireedge onprem hci errors description hci onprem cluster deployment is affected by a couple of minor error typos the provider filed at usr share one oneprovision edge clusters metal provisions onprem d fireedge yml should be onprem instead of oneprem when any of the hosts to run fields are empty it will result in the creation of a host named undefined details affected component fireedge oneprovision hci deployment version additional context add any other context about the problem here progress status code committed testing qa documentation release notes resolved issues compatibility known issues | 1 |
682,914 | 23,362,210,920 | IssuesEvent | 2022-08-10 12:42:26 | feast-dev/feast | https://api.github.com/repos/feast-dev/feast | closed | Wrong condition in to_snowflake method of snowflake provider | kind/bug priority/p2 | ## Expected Behavior
[to_snowflake](https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/offline_stores/snowflake.py#L430) should ignore on_demand_feature_views if on_demand_feature_views is an empty list
## Current Behavior
to_snowflake runs on_demand_feature_views code because on_demand_feature_views is set to [] [here](https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/offline_stores/snowflake.py#L387)
## Steps to reproduce
### Specifications
- Version: 0.23.0
- Platform: feast SDK
- Subsystem:
## Possible Solution
`if self.on_demand_feature_views is not None: ` => `if self.on_demand_feature_views is not []: `
| 1.0 | Wrong condition in to_snowflake method of snowflake provider - ## Expected Behavior
[to_snowflake](https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/offline_stores/snowflake.py#L430) should ignore on_demand_feature_views if on_demand_feature_views is an empty list
## Current Behavior
to_snowflake runs on_demand_feature_views code because on_demand_feature_views is set to [] [here](https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/offline_stores/snowflake.py#L387)
## Steps to reproduce
### Specifications
- Version: 0.23.0
- Platform: feast SDK
- Subsystem:
## Possible Solution
`if self.on_demand_feature_views is not None: ` => `if self.on_demand_feature_views is not []: `
| priority | wrong condition in to snowflake method of snowflake provider expected behavior should ignore on demand feature views if on demand feature views is an empty list current behavior to snowflake runs on demand feature views code because on demand feature views is set to steps to reproduce specifications version platform feast sdk subsystem possible solution if self on demand feature views is not none if self on demand feature views is not | 1 |
281,102 | 30,872,647,784 | IssuesEvent | 2023-08-03 12:26:13 | flatcar/Flatcar | https://api.github.com/repos/flatcar/Flatcar | opened | update: Rust | security advisory | **Name**: rust
**CVEs**: [CVE-2023-38497](https://nvd.nist.gov/vuln/detail/CVE-2023-38497)
**CVSSs**: n/a
**Action Needed**: update to >= 1.71.1
**Summary**:
(quote from https://seclists.org/oss-sec/2023/q3/90)
## Security advisory for Cargo (CVE-2023-38497)
The Rust Security Response WG was notified that Cargo did not respect the umask when extracting crate archives on UNIX-like systems. If the user downloaded a crate containing files writeable by any local user, another local user could exploit this to change the source code compiled and executed by the current user.
This vulnerability has been assigned CVE-2023-38497.
### Overview
In UNIX-like systems, each file has three sets of permissions: for the user owning the file, for the group owning the file, and for all other local users. The "[umask][1]" is configured on most systems to limit those permissions during file creation, removing dangerous ones. For example, the default umask on macOS and most Linux distributions only allow the user owning a file to write to it, preventing the group owning it or other local users from doing the same.
When a dependency is downloaded by Cargo, its source code has to be extracted on disk to allow the Rust compiler to read as part of the build. To improve performance, this extraction only happens the first time a dependency is used, caching the pre-extracted files for future invocations.
Unfortunately, it was discovered that Cargo did not respect the umask during extraction, and propagated the permissions stored in the crate archive as-is. If an archive contained files writeable by any user on the system (and the system configuration didn't prevent writes through other security measures), another local user on the system could replace or tweak the source code of a dependency, potentially achieving code execution the next time the project is compiled.
### Affected Versions
All Rust versions before 1.71.1 on UNIX-like systems (like macOS and Linux) are affected. Note that additional system-dependent security measures configured on the local system might prevent the vulnerability from being exploited.
Users on Windows and other non-UNIX-like systems are not affected.
### Mitigations
We recommend all users to update to Rust 1.71.1, which will be released later today, as it fixes the vulnerability by respecting the umask when extracting crate archives. If you build your own toolchain, patches for 1.71.0 source tarballs are [available here][2].
To prevent existing cached extractions from being exploitable, the Cargo binary included in Rust 1.71.1 or later will purge the caches it tries to access if they were generated by older Cargo versions.
If you cannot update to Rust 1.71.1, we recommend configuring your system to prevent other local users from accessing the Cargo directory, usually located in `~/.cargo`:
```
chmod go= ~/.cargo
```
**refmap.gentoo**: TBD
| True | update: Rust - **Name**: rust
**CVEs**: [CVE-2023-38497](https://nvd.nist.gov/vuln/detail/CVE-2023-38497)
**CVSSs**: n/a
**Action Needed**: update to >= 1.71.1
**Summary**:
(quote from https://seclists.org/oss-sec/2023/q3/90)
## Security advisory for Cargo (CVE-2023-38497)
The Rust Security Response WG was notified that Cargo did not respect the umask when extracting crate archives on UNIX-like systems. If the user downloaded a crate containing files writeable by any local user, another local user could exploit this to change the source code compiled and executed by the current user.
This vulnerability has been assigned CVE-2023-38497.
### Overview
In UNIX-like systems, each file has three sets of permissions: for the user owning the file, for the group owning the file, and for all other local users. The "[umask][1]" is configured on most systems to limit those permissions during file creation, removing dangerous ones. For example, the default umask on macOS and most Linux distributions only allow the user owning a file to write to it, preventing the group owning it or other local users from doing the same.
When a dependency is downloaded by Cargo, its source code has to be extracted on disk to allow the Rust compiler to read as part of the build. To improve performance, this extraction only happens the first time a dependency is used, caching the pre-extracted files for future invocations.
Unfortunately, it was discovered that Cargo did not respect the umask during extraction, and propagated the permissions stored in the crate archive as-is. If an archive contained files writeable by any user on the system (and the system configuration didn't prevent writes through other security measures), another local user on the system could replace or tweak the source code of a dependency, potentially achieving code execution the next time the project is compiled.
### Affected Versions
All Rust versions before 1.71.1 on UNIX-like systems (like macOS and Linux) are affected. Note that additional system-dependent security measures configured on the local system might prevent the vulnerability from being exploited.
Users on Windows and other non-UNIX-like systems are not affected.
### Mitigations
We recommend all users to update to Rust 1.71.1, which will be released later today, as it fixes the vulnerability by respecting the umask when extracting crate archives. If you build your own toolchain, patches for 1.71.0 source tarballs are [available here][2].
To prevent existing cached extractions from being exploitable, the Cargo binary included in Rust 1.71.1 or later will purge the caches it tries to access if they were generated by older Cargo versions.
If you cannot update to Rust 1.71.1, we recommend configuring your system to prevent other local users from accessing the Cargo directory, usually located in `~/.cargo`:
```
chmod go= ~/.cargo
```
**refmap.gentoo**: TBD
| non_priority | update rust name rust cves cvsss n a action needed update to summary quote from security advisory for cargo cve the rust security response wg was notified that cargo did not respect the umask when extracting crate archives on unix like systems if the user downloaded a crate containing files writeable by any local user another local user could exploit this to change the source code compiled and executed by the current user this vulnerability has been assigned cve overview in unix like systems each file has three sets of permissions for the user owning the file for the group owning the file and for all other local users the is configured on most systems to limit those permissions during file creation removing dangerous ones for example the default umask on macos and most linux distributions only allow the user owning a file to write to it preventing the group owning it or other local users from doing the same when a dependency is downloaded by cargo its source code has to be extracted on disk to allow the rust compiler to read as part of the build to improve performance this extraction only happens the first time a dependency is used caching the pre extracted files for future invocations unfortunately it was discovered that cargo did not respect the umask during extraction and propagated the permissions stored in the crate archive as is if an archive contained files writeable by any user on the system and the system configuration didn t prevent writes through other security measures another local user on the system could replace or tweak the source code of a dependency potentially achieving code execution the next time the project is compiled affected versions all rust versions before on unix like systems like macos and linux are affected note that additional system dependent security measures configured on the local system might prevent the vulnerability from being exploited users on windows and other non unix like systems are not affected mitigations we recommend all users to update to rust which will be released later today as it fixes the vulnerability by respecting the umask when extracting crate archives if you build your own toolchain patches for source tarballs are to prevent existing cached extractions from being exploitable the cargo binary included in rust or later will purge the caches it tries to access if they were generated by older cargo versions if you cannot update to rust we recommend configuring your system to prevent other local users from accessing the cargo directory usually located in cargo chmod go cargo refmap gentoo tbd | 0 |
286,584 | 21,580,199,574 | IssuesEvent | 2022-05-02 17:51:46 | open-feature/spec | https://api.github.com/repos/open-feature/spec | opened | Should `after` hooks be allowed to mutate the value? | documentation | Pro: you can do a json -> instance mapping in there, which would be rad.
Con: subsequent hooks don't get the json value anymore, which may be very surprising to them. | 1.0 | Should `after` hooks be allowed to mutate the value? - Pro: you can do a json -> instance mapping in there, which would be rad.
Con: subsequent hooks don't get the json value anymore, which may be very surprising to them. | non_priority | should after hooks be allowed to mutate the value pro you can do a json instance mapping in there which would be rad con subsequent hooks don t get the json value anymore which may be very surprising to them | 0 |
47,029 | 11,947,116,522 | IssuesEvent | 2020-04-03 09:21:30 | Financial-Times/dotcom-page-kit | https://api.github.com/repos/Financial-Times/dotcom-page-kit | closed | Page Kit as a Webpack configuration builder | Build tools CLI tool Proposal Refactor | It has become apparent that the current Page Kit CLI is too complex for our needs and is leading to some confusion about the scope of the project. Because Page Kit provides its own CLI tool it is also worrying users that we are replacing one esoteric system with another.
At its core the current Page Kit CLI enables a complex Webpack configuration to be composed from multiple separate packages. It also makes accessing and modifying specific parts of the configuration simpler and more robust.
However, as discussed in #133 I would like to make it possible to use Webpack's own CLI like this:
```js
// webpack.config.js
module.exports = pageKit(options);
```
Taking this one step further... this would mitigate the need to maintain, document, and support an additional tool. I would also hope that by focusing on the configuration aspect it removes any worries about straying too far from "standard" tools.
This issue supersedes #58 and #123 and #133. | 1.0 | Page Kit as a Webpack configuration builder - It has become apparent that the current Page Kit CLI is too complex for our needs and is leading to some confusion about the scope of the project. Because Page Kit provides its own CLI tool it is also worrying users that we are replacing one esoteric system with another.
At its core the current Page Kit CLI enables a complex Webpack configuration to be composed from multiple separate packages. It also makes accessing and modifying specific parts of the configuration simpler and more robust.
However, as discussed in #133 I would like to make it possible to use Webpack's own CLI like this:
```js
// webpack.config.js
module.exports = pageKit(options);
```
Taking this one step further... this would mitigate the need to maintain, document, and support an additional tool. I would also hope that by focusing on the configuration aspect it removes any worries about straying too far from "standard" tools.
This issue supersedes #58 and #123 and #133. | non_priority | page kit as a webpack configuration builder it has become apparent that the current page kit cli is too complex for our needs and is leading to some confusion about the scope of the project because page kit provides its own cli tool it is also worrying users that we are replacing one esoteric system with another at its core the current page kit cli enables a complex webpack configuration to be composed from multiple separate packages it also makes accessing and modifying specific parts of the configuration simpler and more robust however as discussed in i would like to make it possible to use webpack s own cli like this js webpack config js module exports pagekit options taking this one step further this would mitigate the need to maintain document and support an additional tool i would also hope that by focusing on the configuration aspect it removes any worries about straying too far from standard tools this issue supersedes and and | 0 |
751,267 | 26,237,158,661 | IssuesEvent | 2023-01-05 08:25:08 | zino-hofmann/graphql-flutter | https://api.github.com/repos/zino-hofmann/graphql-flutter | closed | Subscription- Bad State : stream has already been listened to. | needs more info :hourglass: reproduction needed Priority: Waiting to be assigned | Hey, I recently got this problem again.
But this time, 1st time the socket connects and works perfectly fine. But on navigating back and forth to the same widget throws this error.
Are we supposed dispose something even when using hooks generated by codegen ? Am I missing something ?
_Originally posted by @Arpit1496 in https://github.com/zino-hofmann/graphql-flutter/issues/1191#issuecomment-1230804072_ | 1.0 | Subscription- Bad State : stream has already been listened to. - Hey, I recently got this problem again.
But this time, 1st time the socket connects and works perfectly fine. But on navigating back and forth to the same widget throws this error.
Are we supposed dispose something even when using hooks generated by codegen ? Am I missing something ?
_Originally posted by @Arpit1496 in https://github.com/zino-hofmann/graphql-flutter/issues/1191#issuecomment-1230804072_ | priority | subscription bad state stream has already been listened to hey i recently got this problem again but this time time the socket connects and works perfectly fine but on navigating back and forth to the same widget throws this error are we supposed dispose something even when using hooks generated by codegen am i missing something originally posted by in | 1 |
705,008 | 24,218,250,783 | IssuesEvent | 2022-09-26 08:42:10 | wp-media/wp-rocket | https://api.github.com/repos/wp-media/wp-rocket | reopened | RUCSS counter msg isnot displayed without refresh in certain case | type: bug priority: medium effort: [XS] severity: moderate module: remove unused css | **Before submitting an issue please check that youโve completed the following steps:**
- Made sure youโre on the latest version =>. 3.11.4
- Used the search feature to ensure that the bug hasnโt been reported before
**Describe the bug**
Counter msg for RUCSS is not displayed till refreshing the dashboard after activating RUCSS
**To Reproduce**
Steps to reproduce the behavior:
1. Fresh install to wpr
2. Activate RUCSS and save
3. Counter msg not displayed till refreshing dashboard
**Expected behavior**
Counter msg displayed after saving settings with no need to manually refresh the page
**Screenshots**
If applicable, add screenshots to help explain your problem.
https://jmp.sh/zqxRj0D
**Additional context**
Add any other context about the problem here.
- Same with PHP 7.4.3 and 8.1.7
- The same scenario was working fine on 3.11.3
- Another case when permissions are read-only to cache folder then enable RUCSS, once permissions are back, we need to refresh twice so we can see the RUCSS msg => in this case, if we fixed permissions after RUCSS was enabled by > 90sec, the success msg will be displayed not the counter although nothing is completed yet in used CSS table
- Note: clear used CSS, reactivate RUCSS, change safelist => all displaying msg with no need to refresh
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| 1.0 | RUCSS counter msg isnot displayed without refresh in certain case - **Before submitting an issue please check that youโve completed the following steps:**
- Made sure youโre on the latest version =>. 3.11.4
- Used the search feature to ensure that the bug hasnโt been reported before
**Describe the bug**
Counter msg for RUCSS is not displayed till refreshing the dashboard after activating RUCSS
**To Reproduce**
Steps to reproduce the behavior:
1. Fresh install to wpr
2. Activate RUCSS and save
3. Counter msg not displayed till refreshing dashboard
**Expected behavior**
Counter msg displayed after saving settings with no need to manually refresh the page
**Screenshots**
If applicable, add screenshots to help explain your problem.
https://jmp.sh/zqxRj0D
**Additional context**
Add any other context about the problem here.
- Same with PHP 7.4.3 and 8.1.7
- The same scenario was working fine on 3.11.3
- Another case when permissions are read-only to cache folder then enable RUCSS, once permissions are back, we need to refresh twice so we can see the RUCSS msg => in this case, if we fixed permissions after RUCSS was enabled by > 90sec, the success msg will be displayed not the counter although nothing is completed yet in used CSS table
- Note: clear used CSS, reactivate RUCSS, change safelist => all displaying msg with no need to refresh
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| priority | rucss counter msg isnot displayed without refresh in certain case before submitting an issue please check that youโve completed the following steps made sure youโre on the latest version used the search feature to ensure that the bug hasnโt been reported before describe the bug counter msg for rucss is not displayed till refreshing the dashboard after activating rucss to reproduce steps to reproduce the behavior fresh install to wpr activate rucss and save counter msg not displayed till refreshing dashboard expected behavior counter msg displayed after saving settings with no need to manually refresh the page screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here same with php and the same scenario was working fine on another case when permissions are read only to cache folder then enable rucss once permissions are back we need to refresh twice so we can see the rucss msg in this case if we fixed permissions after rucss was enabled by the success msg will be displayed not the counter although nothing is completed yet in used css table note clear used css reactivate rucss change safelist all displaying msg with no need to refresh backlog grooming for wp media dev team use only reproduce the problem identify the root cause scope a solution estimate the effort | 1 |
325,818 | 27,964,356,421 | IssuesEvent | 2023-03-24 18:07:12 | NeurodataWithoutBorders/pynwb | https://api.github.com/repos/NeurodataWithoutBorders/pynwb | closed | test tutorials across python versions | priority: medium topic: docs topic: testing | We are currently limiting the test matrix of the tutorials due to the AllenSDK.
https://github.com/NeurodataWithoutBorders/pynwb/blob/cc5fd34240891e14a6b82552a4b8c3c259fcafae/.github/workflows/run_all_tests.yml#L92-L107
The latest AllenSKD claims to support Python 3.9: https://github.com/AllenInstitute/AllenSDK/blob/1caa779b517eeb2428282919d0fb8a65e7217791/setup.py#L92
And anyway I don't think at this point we should be limiting our entire testing suite based on the Allen SDK.
| 1.0 | test tutorials across python versions - We are currently limiting the test matrix of the tutorials due to the AllenSDK.
https://github.com/NeurodataWithoutBorders/pynwb/blob/cc5fd34240891e14a6b82552a4b8c3c259fcafae/.github/workflows/run_all_tests.yml#L92-L107
The latest AllenSKD claims to support Python 3.9: https://github.com/AllenInstitute/AllenSDK/blob/1caa779b517eeb2428282919d0fb8a65e7217791/setup.py#L92
And anyway I don't think at this point we should be limiting our entire testing suite based on the Allen SDK.
| non_priority | test tutorials across python versions we are currently limiting the test matrix of the tutorials due to the allensdk the latest allenskd claims to support python and anyway i don t think at this point we should be limiting our entire testing suite based on the allen sdk | 0 |
44,696 | 2,910,632,639 | IssuesEvent | 2015-06-21 23:02:38 | SteamDatabase/steamSummerMinigame | https://api.github.com/repos/SteamDatabase/steamSummerMinigame | closed | Ignore raining gold on trash mobs late-game | 2 - Medium Priority Enhancement | Probably anything past level 1500 (or 10 railgun + 10 mouse button + 20 elemental damage levels) isn't worth switching to the raining gold lane unless Max Elemental Damage is also active. Prioritizing quick group levels by focusing on users' elemental spec will yield far more gold on bosses than anything one could gather from trash lane gold clicks. | 1.0 | Ignore raining gold on trash mobs late-game - Probably anything past level 1500 (or 10 railgun + 10 mouse button + 20 elemental damage levels) isn't worth switching to the raining gold lane unless Max Elemental Damage is also active. Prioritizing quick group levels by focusing on users' elemental spec will yield far more gold on bosses than anything one could gather from trash lane gold clicks. | priority | ignore raining gold on trash mobs late game probably anything past level or railgun mouse button elemental damage levels isn t worth switching to the raining gold lane unless max elemental damage is also active prioritizing quick group levels by focusing on users elemental spec will yield far more gold on bosses than anything one could gather from trash lane gold clicks | 1 |
141,616 | 11,427,956,148 | IssuesEvent | 2020-02-04 02:41:03 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | closed | The Clone button shows blank on Clone dialog | :beetle: regression :gear: blobs :gear: files ๐งช testing | **Storage Explorer Version:** 1.12.0
**Build:** [20200201.2](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3434212&view=results)
**Branch:** master
**Platform/OS:** Windows 10/ Linux Ubuntu 18.04
**Architecture:** ia32/x64
**Regression From:** Previous release(1.12.0)
**Steps to reproduce:**
1. Expand one Non-ADLS Gen2 storage account -> Blob Containers.
2. Select one blob container and right click it -> Click 'Clone with New Name...'.
3. Check the Clone button on the Clone dialog.
**Expect Experience:**
The Clone button shows well.
**Actual Experience:**
The Clone button shows blank.

**More Info:**
1. This issue also reproduces for files under one file share & one file share.
2. This issue doesn't reproduce for one blob. | 1.0 | The Clone button shows blank on Clone dialog - **Storage Explorer Version:** 1.12.0
**Build:** [20200201.2](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3434212&view=results)
**Branch:** master
**Platform/OS:** Windows 10/ Linux Ubuntu 18.04
**Architecture:** ia32/x64
**Regression From:** Previous release(1.12.0)
**Steps to reproduce:**
1. Expand one Non-ADLS Gen2 storage account -> Blob Containers.
2. Select one blob container and right click it -> Click 'Clone with New Name...'.
3. Check the Clone button on the Clone dialog.
**Expect Experience:**
The Clone button shows well.
**Actual Experience:**
The Clone button shows blank.

**More Info:**
1. This issue also reproduces for files under one file share & one file share.
2. This issue doesn't reproduce for one blob. | non_priority | the clone button shows blank on clone dialog storage explorer version build branch master platform os windows linux ubuntu architecture regression from previous release steps to reproduce expand one non adls storage account blob containers select one blob container and right click it click clone with new name check the clone button on the clone dialog expect experience the clone button shows well actual experience the clone button shows blank more info this issue also reproduces for files under one file share one file share this issue doesn t reproduce for one blob | 0 |
83,700 | 16,343,582,723 | IssuesEvent | 2021-05-13 03:08:09 | microsoft/AdaptiveCards | https://api.github.com/repos/microsoft/AdaptiveCards | closed | [iOS] Nightly package feed | Status-In Code Review Task | Laundry list.
1. remove .swiftpm from .gitignore
2. generate tags for nightly
3. limit what's included or shown in the package
4. add umbrella header for module map generation | 1.0 | [iOS] Nightly package feed - Laundry list.
1. remove .swiftpm from .gitignore
2. generate tags for nightly
3. limit what's included or shown in the package
4. add umbrella header for module map generation | non_priority | nightly package feed laundry list remove swiftpm from gitignore generate tags for nightly limit what s included or shown in the package add umbrella header for module map generation | 0 |
336,143 | 24,488,282,094 | IssuesEvent | 2022-10-09 18:33:38 | dgtlmoon/changedetection.io | https://api.github.com/repos/dgtlmoon/changedetection.io | closed | Images not showing up in wiki | documentation | I am new to changedetection and was going through the wiki but many pics are missing if you can please fix their links:
- https://github.com/dgtlmoon/changedetection.io/wiki
- https://github.com/dgtlmoon/changedetection.io/wiki/Sharing-a-Watch

| 1.0 | Images not showing up in wiki - I am new to changedetection and was going through the wiki but many pics are missing if you can please fix their links:
- https://github.com/dgtlmoon/changedetection.io/wiki
- https://github.com/dgtlmoon/changedetection.io/wiki/Sharing-a-Watch

| non_priority | images not showing up in wiki i am new to changedetection and was going through the wiki but many pics are missing if you can please fix their links | 0 |
157,048 | 12,344,285,578 | IssuesEvent | 2020-05-15 06:37:26 | celery/celery | https://api.github.com/repos/celery/celery | closed | Weird death / worker not restarting | Component: Prefork Workers Pool Status: Needs Testcase โ Status: Needs Verification โ | Hi,
Two different weird issues started happening. The first one is:
I have workers dying with this stacktrace
```
[2019-08-22 21:03:58,650: ERROR/MainProcess] Process 'ForkPoolWorker-89' pid:101 exited with 'signal 9 (SIGKILL)'
ERROR 2019-08-22 21:03:58,666 [celery.worker.request] <140652482161544> Task handler raised error: WorkerLostError('Worker exited prematurely: signal 9 (SIGKILL).')
Traceback (most recent call last):
File "/venv/lib/python3.7/site-packages/celery/worker/worker.py", line 205, in start
self.blueprint.start(self)
File "/venv/lib/python3.7/site-packages/celery/bootsteps.py", line 119, in start
step.start(parent)
File "/venv/lib/python3.7/site-packages/celery/bootsteps.py", line 369, in start
return self.obj.start()
File "/venv/lib/python3.7/site-packages/celery/worker/consumer/consumer.py", line 318, in start
blueprint.start(self)
File "/venv/lib/python3.7/site-packages/celery/bootsteps.py", line 119, in start
step.start(parent)
File "/venv/lib/python3.7/site-packages/celery/worker/consumer/consumer.py", line 596, in start
c.loop(*c.loop_args())
File "/venv/lib/python3.7/site-packages/celery/worker/loops.py", line 91, in asynloop
next(loop)
File "/venv/lib/python3.7/site-packages/kombu/asynchronous/hub.py", line 362, in create_loop
cb(*cbargs)
File "/venv/lib/python3.7/site-packages/celery/concurrency/asynpool.py", line 298, in on_result_readable
next(it)
File "/venv/lib/python3.7/site-packages/celery/concurrency/asynpool.py", line 279, in _recv_message
message = load(bufv)
UnicodeDecodeError: 'utf-8' codec can't decode bytes in position 46-47: invalid continuation byte
```
Second weird thing:
These workers never get restarted. So one by one my workers die, until it is just the "master" left on the node. A PS gives me something like
```
PID USER TIME COMMAND
1 root 3:53 {celery} /venv/bin/python3.7 /venv/bin/celery -A app_name worker -l INFO -n worker-%n-standard -c 20 -Q slow --without-heartbeat --without-mingle --without-gossip
86 root 0:40 [celery]
87 root 0:28 [celery]
88 root 0:37 [celery]
89 root 0:34 [celery]
90 root 0:38 [celery]
91 root 0:37 [celery]
92 root 0:26 [celery]
93 root 0:12 [celery]
94 root 0:09 [celery]
98 root 0:06 [celery]
99 root 0:05 [celery]
100 root 0:06 [celery]
102 root 0:04 [celery]
103 root 0:02 [celery]
104 root 0:02 [celery]
105 root 0:01 [celery]
106 root 0:01 [celery]
107 root 0:01 [celery]
108 root 0:00 [celery]
109 root 0:00 sh
```
As you can see, the 0 process is the master. The rest are just zombies doing nothing.
A weird byte seems like it should not kill a worker, and killing a worker should not make a zombie. Unsure what to tackle first?
Setup:
celery==4.3.0
billiard==3.6.1.0
kombu==4.5.0
RabbitMQ
Prefork
Running in docker on `python:3.7.3-alpine3.8`
| 1.0 | Weird death / worker not restarting - Hi,
Two different weird issues started happening. The first one is:
I have workers dying with this stacktrace
```
[2019-08-22 21:03:58,650: ERROR/MainProcess] Process 'ForkPoolWorker-89' pid:101 exited with 'signal 9 (SIGKILL)'
ERROR 2019-08-22 21:03:58,666 [celery.worker.request] <140652482161544> Task handler raised error: WorkerLostError('Worker exited prematurely: signal 9 (SIGKILL).')
Traceback (most recent call last):
File "/venv/lib/python3.7/site-packages/celery/worker/worker.py", line 205, in start
self.blueprint.start(self)
File "/venv/lib/python3.7/site-packages/celery/bootsteps.py", line 119, in start
step.start(parent)
File "/venv/lib/python3.7/site-packages/celery/bootsteps.py", line 369, in start
return self.obj.start()
File "/venv/lib/python3.7/site-packages/celery/worker/consumer/consumer.py", line 318, in start
blueprint.start(self)
File "/venv/lib/python3.7/site-packages/celery/bootsteps.py", line 119, in start
step.start(parent)
File "/venv/lib/python3.7/site-packages/celery/worker/consumer/consumer.py", line 596, in start
c.loop(*c.loop_args())
File "/venv/lib/python3.7/site-packages/celery/worker/loops.py", line 91, in asynloop
next(loop)
File "/venv/lib/python3.7/site-packages/kombu/asynchronous/hub.py", line 362, in create_loop
cb(*cbargs)
File "/venv/lib/python3.7/site-packages/celery/concurrency/asynpool.py", line 298, in on_result_readable
next(it)
File "/venv/lib/python3.7/site-packages/celery/concurrency/asynpool.py", line 279, in _recv_message
message = load(bufv)
UnicodeDecodeError: 'utf-8' codec can't decode bytes in position 46-47: invalid continuation byte
```
Second weird thing:
These workers never get restarted. So one by one my workers die, until it is just the "master" left on the node. A PS gives me something like
```
PID USER TIME COMMAND
1 root 3:53 {celery} /venv/bin/python3.7 /venv/bin/celery -A app_name worker -l INFO -n worker-%n-standard -c 20 -Q slow --without-heartbeat --without-mingle --without-gossip
86 root 0:40 [celery]
87 root 0:28 [celery]
88 root 0:37 [celery]
89 root 0:34 [celery]
90 root 0:38 [celery]
91 root 0:37 [celery]
92 root 0:26 [celery]
93 root 0:12 [celery]
94 root 0:09 [celery]
98 root 0:06 [celery]
99 root 0:05 [celery]
100 root 0:06 [celery]
102 root 0:04 [celery]
103 root 0:02 [celery]
104 root 0:02 [celery]
105 root 0:01 [celery]
106 root 0:01 [celery]
107 root 0:01 [celery]
108 root 0:00 [celery]
109 root 0:00 sh
```
As you can see, the 0 process is the master. The rest are just zombies doing nothing.
A weird byte seems like it should not kill a worker, and killing a worker should not make a zombie. Unsure what to tackle first?
Setup:
celery==4.3.0
billiard==3.6.1.0
kombu==4.5.0
RabbitMQ
Prefork
Running in docker on `python:3.7.3-alpine3.8`
| non_priority | weird death worker not restarting hi two different weird issues started happening the first one is i have workers dying with this stacktrace process forkpoolworker pid exited with signal sigkill error task handler raised error workerlosterror worker exited prematurely signal sigkill traceback most recent call last file venv lib site packages celery worker worker py line in start self blueprint start self file venv lib site packages celery bootsteps py line in start step start parent file venv lib site packages celery bootsteps py line in start return self obj start file venv lib site packages celery worker consumer consumer py line in start blueprint start self file venv lib site packages celery bootsteps py line in start step start parent file venv lib site packages celery worker consumer consumer py line in start c loop c loop args file venv lib site packages celery worker loops py line in asynloop next loop file venv lib site packages kombu asynchronous hub py line in create loop cb cbargs file venv lib site packages celery concurrency asynpool py line in on result readable next it file venv lib site packages celery concurrency asynpool py line in recv message message load bufv unicodedecodeerror utf codec can t decode bytes in position invalid continuation byte second weird thing these workers never get restarted so one by one my workers die until it is just the master left on the node a ps gives me something like pid user time command root celery venv bin venv bin celery a app name worker l info n worker n standard c q slow without heartbeat without mingle without gossip root root root root root root root root root root root root root root root root root root root root sh as you can see the process is the master the rest are just zombies doing nothing a weird byte seems like it should not kill a worker and killing a worker should not make a zombie unsure what to tackle first setup celery billiard kombu rabbitmq prefork running in docker on python | 0 |
222,814 | 7,439,064,736 | IssuesEvent | 2018-03-27 04:12:15 | minishift/minishift | https://api.github.com/repos/minishift/minishift | closed | Improvement: allow to assign multiple nameservers for use with network settings assignment | kind/feature priority/major | With #1316 it is currently possible to specify one nameserver on start to use, while the settings and assignment can support multiples of this. To keep the implementation simple only one is used at the moment. This task would like to improve on this by allowing multiple to be used (or at least 2). | 1.0 | Improvement: allow to assign multiple nameservers for use with network settings assignment - With #1316 it is currently possible to specify one nameserver on start to use, while the settings and assignment can support multiples of this. To keep the implementation simple only one is used at the moment. This task would like to improve on this by allowing multiple to be used (or at least 2). | priority | improvement allow to assign multiple nameservers for use with network settings assignment with it is currently possible to specify one nameserver on start to use while the settings and assignment can support multiples of this to keep the implementation simple only one is used at the moment this task would like to improve on this by allowing multiple to be used or at least | 1 |
15,043 | 2,841,940,144 | IssuesEvent | 2015-05-28 05:26:10 | ibus/ibus | https://api.github.com/repos/ibus/ibus | closed | Modify ibus-table to allow table-based IMs to define a symbol | Component-ibus-table Priority-Medium Type-Defect | ```
Input methods based on ibus-table should be able to define a symbol for IM switchers.
This could be done by adding a SYMBOL = line to the .txt table definitions and using
the value to set the symbol in ibus-table's main.py.
```
Original issue reported on code.google.com by `leftmostcat` on 2013-09-05 02:42:30 | 1.0 | Modify ibus-table to allow table-based IMs to define a symbol - ```
Input methods based on ibus-table should be able to define a symbol for IM switchers.
This could be done by adding a SYMBOL = line to the .txt table definitions and using
the value to set the symbol in ibus-table's main.py.
```
Original issue reported on code.google.com by `leftmostcat` on 2013-09-05 02:42:30 | non_priority | modify ibus table to allow table based ims to define a symbol input methods based on ibus table should be able to define a symbol for im switchers this could be done by adding a symbol line to the txt table definitions and using the value to set the symbol in ibus table s main py original issue reported on code google com by leftmostcat on | 0 |
779,792 | 27,366,330,882 | IssuesEvent | 2023-02-27 19:28:33 | bigbluebutton/bigbluebutton | https://api.github.com/repos/bigbluebutton/bigbluebutton | closed | Wb: Move the panning tool to be near the pointer tool | priority: low component: whiteboard | <!--PLEASE DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS.
This issue tracker is only for bbb development related issues.
Search for existing feature requests to avoid creating duplicates.-->
**Is your feature request related to a problem? Please describe.**
The panning tool was moved to the slide control toolbar. This new location is not intuitive. In BBB 2.5 and earlier the two tools were grouped together

**Describe the solution you'd like**

**Describe alternatives you've considered**
keep things as they are
| 1.0 | Wb: Move the panning tool to be near the pointer tool - <!--PLEASE DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS.
This issue tracker is only for bbb development related issues.
Search for existing feature requests to avoid creating duplicates.-->
**Is your feature request related to a problem? Please describe.**
The panning tool was moved to the slide control toolbar. This new location is not intuitive. In BBB 2.5 and earlier the two tools were grouped together

**Describe the solution you'd like**

**Describe alternatives you've considered**
keep things as they are
| priority | wb move the panning tool to be near the pointer tool please do not file issues for general support questions this issue tracker is only for bbb development related issues search for existing feature requests to avoid creating duplicates is your feature request related to a problem please describe the panning tool was moved to the slide control toolbar this new location is not intuitive in bbb and earlier the two tools were grouped together describe the solution you d like describe alternatives you ve considered keep things as they are | 1 |
16,867 | 5,295,025,653 | IssuesEvent | 2017-02-09 12:41:25 | TypeCobolTeam/TypeCobol | https://api.github.com/repos/TypeCobolTeam/TypeCobol | closed | Incorrect Code generation in presence of Preprocessor directives like COPY or REPLACE | bug Cobol Codegen | In presence of Preprocessor directives Like COPY and REPLACE the generator fails to generate original COPY or REPLACE instructions. We must implement another generation model based on generated code injection in the original source code.
The document TypeCobol\Codegen\src\Documentation\CodegenStudy.xlsx
Shows some situations for which the generation is not correctly handled.
| 1.0 | Incorrect Code generation in presence of Preprocessor directives like COPY or REPLACE - In presence of Preprocessor directives Like COPY and REPLACE the generator fails to generate original COPY or REPLACE instructions. We must implement another generation model based on generated code injection in the original source code.
The document TypeCobol\Codegen\src\Documentation\CodegenStudy.xlsx
Shows some situations for which the generation is not correctly handled.
| non_priority | incorrect code generation in presence of preprocessor directives like copy or replace in presence of preprocessor directives like copy and replace the generator fails to generate original copy or replace instructions we must implement another generation model based on generated code injection in the original source code the document typecobol codegen src documentation codegenstudy xlsx shows some situations for which the generation is not correctly handled | 0 |
80,140 | 15,586,268,084 | IssuesEvent | 2021-03-18 01:33:12 | peterwkc85/Spock_Testing_Framework | https://api.github.com/repos/peterwkc85/Spock_Testing_Framework | opened | CVE-2020-36184 (High) detected in jackson-databind-2.5.4.jar | security vulnerability | ## CVE-2020-36184 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.5.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: Spock_Testing_Framework/java-testing-with-spock-master/java-testing-with-spock-master/chapter7/rest-service-example/pom.xml</p>
<p>Path to vulnerable library: 2/repository/com/fasterxml/jackson/core/jackson-databind/2.5.4/jackson-databind-2.5.4.jar,radle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.5.4/5dfa42af84584b4a862ea488da84bbbebbb06c35/jackson-databind-2.5.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.5.4.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184>CVE-2020-36184</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2998">https://github.com/FasterXML/jackson-databind/issues/2998</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-36184 (High) detected in jackson-databind-2.5.4.jar - ## CVE-2020-36184 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.5.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: Spock_Testing_Framework/java-testing-with-spock-master/java-testing-with-spock-master/chapter7/rest-service-example/pom.xml</p>
<p>Path to vulnerable library: 2/repository/com/fasterxml/jackson/core/jackson-databind/2.5.4/jackson-databind-2.5.4.jar,radle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.5.4/5dfa42af84584b4a862ea488da84bbbebbb06c35/jackson-databind-2.5.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.5.4.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184>CVE-2020-36184</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2998">https://github.com/FasterXML/jackson-databind/issues/2998</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file spock testing framework java testing with spock master java testing with spock master rest service example pom xml path to vulnerable library repository com fasterxml jackson core jackson databind jackson databind jar radle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp datasources peruserpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource | 0 |
602,292 | 18,460,287,724 | IssuesEvent | 2021-10-15 23:37:09 | michaelrsweet/pdfio | https://api.github.com/repos/michaelrsweet/pdfio | closed | make all-shared Makefile errors | bug priority-medium | when running `make all-shared`, make fails with several errors
- lines 72,112: `else` is missing trailing backslash
- line 128: `soname` should be `-soname`
- making a shared object requires `-fPIC`
Thanks for the project, it has been extremely useful.
| 1.0 | make all-shared Makefile errors - when running `make all-shared`, make fails with several errors
- lines 72,112: `else` is missing trailing backslash
- line 128: `soname` should be `-soname`
- making a shared object requires `-fPIC`
Thanks for the project, it has been extremely useful.
| priority | make all shared makefile errors when running make all shared make fails with several errors lines else is missing trailing backslash line soname should be soname making a shared object requires fpic thanks for the project it has been extremely useful | 1 |
36,549 | 2,800,966,668 | IssuesEvent | 2015-05-13 13:29:52 | lars-t-hansen/parlib-simple | https://api.github.com/repos/lars-t-hansen/parlib-simple | closed | Occasional non-responsive barrier test | bug High priority | Mac release build with queue as of 16 Feb 2015, current m-i. Fast MBP 4x2. e10s enabled. Load test-barrier.html. Reload it very quickly many times, eventually I get to a state where the workers report "ready" but then nothing happens. Looking at the threads in the content process in lldb, the main thread as well as two of the DOM workers are in futexWait but all the others are in their primary event loop.
Setting the timeout in the main program to 100ms rather than 0 gets rid of the problem. However this does not provide a good explanation for the observed behavior. | 1.0 | Occasional non-responsive barrier test - Mac release build with queue as of 16 Feb 2015, current m-i. Fast MBP 4x2. e10s enabled. Load test-barrier.html. Reload it very quickly many times, eventually I get to a state where the workers report "ready" but then nothing happens. Looking at the threads in the content process in lldb, the main thread as well as two of the DOM workers are in futexWait but all the others are in their primary event loop.
Setting the timeout in the main program to 100ms rather than 0 gets rid of the problem. However this does not provide a good explanation for the observed behavior. | priority | occasional non responsive barrier test mac release build with queue as of feb current m i fast mbp enabled load test barrier html reload it very quickly many times eventually i get to a state where the workers report ready but then nothing happens looking at the threads in the content process in lldb the main thread as well as two of the dom workers are in futexwait but all the others are in their primary event loop setting the timeout in the main program to rather than gets rid of the problem however this does not provide a good explanation for the observed behavior | 1 |
168,003 | 13,055,224,172 | IssuesEvent | 2020-07-30 01:00:59 | n4af/TR4W | https://api.github.com/repos/n4af/TR4W | closed | Use new Icom commands to set inactive VFO | In testing Radio Control enhancement | For Split mode, newer Icom radios now support a command to set the inactive VFO to a frequency and mode (25h) rather than having to switch to VFO B, set the frequency and mode and switch back. This should be implemented as it simplifies the switching and puts it more inline with the way the other radios are done. | 1.0 | Use new Icom commands to set inactive VFO - For Split mode, newer Icom radios now support a command to set the inactive VFO to a frequency and mode (25h) rather than having to switch to VFO B, set the frequency and mode and switch back. This should be implemented as it simplifies the switching and puts it more inline with the way the other radios are done. | non_priority | use new icom commands to set inactive vfo for split mode newer icom radios now support a command to set the inactive vfo to a frequency and mode rather than having to switch to vfo b set the frequency and mode and switch back this should be implemented as it simplifies the switching and puts it more inline with the way the other radios are done | 0 |
116,471 | 17,370,033,368 | IssuesEvent | 2021-07-30 12:49:44 | lukebroganws/Java-Demo | https://api.github.com/repos/lukebroganws/Java-Demo | opened | CVE-2016-10735 (Medium) detected in bootstrap-3.3.7.min.js | security vulnerability | ## CVE-2016-10735 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: Java-Demo/src/main/webapp/dfi/style_bootstrap.html</p>
<p>Path to vulnerable library: Java-Demo/src/main/webapp/dfi/style_bootstrap.html,Java-Demo/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7-3.3.13.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/lukebroganws/Java-Demo/commit/d73a27e2fea07f94b9c092744aef285ec88e27c4">d73a27e2fea07f94b9c092744aef285ec88e27c4</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/issues/20184">https://github.com/twbs/bootstrap/issues/20184</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: 3.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"todc-bootstrap","packageVersion":"3.3.7-3.3.13","packageFilePaths":["/src/main/webapp/dfi/style_bootstrap.html","/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html"],"isTransitiveDependency":false,"dependencyTree":"todc-bootstrap:3.3.7-3.3.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2016-10735","vulnerabilityDetails":"In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2016-10735 (Medium) detected in bootstrap-3.3.7.min.js - ## CVE-2016-10735 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: Java-Demo/src/main/webapp/dfi/style_bootstrap.html</p>
<p>Path to vulnerable library: Java-Demo/src/main/webapp/dfi/style_bootstrap.html,Java-Demo/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7-3.3.13.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/lukebroganws/Java-Demo/commit/d73a27e2fea07f94b9c092744aef285ec88e27c4">d73a27e2fea07f94b9c092744aef285ec88e27c4</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/issues/20184">https://github.com/twbs/bootstrap/issues/20184</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: 3.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"todc-bootstrap","packageVersion":"3.3.7-3.3.13","packageFilePaths":["/src/main/webapp/dfi/style_bootstrap.html","/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html"],"isTransitiveDependency":false,"dependencyTree":"todc-bootstrap:3.3.7-3.3.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2016-10735","vulnerabilityDetails":"In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file java demo src main webapp dfi style bootstrap html path to vulnerable library java demo src main webapp dfi style bootstrap html java demo target easybuggy snapshot dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch main vulnerability details in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree todc bootstrap isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve vulnerabilityurl | 0 |
317,117 | 23,664,740,755 | IssuesEvent | 2022-08-26 19:28:50 | python/mypy | https://api.github.com/repos/python/mypy | closed | globl "type: ignore" ignores too much | bug documentation topic-type-ignore | Structure:
`sample_proj/__init__.py`:
```python
from sample_proj.commons import some_const
from sample_proj.some_file import SomeClass
```
`sample_proj/comons.py`:
```python
some_const = 'qwe'
```
`sample_proj/some_file.py`:
```python
# type: ignore
from sample_proj.commons import some_const
class SomeClass:
def fun(self, param=some_const):
pass
```
```
$ mypy sample_proj
sample_proj/__init__.py:2: error: Module "sample_proj.some_file" has no attribute "SomeClass"
Found 1 error in 1 file (checked 3 source files)
```
I think it's a bug.
Removing `# type: ignore[union-attr]` helps. I wanted this comment to ignore only this exact error in the whole file.
```
$ mypy --version
mypy 0.942
$ python3 --version
Python 3.10.5
``` | 1.0 | globl "type: ignore" ignores too much - Structure:
`sample_proj/__init__.py`:
```python
from sample_proj.commons import some_const
from sample_proj.some_file import SomeClass
```
`sample_proj/comons.py`:
```python
some_const = 'qwe'
```
`sample_proj/some_file.py`:
```python
# type: ignore
from sample_proj.commons import some_const
class SomeClass:
def fun(self, param=some_const):
pass
```
```
$ mypy sample_proj
sample_proj/__init__.py:2: error: Module "sample_proj.some_file" has no attribute "SomeClass"
Found 1 error in 1 file (checked 3 source files)
```
I think it's a bug.
Removing `# type: ignore[union-attr]` helps. I wanted this comment to ignore only this exact error in the whole file.
```
$ mypy --version
mypy 0.942
$ python3 --version
Python 3.10.5
``` | non_priority | globl type ignore ignores too much structure sample proj init py python from sample proj commons import some const from sample proj some file import someclass sample proj comons py python some const qwe sample proj some file py python type ignore from sample proj commons import some const class someclass def fun self param some const pass mypy sample proj sample proj init py error module sample proj some file has no attribute someclass found error in file checked source files i think it s a bug removing type ignore helps i wanted this comment to ignore only this exact error in the whole file mypy version mypy version python | 0 |
360,057 | 10,683,826,931 | IssuesEvent | 2019-10-22 09:11:17 | celo-org/celo-monorepo | https://api.github.com/repos/celo-org/celo-monorepo | closed | Attestation Users SBAT use the new precompile | Priority: P2 enhancement identity protocol triaged | ### Expected Behavior
Use the new precompile
### Current Behavior
Currently use `Validators#getValidators` which is very expensive
| 1.0 | Attestation Users SBAT use the new precompile - ### Expected Behavior
Use the new precompile
### Current Behavior
Currently use `Validators#getValidators` which is very expensive
| priority | attestation users sbat use the new precompile expected behavior use the new precompile current behavior currently use validators getvalidators which is very expensive | 1 |
187,596 | 6,759,497,247 | IssuesEvent | 2017-10-24 17:18:46 | rotorgames/Rg.Plugins.Popup | https://api.github.com/repos/rotorgames/Rg.Plugins.Popup | reopened | CloseWhenBackgroundIsClicked not works on Android with Xamarin.Forms 2.4.0.280 | bug confirmed fixed priority-high | The popup does not close when Background click on Android with Xamarin.Forms 2.4.0.280. This works with version 2.3.x. | 1.0 | CloseWhenBackgroundIsClicked not works on Android with Xamarin.Forms 2.4.0.280 - The popup does not close when Background click on Android with Xamarin.Forms 2.4.0.280. This works with version 2.3.x. | priority | closewhenbackgroundisclicked not works on android with xamarin forms the popup does not close when background click on android with xamarin forms this works with version x | 1 |
39,921 | 8,704,578,819 | IssuesEvent | 2018-12-05 19:48:10 | pgbackrest/pgbackrest | https://api.github.com/repos/pgbackrest/pgbackrest | closed | Archive command can leak s3 access (not secret) key in Postgres logs | enhancement (code) | Under certain circumstances when archiving to s3, the `repo-s3-key` can be leaked in the Postgres logs.
This comes back because s3 returns the request header, which has the credentials in it. Postgres picks that up and logs it to the PG log.
The first time I saw this was when there was an s3 issue and authentication was failing. It's easy to reproduce in test.
Setup pgBackRest to archive to s3. Once it's working, change the `repo-s3-key-secret` to an invalid key and try archiving a transaction log.
pgBackRests logging is configured as:
```
log-level-console=warn
log-level-file=info
log-level-stderr=warn
```
A heavily redacted version of the log output is:
```
2018-08-30 18:54:14 UTC [104]: [12-1] user=,db=,client= LOG: archive command failed with exit code 39
2018-08-30 18:54:14 UTC [104]: [13-1] user=,db=,client= DETAIL: The failed archive command was: pgbackrest --stanza=mydb --config=/conf/pgbackrest.conf archive-push pg_xlog/000000030000000000000007
ERROR: [039]: raised from local-1 process: S3 request error [403] Forbidden
*** request header ***
GET /aa-software/xxxxxxxxxxxx/pgbackrest/archive/mydb/archive.info? HTTP/1.1
authorization: AWS4-HMAC-SHA256 Credential=<repo-s3-key is here>/.....
<headers continue>
``` | 1.0 | Archive command can leak s3 access (not secret) key in Postgres logs - Under certain circumstances when archiving to s3, the `repo-s3-key` can be leaked in the Postgres logs.
This comes back because s3 returns the request header, which has the credentials in it. Postgres picks that up and logs it to the PG log.
The first time I saw this was when there was an s3 issue and authentication was failing. It's easy to reproduce in test.
Setup pgBackRest to archive to s3. Once it's working, change the `repo-s3-key-secret` to an invalid key and try archiving a transaction log.
pgBackRests logging is configured as:
```
log-level-console=warn
log-level-file=info
log-level-stderr=warn
```
A heavily redacted version of the log output is:
```
2018-08-30 18:54:14 UTC [104]: [12-1] user=,db=,client= LOG: archive command failed with exit code 39
2018-08-30 18:54:14 UTC [104]: [13-1] user=,db=,client= DETAIL: The failed archive command was: pgbackrest --stanza=mydb --config=/conf/pgbackrest.conf archive-push pg_xlog/000000030000000000000007
ERROR: [039]: raised from local-1 process: S3 request error [403] Forbidden
*** request header ***
GET /aa-software/xxxxxxxxxxxx/pgbackrest/archive/mydb/archive.info? HTTP/1.1
authorization: AWS4-HMAC-SHA256 Credential=<repo-s3-key is here>/.....
<headers continue>
``` | non_priority | archive command can leak access not secret key in postgres logs under certain circumstances when archiving to the repo key can be leaked in the postgres logs this comes back because returns the request header which has the credentials in it postgres picks that up and logs it to the pg log the first time i saw this was when there was an issue and authentication was failing it s easy to reproduce in test setup pgbackrest to archive to once it s working change the repo key secret to an invalid key and try archiving a transaction log pgbackrests logging is configured as log level console warn log level file info log level stderr warn a heavily redacted version of the log output is utc user db client log archive command failed with exit code utc user db client detail the failed archive command was pgbackrest stanza mydb config conf pgbackrest conf archive push pg xlog error raised from local process request error forbidden request header get aa software xxxxxxxxxxxx pgbackrest archive mydb archive info http authorization hmac credential | 0 |
329,661 | 24,231,192,220 | IssuesEvent | 2022-09-26 18:26:36 | supabase/supabase | https://api.github.com/repos/supabase/supabase | closed | docs: select permissions required for storage upsert | bug documentation | more context - https://github.com/supabase/supabase/discussions/2679
Docs need to be updated | 1.0 | docs: select permissions required for storage upsert - more context - https://github.com/supabase/supabase/discussions/2679
Docs need to be updated | non_priority | docs select permissions required for storage upsert more context docs need to be updated | 0 |
79,285 | 7,703,780,421 | IssuesEvent | 2018-05-21 09:41:03 | DrewAPicture/ensemble | https://api.github.com/repos/DrewAPicture/ensemble | opened | Add actual search support for Contests | ::Contests Enhancement | The search box for the Contests list table is currently disabled because there is only _sort of_ support for searching contests. It would be nice if there was the ability to search a set of filterable contest fields and turn that on (and therefore also support searches via query). | 1.0 | Add actual search support for Contests - The search box for the Contests list table is currently disabled because there is only _sort of_ support for searching contests. It would be nice if there was the ability to search a set of filterable contest fields and turn that on (and therefore also support searches via query). | non_priority | add actual search support for contests the search box for the contests list table is currently disabled because there is only sort of support for searching contests it would be nice if there was the ability to search a set of filterable contest fields and turn that on and therefore also support searches via query | 0 |
125,387 | 4,956,455,005 | IssuesEvent | 2016-12-01 23:33:02 | gravityview/GravityView | https://api.github.com/repos/gravityview/GravityView | opened | Filter & Sort tab not loading in Chrome | Bug Core: Administration Core: UI/UX Priority: Low | Works in Safari.
<img width="809" alt="unknown" src="https://cloud.githubusercontent.com/assets/870979/20817234/b5b10a92-b7e3-11e6-8429-7aa01622cd79.png">
Reported here: https://secure.helpscout.net/conversation/285133290/7556/ | 1.0 | Filter & Sort tab not loading in Chrome - Works in Safari.
<img width="809" alt="unknown" src="https://cloud.githubusercontent.com/assets/870979/20817234/b5b10a92-b7e3-11e6-8429-7aa01622cd79.png">
Reported here: https://secure.helpscout.net/conversation/285133290/7556/ | priority | filter sort tab not loading in chrome works in safari img width alt unknown src reported here | 1 |
639,157 | 20,747,763,368 | IssuesEvent | 2022-03-15 02:11:26 | apache/echarts | https://api.github.com/repos/apache/echarts | closed | Null values with category-based Y-axis don't render correctly | bug priority: high en | ### Version
5.1.2
### Reproduction link
[https://codepen.io/htr3n/pen/YzQKMrq](https://codepen.io/htr3n/pen/YzQKMrq)
### Steps to reproduce
1. Define a chart with the following options
- A-axis with type "category", set `data` to some kind of category such as `["6:00", "6:30", "7:00", "7:30"]`
- Y-axis with type "category", set `data` to some sort of category for instance `["A", "B", "C", "D"]`
- A series with a list of null input data: `data: [null, null, null, null],`
### What is expected?
The chart should not render any values because they are all null.
### What is actually happening?
The chart renders the first value at the position of the first Y-axis category.
---
Replace `null` with `undefined`, the chart seems to render correctly.
<!-- This issue is generated by echarts-issue-helper. DO NOT REMOVE -->
<!-- This issue is in English. DO NOT REMOVE --> | 1.0 | Null values with category-based Y-axis don't render correctly - ### Version
5.1.2
### Reproduction link
[https://codepen.io/htr3n/pen/YzQKMrq](https://codepen.io/htr3n/pen/YzQKMrq)
### Steps to reproduce
1. Define a chart with the following options
- A-axis with type "category", set `data` to some kind of category such as `["6:00", "6:30", "7:00", "7:30"]`
- Y-axis with type "category", set `data` to some sort of category for instance `["A", "B", "C", "D"]`
- A series with a list of null input data: `data: [null, null, null, null],`
### What is expected?
The chart should not render any values because they are all null.
### What is actually happening?
The chart renders the first value at the position of the first Y-axis category.
---
Replace `null` with `undefined`, the chart seems to render correctly.
<!-- This issue is generated by echarts-issue-helper. DO NOT REMOVE -->
<!-- This issue is in English. DO NOT REMOVE --> | priority | null values with category based y axis don t render correctly version reproduction link steps to reproduce define a chart with the following options a axis with type category set data to some kind of category such as y axis with type category set data to some sort of category for instance a series with a list of null input data data what is expected the chart should not render any values because they are all null what is actually happening the chart renders the first value at the position of the first y axis category replace null with undefined the chart seems to render correctly | 1 |
820,510 | 30,775,497,469 | IssuesEvent | 2023-07-31 06:01:48 | LiteLDev/LiteLoader.NET | https://api.github.com/repos/LiteLDev/LiteLoader.NET | closed | Hooks: New HookAPI | priority: normal status: in progress type: enhancement about: new LL.NET | ### Is your feature request related to a problem? Please describe.
Entering a delegate type to create hook's class and register it is not very convenient, and there is a way to get rid of it.

### Describe the solution you'd like
Hook delegate type, its parameters should be defined automatically by `Hook` field. | 1.0 | Hooks: New HookAPI - ### Is your feature request related to a problem? Please describe.
Entering a delegate type to create hook's class and register it is not very convenient, and there is a way to get rid of it.

### Describe the solution you'd like
Hook delegate type, its parameters should be defined automatically by `Hook` field. | priority | hooks new hookapi is your feature request related to a problem please describe entering a delegate type to create hook s class and register it is not very convenient and there is a way to get rid of it describe the solution you d like hook delegate type its parameters should be defined automatically by hook field | 1 |
294,843 | 9,049,296,682 | IssuesEvent | 2019-02-12 03:54:45 | OctopusDeploy/Issues | https://api.github.com/repos/OctopusDeploy/Issues | closed | Service Fabric targets with Certificates fails on cloud-hosted instances | area/cloud priority | # Prerequisites
- [x] I have verified the problem exists in the latest version
- [x] I have searched [open](https://github.com/OctopusDeploy/Issues/issues) and [closed](https://github.com/OctopusDeploy/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already reported
- [x] I have written a descriptive issue title
- [x] I have linked the original source of this report
- [x] I have tagged the issue appropriately (area/*, kind/bug, tag/regression?)
# The bug
When using a cloud-hosted instance of Octopus, if you try to use Service Fabric targets with the Certificates as the security mode, exceptions are occurring.
**PLEASE NOTE** - I was not able to reproduce the same exception the customer has reported, but I received a different exception. When working through this issue, please investigate that the fix will solve the original issue reported by the customer as well.
## Steps to reproduce
1. Create a cloud instance
2. Create an SF cluster on Azure, using certificates and a key vault, wait until the cluster is "Ready"
3. Create an SF target in your cloud instance, setup to use the certificate available on Azure
4. See error during health check (see attached log)
### Screen capture
<img width="1527" alt="screen shot 2018-08-27 at 12 38 43 pm" src="https://user-images.githubusercontent.com/819605/44637712-299cf000-a9f6-11e8-8af7-df93f54fe5fd.png">
### Log exerpt
[ServerTasks-29.log.txt](https://github.com/OctopusDeploy/Issues/files/2322227/ServerTasks-29.log.txt)
## Affected versions
Occurs in latest version of Octopus `2018.7.11` from a cloud-hosted instance.
## Workarounds
Overrride the default certificate store location as part of the step:

## Source
https://help.octopus.com/t/unable-to-create-azure-service-fabric-target/21133 | 1.0 | Service Fabric targets with Certificates fails on cloud-hosted instances - # Prerequisites
- [x] I have verified the problem exists in the latest version
- [x] I have searched [open](https://github.com/OctopusDeploy/Issues/issues) and [closed](https://github.com/OctopusDeploy/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already reported
- [x] I have written a descriptive issue title
- [x] I have linked the original source of this report
- [x] I have tagged the issue appropriately (area/*, kind/bug, tag/regression?)
# The bug
When using a cloud-hosted instance of Octopus, if you try to use Service Fabric targets with the Certificates as the security mode, exceptions are occurring.
**PLEASE NOTE** - I was not able to reproduce the same exception the customer has reported, but I received a different exception. When working through this issue, please investigate that the fix will solve the original issue reported by the customer as well.
## Steps to reproduce
1. Create a cloud instance
2. Create an SF cluster on Azure, using certificates and a key vault, wait until the cluster is "Ready"
3. Create an SF target in your cloud instance, setup to use the certificate available on Azure
4. See error during health check (see attached log)
### Screen capture
<img width="1527" alt="screen shot 2018-08-27 at 12 38 43 pm" src="https://user-images.githubusercontent.com/819605/44637712-299cf000-a9f6-11e8-8af7-df93f54fe5fd.png">
### Log exerpt
[ServerTasks-29.log.txt](https://github.com/OctopusDeploy/Issues/files/2322227/ServerTasks-29.log.txt)
## Affected versions
Occurs in latest version of Octopus `2018.7.11` from a cloud-hosted instance.
## Workarounds
Overrride the default certificate store location as part of the step:

## Source
https://help.octopus.com/t/unable-to-create-azure-service-fabric-target/21133 | priority | service fabric targets with certificates fails on cloud hosted instances prerequisites i have verified the problem exists in the latest version i have searched and issues to make sure it isn t already reported i have written a descriptive issue title i have linked the original source of this report i have tagged the issue appropriately area kind bug tag regression the bug when using a cloud hosted instance of octopus if you try to use service fabric targets with the certificates as the security mode exceptions are occurring please note i was not able to reproduce the same exception the customer has reported but i received a different exception when working through this issue please investigate that the fix will solve the original issue reported by the customer as well steps to reproduce create a cloud instance create an sf cluster on azure using certificates and a key vault wait until the cluster is ready create an sf target in your cloud instance setup to use the certificate available on azure see error during health check see attached log screen capture img width alt screen shot at pm src log exerpt affected versions occurs in latest version of octopus from a cloud hosted instance workarounds overrride the default certificate store location as part of the step source | 1 |
318,528 | 9,693,884,370 | IssuesEvent | 2019-05-24 17:20:37 | CosmiQ/solaris | https://api.github.com/repos/CosmiQ/solaris | closed | Re-write image stitching using torch/tensorflow | Difficulty: Medium Priority: Medium Type: Maintenance | As it stands now, `sol.raster.image.stitch_images` uses numpy to stitch images together. This means that images have to be moved back from the GPU to the CPU to run, and lose any advantage that GPU processing could potentially provide. Particularly since this is almost always done for post-inference images (i.e. `stitch_images()` is called within `sol.nets.infer.Inferer.__call__()`), the objects being merged are likely to often be torch tensors (or could be converted to torch tensors if they're coming from keras). We should therefore implement GPU-based post-processing using `stitch_images()`:
- [ ] re-write `stitch_images()` to use torch tensor operations instead of numpy arrays
- [ ] enable checking for GPU availability and use GPUs if possible
- [ ] enable check to see if data is in a numpy array when it's read in, and if so, convert it to a torch tensor | 1.0 | Re-write image stitching using torch/tensorflow - As it stands now, `sol.raster.image.stitch_images` uses numpy to stitch images together. This means that images have to be moved back from the GPU to the CPU to run, and lose any advantage that GPU processing could potentially provide. Particularly since this is almost always done for post-inference images (i.e. `stitch_images()` is called within `sol.nets.infer.Inferer.__call__()`), the objects being merged are likely to often be torch tensors (or could be converted to torch tensors if they're coming from keras). We should therefore implement GPU-based post-processing using `stitch_images()`:
- [ ] re-write `stitch_images()` to use torch tensor operations instead of numpy arrays
- [ ] enable checking for GPU availability and use GPUs if possible
- [ ] enable check to see if data is in a numpy array when it's read in, and if so, convert it to a torch tensor | priority | re write image stitching using torch tensorflow as it stands now sol raster image stitch images uses numpy to stitch images together this means that images have to be moved back from the gpu to the cpu to run and lose any advantage that gpu processing could potentially provide particularly since this is almost always done for post inference images i e stitch images is called within sol nets infer inferer call the objects being merged are likely to often be torch tensors or could be converted to torch tensors if they re coming from keras we should therefore implement gpu based post processing using stitch images re write stitch images to use torch tensor operations instead of numpy arrays enable checking for gpu availability and use gpus if possible enable check to see if data is in a numpy array when it s read in and if so convert it to a torch tensor | 1 |
28,055 | 5,428,749,306 | IssuesEvent | 2017-03-03 16:38:16 | maryvilledev/codesplainUI | https://api.github.com/repos/maryvilledev/codesplainUI | opened | Project Wikis | Documentation | It would be beneficial to edit our repo's wikis so we can have a centralized base of relevant information | 1.0 | Project Wikis - It would be beneficial to edit our repo's wikis so we can have a centralized base of relevant information | non_priority | project wikis it would be beneficial to edit our repo s wikis so we can have a centralized base of relevant information | 0 |
371,759 | 10,981,078,813 | IssuesEvent | 2019-11-30 18:59:43 | BendroCorp/bendrocorp-app | https://api.github.com/repos/BendroCorp/bendrocorp-app | opened | System Map Search View | api supported effort: high feature priority:high | - [ ] Offer a map view or a searchable list view - switch via toggle and persist preference
- [ ] Search View (will be the default mobile view when added)
- [ ] Searching for something adds it to the recently viewed list at the top of the system map
- [ ] Has a list of all items contained within system map and their essential data
- [ ] Searchable by title
- [ ] Opens to an object details page | 1.0 | System Map Search View - - [ ] Offer a map view or a searchable list view - switch via toggle and persist preference
- [ ] Search View (will be the default mobile view when added)
- [ ] Searching for something adds it to the recently viewed list at the top of the system map
- [ ] Has a list of all items contained within system map and their essential data
- [ ] Searchable by title
- [ ] Opens to an object details page | priority | system map search view offer a map view or a searchable list view switch via toggle and persist preference search view will be the default mobile view when added searching for something adds it to the recently viewed list at the top of the system map has a list of all items contained within system map and their essential data searchable by title opens to an object details page | 1 |
61,974 | 6,767,442,968 | IssuesEvent | 2017-10-26 03:22:32 | Microsoft/vstest | https://api.github.com/repos/Microsoft/vstest | closed | Enable 'dotnet test' to be run on published output | dotnet-test exp: dotnet-cli | _From @kendrahavens on August 31, 2017 17:9_
## Steps to reproduce
After publishing users should be able to run tests on an already compiled executable.
Would address [dotnet/core #917](https://github.com/dotnet/core/issues/917)
This is more important with the development of IL Linker [#915](https://github.com/dotnet/core/issues/915). The [Linker Instructions](https://github.com/dotnet/core/blob/master/samples/linker-instructions.md) mention that the final executables need to be tested after linking.
## Expected behavior
Be able to run 'dotnet test' on published output.
## Actual behavior
You can not today. (Or we need to add documentation if you can!)
@smalpani-msft
_Copied from original issue: dotnet/cli#7570_ | 1.0 | Enable 'dotnet test' to be run on published output - _From @kendrahavens on August 31, 2017 17:9_
## Steps to reproduce
After publishing users should be able to run tests on an already compiled executable.
Would address [dotnet/core #917](https://github.com/dotnet/core/issues/917)
This is more important with the development of IL Linker [#915](https://github.com/dotnet/core/issues/915). The [Linker Instructions](https://github.com/dotnet/core/blob/master/samples/linker-instructions.md) mention that the final executables need to be tested after linking.
## Expected behavior
Be able to run 'dotnet test' on published output.
## Actual behavior
You can not today. (Or we need to add documentation if you can!)
@smalpani-msft
_Copied from original issue: dotnet/cli#7570_ | non_priority | enable dotnet test to be run on published output from kendrahavens on august steps to reproduce after publishing users should be able to run tests on an already compiled executable would address this is more important with the development of il linker the mention that the final executables need to be tested after linking expected behavior be able to run dotnet test on published output actual behavior you can not today or we need to add documentation if you can smalpani msft copied from original issue dotnet cli | 0 |
641,775 | 20,834,249,162 | IssuesEvent | 2022-03-19 23:49:16 | LemonUIbyLemon/LemonUI | https://api.github.com/repos/LemonUIbyLemon/LemonUI | closed | Big Message weapon purchased crash | status: confirmed priority: p1 high type: bug | I get this error when trying to use the Big Message scaleform:
> System.ArgumentException: Unexpected argument type UInt32.
Parameter name: parameters
at LemonUI.Scaleform.BaseScaleform.CallFunctionBase(String function, Object[] parameters)
at LemonUI.Scaleform.BaseScaleform.CallFunction(String function, Object[] parameters)
at LemonUI.Scaleform.BigMessage..ctor(String title, String weapon, WeaponHash hash)
It happens when this line gets executed:
 | 1.0 | Big Message weapon purchased crash - I get this error when trying to use the Big Message scaleform:
> System.ArgumentException: Unexpected argument type UInt32.
Parameter name: parameters
at LemonUI.Scaleform.BaseScaleform.CallFunctionBase(String function, Object[] parameters)
at LemonUI.Scaleform.BaseScaleform.CallFunction(String function, Object[] parameters)
at LemonUI.Scaleform.BigMessage..ctor(String title, String weapon, WeaponHash hash)
It happens when this line gets executed:
 | priority | big message weapon purchased crash i get this error when trying to use the big message scaleform system argumentexception unexpected argument type parameter name parameters at lemonui scaleform basescaleform callfunctionbase string function object parameters at lemonui scaleform basescaleform callfunction string function object parameters at lemonui scaleform bigmessage ctor string title string weapon weaponhash hash it happens when this line gets executed | 1 |
20,992 | 11,573,217,644 | IssuesEvent | 2020-02-21 02:45:38 | Azure/azure-sdk-for-js | https://api.github.com/repos/Azure/azure-sdk-for-js | closed | [@azure/graph] getServicePrincipalsIdByAppId return MissingSubscription | Graph Service Attention customer-reported | - **Package Name**: "@azure/graph"
- **Package Version**: "5.0.0"
- **Operating system**:
- [ ] **nodejs**
- **version**: 10.51
- [ ] **browser**
- **name/version**:
- [ ] **typescript**
- **version**:
- Is the bug related to **documentation** in
- [ ] README.md
- [ ] source code documentation
- [ ] SDK API docs on https://docs.microsoft.com
**Describe the bug**
A clear and concise description of what the bug is.
getServicePrincipalsIdByAppId return 404 error
'{"error":{"code":"MissingSubscription","message":"The request did not have a subscription or a valid tenant level resource provider."}}'
**To Reproduce**
Steps to reproduce the behavior:
1 import pacakge:
const {GraphRbacManagementClient} = require('@azure/graph')
const msRestNodeAuth = require("@azure/ms-rest-nodeauth");
2. sign in with serviceprincipal
msRestNodeAuth.loginWithServicePrincipalSecret(clientId, secret, domain)
3 create graphic client GraphRbacManagementClient
const graphClient = new GraphRbacManagementClient(credentials, domain)
4. get service principalsIdByAppId
graphClient.applications.getServicePrincipalsIdByAppId(clientId, servicecallback)
**Expected behavior**
A clear and concise description of what you expected to happen.
expect to get service principalsId
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Additional context**
Add any other context about the problem here.
{ Error: The request did not have a subscription or a valid tenant level resource provider.
at new RestError (/home/local/TERADICI/fwang/repos/CAM/ms_authorization/node_modules/@azure/ms-rest-js/dist/msRest.node.js:2400:28)
at /home/local/TERADICI/fwang/repos/CAM/ms_authorization/node_modules/@azure/ms-rest-js/dist/msRest.node.js:3544:37
at process._tickCallback (internal/process/next_tick.js:68:7)
code: 'MissingSubscription',
statusCode: 404,
request:
WebResource {
streamResponseBody: false,
url:
'https://management.azure.com/xxxxxx/servicePrincipalsByAppId/xxxxxx/objectId?api-version=1.6',
method: 'GET',
headers: HttpHeaders { _headersMap: [Object] },
body: undefined,
query: undefined,
formData: undefined,
withCredentials: false,
abortSignal: undefined,
timeout: 0,
onUploadProgress: undefined,
onDownloadProgress: undefined,
proxySettings: undefined,
keepAlive: undefined,
operationSpec:
{ httpMethod: 'GET',
path:
'{tenantID}/servicePrincipalsByAppId/{applicationID}/objectId',
urlParameters: [Array],
queryParameters: [Array],
headerParameters: [Array],
responses: [Object],
serializer: [Serializer] } },
response:
{ body:
'{"error":{"code":"MissingSubscription","message":"The request did not have a subscription or a valid tenant level resource provider."}}',
headers: HttpHeaders { _headersMap: [Object] },
status: 404 },
body:
{ error:
{ code: 'MissingSubscription',
message:
'The request did not have a subscription or a valid tenant level resource provider.' } } }
| 1.0 | [@azure/graph] getServicePrincipalsIdByAppId return MissingSubscription - - **Package Name**: "@azure/graph"
- **Package Version**: "5.0.0"
- **Operating system**:
- [ ] **nodejs**
- **version**: 10.51
- [ ] **browser**
- **name/version**:
- [ ] **typescript**
- **version**:
- Is the bug related to **documentation** in
- [ ] README.md
- [ ] source code documentation
- [ ] SDK API docs on https://docs.microsoft.com
**Describe the bug**
A clear and concise description of what the bug is.
getServicePrincipalsIdByAppId return 404 error
'{"error":{"code":"MissingSubscription","message":"The request did not have a subscription or a valid tenant level resource provider."}}'
**To Reproduce**
Steps to reproduce the behavior:
1 import pacakge:
const {GraphRbacManagementClient} = require('@azure/graph')
const msRestNodeAuth = require("@azure/ms-rest-nodeauth");
2. sign in with serviceprincipal
msRestNodeAuth.loginWithServicePrincipalSecret(clientId, secret, domain)
3 create graphic client GraphRbacManagementClient
const graphClient = new GraphRbacManagementClient(credentials, domain)
4. get service principalsIdByAppId
graphClient.applications.getServicePrincipalsIdByAppId(clientId, servicecallback)
**Expected behavior**
A clear and concise description of what you expected to happen.
expect to get service principalsId
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Additional context**
Add any other context about the problem here.
{ Error: The request did not have a subscription or a valid tenant level resource provider.
at new RestError (/home/local/TERADICI/fwang/repos/CAM/ms_authorization/node_modules/@azure/ms-rest-js/dist/msRest.node.js:2400:28)
at /home/local/TERADICI/fwang/repos/CAM/ms_authorization/node_modules/@azure/ms-rest-js/dist/msRest.node.js:3544:37
at process._tickCallback (internal/process/next_tick.js:68:7)
code: 'MissingSubscription',
statusCode: 404,
request:
WebResource {
streamResponseBody: false,
url:
'https://management.azure.com/xxxxxx/servicePrincipalsByAppId/xxxxxx/objectId?api-version=1.6',
method: 'GET',
headers: HttpHeaders { _headersMap: [Object] },
body: undefined,
query: undefined,
formData: undefined,
withCredentials: false,
abortSignal: undefined,
timeout: 0,
onUploadProgress: undefined,
onDownloadProgress: undefined,
proxySettings: undefined,
keepAlive: undefined,
operationSpec:
{ httpMethod: 'GET',
path:
'{tenantID}/servicePrincipalsByAppId/{applicationID}/objectId',
urlParameters: [Array],
queryParameters: [Array],
headerParameters: [Array],
responses: [Object],
serializer: [Serializer] } },
response:
{ body:
'{"error":{"code":"MissingSubscription","message":"The request did not have a subscription or a valid tenant level resource provider."}}',
headers: HttpHeaders { _headersMap: [Object] },
status: 404 },
body:
{ error:
{ code: 'MissingSubscription',
message:
'The request did not have a subscription or a valid tenant level resource provider.' } } }
| non_priority | getserviceprincipalsidbyappid return missingsubscription package name azure graph package version operating system nodejs version browser name version typescript version is the bug related to documentation in readme md source code documentation sdk api docs on describe the bug a clear and concise description of what the bug is getserviceprincipalsidbyappid return error error code missingsubscription message the request did not have a subscription or a valid tenant level resource provider to reproduce steps to reproduce the behavior import pacakge const graphrbacmanagementclient require azure graph const msrestnodeauth require azure ms rest nodeauth sign in with serviceprincipal msrestnodeauth loginwithserviceprincipalsecret clientid secret domain create graphic client graphrbacmanagementclient const graphclient new graphrbacmanagementclient credentials domain get service principalsidbyappid graphclient applications getserviceprincipalsidbyappid clientid servicecallback expected behavior a clear and concise description of what you expected to happen expect to get service principalsid screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here error the request did not have a subscription or a valid tenant level resource provider at new resterror home local teradici fwang repos cam ms authorization node modules azure ms rest js dist msrest node js at home local teradici fwang repos cam ms authorization node modules azure ms rest js dist msrest node js at process tickcallback internal process next tick js code missingsubscription statuscode request webresource streamresponsebody false url method get headers httpheaders headersmap body undefined query undefined formdata undefined withcredentials false abortsignal undefined timeout onuploadprogress undefined ondownloadprogress undefined proxysettings undefined keepalive undefined operationspec httpmethod get path tenantid serviceprincipalsbyappid applicationid objectid urlparameters queryparameters headerparameters responses serializer response body error code missingsubscription message the request did not have a subscription or a valid tenant level resource provider headers httpheaders headersmap status body error code missingsubscription message the request did not have a subscription or a valid tenant level resource provider | 0 |
43,084 | 11,463,532,161 | IssuesEvent | 2020-02-07 16:11:03 | snowplow/snowplow-javascript-tracker | https://api.github.com/repos/snowplow/snowplow-javascript-tracker | closed | Fix non-linear incrementing of session index | type:defect | - lasts for as long as `sessionCookieTimeout`
- doesn't respect `heartbeatDelay` (multiple page pings per second despite an heartbeat of 5s)
- possible leads:
- `configHeartBeatTimer` might somehow be `null` [here](https://github.com/snowplow/snowplow-javascript-tracker/blob/master/src/js/tracker.js#L1291) which would cause the callback to happen relentlessly
- multiple [`setInterval`](https://github.com/snowplow/snowplow-javascript-tracker/blob/master/src/js/tracker.js#L1280) are set, that could be solved by `clearInterval` if we mean to have only one at any time
Session indices look like the following when ordered by `dvce_created_tstamp`:

which would suggest some kind of mad callback.
Tracker initialization looked like the following when the bug occurred:
```js
window._tracker("newTracker", "js", "", {
appId: "",
platform: "web",
post: !0,
cookieLifetime: 31536e4,
contexts: {
webPage: !0,
performanceTiming: !0
}
}),
window._tracker("enableLinkClickTracking", null, !0, !0),
window._tracker("refreshLinkClickTracking"),
window._tracker("enableActivityTracking", 5, 5),
window._tracker("trackPageView"),
``` | 1.0 | Fix non-linear incrementing of session index - - lasts for as long as `sessionCookieTimeout`
- doesn't respect `heartbeatDelay` (multiple page pings per second despite an heartbeat of 5s)
- possible leads:
- `configHeartBeatTimer` might somehow be `null` [here](https://github.com/snowplow/snowplow-javascript-tracker/blob/master/src/js/tracker.js#L1291) which would cause the callback to happen relentlessly
- multiple [`setInterval`](https://github.com/snowplow/snowplow-javascript-tracker/blob/master/src/js/tracker.js#L1280) are set, that could be solved by `clearInterval` if we mean to have only one at any time
Session indices look like the following when ordered by `dvce_created_tstamp`:

which would suggest some kind of mad callback.
Tracker initialization looked like the following when the bug occurred:
```js
window._tracker("newTracker", "js", "", {
appId: "",
platform: "web",
post: !0,
cookieLifetime: 31536e4,
contexts: {
webPage: !0,
performanceTiming: !0
}
}),
window._tracker("enableLinkClickTracking", null, !0, !0),
window._tracker("refreshLinkClickTracking"),
window._tracker("enableActivityTracking", 5, 5),
window._tracker("trackPageView"),
``` | non_priority | fix non linear incrementing of session index lasts for as long as sessioncookietimeout doesn t respect heartbeatdelay multiple page pings per second despite an heartbeat of possible leads configheartbeattimer might somehow be null which would cause the callback to happen relentlessly multiple are set that could be solved by clearinterval if we mean to have only one at any time session indices look like the following when ordered by dvce created tstamp which would suggest some kind of mad callback tracker initialization looked like the following when the bug occurred js window tracker newtracker js appid platform web post cookielifetime contexts webpage performancetiming window tracker enablelinkclicktracking null window tracker refreshlinkclicktracking window tracker enableactivitytracking window tracker trackpageview | 0 |
24,178 | 12,037,483,004 | IssuesEvent | 2020-04-13 21:57:10 | emergenzeHack/covid19gr_issues | https://api.github.com/repos/emergenzeHack/covid19gr_issues | opened | goods donation and services offering | Missing position form goods donation and services offering | <pre><yamldata>
Description: 'Live cameras ฯฮฟฯ
Monterey Bay Aquarium ฯฯฮทฮฝ ฮฮฑฮปฮนฯฯฯฮฝฮนฮฑ.
ฮฮต ฮทฮณฮตฯฮนฮบฮฎ ฯฮฑฯฮฟฯ
ฯฮฏฮฑ ฯฯฮทฮฝ ฮตฮบฯฮฑฮฏฮดฮตฯ
ฯฮท ฮณฮนฮฑ ฯฮฟฯ
ฯ ฯฮบฮตฮฑฮฝฮฟฯฯ, ฯฮฟ Monterey Bay Aquarium
ฮธฮตฯฯฮตฮฏฯฮฑฮน ฮฑฯฯ ฯฮฟฮปฮปฮฟฯฯ ฯฯฮน ฮตฮฏฮฝฮฑฮน ฯฮฟ ฮบฮฑฮปฯฯฮตฯฮฟ ฯฯฮฟฮฝ ฮบฯฯฮผฮฟ.'
Institution: Monteray Bay Aquarium
Legal: OK
Link: https://www.montereybayaquarium.org/animals/live-cams
Title: ฮฮฌฮธฮต ฯฮฟฮฝ ฯ
ฯฮฟฮธฮฑฮปฮฌฯฯฮนฮฟ ฮบฯฯฮผฮฟ!
</yamldata></pre> | 1.0 | goods donation and services offering - <pre><yamldata>
Description: 'Live cameras ฯฮฟฯ
Monterey Bay Aquarium ฯฯฮทฮฝ ฮฮฑฮปฮนฯฯฯฮฝฮนฮฑ.
ฮฮต ฮทฮณฮตฯฮนฮบฮฎ ฯฮฑฯฮฟฯ
ฯฮฏฮฑ ฯฯฮทฮฝ ฮตฮบฯฮฑฮฏฮดฮตฯ
ฯฮท ฮณฮนฮฑ ฯฮฟฯ
ฯ ฯฮบฮตฮฑฮฝฮฟฯฯ, ฯฮฟ Monterey Bay Aquarium
ฮธฮตฯฯฮตฮฏฯฮฑฮน ฮฑฯฯ ฯฮฟฮปฮปฮฟฯฯ ฯฯฮน ฮตฮฏฮฝฮฑฮน ฯฮฟ ฮบฮฑฮปฯฯฮตฯฮฟ ฯฯฮฟฮฝ ฮบฯฯฮผฮฟ.'
Institution: Monteray Bay Aquarium
Legal: OK
Link: https://www.montereybayaquarium.org/animals/live-cams
Title: ฮฮฌฮธฮต ฯฮฟฮฝ ฯ
ฯฮฟฮธฮฑฮปฮฌฯฯฮนฮฟ ฮบฯฯฮผฮฟ!
</yamldata></pre> | non_priority | goods donation and services offering description live cameras ฯฮฟฯ
monterey bay aquarium ฯฯฮทฮฝ ฮบฮฑฮปฮนฯฯฯฮฝฮนฮฑ ฮผฮต ฮทฮณฮตฯฮนฮบฮฎ ฯฮฑฯฮฟฯ
ฯฮฏฮฑ ฯฯฮทฮฝ ฮตฮบฯฮฑฮฏฮดฮตฯ
ฯฮท ฮณฮนฮฑ ฯฮฟฯ
ฯ ฯฮบฮตฮฑฮฝฮฟฯฯ ฯฮฟ monterey bay aquarium ฮธฮตฯฯฮตฮฏฯฮฑฮน ฮฑฯฯ ฯฮฟฮปฮปฮฟฯฯ ฯฯฮน ฮตฮฏฮฝฮฑฮน ฯฮฟ ฮบฮฑฮปฯฯฮตฯฮฟ ฯฯฮฟฮฝ ฮบฯฯฮผฮฟ institution monteray bay aquarium legal ok link title ฮผฮฌฮธฮต ฯฮฟฮฝ ฯ
ฯฮฟฮธฮฑฮปฮฌฯฯฮนฮฟ ฮบฯฯฮผฮฟ | 0 |
4,193 | 2,713,189,950 | IssuesEvent | 2015-04-09 17:52:20 | zaneu/cs373-idb | https://api.github.com/repos/zaneu/cs373-idb | closed | More unit tests | enhancement help wanted testing | I wrote some unit tests for ingredients, could @larry-liu , @jindtang , and @MenglinBrown get to work on writing some more? I was thinking of testing the search and filter by functionality of SQLAlchemy as much as possible.
Note that we need 3+ for each model, but more is always better. | 1.0 | More unit tests - I wrote some unit tests for ingredients, could @larry-liu , @jindtang , and @MenglinBrown get to work on writing some more? I was thinking of testing the search and filter by functionality of SQLAlchemy as much as possible.
Note that we need 3+ for each model, but more is always better. | non_priority | more unit tests i wrote some unit tests for ingredients could larry liu jindtang and menglinbrown get to work on writing some more i was thinking of testing the search and filter by functionality of sqlalchemy as much as possible note that we need for each model but more is always better | 0 |
12,961 | 9,817,730,809 | IssuesEvent | 2019-06-13 17:27:38 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | closed | [DocDB] DocumentDB Needs Tests | Cosmos Question Service Attention Test | - [ ] Add scenario tests for DocumentDB module
- [x] Add DocumentDB to the help document map (PR #3267) | 1.0 | [DocDB] DocumentDB Needs Tests - - [ ] Add scenario tests for DocumentDB module
- [x] Add DocumentDB to the help document map (PR #3267) | non_priority | documentdb needs tests add scenario tests for documentdb module add documentdb to the help document map pr | 0 |
176,724 | 13,650,851,607 | IssuesEvent | 2020-09-26 21:24:05 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Missing banners to re-direct to dashboard for Isito | [zube]: To Test priority/0 team/ui | In monitoring/logging/cis, we have banners added to the top to have links to re-direct to the v2 version.
Can we add these for Istio, alerts and notifiers?
Istio -> `/istio`
`The latest versions of Istio are only available in Cluster Explorer. See docs for details on what's new and how to migrate.`
alerts and notifiers page -> `monitoring` - For this text, can update it to
`We've updated alerts to be directly integrated with monitoring in Cluster Explorer. See docs for details on what's new and how to migrate.`
`We've updated notifiers to be directly integrated with monitoring in Cluster Explorer. See docs for details on what's new and how to migrate.`
<img width="855" alt="Screen Shot 2020-09-22 at 11 18 37 AM" src="https://user-images.githubusercontent.com/11410997/93921409-6450df80-fcc5-11ea-9ba3-b51c978c6834.png">
PR that added the banner for monitoring/logging/cis:
https://github.com/rancher/ui/pull/4147 | 1.0 | Missing banners to re-direct to dashboard for Isito - In monitoring/logging/cis, we have banners added to the top to have links to re-direct to the v2 version.
Can we add these for Istio, alerts and notifiers?
Istio -> `/istio`
`The latest versions of Istio are only available in Cluster Explorer. See docs for details on what's new and how to migrate.`
alerts and notifiers page -> `monitoring` - For this text, can update it to
`We've updated alerts to be directly integrated with monitoring in Cluster Explorer. See docs for details on what's new and how to migrate.`
`We've updated notifiers to be directly integrated with monitoring in Cluster Explorer. See docs for details on what's new and how to migrate.`
<img width="855" alt="Screen Shot 2020-09-22 at 11 18 37 AM" src="https://user-images.githubusercontent.com/11410997/93921409-6450df80-fcc5-11ea-9ba3-b51c978c6834.png">
PR that added the banner for monitoring/logging/cis:
https://github.com/rancher/ui/pull/4147 | non_priority | missing banners to re direct to dashboard for isito in monitoring logging cis we have banners added to the top to have links to re direct to the version can we add these for istio alerts and notifiers istio istio the latest versions of istio are only available in cluster explorer see docs for details on what s new and how to migrate alerts and notifiers page monitoring for this text can update it to we ve updated alerts to be directly integrated with monitoring in cluster explorer see docs for details on what s new and how to migrate we ve updated notifiers to be directly integrated with monitoring in cluster explorer see docs for details on what s new and how to migrate img width alt screen shot at am src pr that added the banner for monitoring logging cis | 0 |
95,890 | 3,961,746,906 | IssuesEvent | 2016-05-02 14:19:52 | Esri/military-tools-webappbuilder | https://api.github.com/repos/Esri/military-tools-webappbuilder | closed | Change the repo description | effort-small priority - high | The description for this repo needs to be changed
Currently:
Web AppBuilder widgets for Military Analyst +
Change to :
Web AppBuilder widgets for Military Tools for ArcGIS | 1.0 | Change the repo description - The description for this repo needs to be changed
Currently:
Web AppBuilder widgets for Military Analyst +
Change to :
Web AppBuilder widgets for Military Tools for ArcGIS | priority | change the repo description the description for this repo needs to be changed currently web appbuilder widgets for military analyst change to web appbuilder widgets for military tools for arcgis | 1 |
113,502 | 4,560,735,120 | IssuesEvent | 2016-09-14 09:07:47 | dotKom/onlineweb4 | https://api.github.com/repos/dotKom/onlineweb4 | opened | Deleting a Careeropportunity in the dashboard does not actually delete | Package: Dashboard Priority: High Status: Available Type: Bug | When trying to delete a career opportunity in the dashboard, it does not actually delete it. | 1.0 | Deleting a Careeropportunity in the dashboard does not actually delete - When trying to delete a career opportunity in the dashboard, it does not actually delete it. | priority | deleting a careeropportunity in the dashboard does not actually delete when trying to delete a career opportunity in the dashboard it does not actually delete it | 1 |
18,743 | 13,090,951,567 | IssuesEvent | 2020-08-03 05:13:07 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | [x64] Error in x64_checked_ubuntu_illink_flow | area-Infrastructure-coreclr | **Job:**
https://ci.dot.net/job/dotnet_coreclr/job/master/view/x64/job/illink/job/x64_checked_ubuntu_illink/648/
**Log:**
```
ERROR: An error occurred in /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/Tools/dotnetcli/dotnet msbuild /nologo /verbosity:minimal /clp:Summary /p:RestoreDefaultOptimizationDataPackage=false /p:PortableBuild=true /p:UsePartialNGENOptimization=false /maxcpucount /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/tests/build.proj /flp:Verbosity=normal;LogFile=/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.log;Append=true /flp1:WarningsOnly;LogFile=/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.wrn;Append=true /flp2:ErrorsOnly;LogFile=/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.err;Append=true /l:BinClashLogger,Tools/Microsoft.DotNet.Build.Tasks.dll;LogFile=binclash.log /p:__BuildArch=x64 /p:__BuildType=Checked /p:__BuildOS=Linux /p:CLRTestPriorityToBuild=1. Check logs under /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink.
Failed to build Managed tests build (build tests). See the build logs:
/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.log
/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.wrn
/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.err
``` | 1.0 | [x64] Error in x64_checked_ubuntu_illink_flow - **Job:**
https://ci.dot.net/job/dotnet_coreclr/job/master/view/x64/job/illink/job/x64_checked_ubuntu_illink/648/
**Log:**
```
ERROR: An error occurred in /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/Tools/dotnetcli/dotnet msbuild /nologo /verbosity:minimal /clp:Summary /p:RestoreDefaultOptimizationDataPackage=false /p:PortableBuild=true /p:UsePartialNGENOptimization=false /maxcpucount /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/tests/build.proj /flp:Verbosity=normal;LogFile=/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.log;Append=true /flp1:WarningsOnly;LogFile=/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.wrn;Append=true /flp2:ErrorsOnly;LogFile=/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.err;Append=true /l:BinClashLogger,Tools/Microsoft.DotNet.Build.Tasks.dll;LogFile=binclash.log /p:__BuildArch=x64 /p:__BuildType=Checked /p:__BuildOS=Linux /p:CLRTestPriorityToBuild=1. Check logs under /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink.
Failed to build Managed tests build (build tests). See the build logs:
/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.log
/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.wrn
/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.err
``` | non_priority | error in checked ubuntu illink flow job log error an error occurred in mnt j workspace dotnet coreclr master illink checked ubuntu illink tools dotnetcli dotnet msbuild nologo verbosity minimal clp summary p restoredefaultoptimizationdatapackage false p portablebuild true p usepartialngenoptimization false maxcpucount mnt j workspace dotnet coreclr master illink checked ubuntu illink tests build proj flp verbosity normal logfile mnt j workspace dotnet coreclr master illink checked ubuntu illink bin logs tests managed linux checked log append true warningsonly logfile mnt j workspace dotnet coreclr master illink checked ubuntu illink bin logs tests managed linux checked wrn append true errorsonly logfile mnt j workspace dotnet coreclr master illink checked ubuntu illink bin logs tests managed linux checked err append true l binclashlogger tools microsoft dotnet build tasks dll logfile binclash log p buildarch p buildtype checked p buildos linux p clrtestprioritytobuild check logs under mnt j workspace dotnet coreclr master illink checked ubuntu illink failed to build managed tests build build tests see the build logs mnt j workspace dotnet coreclr master illink checked ubuntu illink bin logs tests managed linux checked log mnt j workspace dotnet coreclr master illink checked ubuntu illink bin logs tests managed linux checked wrn mnt j workspace dotnet coreclr master illink checked ubuntu illink bin logs tests managed linux checked err | 0 |
156,884 | 5,990,599,402 | IssuesEvent | 2017-06-02 11:58:42 | universAAL/platform | https://api.github.com/repos/universAAL/platform | closed | required testing bundle is not automatically installed | bug imported priority 3 | _Originally Opened: @amedranogil (2014-05-08 11:41:39_)
_Originally Closed: 2014-05-08 12:51:21_
When building an artifact with Integration tests with a clean Maven local repo, a required bundle is not automatically installed.
It is suggested to add this bundle as a forced dependency in top level POMs with scope test, so it gets automatically installed at least the first time.
testComposite(org.universAAL.ontology.test.ArtifactIT) Time elapsed: 0.49 sec <<< ERROR!
java.io.FileNotFoundException: C:\Users\amedrano\.m2\repository\org\slf4j\com.springsource.slf4j.api\1.5.0\com.springsource.slf4j.api-1.5.0.jar (El sistema no puede encontrar la ruta especificada)
at java.io.FileInputStream.open(Native Method)
at java.io.FileInputStream.<init>(Unknown Source)
at org.springframework.core.io.FileSystemResource.getInputStream(FileSystemResource.java:108)
at org.springframework.osgi.test.AbstractOsgiTests.installBundle(AbstractOsgiTests.java:323)
at org.springframework.osgi.test.AbstractOsgiTests.startup(AbstractOsgiTests.java:253)
at org.springframework.osgi.test.AbstractOsgiTests.prepareTestExecution(AbstractOsgiTests.java:374)
at org.springframework.osgi.test.AbstractOsgiTests.runBare(AbstractOsgiTests.java:203)
at org.springframework.osgi.test.AbstractOsgiTests$1.protect(AbstractOsgiTests.java:184)
at junit.framework.TestResult.runProtected(TestResult.java:124)
at org.springframework.osgi.test.AbstractOsgiTests.run(AbstractOsgiTests.java:181)
at junit.framework.TestSuite.runTest(TestSuite.java:230)
at junit.framework.TestSuite.run(TestSuite.java:225)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.apache.maven.surefire.junit.JUnitTestSet.execute(JUnitTestSet.java:95)
at org.apache.maven.surefire.junit.JUnit3Provider.executeTestSet(JUnit3Provider.java:131)
at org.apache.maven.surefire.junit.JUnit3Provider.invoke(JUnit3Provider.java:108)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
Results :
Tests in error:
org.universAAL.ontology.test.ArtifactIT#testComposite FileNotFoundException C:...
--
From: _this issue has been automatically imported from our old issue tracker_ | 1.0 | required testing bundle is not automatically installed - _Originally Opened: @amedranogil (2014-05-08 11:41:39_)
_Originally Closed: 2014-05-08 12:51:21_
When building an artifact with Integration tests with a clean Maven local repo, a required bundle is not automatically installed.
It is suggested to add this bundle as a forced dependency in top level POMs with scope test, so it gets automatically installed at least the first time.
testComposite(org.universAAL.ontology.test.ArtifactIT) Time elapsed: 0.49 sec <<< ERROR!
java.io.FileNotFoundException: C:\Users\amedrano\.m2\repository\org\slf4j\com.springsource.slf4j.api\1.5.0\com.springsource.slf4j.api-1.5.0.jar (El sistema no puede encontrar la ruta especificada)
at java.io.FileInputStream.open(Native Method)
at java.io.FileInputStream.<init>(Unknown Source)
at org.springframework.core.io.FileSystemResource.getInputStream(FileSystemResource.java:108)
at org.springframework.osgi.test.AbstractOsgiTests.installBundle(AbstractOsgiTests.java:323)
at org.springframework.osgi.test.AbstractOsgiTests.startup(AbstractOsgiTests.java:253)
at org.springframework.osgi.test.AbstractOsgiTests.prepareTestExecution(AbstractOsgiTests.java:374)
at org.springframework.osgi.test.AbstractOsgiTests.runBare(AbstractOsgiTests.java:203)
at org.springframework.osgi.test.AbstractOsgiTests$1.protect(AbstractOsgiTests.java:184)
at junit.framework.TestResult.runProtected(TestResult.java:124)
at org.springframework.osgi.test.AbstractOsgiTests.run(AbstractOsgiTests.java:181)
at junit.framework.TestSuite.runTest(TestSuite.java:230)
at junit.framework.TestSuite.run(TestSuite.java:225)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.apache.maven.surefire.junit.JUnitTestSet.execute(JUnitTestSet.java:95)
at org.apache.maven.surefire.junit.JUnit3Provider.executeTestSet(JUnit3Provider.java:131)
at org.apache.maven.surefire.junit.JUnit3Provider.invoke(JUnit3Provider.java:108)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
Results :
Tests in error:
org.universAAL.ontology.test.ArtifactIT#testComposite FileNotFoundException C:...
--
From: _this issue has been automatically imported from our old issue tracker_ | priority | required testing bundle is not automatically installed originally opened amedranogil originally closed when building an artifact with integration tests with a clean maven local repo a required bundle is not automatically installed it is suggested to add this bundle as a forced dependency in top level poms with scope test so it gets automatically installed at least the first time testcomposite org universaal ontology test artifactit time elapsed sec lt lt lt error java io filenotfoundexception c users amedrano repository org com springsource api com springsource api jar el sistema no puede encontrar la ruta especificada at java io fileinputstream open native method at java io fileinputstream lt init gt unknown source at org springframework core io filesystemresource getinputstream filesystemresource java at org springframework osgi test abstractosgitests installbundle abstractosgitests java at org springframework osgi test abstractosgitests startup abstractosgitests java at org springframework osgi test abstractosgitests preparetestexecution abstractosgitests java at org springframework osgi test abstractosgitests runbare abstractosgitests java at org springframework osgi test abstractosgitests protect abstractosgitests java at junit framework testresult runprotected testresult java at org springframework osgi test abstractosgitests run abstractosgitests java at junit framework testsuite runtest testsuite java at junit framework testsuite run testsuite java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source at java lang reflect method invoke unknown source at org apache maven surefire junit junittestset execute junittestset java at org apache maven surefire junit executetestset java at org apache maven surefire junit invoke java at org apache maven surefire booter forkedbooter invokeproviderinsameclassloader forkedbooter java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java results tests in error org universaal ontology test artifactit testcomposite filenotfoundexception c from this issue has been automatically imported from our old issue tracker | 1 |
229,689 | 7,582,728,377 | IssuesEvent | 2018-04-25 06:10:11 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.mclaren.com - site is not usable | browser-firefox priority-important | <!-- @browser: Firefox 58.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: http://www.mclaren.com/formula1/
**Browser / Version**: Firefox 58.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Complete website doesn't load
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2017/12/a1c604c5-a2ef-4c5d-8178-f55c74245748.jpg)
_From [webcompat.com](https://webcompat.com/) with โค๏ธ_ | 1.0 | www.mclaren.com - site is not usable - <!-- @browser: Firefox 58.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: http://www.mclaren.com/formula1/
**Browser / Version**: Firefox 58.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Complete website doesn't load
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2017/12/a1c604c5-a2ef-4c5d-8178-f55c74245748.jpg)
_From [webcompat.com](https://webcompat.com/) with โค๏ธ_ | priority | site is not usable url browser version firefox operating system windows tested another browser yes problem type site is not usable description complete website doesn t load steps to reproduce from with โค๏ธ | 1 |
162,553 | 12,680,489,276 | IssuesEvent | 2020-06-19 13:48:05 | Oldes/Rebol-issues | https://api.github.com/repos/Oldes/Rebol-issues | closed | TO-VALUE function to convert unset to none | Test.written Type.wish | _Submitted by:_ **BrianH**
I want a function that can take one normal parameter of any type, which returns that parameter unevaluated, except for an #[unset!] value. When it is passed an #[unset!](or end! I suppose if one of those escape) then it should return #[none] instead.
This function will need to be native, since it would be used in core control flow to defang unsets. And I have no idea what to call it, so I'm looking for suggestions.
``` rebol
; Rebol code equivalent
func [value [any-type!]] [unless unset? :value [:value]]
```
---
<sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=2003)** [ Version: r3 master Type: Wish Platform: All Category: Native Reproduce: Always Fixed-in:none ]</sup>
<sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/2003</sup>
Comments:
---
> **Rebolbot** commented on Mar 22, 2013:
_Submitted by:_ **BrianH**
Fork suggested DEVOID for this in SO chat, which is metaphorically accurate and less whimsical than DEFANG. The downside is that it makes reference to the "void" type in C-like languages that serves a similar purpose to returning unset in Rebol, but Rebol isn't a C-like language.
Another option would be TO-VALUE since it converts non-values like #[unset!] to values like #[none]. It's a little long, but it fits in with the other TO-\* functions, though without the copying that those do. I don't want to call it VALUE because that is a common variable name.
Other suggestions would be welcome.
---
> **Rebolbot** commented on Mar 22, 2013:
_Submitted by:_ **Ladislav**
I prefer the TO-VALUE alternative.
---
> **Rebolbot** commented on Mar 22, 2013:
_Submitted by:_ **abolka**
I prefer TO-VALUE as well.
---
> **Rebolbot** commented on Mar 6, 2014:
_Submitted by:_ **BrianH**
Implemented as TO-VALUE in https://github.com/rebol/rebol/pull/205
---
> **Rebolbot** mentioned this issue on Jan 12, 2016:
> [Make NONE! values no-op in slots processing results of a conditional expression](https://github.com/Oldes/Rebol-issues/issues/2078)
---
> **Rebolbot** mentioned this issue on Jan 22, 2016:
> [[Epic] Holes in our evaluation model](https://github.com/Oldes/Rebol-issues/issues/2127)
---
> **Hostilefork** mentioned this issue on May 1, 2018:
> [Allow INDEX? to take none as an argument, return none](https://github.com/Oldes/Rebol-issues/issues/1611)
---
> **Rebolbot** added the **Type.wish** on Jan 12, 2016
---
> **Hostilefork** added a commit to **[metaeducation/ren-c](https://github.com/metaeducation/ren-c/)** that referenced this issue on May 1, 2018:
> [Retake TRY for TO-VALUE, make NULL void's API representation](https://github.com/metaeducation/ren-c/commit/04f236d3bf5eac62565d1cb9fe2186e98293c6eb)
---
> **Hostilefork** added a commit to **[hostilefork/rebol](https://github.com/hostilefork/rebol/)** that referenced this issue on May 1, 2018:
> [Retake TRY for TO-VALUE, make NULL void's API representation](https://github.com/hostilefork/rebol/commit/4e44fe58da5417d381e83f3dc779fdc59fd16942)
---
> **Hostilefork** added a commit to **[hostilefork/rebol](https://github.com/hostilefork/rebol/)** that referenced this issue on May 1, 2018:
> [Retake TRY for TO-VALUE, make NULL void's API representation](https://github.com/hostilefork/rebol/commit/20f64be4519e9e6a852329088ec115546abba828)
--- | 1.0 | TO-VALUE function to convert unset to none - _Submitted by:_ **BrianH**
I want a function that can take one normal parameter of any type, which returns that parameter unevaluated, except for an #[unset!] value. When it is passed an #[unset!](or end! I suppose if one of those escape) then it should return #[none] instead.
This function will need to be native, since it would be used in core control flow to defang unsets. And I have no idea what to call it, so I'm looking for suggestions.
``` rebol
; Rebol code equivalent
func [value [any-type!]] [unless unset? :value [:value]]
```
---
<sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=2003)** [ Version: r3 master Type: Wish Platform: All Category: Native Reproduce: Always Fixed-in:none ]</sup>
<sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/2003</sup>
Comments:
---
> **Rebolbot** commented on Mar 22, 2013:
_Submitted by:_ **BrianH**
Fork suggested DEVOID for this in SO chat, which is metaphorically accurate and less whimsical than DEFANG. The downside is that it makes reference to the "void" type in C-like languages that serves a similar purpose to returning unset in Rebol, but Rebol isn't a C-like language.
Another option would be TO-VALUE since it converts non-values like #[unset!] to values like #[none]. It's a little long, but it fits in with the other TO-\* functions, though without the copying that those do. I don't want to call it VALUE because that is a common variable name.
Other suggestions would be welcome.
---
> **Rebolbot** commented on Mar 22, 2013:
_Submitted by:_ **Ladislav**
I prefer the TO-VALUE alternative.
---
> **Rebolbot** commented on Mar 22, 2013:
_Submitted by:_ **abolka**
I prefer TO-VALUE as well.
---
> **Rebolbot** commented on Mar 6, 2014:
_Submitted by:_ **BrianH**
Implemented as TO-VALUE in https://github.com/rebol/rebol/pull/205
---
> **Rebolbot** mentioned this issue on Jan 12, 2016:
> [Make NONE! values no-op in slots processing results of a conditional expression](https://github.com/Oldes/Rebol-issues/issues/2078)
---
> **Rebolbot** mentioned this issue on Jan 22, 2016:
> [[Epic] Holes in our evaluation model](https://github.com/Oldes/Rebol-issues/issues/2127)
---
> **Hostilefork** mentioned this issue on May 1, 2018:
> [Allow INDEX? to take none as an argument, return none](https://github.com/Oldes/Rebol-issues/issues/1611)
---
> **Rebolbot** added the **Type.wish** on Jan 12, 2016
---
> **Hostilefork** added a commit to **[metaeducation/ren-c](https://github.com/metaeducation/ren-c/)** that referenced this issue on May 1, 2018:
> [Retake TRY for TO-VALUE, make NULL void's API representation](https://github.com/metaeducation/ren-c/commit/04f236d3bf5eac62565d1cb9fe2186e98293c6eb)
---
> **Hostilefork** added a commit to **[hostilefork/rebol](https://github.com/hostilefork/rebol/)** that referenced this issue on May 1, 2018:
> [Retake TRY for TO-VALUE, make NULL void's API representation](https://github.com/hostilefork/rebol/commit/4e44fe58da5417d381e83f3dc779fdc59fd16942)
---
> **Hostilefork** added a commit to **[hostilefork/rebol](https://github.com/hostilefork/rebol/)** that referenced this issue on May 1, 2018:
> [Retake TRY for TO-VALUE, make NULL void's API representation](https://github.com/hostilefork/rebol/commit/20f64be4519e9e6a852329088ec115546abba828)
--- | non_priority | to value function to convert unset to none submitted by brianh i want a function that can take one normal parameter of any type which returns that parameter unevaluated except for an value when it is passed an or end i suppose if one of those escape then it should return instead this function will need to be native since it would be used in core control flow to defang unsets and i have no idea what to call it so i m looking for suggestions rebol rebol code equivalent func imported from imported from comments rebolbot commented on mar submitted by brianh fork suggested devoid for this in so chat which is metaphorically accurate and less whimsical than defang the downside is that it makes reference to the void type in c like languages that serves a similar purpose to returning unset in rebol but rebol isn t a c like language another option would be to value since it converts non values like to values like it s a little long but it fits in with the other to functions though without the copying that those do i don t want to call it value because that is a common variable name other suggestions would be welcome rebolbot commented on mar submitted by ladislav i prefer the to value alternative rebolbot commented on mar submitted by abolka i prefer to value as well rebolbot commented on mar submitted by brianh implemented as to value in rebolbot mentioned this issue on jan rebolbot mentioned this issue on jan holes in our evaluation model hostilefork mentioned this issue on may rebolbot added the type wish on jan hostilefork added a commit to that referenced this issue on may hostilefork added a commit to that referenced this issue on may hostilefork added a commit to that referenced this issue on may | 0 |
451,799 | 32,042,143,377 | IssuesEvent | 2023-09-22 20:22:27 | gustavoprezoto/EstagBee | https://api.github.com/repos/gustavoprezoto/EstagBee | closed | Requisitos nรฃo funcionais devem ser mensurรกveis | documentation invalid | Seguindo o padrรฃo passado pela professora, os requisitos nรฃo funcionais devem ser mensurรกveis. Exemplos:
Exemplo ruim de RNF:
[RNF01] As consultas ao sistema devem ser respondidas rapidamente
(o que รฉ rapidamente?)
Exemplo bom de RNF:
[RNF02] As consultas ao sistema devem ser respondidas em menos
de trรชs segundos | 1.0 | Requisitos nรฃo funcionais devem ser mensurรกveis - Seguindo o padrรฃo passado pela professora, os requisitos nรฃo funcionais devem ser mensurรกveis. Exemplos:
Exemplo ruim de RNF:
[RNF01] As consultas ao sistema devem ser respondidas rapidamente
(o que รฉ rapidamente?)
Exemplo bom de RNF:
[RNF02] As consultas ao sistema devem ser respondidas em menos
de trรชs segundos | non_priority | requisitos nรฃo funcionais devem ser mensurรกveis seguindo o padrรฃo passado pela professora os requisitos nรฃo funcionais devem ser mensurรกveis exemplos exemplo ruim de rnf as consultas ao sistema devem ser respondidas rapidamente o que รฉ rapidamente exemplo bom de rnf as consultas ao sistema devem ser respondidas em menos de trรชs segundos | 0 |
682,209 | 23,336,596,122 | IssuesEvent | 2022-08-09 10:29:42 | dnd-side-project/dnd-7th-7-frontend | https://api.github.com/repos/dnd-side-project/dnd-7th-7-frontend | closed | [update] ๊ฒฝ๋ก ๋ฑ๋ก ํญ ๋ค๋น๊ฒ์ด์
์์ ๋ฐ ์ถ๊ฐ | ๐จ๐ผโ๐ง Type: Improvement ๐ด Priority: High | ## ๐ฅ ์ค๋ช
(์ด์์ ๋ํ ์์ธํ ์ค๋ช
์ ์ ์ด์ฃผ์ธ์!)
- ๋์์ธ์ ๋ฐ๋ผ ๋ค์ด๊ฒ์ดํฐ ํญ์ ์์ ํ๋ ์์
์
๋๋ค.
<br>
## ๐ฅ ETC (๊ธฐํ ์ฌํญ์ ์ ์ด์ฃผ์ธ์!)
<img width="461" alt="image" src="https://user-images.githubusercontent.com/80577900/183613007-d58d50a3-7bba-45fb-8d87-56a2a044cf22.png">
<br>
>#### types
>- feat : ์๋ก์ด ๊ธฐ๋ฅ
>- update : ๋ก์ง ๋ณ๊ฒฝ
>- bug : ๋ฒ๊ทธ ์์
>- refactor : ํ์ผ ๋ฐ ํด๋ ๋ฆฌํฉํ ๋ง
>- style : ์ฝ๋ ํ์ ๋ณ๊ฒฝ
>- docs : ๋ฌธ์ ๊ด๋ จ
>- test : ํ
์คํธ ์ฝ๋
>- etc : ๊ธฐํ ๋ณ๊ฒฝ ์ฌํญ
>
| 1.0 | [update] ๊ฒฝ๋ก ๋ฑ๋ก ํญ ๋ค๋น๊ฒ์ด์
์์ ๋ฐ ์ถ๊ฐ - ## ๐ฅ ์ค๋ช
(์ด์์ ๋ํ ์์ธํ ์ค๋ช
์ ์ ์ด์ฃผ์ธ์!)
- ๋์์ธ์ ๋ฐ๋ผ ๋ค์ด๊ฒ์ดํฐ ํญ์ ์์ ํ๋ ์์
์
๋๋ค.
<br>
## ๐ฅ ETC (๊ธฐํ ์ฌํญ์ ์ ์ด์ฃผ์ธ์!)
<img width="461" alt="image" src="https://user-images.githubusercontent.com/80577900/183613007-d58d50a3-7bba-45fb-8d87-56a2a044cf22.png">
<br>
>#### types
>- feat : ์๋ก์ด ๊ธฐ๋ฅ
>- update : ๋ก์ง ๋ณ๊ฒฝ
>- bug : ๋ฒ๊ทธ ์์
>- refactor : ํ์ผ ๋ฐ ํด๋ ๋ฆฌํฉํ ๋ง
>- style : ์ฝ๋ ํ์ ๋ณ๊ฒฝ
>- docs : ๋ฌธ์ ๊ด๋ จ
>- test : ํ
์คํธ ์ฝ๋
>- etc : ๊ธฐํ ๋ณ๊ฒฝ ์ฌํญ
>
| priority | ๊ฒฝ๋ก ๋ฑ๋ก ํญ ๋ค๋น๊ฒ์ด์
์์ ๋ฐ ์ถ๊ฐ ๐ฅ ์ค๋ช
์ด์์ ๋ํ ์์ธํ ์ค๋ช
์ ์ ์ด์ฃผ์ธ์ ๋์์ธ์ ๋ฐ๋ผ ๋ค์ด๊ฒ์ดํฐ ํญ์ ์์ ํ๋ ์์
์
๋๋ค ๐ฅ etc ๊ธฐํ ์ฌํญ์ ์ ์ด์ฃผ์ธ์ img width alt image src types feat ์๋ก์ด ๊ธฐ๋ฅ update ๋ก์ง ๋ณ๊ฒฝ bug ๋ฒ๊ทธ ์์ refactor ํ์ผ ๋ฐ ํด๋ ๋ฆฌํฉํ ๋ง style ์ฝ๋ ํ์ ๋ณ๊ฒฝ docs ๋ฌธ์ ๊ด๋ จ test ํ
์คํธ ์ฝ๋ etc ๊ธฐํ ๋ณ๊ฒฝ ์ฌํญ | 1 |
14,433 | 9,308,404,115 | IssuesEvent | 2019-03-25 14:29:55 | armbrustsamuel/springbootAnnotation | https://api.github.com/repos/armbrustsamuel/springbootAnnotation | opened | CVE-2017-12617 High Severity Vulnerability detected by WhiteSource | security vulnerability | ## CVE-2017-12617 - High Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.14.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>path: /root/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.14/tomcat-embed-core-8.5.14.jar</p>
<p>
<p>Library home page: <a href=http://tomcat.apache.org/>http://tomcat.apache.org/</a></p>
Dependency Hierarchy:
- spring-boot-starter-web-1.5.3.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-1.5.3.RELEASE.jar
- :x: **tomcat-embed-core-8.5.14.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When running Apache Tomcat versions 9.0.0.M1 to 9.0.0, 8.5.0 to 8.5.22, 8.0.0.RC1 to 8.0.46 and 7.0.0 to 7.0.81 with HTTP PUTs enabled (e.g. via setting the readonly initialisation parameter of the Default servlet to false) it was possible to upload a JSP file to the server via a specially crafted request. This JSP could then be requested and any code it contained would be executed by the server.
<p>Publish Date: 2017-10-04
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-12617>CVE-2017-12617</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://www.securitytracker.com/id/1039552">http://www.securitytracker.com/id/1039552</a></p>
<p>Fix Resolution: The vendor has issued a fix (7.0.82, 8.0.47, 9.0.1).
The vendor advisories are available at:
http://tomcat.apache.org/security-7.html#Fixed_in_Apache_Tomcat_7.0.82
http://tomcat.apache.org/security-8.html#Fixed_in_Apache_Tomcat_8.0.47
http://tomcat.apache.org/security-9.html#Fixed_in_Apache_Tomcat_9.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-12617 High Severity Vulnerability detected by WhiteSource - ## CVE-2017-12617 - High Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.14.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>path: /root/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.14/tomcat-embed-core-8.5.14.jar</p>
<p>
<p>Library home page: <a href=http://tomcat.apache.org/>http://tomcat.apache.org/</a></p>
Dependency Hierarchy:
- spring-boot-starter-web-1.5.3.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-1.5.3.RELEASE.jar
- :x: **tomcat-embed-core-8.5.14.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When running Apache Tomcat versions 9.0.0.M1 to 9.0.0, 8.5.0 to 8.5.22, 8.0.0.RC1 to 8.0.46 and 7.0.0 to 7.0.81 with HTTP PUTs enabled (e.g. via setting the readonly initialisation parameter of the Default servlet to false) it was possible to upload a JSP file to the server via a specially crafted request. This JSP could then be requested and any code it contained would be executed by the server.
<p>Publish Date: 2017-10-04
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-12617>CVE-2017-12617</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://www.securitytracker.com/id/1039552">http://www.securitytracker.com/id/1039552</a></p>
<p>Fix Resolution: The vendor has issued a fix (7.0.82, 8.0.47, 9.0.1).
The vendor advisories are available at:
http://tomcat.apache.org/security-7.html#Fixed_in_Apache_Tomcat_7.0.82
http://tomcat.apache.org/security-8.html#Fixed_in_Apache_Tomcat_8.0.47
http://tomcat.apache.org/security-9.html#Fixed_in_Apache_Tomcat_9.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high severity vulnerability detected by whitesource cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation path root repository org apache tomcat embed tomcat embed core tomcat embed core jar library home page a href dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library vulnerability details when running apache tomcat versions to to to and to with http puts enabled e g via setting the readonly initialisation parameter of the default servlet to false it was possible to upload a jsp file to the server via a specially crafted request this jsp could then be requested and any code it contained would be executed by the server publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href fix resolution the vendor has issued a fix the vendor advisories are available at step up your open source security game with whitesource | 0 |
699,059 | 24,002,684,770 | IssuesEvent | 2022-09-14 12:43:55 | status-im/status-desktop | https://api.github.com/repos/status-im/status-desktop | closed | Chat - Pinned messages - Pinned limit reduces after unpinning | bug priority 2: medium E:Bugfixes | # Bug Report
## Description
After a message has been unpinned then the pinned message limit is reduced by 1.
## Steps to reproduce
1. Pin 4 messages (pin limit)
2. 5th pin causes pinned limit reached prompt
3. Unpin 1 message (3 pinned messages)
4. Pin 1 more message - pin limit reached at 3 messages
(pin limit reduces each time)
#### Expected behavior
Pinned message limit remains consistent
#### Actual behavior

### Additional Information
- Status desktop version: https://ci.status.im/job/status-desktop/job/platforms/job/macos/lastSuccessfulBuild/artifact/pkg/StatusIm-Desktop-220824-111243-1c7719.dmg
- Operating System: Mac
| 1.0 | Chat - Pinned messages - Pinned limit reduces after unpinning - # Bug Report
## Description
After a message has been unpinned then the pinned message limit is reduced by 1.
## Steps to reproduce
1. Pin 4 messages (pin limit)
2. 5th pin causes pinned limit reached prompt
3. Unpin 1 message (3 pinned messages)
4. Pin 1 more message - pin limit reached at 3 messages
(pin limit reduces each time)
#### Expected behavior
Pinned message limit remains consistent
#### Actual behavior

### Additional Information
- Status desktop version: https://ci.status.im/job/status-desktop/job/platforms/job/macos/lastSuccessfulBuild/artifact/pkg/StatusIm-Desktop-220824-111243-1c7719.dmg
- Operating System: Mac
| priority | chat pinned messages pinned limit reduces after unpinning bug report description after a message has been unpinned then the pinned message limit is reduced by steps to reproduce pin messages pin limit pin causes pinned limit reached prompt unpin message pinned messages pin more message pin limit reached at messages pin limit reduces each time expected behavior pinned message limit remains consistent actual behavior additional information status desktop version operating system mac | 1 |
709,039 | 24,365,908,811 | IssuesEvent | 2022-10-03 15:08:53 | ramp4-pcar4/story-ramp | https://api.github.com/repos/ramp4-pcar4/story-ramp | closed | Add HTML5 History mode | StoryRAMP Viewer Priority: High | HTML History mode might be something to look into before launch as it removes the `#` in the middle of URLs. https://router.vuejs.org/guide/essentials/history-mode.html
_Originally posted by @spencerwahl in https://github.com/ramp4-pcar4/story-ramp/pull/94#pullrequestreview-849668096_ | 1.0 | Add HTML5 History mode - HTML History mode might be something to look into before launch as it removes the `#` in the middle of URLs. https://router.vuejs.org/guide/essentials/history-mode.html
_Originally posted by @spencerwahl in https://github.com/ramp4-pcar4/story-ramp/pull/94#pullrequestreview-849668096_ | priority | add history mode html history mode might be something to look into before launch as it removes the in the middle of urls originally posted by spencerwahl in | 1 |
627,026 | 19,850,219,176 | IssuesEvent | 2022-01-21 11:25:20 | kubeflow/kubeflow | https://api.github.com/repos/kubeflow/kubeflow | closed | Notebook culling not working with Istio and RBAC | priority/p2 area/jupyter kind/bug | /kind bug
**What steps did you take and what happened:**
We deployed Kubeflow with Istio and Istio RBAC (w/ OIDC) enabled and we were looking to use Notebook culling to cull idle notebooks after a period of time. I should also note that we have mTLS enabled on our Istio config.
Notebook culling is not functional due to multiple issues:
* Notebook controller is not on the service mesh, so fails the mTLS check when talking to notebooks
```
2020-07-13T21:11:51.792Z INFO culler Error talking to http://construction-starts.ns.svc.cluster.local/notebook/dsd/construction-starts/api/status {"error": "Get http://construction-starts.dsd.svc.cluster.local/notebook/dsd/construction-starts/api/status: read tcp 172.21.XX.YY:36004->10.0.XX.YY:80: read: connection reset by peer"}
```
* If we enable injection on the notebook controller, it is not in the ServiceRoleBinding list, so it fails the Istio RBAC check and gets permission denied
We started going down the road of enable istio-injection on the Kubeflow control plane, but this is introduced a bunch of problems. These were the initial steps we tried:
1. Add istio-injection=enabled label to kubeflow namespace
2. Rollout restart all deployments / statefulsets
3. Adding annotation to disable injection on the admission webhook and metadata-envoy-deployment
4. Delete the DestinationRule for the namespace
5. On the svc for ml-pipeline, change http service name to tcp-http
Things generally worked (after disabling injection on a couple of services), but there were random errors using Kubeflow and we didn't know of what other impacts to functionality there were. We also didn't like diverging that much from the direction of Kubeflow (https://github.com/kubeflow/manifests/pull/712).
**What did you expect to happen:**
Notebook culling should function correctly with Istio and Istio RBAC enabled.
**Anything else you would like to add:**
[Miscellaneous information that will assist in solving the issue.]
This was also mentioned in https://github.com/kubeflow/kubeflow/pull/3856#issuecomment-612295807.
**Environment:**
- Kubeflow version: (version number can be found at the bottom left corner of the Kubeflow dashboard): 1.0.2
- kfctl version: (use `kfctl version`): 1.0.2
- Kubernetes platform: (e.g. `minikube`) AKS
- Kubernetes version: (use `kubectl version`): 1.15.7
- OS (e.g. from `/etc/os-release`): Ubuntu 16.04
- Istio: 1.4.9
| 1.0 | Notebook culling not working with Istio and RBAC - /kind bug
**What steps did you take and what happened:**
We deployed Kubeflow with Istio and Istio RBAC (w/ OIDC) enabled and we were looking to use Notebook culling to cull idle notebooks after a period of time. I should also note that we have mTLS enabled on our Istio config.
Notebook culling is not functional due to multiple issues:
* Notebook controller is not on the service mesh, so fails the mTLS check when talking to notebooks
```
2020-07-13T21:11:51.792Z INFO culler Error talking to http://construction-starts.ns.svc.cluster.local/notebook/dsd/construction-starts/api/status {"error": "Get http://construction-starts.dsd.svc.cluster.local/notebook/dsd/construction-starts/api/status: read tcp 172.21.XX.YY:36004->10.0.XX.YY:80: read: connection reset by peer"}
```
* If we enable injection on the notebook controller, it is not in the ServiceRoleBinding list, so it fails the Istio RBAC check and gets permission denied
We started going down the road of enable istio-injection on the Kubeflow control plane, but this is introduced a bunch of problems. These were the initial steps we tried:
1. Add istio-injection=enabled label to kubeflow namespace
2. Rollout restart all deployments / statefulsets
3. Adding annotation to disable injection on the admission webhook and metadata-envoy-deployment
4. Delete the DestinationRule for the namespace
5. On the svc for ml-pipeline, change http service name to tcp-http
Things generally worked (after disabling injection on a couple of services), but there were random errors using Kubeflow and we didn't know of what other impacts to functionality there were. We also didn't like diverging that much from the direction of Kubeflow (https://github.com/kubeflow/manifests/pull/712).
**What did you expect to happen:**
Notebook culling should function correctly with Istio and Istio RBAC enabled.
**Anything else you would like to add:**
[Miscellaneous information that will assist in solving the issue.]
This was also mentioned in https://github.com/kubeflow/kubeflow/pull/3856#issuecomment-612295807.
**Environment:**
- Kubeflow version: (version number can be found at the bottom left corner of the Kubeflow dashboard): 1.0.2
- kfctl version: (use `kfctl version`): 1.0.2
- Kubernetes platform: (e.g. `minikube`) AKS
- Kubernetes version: (use `kubectl version`): 1.15.7
- OS (e.g. from `/etc/os-release`): Ubuntu 16.04
- Istio: 1.4.9
| priority | notebook culling not working with istio and rbac kind bug what steps did you take and what happened we deployed kubeflow with istio and istio rbac w oidc enabled and we were looking to use notebook culling to cull idle notebooks after a period of time i should also note that we have mtls enabled on our istio config notebook culling is not functional due to multiple issues notebook controller is not on the service mesh so fails the mtls check when talking to notebooks info culler error talking to error get read tcp xx yy xx yy read connection reset by peer if we enable injection on the notebook controller it is not in the servicerolebinding list so it fails the istio rbac check and gets permission denied we started going down the road of enable istio injection on the kubeflow control plane but this is introduced a bunch of problems these were the initial steps we tried add istio injection enabled label to kubeflow namespace rollout restart all deployments statefulsets adding annotation to disable injection on the admission webhook and metadata envoy deployment delete the destinationrule for the namespace on the svc for ml pipeline change http service name to tcp http things generally worked after disabling injection on a couple of services but there were random errors using kubeflow and we didn t know of what other impacts to functionality there were we also didn t like diverging that much from the direction of kubeflow what did you expect to happen notebook culling should function correctly with istio and istio rbac enabled anything else you would like to add this was also mentioned in environment kubeflow version version number can be found at the bottom left corner of the kubeflow dashboard kfctl version use kfctl version kubernetes platform e g minikube aks kubernetes version use kubectl version os e g from etc os release ubuntu istio | 1 |
378,911 | 26,344,308,124 | IssuesEvent | 2023-01-10 20:30:51 | Anselmoo/spectrafit | https://api.github.com/repos/Anselmoo/spectrafit | opened | [Docs]: Using mike for versioning docs | documentation | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Missing Information in the Docs
https://squidfunk.github.io/mkdocs-material/setup/setting-up-versioning/
### Anything else?
_No response_
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct | 1.0 | [Docs]: Using mike for versioning docs - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Missing Information in the Docs
https://squidfunk.github.io/mkdocs-material/setup/setting-up-versioning/
### Anything else?
_No response_
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct | non_priority | using mike for versioning docs is there an existing issue for this i have searched the existing issues current missing information in the docs anything else no response code of conduct i agree to follow this project s code of conduct | 0 |
123,109 | 17,772,169,225 | IssuesEvent | 2021-08-30 14:49:02 | kapseliboi/compiler-explorer | https://api.github.com/repos/kapseliboi/compiler-explorer | opened | CVE-2021-3664 (Medium) detected in url-parse-1.5.1.tgz | security vulnerability | ## CVE-2021-3664 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.1.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz</a></p>
<p>Path to dependency file: compiler-explorer/package.json</p>
<p>Path to vulnerable library: compiler-explorer/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.11.2.tgz (Root Library)
- sockjs-client-1.5.1.tgz
- :x: **url-parse-1.5.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/compiler-explorer/commit/e37a5b97c4d98423bce7e1c48beb642adb05b447">e37a5b97c4d98423bce7e1c48beb642adb05b447</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
url-parse is vulnerable to URL Redirection to Untrusted Site
<p>Publish Date: 2021-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3664>CVE-2021-3664</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664</a></p>
<p>Release Date: 2021-07-26</p>
<p>Fix Resolution: url-parse - 1.5.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-3664 (Medium) detected in url-parse-1.5.1.tgz - ## CVE-2021-3664 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.1.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz</a></p>
<p>Path to dependency file: compiler-explorer/package.json</p>
<p>Path to vulnerable library: compiler-explorer/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.11.2.tgz (Root Library)
- sockjs-client-1.5.1.tgz
- :x: **url-parse-1.5.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/compiler-explorer/commit/e37a5b97c4d98423bce7e1c48beb642adb05b447">e37a5b97c4d98423bce7e1c48beb642adb05b447</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
url-parse is vulnerable to URL Redirection to Untrusted Site
<p>Publish Date: 2021-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3664>CVE-2021-3664</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664</a></p>
<p>Release Date: 2021-07-26</p>
<p>Fix Resolution: url-parse - 1.5.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in url parse tgz cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file compiler explorer package json path to vulnerable library compiler explorer node modules url parse package json dependency hierarchy webpack dev server tgz root library sockjs client tgz x url parse tgz vulnerable library found in head commit a href found in base branch main vulnerability details url parse is vulnerable to url redirection to untrusted site publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse step up your open source security game with whitesource | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.