Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
651,449
21,478,944,226
IssuesEvent
2022-04-26 15:51:49
ucb-rit/coldfront
https://api.github.com/repos/ucb-rit/coldfront
closed
Support uploading and processing valid PIDs file in UI
enhancement low priority lrc-only
Every month, the set of valid PIDs changes. The update comes in the form of an `.xlsx` containing the now-valid PIDs. In #354, a CLI tool to process this file was added. It would be convenient for superusers to have a way to upload and process this file directly in the web UI, using the same underlying logic.
1.0
Support uploading and processing valid PIDs file in UI - Every month, the set of valid PIDs changes. The update comes in the form of an `.xlsx` containing the now-valid PIDs. In #354, a CLI tool to process this file was added. It would be convenient for superusers to have a way to upload and process this file directly in the web UI, using the same underlying logic.
priority
support uploading and processing valid pids file in ui every month the set of valid pids changes the update comes in the form of an xlsx containing the now valid pids in a cli tool to process this file was added it would be convenient for superusers to have a way to upload and process this file directly in the web ui using the same underlying logic
1
292,135
8,953,489,670
IssuesEvent
2019-01-25 19:37:17
mathematicalthinking/vmt
https://api.github.com/repos/mathematicalthinking/vmt
closed
Desmos replayer
High Priority enhancement
Desmos replayer has completely broken since implementing tabs and re-composing the workspace This shouldn't be that hard to implement.
1.0
Desmos replayer - Desmos replayer has completely broken since implementing tabs and re-composing the workspace This shouldn't be that hard to implement.
priority
desmos replayer desmos replayer has completely broken since implementing tabs and re composing the workspace this shouldn t be that hard to implement
1
101,299
4,112,497,726
IssuesEvent
2016-06-07 10:42:58
RestComm/mediaserver
https://api.github.com/repos/RestComm/mediaserver
closed
Modify debug log statement for free endpoints to include the endpoint type
enhancement help-wanted Low-Priority
Add the endpoint type to the log statement so it says: "Bridge Free endpoints 49" or "IVR Free endpoints 49"
1.0
Modify debug log statement for free endpoints to include the endpoint type - Add the endpoint type to the log statement so it says: "Bridge Free endpoints 49" or "IVR Free endpoints 49"
priority
modify debug log statement for free endpoints to include the endpoint type add the endpoint type to the log statement so it says bridge free endpoints or ivr free endpoints
1
52,667
6,649,536,161
IssuesEvent
2017-09-28 13:34:31
SecUSo/privacy-friendly-weather
https://api.github.com/repos/SecUSo/privacy-friendly-weather
closed
This location is not already in the list
bug Redesign
Hi, I already put the city 'Annecy' as a permanent location bu I can't select it from the main screen. If I try to add it again, I have the following message: 'This location is already in the list'. As you can see in the screenshot below, 'Annecy' is not here. Thanks. ![fmhgfmfz](https://cloud.githubusercontent.com/assets/5938828/20261300/7acb9e76-aa5d-11e6-96d7-e5870b664234.png)
1.0
This location is not already in the list - Hi, I already put the city 'Annecy' as a permanent location bu I can't select it from the main screen. If I try to add it again, I have the following message: 'This location is already in the list'. As you can see in the screenshot below, 'Annecy' is not here. Thanks. ![fmhgfmfz](https://cloud.githubusercontent.com/assets/5938828/20261300/7acb9e76-aa5d-11e6-96d7-e5870b664234.png)
non_priority
this location is not already in the list hi i already put the city annecy as a permanent location bu i can t select it from the main screen if i try to add it again i have the following message this location is already in the list as you can see in the screenshot below annecy is not here thanks
0
50,181
6,063,898,279
IssuesEvent
2017-06-14 13:12:48
acstech/corkboard
https://api.github.com/repos/acstech/corkboard
opened
Implement Create Post Endpoint
feature MEDIUM testable
Add ability to create a new post, but for signed in users only.
1.0
Implement Create Post Endpoint - Add ability to create a new post, but for signed in users only.
non_priority
implement create post endpoint add ability to create a new post but for signed in users only
0
6,859
3,061,892,617
IssuesEvent
2015-08-16 01:16:44
san650/ember-cli-page-object
https://api.github.com/repos/san650/ember-cli-page-object
closed
Document folder conventions
documentation question
One of the ideas @juanazam and I have is to have some conventions on where to generate the object pages. We would like to propose and document these conventions so everyone can take advantage of them. __Edit: take a look at #27 and #32__ ## Folder structure 1. Put your Page Objects in `<root>/tests/pages/` folder 2. Put you custom components in `<root>/tests/pages/components/` folder An example could be ``` myApp/tests โ””โ”€โ”€ pages ย ย  โ”œโ”€โ”€ team.js ย ย  โ”œโ”€โ”€ users.js ย ย  โ”œโ”€โ”€ users-add.js ย ย  โ””โ”€โ”€ components ย ย  ย ย  โ”œโ”€โ”€ search-box.js ย ย  ย ย  โ””โ”€โ”€ gear-menu.js ``` ## Naming conventions (?) This is something we didn't discuss yet and we would like to hear ideas. For simple routes like `users.index` it's easy to pick a name for the page object `<root>/tests/pages/user.js`, but when you have complex nested routes it's no so trivial to pick a name. It would be great to have some guidelines.
1.0
Document folder conventions - One of the ideas @juanazam and I have is to have some conventions on where to generate the object pages. We would like to propose and document these conventions so everyone can take advantage of them. __Edit: take a look at #27 and #32__ ## Folder structure 1. Put your Page Objects in `<root>/tests/pages/` folder 2. Put you custom components in `<root>/tests/pages/components/` folder An example could be ``` myApp/tests โ””โ”€โ”€ pages ย ย  โ”œโ”€โ”€ team.js ย ย  โ”œโ”€โ”€ users.js ย ย  โ”œโ”€โ”€ users-add.js ย ย  โ””โ”€โ”€ components ย ย  ย ย  โ”œโ”€โ”€ search-box.js ย ย  ย ย  โ””โ”€โ”€ gear-menu.js ``` ## Naming conventions (?) This is something we didn't discuss yet and we would like to hear ideas. For simple routes like `users.index` it's easy to pick a name for the page object `<root>/tests/pages/user.js`, but when you have complex nested routes it's no so trivial to pick a name. It would be great to have some guidelines.
non_priority
document folder conventions one of the ideas juanazam and i have is to have some conventions on where to generate the object pages we would like to propose and document these conventions so everyone can take advantage of them edit take a look at and folder structure put your page objects in tests pages folder put you custom components in tests pages components folder an example could be myapp tests โ””โ”€โ”€ pages ย ย  โ”œโ”€โ”€ team js ย ย  โ”œโ”€โ”€ users js ย ย  โ”œโ”€โ”€ users add js ย ย  โ””โ”€โ”€ components ย ย  ย ย  โ”œโ”€โ”€ search box js ย ย  ย ย  โ””โ”€โ”€ gear menu js naming conventions this is something we didn t discuss yet and we would like to hear ideas for simple routes like users index it s easy to pick a name for the page object tests pages user js but when you have complex nested routes it s no so trivial to pick a name it would be great to have some guidelines
0
579,185
17,185,350,763
IssuesEvent
2021-07-16 00:24:18
apcountryman/picolibrary-microchip-megaavr0
https://api.github.com/repos/apcountryman/picolibrary-microchip-megaavr0
closed
Add TWI multiplexed signals configuration and lookup
priority-normal status-awaiting_review type-enhancement
Add TWI multiplexed signals configuration and lookup. - [x] The ATmega4809 TWI multiplexed signals configuration and lookup functions should be defined in the `::picolibrary::Microchip::megaAVR0::Multiplexed_Signals::ATmega4809` namespace in the `include/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809/twi.h`/`source/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809/twi.cc` header/source file pair - [x] `include/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809/twi.h` should be included by `include/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809.h` The following TWI multiplexed signals configuration and lookup type aliases should be defined: - [x] `TWI_Route = Peripheral::PORTMUX::TWI_Route`: TWI routing configuration The following TWI multiplexed signals configuration and lookup functions should be defined: - [x] `void set_twi_route( Peripheral::TWI const & twi, TWI_Route route ) noexcept`: Set a TWI peripheral's routing configuration - [x] `constexpr auto scl_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's SCL pin number - [x] `constexpr auto scl_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's SCL pin mask - [x] `constexpr auto sda_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's SDA pin number - [x] `constexpr auto sda_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's SDA pin mask - [x] `atuo twi_controller_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's controller pins port - [x] `auto twi_controller_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's controller pins virtual port - [x] `auto controller_scl_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's controller SCL pin port - [x] `auto controller_scl_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's controller SCL pin virtual port - [x] `constexpr auto controller_scl_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's controller SCL pin number - [x] `constexpr auto controller_scl_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's controller SCL pin mask - [x] `auto controller_sda_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's controller SDA pin port - [x] `auto controller_sda_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's controller SDA pin virtual port - [x] `constexpr auto controller_sda_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's controller SDA pin number - [x] `constexpr auto controller_sda_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's controller SDA pin mask - [x] `auto twi_device_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's device pins port - [x] `auto twi_device_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's device pins virtual port - [x] `auto device_scl_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's device SCL pin port - [x] `auto device_scl_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's device SCL pin virtual port - [x] `constexpr auto device_scl_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's device SCL pin number - [x] `constexpr auto device_scl_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's device SCL pin mask - [x] `auto device_sda_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's device SDA pin port - [x] `auto device_sda_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's device SDA pin virtual port - [x] `constexpr auto device_sda_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's device SDA pin number - [x] `constexpr auto device_sda_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's device SDA pin mask
1.0
Add TWI multiplexed signals configuration and lookup - Add TWI multiplexed signals configuration and lookup. - [x] The ATmega4809 TWI multiplexed signals configuration and lookup functions should be defined in the `::picolibrary::Microchip::megaAVR0::Multiplexed_Signals::ATmega4809` namespace in the `include/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809/twi.h`/`source/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809/twi.cc` header/source file pair - [x] `include/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809/twi.h` should be included by `include/picolibrary/microchip/megaavr0/multiplexed_signals/atmega4809.h` The following TWI multiplexed signals configuration and lookup type aliases should be defined: - [x] `TWI_Route = Peripheral::PORTMUX::TWI_Route`: TWI routing configuration The following TWI multiplexed signals configuration and lookup functions should be defined: - [x] `void set_twi_route( Peripheral::TWI const & twi, TWI_Route route ) noexcept`: Set a TWI peripheral's routing configuration - [x] `constexpr auto scl_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's SCL pin number - [x] `constexpr auto scl_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's SCL pin mask - [x] `constexpr auto sda_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's SDA pin number - [x] `constexpr auto sda_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's SDA pin mask - [x] `atuo twi_controller_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's controller pins port - [x] `auto twi_controller_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's controller pins virtual port - [x] `auto controller_scl_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's controller SCL pin port - [x] `auto controller_scl_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's controller SCL pin virtual port - [x] `constexpr auto controller_scl_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's controller SCL pin number - [x] `constexpr auto controller_scl_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's controller SCL pin mask - [x] `auto controller_sda_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's controller SDA pin port - [x] `auto controller_sda_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's controller SDA pin virtual port - [x] `constexpr auto controller_sda_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's controller SDA pin number - [x] `constexpr auto controller_sda_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's controller SDA pin mask - [x] `auto twi_device_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's device pins port - [x] `auto twi_device_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's device pins virtual port - [x] `auto device_scl_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's device SCL pin port - [x] `auto device_scl_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's device SCL pin virtual port - [x] `constexpr auto device_scl_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's device SCL pin number - [x] `constexpr auto device_scl_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's device SCL pin mask - [x] `auto device_sda_port( Peripheral::TWI const & twi ) noexcept -> Peripheral::PORT &`: Lookup a TWI peripheral's device SDA pin port - [x] `auto device_sda_vport( Peripheral::TWI const & twi ) noexcept -> Peripheral::VPORT &`: Lookup a TWI peripheral's device SDA pin virtual port - [x] `constexpr auto device_sda_number( Peripheral::TWI const & twi ) noexcept -> std::uint_fast8_t`: Lookup a TWI peripheral's device SDA pin number - [x] `constexpr auto device_sda_mask( Peripheral::TWI const & twi ) noexcept -> std::uint8_t`: Lookup a TWI peripheral's device SDA pin mask
priority
add twi multiplexed signals configuration and lookup add twi multiplexed signals configuration and lookup the twi multiplexed signals configuration and lookup functions should be defined in the picolibrary microchip multiplexed signals namespace in the include picolibrary microchip multiplexed signals twi h source picolibrary microchip multiplexed signals twi cc header source file pair include picolibrary microchip multiplexed signals twi h should be included by include picolibrary microchip multiplexed signals h the following twi multiplexed signals configuration and lookup type aliases should be defined twi route peripheral portmux twi route twi routing configuration the following twi multiplexed signals configuration and lookup functions should be defined void set twi route peripheral twi const twi twi route route noexcept set a twi peripheral s routing configuration constexpr auto scl number peripheral twi const twi noexcept std uint t lookup a twi peripheral s scl pin number constexpr auto scl mask peripheral twi const twi noexcept std t lookup a twi peripheral s scl pin mask constexpr auto sda number peripheral twi const twi noexcept std uint t lookup a twi peripheral s sda pin number constexpr auto sda mask peripheral twi const twi noexcept std t lookup a twi peripheral s sda pin mask atuo twi controller port peripheral twi const twi noexcept peripheral port lookup a twi peripheral s controller pins port auto twi controller vport peripheral twi const twi noexcept peripheral vport lookup a twi peripheral s controller pins virtual port auto controller scl port peripheral twi const twi noexcept peripheral port lookup a twi peripheral s controller scl pin port auto controller scl vport peripheral twi const twi noexcept peripheral vport lookup a twi peripheral s controller scl pin virtual port constexpr auto controller scl number peripheral twi const twi noexcept std uint t lookup a twi peripheral s controller scl pin number constexpr auto controller scl mask peripheral twi const twi noexcept std t lookup a twi peripheral s controller scl pin mask auto controller sda port peripheral twi const twi noexcept peripheral port lookup a twi peripheral s controller sda pin port auto controller sda vport peripheral twi const twi noexcept peripheral vport lookup a twi peripheral s controller sda pin virtual port constexpr auto controller sda number peripheral twi const twi noexcept std uint t lookup a twi peripheral s controller sda pin number constexpr auto controller sda mask peripheral twi const twi noexcept std t lookup a twi peripheral s controller sda pin mask auto twi device port peripheral twi const twi noexcept peripheral port lookup a twi peripheral s device pins port auto twi device vport peripheral twi const twi noexcept peripheral vport lookup a twi peripheral s device pins virtual port auto device scl port peripheral twi const twi noexcept peripheral port lookup a twi peripheral s device scl pin port auto device scl vport peripheral twi const twi noexcept peripheral vport lookup a twi peripheral s device scl pin virtual port constexpr auto device scl number peripheral twi const twi noexcept std uint t lookup a twi peripheral s device scl pin number constexpr auto device scl mask peripheral twi const twi noexcept std t lookup a twi peripheral s device scl pin mask auto device sda port peripheral twi const twi noexcept peripheral port lookup a twi peripheral s device sda pin port auto device sda vport peripheral twi const twi noexcept peripheral vport lookup a twi peripheral s device sda pin virtual port constexpr auto device sda number peripheral twi const twi noexcept std uint t lookup a twi peripheral s device sda pin number constexpr auto device sda mask peripheral twi const twi noexcept std t lookup a twi peripheral s device sda pin mask
1
58,274
16,458,674,674
IssuesEvent
2021-05-21 15:43:31
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Room Specific functionality VS General functionality
T-Defect
![image](https://user-images.githubusercontent.com/31655082/119161189-4fe20400-ba59-11eb-8bf1-594281ef4f68.png) ### Description It's not logical to have both of these in the same place - entangled in the same area. This causes confusion and users do not know where to look for things. The video and audio call buttons are room specific just like the info button and partially the search button (also the hide widget and message pinning) this gives the user the feeling that also the notifications bell button is room specific. The most important fact though is the search all messages, this is very well hidden and has to be navigated to by going to by going to a room (yes I know users are 99% of the time in a room, but there are some cases where they are not e.g. #/home) and then switching from "This Room" to "All Rooms" which basically renders this useless for normal users. ![image](https://user-images.githubusercontent.com/31655082/119162527-c5020900-ba5a-11eb-8187-162215877e00.png) Suggestions: 1. Replace the filter room with a general search which will also filter rooms. 2. add a search all messages button in #/home ![image](https://user-images.githubusercontent.com/31655082/119163051-570a1180-ba5b-11eb-929d-be972b1c8a71.png) and replace the "Explore rooms" button with a home button (this will also make the web logic more similar to mobile) ![image](https://user-images.githubusercontent.com/31655082/119163235-7ef97500-ba5b-11eb-87bb-116e7a0e45a9.png) 3. The general search could be placed above the other room specific icons, which will also make room for other future non room specific functionality (for example a calendar), this will also solve where to put the Notifications bell button.
1.0
Room Specific functionality VS General functionality - ![image](https://user-images.githubusercontent.com/31655082/119161189-4fe20400-ba59-11eb-8bf1-594281ef4f68.png) ### Description It's not logical to have both of these in the same place - entangled in the same area. This causes confusion and users do not know where to look for things. The video and audio call buttons are room specific just like the info button and partially the search button (also the hide widget and message pinning) this gives the user the feeling that also the notifications bell button is room specific. The most important fact though is the search all messages, this is very well hidden and has to be navigated to by going to by going to a room (yes I know users are 99% of the time in a room, but there are some cases where they are not e.g. #/home) and then switching from "This Room" to "All Rooms" which basically renders this useless for normal users. ![image](https://user-images.githubusercontent.com/31655082/119162527-c5020900-ba5a-11eb-8187-162215877e00.png) Suggestions: 1. Replace the filter room with a general search which will also filter rooms. 2. add a search all messages button in #/home ![image](https://user-images.githubusercontent.com/31655082/119163051-570a1180-ba5b-11eb-929d-be972b1c8a71.png) and replace the "Explore rooms" button with a home button (this will also make the web logic more similar to mobile) ![image](https://user-images.githubusercontent.com/31655082/119163235-7ef97500-ba5b-11eb-87bb-116e7a0e45a9.png) 3. The general search could be placed above the other room specific icons, which will also make room for other future non room specific functionality (for example a calendar), this will also solve where to put the Notifications bell button.
non_priority
room specific functionality vs general functionality description it s not logical to have both of these in the same place entangled in the same area this causes confusion and users do not know where to look for things the video and audio call buttons are room specific just like the info button and partially the search button also the hide widget and message pinning this gives the user the feeling that also the notifications bell button is room specific the most important fact though is the search all messages this is very well hidden and has to be navigated to by going to by going to a room yes i know users are of the time in a room but there are some cases where they are not e g home and then switching from this room to all rooms which basically renders this useless for normal users suggestions replace the filter room with a general search which will also filter rooms add a search all messages button in home and replace the explore rooms button with a home button this will also make the web logic more similar to mobile the general search could be placed above the other room specific icons which will also make room for other future non room specific functionality for example a calendar this will also solve where to put the notifications bell button
0
99,311
8,696,946,334
IssuesEvent
2018-12-04 18:54:31
nebula-plugins/gradle-nebula-integration
https://api.github.com/repos/nebula-plugins/gradle-nebula-integration
closed
Dependency insight with preferred dependency
priority testable
Tracking gradle/gradle#7334 Related to dependency insight not showing full information on how/ where a preferred dependency contributes to the selection reason.
1.0
Dependency insight with preferred dependency - Tracking gradle/gradle#7334 Related to dependency insight not showing full information on how/ where a preferred dependency contributes to the selection reason.
non_priority
dependency insight with preferred dependency tracking gradle gradle related to dependency insight not showing full information on how where a preferred dependency contributes to the selection reason
0
201,905
15,816,491,071
IssuesEvent
2021-04-05 13:12:00
microsoft/code-with-engineering-playbook
https://api.github.com/repos/microsoft/code-with-engineering-playbook
opened
Add "Keeping documentation updated" best practices to Documentation section
documentation
After the new Documentation section is created (#533), add a new "Keeping documentation updated" document under "Best practices".
1.0
Add "Keeping documentation updated" best practices to Documentation section - After the new Documentation section is created (#533), add a new "Keeping documentation updated" document under "Best practices".
non_priority
add keeping documentation updated best practices to documentation section after the new documentation section is created add a new keeping documentation updated document under best practices
0
657,721
21,802,468,679
IssuesEvent
2022-05-16 07:12:38
teogor/teogor.github.io
https://api.github.com/repos/teogor/teogor.github.io
opened
Sitemap is not updated
@bug @priority-critical
The sitemap is not updated based on the latest values - [sitemap](teogor.github.io/sitemap)
1.0
Sitemap is not updated - The sitemap is not updated based on the latest values - [sitemap](teogor.github.io/sitemap)
priority
sitemap is not updated the sitemap is not updated based on the latest values teogor github io sitemap
1
203,266
23,131,775,238
IssuesEvent
2022-07-28 11:02:30
Gal-Doron/-gradle-main-app-gal2
https://api.github.com/repos/Gal-Doron/-gradle-main-app-gal2
closed
guava-28.1-jre.jar: 1 vulnerabilities (highest severity is: 3.3) - autoclosed
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>guava-28.1-jre.jar</b></p></summary> <p>Guava is a suite of core and expanded libraries that include utility classes, google's collections, io classes, and much much more.</p> <p>Library home page: <a href="https://github.com/google/guava">https://github.com/google/guava</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /modules-2/files-2.1/com.google.guava/guava/28.1-jre/b0e91dcb6a44ffb6221b5027e12a5cb34b841145/guava-28.1-jre.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/-gradle-main-app-gal2/commit/13111bc5f66b724f62da31e0515a2c958de72ac5">13111bc5f66b724f62da31e0515a2c958de72ac5</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-8908](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.3 | guava-28.1-jre.jar | Direct | 30.0-android | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2020-8908</summary> ### Vulnerable Library - <b>guava-28.1-jre.jar</b></p> <p>Guava is a suite of core and expanded libraries that include utility classes, google's collections, io classes, and much much more.</p> <p>Library home page: <a href="https://github.com/google/guava">https://github.com/google/guava</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /modules-2/files-2.1/com.google.guava/guava/28.1-jre/b0e91dcb6a44ffb6221b5027e12a5cb34b841145/guava-28.1-jre.jar</p> <p> Dependency Hierarchy: - :x: **guava-28.1-jre.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/-gradle-main-app-gal2/commit/13111bc5f66b724f62da31e0515a2c958de72ac5">13111bc5f66b724f62da31e0515a2c958de72ac5</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> A temp directory creation vulnerability exists in all versions of Guava, allowing an attacker with access to the machine to potentially access data in a temporary directory created by the Guava API com.google.common.io.Files.createTempDir(). By default, on unix-like systems, the created directory is world-readable (readable by an attacker with access to the system). The method in question has been marked @Deprecated in versions 30.0 and later and should not be used. For Android developers, we recommend choosing a temporary directory API provided by Android, such as context.getCacheDir(). For other Java developers, we recommend migrating to the Java 7 API java.nio.file.Files.createTempDirectory() which explicitly configures permissions of 700, or configuring the Java runtime's java.io.tmpdir system property to point to a location whose permissions are appropriately configured. <p>Publish Date: 2020-12-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908>CVE-2020-8908</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>3.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908</a></p> <p>Release Date: 2020-12-10</p> <p>Fix Resolution: 30.0-android</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
True
guava-28.1-jre.jar: 1 vulnerabilities (highest severity is: 3.3) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>guava-28.1-jre.jar</b></p></summary> <p>Guava is a suite of core and expanded libraries that include utility classes, google's collections, io classes, and much much more.</p> <p>Library home page: <a href="https://github.com/google/guava">https://github.com/google/guava</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /modules-2/files-2.1/com.google.guava/guava/28.1-jre/b0e91dcb6a44ffb6221b5027e12a5cb34b841145/guava-28.1-jre.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/-gradle-main-app-gal2/commit/13111bc5f66b724f62da31e0515a2c958de72ac5">13111bc5f66b724f62da31e0515a2c958de72ac5</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-8908](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.3 | guava-28.1-jre.jar | Direct | 30.0-android | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2020-8908</summary> ### Vulnerable Library - <b>guava-28.1-jre.jar</b></p> <p>Guava is a suite of core and expanded libraries that include utility classes, google's collections, io classes, and much much more.</p> <p>Library home page: <a href="https://github.com/google/guava">https://github.com/google/guava</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /modules-2/files-2.1/com.google.guava/guava/28.1-jre/b0e91dcb6a44ffb6221b5027e12a5cb34b841145/guava-28.1-jre.jar</p> <p> Dependency Hierarchy: - :x: **guava-28.1-jre.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/-gradle-main-app-gal2/commit/13111bc5f66b724f62da31e0515a2c958de72ac5">13111bc5f66b724f62da31e0515a2c958de72ac5</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> A temp directory creation vulnerability exists in all versions of Guava, allowing an attacker with access to the machine to potentially access data in a temporary directory created by the Guava API com.google.common.io.Files.createTempDir(). By default, on unix-like systems, the created directory is world-readable (readable by an attacker with access to the system). The method in question has been marked @Deprecated in versions 30.0 and later and should not be used. For Android developers, we recommend choosing a temporary directory API provided by Android, such as context.getCacheDir(). For other Java developers, we recommend migrating to the Java 7 API java.nio.file.Files.createTempDirectory() which explicitly configures permissions of 700, or configuring the Java runtime's java.io.tmpdir system property to point to a location whose permissions are appropriately configured. <p>Publish Date: 2020-12-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908>CVE-2020-8908</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>3.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908</a></p> <p>Release Date: 2020-12-10</p> <p>Fix Resolution: 30.0-android</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
non_priority
guava jre jar vulnerabilities highest severity is autoclosed vulnerable library guava jre jar guava is a suite of core and expanded libraries that include utility classes google s collections io classes and much much more library home page a href path to dependency file build gradle path to vulnerable library modules files com google guava guava jre guava jre jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available low guava jre jar direct android details cve vulnerable library guava jre jar guava is a suite of core and expanded libraries that include utility classes google s collections io classes and much much more library home page a href path to dependency file build gradle path to vulnerable library modules files com google guava guava jre guava jre jar dependency hierarchy x guava jre jar vulnerable library found in head commit a href found in base branch main vulnerability details a temp directory creation vulnerability exists in all versions of guava allowing an attacker with access to the machine to potentially access data in a temporary directory created by the guava api com google common io files createtempdir by default on unix like systems the created directory is world readable readable by an attacker with access to the system the method in question has been marked deprecated in versions and later and should not be used for android developers we recommend choosing a temporary directory api provided by android such as context getcachedir for other java developers we recommend migrating to the java api java nio file files createtempdirectory which explicitly configures permissions of or configuring the java runtime s java io tmpdir system property to point to a location whose permissions are appropriately configured publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
0
11,977
3,244,706,531
IssuesEvent
2015-10-16 05:10:33
commitizen/cz-cli
https://api.github.com/repos/commitizen/cz-cli
closed
Need a test suite
enhancement in progress tests
There is currently no test suite for commitizen or a way for adapters to smoke test.
1.0
Need a test suite - There is currently no test suite for commitizen or a way for adapters to smoke test.
non_priority
need a test suite there is currently no test suite for commitizen or a way for adapters to smoke test
0
259,799
27,725,073,192
IssuesEvent
2023-03-15 01:06:25
safat/conductor
https://api.github.com/repos/safat/conductor
opened
CVE-2019-20920 (High) detected in handlebars-4.0.5.js
Mend: dependency security vulnerability
## CVE-2019-20920 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.5.js</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.js</a></p> <p>Path to dependency file: /server/build/resources/main/swagger-ui/index.html</p> <p>Path to vulnerable library: /d/resources/main/swagger-ui/lib/handlebars-4.0.5.js,/server/src/main/resources/swagger-ui/lib/handlebars-4.0.5.js,/main/resources/swagger-ui/lib/handlebars-4.0.5.js,/server/build/resources/main/swagger-ui/lib/handlebars-4.0.5.js</p> <p> Dependency Hierarchy: - :x: **handlebars-4.0.5.js** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Handlebars before 3.0.8 and 4.x before 4.5.3 is vulnerable to Arbitrary Code Execution. The lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript. This can be used to run arbitrary code on a server processing Handlebars templates or in a victim's browser (effectively serving as XSS). <p>Publish Date: 2020-09-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-20920>CVE-2019-20920</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1316">https://www.npmjs.com/advisories/1316</a></p> <p>Release Date: 2020-10-15</p> <p>Fix Resolution: handlebars - 4.5.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-20920 (High) detected in handlebars-4.0.5.js - ## CVE-2019-20920 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.5.js</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.js</a></p> <p>Path to dependency file: /server/build/resources/main/swagger-ui/index.html</p> <p>Path to vulnerable library: /d/resources/main/swagger-ui/lib/handlebars-4.0.5.js,/server/src/main/resources/swagger-ui/lib/handlebars-4.0.5.js,/main/resources/swagger-ui/lib/handlebars-4.0.5.js,/server/build/resources/main/swagger-ui/lib/handlebars-4.0.5.js</p> <p> Dependency Hierarchy: - :x: **handlebars-4.0.5.js** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Handlebars before 3.0.8 and 4.x before 4.5.3 is vulnerable to Arbitrary Code Execution. The lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript. This can be used to run arbitrary code on a server processing Handlebars templates or in a victim's browser (effectively serving as XSS). <p>Publish Date: 2020-09-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-20920>CVE-2019-20920</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1316">https://www.npmjs.com/advisories/1316</a></p> <p>Release Date: 2020-10-15</p> <p>Fix Resolution: handlebars - 4.5.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in handlebars js cve high severity vulnerability vulnerable library handlebars js handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file server build resources main swagger ui index html path to vulnerable library d resources main swagger ui lib handlebars js server src main resources swagger ui lib handlebars js main resources swagger ui lib handlebars js server build resources main swagger ui lib handlebars js dependency hierarchy x handlebars js vulnerable library found in base branch master vulnerability details handlebars before and x before is vulnerable to arbitrary code execution the lookup helper fails to properly validate templates allowing attackers to submit templates that execute arbitrary javascript this can be used to run arbitrary code on a server processing handlebars templates or in a victim s browser effectively serving as xss publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars step up your open source security game with mend
0
261,556
19,716,482,317
IssuesEvent
2022-01-13 11:29:58
obss/sahi
https://api.github.com/repos/obss/sahi
reopened
Evaluation & Error Analysis
documentation
โœ”๏ธ Create COCO formatted prediction results using COCO formatted dataset: ![coco_prediction](https://user-images.githubusercontent.com/34196005/149318888-71e693ae-8b23-4ee4-80a0-b0107db7f6e7.gif) ```bash sahi predict --source image_dir/ --dataset_json_path dataset.json --model_type detectron2 --model_path weight.pt --config_path config.yaml --no_sliced_prediction ``` โœ”๏ธ Calculate COCO evaluation metrics using the created result.json: ![evaluate_1](https://user-images.githubusercontent.com/34196005/149319312-ba861c08-e46f-4c66-980c-00e67b3de626.gif) ```bash sahi coco evaluate--dataset_json_path dataset.json --result_json_path result.json ``` โœ”๏ธ Create error analysis plots using the created result.json: ![analyse](https://user-images.githubusercontent.com/34196005/149319070-5988b4ef-3be7-4c85-9b1d-e878ba4d89b5.gif) ```bash sahi coco analyse --dataset_json_path dataset.json --result_json_path result.json ``` ๐ŸŽฏ **Meaning of the metrics:** **C75:** Results at 0.75 IOU threshod **C50:** Results at 0.75 IOU threshold **Loc:** Results after ignoring localization errors **Sim:** Results after ignoring supercategory false positives **Oth:** Results after ignoring all category confusions **BG:** Results after ignoring all false positives **FN:** Results after ignoring all false negatives ๐Ÿ“ˆ **Possible model improvements:** **C75-C50 and C50-Loc**=Potential gain with more accurate bounding box prediction **Loc-Sim**=Potential gain after fixing supercategory confusions **Loc-Oth**=Potential gain after fixing category confusions **Oth-BG**=Potential gain after fixing all false positives **BG-FN**=Potential gain after fixing all false negatives
1.0
Evaluation & Error Analysis - โœ”๏ธ Create COCO formatted prediction results using COCO formatted dataset: ![coco_prediction](https://user-images.githubusercontent.com/34196005/149318888-71e693ae-8b23-4ee4-80a0-b0107db7f6e7.gif) ```bash sahi predict --source image_dir/ --dataset_json_path dataset.json --model_type detectron2 --model_path weight.pt --config_path config.yaml --no_sliced_prediction ``` โœ”๏ธ Calculate COCO evaluation metrics using the created result.json: ![evaluate_1](https://user-images.githubusercontent.com/34196005/149319312-ba861c08-e46f-4c66-980c-00e67b3de626.gif) ```bash sahi coco evaluate--dataset_json_path dataset.json --result_json_path result.json ``` โœ”๏ธ Create error analysis plots using the created result.json: ![analyse](https://user-images.githubusercontent.com/34196005/149319070-5988b4ef-3be7-4c85-9b1d-e878ba4d89b5.gif) ```bash sahi coco analyse --dataset_json_path dataset.json --result_json_path result.json ``` ๐ŸŽฏ **Meaning of the metrics:** **C75:** Results at 0.75 IOU threshod **C50:** Results at 0.75 IOU threshold **Loc:** Results after ignoring localization errors **Sim:** Results after ignoring supercategory false positives **Oth:** Results after ignoring all category confusions **BG:** Results after ignoring all false positives **FN:** Results after ignoring all false negatives ๐Ÿ“ˆ **Possible model improvements:** **C75-C50 and C50-Loc**=Potential gain with more accurate bounding box prediction **Loc-Sim**=Potential gain after fixing supercategory confusions **Loc-Oth**=Potential gain after fixing category confusions **Oth-BG**=Potential gain after fixing all false positives **BG-FN**=Potential gain after fixing all false negatives
non_priority
evaluation error analysis โœ”๏ธ create coco formatted prediction results using coco formatted dataset bash sahi predict source image dir dataset json path dataset json model type model path weight pt config path config yaml no sliced prediction โœ”๏ธ calculate coco evaluation metrics using the created result json bash sahi coco evaluate dataset json path dataset json result json path result json โœ”๏ธ create error analysis plots using the created result json bash sahi coco analyse dataset json path dataset json result json path result json ๐ŸŽฏ meaning of the metrics results at iou threshod results at iou threshold loc results after ignoring localization errors sim results after ignoring supercategory false positives oth results after ignoring all category confusions bg results after ignoring all false positives fn results after ignoring all false negatives ๐Ÿ“ˆ possible model improvements and loc potential gain with more accurate bounding box prediction loc sim potential gain after fixing supercategory confusions loc oth potential gain after fixing category confusions oth bg potential gain after fixing all false positives bg fn potential gain after fixing all false negatives
0
390,335
26,858,692,689
IssuesEvent
2023-02-03 16:31:47
tmforum-oda/oda-ca
https://api.github.com/repos/tmforum-oda/oda-ca
closed
Docs: Add Prometheus to API Operator documentation
documentation
## Description The API apiOperatorIstio includes handling Prometheus APIs and integrating to the Prometheus operator. Need to add documentation describing how this integration works.
1.0
Docs: Add Prometheus to API Operator documentation - ## Description The API apiOperatorIstio includes handling Prometheus APIs and integrating to the Prometheus operator. Need to add documentation describing how this integration works.
non_priority
docs add prometheus to api operator documentation description the api apioperatoristio includes handling prometheus apis and integrating to the prometheus operator need to add documentation describing how this integration works
0
93,668
27,012,207,630
IssuesEvent
2023-02-10 16:15:00
apache/camel-k
https://api.github.com/repos/apache/camel-k
closed
Integration Kit image build runs into an endless loop when a lower level image can't be pulled from Registry
kind/bug status/waiting-for-feedback area/build-system
Hi all. We're still working with Camel-K 1.8.0. We did experience that an Integration Kit build runs into an endless loop when a lower level image can't be pulled from Registry. The build will be tried again and again. We must delete the Operator, afterwards the Integration Kit build was running properly again. I'd expect in such a case (lower level image not found), that the Operator does have a mitigation strategy built in (e.g. build the Integration Kit based on an other one existing or from scratch). So far the question, does the Integration Kit build process be re-worked in 1.9.x. And .. we'd expect a bit more detailed log entries in the Operator logs. Best Gerald
1.0
Integration Kit image build runs into an endless loop when a lower level image can't be pulled from Registry - Hi all. We're still working with Camel-K 1.8.0. We did experience that an Integration Kit build runs into an endless loop when a lower level image can't be pulled from Registry. The build will be tried again and again. We must delete the Operator, afterwards the Integration Kit build was running properly again. I'd expect in such a case (lower level image not found), that the Operator does have a mitigation strategy built in (e.g. build the Integration Kit based on an other one existing or from scratch). So far the question, does the Integration Kit build process be re-worked in 1.9.x. And .. we'd expect a bit more detailed log entries in the Operator logs. Best Gerald
non_priority
integration kit image build runs into an endless loop when a lower level image can t be pulled from registry hi all we re still working with camel k we did experience that an integration kit build runs into an endless loop when a lower level image can t be pulled from registry the build will be tried again and again we must delete the operator afterwards the integration kit build was running properly again i d expect in such a case lower level image not found that the operator does have a mitigation strategy built in e g build the integration kit based on an other one existing or from scratch so far the question does the integration kit build process be re worked in x and we d expect a bit more detailed log entries in the operator logs best gerald
0
162,538
6,154,982,119
IssuesEvent
2017-06-28 13:54:06
k0shk0sh/FastHub
https://api.github.com/repos/k0shk0sh/FastHub
closed
Won't mark notification as read
Priority: Medium Status: Accepted Status: Completed
**App Version: 3.2.0** **OS Version: 25** **Model: LGE-Nexus 5X** I have the settings to __not__ mark as read notifications when I click on them. That way I can check them out and still have them when I reach a desktop version of GitHub. The problem is that sometimes I want to mark a notification as read because I don't have to see it when I get to the desktop, so I press the โ˜‘๏ธ to mark it and it still won't get marked. I'm guessing it happens because I have the settings set to not mark as read. _Sent from my LGE Nexus 5X using [FastHub](https://play.google.com/store/apps/details?id=com.fastaccess.github)_
1.0
Won't mark notification as read - **App Version: 3.2.0** **OS Version: 25** **Model: LGE-Nexus 5X** I have the settings to __not__ mark as read notifications when I click on them. That way I can check them out and still have them when I reach a desktop version of GitHub. The problem is that sometimes I want to mark a notification as read because I don't have to see it when I get to the desktop, so I press the โ˜‘๏ธ to mark it and it still won't get marked. I'm guessing it happens because I have the settings set to not mark as read. _Sent from my LGE Nexus 5X using [FastHub](https://play.google.com/store/apps/details?id=com.fastaccess.github)_
priority
won t mark notification as read app version os version model lge nexus i have the settings to not mark as read notifications when i click on them that way i can check them out and still have them when i reach a desktop version of github the problem is that sometimes i want to mark a notification as read because i don t have to see it when i get to the desktop so i press the โ˜‘๏ธ to mark it and it still won t get marked i m guessing it happens because i have the settings set to not mark as read sent from my lge nexus using
1
801,886
28,506,236,316
IssuesEvent
2023-04-18 21:45:20
envoyproxy/gateway
https://api.github.com/repos/envoyproxy/gateway
closed
Release 0.4.0-rc.1 install.yaml installs EG in default namespace
bug help wanted priority/high
*Description*: The install yaml in the 0.4.0-rc.1 release https://github.com/envoyproxy/gateway/releases/tag/v0.4.0-rc.1 does not have the metadata.Namespace field set to `envoy-gateway-system` installing all EG resources in the `default` namespace *Repro steps*: > Include sample requests, environment, etc. All data and inputs required to reproduce the bug. `kubectl apply -f https://github.com/envoyproxy/gateway/releases/download/v0.4.0-rc.1/install.yaml` >**Note**: If there are privacy concerns, sanitize the data prior to sharing. *Environment*: >Include the environment like gateway version, envoy version and so on. *Logs*: >Include the access logs and the Envoy logs.
1.0
Release 0.4.0-rc.1 install.yaml installs EG in default namespace - *Description*: The install yaml in the 0.4.0-rc.1 release https://github.com/envoyproxy/gateway/releases/tag/v0.4.0-rc.1 does not have the metadata.Namespace field set to `envoy-gateway-system` installing all EG resources in the `default` namespace *Repro steps*: > Include sample requests, environment, etc. All data and inputs required to reproduce the bug. `kubectl apply -f https://github.com/envoyproxy/gateway/releases/download/v0.4.0-rc.1/install.yaml` >**Note**: If there are privacy concerns, sanitize the data prior to sharing. *Environment*: >Include the environment like gateway version, envoy version and so on. *Logs*: >Include the access logs and the Envoy logs.
priority
release rc install yaml installs eg in default namespace description the install yaml in the rc release does not have the metadata namespace field set to envoy gateway system installing all eg resources in the default namespace repro steps include sample requests environment etc all data and inputs required to reproduce the bug kubectl apply f note if there are privacy concerns sanitize the data prior to sharing environment include the environment like gateway version envoy version and so on logs include the access logs and the envoy logs
1
447,459
31,711,290,553
IssuesEvent
2023-09-09 09:57:27
scrapfly/scrapfly-scrapers
https://api.github.com/repos/scrapfly/scrapfly-scrapers
closed
Trouble following instructions
documentation
Hi there, I'm having trouble following instructions from https://github.com/scrapfly/scrapfly-scrapers/tree/main/bookingcom-scraper At point 2 when trying `git clone git@github.com:scrapfly/scrapfly-scrapers.git` I got: ```git@github.com: Permission denied (publickey). fatal: Could not read from remote repository. Please make sure you have the correct access rights and the repository exists. ``` As a workaround I did `git clone https://github.com/scrapfly/scrapfly-scrapers.git` Next step is `poetry install .` I got: `-bash: poetry: command not found` I go to https://scrapfly.io/blog/how-to-scrape-bookingcom/ and as stated, I try `pip install "httpx[http2,brotli]" parsel` I got: `-bash: pip: command not found` As a workaround I did `sudo apt install python3-pip` and then `pip install "httpx[http2,brotli]" parsel` I got: ``` error: externally-managed-environment ร— This environment is externally managed โ•ฐโ”€> To install Python packages system-wide, try apt install python3-xyz, where xyz is the package you are trying to install. If you wish to install a non-Debian-packaged Python package, create a virtual environment using python3 -m venv path/to/venv. Then use path/to/venv/bin/python and path/to/venv/bin/pip. Make sure you have python3-full installed. If you wish to install a non-Debian packaged Python application, it may be easiest to use pipx install xyz, which will manage a virtual environment for you. Make sure you have pipx installed. See /usr/share/doc/python3.11/README.venv for more information. note: If you believe this is a mistake, please contact your Python installation or OS distribution provider. You can override this, at the risk of breaking your Python installation or OS, by passing --break-system-packages. hint: See PEP 668 for the detailed specification. ``` After searching how to install poetry I did: `curl -sSL https://install.python-poetry.org | python3 -`, added Poetry's bin directory in my PATH environment and tried again `poetry install .` I got: ``` Creating virtualenv scrapfly-booking-tyZw0pBk-py3.11 in /home/xxxx/.cache/pypoetry/virtualenvs No arguments expected for "install" command, got "." ``` I decided to go for step 3: `poetry run python run.py` I got: ``` Traceback (most recent call last): File "/home/esaumell/scrapfly-scrapers/bookingcom-scraper/run.py", line 12, in <module> import bookingcom File "/home/esaumell/scrapfly-scrapers/bookingcom-scraper/bookingcom.py", line 19, in <module> from loguru import logger as log ModuleNotFoundError: No module named 'loguru' ``` So brave and I went for step 4: `poetry install --with dev` I got: ``` Installing dependencies from lock file Package operations: 67 installs, 1 update, 0 removals โ€ข Installing certifi (2023.5.7) โ€ข Installing charset-normalizer (3.1.0) โ€ข Installing idna (3.4) โ€ข Installing pycparser (2.21) โ€ข Installing six (1.16.0) โ€ข Installing urllib3 (2.0.2) โ€ข Installing attrs (23.1.0) โ€ข Installing cffi (1.15.1) โ€ข Installing cssselect (1.2.0) โ€ข Installing jmespath (1.0.1) โ€ข Installing isodate (0.6.1) โ€ข Installing lxml (4.9.2) โ€ข Installing packaging (23.1) โ€ข Installing pyasn1 (0.5.0) โ€ข Installing pyparsing (3.0.9) โ€ข Installing requests (2.31.0) โ€ข Downgrading setuptools (68.1.2 -> 67.8.0) โ€ข Installing soupsieve (2.4.1) โ€ข Installing w3lib (2.1.1) โ€ข Installing webencodings (0.5.1) โ€ข Installing automat (22.10.0) โ€ข Installing beautifulsoup4 (4.12.2) โ€ข Installing constantly (15.1.0) โ€ข Installing cryptography (40.0.2) โ€ข Installing filelock (3.12.0) โ€ข Installing html5lib (1.1) โ€ข Installing hyperlink (21.0.0) โ€ข Installing incremental (22.10.0) โ€ข Installing itemadapter (0.8.0) โ€ข Installing parsel (1.8.1) โ€ข Installing pyasn1-modules (0.3.0) โ€ข Installing rdflib (6.3.2) โ€ข Installing requests-file (1.5.1) โ€ข Installing typing-extensions (4.6.1) โ€ข Installing zope-interface (6.0) โ€ข Installing html-text (0.5.2) โ€ข Installing iniconfig (2.0.0) โ€ข Installing itemloaders (1.1.0) โ€ข Installing jstyleson (0.0.2) โ€ข Installing mf2py (1.1.2) โ€ข Installing pluggy (1.0.0) โ€ข Installing protego (0.2.1) โ€ข Installing pydispatcher (2.0.7) โ€ข Installing pyopenssl (23.1.1) โ€ข Installing pyrdfa3 (3.5.3) โ€ข Installing queuelib (1.6.2) โ€ข Installing service-identity (21.1.0) โ€ข Installing tldextract (3.4.4) โ€ข Installing twisted (22.10.0) โ€ข Installing backoff (2.2.1) โ€ข Installing brotlipy (0.7.0) โ€ข Installing cchardet (2.1.7): Failed ChefBuildError Backend subprocess exited when trying to invoke build_wheel running bdist_wheel running build running build_py creating build creating build/lib.linux-aarch64-cpython-311 creating build/lib.linux-aarch64-cpython-311/cchardet copying src/cchardet/version.py -> build/lib.linux-aarch64-cpython-311/cchardet copying src/cchardet/__init__.py -> build/lib.linux-aarch64-cpython-311/cchardet running build_ext building 'cchardet._cchardet' extension creating build/temp.linux-aarch64-cpython-311 creating build/temp.linux-aarch64-cpython-311/src creating build/temp.linux-aarch64-cpython-311/src/cchardet creating build/temp.linux-aarch64-cpython-311/src/ext creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src/LangModels aarch64-linux-gnu-gcc -Wsign-compare -DNDEBUG -g -fwrapv -O2 -Wall -g -fstack-protector-strong -Wformat -Werror=format-security -g -fwrapv -O2 -fPIC -Isrc/ext/uchardet/src -I/tmp/tmp2edcn7ai/.venv/include -I/usr/include/python3.11 -c src/cchardet/_cchardet.cpp -o build/temp.linux-aarch64-cpython-311/src/cchardet/_cchardet.o src/cchardet/_cchardet.cpp:4:10: fatal error: Python.h: No such file or directory 4 | #include "Python.h" | ^~~~~~~~~~ compilation terminated. error: command '/usr/bin/aarch64-linux-gnu-gcc' failed with exit code 1 at ~/.local/share/pypoetry/venv/lib/python3.11/site-packages/poetry/installation/chef.py:147 in _prepare 143โ”‚ 144โ”‚ error = ChefBuildError("\n\n".join(message_parts)) 145โ”‚ 146โ”‚ if error is not None: โ†’ 147โ”‚ raise error from None 148โ”‚ 149โ”‚ return path 150โ”‚ 151โ”‚ def _prepare_sdist(self, archive: Path, destination: Path | None = None) -> Path: Note: This error originates from the build backend, and is likely not a problem with poetry but with cchardet (2.1.7) not supporting PEP 517 builds. You can verify this by running 'pip wheel --use-pep517 "cchardet (==2.1.7)"'. โ€ข Installing click (8.1.3) โ€ข Installing decorator (5.1.1) โ€ข Installing extruct (0.14.0) โ€ข Installing loguru (0.7.0) โ€ข Installing msgpack (1.0.5) โ€ข Installing mypy-extensions (1.0.0) โ€ข Installing pathspec (0.11.1) โ€ข Installing platformdirs (3.5.1) โ€ข Installing pytest (7.3.1) โ€ข Installing python-dateutil (2.8.2) โ€ข Installing scrapy (2.9.0) ``` As stated on the output I tried `pip wheel --use-pep517 "cchardet (==2.1.7)"` I got: ``` Collecting cchardet==2.1.7 Downloading cchardet-2.1.7.tar.gz (653 kB) โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 653.6/653.6 kB 13.5 MB/s eta 0:00:00 Installing build dependencies ... done Getting requirements to build wheel ... done Preparing metadata (pyproject.toml) ... done Building wheels for collected packages: cchardet Building wheel for cchardet (pyproject.toml) ... error error: subprocess-exited-with-error ร— Building wheel for cchardet (pyproject.toml) did not run successfully. โ”‚ exit code: 1 โ•ฐโ”€> [23 lines of output] running bdist_wheel running build running build_py creating build creating build/lib.linux-aarch64-cpython-311 creating build/lib.linux-aarch64-cpython-311/cchardet copying src/cchardet/version.py -> build/lib.linux-aarch64-cpython-311/cchardet copying src/cchardet/__init__.py -> build/lib.linux-aarch64-cpython-311/cchardet running build_ext building 'cchardet._cchardet' extension creating build/temp.linux-aarch64-cpython-311 creating build/temp.linux-aarch64-cpython-311/src creating build/temp.linux-aarch64-cpython-311/src/cchardet creating build/temp.linux-aarch64-cpython-311/src/ext creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src/LangModels aarch64-linux-gnu-gcc -Wsign-compare -DNDEBUG -g -fwrapv -O2 -Wall -g -fstack-protector-strong -Wformat -Werror=format-security -g -fwrapv -O2 -fPIC -Isrc/ext/uchardet/src -I/usr/include/python3.11 -c src/cchardet/_cchardet.cpp -o build/temp.linux-aarch64-cpython-311/src/cchardet/_cchardet.o src/cchardet/_cchardet.cpp:4:10: fatal error: Python.h: No such file or directory 4 | #include "Python.h" | ^~~~~~~~~~ compilation terminated. error: command '/usr/bin/aarch64-linux-gnu-gcc' failed with exit code 1 [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. ERROR: Failed building wheel for cchardet Failed to build cchardet ERROR: Failed to build one or more wheels ``` This is running on a fresh Debian 12 setup and I have installed poetry 1.6.1 Searching for the problem I'm suggested to try to downgrade poetry, so I try `poetry self update 1.4` but `poetry install --with dev` fails again on cchardet I also tried installing `python3-dev` without success. I don't know where to go from here. Any help would be appreciated.
1.0
Trouble following instructions - Hi there, I'm having trouble following instructions from https://github.com/scrapfly/scrapfly-scrapers/tree/main/bookingcom-scraper At point 2 when trying `git clone git@github.com:scrapfly/scrapfly-scrapers.git` I got: ```git@github.com: Permission denied (publickey). fatal: Could not read from remote repository. Please make sure you have the correct access rights and the repository exists. ``` As a workaround I did `git clone https://github.com/scrapfly/scrapfly-scrapers.git` Next step is `poetry install .` I got: `-bash: poetry: command not found` I go to https://scrapfly.io/blog/how-to-scrape-bookingcom/ and as stated, I try `pip install "httpx[http2,brotli]" parsel` I got: `-bash: pip: command not found` As a workaround I did `sudo apt install python3-pip` and then `pip install "httpx[http2,brotli]" parsel` I got: ``` error: externally-managed-environment ร— This environment is externally managed โ•ฐโ”€> To install Python packages system-wide, try apt install python3-xyz, where xyz is the package you are trying to install. If you wish to install a non-Debian-packaged Python package, create a virtual environment using python3 -m venv path/to/venv. Then use path/to/venv/bin/python and path/to/venv/bin/pip. Make sure you have python3-full installed. If you wish to install a non-Debian packaged Python application, it may be easiest to use pipx install xyz, which will manage a virtual environment for you. Make sure you have pipx installed. See /usr/share/doc/python3.11/README.venv for more information. note: If you believe this is a mistake, please contact your Python installation or OS distribution provider. You can override this, at the risk of breaking your Python installation or OS, by passing --break-system-packages. hint: See PEP 668 for the detailed specification. ``` After searching how to install poetry I did: `curl -sSL https://install.python-poetry.org | python3 -`, added Poetry's bin directory in my PATH environment and tried again `poetry install .` I got: ``` Creating virtualenv scrapfly-booking-tyZw0pBk-py3.11 in /home/xxxx/.cache/pypoetry/virtualenvs No arguments expected for "install" command, got "." ``` I decided to go for step 3: `poetry run python run.py` I got: ``` Traceback (most recent call last): File "/home/esaumell/scrapfly-scrapers/bookingcom-scraper/run.py", line 12, in <module> import bookingcom File "/home/esaumell/scrapfly-scrapers/bookingcom-scraper/bookingcom.py", line 19, in <module> from loguru import logger as log ModuleNotFoundError: No module named 'loguru' ``` So brave and I went for step 4: `poetry install --with dev` I got: ``` Installing dependencies from lock file Package operations: 67 installs, 1 update, 0 removals โ€ข Installing certifi (2023.5.7) โ€ข Installing charset-normalizer (3.1.0) โ€ข Installing idna (3.4) โ€ข Installing pycparser (2.21) โ€ข Installing six (1.16.0) โ€ข Installing urllib3 (2.0.2) โ€ข Installing attrs (23.1.0) โ€ข Installing cffi (1.15.1) โ€ข Installing cssselect (1.2.0) โ€ข Installing jmespath (1.0.1) โ€ข Installing isodate (0.6.1) โ€ข Installing lxml (4.9.2) โ€ข Installing packaging (23.1) โ€ข Installing pyasn1 (0.5.0) โ€ข Installing pyparsing (3.0.9) โ€ข Installing requests (2.31.0) โ€ข Downgrading setuptools (68.1.2 -> 67.8.0) โ€ข Installing soupsieve (2.4.1) โ€ข Installing w3lib (2.1.1) โ€ข Installing webencodings (0.5.1) โ€ข Installing automat (22.10.0) โ€ข Installing beautifulsoup4 (4.12.2) โ€ข Installing constantly (15.1.0) โ€ข Installing cryptography (40.0.2) โ€ข Installing filelock (3.12.0) โ€ข Installing html5lib (1.1) โ€ข Installing hyperlink (21.0.0) โ€ข Installing incremental (22.10.0) โ€ข Installing itemadapter (0.8.0) โ€ข Installing parsel (1.8.1) โ€ข Installing pyasn1-modules (0.3.0) โ€ข Installing rdflib (6.3.2) โ€ข Installing requests-file (1.5.1) โ€ข Installing typing-extensions (4.6.1) โ€ข Installing zope-interface (6.0) โ€ข Installing html-text (0.5.2) โ€ข Installing iniconfig (2.0.0) โ€ข Installing itemloaders (1.1.0) โ€ข Installing jstyleson (0.0.2) โ€ข Installing mf2py (1.1.2) โ€ข Installing pluggy (1.0.0) โ€ข Installing protego (0.2.1) โ€ข Installing pydispatcher (2.0.7) โ€ข Installing pyopenssl (23.1.1) โ€ข Installing pyrdfa3 (3.5.3) โ€ข Installing queuelib (1.6.2) โ€ข Installing service-identity (21.1.0) โ€ข Installing tldextract (3.4.4) โ€ข Installing twisted (22.10.0) โ€ข Installing backoff (2.2.1) โ€ข Installing brotlipy (0.7.0) โ€ข Installing cchardet (2.1.7): Failed ChefBuildError Backend subprocess exited when trying to invoke build_wheel running bdist_wheel running build running build_py creating build creating build/lib.linux-aarch64-cpython-311 creating build/lib.linux-aarch64-cpython-311/cchardet copying src/cchardet/version.py -> build/lib.linux-aarch64-cpython-311/cchardet copying src/cchardet/__init__.py -> build/lib.linux-aarch64-cpython-311/cchardet running build_ext building 'cchardet._cchardet' extension creating build/temp.linux-aarch64-cpython-311 creating build/temp.linux-aarch64-cpython-311/src creating build/temp.linux-aarch64-cpython-311/src/cchardet creating build/temp.linux-aarch64-cpython-311/src/ext creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src/LangModels aarch64-linux-gnu-gcc -Wsign-compare -DNDEBUG -g -fwrapv -O2 -Wall -g -fstack-protector-strong -Wformat -Werror=format-security -g -fwrapv -O2 -fPIC -Isrc/ext/uchardet/src -I/tmp/tmp2edcn7ai/.venv/include -I/usr/include/python3.11 -c src/cchardet/_cchardet.cpp -o build/temp.linux-aarch64-cpython-311/src/cchardet/_cchardet.o src/cchardet/_cchardet.cpp:4:10: fatal error: Python.h: No such file or directory 4 | #include "Python.h" | ^~~~~~~~~~ compilation terminated. error: command '/usr/bin/aarch64-linux-gnu-gcc' failed with exit code 1 at ~/.local/share/pypoetry/venv/lib/python3.11/site-packages/poetry/installation/chef.py:147 in _prepare 143โ”‚ 144โ”‚ error = ChefBuildError("\n\n".join(message_parts)) 145โ”‚ 146โ”‚ if error is not None: โ†’ 147โ”‚ raise error from None 148โ”‚ 149โ”‚ return path 150โ”‚ 151โ”‚ def _prepare_sdist(self, archive: Path, destination: Path | None = None) -> Path: Note: This error originates from the build backend, and is likely not a problem with poetry but with cchardet (2.1.7) not supporting PEP 517 builds. You can verify this by running 'pip wheel --use-pep517 "cchardet (==2.1.7)"'. โ€ข Installing click (8.1.3) โ€ข Installing decorator (5.1.1) โ€ข Installing extruct (0.14.0) โ€ข Installing loguru (0.7.0) โ€ข Installing msgpack (1.0.5) โ€ข Installing mypy-extensions (1.0.0) โ€ข Installing pathspec (0.11.1) โ€ข Installing platformdirs (3.5.1) โ€ข Installing pytest (7.3.1) โ€ข Installing python-dateutil (2.8.2) โ€ข Installing scrapy (2.9.0) ``` As stated on the output I tried `pip wheel --use-pep517 "cchardet (==2.1.7)"` I got: ``` Collecting cchardet==2.1.7 Downloading cchardet-2.1.7.tar.gz (653 kB) โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 653.6/653.6 kB 13.5 MB/s eta 0:00:00 Installing build dependencies ... done Getting requirements to build wheel ... done Preparing metadata (pyproject.toml) ... done Building wheels for collected packages: cchardet Building wheel for cchardet (pyproject.toml) ... error error: subprocess-exited-with-error ร— Building wheel for cchardet (pyproject.toml) did not run successfully. โ”‚ exit code: 1 โ•ฐโ”€> [23 lines of output] running bdist_wheel running build running build_py creating build creating build/lib.linux-aarch64-cpython-311 creating build/lib.linux-aarch64-cpython-311/cchardet copying src/cchardet/version.py -> build/lib.linux-aarch64-cpython-311/cchardet copying src/cchardet/__init__.py -> build/lib.linux-aarch64-cpython-311/cchardet running build_ext building 'cchardet._cchardet' extension creating build/temp.linux-aarch64-cpython-311 creating build/temp.linux-aarch64-cpython-311/src creating build/temp.linux-aarch64-cpython-311/src/cchardet creating build/temp.linux-aarch64-cpython-311/src/ext creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src creating build/temp.linux-aarch64-cpython-311/src/ext/uchardet/src/LangModels aarch64-linux-gnu-gcc -Wsign-compare -DNDEBUG -g -fwrapv -O2 -Wall -g -fstack-protector-strong -Wformat -Werror=format-security -g -fwrapv -O2 -fPIC -Isrc/ext/uchardet/src -I/usr/include/python3.11 -c src/cchardet/_cchardet.cpp -o build/temp.linux-aarch64-cpython-311/src/cchardet/_cchardet.o src/cchardet/_cchardet.cpp:4:10: fatal error: Python.h: No such file or directory 4 | #include "Python.h" | ^~~~~~~~~~ compilation terminated. error: command '/usr/bin/aarch64-linux-gnu-gcc' failed with exit code 1 [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. ERROR: Failed building wheel for cchardet Failed to build cchardet ERROR: Failed to build one or more wheels ``` This is running on a fresh Debian 12 setup and I have installed poetry 1.6.1 Searching for the problem I'm suggested to try to downgrade poetry, so I try `poetry self update 1.4` but `poetry install --with dev` fails again on cchardet I also tried installing `python3-dev` without success. I don't know where to go from here. Any help would be appreciated.
non_priority
trouble following instructions hi there i m having trouble following instructions from at point when trying git clone git github com scrapfly scrapfly scrapers git i got git github com permission denied publickey fatal could not read from remote repository please make sure you have the correct access rights and the repository exists as a workaround i did git clone next step is poetry install i got bash poetry command not found i go to and as stated i try pip install httpx parsel i got bash pip command not found as a workaround i did sudo apt install pip and then pip install httpx parsel i got error externally managed environment ร— this environment is externally managed โ•ฐโ”€ to install python packages system wide try apt install xyz where xyz is the package you are trying to install if you wish to install a non debian packaged python package create a virtual environment using m venv path to venv then use path to venv bin python and path to venv bin pip make sure you have full installed if you wish to install a non debian packaged python application it may be easiest to use pipx install xyz which will manage a virtual environment for you make sure you have pipx installed see usr share doc readme venv for more information note if you believe this is a mistake please contact your python installation or os distribution provider you can override this at the risk of breaking your python installation or os by passing break system packages hint see pep for the detailed specification after searching how to install poetry i did curl ssl added poetry s bin directory in my path environment and tried again poetry install i got creating virtualenv scrapfly booking in home xxxx cache pypoetry virtualenvs no arguments expected for install command got i decided to go for step poetry run python run py i got traceback most recent call last file home esaumell scrapfly scrapers bookingcom scraper run py line in import bookingcom file home esaumell scrapfly scrapers bookingcom scraper bookingcom py line in from loguru import logger as log modulenotfounderror no module named loguru so brave and i went for step poetry install with dev i got installing dependencies from lock file package operations installs update removals โ€ข installing certifi โ€ข installing charset normalizer โ€ข installing idna โ€ข installing pycparser โ€ข installing six โ€ข installing โ€ข installing attrs โ€ข installing cffi โ€ข installing cssselect โ€ข installing jmespath โ€ข installing isodate โ€ข installing lxml โ€ข installing packaging โ€ข installing โ€ข installing pyparsing โ€ข installing requests โ€ข downgrading setuptools โ€ข installing soupsieve โ€ข installing โ€ข installing webencodings โ€ข installing automat โ€ข installing โ€ข installing constantly โ€ข installing cryptography โ€ข installing filelock โ€ข installing โ€ข installing hyperlink โ€ข installing incremental โ€ข installing itemadapter โ€ข installing parsel โ€ข installing modules โ€ข installing rdflib โ€ข installing requests file โ€ข installing typing extensions โ€ข installing zope interface โ€ข installing html text โ€ข installing iniconfig โ€ข installing itemloaders โ€ข installing jstyleson โ€ข installing โ€ข installing pluggy โ€ข installing protego โ€ข installing pydispatcher โ€ข installing pyopenssl โ€ข installing โ€ข installing queuelib โ€ข installing service identity โ€ข installing tldextract โ€ข installing twisted โ€ข installing backoff โ€ข installing brotlipy โ€ข installing cchardet failed chefbuilderror backend subprocess exited when trying to invoke build wheel running bdist wheel running build running build py creating build creating build lib linux cpython creating build lib linux cpython cchardet copying src cchardet version py build lib linux cpython cchardet copying src cchardet init py build lib linux cpython cchardet running build ext building cchardet cchardet extension creating build temp linux cpython creating build temp linux cpython src creating build temp linux cpython src cchardet creating build temp linux cpython src ext creating build temp linux cpython src ext uchardet creating build temp linux cpython src ext uchardet src creating build temp linux cpython src ext uchardet src langmodels linux gnu gcc wsign compare dndebug g fwrapv wall g fstack protector strong wformat werror format security g fwrapv fpic isrc ext uchardet src i tmp venv include i usr include c src cchardet cchardet cpp o build temp linux cpython src cchardet cchardet o src cchardet cchardet cpp fatal error python h no such file or directory include python h compilation terminated error command usr bin linux gnu gcc failed with exit code at local share pypoetry venv lib site packages poetry installation chef py in prepare โ”‚ โ”‚ error chefbuilderror n n join message parts โ”‚ โ”‚ if error is not none โ†’ โ”‚ raise error from none โ”‚ โ”‚ return path โ”‚ โ”‚ def prepare sdist self archive path destination path none none path note this error originates from the build backend and is likely not a problem with poetry but with cchardet not supporting pep builds you can verify this by running pip wheel use cchardet โ€ข installing click โ€ข installing decorator โ€ข installing extruct โ€ข installing loguru โ€ข installing msgpack โ€ข installing mypy extensions โ€ข installing pathspec โ€ข installing platformdirs โ€ข installing pytest โ€ข installing python dateutil โ€ข installing scrapy as stated on the output i tried pip wheel use cchardet i got collecting cchardet downloading cchardet tar gz kb โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” kb mb s eta installing build dependencies done getting requirements to build wheel done preparing metadata pyproject toml done building wheels for collected packages cchardet building wheel for cchardet pyproject toml error error subprocess exited with error ร— building wheel for cchardet pyproject toml did not run successfully โ”‚ exit code โ•ฐโ”€ running bdist wheel running build running build py creating build creating build lib linux cpython creating build lib linux cpython cchardet copying src cchardet version py build lib linux cpython cchardet copying src cchardet init py build lib linux cpython cchardet running build ext building cchardet cchardet extension creating build temp linux cpython creating build temp linux cpython src creating build temp linux cpython src cchardet creating build temp linux cpython src ext creating build temp linux cpython src ext uchardet creating build temp linux cpython src ext uchardet src creating build temp linux cpython src ext uchardet src langmodels linux gnu gcc wsign compare dndebug g fwrapv wall g fstack protector strong wformat werror format security g fwrapv fpic isrc ext uchardet src i usr include c src cchardet cchardet cpp o build temp linux cpython src cchardet cchardet o src cchardet cchardet cpp fatal error python h no such file or directory include python h compilation terminated error command usr bin linux gnu gcc failed with exit code note this error originates from a subprocess and is likely not a problem with pip error failed building wheel for cchardet failed to build cchardet error failed to build one or more wheels this is running on a fresh debian setup and i have installed poetry searching for the problem i m suggested to try to downgrade poetry so i try poetry self update but poetry install with dev fails again on cchardet i also tried installing dev without success i don t know where to go from here any help would be appreciated
0
155,166
13,613,509,672
IssuesEvent
2020-09-23 11:59:59
fluid-project/floeproject.org
https://api.github.com/repos/fluid-project/floeproject.org
closed
Add screen reader or other AT details to the Technical Details section
documentation enhancement
## Describe the solution you'd like Add a section to list ATs and their versions in the technical details section of a bug report.
1.0
Add screen reader or other AT details to the Technical Details section - ## Describe the solution you'd like Add a section to list ATs and their versions in the technical details section of a bug report.
non_priority
add screen reader or other at details to the technical details section describe the solution you d like add a section to list ats and their versions in the technical details section of a bug report
0
189,075
22,046,977,735
IssuesEvent
2022-05-30 03:38:49
praneethpanasala/linux
https://api.github.com/repos/praneethpanasala/linux
closed
CVE-2019-19534 (Low) detected in linuxlinux-4.19.6 - autoclosed
security vulnerability
## CVE-2019-19534 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/praneethpanasala/linux/commits/d80c4f847c91020292cb280132b15e2ea147f1a3">d80c4f847c91020292cb280132b15e2ea147f1a3</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/can/usb/peak_usb/pcan_usb_core.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/can/usb/peak_usb/pcan_usb_core.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel before 5.3.11, there is an info-leak bug that can be caused by a malicious USB device in the drivers/net/can/usb/peak_usb/pcan_usb_core.c driver, aka CID-f7a1337f0d29. <p>Publish Date: 2019-12-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19534>CVE-2019-19534</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2019-19534">https://www.linuxkernelcves.com/cves/CVE-2019-19534</a></p> <p>Release Date: 2019-12-03</p> <p>Fix Resolution: v5.4-rc7,v3.16.79,v4.14.154,v4.19.84,v4.4.201,v4.9.201,v5.3.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-19534 (Low) detected in linuxlinux-4.19.6 - autoclosed - ## CVE-2019-19534 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/praneethpanasala/linux/commits/d80c4f847c91020292cb280132b15e2ea147f1a3">d80c4f847c91020292cb280132b15e2ea147f1a3</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/can/usb/peak_usb/pcan_usb_core.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/can/usb/peak_usb/pcan_usb_core.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel before 5.3.11, there is an info-leak bug that can be caused by a malicious USB device in the drivers/net/can/usb/peak_usb/pcan_usb_core.c driver, aka CID-f7a1337f0d29. <p>Publish Date: 2019-12-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19534>CVE-2019-19534</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2019-19534">https://www.linuxkernelcves.com/cves/CVE-2019-19534</a></p> <p>Release Date: 2019-12-03</p> <p>Fix Resolution: v5.4-rc7,v3.16.79,v4.14.154,v4.19.84,v4.4.201,v4.9.201,v5.3.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve low detected in linuxlinux autoclosed cve low severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files drivers net can usb peak usb pcan usb core c drivers net can usb peak usb pcan usb core c vulnerability details in the linux kernel before there is an info leak bug that can be caused by a malicious usb device in the drivers net can usb peak usb pcan usb core c driver aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
644,732
20,986,104,710
IssuesEvent
2022-03-29 03:27:01
EspressoSystems/cape
https://api.github.com/repos/EspressoSystems/cape
opened
Wallet :name parameter
bug priority: high wallet security wallet API unplanned
The `:name` parameter is used directly as the wallet file name, so this should be changed to something assured to be a valid file name, e.g. no path separators and length supported across platforms Linux, MacOS, and Windows.
1.0
Wallet :name parameter - The `:name` parameter is used directly as the wallet file name, so this should be changed to something assured to be a valid file name, e.g. no path separators and length supported across platforms Linux, MacOS, and Windows.
priority
wallet name parameter the name parameter is used directly as the wallet file name so this should be changed to something assured to be a valid file name e g no path separators and length supported across platforms linux macos and windows
1
18,174
10,889,362,513
IssuesEvent
2019-11-18 18:04:53
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
closed
Preliminary research
Epic Impact: 4-None Project: Warehouse Inventory Service: PM Type: Research Workgroup: AMD Workgroup: SMB
### [Warehouse Inventory Application comparison matrix](https://docs.google.com/spreadsheets/d/1r2l_KFgjygCCp1E4oaBa5lyK2QzkVcP5qCu2u85UwcY/edit#gid=2118741241) ### Candidates [Wasp Inventory Control](https://www.waspbarcode.com/inventory-control) - [x] Form submitted - [x] Phone call #173 - [x] Evaluate trial account #174 [Finale Inventory](https://www.finaleinventory.com/) - [x] Email inquiry with basic requirements & demo request - [x] [Talk with sales rep](https://github.com/cityofaustin/atd-data-tech/issues/156#issuecomment-499175173) - [x] Evaluate trial account #176 [Asset Panda](https://www.assetpanda.com/asset-tracking-government-2/) - [x] Form submitted - [x] [Talk with sales rep](https://github.com/cityofaustin/atd-data-tech/issues/156#issuecomment-499284953) - [x] Evaluate trial account #178 [Dude Solutions](https://www.dudesolutions.com/solutions/work-asset) - [x] Call with sales rep - [x] Demo #233 ------ ### Ruled out [Cin7](https://www.cin7.com/) โ€” Expensive, too tailored to ecommerce [SKU Vault](https://www.skuvault.com/) โ€” Expensive, too tailored to ecommerce [Infoplus](https://www.infopluscommerce.com/) โ€” Expensive, too tailored to ecommerce [Grainger](https://www.grainger.com/content/keepstock) โ€” Only tracks Grainger products
1.0
Preliminary research - ### [Warehouse Inventory Application comparison matrix](https://docs.google.com/spreadsheets/d/1r2l_KFgjygCCp1E4oaBa5lyK2QzkVcP5qCu2u85UwcY/edit#gid=2118741241) ### Candidates [Wasp Inventory Control](https://www.waspbarcode.com/inventory-control) - [x] Form submitted - [x] Phone call #173 - [x] Evaluate trial account #174 [Finale Inventory](https://www.finaleinventory.com/) - [x] Email inquiry with basic requirements & demo request - [x] [Talk with sales rep](https://github.com/cityofaustin/atd-data-tech/issues/156#issuecomment-499175173) - [x] Evaluate trial account #176 [Asset Panda](https://www.assetpanda.com/asset-tracking-government-2/) - [x] Form submitted - [x] [Talk with sales rep](https://github.com/cityofaustin/atd-data-tech/issues/156#issuecomment-499284953) - [x] Evaluate trial account #178 [Dude Solutions](https://www.dudesolutions.com/solutions/work-asset) - [x] Call with sales rep - [x] Demo #233 ------ ### Ruled out [Cin7](https://www.cin7.com/) โ€” Expensive, too tailored to ecommerce [SKU Vault](https://www.skuvault.com/) โ€” Expensive, too tailored to ecommerce [Infoplus](https://www.infopluscommerce.com/) โ€” Expensive, too tailored to ecommerce [Grainger](https://www.grainger.com/content/keepstock) โ€” Only tracks Grainger products
non_priority
preliminary research candidates form submitted phone call evaluate trial account email inquiry with basic requirements demo request evaluate trial account form submitted evaluate trial account call with sales rep demo ruled out โ€” expensive too tailored to ecommerce โ€” expensive too tailored to ecommerce โ€” expensive too tailored to ecommerce โ€” only tracks grainger products
0
66,124
6,989,255,021
IssuesEvent
2017-12-14 15:38:44
khartec/waltz
https://api.github.com/repos/khartec/waltz
closed
Increase Bookmark.URL field length
DDL change fixed (test & close) small change
Currently is it 255, suggestion is to increase to 500 to accommodate longer URLs
1.0
Increase Bookmark.URL field length - Currently is it 255, suggestion is to increase to 500 to accommodate longer URLs
non_priority
increase bookmark url field length currently is it suggestion is to increase to to accommodate longer urls
0
17,056
2,615,129,869
IssuesEvent
2015-03-01 05:59:38
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
closed
setPlusRequestInitializer
auto-migrated Component-Google-APIs Milestone-Version1.12.0 Priority-High Type-Enhancement
``` External references, such as a standards document, or specification? Java environments (e.g. Java 6, Android 2.3, App Engine, or All)? All Please describe the feature requested. Here's your typical builder structure: Plus plus = new Plus.Builder(httpTransport, jsonFactory, null).setGoogleClientRequestInitializer( new PlusRequestInitializer(API_KEY)).build(); But it would be easier to use if it were: Plus plus = new Plus.Builder(httpTransport, jsonFactory, null).setPlusRequestInitializer( new PlusRequestInitializer(API_KEY)).build(); In other words, add a setPlusRequestInitializer that looks something like: public Builder setPlusRequestInitializer(PlusRequestInitializer initializer) { return (Builder) setGoogleClientRequestInitializer(initializer); } Note that we'd still keep setGoogleClientRequestInitializer. ``` Original issue reported on code.google.com by `yan...@google.com` on 31 Oct 2012 at 4:10
1.0
setPlusRequestInitializer - ``` External references, such as a standards document, or specification? Java environments (e.g. Java 6, Android 2.3, App Engine, or All)? All Please describe the feature requested. Here's your typical builder structure: Plus plus = new Plus.Builder(httpTransport, jsonFactory, null).setGoogleClientRequestInitializer( new PlusRequestInitializer(API_KEY)).build(); But it would be easier to use if it were: Plus plus = new Plus.Builder(httpTransport, jsonFactory, null).setPlusRequestInitializer( new PlusRequestInitializer(API_KEY)).build(); In other words, add a setPlusRequestInitializer that looks something like: public Builder setPlusRequestInitializer(PlusRequestInitializer initializer) { return (Builder) setGoogleClientRequestInitializer(initializer); } Note that we'd still keep setGoogleClientRequestInitializer. ``` Original issue reported on code.google.com by `yan...@google.com` on 31 Oct 2012 at 4:10
priority
setplusrequestinitializer external references such as a standards document or specification java environments e g java android app engine or all all please describe the feature requested here s your typical builder structure plus plus new plus builder httptransport jsonfactory null setgoogleclientrequestinitializer new plusrequestinitializer api key build but it would be easier to use if it were plus plus new plus builder httptransport jsonfactory null setplusrequestinitializer new plusrequestinitializer api key build in other words add a setplusrequestinitializer that looks something like public builder setplusrequestinitializer plusrequestinitializer initializer return builder setgoogleclientrequestinitializer initializer note that we d still keep setgoogleclientrequestinitializer original issue reported on code google com by yan google com on oct at
1
81,480
3,591,433,657
IssuesEvent
2016-02-01 11:45:04
sa31fh/ims
https://api.github.com/repos/sa31fh/ims
closed
Expected sales keeps moving down
bug High Priority
On category_status.php, if I click on one of the categories e.g. 'Drinks' for three or more times, the 'Expected Sales' field below the table starts moving down one line with every click on a category. It doesn't matter which category/ies are clicked on and in which order.
1.0
Expected sales keeps moving down - On category_status.php, if I click on one of the categories e.g. 'Drinks' for three or more times, the 'Expected Sales' field below the table starts moving down one line with every click on a category. It doesn't matter which category/ies are clicked on and in which order.
priority
expected sales keeps moving down on category status php if i click on one of the categories e g drinks for three or more times the expected sales field below the table starts moving down one line with every click on a category it doesn t matter which category ies are clicked on and in which order
1
700,879
24,076,569,356
IssuesEvent
2022-09-18 21:45:43
json-schema-org/json-schema-spec
https://api.github.com/repos/json-schema-org/json-schema-spec
closed
Review the concept of how annotations can impact validation
Priority: High core annotation
> I'm wondering if the entire concept of annotations should be revised to something different - @karenetheridge > This is the direction I'd like to go. - @jdesrosiers > I'm happy to do that, too, but not in a patch to this version of the spec. Let's open a new issue to discuss that for the next one. - @gregsdennis _Originally posted in https://github.com/json-schema-org/json-schema-spec/issues/1172#issuecomment-1068698286_ This may be best discussed in a [GH Discussion](https://github.com/json-schema-org/community/discussions). One should be created.
1.0
Review the concept of how annotations can impact validation - > I'm wondering if the entire concept of annotations should be revised to something different - @karenetheridge > This is the direction I'd like to go. - @jdesrosiers > I'm happy to do that, too, but not in a patch to this version of the spec. Let's open a new issue to discuss that for the next one. - @gregsdennis _Originally posted in https://github.com/json-schema-org/json-schema-spec/issues/1172#issuecomment-1068698286_ This may be best discussed in a [GH Discussion](https://github.com/json-schema-org/community/discussions). One should be created.
priority
review the concept of how annotations can impact validation i m wondering if the entire concept of annotations should be revised to something different karenetheridge this is the direction i d like to go jdesrosiers i m happy to do that too but not in a patch to this version of the spec let s open a new issue to discuss that for the next one gregsdennis originally posted in this may be best discussed in a one should be created
1
255,551
8,125,417,354
IssuesEvent
2018-08-16 20:52:14
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
StreetLight shadows missing
Art Medium Priority
Looks like cast shadows got turned off on them for some reason? ![image](https://user-images.githubusercontent.com/774628/44125527-0bfb2a5c-9fe8-11e8-92a8-f8c9dae4a907.png)
1.0
StreetLight shadows missing - Looks like cast shadows got turned off on them for some reason? ![image](https://user-images.githubusercontent.com/774628/44125527-0bfb2a5c-9fe8-11e8-92a8-f8c9dae4a907.png)
priority
streetlight shadows missing looks like cast shadows got turned off on them for some reason
1
187,742
6,760,834,149
IssuesEvent
2017-10-24 22:10:11
b3aver/Automate
https://api.github.com/repos/b3aver/Automate
opened
In viewMode show Actions without the input fields
enhancement priority:minor time:medium topic:ui
Insert span tags with the saved informations. Substitute also the select fields.
1.0
In viewMode show Actions without the input fields - Insert span tags with the saved informations. Substitute also the select fields.
priority
in viewmode show actions without the input fields insert span tags with the saved informations substitute also the select fields
1
665,748
22,327,558,311
IssuesEvent
2022-06-14 12:02:30
laurencedawson/sync-for-reddit
https://api.github.com/repos/laurencedawson/sync-for-reddit
closed
[BUG] When a bottomsheet dialog is shown and the display is off, it doesn't theme
bug low-priority
For example if the app shows the updated dialog while the screen is off it may use the system coloring...
1.0
[BUG] When a bottomsheet dialog is shown and the display is off, it doesn't theme - For example if the app shows the updated dialog while the screen is off it may use the system coloring...
priority
when a bottomsheet dialog is shown and the display is off it doesn t theme for example if the app shows the updated dialog while the screen is off it may use the system coloring
1
255,113
8,108,882,352
IssuesEvent
2018-08-14 04:35:30
RhoInc/web-codebook
https://api.github.com/repos/RhoInc/web-codebook
closed
Use fixed width font to prevent y-axis label text from getting clipped.
invalid low priority
@brittsikora noticed in testing #62
1.0
Use fixed width font to prevent y-axis label text from getting clipped. - @brittsikora noticed in testing #62
priority
use fixed width font to prevent y axis label text from getting clipped brittsikora noticed in testing
1
379,060
26,351,440,800
IssuesEvent
2023-01-11 05:27:12
tossicat/tossicat-core
https://api.github.com/repos/tossicat/tossicat-core
opened
`README.md`์— ์ด ํ”„๋กœ์ ํŠธ ์žฅ์  ์ถ”๊ฐ€
documentation
ํ˜„์žฌ ์ด ํ”„๋กœ์ ํŠธ๋งŒ์˜ ์žฅ์ ์„ ์†Œ๊ฐœํ•˜๋Š” ๋ถ€๋ถ„์ด ์—†์Šต๋‹ˆ๋‹ค. ์ œ๊ฐ€ ์ƒ๊ฐํ•˜๊ธฐ์—๋Š” ๋‹ค์Œ๊ณผ ๊ฐ™์€ ์žฅ์ ์ด ์žˆ๋‹ค๊ณ  ์ƒ๊ฐํ•ฉ๋‹ˆ๋‹ค. ๋ฌผ๋ก  ์•„์ง๊นŒ์ง€ ์™„๋ฒฝํ•˜์ง€๋Š” ์•Š์ง€๋งŒ, ๊ณง ์•„๋ž˜ ๋‚ด์šฉ์„ ๋‹ค ๊ตฌํ˜„ํ•  ์ˆ˜ ์žˆ๋‹ค๊ณ  ์ƒ๊ฐํ•ฉ๋‹ˆ๋‹ค. ๋” ์ถ”๊ฐ€ํ•  ๋ถ€๋ถ„์ด๋‚˜ ์•„๋ž˜ ๋‚ด์šฉ์—์„œ ๊ณ ์น  ๋ถ€๋ถ„์ด ์žˆ๋‹ค๋ฉด ์•„๋ž˜์— ์จ์ฃผ์„ธ์š”. ```md ## ์ด ํ”„๋กœ์ ํŠธ์˜ ์žฅ์ (์•„์ง ๋‹ค ๊ตฌํ˜„ํ•˜) 1. ๋Ÿฌ์ŠคํŠธ๋กœ ๋งŒ๋“ค์–ด์„œ ๋น ๋ฆ…๋‹ˆ๋‹ค. 2. ๋ณ€ํ™˜ํ•  ๊ฒƒ๋“ค์ด ์—ฌ๋Ÿฌ ๊ฐœ ํฌํ•จ๋œ ๋ฌธ์žฅ๋„ ์ ์ ˆํ•œ ํ˜•์‹์œผ๋กœ ์ž…๋ ฅํ•˜์‹œ๋ฉด ํ•œ ๋ฒˆ์— ์ฒ˜๋ฆฌํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. 3. ์™œ ๋ณ€ํ™˜์ด ๋˜์ง€ ์•Š๋Š” ์ ์ ˆํ•˜๊ฒŒ ์—๋Ÿฌ ๋ฉ”์„ธ์ง€๋ฅผ ๋ฐ˜ํ™˜ํ•ฉ๋‹ˆ๋‹ค. ```
1.0
`README.md`์— ์ด ํ”„๋กœ์ ํŠธ ์žฅ์  ์ถ”๊ฐ€ - ํ˜„์žฌ ์ด ํ”„๋กœ์ ํŠธ๋งŒ์˜ ์žฅ์ ์„ ์†Œ๊ฐœํ•˜๋Š” ๋ถ€๋ถ„์ด ์—†์Šต๋‹ˆ๋‹ค. ์ œ๊ฐ€ ์ƒ๊ฐํ•˜๊ธฐ์—๋Š” ๋‹ค์Œ๊ณผ ๊ฐ™์€ ์žฅ์ ์ด ์žˆ๋‹ค๊ณ  ์ƒ๊ฐํ•ฉ๋‹ˆ๋‹ค. ๋ฌผ๋ก  ์•„์ง๊นŒ์ง€ ์™„๋ฒฝํ•˜์ง€๋Š” ์•Š์ง€๋งŒ, ๊ณง ์•„๋ž˜ ๋‚ด์šฉ์„ ๋‹ค ๊ตฌํ˜„ํ•  ์ˆ˜ ์žˆ๋‹ค๊ณ  ์ƒ๊ฐํ•ฉ๋‹ˆ๋‹ค. ๋” ์ถ”๊ฐ€ํ•  ๋ถ€๋ถ„์ด๋‚˜ ์•„๋ž˜ ๋‚ด์šฉ์—์„œ ๊ณ ์น  ๋ถ€๋ถ„์ด ์žˆ๋‹ค๋ฉด ์•„๋ž˜์— ์จ์ฃผ์„ธ์š”. ```md ## ์ด ํ”„๋กœ์ ํŠธ์˜ ์žฅ์ (์•„์ง ๋‹ค ๊ตฌํ˜„ํ•˜) 1. ๋Ÿฌ์ŠคํŠธ๋กœ ๋งŒ๋“ค์–ด์„œ ๋น ๋ฆ…๋‹ˆ๋‹ค. 2. ๋ณ€ํ™˜ํ•  ๊ฒƒ๋“ค์ด ์—ฌ๋Ÿฌ ๊ฐœ ํฌํ•จ๋œ ๋ฌธ์žฅ๋„ ์ ์ ˆํ•œ ํ˜•์‹์œผ๋กœ ์ž…๋ ฅํ•˜์‹œ๋ฉด ํ•œ ๋ฒˆ์— ์ฒ˜๋ฆฌํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. 3. ์™œ ๋ณ€ํ™˜์ด ๋˜์ง€ ์•Š๋Š” ์ ์ ˆํ•˜๊ฒŒ ์—๋Ÿฌ ๋ฉ”์„ธ์ง€๋ฅผ ๋ฐ˜ํ™˜ํ•ฉ๋‹ˆ๋‹ค. ```
non_priority
readme md ์— ์ด ํ”„๋กœ์ ํŠธ ์žฅ์  ์ถ”๊ฐ€ ํ˜„์žฌ ์ด ํ”„๋กœ์ ํŠธ๋งŒ์˜ ์žฅ์ ์„ ์†Œ๊ฐœํ•˜๋Š” ๋ถ€๋ถ„์ด ์—†์Šต๋‹ˆ๋‹ค ์ œ๊ฐ€ ์ƒ๊ฐํ•˜๊ธฐ์—๋Š” ๋‹ค์Œ๊ณผ ๊ฐ™์€ ์žฅ์ ์ด ์žˆ๋‹ค๊ณ  ์ƒ๊ฐํ•ฉ๋‹ˆ๋‹ค ๋ฌผ๋ก  ์•„์ง๊นŒ์ง€ ์™„๋ฒฝํ•˜์ง€๋Š” ์•Š์ง€๋งŒ ๊ณง ์•„๋ž˜ ๋‚ด์šฉ์„ ๋‹ค ๊ตฌํ˜„ํ•  ์ˆ˜ ์žˆ๋‹ค๊ณ  ์ƒ๊ฐํ•ฉ๋‹ˆ๋‹ค ๋” ์ถ”๊ฐ€ํ•  ๋ถ€๋ถ„์ด๋‚˜ ์•„๋ž˜ ๋‚ด์šฉ์—์„œ ๊ณ ์น  ๋ถ€๋ถ„์ด ์žˆ๋‹ค๋ฉด ์•„๋ž˜์— ์จ์ฃผ์„ธ์š” md ์ด ํ”„๋กœ์ ํŠธ์˜ ์žฅ์  ์•„์ง ๋‹ค ๊ตฌํ˜„ํ•˜ ๋Ÿฌ์ŠคํŠธ๋กœ ๋งŒ๋“ค์–ด์„œ ๋น ๋ฆ…๋‹ˆ๋‹ค ๋ณ€ํ™˜ํ•  ๊ฒƒ๋“ค์ด ์—ฌ๋Ÿฌ ๊ฐœ ํฌํ•จ๋œ ๋ฌธ์žฅ๋„ ์ ์ ˆํ•œ ํ˜•์‹์œผ๋กœ ์ž…๋ ฅํ•˜์‹œ๋ฉด ํ•œ ๋ฒˆ์— ์ฒ˜๋ฆฌํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค ์™œ ๋ณ€ํ™˜์ด ๋˜์ง€ ์•Š๋Š” ์ ์ ˆํ•˜๊ฒŒ ์—๋Ÿฌ ๋ฉ”์„ธ์ง€๋ฅผ ๋ฐ˜ํ™˜ํ•ฉ๋‹ˆ๋‹ค
0
75,831
21,007,491,702
IssuesEvent
2022-03-30 01:00:17
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
Build fails to make use of bazelisk, if available.
type:build/install subtype:windows TF 2.8
<em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em> **System information** - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): WINDOWS - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: - TensorFlow installed from (source or binary): https://github.com/tensorflow/tensorflow.git - TensorFlow version: r2. - Python version: 3.10 - Installed using virtualenv? pip? conda?: GIT - Bazel version (if compiling from source): 5.0 - GCC/Compiler version (if compiling from source): 11.2 - CUDA/cuDNN version: NA - GPU model and memory: NA **Describe the problem** When using the [build from source](https://www.tensorflow.org/install/source_windows) https://www.tensorflow.org/install/source_windows , Even when 'bazelisk' is installed, the install run fails from config.py stating 'Cannot find bazel. Please install bazel.', This could easly be fixed, like ``` if bazel_executable is None: bazel_executable = which('bazelisk') print('bazel_executable not found, instead using bazelisk @ ',bazel_executable) if bazel_executable is None: print('Cannot find bazel. Please install bazel.') sys.exit(1) ``` **Provide the exact sequence of commands / steps that you executed before running into the problem** Remove from path variable /Un Install bazel. Install bazelisk. execute ./config or python config.py **Any other info / logs** Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached. **Error**: 'Cannot find bazel. Please install bazel.'
1.0
Build fails to make use of bazelisk, if available. - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em> **System information** - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): WINDOWS - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: - TensorFlow installed from (source or binary): https://github.com/tensorflow/tensorflow.git - TensorFlow version: r2. - Python version: 3.10 - Installed using virtualenv? pip? conda?: GIT - Bazel version (if compiling from source): 5.0 - GCC/Compiler version (if compiling from source): 11.2 - CUDA/cuDNN version: NA - GPU model and memory: NA **Describe the problem** When using the [build from source](https://www.tensorflow.org/install/source_windows) https://www.tensorflow.org/install/source_windows , Even when 'bazelisk' is installed, the install run fails from config.py stating 'Cannot find bazel. Please install bazel.', This could easly be fixed, like ``` if bazel_executable is None: bazel_executable = which('bazelisk') print('bazel_executable not found, instead using bazelisk @ ',bazel_executable) if bazel_executable is None: print('Cannot find bazel. Please install bazel.') sys.exit(1) ``` **Provide the exact sequence of commands / steps that you executed before running into the problem** Remove from path variable /Un Install bazel. Install bazelisk. execute ./config or python config.py **Any other info / logs** Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached. **Error**: 'Cannot find bazel. Please install bazel.'
non_priority
build fails to make use of bazelisk if available please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution e g linux ubuntu windows mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device tensorflow installed from source or binary tensorflow version python version installed using virtualenv pip conda git bazel version if compiling from source gcc compiler version if compiling from source cuda cudnn version na gpu model and memory na describe the problem when using the even when bazelisk is installed the install run fails from config py stating cannot find bazel please install bazel this could easly be fixed like if bazel executable is none bazel executable which bazelisk print bazel executable not found instead using bazelisk bazel executable if bazel executable is none print cannot find bazel please install bazel sys exit provide the exact sequence of commands steps that you executed before running into the problem remove from path variable un install bazel install bazelisk execute config or python config py any other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached error cannot find bazel please install bazel
0
150,461
5,773,143,350
IssuesEvent
2017-04-28 00:40:08
coreos/flannel
https://api.github.com/repos/coreos/flannel
closed
Add consistent flannel configuration.
area/configuration help wanted priority/Pmaybe
Hi, it is extremely annoying to configure flannel from cloud-config. When doing flannel configuration, you have to think about YAML, JSON, SHELL syntaxes and catch bugs in different files. Please add consistent configuration for flannel in single file.
1.0
Add consistent flannel configuration. - Hi, it is extremely annoying to configure flannel from cloud-config. When doing flannel configuration, you have to think about YAML, JSON, SHELL syntaxes and catch bugs in different files. Please add consistent configuration for flannel in single file.
priority
add consistent flannel configuration hi it is extremely annoying to configure flannel from cloud config when doing flannel configuration you have to think about yaml json shell syntaxes and catch bugs in different files please add consistent configuration for flannel in single file
1
174,424
14,481,915,642
IssuesEvent
2020-12-10 13:18:18
OpenEnergyPlatform/open-MaStR
https://api.github.com/repos/OpenEnergyPlatform/open-MaStR
closed
streamline docs
:book: documentation
### Tasks - [x] Move README content partially to RTD - [x] Think about including remaining README into index.rst - [x] Sort content in sections download, post-processing, analyze with content in getting started
1.0
streamline docs - ### Tasks - [x] Move README content partially to RTD - [x] Think about including remaining README into index.rst - [x] Sort content in sections download, post-processing, analyze with content in getting started
non_priority
streamline docs tasks move readme content partially to rtd think about including remaining readme into index rst sort content in sections download post processing analyze with content in getting started
0
502,599
14,562,721,282
IssuesEvent
2020-12-17 00:44:17
Sage-Bionetworks/sageseqr
https://api.github.com/repos/Sage-Bionetworks/sageseqr
closed
Use githubr to version control markdown
low priority wontfix
The `githubr` package can point to a specific commit of a file in a repository. The package loads the markdown code into the users working directory and is currently hard coded as "sageseqr-report.Rmd". To track the correct version via a commit using `githubr`, the user would need to push the changed file to their repo before running the plan. This is difficult to message to the user and may over complicate the workflow. On the other hand, it might be possible to either `push` the file as a step in the workflow if the repository argument is provided. For now - marking this as low priority since the `sageseqr` package version is tracked in the activity provenance and I think the general use case for the workflow will be to run it with minimal modifications to the markdown document.
1.0
Use githubr to version control markdown - The `githubr` package can point to a specific commit of a file in a repository. The package loads the markdown code into the users working directory and is currently hard coded as "sageseqr-report.Rmd". To track the correct version via a commit using `githubr`, the user would need to push the changed file to their repo before running the plan. This is difficult to message to the user and may over complicate the workflow. On the other hand, it might be possible to either `push` the file as a step in the workflow if the repository argument is provided. For now - marking this as low priority since the `sageseqr` package version is tracked in the activity provenance and I think the general use case for the workflow will be to run it with minimal modifications to the markdown document.
priority
use githubr to version control markdown the githubr package can point to a specific commit of a file in a repository the package loads the markdown code into the users working directory and is currently hard coded as sageseqr report rmd to track the correct version via a commit using githubr the user would need to push the changed file to their repo before running the plan this is difficult to message to the user and may over complicate the workflow on the other hand it might be possible to either push the file as a step in the workflow if the repository argument is provided for now marking this as low priority since the sageseqr package version is tracked in the activity provenance and i think the general use case for the workflow will be to run it with minimal modifications to the markdown document
1
468,315
13,465,137,065
IssuesEvent
2020-09-09 20:21:32
googleapis/python-bigquery
https://api.github.com/repos/googleapis/python-bigquery
closed
Using string array with magic block results in whitespace removal from first entry of the array
api: bigquery priority: p2 type: bug
On passing a dict ` { 'cats': ['apple orange', 'pear plum']}` as params in bigquery magic cell, the first value is changed to `appleorange` - the space character is filtered out. #### Environment details - OS type and version: colab - Python version: Python 3.6.9 - pip version: 19.3.1 - `google-cloud-bigquery` version: 1.21.0 #### Steps to reproduce A colab notebook illustrating the error: https://colab.research.google.com/gist/amardeep/63ec303ba8bac3db9849f4044cd19ff1/test-bigquery-array-parameter-bug.ipynb #### Code example ```python params = { 'cats': ['apple orange', 'pear plum'] } %%bigquery --params $params SELECT * FROM UNNEST(@cats) ``` This results in the output: ``` f0_ 0 appleorange 1 pear plum ```
1.0
Using string array with magic block results in whitespace removal from first entry of the array - On passing a dict ` { 'cats': ['apple orange', 'pear plum']}` as params in bigquery magic cell, the first value is changed to `appleorange` - the space character is filtered out. #### Environment details - OS type and version: colab - Python version: Python 3.6.9 - pip version: 19.3.1 - `google-cloud-bigquery` version: 1.21.0 #### Steps to reproduce A colab notebook illustrating the error: https://colab.research.google.com/gist/amardeep/63ec303ba8bac3db9849f4044cd19ff1/test-bigquery-array-parameter-bug.ipynb #### Code example ```python params = { 'cats': ['apple orange', 'pear plum'] } %%bigquery --params $params SELECT * FROM UNNEST(@cats) ``` This results in the output: ``` f0_ 0 appleorange 1 pear plum ```
priority
using string array with magic block results in whitespace removal from first entry of the array on passing a dict cats as params in bigquery magic cell the first value is changed to appleorange the space character is filtered out environment details os type and version colab python version python pip version google cloud bigquery version steps to reproduce a colab notebook illustrating the error code example python params cats bigquery params params select from unnest cats this results in the output appleorange pear plum
1
689,006
23,603,752,717
IssuesEvent
2022-08-24 06:13:49
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Missing information on the Dry run report generation configuration in migration guide
Priority/Low docs Severity/Major Component/Migration 6.0.0-Migration Affected-6.0.0 QA-Reported 5.11-migration 6.0.0-post-ga
**Describe the issue:** During a Migration testing if you want to generate dry run reports that can be done from the migration-config.yaml. But migration guide doesnt include the correct steps with example on how to do that. https://is.docs.wso2.com/en/latest/setup/migrating-userstore-managers/#how-to-run Dry Run section To generate dry run reports we have to pass the reportPath as a string value and this reportPath always needs to be defined under a **parameter** config. If we dont pass it as a string or dont include it under a **parameter** config report generatio will not get success and you will end up in error saying migration process stopped. So lets include a sample config for the correct way of doing it <IS_HOME>/migration-resources/migration-config.yaml ``` name: "GroupsAndRolesMigrator" order: 4 parameters: # Absolute path for the dry report. This is required in the dry run mode. reportPath: "/home/ubuntu/WSO2/IS600/MigrationTesting/reports" ``` - [x] Mention you need to pass the reportPath as a string value - [x] Mention you need to define reportPath under a **parameter** config - [x] Mention that you need to uncomment and include reportPath in the migration-config.yaml uner each migrator that you wants to get the reports generated ( At the moment there are 5 occurences in the migration-config.yaml to uncomment) Attaching sample config file for reference [resources.zip](https://github.com/wso2/product-is/files/8888114/resources.zip)
1.0
Missing information on the Dry run report generation configuration in migration guide - **Describe the issue:** During a Migration testing if you want to generate dry run reports that can be done from the migration-config.yaml. But migration guide doesnt include the correct steps with example on how to do that. https://is.docs.wso2.com/en/latest/setup/migrating-userstore-managers/#how-to-run Dry Run section To generate dry run reports we have to pass the reportPath as a string value and this reportPath always needs to be defined under a **parameter** config. If we dont pass it as a string or dont include it under a **parameter** config report generatio will not get success and you will end up in error saying migration process stopped. So lets include a sample config for the correct way of doing it <IS_HOME>/migration-resources/migration-config.yaml ``` name: "GroupsAndRolesMigrator" order: 4 parameters: # Absolute path for the dry report. This is required in the dry run mode. reportPath: "/home/ubuntu/WSO2/IS600/MigrationTesting/reports" ``` - [x] Mention you need to pass the reportPath as a string value - [x] Mention you need to define reportPath under a **parameter** config - [x] Mention that you need to uncomment and include reportPath in the migration-config.yaml uner each migrator that you wants to get the reports generated ( At the moment there are 5 occurences in the migration-config.yaml to uncomment) Attaching sample config file for reference [resources.zip](https://github.com/wso2/product-is/files/8888114/resources.zip)
priority
missing information on the dry run report generation configuration in migration guide describe the issue during a migration testing if you want to generate dry run reports that can be done from the migration config yaml but migration guide doesnt include the correct steps with example on how to do that dry run section to generate dry run reports we have to pass the reportpath as a string value and this reportpath always needs to be defined under a parameter config if we dont pass it as a string or dont include it under a parameter config report generatio will not get success and you will end up in error saying migration process stopped so lets include a sample config for the correct way of doing it migration resources migration config yaml name groupsandrolesmigrator order parameters absolute path for the dry report this is required in the dry run mode reportpath home ubuntu migrationtesting reports mention you need to pass the reportpath as a string value mention you need to define reportpath under a parameter config mention that you need to uncomment and include reportpath in the migration config yaml uner each migrator that you wants to get the reports generated at the moment there are occurences in the migration config yaml to uncomment attaching sample config file for reference
1
266,259
20,143,479,793
IssuesEvent
2022-02-09 03:24:14
Websoft9/ansible-ruby
https://api.github.com/repos/Websoft9/ansible-ruby
opened
Some pages are not fully translated
documentation
## Language - [ ] Chinese - [x] English ## Content error * URL: https://support.websoft9.com/docs/ruby/solution-upgrade.html#upgrade-ruby * ![image](https://user-images.githubusercontent.com/54173243/153116018-4add2a89-0791-4e4c-82ff-58cdc5a46480.png) ## Content optimization Provide us with content optimization, e.g easy to read, simplify the steps, step by step, precise expression. * URL: * Description: ## New content Describe important content what do you think is important in this documentation. **What content do you want to add?** (describe here) **Steps for the content** 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. Modify file '....'
1.0
Some pages are not fully translated - ## Language - [ ] Chinese - [x] English ## Content error * URL: https://support.websoft9.com/docs/ruby/solution-upgrade.html#upgrade-ruby * ![image](https://user-images.githubusercontent.com/54173243/153116018-4add2a89-0791-4e4c-82ff-58cdc5a46480.png) ## Content optimization Provide us with content optimization, e.g easy to read, simplify the steps, step by step, precise expression. * URL: * Description: ## New content Describe important content what do you think is important in this documentation. **What content do you want to add?** (describe here) **Steps for the content** 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. Modify file '....'
non_priority
some pages are not fully translated language chinese english content error url content optimization provide us with content optimization e g easy to read simplify the steps step by step precise expression url description new content describe important content what do you think is important in this documentation what content do you want to add describe here steps for the content go to click on scroll down to modify file
0
8,063
5,376,205,928
IssuesEvent
2017-02-23 08:23:13
apinf/platform
https://api.github.com/repos/apinf/platform
closed
In Suomi setting, Filtering option in API catalog needs Finnish translations
i18n/l10n in progress Usability Issue
_From @Nazarah on February 1, 2017 7:30_ ![image](https://cloud.githubusercontent.com/assets/17976243/22498444/2bab81ac-e861-11e6-9c72-94e22f8f8ff0.png) ![image](https://cloud.githubusercontent.com/assets/17976243/22498472/46154140-e861-11e6-9a06-6a166012ffb3.png) _Copied from original issue: Digipalvelutehdas/APIKA#337_
True
In Suomi setting, Filtering option in API catalog needs Finnish translations - _From @Nazarah on February 1, 2017 7:30_ ![image](https://cloud.githubusercontent.com/assets/17976243/22498444/2bab81ac-e861-11e6-9c72-94e22f8f8ff0.png) ![image](https://cloud.githubusercontent.com/assets/17976243/22498472/46154140-e861-11e6-9a06-6a166012ffb3.png) _Copied from original issue: Digipalvelutehdas/APIKA#337_
non_priority
in suomi setting filtering option in api catalog needs finnish translations from nazarah on february copied from original issue digipalvelutehdas apika
0
264,910
8,321,332,077
IssuesEvent
2018-09-25 23:35:02
phenoscape/phenoscape-kb-services
https://api.github.com/repos/phenoscape/phenoscape-kb-services
closed
Apiary "Console" view doesn't work for Term Search in Firefox
bug low priority
In Firefox, the other API methods do work, just [the term search](http://docs.phenoscapekb.apiary.io/#reference/terms/term-search/search-for-terms-in-the-kb-by-text-match-on-a-property-value.-`rdfs:label`-is-searched-by-default.?console=1) doesn't. It does work in Chrome and Safari (cc @xu-hong). Not sure why Firefox is an issue. Given that this is specific to one browser and is mostly controlled by Apiary, I don't think it qualifies as a priority. But perhaps there is something that you can see as a difference in the Term Search that might be responsible for this.
1.0
Apiary "Console" view doesn't work for Term Search in Firefox - In Firefox, the other API methods do work, just [the term search](http://docs.phenoscapekb.apiary.io/#reference/terms/term-search/search-for-terms-in-the-kb-by-text-match-on-a-property-value.-`rdfs:label`-is-searched-by-default.?console=1) doesn't. It does work in Chrome and Safari (cc @xu-hong). Not sure why Firefox is an issue. Given that this is specific to one browser and is mostly controlled by Apiary, I don't think it qualifies as a priority. But perhaps there is something that you can see as a difference in the Term Search that might be responsible for this.
priority
apiary console view doesn t work for term search in firefox in firefox the other api methods do work just doesn t it does work in chrome and safari cc xu hong not sure why firefox is an issue given that this is specific to one browser and is mostly controlled by apiary i don t think it qualifies as a priority but perhaps there is something that you can see as a difference in the term search that might be responsible for this
1
347,767
24,898,371,024
IssuesEvent
2022-10-28 18:05:55
SeekyCt/ppcdis
https://api.github.com/repos/SeekyCt/ppcdis
opened
Document .ctors and .dtors weirdness
documentation high priority
There's a bunch of features / methods to work around the MWLD weirdness with the .ctors and .dtors sections that aren't really explained anywhere yet
1.0
Document .ctors and .dtors weirdness - There's a bunch of features / methods to work around the MWLD weirdness with the .ctors and .dtors sections that aren't really explained anywhere yet
non_priority
document ctors and dtors weirdness there s a bunch of features methods to work around the mwld weirdness with the ctors and dtors sections that aren t really explained anywhere yet
0
663,782
22,206,461,399
IssuesEvent
2022-06-07 15:14:49
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
closed
FireEdge onprem HCI errors
Type: Bug Status: Accepted Priority: Normal Category: Provision Category: FireEdge
**Description** HCI onprem cluster deployment is affected by a couple of minor error/typos: - The provider filed at `usr/share/one/oneprovision/edge-clusters/metal/provisions/onprem.d/fireedge.yml` should be `onprem` instead of `oneprem`. - When any of the `Hosts to run ...` fields are empty it will result in the creation of a host named `undefined`. **Details** - Affected Component: FireEdge - OneProvision - HCI Deployment - Version: 6.4.0 **Additional context** Add any other context about the problem here. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Code committed - [ ] Testing - QA - [ ] Documentation (Release notes - resolved issues, compatibility, known issues)
1.0
FireEdge onprem HCI errors - **Description** HCI onprem cluster deployment is affected by a couple of minor error/typos: - The provider filed at `usr/share/one/oneprovision/edge-clusters/metal/provisions/onprem.d/fireedge.yml` should be `onprem` instead of `oneprem`. - When any of the `Hosts to run ...` fields are empty it will result in the creation of a host named `undefined`. **Details** - Affected Component: FireEdge - OneProvision - HCI Deployment - Version: 6.4.0 **Additional context** Add any other context about the problem here. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Code committed - [ ] Testing - QA - [ ] Documentation (Release notes - resolved issues, compatibility, known issues)
priority
fireedge onprem hci errors description hci onprem cluster deployment is affected by a couple of minor error typos the provider filed at usr share one oneprovision edge clusters metal provisions onprem d fireedge yml should be onprem instead of oneprem when any of the hosts to run fields are empty it will result in the creation of a host named undefined details affected component fireedge oneprovision hci deployment version additional context add any other context about the problem here progress status code committed testing qa documentation release notes resolved issues compatibility known issues
1
682,914
23,362,210,920
IssuesEvent
2022-08-10 12:42:26
feast-dev/feast
https://api.github.com/repos/feast-dev/feast
closed
Wrong condition in to_snowflake method of snowflake provider
kind/bug priority/p2
## Expected Behavior [to_snowflake](https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/offline_stores/snowflake.py#L430) should ignore on_demand_feature_views if on_demand_feature_views is an empty list ## Current Behavior to_snowflake runs on_demand_feature_views code because on_demand_feature_views is set to [] [here](https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/offline_stores/snowflake.py#L387) ## Steps to reproduce ### Specifications - Version: 0.23.0 - Platform: feast SDK - Subsystem: ## Possible Solution `if self.on_demand_feature_views is not None: ` => `if self.on_demand_feature_views is not []: `
1.0
Wrong condition in to_snowflake method of snowflake provider - ## Expected Behavior [to_snowflake](https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/offline_stores/snowflake.py#L430) should ignore on_demand_feature_views if on_demand_feature_views is an empty list ## Current Behavior to_snowflake runs on_demand_feature_views code because on_demand_feature_views is set to [] [here](https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/offline_stores/snowflake.py#L387) ## Steps to reproduce ### Specifications - Version: 0.23.0 - Platform: feast SDK - Subsystem: ## Possible Solution `if self.on_demand_feature_views is not None: ` => `if self.on_demand_feature_views is not []: `
priority
wrong condition in to snowflake method of snowflake provider expected behavior should ignore on demand feature views if on demand feature views is an empty list current behavior to snowflake runs on demand feature views code because on demand feature views is set to steps to reproduce specifications version platform feast sdk subsystem possible solution if self on demand feature views is not none if self on demand feature views is not
1
281,102
30,872,647,784
IssuesEvent
2023-08-03 12:26:13
flatcar/Flatcar
https://api.github.com/repos/flatcar/Flatcar
opened
update: Rust
security advisory
**Name**: rust **CVEs**: [CVE-2023-38497](https://nvd.nist.gov/vuln/detail/CVE-2023-38497) **CVSSs**: n/a **Action Needed**: update to >= 1.71.1 **Summary**: (quote from https://seclists.org/oss-sec/2023/q3/90) ## Security advisory for Cargo (CVE-2023-38497) The Rust Security Response WG was notified that Cargo did not respect the umask when extracting crate archives on UNIX-like systems. If the user downloaded a crate containing files writeable by any local user, another local user could exploit this to change the source code compiled and executed by the current user. This vulnerability has been assigned CVE-2023-38497. ### Overview In UNIX-like systems, each file has three sets of permissions: for the user owning the file, for the group owning the file, and for all other local users. The "[umask][1]" is configured on most systems to limit those permissions during file creation, removing dangerous ones. For example, the default umask on macOS and most Linux distributions only allow the user owning a file to write to it, preventing the group owning it or other local users from doing the same. When a dependency is downloaded by Cargo, its source code has to be extracted on disk to allow the Rust compiler to read as part of the build. To improve performance, this extraction only happens the first time a dependency is used, caching the pre-extracted files for future invocations. Unfortunately, it was discovered that Cargo did not respect the umask during extraction, and propagated the permissions stored in the crate archive as-is. If an archive contained files writeable by any user on the system (and the system configuration didn't prevent writes through other security measures), another local user on the system could replace or tweak the source code of a dependency, potentially achieving code execution the next time the project is compiled. ### Affected Versions All Rust versions before 1.71.1 on UNIX-like systems (like macOS and Linux) are affected. Note that additional system-dependent security measures configured on the local system might prevent the vulnerability from being exploited. Users on Windows and other non-UNIX-like systems are not affected. ### Mitigations We recommend all users to update to Rust 1.71.1, which will be released later today, as it fixes the vulnerability by respecting the umask when extracting crate archives. If you build your own toolchain, patches for 1.71.0 source tarballs are [available here][2]. To prevent existing cached extractions from being exploitable, the Cargo binary included in Rust 1.71.1 or later will purge the caches it tries to access if they were generated by older Cargo versions. If you cannot update to Rust 1.71.1, we recommend configuring your system to prevent other local users from accessing the Cargo directory, usually located in `~/.cargo`: ``` chmod go= ~/.cargo ``` **refmap.gentoo**: TBD
True
update: Rust - **Name**: rust **CVEs**: [CVE-2023-38497](https://nvd.nist.gov/vuln/detail/CVE-2023-38497) **CVSSs**: n/a **Action Needed**: update to >= 1.71.1 **Summary**: (quote from https://seclists.org/oss-sec/2023/q3/90) ## Security advisory for Cargo (CVE-2023-38497) The Rust Security Response WG was notified that Cargo did not respect the umask when extracting crate archives on UNIX-like systems. If the user downloaded a crate containing files writeable by any local user, another local user could exploit this to change the source code compiled and executed by the current user. This vulnerability has been assigned CVE-2023-38497. ### Overview In UNIX-like systems, each file has three sets of permissions: for the user owning the file, for the group owning the file, and for all other local users. The "[umask][1]" is configured on most systems to limit those permissions during file creation, removing dangerous ones. For example, the default umask on macOS and most Linux distributions only allow the user owning a file to write to it, preventing the group owning it or other local users from doing the same. When a dependency is downloaded by Cargo, its source code has to be extracted on disk to allow the Rust compiler to read as part of the build. To improve performance, this extraction only happens the first time a dependency is used, caching the pre-extracted files for future invocations. Unfortunately, it was discovered that Cargo did not respect the umask during extraction, and propagated the permissions stored in the crate archive as-is. If an archive contained files writeable by any user on the system (and the system configuration didn't prevent writes through other security measures), another local user on the system could replace or tweak the source code of a dependency, potentially achieving code execution the next time the project is compiled. ### Affected Versions All Rust versions before 1.71.1 on UNIX-like systems (like macOS and Linux) are affected. Note that additional system-dependent security measures configured on the local system might prevent the vulnerability from being exploited. Users on Windows and other non-UNIX-like systems are not affected. ### Mitigations We recommend all users to update to Rust 1.71.1, which will be released later today, as it fixes the vulnerability by respecting the umask when extracting crate archives. If you build your own toolchain, patches for 1.71.0 source tarballs are [available here][2]. To prevent existing cached extractions from being exploitable, the Cargo binary included in Rust 1.71.1 or later will purge the caches it tries to access if they were generated by older Cargo versions. If you cannot update to Rust 1.71.1, we recommend configuring your system to prevent other local users from accessing the Cargo directory, usually located in `~/.cargo`: ``` chmod go= ~/.cargo ``` **refmap.gentoo**: TBD
non_priority
update rust name rust cves cvsss n a action needed update to summary quote from security advisory for cargo cve the rust security response wg was notified that cargo did not respect the umask when extracting crate archives on unix like systems if the user downloaded a crate containing files writeable by any local user another local user could exploit this to change the source code compiled and executed by the current user this vulnerability has been assigned cve overview in unix like systems each file has three sets of permissions for the user owning the file for the group owning the file and for all other local users the is configured on most systems to limit those permissions during file creation removing dangerous ones for example the default umask on macos and most linux distributions only allow the user owning a file to write to it preventing the group owning it or other local users from doing the same when a dependency is downloaded by cargo its source code has to be extracted on disk to allow the rust compiler to read as part of the build to improve performance this extraction only happens the first time a dependency is used caching the pre extracted files for future invocations unfortunately it was discovered that cargo did not respect the umask during extraction and propagated the permissions stored in the crate archive as is if an archive contained files writeable by any user on the system and the system configuration didn t prevent writes through other security measures another local user on the system could replace or tweak the source code of a dependency potentially achieving code execution the next time the project is compiled affected versions all rust versions before on unix like systems like macos and linux are affected note that additional system dependent security measures configured on the local system might prevent the vulnerability from being exploited users on windows and other non unix like systems are not affected mitigations we recommend all users to update to rust which will be released later today as it fixes the vulnerability by respecting the umask when extracting crate archives if you build your own toolchain patches for source tarballs are to prevent existing cached extractions from being exploitable the cargo binary included in rust or later will purge the caches it tries to access if they were generated by older cargo versions if you cannot update to rust we recommend configuring your system to prevent other local users from accessing the cargo directory usually located in cargo chmod go cargo refmap gentoo tbd
0
286,584
21,580,199,574
IssuesEvent
2022-05-02 17:51:46
open-feature/spec
https://api.github.com/repos/open-feature/spec
opened
Should `after` hooks be allowed to mutate the value?
documentation
Pro: you can do a json -> instance mapping in there, which would be rad. Con: subsequent hooks don't get the json value anymore, which may be very surprising to them.
1.0
Should `after` hooks be allowed to mutate the value? - Pro: you can do a json -> instance mapping in there, which would be rad. Con: subsequent hooks don't get the json value anymore, which may be very surprising to them.
non_priority
should after hooks be allowed to mutate the value pro you can do a json instance mapping in there which would be rad con subsequent hooks don t get the json value anymore which may be very surprising to them
0
47,029
11,947,116,522
IssuesEvent
2020-04-03 09:21:30
Financial-Times/dotcom-page-kit
https://api.github.com/repos/Financial-Times/dotcom-page-kit
closed
Page Kit as a Webpack configuration builder
Build tools CLI tool Proposal Refactor
It has become apparent that the current Page Kit CLI is too complex for our needs and is leading to some confusion about the scope of the project. Because Page Kit provides its own CLI tool it is also worrying users that we are replacing one esoteric system with another. At its core the current Page Kit CLI enables a complex Webpack configuration to be composed from multiple separate packages. It also makes accessing and modifying specific parts of the configuration simpler and more robust. However, as discussed in #133 I would like to make it possible to use Webpack's own CLI like this: ```js // webpack.config.js module.exports = pageKit(options); ``` Taking this one step further... this would mitigate the need to maintain, document, and support an additional tool. I would also hope that by focusing on the configuration aspect it removes any worries about straying too far from "standard" tools. This issue supersedes #58 and #123 and #133.
1.0
Page Kit as a Webpack configuration builder - It has become apparent that the current Page Kit CLI is too complex for our needs and is leading to some confusion about the scope of the project. Because Page Kit provides its own CLI tool it is also worrying users that we are replacing one esoteric system with another. At its core the current Page Kit CLI enables a complex Webpack configuration to be composed from multiple separate packages. It also makes accessing and modifying specific parts of the configuration simpler and more robust. However, as discussed in #133 I would like to make it possible to use Webpack's own CLI like this: ```js // webpack.config.js module.exports = pageKit(options); ``` Taking this one step further... this would mitigate the need to maintain, document, and support an additional tool. I would also hope that by focusing on the configuration aspect it removes any worries about straying too far from "standard" tools. This issue supersedes #58 and #123 and #133.
non_priority
page kit as a webpack configuration builder it has become apparent that the current page kit cli is too complex for our needs and is leading to some confusion about the scope of the project because page kit provides its own cli tool it is also worrying users that we are replacing one esoteric system with another at its core the current page kit cli enables a complex webpack configuration to be composed from multiple separate packages it also makes accessing and modifying specific parts of the configuration simpler and more robust however as discussed in i would like to make it possible to use webpack s own cli like this js webpack config js module exports pagekit options taking this one step further this would mitigate the need to maintain document and support an additional tool i would also hope that by focusing on the configuration aspect it removes any worries about straying too far from standard tools this issue supersedes and and
0
751,267
26,237,158,661
IssuesEvent
2023-01-05 08:25:08
zino-hofmann/graphql-flutter
https://api.github.com/repos/zino-hofmann/graphql-flutter
closed
Subscription- Bad State : stream has already been listened to.
needs more info :hourglass: reproduction needed Priority: Waiting to be assigned
Hey, I recently got this problem again. But this time, 1st time the socket connects and works perfectly fine. But on navigating back and forth to the same widget throws this error. Are we supposed dispose something even when using hooks generated by codegen ? Am I missing something ? _Originally posted by @Arpit1496 in https://github.com/zino-hofmann/graphql-flutter/issues/1191#issuecomment-1230804072_
1.0
Subscription- Bad State : stream has already been listened to. - Hey, I recently got this problem again. But this time, 1st time the socket connects and works perfectly fine. But on navigating back and forth to the same widget throws this error. Are we supposed dispose something even when using hooks generated by codegen ? Am I missing something ? _Originally posted by @Arpit1496 in https://github.com/zino-hofmann/graphql-flutter/issues/1191#issuecomment-1230804072_
priority
subscription bad state stream has already been listened to hey i recently got this problem again but this time time the socket connects and works perfectly fine but on navigating back and forth to the same widget throws this error are we supposed dispose something even when using hooks generated by codegen am i missing something originally posted by in
1
705,008
24,218,250,783
IssuesEvent
2022-09-26 08:42:10
wp-media/wp-rocket
https://api.github.com/repos/wp-media/wp-rocket
reopened
RUCSS counter msg isnot displayed without refresh in certain case
type: bug priority: medium effort: [XS] severity: moderate module: remove unused css
**Before submitting an issue please check that youโ€™ve completed the following steps:** - Made sure youโ€™re on the latest version =>. 3.11.4 - Used the search feature to ensure that the bug hasnโ€™t been reported before **Describe the bug** Counter msg for RUCSS is not displayed till refreshing the dashboard after activating RUCSS **To Reproduce** Steps to reproduce the behavior: 1. Fresh install to wpr 2. Activate RUCSS and save 3. Counter msg not displayed till refreshing dashboard **Expected behavior** Counter msg displayed after saving settings with no need to manually refresh the page **Screenshots** If applicable, add screenshots to help explain your problem. https://jmp.sh/zqxRj0D **Additional context** Add any other context about the problem here. - Same with PHP 7.4.3 and 8.1.7 - The same scenario was working fine on 3.11.3 - Another case when permissions are read-only to cache folder then enable RUCSS, once permissions are back, we need to refresh twice so we can see the RUCSS msg => in this case, if we fixed permissions after RUCSS was enabled by > 90sec, the success msg will be displayed not the counter although nothing is completed yet in used CSS table - Note: clear used CSS, reactivate RUCSS, change safelist => all displaying msg with no need to refresh **Backlog Grooming (for WP Media dev team use only)** - [ ] Reproduce the problem - [ ] Identify the root cause - [ ] Scope a solution - [ ] Estimate the effort
1.0
RUCSS counter msg isnot displayed without refresh in certain case - **Before submitting an issue please check that youโ€™ve completed the following steps:** - Made sure youโ€™re on the latest version =>. 3.11.4 - Used the search feature to ensure that the bug hasnโ€™t been reported before **Describe the bug** Counter msg for RUCSS is not displayed till refreshing the dashboard after activating RUCSS **To Reproduce** Steps to reproduce the behavior: 1. Fresh install to wpr 2. Activate RUCSS and save 3. Counter msg not displayed till refreshing dashboard **Expected behavior** Counter msg displayed after saving settings with no need to manually refresh the page **Screenshots** If applicable, add screenshots to help explain your problem. https://jmp.sh/zqxRj0D **Additional context** Add any other context about the problem here. - Same with PHP 7.4.3 and 8.1.7 - The same scenario was working fine on 3.11.3 - Another case when permissions are read-only to cache folder then enable RUCSS, once permissions are back, we need to refresh twice so we can see the RUCSS msg => in this case, if we fixed permissions after RUCSS was enabled by > 90sec, the success msg will be displayed not the counter although nothing is completed yet in used CSS table - Note: clear used CSS, reactivate RUCSS, change safelist => all displaying msg with no need to refresh **Backlog Grooming (for WP Media dev team use only)** - [ ] Reproduce the problem - [ ] Identify the root cause - [ ] Scope a solution - [ ] Estimate the effort
priority
rucss counter msg isnot displayed without refresh in certain case before submitting an issue please check that youโ€™ve completed the following steps made sure youโ€™re on the latest version used the search feature to ensure that the bug hasnโ€™t been reported before describe the bug counter msg for rucss is not displayed till refreshing the dashboard after activating rucss to reproduce steps to reproduce the behavior fresh install to wpr activate rucss and save counter msg not displayed till refreshing dashboard expected behavior counter msg displayed after saving settings with no need to manually refresh the page screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here same with php and the same scenario was working fine on another case when permissions are read only to cache folder then enable rucss once permissions are back we need to refresh twice so we can see the rucss msg in this case if we fixed permissions after rucss was enabled by the success msg will be displayed not the counter although nothing is completed yet in used css table note clear used css reactivate rucss change safelist all displaying msg with no need to refresh backlog grooming for wp media dev team use only reproduce the problem identify the root cause scope a solution estimate the effort
1
325,818
27,964,356,421
IssuesEvent
2023-03-24 18:07:12
NeurodataWithoutBorders/pynwb
https://api.github.com/repos/NeurodataWithoutBorders/pynwb
closed
test tutorials across python versions
priority: medium topic: docs topic: testing
We are currently limiting the test matrix of the tutorials due to the AllenSDK. https://github.com/NeurodataWithoutBorders/pynwb/blob/cc5fd34240891e14a6b82552a4b8c3c259fcafae/.github/workflows/run_all_tests.yml#L92-L107 The latest AllenSKD claims to support Python 3.9: https://github.com/AllenInstitute/AllenSDK/blob/1caa779b517eeb2428282919d0fb8a65e7217791/setup.py#L92 And anyway I don't think at this point we should be limiting our entire testing suite based on the Allen SDK.
1.0
test tutorials across python versions - We are currently limiting the test matrix of the tutorials due to the AllenSDK. https://github.com/NeurodataWithoutBorders/pynwb/blob/cc5fd34240891e14a6b82552a4b8c3c259fcafae/.github/workflows/run_all_tests.yml#L92-L107 The latest AllenSKD claims to support Python 3.9: https://github.com/AllenInstitute/AllenSDK/blob/1caa779b517eeb2428282919d0fb8a65e7217791/setup.py#L92 And anyway I don't think at this point we should be limiting our entire testing suite based on the Allen SDK.
non_priority
test tutorials across python versions we are currently limiting the test matrix of the tutorials due to the allensdk the latest allenskd claims to support python and anyway i don t think at this point we should be limiting our entire testing suite based on the allen sdk
0
44,696
2,910,632,639
IssuesEvent
2015-06-21 23:02:38
SteamDatabase/steamSummerMinigame
https://api.github.com/repos/SteamDatabase/steamSummerMinigame
closed
Ignore raining gold on trash mobs late-game
2 - Medium Priority Enhancement
Probably anything past level 1500 (or 10 railgun + 10 mouse button + 20 elemental damage levels) isn't worth switching to the raining gold lane unless Max Elemental Damage is also active. Prioritizing quick group levels by focusing on users' elemental spec will yield far more gold on bosses than anything one could gather from trash lane gold clicks.
1.0
Ignore raining gold on trash mobs late-game - Probably anything past level 1500 (or 10 railgun + 10 mouse button + 20 elemental damage levels) isn't worth switching to the raining gold lane unless Max Elemental Damage is also active. Prioritizing quick group levels by focusing on users' elemental spec will yield far more gold on bosses than anything one could gather from trash lane gold clicks.
priority
ignore raining gold on trash mobs late game probably anything past level or railgun mouse button elemental damage levels isn t worth switching to the raining gold lane unless max elemental damage is also active prioritizing quick group levels by focusing on users elemental spec will yield far more gold on bosses than anything one could gather from trash lane gold clicks
1
141,616
11,427,956,148
IssuesEvent
2020-02-04 02:41:03
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
The Clone button shows blank on Clone dialog
:beetle: regression :gear: blobs :gear: files ๐Ÿงช testing
**Storage Explorer Version:** 1.12.0 **Build:** [20200201.2](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3434212&view=results) **Branch:** master **Platform/OS:** Windows 10/ Linux Ubuntu 18.04 **Architecture:** ia32/x64 **Regression From:** Previous release(1.12.0) **Steps to reproduce:** 1. Expand one Non-ADLS Gen2 storage account -> Blob Containers. 2. Select one blob container and right click it -> Click 'Clone with New Name...'. 3. Check the Clone button on the Clone dialog. **Expect Experience:** The Clone button shows well. **Actual Experience:** The Clone button shows blank. ![image](https://user-images.githubusercontent.com/41351993/73641600-36be1400-46ab-11ea-9c53-16725a7da9d9.png) **More Info:** 1. This issue also reproduces for files under one file share & one file share. 2. This issue doesn't reproduce for one blob.
1.0
The Clone button shows blank on Clone dialog - **Storage Explorer Version:** 1.12.0 **Build:** [20200201.2](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3434212&view=results) **Branch:** master **Platform/OS:** Windows 10/ Linux Ubuntu 18.04 **Architecture:** ia32/x64 **Regression From:** Previous release(1.12.0) **Steps to reproduce:** 1. Expand one Non-ADLS Gen2 storage account -> Blob Containers. 2. Select one blob container and right click it -> Click 'Clone with New Name...'. 3. Check the Clone button on the Clone dialog. **Expect Experience:** The Clone button shows well. **Actual Experience:** The Clone button shows blank. ![image](https://user-images.githubusercontent.com/41351993/73641600-36be1400-46ab-11ea-9c53-16725a7da9d9.png) **More Info:** 1. This issue also reproduces for files under one file share & one file share. 2. This issue doesn't reproduce for one blob.
non_priority
the clone button shows blank on clone dialog storage explorer version build branch master platform os windows linux ubuntu architecture regression from previous release steps to reproduce expand one non adls storage account blob containers select one blob container and right click it click clone with new name check the clone button on the clone dialog expect experience the clone button shows well actual experience the clone button shows blank more info this issue also reproduces for files under one file share one file share this issue doesn t reproduce for one blob
0
83,700
16,343,582,723
IssuesEvent
2021-05-13 03:08:09
microsoft/AdaptiveCards
https://api.github.com/repos/microsoft/AdaptiveCards
closed
[iOS] Nightly package feed
Status-In Code Review Task
Laundry list. 1. remove .swiftpm from .gitignore 2. generate tags for nightly 3. limit what's included or shown in the package 4. add umbrella header for module map generation
1.0
[iOS] Nightly package feed - Laundry list. 1. remove .swiftpm from .gitignore 2. generate tags for nightly 3. limit what's included or shown in the package 4. add umbrella header for module map generation
non_priority
nightly package feed laundry list remove swiftpm from gitignore generate tags for nightly limit what s included or shown in the package add umbrella header for module map generation
0
336,143
24,488,282,094
IssuesEvent
2022-10-09 18:33:38
dgtlmoon/changedetection.io
https://api.github.com/repos/dgtlmoon/changedetection.io
closed
Images not showing up in wiki
documentation
I am new to changedetection and was going through the wiki but many pics are missing if you can please fix their links: - https://github.com/dgtlmoon/changedetection.io/wiki - https://github.com/dgtlmoon/changedetection.io/wiki/Sharing-a-Watch ![image](https://user-images.githubusercontent.com/794673/190856272-c8609dc2-327e-4b5c-a47d-4e60d587fef6.png)
1.0
Images not showing up in wiki - I am new to changedetection and was going through the wiki but many pics are missing if you can please fix their links: - https://github.com/dgtlmoon/changedetection.io/wiki - https://github.com/dgtlmoon/changedetection.io/wiki/Sharing-a-Watch ![image](https://user-images.githubusercontent.com/794673/190856272-c8609dc2-327e-4b5c-a47d-4e60d587fef6.png)
non_priority
images not showing up in wiki i am new to changedetection and was going through the wiki but many pics are missing if you can please fix their links
0
157,048
12,344,285,578
IssuesEvent
2020-05-15 06:37:26
celery/celery
https://api.github.com/repos/celery/celery
closed
Weird death / worker not restarting
Component: Prefork Workers Pool Status: Needs Testcase โœ˜ Status: Needs Verification โœ˜
Hi, Two different weird issues started happening. The first one is: I have workers dying with this stacktrace ``` [2019-08-22 21:03:58,650: ERROR/MainProcess] Process 'ForkPoolWorker-89' pid:101 exited with 'signal 9 (SIGKILL)' ERROR 2019-08-22 21:03:58,666 [celery.worker.request] <140652482161544> Task handler raised error: WorkerLostError('Worker exited prematurely: signal 9 (SIGKILL).') Traceback (most recent call last): File "/venv/lib/python3.7/site-packages/celery/worker/worker.py", line 205, in start self.blueprint.start(self) File "/venv/lib/python3.7/site-packages/celery/bootsteps.py", line 119, in start step.start(parent) File "/venv/lib/python3.7/site-packages/celery/bootsteps.py", line 369, in start return self.obj.start() File "/venv/lib/python3.7/site-packages/celery/worker/consumer/consumer.py", line 318, in start blueprint.start(self) File "/venv/lib/python3.7/site-packages/celery/bootsteps.py", line 119, in start step.start(parent) File "/venv/lib/python3.7/site-packages/celery/worker/consumer/consumer.py", line 596, in start c.loop(*c.loop_args()) File "/venv/lib/python3.7/site-packages/celery/worker/loops.py", line 91, in asynloop next(loop) File "/venv/lib/python3.7/site-packages/kombu/asynchronous/hub.py", line 362, in create_loop cb(*cbargs) File "/venv/lib/python3.7/site-packages/celery/concurrency/asynpool.py", line 298, in on_result_readable next(it) File "/venv/lib/python3.7/site-packages/celery/concurrency/asynpool.py", line 279, in _recv_message message = load(bufv) UnicodeDecodeError: 'utf-8' codec can't decode bytes in position 46-47: invalid continuation byte ``` Second weird thing: These workers never get restarted. So one by one my workers die, until it is just the "master" left on the node. A PS gives me something like ``` PID USER TIME COMMAND 1 root 3:53 {celery} /venv/bin/python3.7 /venv/bin/celery -A app_name worker -l INFO -n worker-%n-standard -c 20 -Q slow --without-heartbeat --without-mingle --without-gossip 86 root 0:40 [celery] 87 root 0:28 [celery] 88 root 0:37 [celery] 89 root 0:34 [celery] 90 root 0:38 [celery] 91 root 0:37 [celery] 92 root 0:26 [celery] 93 root 0:12 [celery] 94 root 0:09 [celery] 98 root 0:06 [celery] 99 root 0:05 [celery] 100 root 0:06 [celery] 102 root 0:04 [celery] 103 root 0:02 [celery] 104 root 0:02 [celery] 105 root 0:01 [celery] 106 root 0:01 [celery] 107 root 0:01 [celery] 108 root 0:00 [celery] 109 root 0:00 sh ``` As you can see, the 0 process is the master. The rest are just zombies doing nothing. A weird byte seems like it should not kill a worker, and killing a worker should not make a zombie. Unsure what to tackle first? Setup: celery==4.3.0 billiard==3.6.1.0 kombu==4.5.0 RabbitMQ Prefork Running in docker on `python:3.7.3-alpine3.8`
1.0
Weird death / worker not restarting - Hi, Two different weird issues started happening. The first one is: I have workers dying with this stacktrace ``` [2019-08-22 21:03:58,650: ERROR/MainProcess] Process 'ForkPoolWorker-89' pid:101 exited with 'signal 9 (SIGKILL)' ERROR 2019-08-22 21:03:58,666 [celery.worker.request] <140652482161544> Task handler raised error: WorkerLostError('Worker exited prematurely: signal 9 (SIGKILL).') Traceback (most recent call last): File "/venv/lib/python3.7/site-packages/celery/worker/worker.py", line 205, in start self.blueprint.start(self) File "/venv/lib/python3.7/site-packages/celery/bootsteps.py", line 119, in start step.start(parent) File "/venv/lib/python3.7/site-packages/celery/bootsteps.py", line 369, in start return self.obj.start() File "/venv/lib/python3.7/site-packages/celery/worker/consumer/consumer.py", line 318, in start blueprint.start(self) File "/venv/lib/python3.7/site-packages/celery/bootsteps.py", line 119, in start step.start(parent) File "/venv/lib/python3.7/site-packages/celery/worker/consumer/consumer.py", line 596, in start c.loop(*c.loop_args()) File "/venv/lib/python3.7/site-packages/celery/worker/loops.py", line 91, in asynloop next(loop) File "/venv/lib/python3.7/site-packages/kombu/asynchronous/hub.py", line 362, in create_loop cb(*cbargs) File "/venv/lib/python3.7/site-packages/celery/concurrency/asynpool.py", line 298, in on_result_readable next(it) File "/venv/lib/python3.7/site-packages/celery/concurrency/asynpool.py", line 279, in _recv_message message = load(bufv) UnicodeDecodeError: 'utf-8' codec can't decode bytes in position 46-47: invalid continuation byte ``` Second weird thing: These workers never get restarted. So one by one my workers die, until it is just the "master" left on the node. A PS gives me something like ``` PID USER TIME COMMAND 1 root 3:53 {celery} /venv/bin/python3.7 /venv/bin/celery -A app_name worker -l INFO -n worker-%n-standard -c 20 -Q slow --without-heartbeat --without-mingle --without-gossip 86 root 0:40 [celery] 87 root 0:28 [celery] 88 root 0:37 [celery] 89 root 0:34 [celery] 90 root 0:38 [celery] 91 root 0:37 [celery] 92 root 0:26 [celery] 93 root 0:12 [celery] 94 root 0:09 [celery] 98 root 0:06 [celery] 99 root 0:05 [celery] 100 root 0:06 [celery] 102 root 0:04 [celery] 103 root 0:02 [celery] 104 root 0:02 [celery] 105 root 0:01 [celery] 106 root 0:01 [celery] 107 root 0:01 [celery] 108 root 0:00 [celery] 109 root 0:00 sh ``` As you can see, the 0 process is the master. The rest are just zombies doing nothing. A weird byte seems like it should not kill a worker, and killing a worker should not make a zombie. Unsure what to tackle first? Setup: celery==4.3.0 billiard==3.6.1.0 kombu==4.5.0 RabbitMQ Prefork Running in docker on `python:3.7.3-alpine3.8`
non_priority
weird death worker not restarting hi two different weird issues started happening the first one is i have workers dying with this stacktrace process forkpoolworker pid exited with signal sigkill error task handler raised error workerlosterror worker exited prematurely signal sigkill traceback most recent call last file venv lib site packages celery worker worker py line in start self blueprint start self file venv lib site packages celery bootsteps py line in start step start parent file venv lib site packages celery bootsteps py line in start return self obj start file venv lib site packages celery worker consumer consumer py line in start blueprint start self file venv lib site packages celery bootsteps py line in start step start parent file venv lib site packages celery worker consumer consumer py line in start c loop c loop args file venv lib site packages celery worker loops py line in asynloop next loop file venv lib site packages kombu asynchronous hub py line in create loop cb cbargs file venv lib site packages celery concurrency asynpool py line in on result readable next it file venv lib site packages celery concurrency asynpool py line in recv message message load bufv unicodedecodeerror utf codec can t decode bytes in position invalid continuation byte second weird thing these workers never get restarted so one by one my workers die until it is just the master left on the node a ps gives me something like pid user time command root celery venv bin venv bin celery a app name worker l info n worker n standard c q slow without heartbeat without mingle without gossip root root root root root root root root root root root root root root root root root root root root sh as you can see the process is the master the rest are just zombies doing nothing a weird byte seems like it should not kill a worker and killing a worker should not make a zombie unsure what to tackle first setup celery billiard kombu rabbitmq prefork running in docker on python
0
222,814
7,439,064,736
IssuesEvent
2018-03-27 04:12:15
minishift/minishift
https://api.github.com/repos/minishift/minishift
closed
Improvement: allow to assign multiple nameservers for use with network settings assignment
kind/feature priority/major
With #1316 it is currently possible to specify one nameserver on start to use, while the settings and assignment can support multiples of this. To keep the implementation simple only one is used at the moment. This task would like to improve on this by allowing multiple to be used (or at least 2).
1.0
Improvement: allow to assign multiple nameservers for use with network settings assignment - With #1316 it is currently possible to specify one nameserver on start to use, while the settings and assignment can support multiples of this. To keep the implementation simple only one is used at the moment. This task would like to improve on this by allowing multiple to be used (or at least 2).
priority
improvement allow to assign multiple nameservers for use with network settings assignment with it is currently possible to specify one nameserver on start to use while the settings and assignment can support multiples of this to keep the implementation simple only one is used at the moment this task would like to improve on this by allowing multiple to be used or at least
1
15,043
2,841,940,144
IssuesEvent
2015-05-28 05:26:10
ibus/ibus
https://api.github.com/repos/ibus/ibus
closed
Modify ibus-table to allow table-based IMs to define a symbol
Component-ibus-table Priority-Medium Type-Defect
``` Input methods based on ibus-table should be able to define a symbol for IM switchers. This could be done by adding a SYMBOL = line to the .txt table definitions and using the value to set the symbol in ibus-table's main.py. ``` Original issue reported on code.google.com by `leftmostcat` on 2013-09-05 02:42:30
1.0
Modify ibus-table to allow table-based IMs to define a symbol - ``` Input methods based on ibus-table should be able to define a symbol for IM switchers. This could be done by adding a SYMBOL = line to the .txt table definitions and using the value to set the symbol in ibus-table's main.py. ``` Original issue reported on code.google.com by `leftmostcat` on 2013-09-05 02:42:30
non_priority
modify ibus table to allow table based ims to define a symbol input methods based on ibus table should be able to define a symbol for im switchers this could be done by adding a symbol line to the txt table definitions and using the value to set the symbol in ibus table s main py original issue reported on code google com by leftmostcat on
0
779,792
27,366,330,882
IssuesEvent
2023-02-27 19:28:33
bigbluebutton/bigbluebutton
https://api.github.com/repos/bigbluebutton/bigbluebutton
closed
Wb: Move the panning tool to be near the pointer tool
priority: low component: whiteboard
<!--PLEASE DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS. This issue tracker is only for bbb development related issues. Search for existing feature requests to avoid creating duplicates.--> **Is your feature request related to a problem? Please describe.** The panning tool was moved to the slide control toolbar. This new location is not intuitive. In BBB 2.5 and earlier the two tools were grouped together ![Screenshot from 2023-02-07 09-26-50](https://user-images.githubusercontent.com/6312397/217277653-ab288dd4-066d-4d62-9426-a85e7e4153b4.png) **Describe the solution you'd like** ![Screenshot from 2023-02-01 12-39-38](https://user-images.githubusercontent.com/6312397/217277695-ae3ec1c5-5e40-4e45-a0d3-f1763330507e.png) **Describe alternatives you've considered** keep things as they are
1.0
Wb: Move the panning tool to be near the pointer tool - <!--PLEASE DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS. This issue tracker is only for bbb development related issues. Search for existing feature requests to avoid creating duplicates.--> **Is your feature request related to a problem? Please describe.** The panning tool was moved to the slide control toolbar. This new location is not intuitive. In BBB 2.5 and earlier the two tools were grouped together ![Screenshot from 2023-02-07 09-26-50](https://user-images.githubusercontent.com/6312397/217277653-ab288dd4-066d-4d62-9426-a85e7e4153b4.png) **Describe the solution you'd like** ![Screenshot from 2023-02-01 12-39-38](https://user-images.githubusercontent.com/6312397/217277695-ae3ec1c5-5e40-4e45-a0d3-f1763330507e.png) **Describe alternatives you've considered** keep things as they are
priority
wb move the panning tool to be near the pointer tool please do not file issues for general support questions this issue tracker is only for bbb development related issues search for existing feature requests to avoid creating duplicates is your feature request related to a problem please describe the panning tool was moved to the slide control toolbar this new location is not intuitive in bbb and earlier the two tools were grouped together describe the solution you d like describe alternatives you ve considered keep things as they are
1
16,867
5,295,025,653
IssuesEvent
2017-02-09 12:41:25
TypeCobolTeam/TypeCobol
https://api.github.com/repos/TypeCobolTeam/TypeCobol
closed
Incorrect Code generation in presence of Preprocessor directives like COPY or REPLACE
bug Cobol Codegen
In presence of Preprocessor directives Like COPY and REPLACE the generator fails to generate original COPY or REPLACE instructions. We must implement another generation model based on generated code injection in the original source code. The document TypeCobol\Codegen\src\Documentation\CodegenStudy.xlsx Shows some situations for which the generation is not correctly handled.
1.0
Incorrect Code generation in presence of Preprocessor directives like COPY or REPLACE - In presence of Preprocessor directives Like COPY and REPLACE the generator fails to generate original COPY or REPLACE instructions. We must implement another generation model based on generated code injection in the original source code. The document TypeCobol\Codegen\src\Documentation\CodegenStudy.xlsx Shows some situations for which the generation is not correctly handled.
non_priority
incorrect code generation in presence of preprocessor directives like copy or replace in presence of preprocessor directives like copy and replace the generator fails to generate original copy or replace instructions we must implement another generation model based on generated code injection in the original source code the document typecobol codegen src documentation codegenstudy xlsx shows some situations for which the generation is not correctly handled
0
80,140
15,586,268,084
IssuesEvent
2021-03-18 01:33:12
peterwkc85/Spock_Testing_Framework
https://api.github.com/repos/peterwkc85/Spock_Testing_Framework
opened
CVE-2020-36184 (High) detected in jackson-databind-2.5.4.jar
security vulnerability
## CVE-2020-36184 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.5.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Spock_Testing_Framework/java-testing-with-spock-master/java-testing-with-spock-master/chapter7/rest-service-example/pom.xml</p> <p>Path to vulnerable library: 2/repository/com/fasterxml/jackson/core/jackson-databind/2.5.4/jackson-databind-2.5.4.jar,radle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.5.4/5dfa42af84584b4a862ea488da84bbbebbb06c35/jackson-databind-2.5.4.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.5.4.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184>CVE-2020-36184</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2998">https://github.com/FasterXML/jackson-databind/issues/2998</a></p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-36184 (High) detected in jackson-databind-2.5.4.jar - ## CVE-2020-36184 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.5.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Spock_Testing_Framework/java-testing-with-spock-master/java-testing-with-spock-master/chapter7/rest-service-example/pom.xml</p> <p>Path to vulnerable library: 2/repository/com/fasterxml/jackson/core/jackson-databind/2.5.4/jackson-databind-2.5.4.jar,radle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.5.4/5dfa42af84584b4a862ea488da84bbbebbb06c35/jackson-databind-2.5.4.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.5.4.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184>CVE-2020-36184</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2998">https://github.com/FasterXML/jackson-databind/issues/2998</a></p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file spock testing framework java testing with spock master java testing with spock master rest service example pom xml path to vulnerable library repository com fasterxml jackson core jackson databind jackson databind jar radle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp datasources peruserpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
602,292
18,460,287,724
IssuesEvent
2021-10-15 23:37:09
michaelrsweet/pdfio
https://api.github.com/repos/michaelrsweet/pdfio
closed
make all-shared Makefile errors
bug priority-medium
when running `make all-shared`, make fails with several errors - lines 72,112: `else` is missing trailing backslash - line 128: `soname` should be `-soname` - making a shared object requires `-fPIC` Thanks for the project, it has been extremely useful.
1.0
make all-shared Makefile errors - when running `make all-shared`, make fails with several errors - lines 72,112: `else` is missing trailing backslash - line 128: `soname` should be `-soname` - making a shared object requires `-fPIC` Thanks for the project, it has been extremely useful.
priority
make all shared makefile errors when running make all shared make fails with several errors lines else is missing trailing backslash line soname should be soname making a shared object requires fpic thanks for the project it has been extremely useful
1
36,549
2,800,966,668
IssuesEvent
2015-05-13 13:29:52
lars-t-hansen/parlib-simple
https://api.github.com/repos/lars-t-hansen/parlib-simple
closed
Occasional non-responsive barrier test
bug High priority
Mac release build with queue as of 16 Feb 2015, current m-i. Fast MBP 4x2. e10s enabled. Load test-barrier.html. Reload it very quickly many times, eventually I get to a state where the workers report "ready" but then nothing happens. Looking at the threads in the content process in lldb, the main thread as well as two of the DOM workers are in futexWait but all the others are in their primary event loop. Setting the timeout in the main program to 100ms rather than 0 gets rid of the problem. However this does not provide a good explanation for the observed behavior.
1.0
Occasional non-responsive barrier test - Mac release build with queue as of 16 Feb 2015, current m-i. Fast MBP 4x2. e10s enabled. Load test-barrier.html. Reload it very quickly many times, eventually I get to a state where the workers report "ready" but then nothing happens. Looking at the threads in the content process in lldb, the main thread as well as two of the DOM workers are in futexWait but all the others are in their primary event loop. Setting the timeout in the main program to 100ms rather than 0 gets rid of the problem. However this does not provide a good explanation for the observed behavior.
priority
occasional non responsive barrier test mac release build with queue as of feb current m i fast mbp enabled load test barrier html reload it very quickly many times eventually i get to a state where the workers report ready but then nothing happens looking at the threads in the content process in lldb the main thread as well as two of the dom workers are in futexwait but all the others are in their primary event loop setting the timeout in the main program to rather than gets rid of the problem however this does not provide a good explanation for the observed behavior
1
168,003
13,055,224,172
IssuesEvent
2020-07-30 01:00:59
n4af/TR4W
https://api.github.com/repos/n4af/TR4W
closed
Use new Icom commands to set inactive VFO
In testing Radio Control enhancement
For Split mode, newer Icom radios now support a command to set the inactive VFO to a frequency and mode (25h) rather than having to switch to VFO B, set the frequency and mode and switch back. This should be implemented as it simplifies the switching and puts it more inline with the way the other radios are done.
1.0
Use new Icom commands to set inactive VFO - For Split mode, newer Icom radios now support a command to set the inactive VFO to a frequency and mode (25h) rather than having to switch to VFO B, set the frequency and mode and switch back. This should be implemented as it simplifies the switching and puts it more inline with the way the other radios are done.
non_priority
use new icom commands to set inactive vfo for split mode newer icom radios now support a command to set the inactive vfo to a frequency and mode rather than having to switch to vfo b set the frequency and mode and switch back this should be implemented as it simplifies the switching and puts it more inline with the way the other radios are done
0
116,471
17,370,033,368
IssuesEvent
2021-07-30 12:49:44
lukebroganws/Java-Demo
https://api.github.com/repos/lukebroganws/Java-Demo
opened
CVE-2016-10735 (Medium) detected in bootstrap-3.3.7.min.js
security vulnerability
## CVE-2016-10735 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: Java-Demo/src/main/webapp/dfi/style_bootstrap.html</p> <p>Path to vulnerable library: Java-Demo/src/main/webapp/dfi/style_bootstrap.html,Java-Demo/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7-3.3.13.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/Java-Demo/commit/d73a27e2fea07f94b9c092744aef285ec88e27c4">d73a27e2fea07f94b9c092744aef285ec88e27c4</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/issues/20184">https://github.com/twbs/bootstrap/issues/20184</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: 3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"todc-bootstrap","packageVersion":"3.3.7-3.3.13","packageFilePaths":["/src/main/webapp/dfi/style_bootstrap.html","/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html"],"isTransitiveDependency":false,"dependencyTree":"todc-bootstrap:3.3.7-3.3.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2016-10735","vulnerabilityDetails":"In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-10735 (Medium) detected in bootstrap-3.3.7.min.js - ## CVE-2016-10735 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: Java-Demo/src/main/webapp/dfi/style_bootstrap.html</p> <p>Path to vulnerable library: Java-Demo/src/main/webapp/dfi/style_bootstrap.html,Java-Demo/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7-3.3.13.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/Java-Demo/commit/d73a27e2fea07f94b9c092744aef285ec88e27c4">d73a27e2fea07f94b9c092744aef285ec88e27c4</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/issues/20184">https://github.com/twbs/bootstrap/issues/20184</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: 3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"todc-bootstrap","packageVersion":"3.3.7-3.3.13","packageFilePaths":["/src/main/webapp/dfi/style_bootstrap.html","/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html"],"isTransitiveDependency":false,"dependencyTree":"todc-bootstrap:3.3.7-3.3.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2016-10735","vulnerabilityDetails":"In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_priority
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file java demo src main webapp dfi style bootstrap html path to vulnerable library java demo src main webapp dfi style bootstrap html java demo target easybuggy snapshot dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch main vulnerability details in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree todc bootstrap isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve vulnerabilityurl
0
317,117
23,664,740,755
IssuesEvent
2022-08-26 19:28:50
python/mypy
https://api.github.com/repos/python/mypy
closed
globl "type: ignore" ignores too much
bug documentation topic-type-ignore
Structure: `sample_proj/__init__.py`: ```python from sample_proj.commons import some_const from sample_proj.some_file import SomeClass ``` `sample_proj/comons.py`: ```python some_const = 'qwe' ``` `sample_proj/some_file.py`: ```python # type: ignore from sample_proj.commons import some_const class SomeClass: def fun(self, param=some_const): pass ``` ``` $ mypy sample_proj sample_proj/__init__.py:2: error: Module "sample_proj.some_file" has no attribute "SomeClass" Found 1 error in 1 file (checked 3 source files) ``` I think it's a bug. Removing `# type: ignore[union-attr]` helps. I wanted this comment to ignore only this exact error in the whole file. ``` $ mypy --version mypy 0.942 $ python3 --version Python 3.10.5 ```
1.0
globl "type: ignore" ignores too much - Structure: `sample_proj/__init__.py`: ```python from sample_proj.commons import some_const from sample_proj.some_file import SomeClass ``` `sample_proj/comons.py`: ```python some_const = 'qwe' ``` `sample_proj/some_file.py`: ```python # type: ignore from sample_proj.commons import some_const class SomeClass: def fun(self, param=some_const): pass ``` ``` $ mypy sample_proj sample_proj/__init__.py:2: error: Module "sample_proj.some_file" has no attribute "SomeClass" Found 1 error in 1 file (checked 3 source files) ``` I think it's a bug. Removing `# type: ignore[union-attr]` helps. I wanted this comment to ignore only this exact error in the whole file. ``` $ mypy --version mypy 0.942 $ python3 --version Python 3.10.5 ```
non_priority
globl type ignore ignores too much structure sample proj init py python from sample proj commons import some const from sample proj some file import someclass sample proj comons py python some const qwe sample proj some file py python type ignore from sample proj commons import some const class someclass def fun self param some const pass mypy sample proj sample proj init py error module sample proj some file has no attribute someclass found error in file checked source files i think it s a bug removing type ignore helps i wanted this comment to ignore only this exact error in the whole file mypy version mypy version python
0
360,057
10,683,826,931
IssuesEvent
2019-10-22 09:11:17
celo-org/celo-monorepo
https://api.github.com/repos/celo-org/celo-monorepo
closed
Attestation Users SBAT use the new precompile
Priority: P2 enhancement identity protocol triaged
### Expected Behavior Use the new precompile ### Current Behavior Currently use `Validators#getValidators` which is very expensive
1.0
Attestation Users SBAT use the new precompile - ### Expected Behavior Use the new precompile ### Current Behavior Currently use `Validators#getValidators` which is very expensive
priority
attestation users sbat use the new precompile expected behavior use the new precompile current behavior currently use validators getvalidators which is very expensive
1
187,596
6,759,497,247
IssuesEvent
2017-10-24 17:18:46
rotorgames/Rg.Plugins.Popup
https://api.github.com/repos/rotorgames/Rg.Plugins.Popup
reopened
CloseWhenBackgroundIsClicked not works on Android with Xamarin.Forms 2.4.0.280
bug confirmed fixed priority-high
The popup does not close when Background click on Android with Xamarin.Forms 2.4.0.280. This works with version 2.3.x.
1.0
CloseWhenBackgroundIsClicked not works on Android with Xamarin.Forms 2.4.0.280 - The popup does not close when Background click on Android with Xamarin.Forms 2.4.0.280. This works with version 2.3.x.
priority
closewhenbackgroundisclicked not works on android with xamarin forms the popup does not close when background click on android with xamarin forms this works with version x
1
39,921
8,704,578,819
IssuesEvent
2018-12-05 19:48:10
pgbackrest/pgbackrest
https://api.github.com/repos/pgbackrest/pgbackrest
closed
Archive command can leak s3 access (not secret) key in Postgres logs
enhancement (code)
Under certain circumstances when archiving to s3, the `repo-s3-key` can be leaked in the Postgres logs. This comes back because s3 returns the request header, which has the credentials in it. Postgres picks that up and logs it to the PG log. The first time I saw this was when there was an s3 issue and authentication was failing. It's easy to reproduce in test. Setup pgBackRest to archive to s3. Once it's working, change the `repo-s3-key-secret` to an invalid key and try archiving a transaction log. pgBackRests logging is configured as: ``` log-level-console=warn log-level-file=info log-level-stderr=warn ``` A heavily redacted version of the log output is: ``` 2018-08-30 18:54:14 UTC [104]: [12-1] user=,db=,client= LOG: archive command failed with exit code 39 2018-08-30 18:54:14 UTC [104]: [13-1] user=,db=,client= DETAIL: The failed archive command was: pgbackrest --stanza=mydb --config=/conf/pgbackrest.conf archive-push pg_xlog/000000030000000000000007 ERROR: [039]: raised from local-1 process: S3 request error [403] Forbidden *** request header *** GET /aa-software/xxxxxxxxxxxx/pgbackrest/archive/mydb/archive.info? HTTP/1.1 authorization: AWS4-HMAC-SHA256 Credential=<repo-s3-key is here>/..... <headers continue> ```
1.0
Archive command can leak s3 access (not secret) key in Postgres logs - Under certain circumstances when archiving to s3, the `repo-s3-key` can be leaked in the Postgres logs. This comes back because s3 returns the request header, which has the credentials in it. Postgres picks that up and logs it to the PG log. The first time I saw this was when there was an s3 issue and authentication was failing. It's easy to reproduce in test. Setup pgBackRest to archive to s3. Once it's working, change the `repo-s3-key-secret` to an invalid key and try archiving a transaction log. pgBackRests logging is configured as: ``` log-level-console=warn log-level-file=info log-level-stderr=warn ``` A heavily redacted version of the log output is: ``` 2018-08-30 18:54:14 UTC [104]: [12-1] user=,db=,client= LOG: archive command failed with exit code 39 2018-08-30 18:54:14 UTC [104]: [13-1] user=,db=,client= DETAIL: The failed archive command was: pgbackrest --stanza=mydb --config=/conf/pgbackrest.conf archive-push pg_xlog/000000030000000000000007 ERROR: [039]: raised from local-1 process: S3 request error [403] Forbidden *** request header *** GET /aa-software/xxxxxxxxxxxx/pgbackrest/archive/mydb/archive.info? HTTP/1.1 authorization: AWS4-HMAC-SHA256 Credential=<repo-s3-key is here>/..... <headers continue> ```
non_priority
archive command can leak access not secret key in postgres logs under certain circumstances when archiving to the repo key can be leaked in the postgres logs this comes back because returns the request header which has the credentials in it postgres picks that up and logs it to the pg log the first time i saw this was when there was an issue and authentication was failing it s easy to reproduce in test setup pgbackrest to archive to once it s working change the repo key secret to an invalid key and try archiving a transaction log pgbackrests logging is configured as log level console warn log level file info log level stderr warn a heavily redacted version of the log output is utc user db client log archive command failed with exit code utc user db client detail the failed archive command was pgbackrest stanza mydb config conf pgbackrest conf archive push pg xlog error raised from local process request error forbidden request header get aa software xxxxxxxxxxxx pgbackrest archive mydb archive info http authorization hmac credential
0
329,661
24,231,192,220
IssuesEvent
2022-09-26 18:26:36
supabase/supabase
https://api.github.com/repos/supabase/supabase
closed
docs: select permissions required for storage upsert
bug documentation
more context - https://github.com/supabase/supabase/discussions/2679 Docs need to be updated
1.0
docs: select permissions required for storage upsert - more context - https://github.com/supabase/supabase/discussions/2679 Docs need to be updated
non_priority
docs select permissions required for storage upsert more context docs need to be updated
0
79,285
7,703,780,421
IssuesEvent
2018-05-21 09:41:03
DrewAPicture/ensemble
https://api.github.com/repos/DrewAPicture/ensemble
opened
Add actual search support for Contests
::Contests Enhancement
The search box for the Contests list table is currently disabled because there is only _sort of_ support for searching contests. It would be nice if there was the ability to search a set of filterable contest fields and turn that on (and therefore also support searches via query).
1.0
Add actual search support for Contests - The search box for the Contests list table is currently disabled because there is only _sort of_ support for searching contests. It would be nice if there was the ability to search a set of filterable contest fields and turn that on (and therefore also support searches via query).
non_priority
add actual search support for contests the search box for the contests list table is currently disabled because there is only sort of support for searching contests it would be nice if there was the ability to search a set of filterable contest fields and turn that on and therefore also support searches via query
0
125,387
4,956,455,005
IssuesEvent
2016-12-01 23:33:02
gravityview/GravityView
https://api.github.com/repos/gravityview/GravityView
opened
Filter & Sort tab not loading in Chrome
Bug Core: Administration Core: UI/UX Priority: Low
Works in Safari. <img width="809" alt="unknown" src="https://cloud.githubusercontent.com/assets/870979/20817234/b5b10a92-b7e3-11e6-8429-7aa01622cd79.png"> Reported here: https://secure.helpscout.net/conversation/285133290/7556/
1.0
Filter & Sort tab not loading in Chrome - Works in Safari. <img width="809" alt="unknown" src="https://cloud.githubusercontent.com/assets/870979/20817234/b5b10a92-b7e3-11e6-8429-7aa01622cd79.png"> Reported here: https://secure.helpscout.net/conversation/285133290/7556/
priority
filter sort tab not loading in chrome works in safari img width alt unknown src reported here
1
639,157
20,747,763,368
IssuesEvent
2022-03-15 02:11:26
apache/echarts
https://api.github.com/repos/apache/echarts
closed
Null values with category-based Y-axis don't render correctly
bug priority: high en
### Version 5.1.2 ### Reproduction link [https://codepen.io/htr3n/pen/YzQKMrq](https://codepen.io/htr3n/pen/YzQKMrq) ### Steps to reproduce 1. Define a chart with the following options - A-axis with type "category", set `data` to some kind of category such as `["6:00", "6:30", "7:00", "7:30"]` - Y-axis with type "category", set `data` to some sort of category for instance `["A", "B", "C", "D"]` - A series with a list of null input data: `data: [null, null, null, null],` ### What is expected? The chart should not render any values because they are all null. ### What is actually happening? The chart renders the first value at the position of the first Y-axis category. --- Replace `null` with `undefined`, the chart seems to render correctly. <!-- This issue is generated by echarts-issue-helper. DO NOT REMOVE --> <!-- This issue is in English. DO NOT REMOVE -->
1.0
Null values with category-based Y-axis don't render correctly - ### Version 5.1.2 ### Reproduction link [https://codepen.io/htr3n/pen/YzQKMrq](https://codepen.io/htr3n/pen/YzQKMrq) ### Steps to reproduce 1. Define a chart with the following options - A-axis with type "category", set `data` to some kind of category such as `["6:00", "6:30", "7:00", "7:30"]` - Y-axis with type "category", set `data` to some sort of category for instance `["A", "B", "C", "D"]` - A series with a list of null input data: `data: [null, null, null, null],` ### What is expected? The chart should not render any values because they are all null. ### What is actually happening? The chart renders the first value at the position of the first Y-axis category. --- Replace `null` with `undefined`, the chart seems to render correctly. <!-- This issue is generated by echarts-issue-helper. DO NOT REMOVE --> <!-- This issue is in English. DO NOT REMOVE -->
priority
null values with category based y axis don t render correctly version reproduction link steps to reproduce define a chart with the following options a axis with type category set data to some kind of category such as y axis with type category set data to some sort of category for instance a series with a list of null input data data what is expected the chart should not render any values because they are all null what is actually happening the chart renders the first value at the position of the first y axis category replace null with undefined the chart seems to render correctly
1
820,510
30,775,497,469
IssuesEvent
2023-07-31 06:01:48
LiteLDev/LiteLoader.NET
https://api.github.com/repos/LiteLDev/LiteLoader.NET
closed
Hooks: New HookAPI
priority: normal status: in progress type: enhancement about: new LL.NET
### Is your feature request related to a problem? Please describe. Entering a delegate type to create hook's class and register it is not very convenient, and there is a way to get rid of it. ![image](https://user-images.githubusercontent.com/49648416/194613190-e993703b-4c36-4e2e-9b64-3df9f07f37f3.png) ### Describe the solution you'd like Hook delegate type, its parameters should be defined automatically by `Hook` field.
1.0
Hooks: New HookAPI - ### Is your feature request related to a problem? Please describe. Entering a delegate type to create hook's class and register it is not very convenient, and there is a way to get rid of it. ![image](https://user-images.githubusercontent.com/49648416/194613190-e993703b-4c36-4e2e-9b64-3df9f07f37f3.png) ### Describe the solution you'd like Hook delegate type, its parameters should be defined automatically by `Hook` field.
priority
hooks new hookapi is your feature request related to a problem please describe entering a delegate type to create hook s class and register it is not very convenient and there is a way to get rid of it describe the solution you d like hook delegate type its parameters should be defined automatically by hook field
1
294,843
9,049,296,682
IssuesEvent
2019-02-12 03:54:45
OctopusDeploy/Issues
https://api.github.com/repos/OctopusDeploy/Issues
closed
Service Fabric targets with Certificates fails on cloud-hosted instances
area/cloud priority
# Prerequisites - [x] I have verified the problem exists in the latest version - [x] I have searched [open](https://github.com/OctopusDeploy/Issues/issues) and [closed](https://github.com/OctopusDeploy/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already reported - [x] I have written a descriptive issue title - [x] I have linked the original source of this report - [x] I have tagged the issue appropriately (area/*, kind/bug, tag/regression?) # The bug When using a cloud-hosted instance of Octopus, if you try to use Service Fabric targets with the Certificates as the security mode, exceptions are occurring. **PLEASE NOTE** - I was not able to reproduce the same exception the customer has reported, but I received a different exception. When working through this issue, please investigate that the fix will solve the original issue reported by the customer as well. ## Steps to reproduce 1. Create a cloud instance 2. Create an SF cluster on Azure, using certificates and a key vault, wait until the cluster is "Ready" 3. Create an SF target in your cloud instance, setup to use the certificate available on Azure 4. See error during health check (see attached log) ### Screen capture <img width="1527" alt="screen shot 2018-08-27 at 12 38 43 pm" src="https://user-images.githubusercontent.com/819605/44637712-299cf000-a9f6-11e8-8af7-df93f54fe5fd.png"> ### Log exerpt [ServerTasks-29.log.txt](https://github.com/OctopusDeploy/Issues/files/2322227/ServerTasks-29.log.txt) ## Affected versions Occurs in latest version of Octopus `2018.7.11` from a cloud-hosted instance. ## Workarounds Overrride the default certificate store location as part of the step: ![image](https://user-images.githubusercontent.com/122651/52610588-879e9080-2ecd-11e9-800d-8e3ddeba2466.png) ## Source https://help.octopus.com/t/unable-to-create-azure-service-fabric-target/21133
1.0
Service Fabric targets with Certificates fails on cloud-hosted instances - # Prerequisites - [x] I have verified the problem exists in the latest version - [x] I have searched [open](https://github.com/OctopusDeploy/Issues/issues) and [closed](https://github.com/OctopusDeploy/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already reported - [x] I have written a descriptive issue title - [x] I have linked the original source of this report - [x] I have tagged the issue appropriately (area/*, kind/bug, tag/regression?) # The bug When using a cloud-hosted instance of Octopus, if you try to use Service Fabric targets with the Certificates as the security mode, exceptions are occurring. **PLEASE NOTE** - I was not able to reproduce the same exception the customer has reported, but I received a different exception. When working through this issue, please investigate that the fix will solve the original issue reported by the customer as well. ## Steps to reproduce 1. Create a cloud instance 2. Create an SF cluster on Azure, using certificates and a key vault, wait until the cluster is "Ready" 3. Create an SF target in your cloud instance, setup to use the certificate available on Azure 4. See error during health check (see attached log) ### Screen capture <img width="1527" alt="screen shot 2018-08-27 at 12 38 43 pm" src="https://user-images.githubusercontent.com/819605/44637712-299cf000-a9f6-11e8-8af7-df93f54fe5fd.png"> ### Log exerpt [ServerTasks-29.log.txt](https://github.com/OctopusDeploy/Issues/files/2322227/ServerTasks-29.log.txt) ## Affected versions Occurs in latest version of Octopus `2018.7.11` from a cloud-hosted instance. ## Workarounds Overrride the default certificate store location as part of the step: ![image](https://user-images.githubusercontent.com/122651/52610588-879e9080-2ecd-11e9-800d-8e3ddeba2466.png) ## Source https://help.octopus.com/t/unable-to-create-azure-service-fabric-target/21133
priority
service fabric targets with certificates fails on cloud hosted instances prerequisites i have verified the problem exists in the latest version i have searched and issues to make sure it isn t already reported i have written a descriptive issue title i have linked the original source of this report i have tagged the issue appropriately area kind bug tag regression the bug when using a cloud hosted instance of octopus if you try to use service fabric targets with the certificates as the security mode exceptions are occurring please note i was not able to reproduce the same exception the customer has reported but i received a different exception when working through this issue please investigate that the fix will solve the original issue reported by the customer as well steps to reproduce create a cloud instance create an sf cluster on azure using certificates and a key vault wait until the cluster is ready create an sf target in your cloud instance setup to use the certificate available on azure see error during health check see attached log screen capture img width alt screen shot at pm src log exerpt affected versions occurs in latest version of octopus from a cloud hosted instance workarounds overrride the default certificate store location as part of the step source
1
318,528
9,693,884,370
IssuesEvent
2019-05-24 17:20:37
CosmiQ/solaris
https://api.github.com/repos/CosmiQ/solaris
closed
Re-write image stitching using torch/tensorflow
Difficulty: Medium Priority: Medium Type: Maintenance
As it stands now, `sol.raster.image.stitch_images` uses numpy to stitch images together. This means that images have to be moved back from the GPU to the CPU to run, and lose any advantage that GPU processing could potentially provide. Particularly since this is almost always done for post-inference images (i.e. `stitch_images()` is called within `sol.nets.infer.Inferer.__call__()`), the objects being merged are likely to often be torch tensors (or could be converted to torch tensors if they're coming from keras). We should therefore implement GPU-based post-processing using `stitch_images()`: - [ ] re-write `stitch_images()` to use torch tensor operations instead of numpy arrays - [ ] enable checking for GPU availability and use GPUs if possible - [ ] enable check to see if data is in a numpy array when it's read in, and if so, convert it to a torch tensor
1.0
Re-write image stitching using torch/tensorflow - As it stands now, `sol.raster.image.stitch_images` uses numpy to stitch images together. This means that images have to be moved back from the GPU to the CPU to run, and lose any advantage that GPU processing could potentially provide. Particularly since this is almost always done for post-inference images (i.e. `stitch_images()` is called within `sol.nets.infer.Inferer.__call__()`), the objects being merged are likely to often be torch tensors (or could be converted to torch tensors if they're coming from keras). We should therefore implement GPU-based post-processing using `stitch_images()`: - [ ] re-write `stitch_images()` to use torch tensor operations instead of numpy arrays - [ ] enable checking for GPU availability and use GPUs if possible - [ ] enable check to see if data is in a numpy array when it's read in, and if so, convert it to a torch tensor
priority
re write image stitching using torch tensorflow as it stands now sol raster image stitch images uses numpy to stitch images together this means that images have to be moved back from the gpu to the cpu to run and lose any advantage that gpu processing could potentially provide particularly since this is almost always done for post inference images i e stitch images is called within sol nets infer inferer call the objects being merged are likely to often be torch tensors or could be converted to torch tensors if they re coming from keras we should therefore implement gpu based post processing using stitch images re write stitch images to use torch tensor operations instead of numpy arrays enable checking for gpu availability and use gpus if possible enable check to see if data is in a numpy array when it s read in and if so convert it to a torch tensor
1
28,055
5,428,749,306
IssuesEvent
2017-03-03 16:38:16
maryvilledev/codesplainUI
https://api.github.com/repos/maryvilledev/codesplainUI
opened
Project Wikis
Documentation
It would be beneficial to edit our repo's wikis so we can have a centralized base of relevant information
1.0
Project Wikis - It would be beneficial to edit our repo's wikis so we can have a centralized base of relevant information
non_priority
project wikis it would be beneficial to edit our repo s wikis so we can have a centralized base of relevant information
0
371,759
10,981,078,813
IssuesEvent
2019-11-30 18:59:43
BendroCorp/bendrocorp-app
https://api.github.com/repos/BendroCorp/bendrocorp-app
opened
System Map Search View
api supported effort: high feature priority:high
- [ ] Offer a map view or a searchable list view - switch via toggle and persist preference - [ ] Search View (will be the default mobile view when added) - [ ] Searching for something adds it to the recently viewed list at the top of the system map - [ ] Has a list of all items contained within system map and their essential data - [ ] Searchable by title - [ ] Opens to an object details page
1.0
System Map Search View - - [ ] Offer a map view or a searchable list view - switch via toggle and persist preference - [ ] Search View (will be the default mobile view when added) - [ ] Searching for something adds it to the recently viewed list at the top of the system map - [ ] Has a list of all items contained within system map and their essential data - [ ] Searchable by title - [ ] Opens to an object details page
priority
system map search view offer a map view or a searchable list view switch via toggle and persist preference search view will be the default mobile view when added searching for something adds it to the recently viewed list at the top of the system map has a list of all items contained within system map and their essential data searchable by title opens to an object details page
1
61,974
6,767,442,968
IssuesEvent
2017-10-26 03:22:32
Microsoft/vstest
https://api.github.com/repos/Microsoft/vstest
closed
Enable 'dotnet test' to be run on published output
dotnet-test exp: dotnet-cli
_From @kendrahavens on August 31, 2017 17:9_ ## Steps to reproduce After publishing users should be able to run tests on an already compiled executable. Would address [dotnet/core #917](https://github.com/dotnet/core/issues/917) This is more important with the development of IL Linker [#915](https://github.com/dotnet/core/issues/915). The [Linker Instructions](https://github.com/dotnet/core/blob/master/samples/linker-instructions.md) mention that the final executables need to be tested after linking. ## Expected behavior Be able to run 'dotnet test' on published output. ## Actual behavior You can not today. (Or we need to add documentation if you can!) @smalpani-msft _Copied from original issue: dotnet/cli#7570_
1.0
Enable 'dotnet test' to be run on published output - _From @kendrahavens on August 31, 2017 17:9_ ## Steps to reproduce After publishing users should be able to run tests on an already compiled executable. Would address [dotnet/core #917](https://github.com/dotnet/core/issues/917) This is more important with the development of IL Linker [#915](https://github.com/dotnet/core/issues/915). The [Linker Instructions](https://github.com/dotnet/core/blob/master/samples/linker-instructions.md) mention that the final executables need to be tested after linking. ## Expected behavior Be able to run 'dotnet test' on published output. ## Actual behavior You can not today. (Or we need to add documentation if you can!) @smalpani-msft _Copied from original issue: dotnet/cli#7570_
non_priority
enable dotnet test to be run on published output from kendrahavens on august steps to reproduce after publishing users should be able to run tests on an already compiled executable would address this is more important with the development of il linker the mention that the final executables need to be tested after linking expected behavior be able to run dotnet test on published output actual behavior you can not today or we need to add documentation if you can smalpani msft copied from original issue dotnet cli
0
641,775
20,834,249,162
IssuesEvent
2022-03-19 23:49:16
LemonUIbyLemon/LemonUI
https://api.github.com/repos/LemonUIbyLemon/LemonUI
closed
Big Message weapon purchased crash
status: confirmed priority: p1 high type: bug
I get this error when trying to use the Big Message scaleform: > System.ArgumentException: Unexpected argument type UInt32. Parameter name: parameters at LemonUI.Scaleform.BaseScaleform.CallFunctionBase(String function, Object[] parameters) at LemonUI.Scaleform.BaseScaleform.CallFunction(String function, Object[] parameters) at LemonUI.Scaleform.BigMessage..ctor(String title, String weapon, WeaponHash hash) It happens when this line gets executed: ![https://i.imgur.com/2J0yTYd.png](https://i.imgur.com/2J0yTYd.png)
1.0
Big Message weapon purchased crash - I get this error when trying to use the Big Message scaleform: > System.ArgumentException: Unexpected argument type UInt32. Parameter name: parameters at LemonUI.Scaleform.BaseScaleform.CallFunctionBase(String function, Object[] parameters) at LemonUI.Scaleform.BaseScaleform.CallFunction(String function, Object[] parameters) at LemonUI.Scaleform.BigMessage..ctor(String title, String weapon, WeaponHash hash) It happens when this line gets executed: ![https://i.imgur.com/2J0yTYd.png](https://i.imgur.com/2J0yTYd.png)
priority
big message weapon purchased crash i get this error when trying to use the big message scaleform system argumentexception unexpected argument type parameter name parameters at lemonui scaleform basescaleform callfunctionbase string function object parameters at lemonui scaleform basescaleform callfunction string function object parameters at lemonui scaleform bigmessage ctor string title string weapon weaponhash hash it happens when this line gets executed
1
20,992
11,573,217,644
IssuesEvent
2020-02-21 02:45:38
Azure/azure-sdk-for-js
https://api.github.com/repos/Azure/azure-sdk-for-js
closed
[@azure/graph] getServicePrincipalsIdByAppId return MissingSubscription
Graph Service Attention customer-reported
- **Package Name**: "@azure/graph" - **Package Version**: "5.0.0" - **Operating system**: - [ ] **nodejs** - **version**: 10.51 - [ ] **browser** - **name/version**: - [ ] **typescript** - **version**: - Is the bug related to **documentation** in - [ ] README.md - [ ] source code documentation - [ ] SDK API docs on https://docs.microsoft.com **Describe the bug** A clear and concise description of what the bug is. getServicePrincipalsIdByAppId return 404 error '{"error":{"code":"MissingSubscription","message":"The request did not have a subscription or a valid tenant level resource provider."}}' **To Reproduce** Steps to reproduce the behavior: 1 import pacakge: const {GraphRbacManagementClient} = require('@azure/graph') const msRestNodeAuth = require("@azure/ms-rest-nodeauth"); 2. sign in with serviceprincipal msRestNodeAuth.loginWithServicePrincipalSecret(clientId, secret, domain) 3 create graphic client GraphRbacManagementClient const graphClient = new GraphRbacManagementClient(credentials, domain) 4. get service principalsIdByAppId graphClient.applications.getServicePrincipalsIdByAppId(clientId, servicecallback) **Expected behavior** A clear and concise description of what you expected to happen. expect to get service principalsId **Screenshots** If applicable, add screenshots to help explain your problem. **Additional context** Add any other context about the problem here. { Error: The request did not have a subscription or a valid tenant level resource provider. at new RestError (/home/local/TERADICI/fwang/repos/CAM/ms_authorization/node_modules/@azure/ms-rest-js/dist/msRest.node.js:2400:28) at /home/local/TERADICI/fwang/repos/CAM/ms_authorization/node_modules/@azure/ms-rest-js/dist/msRest.node.js:3544:37 at process._tickCallback (internal/process/next_tick.js:68:7) code: 'MissingSubscription', statusCode: 404, request: WebResource { streamResponseBody: false, url: 'https://management.azure.com/xxxxxx/servicePrincipalsByAppId/xxxxxx/objectId?api-version=1.6', method: 'GET', headers: HttpHeaders { _headersMap: [Object] }, body: undefined, query: undefined, formData: undefined, withCredentials: false, abortSignal: undefined, timeout: 0, onUploadProgress: undefined, onDownloadProgress: undefined, proxySettings: undefined, keepAlive: undefined, operationSpec: { httpMethod: 'GET', path: '{tenantID}/servicePrincipalsByAppId/{applicationID}/objectId', urlParameters: [Array], queryParameters: [Array], headerParameters: [Array], responses: [Object], serializer: [Serializer] } }, response: { body: '{"error":{"code":"MissingSubscription","message":"The request did not have a subscription or a valid tenant level resource provider."}}', headers: HttpHeaders { _headersMap: [Object] }, status: 404 }, body: { error: { code: 'MissingSubscription', message: 'The request did not have a subscription or a valid tenant level resource provider.' } } }
1.0
[@azure/graph] getServicePrincipalsIdByAppId return MissingSubscription - - **Package Name**: "@azure/graph" - **Package Version**: "5.0.0" - **Operating system**: - [ ] **nodejs** - **version**: 10.51 - [ ] **browser** - **name/version**: - [ ] **typescript** - **version**: - Is the bug related to **documentation** in - [ ] README.md - [ ] source code documentation - [ ] SDK API docs on https://docs.microsoft.com **Describe the bug** A clear and concise description of what the bug is. getServicePrincipalsIdByAppId return 404 error '{"error":{"code":"MissingSubscription","message":"The request did not have a subscription or a valid tenant level resource provider."}}' **To Reproduce** Steps to reproduce the behavior: 1 import pacakge: const {GraphRbacManagementClient} = require('@azure/graph') const msRestNodeAuth = require("@azure/ms-rest-nodeauth"); 2. sign in with serviceprincipal msRestNodeAuth.loginWithServicePrincipalSecret(clientId, secret, domain) 3 create graphic client GraphRbacManagementClient const graphClient = new GraphRbacManagementClient(credentials, domain) 4. get service principalsIdByAppId graphClient.applications.getServicePrincipalsIdByAppId(clientId, servicecallback) **Expected behavior** A clear and concise description of what you expected to happen. expect to get service principalsId **Screenshots** If applicable, add screenshots to help explain your problem. **Additional context** Add any other context about the problem here. { Error: The request did not have a subscription or a valid tenant level resource provider. at new RestError (/home/local/TERADICI/fwang/repos/CAM/ms_authorization/node_modules/@azure/ms-rest-js/dist/msRest.node.js:2400:28) at /home/local/TERADICI/fwang/repos/CAM/ms_authorization/node_modules/@azure/ms-rest-js/dist/msRest.node.js:3544:37 at process._tickCallback (internal/process/next_tick.js:68:7) code: 'MissingSubscription', statusCode: 404, request: WebResource { streamResponseBody: false, url: 'https://management.azure.com/xxxxxx/servicePrincipalsByAppId/xxxxxx/objectId?api-version=1.6', method: 'GET', headers: HttpHeaders { _headersMap: [Object] }, body: undefined, query: undefined, formData: undefined, withCredentials: false, abortSignal: undefined, timeout: 0, onUploadProgress: undefined, onDownloadProgress: undefined, proxySettings: undefined, keepAlive: undefined, operationSpec: { httpMethod: 'GET', path: '{tenantID}/servicePrincipalsByAppId/{applicationID}/objectId', urlParameters: [Array], queryParameters: [Array], headerParameters: [Array], responses: [Object], serializer: [Serializer] } }, response: { body: '{"error":{"code":"MissingSubscription","message":"The request did not have a subscription or a valid tenant level resource provider."}}', headers: HttpHeaders { _headersMap: [Object] }, status: 404 }, body: { error: { code: 'MissingSubscription', message: 'The request did not have a subscription or a valid tenant level resource provider.' } } }
non_priority
getserviceprincipalsidbyappid return missingsubscription package name azure graph package version operating system nodejs version browser name version typescript version is the bug related to documentation in readme md source code documentation sdk api docs on describe the bug a clear and concise description of what the bug is getserviceprincipalsidbyappid return error error code missingsubscription message the request did not have a subscription or a valid tenant level resource provider to reproduce steps to reproduce the behavior import pacakge const graphrbacmanagementclient require azure graph const msrestnodeauth require azure ms rest nodeauth sign in with serviceprincipal msrestnodeauth loginwithserviceprincipalsecret clientid secret domain create graphic client graphrbacmanagementclient const graphclient new graphrbacmanagementclient credentials domain get service principalsidbyappid graphclient applications getserviceprincipalsidbyappid clientid servicecallback expected behavior a clear and concise description of what you expected to happen expect to get service principalsid screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here error the request did not have a subscription or a valid tenant level resource provider at new resterror home local teradici fwang repos cam ms authorization node modules azure ms rest js dist msrest node js at home local teradici fwang repos cam ms authorization node modules azure ms rest js dist msrest node js at process tickcallback internal process next tick js code missingsubscription statuscode request webresource streamresponsebody false url method get headers httpheaders headersmap body undefined query undefined formdata undefined withcredentials false abortsignal undefined timeout onuploadprogress undefined ondownloadprogress undefined proxysettings undefined keepalive undefined operationspec httpmethod get path tenantid serviceprincipalsbyappid applicationid objectid urlparameters queryparameters headerparameters responses serializer response body error code missingsubscription message the request did not have a subscription or a valid tenant level resource provider headers httpheaders headersmap status body error code missingsubscription message the request did not have a subscription or a valid tenant level resource provider
0
43,084
11,463,532,161
IssuesEvent
2020-02-07 16:11:03
snowplow/snowplow-javascript-tracker
https://api.github.com/repos/snowplow/snowplow-javascript-tracker
closed
Fix non-linear incrementing of session index
type:defect
- lasts for as long as `sessionCookieTimeout` - doesn't respect `heartbeatDelay` (multiple page pings per second despite an heartbeat of 5s) - possible leads: - `configHeartBeatTimer` might somehow be `null` [here](https://github.com/snowplow/snowplow-javascript-tracker/blob/master/src/js/tracker.js#L1291) which would cause the callback to happen relentlessly - multiple [`setInterval`](https://github.com/snowplow/snowplow-javascript-tracker/blob/master/src/js/tracker.js#L1280) are set, that could be solved by `clearInterval` if we mean to have only one at any time Session indices look like the following when ordered by `dvce_created_tstamp`: ![image](https://cloud.githubusercontent.com/assets/1737211/24855257/ee7e80a6-1dd8-11e7-87d1-0a5567620839.png) which would suggest some kind of mad callback. Tracker initialization looked like the following when the bug occurred: ```js window._tracker("newTracker", "js", "", { appId: "", platform: "web", post: !0, cookieLifetime: 31536e4, contexts: { webPage: !0, performanceTiming: !0 } }), window._tracker("enableLinkClickTracking", null, !0, !0), window._tracker("refreshLinkClickTracking"), window._tracker("enableActivityTracking", 5, 5), window._tracker("trackPageView"), ```
1.0
Fix non-linear incrementing of session index - - lasts for as long as `sessionCookieTimeout` - doesn't respect `heartbeatDelay` (multiple page pings per second despite an heartbeat of 5s) - possible leads: - `configHeartBeatTimer` might somehow be `null` [here](https://github.com/snowplow/snowplow-javascript-tracker/blob/master/src/js/tracker.js#L1291) which would cause the callback to happen relentlessly - multiple [`setInterval`](https://github.com/snowplow/snowplow-javascript-tracker/blob/master/src/js/tracker.js#L1280) are set, that could be solved by `clearInterval` if we mean to have only one at any time Session indices look like the following when ordered by `dvce_created_tstamp`: ![image](https://cloud.githubusercontent.com/assets/1737211/24855257/ee7e80a6-1dd8-11e7-87d1-0a5567620839.png) which would suggest some kind of mad callback. Tracker initialization looked like the following when the bug occurred: ```js window._tracker("newTracker", "js", "", { appId: "", platform: "web", post: !0, cookieLifetime: 31536e4, contexts: { webPage: !0, performanceTiming: !0 } }), window._tracker("enableLinkClickTracking", null, !0, !0), window._tracker("refreshLinkClickTracking"), window._tracker("enableActivityTracking", 5, 5), window._tracker("trackPageView"), ```
non_priority
fix non linear incrementing of session index lasts for as long as sessioncookietimeout doesn t respect heartbeatdelay multiple page pings per second despite an heartbeat of possible leads configheartbeattimer might somehow be null which would cause the callback to happen relentlessly multiple are set that could be solved by clearinterval if we mean to have only one at any time session indices look like the following when ordered by dvce created tstamp which would suggest some kind of mad callback tracker initialization looked like the following when the bug occurred js window tracker newtracker js appid platform web post cookielifetime contexts webpage performancetiming window tracker enablelinkclicktracking null window tracker refreshlinkclicktracking window tracker enableactivitytracking window tracker trackpageview
0
24,178
12,037,483,004
IssuesEvent
2020-04-13 21:57:10
emergenzeHack/covid19gr_issues
https://api.github.com/repos/emergenzeHack/covid19gr_issues
opened
goods donation and services offering
Missing position form goods donation and services offering
<pre><yamldata> Description: 'Live cameras ฯ„ฮฟฯ… Monterey Bay Aquarium ฯƒฯ„ฮทฮฝ ฮšฮฑฮปฮนฯ†ฯŒฯฮฝฮนฮฑ. ฮœฮต ฮทฮณฮตฯ„ฮนฮบฮฎ ฯ€ฮฑฯฮฟฯ…ฯƒฮฏฮฑ ฯƒฯ„ฮทฮฝ ฮตฮบฯ€ฮฑฮฏฮดฮตฯ…ฯƒฮท ฮณฮนฮฑ ฯ„ฮฟฯ…ฯ‚ ฯ‰ฮบฮตฮฑฮฝฮฟฯฯ‚, ฯ„ฮฟ Monterey Bay Aquarium ฮธฮตฯ‰ฯฮตฮฏฯ„ฮฑฮน ฮฑฯ€ฯŒ ฯ€ฮฟฮปฮปฮฟฯฯ‚ ฯŒฯ„ฮน ฮตฮฏฮฝฮฑฮน ฯ„ฮฟ ฮบฮฑฮปฯฯ„ฮตฯฮฟ ฯƒฯ„ฮฟฮฝ ฮบฯŒฯƒฮผฮฟ.' Institution: Monteray Bay Aquarium Legal: OK Link: https://www.montereybayaquarium.org/animals/live-cams Title: ฮœฮฌฮธฮต ฯ„ฮฟฮฝ ฯ…ฯ€ฮฟฮธฮฑฮปฮฌฯƒฯƒฮนฮฟ ฮบฯŒฯƒฮผฮฟ! </yamldata></pre>
1.0
goods donation and services offering - <pre><yamldata> Description: 'Live cameras ฯ„ฮฟฯ… Monterey Bay Aquarium ฯƒฯ„ฮทฮฝ ฮšฮฑฮปฮนฯ†ฯŒฯฮฝฮนฮฑ. ฮœฮต ฮทฮณฮตฯ„ฮนฮบฮฎ ฯ€ฮฑฯฮฟฯ…ฯƒฮฏฮฑ ฯƒฯ„ฮทฮฝ ฮตฮบฯ€ฮฑฮฏฮดฮตฯ…ฯƒฮท ฮณฮนฮฑ ฯ„ฮฟฯ…ฯ‚ ฯ‰ฮบฮตฮฑฮฝฮฟฯฯ‚, ฯ„ฮฟ Monterey Bay Aquarium ฮธฮตฯ‰ฯฮตฮฏฯ„ฮฑฮน ฮฑฯ€ฯŒ ฯ€ฮฟฮปฮปฮฟฯฯ‚ ฯŒฯ„ฮน ฮตฮฏฮฝฮฑฮน ฯ„ฮฟ ฮบฮฑฮปฯฯ„ฮตฯฮฟ ฯƒฯ„ฮฟฮฝ ฮบฯŒฯƒฮผฮฟ.' Institution: Monteray Bay Aquarium Legal: OK Link: https://www.montereybayaquarium.org/animals/live-cams Title: ฮœฮฌฮธฮต ฯ„ฮฟฮฝ ฯ…ฯ€ฮฟฮธฮฑฮปฮฌฯƒฯƒฮนฮฟ ฮบฯŒฯƒฮผฮฟ! </yamldata></pre>
non_priority
goods donation and services offering description live cameras ฯ„ฮฟฯ… monterey bay aquarium ฯƒฯ„ฮทฮฝ ฮบฮฑฮปฮนฯ†ฯŒฯฮฝฮนฮฑ ฮผฮต ฮทฮณฮตฯ„ฮนฮบฮฎ ฯ€ฮฑฯฮฟฯ…ฯƒฮฏฮฑ ฯƒฯ„ฮทฮฝ ฮตฮบฯ€ฮฑฮฏฮดฮตฯ…ฯƒฮท ฮณฮนฮฑ ฯ„ฮฟฯ…ฯ‚ ฯ‰ฮบฮตฮฑฮฝฮฟฯฯ‚ ฯ„ฮฟ monterey bay aquarium ฮธฮตฯ‰ฯฮตฮฏฯ„ฮฑฮน ฮฑฯ€ฯŒ ฯ€ฮฟฮปฮปฮฟฯฯ‚ ฯŒฯ„ฮน ฮตฮฏฮฝฮฑฮน ฯ„ฮฟ ฮบฮฑฮปฯฯ„ฮตฯฮฟ ฯƒฯ„ฮฟฮฝ ฮบฯŒฯƒฮผฮฟ institution monteray bay aquarium legal ok link title ฮผฮฌฮธฮต ฯ„ฮฟฮฝ ฯ…ฯ€ฮฟฮธฮฑฮปฮฌฯƒฯƒฮนฮฟ ฮบฯŒฯƒฮผฮฟ
0
4,193
2,713,189,950
IssuesEvent
2015-04-09 17:52:20
zaneu/cs373-idb
https://api.github.com/repos/zaneu/cs373-idb
closed
More unit tests
enhancement help wanted testing
I wrote some unit tests for ingredients, could @larry-liu , @jindtang , and @MenglinBrown get to work on writing some more? I was thinking of testing the search and filter by functionality of SQLAlchemy as much as possible. Note that we need 3+ for each model, but more is always better.
1.0
More unit tests - I wrote some unit tests for ingredients, could @larry-liu , @jindtang , and @MenglinBrown get to work on writing some more? I was thinking of testing the search and filter by functionality of SQLAlchemy as much as possible. Note that we need 3+ for each model, but more is always better.
non_priority
more unit tests i wrote some unit tests for ingredients could larry liu jindtang and menglinbrown get to work on writing some more i was thinking of testing the search and filter by functionality of sqlalchemy as much as possible note that we need for each model but more is always better
0
12,961
9,817,730,809
IssuesEvent
2019-06-13 17:27:38
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
[DocDB] DocumentDB Needs Tests
Cosmos Question Service Attention Test
- [ ] Add scenario tests for DocumentDB module - [x] Add DocumentDB to the help document map (PR #3267)
1.0
[DocDB] DocumentDB Needs Tests - - [ ] Add scenario tests for DocumentDB module - [x] Add DocumentDB to the help document map (PR #3267)
non_priority
documentdb needs tests add scenario tests for documentdb module add documentdb to the help document map pr
0
176,724
13,650,851,607
IssuesEvent
2020-09-26 21:24:05
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Missing banners to re-direct to dashboard for Isito
[zube]: To Test priority/0 team/ui
In monitoring/logging/cis, we have banners added to the top to have links to re-direct to the v2 version. Can we add these for Istio, alerts and notifiers? Istio -> `/istio` `The latest versions of Istio are only available in Cluster Explorer. See docs for details on what's new and how to migrate.` alerts and notifiers page -> `monitoring` - For this text, can update it to `We've updated alerts to be directly integrated with monitoring in Cluster Explorer. See docs for details on what's new and how to migrate.` `We've updated notifiers to be directly integrated with monitoring in Cluster Explorer. See docs for details on what's new and how to migrate.` <img width="855" alt="Screen Shot 2020-09-22 at 11 18 37 AM" src="https://user-images.githubusercontent.com/11410997/93921409-6450df80-fcc5-11ea-9ba3-b51c978c6834.png"> PR that added the banner for monitoring/logging/cis: https://github.com/rancher/ui/pull/4147
1.0
Missing banners to re-direct to dashboard for Isito - In monitoring/logging/cis, we have banners added to the top to have links to re-direct to the v2 version. Can we add these for Istio, alerts and notifiers? Istio -> `/istio` `The latest versions of Istio are only available in Cluster Explorer. See docs for details on what's new and how to migrate.` alerts and notifiers page -> `monitoring` - For this text, can update it to `We've updated alerts to be directly integrated with monitoring in Cluster Explorer. See docs for details on what's new and how to migrate.` `We've updated notifiers to be directly integrated with monitoring in Cluster Explorer. See docs for details on what's new and how to migrate.` <img width="855" alt="Screen Shot 2020-09-22 at 11 18 37 AM" src="https://user-images.githubusercontent.com/11410997/93921409-6450df80-fcc5-11ea-9ba3-b51c978c6834.png"> PR that added the banner for monitoring/logging/cis: https://github.com/rancher/ui/pull/4147
non_priority
missing banners to re direct to dashboard for isito in monitoring logging cis we have banners added to the top to have links to re direct to the version can we add these for istio alerts and notifiers istio istio the latest versions of istio are only available in cluster explorer see docs for details on what s new and how to migrate alerts and notifiers page monitoring for this text can update it to we ve updated alerts to be directly integrated with monitoring in cluster explorer see docs for details on what s new and how to migrate we ve updated notifiers to be directly integrated with monitoring in cluster explorer see docs for details on what s new and how to migrate img width alt screen shot at am src pr that added the banner for monitoring logging cis
0
95,890
3,961,746,906
IssuesEvent
2016-05-02 14:19:52
Esri/military-tools-webappbuilder
https://api.github.com/repos/Esri/military-tools-webappbuilder
closed
Change the repo description
effort-small priority - high
The description for this repo needs to be changed Currently: Web AppBuilder widgets for Military Analyst + Change to : Web AppBuilder widgets for Military Tools for ArcGIS
1.0
Change the repo description - The description for this repo needs to be changed Currently: Web AppBuilder widgets for Military Analyst + Change to : Web AppBuilder widgets for Military Tools for ArcGIS
priority
change the repo description the description for this repo needs to be changed currently web appbuilder widgets for military analyst change to web appbuilder widgets for military tools for arcgis
1
113,502
4,560,735,120
IssuesEvent
2016-09-14 09:07:47
dotKom/onlineweb4
https://api.github.com/repos/dotKom/onlineweb4
opened
Deleting a Careeropportunity in the dashboard does not actually delete
Package: Dashboard Priority: High Status: Available Type: Bug
When trying to delete a career opportunity in the dashboard, it does not actually delete it.
1.0
Deleting a Careeropportunity in the dashboard does not actually delete - When trying to delete a career opportunity in the dashboard, it does not actually delete it.
priority
deleting a careeropportunity in the dashboard does not actually delete when trying to delete a career opportunity in the dashboard it does not actually delete it
1
18,743
13,090,951,567
IssuesEvent
2020-08-03 05:13:07
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
[x64] Error in x64_checked_ubuntu_illink_flow
area-Infrastructure-coreclr
**Job:** https://ci.dot.net/job/dotnet_coreclr/job/master/view/x64/job/illink/job/x64_checked_ubuntu_illink/648/ **Log:** ``` ERROR: An error occurred in /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/Tools/dotnetcli/dotnet msbuild /nologo /verbosity:minimal /clp:Summary /p:RestoreDefaultOptimizationDataPackage=false /p:PortableBuild=true /p:UsePartialNGENOptimization=false /maxcpucount /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/tests/build.proj /flp:Verbosity=normal;LogFile=/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.log;Append=true /flp1:WarningsOnly;LogFile=/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.wrn;Append=true /flp2:ErrorsOnly;LogFile=/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.err;Append=true /l:BinClashLogger,Tools/Microsoft.DotNet.Build.Tasks.dll;LogFile=binclash.log /p:__BuildArch=x64 /p:__BuildType=Checked /p:__BuildOS=Linux /p:CLRTestPriorityToBuild=1. Check logs under /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink. Failed to build Managed tests build (build tests). See the build logs: /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.log /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.wrn /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.err ```
1.0
[x64] Error in x64_checked_ubuntu_illink_flow - **Job:** https://ci.dot.net/job/dotnet_coreclr/job/master/view/x64/job/illink/job/x64_checked_ubuntu_illink/648/ **Log:** ``` ERROR: An error occurred in /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/Tools/dotnetcli/dotnet msbuild /nologo /verbosity:minimal /clp:Summary /p:RestoreDefaultOptimizationDataPackage=false /p:PortableBuild=true /p:UsePartialNGENOptimization=false /maxcpucount /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/tests/build.proj /flp:Verbosity=normal;LogFile=/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.log;Append=true /flp1:WarningsOnly;LogFile=/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.wrn;Append=true /flp2:ErrorsOnly;LogFile=/mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.err;Append=true /l:BinClashLogger,Tools/Microsoft.DotNet.Build.Tasks.dll;LogFile=binclash.log /p:__BuildArch=x64 /p:__BuildType=Checked /p:__BuildOS=Linux /p:CLRTestPriorityToBuild=1. Check logs under /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink. Failed to build Managed tests build (build tests). See the build logs: /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.log /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.wrn /mnt/j/workspace/dotnet_coreclr/master/illink/x64_checked_ubuntu_illink/bin/Logs/Tests_Managed.Linux.x64.Checked.err ```
non_priority
error in checked ubuntu illink flow job log error an error occurred in mnt j workspace dotnet coreclr master illink checked ubuntu illink tools dotnetcli dotnet msbuild nologo verbosity minimal clp summary p restoredefaultoptimizationdatapackage false p portablebuild true p usepartialngenoptimization false maxcpucount mnt j workspace dotnet coreclr master illink checked ubuntu illink tests build proj flp verbosity normal logfile mnt j workspace dotnet coreclr master illink checked ubuntu illink bin logs tests managed linux checked log append true warningsonly logfile mnt j workspace dotnet coreclr master illink checked ubuntu illink bin logs tests managed linux checked wrn append true errorsonly logfile mnt j workspace dotnet coreclr master illink checked ubuntu illink bin logs tests managed linux checked err append true l binclashlogger tools microsoft dotnet build tasks dll logfile binclash log p buildarch p buildtype checked p buildos linux p clrtestprioritytobuild check logs under mnt j workspace dotnet coreclr master illink checked ubuntu illink failed to build managed tests build build tests see the build logs mnt j workspace dotnet coreclr master illink checked ubuntu illink bin logs tests managed linux checked log mnt j workspace dotnet coreclr master illink checked ubuntu illink bin logs tests managed linux checked wrn mnt j workspace dotnet coreclr master illink checked ubuntu illink bin logs tests managed linux checked err
0
156,884
5,990,599,402
IssuesEvent
2017-06-02 11:58:42
universAAL/platform
https://api.github.com/repos/universAAL/platform
closed
required testing bundle is not automatically installed
bug imported priority 3
_Originally Opened: @amedranogil (2014-05-08 11:41:39_) _Originally Closed: 2014-05-08 12:51:21_ When building an artifact with Integration tests with a clean Maven local repo, a required bundle is not automatically installed. It is suggested to add this bundle as a forced dependency in top level POMs with scope test, so it gets automatically installed at least the first time. testComposite(org.universAAL.ontology.test.ArtifactIT) Time elapsed: 0.49 sec &lt;&lt;&lt; ERROR! java.io.FileNotFoundException: C:\Users\amedrano\.m2\repository\org\slf4j\com.springsource.slf4j.api\1.5.0\com.springsource.slf4j.api-1.5.0.jar (El sistema no puede encontrar la ruta especificada) at java.io.FileInputStream.open(Native Method) at java.io.FileInputStream.&lt;init&gt;(Unknown Source) at org.springframework.core.io.FileSystemResource.getInputStream(FileSystemResource.java:108) at org.springframework.osgi.test.AbstractOsgiTests.installBundle(AbstractOsgiTests.java:323) at org.springframework.osgi.test.AbstractOsgiTests.startup(AbstractOsgiTests.java:253) at org.springframework.osgi.test.AbstractOsgiTests.prepareTestExecution(AbstractOsgiTests.java:374) at org.springframework.osgi.test.AbstractOsgiTests.runBare(AbstractOsgiTests.java:203) at org.springframework.osgi.test.AbstractOsgiTests$1.protect(AbstractOsgiTests.java:184) at junit.framework.TestResult.runProtected(TestResult.java:124) at org.springframework.osgi.test.AbstractOsgiTests.run(AbstractOsgiTests.java:181) at junit.framework.TestSuite.runTest(TestSuite.java:230) at junit.framework.TestSuite.run(TestSuite.java:225) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.apache.maven.surefire.junit.JUnitTestSet.execute(JUnitTestSet.java:95) at org.apache.maven.surefire.junit.JUnit3Provider.executeTestSet(JUnit3Provider.java:131) at org.apache.maven.surefire.junit.JUnit3Provider.invoke(JUnit3Provider.java:108) at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103) Results : Tests in error: org.universAAL.ontology.test.ArtifactIT#testComposite FileNotFoundException C:... -- From: _this issue has been automatically imported from our old issue tracker_
1.0
required testing bundle is not automatically installed - _Originally Opened: @amedranogil (2014-05-08 11:41:39_) _Originally Closed: 2014-05-08 12:51:21_ When building an artifact with Integration tests with a clean Maven local repo, a required bundle is not automatically installed. It is suggested to add this bundle as a forced dependency in top level POMs with scope test, so it gets automatically installed at least the first time. testComposite(org.universAAL.ontology.test.ArtifactIT) Time elapsed: 0.49 sec &lt;&lt;&lt; ERROR! java.io.FileNotFoundException: C:\Users\amedrano\.m2\repository\org\slf4j\com.springsource.slf4j.api\1.5.0\com.springsource.slf4j.api-1.5.0.jar (El sistema no puede encontrar la ruta especificada) at java.io.FileInputStream.open(Native Method) at java.io.FileInputStream.&lt;init&gt;(Unknown Source) at org.springframework.core.io.FileSystemResource.getInputStream(FileSystemResource.java:108) at org.springframework.osgi.test.AbstractOsgiTests.installBundle(AbstractOsgiTests.java:323) at org.springframework.osgi.test.AbstractOsgiTests.startup(AbstractOsgiTests.java:253) at org.springframework.osgi.test.AbstractOsgiTests.prepareTestExecution(AbstractOsgiTests.java:374) at org.springframework.osgi.test.AbstractOsgiTests.runBare(AbstractOsgiTests.java:203) at org.springframework.osgi.test.AbstractOsgiTests$1.protect(AbstractOsgiTests.java:184) at junit.framework.TestResult.runProtected(TestResult.java:124) at org.springframework.osgi.test.AbstractOsgiTests.run(AbstractOsgiTests.java:181) at junit.framework.TestSuite.runTest(TestSuite.java:230) at junit.framework.TestSuite.run(TestSuite.java:225) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.apache.maven.surefire.junit.JUnitTestSet.execute(JUnitTestSet.java:95) at org.apache.maven.surefire.junit.JUnit3Provider.executeTestSet(JUnit3Provider.java:131) at org.apache.maven.surefire.junit.JUnit3Provider.invoke(JUnit3Provider.java:108) at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103) Results : Tests in error: org.universAAL.ontology.test.ArtifactIT#testComposite FileNotFoundException C:... -- From: _this issue has been automatically imported from our old issue tracker_
priority
required testing bundle is not automatically installed originally opened amedranogil originally closed when building an artifact with integration tests with a clean maven local repo a required bundle is not automatically installed it is suggested to add this bundle as a forced dependency in top level poms with scope test so it gets automatically installed at least the first time testcomposite org universaal ontology test artifactit time elapsed sec lt lt lt error java io filenotfoundexception c users amedrano repository org com springsource api com springsource api jar el sistema no puede encontrar la ruta especificada at java io fileinputstream open native method at java io fileinputstream lt init gt unknown source at org springframework core io filesystemresource getinputstream filesystemresource java at org springframework osgi test abstractosgitests installbundle abstractosgitests java at org springframework osgi test abstractosgitests startup abstractosgitests java at org springframework osgi test abstractosgitests preparetestexecution abstractosgitests java at org springframework osgi test abstractosgitests runbare abstractosgitests java at org springframework osgi test abstractosgitests protect abstractosgitests java at junit framework testresult runprotected testresult java at org springframework osgi test abstractosgitests run abstractosgitests java at junit framework testsuite runtest testsuite java at junit framework testsuite run testsuite java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source at java lang reflect method invoke unknown source at org apache maven surefire junit junittestset execute junittestset java at org apache maven surefire junit executetestset java at org apache maven surefire junit invoke java at org apache maven surefire booter forkedbooter invokeproviderinsameclassloader forkedbooter java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java results tests in error org universaal ontology test artifactit testcomposite filenotfoundexception c from this issue has been automatically imported from our old issue tracker
1
229,689
7,582,728,377
IssuesEvent
2018-04-25 06:10:11
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.mclaren.com - site is not usable
browser-firefox priority-important
<!-- @browser: Firefox 58.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 --> <!-- @reported_with: desktop-reporter --> **URL**: http://www.mclaren.com/formula1/ **Browser / Version**: Firefox 58.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Complete website doesn't load **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2017/12/a1c604c5-a2ef-4c5d-8178-f55c74245748-thumb.jpg)](https://webcompat.com/uploads/2017/12/a1c604c5-a2ef-4c5d-8178-f55c74245748.jpg) _From [webcompat.com](https://webcompat.com/) with โค๏ธ_
1.0
www.mclaren.com - site is not usable - <!-- @browser: Firefox 58.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 --> <!-- @reported_with: desktop-reporter --> **URL**: http://www.mclaren.com/formula1/ **Browser / Version**: Firefox 58.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Complete website doesn't load **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2017/12/a1c604c5-a2ef-4c5d-8178-f55c74245748-thumb.jpg)](https://webcompat.com/uploads/2017/12/a1c604c5-a2ef-4c5d-8178-f55c74245748.jpg) _From [webcompat.com](https://webcompat.com/) with โค๏ธ_
priority
site is not usable url browser version firefox operating system windows tested another browser yes problem type site is not usable description complete website doesn t load steps to reproduce from with โค๏ธ
1
162,553
12,680,489,276
IssuesEvent
2020-06-19 13:48:05
Oldes/Rebol-issues
https://api.github.com/repos/Oldes/Rebol-issues
closed
TO-VALUE function to convert unset to none
Test.written Type.wish
_Submitted by:_ **BrianH** I want a function that can take one normal parameter of any type, which returns that parameter unevaluated, except for an #[unset!] value. When it is passed an #[unset!](or end! I suppose if one of those escape) then it should return #[none] instead. This function will need to be native, since it would be used in core control flow to defang unsets. And I have no idea what to call it, so I'm looking for suggestions. ``` rebol ; Rebol code equivalent func [value [any-type!]] [unless unset? :value [:value]] ``` --- <sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=2003)** [ Version: r3 master Type: Wish Platform: All Category: Native Reproduce: Always Fixed-in:none ]</sup> <sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/2003</sup> Comments: --- > **Rebolbot** commented on Mar 22, 2013: _Submitted by:_ **BrianH** Fork suggested DEVOID for this in SO chat, which is metaphorically accurate and less whimsical than DEFANG. The downside is that it makes reference to the "void" type in C-like languages that serves a similar purpose to returning unset in Rebol, but Rebol isn't a C-like language. Another option would be TO-VALUE since it converts non-values like #[unset!] to values like #[none]. It's a little long, but it fits in with the other TO-\* functions, though without the copying that those do. I don't want to call it VALUE because that is a common variable name. Other suggestions would be welcome. --- > **Rebolbot** commented on Mar 22, 2013: _Submitted by:_ **Ladislav** I prefer the TO-VALUE alternative. --- > **Rebolbot** commented on Mar 22, 2013: _Submitted by:_ **abolka** I prefer TO-VALUE as well. --- > **Rebolbot** commented on Mar 6, 2014: _Submitted by:_ **BrianH** Implemented as TO-VALUE in https://github.com/rebol/rebol/pull/205 --- > **Rebolbot** mentioned this issue on Jan 12, 2016: > [Make NONE! values no-op in slots processing results of a conditional expression](https://github.com/Oldes/Rebol-issues/issues/2078) --- > **Rebolbot** mentioned this issue on Jan 22, 2016: > [[Epic] Holes in our evaluation model](https://github.com/Oldes/Rebol-issues/issues/2127) --- > **Hostilefork** mentioned this issue on May 1, 2018: > [Allow INDEX? to take none as an argument, return none](https://github.com/Oldes/Rebol-issues/issues/1611) --- > **Rebolbot** added the **Type.wish** on Jan 12, 2016 --- > **Hostilefork** added a commit to **[metaeducation/ren-c](https://github.com/metaeducation/ren-c/)** that referenced this issue on May 1, 2018: > [Retake TRY for TO-VALUE, make NULL void's API representation](https://github.com/metaeducation/ren-c/commit/04f236d3bf5eac62565d1cb9fe2186e98293c6eb) --- > **Hostilefork** added a commit to **[hostilefork/rebol](https://github.com/hostilefork/rebol/)** that referenced this issue on May 1, 2018: > [Retake TRY for TO-VALUE, make NULL void's API representation](https://github.com/hostilefork/rebol/commit/4e44fe58da5417d381e83f3dc779fdc59fd16942) --- > **Hostilefork** added a commit to **[hostilefork/rebol](https://github.com/hostilefork/rebol/)** that referenced this issue on May 1, 2018: > [Retake TRY for TO-VALUE, make NULL void's API representation](https://github.com/hostilefork/rebol/commit/20f64be4519e9e6a852329088ec115546abba828) ---
1.0
TO-VALUE function to convert unset to none - _Submitted by:_ **BrianH** I want a function that can take one normal parameter of any type, which returns that parameter unevaluated, except for an #[unset!] value. When it is passed an #[unset!](or end! I suppose if one of those escape) then it should return #[none] instead. This function will need to be native, since it would be used in core control flow to defang unsets. And I have no idea what to call it, so I'm looking for suggestions. ``` rebol ; Rebol code equivalent func [value [any-type!]] [unless unset? :value [:value]] ``` --- <sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=2003)** [ Version: r3 master Type: Wish Platform: All Category: Native Reproduce: Always Fixed-in:none ]</sup> <sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/2003</sup> Comments: --- > **Rebolbot** commented on Mar 22, 2013: _Submitted by:_ **BrianH** Fork suggested DEVOID for this in SO chat, which is metaphorically accurate and less whimsical than DEFANG. The downside is that it makes reference to the "void" type in C-like languages that serves a similar purpose to returning unset in Rebol, but Rebol isn't a C-like language. Another option would be TO-VALUE since it converts non-values like #[unset!] to values like #[none]. It's a little long, but it fits in with the other TO-\* functions, though without the copying that those do. I don't want to call it VALUE because that is a common variable name. Other suggestions would be welcome. --- > **Rebolbot** commented on Mar 22, 2013: _Submitted by:_ **Ladislav** I prefer the TO-VALUE alternative. --- > **Rebolbot** commented on Mar 22, 2013: _Submitted by:_ **abolka** I prefer TO-VALUE as well. --- > **Rebolbot** commented on Mar 6, 2014: _Submitted by:_ **BrianH** Implemented as TO-VALUE in https://github.com/rebol/rebol/pull/205 --- > **Rebolbot** mentioned this issue on Jan 12, 2016: > [Make NONE! values no-op in slots processing results of a conditional expression](https://github.com/Oldes/Rebol-issues/issues/2078) --- > **Rebolbot** mentioned this issue on Jan 22, 2016: > [[Epic] Holes in our evaluation model](https://github.com/Oldes/Rebol-issues/issues/2127) --- > **Hostilefork** mentioned this issue on May 1, 2018: > [Allow INDEX? to take none as an argument, return none](https://github.com/Oldes/Rebol-issues/issues/1611) --- > **Rebolbot** added the **Type.wish** on Jan 12, 2016 --- > **Hostilefork** added a commit to **[metaeducation/ren-c](https://github.com/metaeducation/ren-c/)** that referenced this issue on May 1, 2018: > [Retake TRY for TO-VALUE, make NULL void's API representation](https://github.com/metaeducation/ren-c/commit/04f236d3bf5eac62565d1cb9fe2186e98293c6eb) --- > **Hostilefork** added a commit to **[hostilefork/rebol](https://github.com/hostilefork/rebol/)** that referenced this issue on May 1, 2018: > [Retake TRY for TO-VALUE, make NULL void's API representation](https://github.com/hostilefork/rebol/commit/4e44fe58da5417d381e83f3dc779fdc59fd16942) --- > **Hostilefork** added a commit to **[hostilefork/rebol](https://github.com/hostilefork/rebol/)** that referenced this issue on May 1, 2018: > [Retake TRY for TO-VALUE, make NULL void's API representation](https://github.com/hostilefork/rebol/commit/20f64be4519e9e6a852329088ec115546abba828) ---
non_priority
to value function to convert unset to none submitted by brianh i want a function that can take one normal parameter of any type which returns that parameter unevaluated except for an value when it is passed an or end i suppose if one of those escape then it should return instead this function will need to be native since it would be used in core control flow to defang unsets and i have no idea what to call it so i m looking for suggestions rebol rebol code equivalent func imported from imported from comments rebolbot commented on mar submitted by brianh fork suggested devoid for this in so chat which is metaphorically accurate and less whimsical than defang the downside is that it makes reference to the void type in c like languages that serves a similar purpose to returning unset in rebol but rebol isn t a c like language another option would be to value since it converts non values like to values like it s a little long but it fits in with the other to functions though without the copying that those do i don t want to call it value because that is a common variable name other suggestions would be welcome rebolbot commented on mar submitted by ladislav i prefer the to value alternative rebolbot commented on mar submitted by abolka i prefer to value as well rebolbot commented on mar submitted by brianh implemented as to value in rebolbot mentioned this issue on jan rebolbot mentioned this issue on jan holes in our evaluation model hostilefork mentioned this issue on may rebolbot added the type wish on jan hostilefork added a commit to that referenced this issue on may hostilefork added a commit to that referenced this issue on may hostilefork added a commit to that referenced this issue on may
0
451,799
32,042,143,377
IssuesEvent
2023-09-22 20:22:27
gustavoprezoto/EstagBee
https://api.github.com/repos/gustavoprezoto/EstagBee
closed
Requisitos nรฃo funcionais devem ser mensurรกveis
documentation invalid
Seguindo o padrรฃo passado pela professora, os requisitos nรฃo funcionais devem ser mensurรกveis. Exemplos: Exemplo ruim de RNF: [RNF01] As consultas ao sistema devem ser respondidas rapidamente (o que รฉ rapidamente?) Exemplo bom de RNF: [RNF02] As consultas ao sistema devem ser respondidas em menos de trรชs segundos
1.0
Requisitos nรฃo funcionais devem ser mensurรกveis - Seguindo o padrรฃo passado pela professora, os requisitos nรฃo funcionais devem ser mensurรกveis. Exemplos: Exemplo ruim de RNF: [RNF01] As consultas ao sistema devem ser respondidas rapidamente (o que รฉ rapidamente?) Exemplo bom de RNF: [RNF02] As consultas ao sistema devem ser respondidas em menos de trรชs segundos
non_priority
requisitos nรฃo funcionais devem ser mensurรกveis seguindo o padrรฃo passado pela professora os requisitos nรฃo funcionais devem ser mensurรกveis exemplos exemplo ruim de rnf as consultas ao sistema devem ser respondidas rapidamente o que รฉ rapidamente exemplo bom de rnf as consultas ao sistema devem ser respondidas em menos de trรชs segundos
0
682,209
23,336,596,122
IssuesEvent
2022-08-09 10:29:42
dnd-side-project/dnd-7th-7-frontend
https://api.github.com/repos/dnd-side-project/dnd-7th-7-frontend
closed
[update] ๊ฒฝ๋กœ ๋“ฑ๋ก ํƒญ ๋„ค๋น„๊ฒŒ์ด์…˜ ์ˆ˜์ • ๋ฐ ์ถ”๊ฐ€
๐Ÿ‘จ๐Ÿผโ€๐Ÿ”ง Type: Improvement ๐Ÿ”ด Priority: High
## ๐Ÿ”ฅ ์„ค๋ช… (์ด์Šˆ์— ๋Œ€ํ•œ ์ž์„ธํ•œ ์„ค๋ช…์„ ์ ์–ด์ฃผ์„ธ์š”!) - ๋””์ž์ธ์— ๋”ฐ๋ผ ๋„ค์ด๊ฒŒ์ดํ„ฐ ํƒญ์„ ์ˆ˜์ •ํ•˜๋Š” ์ž‘์—…์ž…๋‹ˆ๋‹ค. <br> ## ๐Ÿ”ฅ ETC (๊ธฐํƒ€ ์‚ฌํ•ญ์„ ์ ์–ด์ฃผ์„ธ์š”!) <img width="461" alt="image" src="https://user-images.githubusercontent.com/80577900/183613007-d58d50a3-7bba-45fb-8d87-56a2a044cf22.png"> <br> >#### types >- feat : ์ƒˆ๋กœ์šด ๊ธฐ๋Šฅ >- update : ๋กœ์ง ๋ณ€๊ฒฝ >- bug : ๋ฒ„๊ทธ ์ˆ˜์ • >- refactor : ํŒŒ์ผ ๋ฐ ํด๋” ๋ฆฌํŒฉํ† ๋ง >- style : ์ฝ”๋“œ ํ˜•์‹ ๋ณ€๊ฒฝ >- docs : ๋ฌธ์„œ ๊ด€๋ จ >- test : ํ…Œ์ŠคํŠธ ์ฝ”๋“œ >- etc : ๊ธฐํƒ€ ๋ณ€๊ฒฝ ์‚ฌํ•ญ >
1.0
[update] ๊ฒฝ๋กœ ๋“ฑ๋ก ํƒญ ๋„ค๋น„๊ฒŒ์ด์…˜ ์ˆ˜์ • ๋ฐ ์ถ”๊ฐ€ - ## ๐Ÿ”ฅ ์„ค๋ช… (์ด์Šˆ์— ๋Œ€ํ•œ ์ž์„ธํ•œ ์„ค๋ช…์„ ์ ์–ด์ฃผ์„ธ์š”!) - ๋””์ž์ธ์— ๋”ฐ๋ผ ๋„ค์ด๊ฒŒ์ดํ„ฐ ํƒญ์„ ์ˆ˜์ •ํ•˜๋Š” ์ž‘์—…์ž…๋‹ˆ๋‹ค. <br> ## ๐Ÿ”ฅ ETC (๊ธฐํƒ€ ์‚ฌํ•ญ์„ ์ ์–ด์ฃผ์„ธ์š”!) <img width="461" alt="image" src="https://user-images.githubusercontent.com/80577900/183613007-d58d50a3-7bba-45fb-8d87-56a2a044cf22.png"> <br> >#### types >- feat : ์ƒˆ๋กœ์šด ๊ธฐ๋Šฅ >- update : ๋กœ์ง ๋ณ€๊ฒฝ >- bug : ๋ฒ„๊ทธ ์ˆ˜์ • >- refactor : ํŒŒ์ผ ๋ฐ ํด๋” ๋ฆฌํŒฉํ† ๋ง >- style : ์ฝ”๋“œ ํ˜•์‹ ๋ณ€๊ฒฝ >- docs : ๋ฌธ์„œ ๊ด€๋ จ >- test : ํ…Œ์ŠคํŠธ ์ฝ”๋“œ >- etc : ๊ธฐํƒ€ ๋ณ€๊ฒฝ ์‚ฌํ•ญ >
priority
๊ฒฝ๋กœ ๋“ฑ๋ก ํƒญ ๋„ค๋น„๊ฒŒ์ด์…˜ ์ˆ˜์ • ๋ฐ ์ถ”๊ฐ€ ๐Ÿ”ฅ ์„ค๋ช… ์ด์Šˆ์— ๋Œ€ํ•œ ์ž์„ธํ•œ ์„ค๋ช…์„ ์ ์–ด์ฃผ์„ธ์š” ๋””์ž์ธ์— ๋”ฐ๋ผ ๋„ค์ด๊ฒŒ์ดํ„ฐ ํƒญ์„ ์ˆ˜์ •ํ•˜๋Š” ์ž‘์—…์ž…๋‹ˆ๋‹ค ๐Ÿ”ฅ etc ๊ธฐํƒ€ ์‚ฌํ•ญ์„ ์ ์–ด์ฃผ์„ธ์š” img width alt image src types feat ์ƒˆ๋กœ์šด ๊ธฐ๋Šฅ update ๋กœ์ง ๋ณ€๊ฒฝ bug ๋ฒ„๊ทธ ์ˆ˜์ • refactor ํŒŒ์ผ ๋ฐ ํด๋” ๋ฆฌํŒฉํ† ๋ง style ์ฝ”๋“œ ํ˜•์‹ ๋ณ€๊ฒฝ docs ๋ฌธ์„œ ๊ด€๋ จ test ํ…Œ์ŠคํŠธ ์ฝ”๋“œ etc ๊ธฐํƒ€ ๋ณ€๊ฒฝ ์‚ฌํ•ญ
1
14,433
9,308,404,115
IssuesEvent
2019-03-25 14:29:55
armbrustsamuel/springbootAnnotation
https://api.github.com/repos/armbrustsamuel/springbootAnnotation
opened
CVE-2017-12617 High Severity Vulnerability detected by WhiteSource
security vulnerability
## CVE-2017-12617 - High Severity Vulnerability <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.14.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>path: /root/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.14/tomcat-embed-core-8.5.14.jar</p> <p> <p>Library home page: <a href=http://tomcat.apache.org/>http://tomcat.apache.org/</a></p> Dependency Hierarchy: - spring-boot-starter-web-1.5.3.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-1.5.3.RELEASE.jar - :x: **tomcat-embed-core-8.5.14.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When running Apache Tomcat versions 9.0.0.M1 to 9.0.0, 8.5.0 to 8.5.22, 8.0.0.RC1 to 8.0.46 and 7.0.0 to 7.0.81 with HTTP PUTs enabled (e.g. via setting the readonly initialisation parameter of the Default servlet to false) it was possible to upload a JSP file to the server via a specially crafted request. This JSP could then be requested and any code it contained would be executed by the server. <p>Publish Date: 2017-10-04 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-12617>CVE-2017-12617</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://www.securitytracker.com/id/1039552">http://www.securitytracker.com/id/1039552</a></p> <p>Fix Resolution: The vendor has issued a fix (7.0.82, 8.0.47, 9.0.1). The vendor advisories are available at: http://tomcat.apache.org/security-7.html#Fixed_in_Apache_Tomcat_7.0.82 http://tomcat.apache.org/security-8.html#Fixed_in_Apache_Tomcat_8.0.47 http://tomcat.apache.org/security-9.html#Fixed_in_Apache_Tomcat_9.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-12617 High Severity Vulnerability detected by WhiteSource - ## CVE-2017-12617 - High Severity Vulnerability <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.14.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>path: /root/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.14/tomcat-embed-core-8.5.14.jar</p> <p> <p>Library home page: <a href=http://tomcat.apache.org/>http://tomcat.apache.org/</a></p> Dependency Hierarchy: - spring-boot-starter-web-1.5.3.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-1.5.3.RELEASE.jar - :x: **tomcat-embed-core-8.5.14.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When running Apache Tomcat versions 9.0.0.M1 to 9.0.0, 8.5.0 to 8.5.22, 8.0.0.RC1 to 8.0.46 and 7.0.0 to 7.0.81 with HTTP PUTs enabled (e.g. via setting the readonly initialisation parameter of the Default servlet to false) it was possible to upload a JSP file to the server via a specially crafted request. This JSP could then be requested and any code it contained would be executed by the server. <p>Publish Date: 2017-10-04 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-12617>CVE-2017-12617</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://www.securitytracker.com/id/1039552">http://www.securitytracker.com/id/1039552</a></p> <p>Fix Resolution: The vendor has issued a fix (7.0.82, 8.0.47, 9.0.1). The vendor advisories are available at: http://tomcat.apache.org/security-7.html#Fixed_in_Apache_Tomcat_7.0.82 http://tomcat.apache.org/security-8.html#Fixed_in_Apache_Tomcat_8.0.47 http://tomcat.apache.org/security-9.html#Fixed_in_Apache_Tomcat_9.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high severity vulnerability detected by whitesource cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation path root repository org apache tomcat embed tomcat embed core tomcat embed core jar library home page a href dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library vulnerability details when running apache tomcat versions to to to and to with http puts enabled e g via setting the readonly initialisation parameter of the default servlet to false it was possible to upload a jsp file to the server via a specially crafted request this jsp could then be requested and any code it contained would be executed by the server publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href fix resolution the vendor has issued a fix the vendor advisories are available at step up your open source security game with whitesource
0
699,059
24,002,684,770
IssuesEvent
2022-09-14 12:43:55
status-im/status-desktop
https://api.github.com/repos/status-im/status-desktop
closed
Chat - Pinned messages - Pinned limit reduces after unpinning
bug priority 2: medium E:Bugfixes
# Bug Report ## Description After a message has been unpinned then the pinned message limit is reduced by 1. ## Steps to reproduce 1. Pin 4 messages (pin limit) 2. 5th pin causes pinned limit reached prompt 3. Unpin 1 message (3 pinned messages) 4. Pin 1 more message - pin limit reached at 3 messages (pin limit reduces each time) #### Expected behavior Pinned message limit remains consistent #### Actual behavior ![image](https://user-images.githubusercontent.com/50769329/186421353-8e586926-e4e8-4694-8390-6d8ea4f166e3.png) ### Additional Information - Status desktop version: https://ci.status.im/job/status-desktop/job/platforms/job/macos/lastSuccessfulBuild/artifact/pkg/StatusIm-Desktop-220824-111243-1c7719.dmg - Operating System: Mac
1.0
Chat - Pinned messages - Pinned limit reduces after unpinning - # Bug Report ## Description After a message has been unpinned then the pinned message limit is reduced by 1. ## Steps to reproduce 1. Pin 4 messages (pin limit) 2. 5th pin causes pinned limit reached prompt 3. Unpin 1 message (3 pinned messages) 4. Pin 1 more message - pin limit reached at 3 messages (pin limit reduces each time) #### Expected behavior Pinned message limit remains consistent #### Actual behavior ![image](https://user-images.githubusercontent.com/50769329/186421353-8e586926-e4e8-4694-8390-6d8ea4f166e3.png) ### Additional Information - Status desktop version: https://ci.status.im/job/status-desktop/job/platforms/job/macos/lastSuccessfulBuild/artifact/pkg/StatusIm-Desktop-220824-111243-1c7719.dmg - Operating System: Mac
priority
chat pinned messages pinned limit reduces after unpinning bug report description after a message has been unpinned then the pinned message limit is reduced by steps to reproduce pin messages pin limit pin causes pinned limit reached prompt unpin message pinned messages pin more message pin limit reached at messages pin limit reduces each time expected behavior pinned message limit remains consistent actual behavior additional information status desktop version operating system mac
1
709,039
24,365,908,811
IssuesEvent
2022-10-03 15:08:53
ramp4-pcar4/story-ramp
https://api.github.com/repos/ramp4-pcar4/story-ramp
closed
Add HTML5 History mode
StoryRAMP Viewer Priority: High
HTML History mode might be something to look into before launch as it removes the `#` in the middle of URLs. https://router.vuejs.org/guide/essentials/history-mode.html _Originally posted by @spencerwahl in https://github.com/ramp4-pcar4/story-ramp/pull/94#pullrequestreview-849668096_
1.0
Add HTML5 History mode - HTML History mode might be something to look into before launch as it removes the `#` in the middle of URLs. https://router.vuejs.org/guide/essentials/history-mode.html _Originally posted by @spencerwahl in https://github.com/ramp4-pcar4/story-ramp/pull/94#pullrequestreview-849668096_
priority
add history mode html history mode might be something to look into before launch as it removes the in the middle of urls originally posted by spencerwahl in
1
627,026
19,850,219,176
IssuesEvent
2022-01-21 11:25:20
kubeflow/kubeflow
https://api.github.com/repos/kubeflow/kubeflow
closed
Notebook culling not working with Istio and RBAC
priority/p2 area/jupyter kind/bug
/kind bug **What steps did you take and what happened:** We deployed Kubeflow with Istio and Istio RBAC (w/ OIDC) enabled and we were looking to use Notebook culling to cull idle notebooks after a period of time. I should also note that we have mTLS enabled on our Istio config. Notebook culling is not functional due to multiple issues: * Notebook controller is not on the service mesh, so fails the mTLS check when talking to notebooks ``` 2020-07-13T21:11:51.792Z INFO culler Error talking to http://construction-starts.ns.svc.cluster.local/notebook/dsd/construction-starts/api/status {"error": "Get http://construction-starts.dsd.svc.cluster.local/notebook/dsd/construction-starts/api/status: read tcp 172.21.XX.YY:36004->10.0.XX.YY:80: read: connection reset by peer"} ``` * If we enable injection on the notebook controller, it is not in the ServiceRoleBinding list, so it fails the Istio RBAC check and gets permission denied We started going down the road of enable istio-injection on the Kubeflow control plane, but this is introduced a bunch of problems. These were the initial steps we tried: 1. Add istio-injection=enabled label to kubeflow namespace 2. Rollout restart all deployments / statefulsets 3. Adding annotation to disable injection on the admission webhook and metadata-envoy-deployment 4. Delete the DestinationRule for the namespace 5. On the svc for ml-pipeline, change http service name to tcp-http Things generally worked (after disabling injection on a couple of services), but there were random errors using Kubeflow and we didn't know of what other impacts to functionality there were. We also didn't like diverging that much from the direction of Kubeflow (https://github.com/kubeflow/manifests/pull/712). **What did you expect to happen:** Notebook culling should function correctly with Istio and Istio RBAC enabled. **Anything else you would like to add:** [Miscellaneous information that will assist in solving the issue.] This was also mentioned in https://github.com/kubeflow/kubeflow/pull/3856#issuecomment-612295807. **Environment:** - Kubeflow version: (version number can be found at the bottom left corner of the Kubeflow dashboard): 1.0.2 - kfctl version: (use `kfctl version`): 1.0.2 - Kubernetes platform: (e.g. `minikube`) AKS - Kubernetes version: (use `kubectl version`): 1.15.7 - OS (e.g. from `/etc/os-release`): Ubuntu 16.04 - Istio: 1.4.9
1.0
Notebook culling not working with Istio and RBAC - /kind bug **What steps did you take and what happened:** We deployed Kubeflow with Istio and Istio RBAC (w/ OIDC) enabled and we were looking to use Notebook culling to cull idle notebooks after a period of time. I should also note that we have mTLS enabled on our Istio config. Notebook culling is not functional due to multiple issues: * Notebook controller is not on the service mesh, so fails the mTLS check when talking to notebooks ``` 2020-07-13T21:11:51.792Z INFO culler Error talking to http://construction-starts.ns.svc.cluster.local/notebook/dsd/construction-starts/api/status {"error": "Get http://construction-starts.dsd.svc.cluster.local/notebook/dsd/construction-starts/api/status: read tcp 172.21.XX.YY:36004->10.0.XX.YY:80: read: connection reset by peer"} ``` * If we enable injection on the notebook controller, it is not in the ServiceRoleBinding list, so it fails the Istio RBAC check and gets permission denied We started going down the road of enable istio-injection on the Kubeflow control plane, but this is introduced a bunch of problems. These were the initial steps we tried: 1. Add istio-injection=enabled label to kubeflow namespace 2. Rollout restart all deployments / statefulsets 3. Adding annotation to disable injection on the admission webhook and metadata-envoy-deployment 4. Delete the DestinationRule for the namespace 5. On the svc for ml-pipeline, change http service name to tcp-http Things generally worked (after disabling injection on a couple of services), but there were random errors using Kubeflow and we didn't know of what other impacts to functionality there were. We also didn't like diverging that much from the direction of Kubeflow (https://github.com/kubeflow/manifests/pull/712). **What did you expect to happen:** Notebook culling should function correctly with Istio and Istio RBAC enabled. **Anything else you would like to add:** [Miscellaneous information that will assist in solving the issue.] This was also mentioned in https://github.com/kubeflow/kubeflow/pull/3856#issuecomment-612295807. **Environment:** - Kubeflow version: (version number can be found at the bottom left corner of the Kubeflow dashboard): 1.0.2 - kfctl version: (use `kfctl version`): 1.0.2 - Kubernetes platform: (e.g. `minikube`) AKS - Kubernetes version: (use `kubectl version`): 1.15.7 - OS (e.g. from `/etc/os-release`): Ubuntu 16.04 - Istio: 1.4.9
priority
notebook culling not working with istio and rbac kind bug what steps did you take and what happened we deployed kubeflow with istio and istio rbac w oidc enabled and we were looking to use notebook culling to cull idle notebooks after a period of time i should also note that we have mtls enabled on our istio config notebook culling is not functional due to multiple issues notebook controller is not on the service mesh so fails the mtls check when talking to notebooks info culler error talking to error get read tcp xx yy xx yy read connection reset by peer if we enable injection on the notebook controller it is not in the servicerolebinding list so it fails the istio rbac check and gets permission denied we started going down the road of enable istio injection on the kubeflow control plane but this is introduced a bunch of problems these were the initial steps we tried add istio injection enabled label to kubeflow namespace rollout restart all deployments statefulsets adding annotation to disable injection on the admission webhook and metadata envoy deployment delete the destinationrule for the namespace on the svc for ml pipeline change http service name to tcp http things generally worked after disabling injection on a couple of services but there were random errors using kubeflow and we didn t know of what other impacts to functionality there were we also didn t like diverging that much from the direction of kubeflow what did you expect to happen notebook culling should function correctly with istio and istio rbac enabled anything else you would like to add this was also mentioned in environment kubeflow version version number can be found at the bottom left corner of the kubeflow dashboard kfctl version use kfctl version kubernetes platform e g minikube aks kubernetes version use kubectl version os e g from etc os release ubuntu istio
1
378,911
26,344,308,124
IssuesEvent
2023-01-10 20:30:51
Anselmoo/spectrafit
https://api.github.com/repos/Anselmoo/spectrafit
opened
[Docs]: Using mike for versioning docs
documentation
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Missing Information in the Docs https://squidfunk.github.io/mkdocs-material/setup/setting-up-versioning/ ### Anything else? _No response_ ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
[Docs]: Using mike for versioning docs - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Missing Information in the Docs https://squidfunk.github.io/mkdocs-material/setup/setting-up-versioning/ ### Anything else? _No response_ ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
non_priority
using mike for versioning docs is there an existing issue for this i have searched the existing issues current missing information in the docs anything else no response code of conduct i agree to follow this project s code of conduct
0
123,109
17,772,169,225
IssuesEvent
2021-08-30 14:49:02
kapseliboi/compiler-explorer
https://api.github.com/repos/kapseliboi/compiler-explorer
opened
CVE-2021-3664 (Medium) detected in url-parse-1.5.1.tgz
security vulnerability
## CVE-2021-3664 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.1.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz</a></p> <p>Path to dependency file: compiler-explorer/package.json</p> <p>Path to vulnerable library: compiler-explorer/node_modules/url-parse/package.json</p> <p> Dependency Hierarchy: - webpack-dev-server-3.11.2.tgz (Root Library) - sockjs-client-1.5.1.tgz - :x: **url-parse-1.5.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/compiler-explorer/commit/e37a5b97c4d98423bce7e1c48beb642adb05b447">e37a5b97c4d98423bce7e1c48beb642adb05b447</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> url-parse is vulnerable to URL Redirection to Untrusted Site <p>Publish Date: 2021-07-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3664>CVE-2021-3664</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664</a></p> <p>Release Date: 2021-07-26</p> <p>Fix Resolution: url-parse - 1.5.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-3664 (Medium) detected in url-parse-1.5.1.tgz - ## CVE-2021-3664 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.1.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz</a></p> <p>Path to dependency file: compiler-explorer/package.json</p> <p>Path to vulnerable library: compiler-explorer/node_modules/url-parse/package.json</p> <p> Dependency Hierarchy: - webpack-dev-server-3.11.2.tgz (Root Library) - sockjs-client-1.5.1.tgz - :x: **url-parse-1.5.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/compiler-explorer/commit/e37a5b97c4d98423bce7e1c48beb642adb05b447">e37a5b97c4d98423bce7e1c48beb642adb05b447</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> url-parse is vulnerable to URL Redirection to Untrusted Site <p>Publish Date: 2021-07-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3664>CVE-2021-3664</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664</a></p> <p>Release Date: 2021-07-26</p> <p>Fix Resolution: url-parse - 1.5.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in url parse tgz cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file compiler explorer package json path to vulnerable library compiler explorer node modules url parse package json dependency hierarchy webpack dev server tgz root library sockjs client tgz x url parse tgz vulnerable library found in head commit a href found in base branch main vulnerability details url parse is vulnerable to url redirection to untrusted site publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse step up your open source security game with whitesource
0