Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
957
labels
stringlengths
4
795
body
stringlengths
1
259k
index
stringclasses
12 values
text_combine
stringlengths
96
259k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
438,071
12,610,637,765
IssuesEvent
2020-06-12 05:37:33
confidantstation/Confidant-Station
https://api.github.com/repos/confidantstation/Confidant-Station
closed
Optimize confidant friends, file transfer data storage structure
Priority: Medium Status: In Progress Type: Enhancement
Currently, a user reserves n friends. The number of file transfers is changed to a global pool
1.0
Optimize confidant friends, file transfer data storage structure - Currently, a user reserves n friends. The number of file transfers is changed to a global pool
priority
optimize confidant friends file transfer data storage structure currently a user reserves n friends the number of file transfers is changed to a global pool
1
554,487
16,430,524,636
IssuesEvent
2021-05-20 00:28:23
MeAmAnUsername/pie
https://api.github.com/repos/MeAmAnUsername/pie
opened
Remove warning on single element in multi-import if element is renamed
Component: editor Priority: medium Status: specified Type: bug
Remove warning on single element in multi-import if element is renamed. Could also add a note suggesting a rewrite if the renaming happens for a single import. Example of how it should work ``` import a:b:{c}:d:{someFunc, someOtherFunc} // warning: single element multi-import (quick fix: remove braces) import a:b:{c as something}:d:{someFunc, someOtherFunc} // no warning, currently there is a warning import a:b:{c as something}:d:someFunc // optional note: could use a:b:c:d:someFunc as something:someFunc ```
1.0
Remove warning on single element in multi-import if element is renamed - Remove warning on single element in multi-import if element is renamed. Could also add a note suggesting a rewrite if the renaming happens for a single import. Example of how it should work ``` import a:b:{c}:d:{someFunc, someOtherFunc} // warning: single element multi-import (quick fix: remove braces) import a:b:{c as something}:d:{someFunc, someOtherFunc} // no warning, currently there is a warning import a:b:{c as something}:d:someFunc // optional note: could use a:b:c:d:someFunc as something:someFunc ```
priority
remove warning on single element in multi import if element is renamed remove warning on single element in multi import if element is renamed could also add a note suggesting a rewrite if the renaming happens for a single import example of how it should work import a b c d somefunc someotherfunc warning single element multi import quick fix remove braces import a b c as something d somefunc someotherfunc no warning currently there is a warning import a b c as something d somefunc optional note could use a b c d somefunc as something somefunc
1
427,511
12,396,183,081
IssuesEvent
2020-05-20 20:03:02
react-figma/react-figma
https://api.github.com/repos/react-figma/react-figma
closed
Create polyfills for functions as ’fetch’ in main thread
complexity: hard priority: medium type: feature or enhancement
Currently things such as `fetch` are undefined in the main execution thread where all components are rendered. We need to polyfill these functions using a bridge between UI and main threads
1.0
Create polyfills for functions as ’fetch’ in main thread - Currently things such as `fetch` are undefined in the main execution thread where all components are rendered. We need to polyfill these functions using a bridge between UI and main threads
priority
create polyfills for functions as ’fetch’ in main thread currently things such as fetch are undefined in the main execution thread where all components are rendered we need to polyfill these functions using a bridge between ui and main threads
1
791,994
27,884,351,923
IssuesEvent
2023-03-21 22:13:25
agrc/electrofishing
https://api.github.com/repos/agrc/electrofishing
closed
Gut Check Metrics Before Submission
waiting medium priority
>When submitting a report create a summary that allows for easy QA/QC using metrics like Condition Factor, Average length, Maximum/Minimum length by species that could serve as red flags. Prior to submitting report the verification of the data includes some measure of condition based on length and weight relationship for fish collected. Ref 54) from https://docs.google.com/document/d/1pTsCfe0-Kmp20vSLy9DJtVIEFB_PwfoabOojmtVQXYM/edit @ToddVolkening: Can you get us specifics on what they want?
1.0
Gut Check Metrics Before Submission - >When submitting a report create a summary that allows for easy QA/QC using metrics like Condition Factor, Average length, Maximum/Minimum length by species that could serve as red flags. Prior to submitting report the verification of the data includes some measure of condition based on length and weight relationship for fish collected. Ref 54) from https://docs.google.com/document/d/1pTsCfe0-Kmp20vSLy9DJtVIEFB_PwfoabOojmtVQXYM/edit @ToddVolkening: Can you get us specifics on what they want?
priority
gut check metrics before submission when submitting a report create a summary that allows for easy qa qc using metrics like condition factor average length maximum minimum length by species that could serve as red flags prior to submitting report the verification of the data includes some measure of condition based on length and weight relationship for fish collected ref from toddvolkening can you get us specifics on what they want
1
57,809
3,083,990,705
IssuesEvent
2015-08-24 12:45:07
StefanIsidorovic/salira
https://api.github.com/repos/StefanIsidorovic/salira
closed
Unary minus
auto-migrated Priority-Medium Type-Other
``` Look at PPJ->ispit->aritmeticki izrazi UNARY MINUS (-1) have to be interpreted like negative number and (n-1) like functor("minus", ...) ``` Original issue reported on code.google.com by `missuchi...@gmail.com` on 7 Jun 2015 at 2:13
1.0
Unary minus - ``` Look at PPJ->ispit->aritmeticki izrazi UNARY MINUS (-1) have to be interpreted like negative number and (n-1) like functor("minus", ...) ``` Original issue reported on code.google.com by `missuchi...@gmail.com` on 7 Jun 2015 at 2:13
priority
unary minus look at ppj ispit aritmeticki izrazi unary minus have to be interpreted like negative number and n like functor minus original issue reported on code google com by missuchi gmail com on jun at
1
339,862
10,263,332,877
IssuesEvent
2019-08-22 14:10:23
dmwm/WMCore
https://api.github.com/repos/dmwm/WMCore
closed
Separate MSTransferor and MSMonitor in different process
Enhancement Medium Priority New Feature ReqMgr2MS
**Impact of the new feature** ReqMgr2MS **Is your feature request related to a problem? Please describe.** Once ReqMgr2MS hit production, it will be under heavy load and we better allocate a different CPU for each of those threads. With the thread/service separation, we also need to take into consideration the log file that each of them is going to write to. **Describe the solution you'd like** Start two different process, one for MSTransferor and one for MSMonitor. **Describe alternatives you've considered** **Additional context** Changes will also be required to the deployment repository.
1.0
Separate MSTransferor and MSMonitor in different process - **Impact of the new feature** ReqMgr2MS **Is your feature request related to a problem? Please describe.** Once ReqMgr2MS hit production, it will be under heavy load and we better allocate a different CPU for each of those threads. With the thread/service separation, we also need to take into consideration the log file that each of them is going to write to. **Describe the solution you'd like** Start two different process, one for MSTransferor and one for MSMonitor. **Describe alternatives you've considered** **Additional context** Changes will also be required to the deployment repository.
priority
separate mstransferor and msmonitor in different process impact of the new feature is your feature request related to a problem please describe once hit production it will be under heavy load and we better allocate a different cpu for each of those threads with the thread service separation we also need to take into consideration the log file that each of them is going to write to describe the solution you d like start two different process one for mstransferor and one for msmonitor describe alternatives you ve considered additional context changes will also be required to the deployment repository
1
517,384
15,008,360,404
IssuesEvent
2021-01-31 09:44:22
bounswe/bounswe2020group9
https://api.github.com/repos/bounswe/bounswe2020group9
closed
iOS - Customer/ Add Review
Estimation - Medium Mobile Priority - High Status - Completed
Implement "adding reviews" feature. As discussed in the meeting, customer shall be able to add review only for the products they have already purchased. Therefore, it should be carried out on the Orders Page. **Deadline: 25.01.2021**
1.0
iOS - Customer/ Add Review - Implement "adding reviews" feature. As discussed in the meeting, customer shall be able to add review only for the products they have already purchased. Therefore, it should be carried out on the Orders Page. **Deadline: 25.01.2021**
priority
ios customer add review implement adding reviews feature as discussed in the meeting customer shall be able to add review only for the products they have already purchased therefore it should be carried out on the orders page deadline
1
642,057
20,866,178,239
IssuesEvent
2022-03-22 07:26:31
AY2122S2-CS2103T-T17-4/tp
https://api.github.com/repos/AY2122S2-CS2103T-T17-4/tp
closed
Add status for each person
type.Story priority.Medium
As an advanced user, I can check the status of a person so that I can focus on contacting people that have not yet been contacted.
1.0
Add status for each person - As an advanced user, I can check the status of a person so that I can focus on contacting people that have not yet been contacted.
priority
add status for each person as an advanced user i can check the status of a person so that i can focus on contacting people that have not yet been contacted
1
246,493
7,895,376,819
IssuesEvent
2018-06-29 02:52:56
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
Make IceT the default when available.
Expected Use: 3 - Occasional Feature Impact: 3 - Medium OS: All Priority: Normal Support Group: Any
Several developers (Mark Miller, Cyrus Harrison, Brad Whitlock, Tom Fogal and Eric Brugger) have been discussing making IceT the default when available. The consensus was that we should. Mark and Tom suggested it would still be nice to have a way to disable it for debugging and other purposes. Mark proposed adding a flag to disable it on the command line as well as control it through the GUI and CLI. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Eric Brugger Original creation: 01/11/2016 10:44 am Original update: 02/18/2016 12:45 pm Ticket number: 2506
1.0
Make IceT the default when available. - Several developers (Mark Miller, Cyrus Harrison, Brad Whitlock, Tom Fogal and Eric Brugger) have been discussing making IceT the default when available. The consensus was that we should. Mark and Tom suggested it would still be nice to have a way to disable it for debugging and other purposes. Mark proposed adding a flag to disable it on the command line as well as control it through the GUI and CLI. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Eric Brugger Original creation: 01/11/2016 10:44 am Original update: 02/18/2016 12:45 pm Ticket number: 2506
priority
make icet the default when available several developers mark miller cyrus harrison brad whitlock tom fogal and eric brugger have been discussing making icet the default when available the consensus was that we should mark and tom suggested it would still be nice to have a way to disable it for debugging and other purposes mark proposed adding a flag to disable it on the command line as well as control it through the gui and cli redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author eric brugger original creation am original update pm ticket number
1
48,013
2,990,117,627
IssuesEvent
2015-07-21 07:02:47
jayway/rest-assured
https://api.github.com/repos/jayway/rest-assured
closed
Lazily merge path arguments
bug imported invalid Priority-Medium
_From [johan.ha...@gmail.com](https://code.google.com/u/105676376875942041029/) on May 02, 2013 08:45:15_ Make this work: ... rootPath("x.y.%s.z").body("w", withArguments("u"), .. ). _Original issue: http://code.google.com/p/rest-assured/issues/detail?id=232_
1.0
Lazily merge path arguments - _From [johan.ha...@gmail.com](https://code.google.com/u/105676376875942041029/) on May 02, 2013 08:45:15_ Make this work: ... rootPath("x.y.%s.z").body("w", withArguments("u"), .. ). _Original issue: http://code.google.com/p/rest-assured/issues/detail?id=232_
priority
lazily merge path arguments from on may make this work rootpath x y s z body w witharguments u original issue
1
525,033
15,228,015,416
IssuesEvent
2021-02-18 10:56:03
ivpn/ios-app
https://api.github.com/repos/ivpn/ios-app
closed
UI issue regarding the circles animation when changing networks
Network Protection priority: medium type: bug
**Description:** In the App Store build 2.0.4, as well as in the latest beta 2.1.0 (23), there is an UI issue with the connecting circle animation while changing networks and having the following network trust settings (Mobile data: Untrusted, WIFI: Trusted or vide versa). When changing networks, it is observed two circles at the same time, one for the disconnected server and another one for the server connected to. **Note:** See attached video for further details. **Actual Result:** Two connecting circles when changing networks while having different trust status set from every network. **Expected Result:** Only one circle animation should be shown when connecting or disconnecting. **Steps to reproduce:** 1. Over WIFI, install App Store version 2.0.4. 2. Login. 3. Select e.g. OpenVPN as protocol and connect to any server. 4. Enable Network Protection. 5. Set Mobile Data to Untrusted and WIFI network to Trusted. 6. Proceed to the main screen. 7. Change network from WIFI to Mobile Data. 8. Observe that two circles are displayed when connecting to the server. 9. Change network from Mobile Data to WIFI. 10. Observe that two circles are displayed when disconnecting from the server. **Extra Info:** * Device: iPhone XR * OS name and version: iOS 14.3 * IVPN app version: App Store 2.0.4, Beta 2.1.0 (23) https://user-images.githubusercontent.com/16647612/105837005-56420100-5fce-11eb-9f62-c9981cf373f6.MP4
1.0
UI issue regarding the circles animation when changing networks - **Description:** In the App Store build 2.0.4, as well as in the latest beta 2.1.0 (23), there is an UI issue with the connecting circle animation while changing networks and having the following network trust settings (Mobile data: Untrusted, WIFI: Trusted or vide versa). When changing networks, it is observed two circles at the same time, one for the disconnected server and another one for the server connected to. **Note:** See attached video for further details. **Actual Result:** Two connecting circles when changing networks while having different trust status set from every network. **Expected Result:** Only one circle animation should be shown when connecting or disconnecting. **Steps to reproduce:** 1. Over WIFI, install App Store version 2.0.4. 2. Login. 3. Select e.g. OpenVPN as protocol and connect to any server. 4. Enable Network Protection. 5. Set Mobile Data to Untrusted and WIFI network to Trusted. 6. Proceed to the main screen. 7. Change network from WIFI to Mobile Data. 8. Observe that two circles are displayed when connecting to the server. 9. Change network from Mobile Data to WIFI. 10. Observe that two circles are displayed when disconnecting from the server. **Extra Info:** * Device: iPhone XR * OS name and version: iOS 14.3 * IVPN app version: App Store 2.0.4, Beta 2.1.0 (23) https://user-images.githubusercontent.com/16647612/105837005-56420100-5fce-11eb-9f62-c9981cf373f6.MP4
priority
ui issue regarding the circles animation when changing networks description in the app store build as well as in the latest beta there is an ui issue with the connecting circle animation while changing networks and having the following network trust settings mobile data untrusted wifi trusted or vide versa when changing networks it is observed two circles at the same time one for the disconnected server and another one for the server connected to note see attached video for further details actual result two connecting circles when changing networks while having different trust status set from every network expected result only one circle animation should be shown when connecting or disconnecting steps to reproduce over wifi install app store version login select e g openvpn as protocol and connect to any server enable network protection set mobile data to untrusted and wifi network to trusted proceed to the main screen change network from wifi to mobile data observe that two circles are displayed when connecting to the server change network from mobile data to wifi observe that two circles are displayed when disconnecting from the server extra info device iphone xr os name and version ios ivpn app version app store beta
1
819,070
30,718,765,989
IssuesEvent
2023-07-27 14:36:13
containrrr/watchtower
https://api.github.com/repos/containrrr/watchtower
opened
ghcr Private Repo Issue
Type: Bug Priority: Medium Status: Available
### Describe the bug I become a 404 Error because the URL what Watchtower tries to reach is unavailable. ### Steps to reproduce 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error ### Expected behavior That WatchTower checks the repo and loads updates. ### Screenshots _No response_ ### Environment - Platform Synology DSM - Architecture x64 - Docker Version 20.10.23 ### Your logs ```text 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="No new images found for /ghcr-io-derrobin154-lurchbot-1" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="No pull needed. Skipping image." 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Found a match" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg=Comparing local="sha256:81c94cc8d71a4763316bd6623ee89e981d37936bdb7cc9eb76485a31bd44ee81" remote="sha256:81c94cc8d71a4763316bd6623ee89e981d37936bdb7cc9eb76485a31bd44ee81" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Found a remote digest to compare with" remote="sha256:81c94cc8d71a4763316bd6623ee89e981d37936bdb7cc9eb76485a31bd44ee81" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Doing a HEAD request to fetch a digest" url="https://ghcr.io/v2/derrobin154/lurchbot/manifests/master" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Parsing image ref" host=ghcr.io image=derrobin154/lurchbot normalized="ghcr.io/derrobin154/lurchbot:master" tag=master 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Credentials found." 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Setting scope for auth token" image=ghcr.io/derrobin154/lurchbot scope="repository:ghcr.io/derrobin154/lurchbot:pull" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Checking challenge header content" realm="https://ghcr.io/token" service=ghcr.io 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Got response to challenge request" header="Bearer realm=\"https://ghcr.io/token\",service=\"ghcr.io\",scope=\"repository:user/image:pull\"" status="401 Unauthorized" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Building challenge URL" URL="https://ghcr.io/v2/" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Checking if pull is needed" container=/ghcr-io-derrobin154-lurchbot-1 image="ghcr.io/derrobin154/lurchbot:master" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Credentials loaded" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Got image name: ghcr.io/derrobin154/lurchbot:master" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Loaded auth credentials for user derrobin154, on registry ghcr.io/derrobin154/lurchbot:master, from file /config.json" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Trying to load authentication credentials." container=/ghcr-io-derrobin154-lurchbot-1 image="ghcr.io/derrobin154/lurchbot:master" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Retrieving running, stopped and exited containers" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Checking containers for updated images" ``` ### Additional context I take the Auth Config from docker login and put it in config.json. I also have the Problem, i download ex. the Watchtower Image from DockerHub over DSM and in Watchtower i become this error: 2023/07/27 16:32:27 | stderr | time="2023-07-27T14:32:27Z" level=debug msg="No credentials for containrrr found" config_file=/config.json 2023/07/27 16:32:27 | stderr | time="2023-07-27T14:32:27Z" level=debug msg="Trying to load authentication credentials." container=/watchtower image="containrrr/watchtower:latest" because the Image name is containerrr/watchtower and in the config.json containerrr is not set because it is from dockerhub.
1.0
ghcr Private Repo Issue - ### Describe the bug I become a 404 Error because the URL what Watchtower tries to reach is unavailable. ### Steps to reproduce 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error ### Expected behavior That WatchTower checks the repo and loads updates. ### Screenshots _No response_ ### Environment - Platform Synology DSM - Architecture x64 - Docker Version 20.10.23 ### Your logs ```text 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="No new images found for /ghcr-io-derrobin154-lurchbot-1" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="No pull needed. Skipping image." 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Found a match" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg=Comparing local="sha256:81c94cc8d71a4763316bd6623ee89e981d37936bdb7cc9eb76485a31bd44ee81" remote="sha256:81c94cc8d71a4763316bd6623ee89e981d37936bdb7cc9eb76485a31bd44ee81" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Found a remote digest to compare with" remote="sha256:81c94cc8d71a4763316bd6623ee89e981d37936bdb7cc9eb76485a31bd44ee81" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Doing a HEAD request to fetch a digest" url="https://ghcr.io/v2/derrobin154/lurchbot/manifests/master" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Parsing image ref" host=ghcr.io image=derrobin154/lurchbot normalized="ghcr.io/derrobin154/lurchbot:master" tag=master 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Credentials found." 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Setting scope for auth token" image=ghcr.io/derrobin154/lurchbot scope="repository:ghcr.io/derrobin154/lurchbot:pull" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Checking challenge header content" realm="https://ghcr.io/token" service=ghcr.io 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Got response to challenge request" header="Bearer realm=\"https://ghcr.io/token\",service=\"ghcr.io\",scope=\"repository:user/image:pull\"" status="401 Unauthorized" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Building challenge URL" URL="https://ghcr.io/v2/" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Checking if pull is needed" container=/ghcr-io-derrobin154-lurchbot-1 image="ghcr.io/derrobin154/lurchbot:master" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Credentials loaded" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Got image name: ghcr.io/derrobin154/lurchbot:master" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Loaded auth credentials for user derrobin154, on registry ghcr.io/derrobin154/lurchbot:master, from file /config.json" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Trying to load authentication credentials." container=/ghcr-io-derrobin154-lurchbot-1 image="ghcr.io/derrobin154/lurchbot:master" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Retrieving running, stopped and exited containers" 2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Checking containers for updated images" ``` ### Additional context I take the Auth Config from docker login and put it in config.json. I also have the Problem, i download ex. the Watchtower Image from DockerHub over DSM and in Watchtower i become this error: 2023/07/27 16:32:27 | stderr | time="2023-07-27T14:32:27Z" level=debug msg="No credentials for containrrr found" config_file=/config.json 2023/07/27 16:32:27 | stderr | time="2023-07-27T14:32:27Z" level=debug msg="Trying to load authentication credentials." container=/watchtower image="containrrr/watchtower:latest" because the Image name is containerrr/watchtower and in the config.json containerrr is not set because it is from dockerhub.
priority
ghcr private repo issue describe the bug i become a error because the url what watchtower tries to reach is unavailable steps to reproduce go to click on scroll down to see error expected behavior that watchtower checks the repo and loads updates screenshots no response environment platform synology dsm architecture docker version your logs text stderr time level debug msg no new images found for ghcr io lurchbot stderr time level debug msg no pull needed skipping image stderr time level debug msg found a match stderr time level debug msg comparing local remote stderr time level debug msg found a remote digest to compare with remote stderr time level debug msg doing a head request to fetch a digest url stderr time level debug msg parsing image ref host ghcr io image lurchbot normalized ghcr io lurchbot master tag master stderr time level debug msg credentials found stderr time level debug msg setting scope for auth token image ghcr io lurchbot scope repository ghcr io lurchbot pull stderr time level debug msg checking challenge header content realm service ghcr io stderr time level debug msg got response to challenge request header bearer realm status unauthorized stderr time level debug msg building challenge url url stderr time level debug msg checking if pull is needed container ghcr io lurchbot image ghcr io lurchbot master stderr time level debug msg credentials loaded stderr time level debug msg got image name ghcr io lurchbot master stderr time level debug msg loaded auth credentials for user on registry ghcr io lurchbot master from file config json stderr time level debug msg trying to load authentication credentials container ghcr io lurchbot image ghcr io lurchbot master stderr time level debug msg retrieving running stopped and exited containers stderr time level debug msg checking containers for updated images additional context i take the auth config from docker login and put it in config json i also have the problem i download ex the watchtower image from dockerhub over dsm and in watchtower i become this error   stderr time  level debug msg no credentials for containrrr found  config file config json   stderr time  level debug msg trying to load authentication credentials  container watchtower image containrrr watchtower latest because the image name is containerrr watchtower and in the config json containerrr is not set because it is from dockerhub
1
69,965
3,316,353,453
IssuesEvent
2015-11-06 16:33:13
TeselaGen/Peony-Issue-Tracking
https://api.github.com/repos/TeselaGen/Peony-Issue-Tracking
opened
Rationalize right-click menu for all library views
Customer: DAS Phase I Priority: Medium Status: In Progress Type: Enhancement
_From @mfero on September 24, 2015 20:37_ We should rationalize the right-click menu across all library views. Menu items specific to a particular library can sit below a separator. My Protocols: Rename, Edit, Delete, Create Copy, Export My Strains: Rename, Edit, Delete, Create Copy, Export My Sequence: Rename, Edit, Delete, Create Copy, Export || Create Parts from Sequences, Create Parts from Sequence Features My Amino Acid Sequence: Rename, Edit, Delete, Create Copy, Export My Parts: Rename, Edit, Delete, Create Copy, Export My Projects: Rename, Edit, Delete, Create Copy, Export _Copied from original issue: TeselaGen/ve#1390_
1.0
Rationalize right-click menu for all library views - _From @mfero on September 24, 2015 20:37_ We should rationalize the right-click menu across all library views. Menu items specific to a particular library can sit below a separator. My Protocols: Rename, Edit, Delete, Create Copy, Export My Strains: Rename, Edit, Delete, Create Copy, Export My Sequence: Rename, Edit, Delete, Create Copy, Export || Create Parts from Sequences, Create Parts from Sequence Features My Amino Acid Sequence: Rename, Edit, Delete, Create Copy, Export My Parts: Rename, Edit, Delete, Create Copy, Export My Projects: Rename, Edit, Delete, Create Copy, Export _Copied from original issue: TeselaGen/ve#1390_
priority
rationalize right click menu for all library views from mfero on september we should rationalize the right click menu across all library views menu items specific to a particular library can sit below a separator my protocols rename edit delete create copy export my strains rename edit delete create copy export my sequence rename edit delete create copy export create parts from sequences create parts from sequence features my amino acid sequence rename edit delete create copy export my parts rename edit delete create copy export my projects rename edit delete create copy export copied from original issue teselagen ve
1
685,290
23,451,453,556
IssuesEvent
2022-08-16 03:31:39
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[YSQL] Inequality is shown to be getting pushed down to docDB when yb_enable_optimizer_statistics is set to true
kind/bug area/ysql priority/medium
Jira Link: [DB-2897](https://yugabyte.atlassian.net/browse/DB-2897) ### Description When `yb_enable_optimizer_statistics` is set to true, an inequality filter on the hash indexed primary key seems to be pushed down to docDB. The query plan shows an Index Scan. A range index scan cannot be performed on the hash index and so this is not correct. ```sql CREATE TABLE test(k int, v int, primary key (k HASH)) SPLIT INTO 10 TABLETS; INSERT INTO test SELECT * FROM generate_series(1,1000); -- yb_enable_optimizer_statistics is false by default SET yb_enable_optimizer_statistics = false; EXPLAIN SELECT * FROM test where k > 500; ``` ``` QUERY PLAN ---------------------------------------------------- Seq Scan on test (cost=0.00..0.00 rows=1 width=8) Filter: (k > 500) ``` ^ This is the correct plan. ```sql SET yb_enable_optimizer_statistics = true; ANALYSE test; EXPLAIN SELECT * FROM test where k > 500; ``` ``` QUERY PLAN ------------------------------------------------------------------------- Index Scan using test_pkey on test (cost=0.00..60.25 rows=500 width=8) Index Cond: (k > 500) ``` ^ After enabling the feature and calling ANALYSE to compute statistics, the query planner proposes an Index Scan which is not possible. However, the results of the query seem correct. The request sent over to docDB also shows that we do not pursue the Index Scan and indeed perform a Sequential Scan.
1.0
[YSQL] Inequality is shown to be getting pushed down to docDB when yb_enable_optimizer_statistics is set to true - Jira Link: [DB-2897](https://yugabyte.atlassian.net/browse/DB-2897) ### Description When `yb_enable_optimizer_statistics` is set to true, an inequality filter on the hash indexed primary key seems to be pushed down to docDB. The query plan shows an Index Scan. A range index scan cannot be performed on the hash index and so this is not correct. ```sql CREATE TABLE test(k int, v int, primary key (k HASH)) SPLIT INTO 10 TABLETS; INSERT INTO test SELECT * FROM generate_series(1,1000); -- yb_enable_optimizer_statistics is false by default SET yb_enable_optimizer_statistics = false; EXPLAIN SELECT * FROM test where k > 500; ``` ``` QUERY PLAN ---------------------------------------------------- Seq Scan on test (cost=0.00..0.00 rows=1 width=8) Filter: (k > 500) ``` ^ This is the correct plan. ```sql SET yb_enable_optimizer_statistics = true; ANALYSE test; EXPLAIN SELECT * FROM test where k > 500; ``` ``` QUERY PLAN ------------------------------------------------------------------------- Index Scan using test_pkey on test (cost=0.00..60.25 rows=500 width=8) Index Cond: (k > 500) ``` ^ After enabling the feature and calling ANALYSE to compute statistics, the query planner proposes an Index Scan which is not possible. However, the results of the query seem correct. The request sent over to docDB also shows that we do not pursue the Index Scan and indeed perform a Sequential Scan.
priority
inequality is shown to be getting pushed down to docdb when yb enable optimizer statistics is set to true jira link description when yb enable optimizer statistics is set to true an inequality filter on the hash indexed primary key seems to be pushed down to docdb the query plan shows an index scan a range index scan cannot be performed on the hash index and so this is not correct sql create table test k int v int primary key k hash split into tablets insert into test select from generate series yb enable optimizer statistics is false by default set yb enable optimizer statistics false explain select from test where k query plan seq scan on test cost rows width filter k this is the correct plan sql set yb enable optimizer statistics true analyse test explain select from test where k query plan index scan using test pkey on test cost rows width index cond k after enabling the feature and calling analyse to compute statistics the query planner proposes an index scan which is not possible however the results of the query seem correct the request sent over to docdb also shows that we do not pursue the index scan and indeed perform a sequential scan
1
213,178
7,246,466,448
IssuesEvent
2018-02-14 21:45:59
Motoxpro/WorldCupStatsSite
https://api.github.com/repos/Motoxpro/WorldCupStatsSite
closed
Need to get track length for world championships
Medium Priority Data Issue
need to download and parse the worlds pdfs for track length
1.0
Need to get track length for world championships - need to download and parse the worlds pdfs for track length
priority
need to get track length for world championships need to download and parse the worlds pdfs for track length
1
160,765
6,102,038,199
IssuesEvent
2017-06-20 15:41:47
OperationCode/operationcode_frontend
https://api.github.com/repos/OperationCode/operationcode_frontend
closed
Add sentry
beginner friendly Priority: Medium Status: Available Type: Feature
<!-- Please fill out one of the sections below based on the type of issue you're creating --> # Feature ## Why is this feature being added? <!-- What problem is it solving? What value does it add? --> Sentry helps track errors. ## What should your feature do? Follow https://sentry.io/for/react/ Our public DN is `https://23e9c0c37a3841b8a692c98b8978f515@sentry.io/147247`
1.0
Add sentry - <!-- Please fill out one of the sections below based on the type of issue you're creating --> # Feature ## Why is this feature being added? <!-- What problem is it solving? What value does it add? --> Sentry helps track errors. ## What should your feature do? Follow https://sentry.io/for/react/ Our public DN is `https://23e9c0c37a3841b8a692c98b8978f515@sentry.io/147247`
priority
add sentry feature why is this feature being added sentry helps track errors what should your feature do follow our public dn is
1
659,086
21,916,230,207
IssuesEvent
2022-05-21 21:42:06
SkriptLang/Skript
https://api.github.com/repos/SkriptLang/Skript
closed
Cant remove saturation effect
bug priority: medium completed
### Skript/Server Version ``` [23:45:09 INFO]: [Skript] Skript's aliases can be found here: https://github.com/SkriptLang/skript-aliases [23:45:09 INFO]: [Skript] Skript's documentation can be found here: https://skriptlang.github.io/Skript [23:45:09 INFO]: [Skript] Server Version: git-Purpur-"4ff0630" (MC: 1.17.1) [23:45:09 INFO]: [Skript] Skript Version: 2.6.1 [23:45:09 INFO]: [Skript] Installed Skript Addons: [23:45:09 INFO]: [Skript] - skript-yaml v1.4 [23:45:09 INFO]: [Skript] - skript-reflect v2.2.3 (https://github.com/TPGamesNL/skript-reflect) [23:45:09 INFO]: [Skript] - Skore v2.0.2-BETA (1.17 only) [23:45:09 INFO]: [Skript] Installed dependencies: None ``` ### Bug Description i can give/apply the saturation without any problems. but I can't remove the effect from the player. it just throws a `Can't understand this condition/effect: remove saturation from player (test.sk, line 2: remove saturation from player')` it should be fixed in https://github.com/SkriptLang/Skript/pull/1142 but its not. (see issue https://github.com/SkriptLang/Skript/issues/692) The Expession below probably conflicts with the effect name. https://github.com/SkriptLang/Skript/blob/d09320866fa15bd6f2fc77b13f3e0cfd2c46ef36/src/main/java/ch/njol/skript/expressions/ExprSaturation.java also the potion effect isnt listed in the aliases https://github.com/SkriptLang/skript-aliases/blob/master/brewing.sk Potion Effect Types: https://skripthub.net/docs/?id=2161 Saturation Expression: https://skripthub.net/docs/?id=876 ### Expected Behavior remove the potion effect from the player ### Steps to Reproduce ``` on rightclick with any pickaxe: set {_p} to player remove saturation from {_p} # gives no errors, but none of them are working remove saturation from player ``` ### Errors or Screenshots ``` [23:44:37 ERROR]: [Skript] Line 20: (Friends.sk) Can't understand this condition/effect: remove saturation from player Line: remove saturation from player ``` ### Other _No response_ ### Agreement - [x] I have read the guidelines above and confirm I am following them with this report.
1.0
Cant remove saturation effect - ### Skript/Server Version ``` [23:45:09 INFO]: [Skript] Skript's aliases can be found here: https://github.com/SkriptLang/skript-aliases [23:45:09 INFO]: [Skript] Skript's documentation can be found here: https://skriptlang.github.io/Skript [23:45:09 INFO]: [Skript] Server Version: git-Purpur-"4ff0630" (MC: 1.17.1) [23:45:09 INFO]: [Skript] Skript Version: 2.6.1 [23:45:09 INFO]: [Skript] Installed Skript Addons: [23:45:09 INFO]: [Skript] - skript-yaml v1.4 [23:45:09 INFO]: [Skript] - skript-reflect v2.2.3 (https://github.com/TPGamesNL/skript-reflect) [23:45:09 INFO]: [Skript] - Skore v2.0.2-BETA (1.17 only) [23:45:09 INFO]: [Skript] Installed dependencies: None ``` ### Bug Description i can give/apply the saturation without any problems. but I can't remove the effect from the player. it just throws a `Can't understand this condition/effect: remove saturation from player (test.sk, line 2: remove saturation from player')` it should be fixed in https://github.com/SkriptLang/Skript/pull/1142 but its not. (see issue https://github.com/SkriptLang/Skript/issues/692) The Expession below probably conflicts with the effect name. https://github.com/SkriptLang/Skript/blob/d09320866fa15bd6f2fc77b13f3e0cfd2c46ef36/src/main/java/ch/njol/skript/expressions/ExprSaturation.java also the potion effect isnt listed in the aliases https://github.com/SkriptLang/skript-aliases/blob/master/brewing.sk Potion Effect Types: https://skripthub.net/docs/?id=2161 Saturation Expression: https://skripthub.net/docs/?id=876 ### Expected Behavior remove the potion effect from the player ### Steps to Reproduce ``` on rightclick with any pickaxe: set {_p} to player remove saturation from {_p} # gives no errors, but none of them are working remove saturation from player ``` ### Errors or Screenshots ``` [23:44:37 ERROR]: [Skript] Line 20: (Friends.sk) Can't understand this condition/effect: remove saturation from player Line: remove saturation from player ``` ### Other _No response_ ### Agreement - [x] I have read the guidelines above and confirm I am following them with this report.
priority
cant remove saturation effect skript server version skript s aliases can be found here skript s documentation can be found here server version git purpur mc skript version installed skript addons skript yaml skript reflect skore beta only installed dependencies none bug description i can give apply the saturation without any problems but i can t remove the effect from the player it just throws a can t understand this condition effect remove saturation from player test sk line remove saturation from player it should be fixed in but its not see issue the expession below probably conflicts with the effect name also the potion effect isnt listed in the aliases potion effect types saturation expression expected behavior remove the potion effect from the player steps to reproduce on rightclick with any pickaxe set p to player remove saturation from p gives no errors but none of them are working remove saturation from player errors or screenshots line friends sk can t understand this condition effect remove saturation from player line remove saturation from player other no response agreement i have read the guidelines above and confirm i am following them with this report
1
749,535
26,166,800,480
IssuesEvent
2023-01-01 11:45:47
docker-mailserver/docker-mailserver
https://api.github.com/repos/docker-mailserver/docker-mailserver
opened
[BUG] postfix: reject_unknown_client_hostname prevents legitimate mail from being received
kind/bug meta/needs triage priority/medium
### Miscellaneous first checks - [X] I checked that all ports are open and not blocked by my ISP / hosting provider. - [X] I know that SSL errors are likely the result of a wrong setup on the user side and not caused by DMS itself. I'm confident my setup is correct. ### Affected Component(s) postfix ### What happened and when does this occur? ```Markdown mailserver | Jan 1 12:35:19 mail postfix/smtpd[1556]: NOQUEUE: reject: RCPT from unknown[54.240.85.218]: 450 4.7.25 Client host rejected: cannot find your hostname, [54.240.85.218]; from=<010101856d1d37d3-cb9c308b-0000-0000-0000-a21593c7994e-000000@bounce.stripe.com> to=<my-localpart+stripe@example.com> proto=ESMTP helo=<a85-218.smtp-out.us-west-2.amazonses.com> ``` ``` ### What did you expect to happen? ```Markdown I expected an option to turn `reject_unknown_client_hostname` off, if it is not considered necessary. I can overwrite the config in the docker-compose, but that feels hacky. ``` ### How do we replicate the issue? ```Markdown I have absolutely no idea. I don't think this happens for all SES mails. ``` ### DMS version latest (d5db200744f1) ### What operating system is DMS running on? Linux ### Which operating system version? Archlinux ### What instruction set architecture is DMS running on? AMD64 / x86_64 ### What container orchestration tool are you using? Docker Compose ### docker-compose.yml ```yml moritz@pan ❯ cat docker-compose.yml docker-mailserver -> master + ? ! $ services: mailserver: image: docker.io/mailserver/docker-mailserver:latest #entrypoint: ["/bin/sh", "-c", "while true; do sleep 10; done"] container_name: mailserver # If the FQDN for your mail-server is only two labels (eg: example.com), # you can assign this entirely to `hostname` and remove `domainname`. hostname: mail domainname: moritz.sh env_file: mailserver.env network_mode: host # More information about the mail-server ports: # https://docker-mailserver.github.io/docker-mailserver/edge/config/security/understanding-the-ports/ # To avoid conflicts with yaml base-60 float, DO NOT remove the quotation marks. ports: - "25:25" # SMTP (explicit TLS => STARTTLS) - "143:143" # IMAP4 (explicit TLS => STARTTLS) - "465:465" # ESMTP (implicit TLS) - "587:587" # ESMTP (explicit TLS => STARTTLS) - "993:993" # IMAP4 (implicit TLS) - "4190:4190" # MANAGE SIEVE volumes: - ./docker-data/dms/mail-data/:/var/mail/ - ./docker-data/dms/mail-state/:/var/mail-state/ - ./docker-data/dms/mail-logs/:/var/log/mail/ - ./config/:/tmp/docker-mailserver/ - ./docker-data/dms/config/dovecot/fts-xapian-plugin.conf:/etc/dovecot/conf.d/10-plugin.conf:ro - /etc/localtime:/etc/localtime:ro - /etc/letsencrypt:/etc/letsencrypt restart: always stop_grace_period: 1m cap_add: - NET_ADMIN healthcheck: test: "ss --listening --tcp | grep -P 'LISTEN.+:smtp' || exit 1" timeout: 3s retries: 0 ``` ### Relevant log output ```bash mailserver | Jan 1 12:35:19 mail postfix/smtpd[1556]: NOQUEUE: reject: RCPT from unknown[54.240.85.218]: 450 4.7.25 Client host rejected: cannot find your hostname, [54.240.85.218]; from=<010101856d1d37d3-cb9c308b-0000-0000-0000-a21593c7994e-000000@bounce.stripe.com> to=<my-localpart+stripe@example.com> proto=ESMTP helo=<a85-218.smtp-out.us-west-2.amazonses.com> ``` ``` ### Other relevant information _No response_ ### What level of experience do you have with Docker and mail servers? - [ ] I am inexperienced with docker - [X] I am rather experienced with docker - [ ] I am inexperienced with mail servers - [ ] I am rather experienced with mail servers - [ ] I am uncomfortable with the CLI - [X] I am rather comfortable with the CLI ### Code of conduct - [X] I have read this project's [Code of Conduct](https://github.com/docker-mailserver/docker-mailserver/blob/master/CODE_OF_CONDUCT.md) and I agree - [X] I have read the [README](https://github.com/docker-mailserver/docker-mailserver/blob/master/README.md) and the [documentation](https://docker-mailserver.github.io/docker-mailserver/edge/) and I searched the [issue tracker](https://github.com/docker-mailserver/docker-mailserver/issues?q=is%3Aissue) but could not find a solution ### Improvements to this form? _No response_
1.0
[BUG] postfix: reject_unknown_client_hostname prevents legitimate mail from being received - ### Miscellaneous first checks - [X] I checked that all ports are open and not blocked by my ISP / hosting provider. - [X] I know that SSL errors are likely the result of a wrong setup on the user side and not caused by DMS itself. I'm confident my setup is correct. ### Affected Component(s) postfix ### What happened and when does this occur? ```Markdown mailserver | Jan 1 12:35:19 mail postfix/smtpd[1556]: NOQUEUE: reject: RCPT from unknown[54.240.85.218]: 450 4.7.25 Client host rejected: cannot find your hostname, [54.240.85.218]; from=<010101856d1d37d3-cb9c308b-0000-0000-0000-a21593c7994e-000000@bounce.stripe.com> to=<my-localpart+stripe@example.com> proto=ESMTP helo=<a85-218.smtp-out.us-west-2.amazonses.com> ``` ``` ### What did you expect to happen? ```Markdown I expected an option to turn `reject_unknown_client_hostname` off, if it is not considered necessary. I can overwrite the config in the docker-compose, but that feels hacky. ``` ### How do we replicate the issue? ```Markdown I have absolutely no idea. I don't think this happens for all SES mails. ``` ### DMS version latest (d5db200744f1) ### What operating system is DMS running on? Linux ### Which operating system version? Archlinux ### What instruction set architecture is DMS running on? AMD64 / x86_64 ### What container orchestration tool are you using? Docker Compose ### docker-compose.yml ```yml moritz@pan ❯ cat docker-compose.yml docker-mailserver -> master + ? ! $ services: mailserver: image: docker.io/mailserver/docker-mailserver:latest #entrypoint: ["/bin/sh", "-c", "while true; do sleep 10; done"] container_name: mailserver # If the FQDN for your mail-server is only two labels (eg: example.com), # you can assign this entirely to `hostname` and remove `domainname`. hostname: mail domainname: moritz.sh env_file: mailserver.env network_mode: host # More information about the mail-server ports: # https://docker-mailserver.github.io/docker-mailserver/edge/config/security/understanding-the-ports/ # To avoid conflicts with yaml base-60 float, DO NOT remove the quotation marks. ports: - "25:25" # SMTP (explicit TLS => STARTTLS) - "143:143" # IMAP4 (explicit TLS => STARTTLS) - "465:465" # ESMTP (implicit TLS) - "587:587" # ESMTP (explicit TLS => STARTTLS) - "993:993" # IMAP4 (implicit TLS) - "4190:4190" # MANAGE SIEVE volumes: - ./docker-data/dms/mail-data/:/var/mail/ - ./docker-data/dms/mail-state/:/var/mail-state/ - ./docker-data/dms/mail-logs/:/var/log/mail/ - ./config/:/tmp/docker-mailserver/ - ./docker-data/dms/config/dovecot/fts-xapian-plugin.conf:/etc/dovecot/conf.d/10-plugin.conf:ro - /etc/localtime:/etc/localtime:ro - /etc/letsencrypt:/etc/letsencrypt restart: always stop_grace_period: 1m cap_add: - NET_ADMIN healthcheck: test: "ss --listening --tcp | grep -P 'LISTEN.+:smtp' || exit 1" timeout: 3s retries: 0 ``` ### Relevant log output ```bash mailserver | Jan 1 12:35:19 mail postfix/smtpd[1556]: NOQUEUE: reject: RCPT from unknown[54.240.85.218]: 450 4.7.25 Client host rejected: cannot find your hostname, [54.240.85.218]; from=<010101856d1d37d3-cb9c308b-0000-0000-0000-a21593c7994e-000000@bounce.stripe.com> to=<my-localpart+stripe@example.com> proto=ESMTP helo=<a85-218.smtp-out.us-west-2.amazonses.com> ``` ``` ### Other relevant information _No response_ ### What level of experience do you have with Docker and mail servers? - [ ] I am inexperienced with docker - [X] I am rather experienced with docker - [ ] I am inexperienced with mail servers - [ ] I am rather experienced with mail servers - [ ] I am uncomfortable with the CLI - [X] I am rather comfortable with the CLI ### Code of conduct - [X] I have read this project's [Code of Conduct](https://github.com/docker-mailserver/docker-mailserver/blob/master/CODE_OF_CONDUCT.md) and I agree - [X] I have read the [README](https://github.com/docker-mailserver/docker-mailserver/blob/master/README.md) and the [documentation](https://docker-mailserver.github.io/docker-mailserver/edge/) and I searched the [issue tracker](https://github.com/docker-mailserver/docker-mailserver/issues?q=is%3Aissue) but could not find a solution ### Improvements to this form? _No response_
priority
postfix reject unknown client hostname prevents legitimate mail from being received miscellaneous first checks i checked that all ports are open and not blocked by my isp hosting provider i know that ssl errors are likely the result of a wrong setup on the user side and not caused by dms itself i m confident my setup is correct affected component s postfix what happened and when does this occur markdown mailserver jan mail postfix smtpd noqueue reject rcpt from unknown client host rejected cannot find your hostname from to proto esmtp helo what did you expect to happen markdown i expected an option to turn reject unknown client hostname off if it is not considered necessary i can overwrite the config in the docker compose but that feels hacky how do we replicate the issue markdown i have absolutely no idea i don t think this happens for all ses mails dms version latest what operating system is dms running on linux which operating system version archlinux what instruction set architecture is dms running on what container orchestration tool are you using docker compose docker compose yml yml moritz pan ❯ cat docker compose yml docker mailserver master services mailserver image docker io mailserver docker mailserver latest entrypoint container name mailserver if the fqdn for your mail server is only two labels eg example com you can assign this entirely to hostname and remove domainname hostname mail domainname moritz sh env file mailserver env network mode host more information about the mail server ports to avoid conflicts with yaml base float do not remove the quotation marks ports smtp explicit tls starttls explicit tls starttls esmtp implicit tls esmtp explicit tls starttls implicit tls manage sieve volumes docker data dms mail data var mail docker data dms mail state var mail state docker data dms mail logs var log mail config tmp docker mailserver docker data dms config dovecot fts xapian plugin conf etc dovecot conf d plugin conf ro etc localtime etc localtime ro etc letsencrypt etc letsencrypt restart always stop grace period cap add net admin healthcheck test ss listening tcp grep p listen smtp exit timeout retries relevant log output bash mailserver jan mail postfix smtpd noqueue reject rcpt from unknown client host rejected cannot find your hostname from to proto esmtp helo other relevant information no response what level of experience do you have with docker and mail servers i am inexperienced with docker i am rather experienced with docker i am inexperienced with mail servers i am rather experienced with mail servers i am uncomfortable with the cli i am rather comfortable with the cli code of conduct i have read this project s and i agree i have read the and the and i searched the but could not find a solution improvements to this form no response
1
617,243
19,345,995,265
IssuesEvent
2021-12-15 10:51:46
google/android-fhir
https://api.github.com/repos/google/android-fhir
closed
Support evaluation of FHIRPath expressions and calculation within Questionnaire
enhancement medium priority Q4 2021
**Is your feature request related to a problem? Please describe.** There is a need to be able to evaluate expressions within a Questionnaire. This is currently not supported by the data-capture library **Describe the solution you'd like** Support for FHIRPath expressions **Describe alternatives you've considered** Other options include CQL. **Additional context** @maimoonak - can you provide some example use cases for the types of expressions you are trying to evaluate
1.0
Support evaluation of FHIRPath expressions and calculation within Questionnaire - **Is your feature request related to a problem? Please describe.** There is a need to be able to evaluate expressions within a Questionnaire. This is currently not supported by the data-capture library **Describe the solution you'd like** Support for FHIRPath expressions **Describe alternatives you've considered** Other options include CQL. **Additional context** @maimoonak - can you provide some example use cases for the types of expressions you are trying to evaluate
priority
support evaluation of fhirpath expressions and calculation within questionnaire is your feature request related to a problem please describe there is a need to be able to evaluate expressions within a questionnaire this is currently not supported by the data capture library describe the solution you d like support for fhirpath expressions describe alternatives you ve considered other options include cql additional context maimoonak can you provide some example use cases for the types of expressions you are trying to evaluate
1
684,158
23,409,453,335
IssuesEvent
2022-08-12 15:53:17
Kong/kubernetes-ingress-controller
https://api.github.com/repos/Kong/kubernetes-ingress-controller
closed
Make Gateway API enabled by default
priority/medium area/gateway-api
### Problem Statement Now that [Gateway API](https://github.com/kubernetes-sigs/gateway-api) has APIs in `v1beta1` with the release of `v0.5.0` we are ready to call our Gateway API implementation beta as well. The purpose of this task is to mark the beta APIs as beta and enable them by default for future KIC releases. Alpha APIs should remain in their current alpha state behind a feature gate that is off. ### Proposed Solution Create a new feature gate called `GatewayAlpha` which implicates support for the alpha apis. Put anything that's pre-beta under that gate, and turn it off by default. Mark `Gateway` feature gate as beta and turn it on by default. Update relevant documentation. ### Acceptance Criteria - [x] beta gateway APIs are supported and on by default in KIC https://github.com/Kong/kubernetes-ingress-controller/pull/2781 - [x] alpha gateway APIs are off by default in KIC https://github.com/Kong/kubernetes-ingress-controller/pull/2781 - [x] documentation is updated to adjust for these changes
1.0
Make Gateway API enabled by default - ### Problem Statement Now that [Gateway API](https://github.com/kubernetes-sigs/gateway-api) has APIs in `v1beta1` with the release of `v0.5.0` we are ready to call our Gateway API implementation beta as well. The purpose of this task is to mark the beta APIs as beta and enable them by default for future KIC releases. Alpha APIs should remain in their current alpha state behind a feature gate that is off. ### Proposed Solution Create a new feature gate called `GatewayAlpha` which implicates support for the alpha apis. Put anything that's pre-beta under that gate, and turn it off by default. Mark `Gateway` feature gate as beta and turn it on by default. Update relevant documentation. ### Acceptance Criteria - [x] beta gateway APIs are supported and on by default in KIC https://github.com/Kong/kubernetes-ingress-controller/pull/2781 - [x] alpha gateway APIs are off by default in KIC https://github.com/Kong/kubernetes-ingress-controller/pull/2781 - [x] documentation is updated to adjust for these changes
priority
make gateway api enabled by default problem statement now that has apis in with the release of we are ready to call our gateway api implementation beta as well the purpose of this task is to mark the beta apis as beta and enable them by default for future kic releases alpha apis should remain in their current alpha state behind a feature gate that is off proposed solution create a new feature gate called gatewayalpha which implicates support for the alpha apis put anything that s pre beta under that gate and turn it off by default mark gateway feature gate as beta and turn it on by default update relevant documentation acceptance criteria beta gateway apis are supported and on by default in kic alpha gateway apis are off by default in kic documentation is updated to adjust for these changes
1
671,620
22,769,284,204
IssuesEvent
2022-07-08 08:27:39
canonical-web-and-design/ubuntu.com
https://api.github.com/repos/canonical-web-and-design/ubuntu.com
closed
Blog cards are missing the coloured strip at the top
Priority: Medium
There is a fallback class for cards that don't have a specific colour assigned that is over riding everything https://github.com/canonical-web-and-design/ubuntu.com/blob/main/static/sass/_pattern_blog-card.scss#L119 ![image.png](https://images.zenhubusercontent.com/609109f935b2ad43af7db68e/a44a2963-93dc-40ff-8c05-b1e04e148aff)
1.0
Blog cards are missing the coloured strip at the top - There is a fallback class for cards that don't have a specific colour assigned that is over riding everything https://github.com/canonical-web-and-design/ubuntu.com/blob/main/static/sass/_pattern_blog-card.scss#L119 ![image.png](https://images.zenhubusercontent.com/609109f935b2ad43af7db68e/a44a2963-93dc-40ff-8c05-b1e04e148aff)
priority
blog cards are missing the coloured strip at the top there is a fallback class for cards that don t have a specific colour assigned that is over riding everything
1
528,102
15,360,136,941
IssuesEvent
2021-03-01 16:36:12
ChainSafe/gossamer
https://api.github.com/repos/ChainSafe/gossamer
closed
update runtime keystore to use GlobalKeystore, use key type IDs in ext_crypto funcs
Priority: 3 - Medium approved wasm
## Task summary <!-- A clear and concise description of what the task is. --> - currently the runtime uses only the `Acco` keystore - it should be using the `GlobalKeystore` instead (see dot/node.go `createRuntime`) - the runtime should then use the correct keystore based on the `keyTypeID` that's passed into ext_crypto_... funcs - the `keyTypeID` corresponds to the keystore's `Name` field ## Specification reference <!-- Provide a reference to the specification as to what is being implemented. --> - spec appendix E ext_crypto funcs ## Other information and links <!-- Add any other context, existing implementation reference or screenshots about the task here. --> - <!-- Thank you 💪 -->
1.0
update runtime keystore to use GlobalKeystore, use key type IDs in ext_crypto funcs - ## Task summary <!-- A clear and concise description of what the task is. --> - currently the runtime uses only the `Acco` keystore - it should be using the `GlobalKeystore` instead (see dot/node.go `createRuntime`) - the runtime should then use the correct keystore based on the `keyTypeID` that's passed into ext_crypto_... funcs - the `keyTypeID` corresponds to the keystore's `Name` field ## Specification reference <!-- Provide a reference to the specification as to what is being implemented. --> - spec appendix E ext_crypto funcs ## Other information and links <!-- Add any other context, existing implementation reference or screenshots about the task here. --> - <!-- Thank you 💪 -->
priority
update runtime keystore to use globalkeystore use key type ids in ext crypto funcs task summary currently the runtime uses only the acco keystore it should be using the globalkeystore instead see dot node go createruntime the runtime should then use the correct keystore based on the keytypeid that s passed into ext crypto funcs the keytypeid corresponds to the keystore s name field specification reference spec appendix e ext crypto funcs other information and links
1
707,528
24,309,123,912
IssuesEvent
2022-09-29 20:19:46
georchestra/georchestra
https://api.github.com/repos/georchestra/georchestra
closed
mapfishapp: produce SLD v1.1.0 compliant documents
feature 0 - Backlog priority-medium
... targeted to WMS >= 1.3.0 servers. Keep current SLD 1.0 service for WMS version <= 1.1.1 servers.
1.0
mapfishapp: produce SLD v1.1.0 compliant documents - ... targeted to WMS >= 1.3.0 servers. Keep current SLD 1.0 service for WMS version <= 1.1.1 servers.
priority
mapfishapp produce sld compliant documents targeted to wms servers keep current sld service for wms version servers
1
509,570
14,739,817,832
IssuesEvent
2021-01-07 07:59:23
konveyor/forklift-ui
https://api.github.com/repos/konveyor/forklift-ui
closed
Non-ready providers should be excluded from any provider selections in forms
medium-priority
If a provider does not have the Ready condition, inventory API requests for resources in that provider will fail. We should either filter out or disable those options in Select fields for providers (and if filtering out, show a message when no providers remain instead of showing an empty dropdown). This might also be a good opportunity to factor out a `ProviderSelect` common component, since we have some duplicated code for getting the available provider options and setting up that Select.
1.0
Non-ready providers should be excluded from any provider selections in forms - If a provider does not have the Ready condition, inventory API requests for resources in that provider will fail. We should either filter out or disable those options in Select fields for providers (and if filtering out, show a message when no providers remain instead of showing an empty dropdown). This might also be a good opportunity to factor out a `ProviderSelect` common component, since we have some duplicated code for getting the available provider options and setting up that Select.
priority
non ready providers should be excluded from any provider selections in forms if a provider does not have the ready condition inventory api requests for resources in that provider will fail we should either filter out or disable those options in select fields for providers and if filtering out show a message when no providers remain instead of showing an empty dropdown this might also be a good opportunity to factor out a providerselect common component since we have some duplicated code for getting the available provider options and setting up that select
1
642,165
20,868,857,197
IssuesEvent
2022-03-22 10:01:23
LiskHQ/lisk-desktop
https://api.github.com/repos/LiskHQ/lisk-desktop
closed
Distorted UI on wallet balance card when in discrete mode
type: bug unplanned priority: medium
### Expected behavior There should be no distorted view on the balance card when discrete mode is toggled on. ### Actual behavior Switching to discrete mode cause a misplacement in the content of the balance card as seen in the screenshot. ![image](https://user-images.githubusercontent.com/32064769/159011123-e82127b7-3150-4e6f-8662-304a292d44f4.png) ### Steps to reproduce - Navigate to the wallets page - Toggle to discrete mode ( It would be observed that the balance card content UI is distorted) ### Which version(s) does this affect? (Environment, OS, etc...) v2.3.0-rc-1
1.0
Distorted UI on wallet balance card when in discrete mode - ### Expected behavior There should be no distorted view on the balance card when discrete mode is toggled on. ### Actual behavior Switching to discrete mode cause a misplacement in the content of the balance card as seen in the screenshot. ![image](https://user-images.githubusercontent.com/32064769/159011123-e82127b7-3150-4e6f-8662-304a292d44f4.png) ### Steps to reproduce - Navigate to the wallets page - Toggle to discrete mode ( It would be observed that the balance card content UI is distorted) ### Which version(s) does this affect? (Environment, OS, etc...) v2.3.0-rc-1
priority
distorted ui on wallet balance card when in discrete mode expected behavior there should be no distorted view on the balance card when discrete mode is toggled on actual behavior switching to discrete mode cause a misplacement in the content of the balance card as seen in the screenshot steps to reproduce navigate to the wallets page toggle to discrete mode it would be observed that the balance card content ui is distorted which version s does this affect environment os etc rc
1
473,705
13,646,320,093
IssuesEvent
2020-09-25 22:50:21
shahednasser/sbuttons
https://api.github.com/repos/shahednasser/sbuttons
opened
Change font for buttons in header
Hacktoberfest Priority: Medium good first issue help wanted up-for-grabs website
Change the font for the "Download" and "Documentation" links in the header to "Syne"
1.0
Change font for buttons in header - Change the font for the "Download" and "Documentation" links in the header to "Syne"
priority
change font for buttons in header change the font for the download and documentation links in the header to syne
1
503,985
14,610,588,559
IssuesEvent
2020-12-22 00:49:16
ableplayer/ableplayer
https://api.github.com/repos/ableplayer/ableplayer
closed
Eliminate options related to description
enhancement - medium priority
User testing with screen reader users has led to some insights related to description: - If description is available as _both_ a separate described version of the video and text-based description (i.e., a WebVTT description track) nearly 100% of users would choose the separate described version. - It is extremely rare for videos to be made available with description at all, and especially rare and unlikely that videos will be made available with description in both formats supported by Able Player. - The Audio Description Preferences dialog is verbose and confusing, and arguably unnecessary given the preceding two items. Therefore, the Preferred Format option should be eliminated, and should always default to "Alternative described version of the video" if both formats are available. With that option gone, the explanatory text in the dialog can be scaled back significantly.
1.0
Eliminate options related to description - User testing with screen reader users has led to some insights related to description: - If description is available as _both_ a separate described version of the video and text-based description (i.e., a WebVTT description track) nearly 100% of users would choose the separate described version. - It is extremely rare for videos to be made available with description at all, and especially rare and unlikely that videos will be made available with description in both formats supported by Able Player. - The Audio Description Preferences dialog is verbose and confusing, and arguably unnecessary given the preceding two items. Therefore, the Preferred Format option should be eliminated, and should always default to "Alternative described version of the video" if both formats are available. With that option gone, the explanatory text in the dialog can be scaled back significantly.
priority
eliminate options related to description user testing with screen reader users has led to some insights related to description if description is available as both a separate described version of the video and text based description i e a webvtt description track nearly of users would choose the separate described version it is extremely rare for videos to be made available with description at all and especially rare and unlikely that videos will be made available with description in both formats supported by able player the audio description preferences dialog is verbose and confusing and arguably unnecessary given the preceding two items therefore the preferred format option should be eliminated and should always default to alternative described version of the video if both formats are available with that option gone the explanatory text in the dialog can be scaled back significantly
1
548,352
16,062,580,992
IssuesEvent
2021-04-23 14:27:49
enso-org/ide
https://api.github.com/repos/enso-org/ide
opened
Strange things happens when putting colon as a separator in list.
Category: Controllers Priority: Medium Type: Bug
<!-- Please ensure that you are using the latest version of Enso IDE before reporting the bug! It may have been fixed since. --> ### What did you do? I put `:` as a separator list by mistake, trying to make node `["x": [4.0]]` ### What did you expect to see? The node with that exact expression, probably with syntax error message ### What did you see instead? An expression without last ] ### Enso Version alpha-3 ### Additional notes Trying to put such expressions in code editor also works strangely.
1.0
Strange things happens when putting colon as a separator in list. - <!-- Please ensure that you are using the latest version of Enso IDE before reporting the bug! It may have been fixed since. --> ### What did you do? I put `:` as a separator list by mistake, trying to make node `["x": [4.0]]` ### What did you expect to see? The node with that exact expression, probably with syntax error message ### What did you see instead? An expression without last ] ### Enso Version alpha-3 ### Additional notes Trying to put such expressions in code editor also works strangely.
priority
strange things happens when putting colon as a separator in list please ensure that you are using the latest version of enso ide before reporting the bug it may have been fixed since what did you do i put as a separator list by mistake trying to make node what did you expect to see the node with that exact expression probably with syntax error message what did you see instead an expression without last enso version alpha additional notes trying to put such expressions in code editor also works strangely
1
738,649
25,570,643,406
IssuesEvent
2022-11-30 17:23:43
uhh-cms/columnflow
https://api.github.com/repos/uhh-cms/columnflow
opened
Allow selection masks being passed to muon / electron weight producers
enhancement medium-priority
As it is right now, the muon and electron weight producers in https://github.com/uhh-cms/columnflow/blob/master/columnflow/production/muon.py and https://github.com/uhh-cms/columnflow/blob/master/columnflow/production/electron.py are very generic, but still they automatically read all leptons available. In some use cases, where the unwanted leptons were already removed in an upstream reduction step, this might work, but in case some (e.g.) veto leptons are kept, we need to be able to select those we want to obtain the weights with. A similar approach has been already followed in the [btag weight producer](https://github.com/uhh-cms/columnflow/blob/16c8a35f601de294a0ac3dd7501d4bf3f2e70080/columnflow/production/btag.py#L26), so adding this to the lepton weights as well should be straight forward.
1.0
Allow selection masks being passed to muon / electron weight producers - As it is right now, the muon and electron weight producers in https://github.com/uhh-cms/columnflow/blob/master/columnflow/production/muon.py and https://github.com/uhh-cms/columnflow/blob/master/columnflow/production/electron.py are very generic, but still they automatically read all leptons available. In some use cases, where the unwanted leptons were already removed in an upstream reduction step, this might work, but in case some (e.g.) veto leptons are kept, we need to be able to select those we want to obtain the weights with. A similar approach has been already followed in the [btag weight producer](https://github.com/uhh-cms/columnflow/blob/16c8a35f601de294a0ac3dd7501d4bf3f2e70080/columnflow/production/btag.py#L26), so adding this to the lepton weights as well should be straight forward.
priority
allow selection masks being passed to muon electron weight producers as it is right now the muon and electron weight producers in and are very generic but still they automatically read all leptons available in some use cases where the unwanted leptons were already removed in an upstream reduction step this might work but in case some e g veto leptons are kept we need to be able to select those we want to obtain the weights with a similar approach has been already followed in the so adding this to the lepton weights as well should be straight forward
1
713,679
24,535,334,769
IssuesEvent
2022-10-11 20:11:32
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
closed
Install Python auto-complete helper files (`*.pyi`)
component: distribution priority: medium
See https://github.com/pybind/pybind11/issues/2350#issuecomment-668879301. As I understand it, some IDEs will not help the user auto-complete their function-call arguments unless we also install `*.pyi` files for our native modules. As part of our build process, we should import `mypy` (maybe `python3-mypy` if it's new enough, otherwise the wheel) and then invoke it to create pyi files, and then install them alongside our modules.
1.0
Install Python auto-complete helper files (`*.pyi`) - See https://github.com/pybind/pybind11/issues/2350#issuecomment-668879301. As I understand it, some IDEs will not help the user auto-complete their function-call arguments unless we also install `*.pyi` files for our native modules. As part of our build process, we should import `mypy` (maybe `python3-mypy` if it's new enough, otherwise the wheel) and then invoke it to create pyi files, and then install them alongside our modules.
priority
install python auto complete helper files pyi see as i understand it some ides will not help the user auto complete their function call arguments unless we also install pyi files for our native modules as part of our build process we should import mypy maybe mypy if it s new enough otherwise the wheel and then invoke it to create pyi files and then install them alongside our modules
1
1,566
2,515,613,855
IssuesEvent
2015-01-15 19:50:18
adobe/brackets
https://api.github.com/repos/adobe/brackets
reopened
LiveDevelopmentMultiBrowser unit test issues
F Live Preview MultiBrowser medium priority
**1)** <del>The Jasmine test-runner window always logs an exception when it is first loaded, before any tests are run:</del> [console messages hidden] <!-- ``` TypeError: undefined is not a function at _showStatusChangeReason (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/main.js:145:21) at statusChange (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/main.js:185:13) at Object.trigger (file:///C:/code/Brackets/brackets-app/brackets/src/utils/EventDispatcher.js:222:40) at _setStatus (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/LiveDevelopment.js:328:17) at Object.init (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/LiveDevelopment.js:1497:9) at file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/main.js:272:25 at _callHandler (file:///C:/code/Brackets/brackets-app/brackets/src/utils/AppInit.js:93:13) at Object._dispatchReady (file:///C:/code/Brackets/brackets-app/brackets/src/utils/AppInit.js:113:13) at _documentReadyHandler (file:///C:/code/Brackets/brackets-app/brackets/test/SpecRunner.js:166:17) ``` --> <del>(because bootstrap-twipsy-mod is not loaded in the Jasmine window)</del> Fixed in #10285 **2)** Several LiveDevelopmentMultiBrowser tests fail for me with "timeout: timed out after 5000 msec waiting for livedevelopment.done.opened" (usually 2-3 tests fail per run). **3)** The console is littered with error messages after running this test suite: ``` [testWindow] [node-error 10:14:11 AM] nodeSocketTransport: Socket closed, but couldn't locate client SpecRunnerUtils.js:543 runs.forEach._testWindow.console.(anonymous function) SpecRunnerUtils.js:543 handleLogEvent src/extensions/default/DebugCommands/NodeDebugUtils.js:114 trigger EventDispatcher.js:222 triggerWithArray EventDispatcher.js:260 NodeConnection._receive NodeConnection.js:479 ``` (twice) ``` [testWindow] Exception in 'base:log' listener on NodeConnection {domains: Object, _registeredModules: Array[0], _pendingInterfaceRefreshDeferreds: Array[0], _pendingCommandDeferreds: Array[0], _autoReconnect: true…} TypeError: Cannot read property 'console' of null at Console.runs.forEach._testWindow.console.(anonymous function) [as log] (file:///C:/code/Brackets/brackets-app/brackets/test/spec/SpecRunnerUtils.js:544:53) at handleLogEvent (file:///C:/code/Brackets/brackets-app/brackets/src/extensions/default/DebugCommands/NodeDebugUtils.js:117:21) at NodeConnection.trigger (file:///C:/code/Brackets/brackets-app/brackets/src/utils/EventDispatcher.js:222:40) at Object.triggerWithArray (file:///C:/code/Brackets/brackets-app/brackets/src/utils/EventDispatcher.js:260:28) at NodeConnection._receive (file:///C:/code/Brackets/brackets-app/brackets/src/utils/NodeConnection.js:479:29) ``` (13 times - various listeners all with the same TypeError) Even if these are caused by the timeouts above, it suggests the code does not fail gracefully in the event of a connection timeout. **4)** One time this test suite seemed to get stuck with the Node process hanging onto file handles in the brackets/test/temp folder, which caused many other unit tests to fail since they were unable to clear that folder. Restating the Node process released the file handles. Depending on how easy it is to hit this failure case, we may want to do more to increase robustness...
1.0
LiveDevelopmentMultiBrowser unit test issues - **1)** <del>The Jasmine test-runner window always logs an exception when it is first loaded, before any tests are run:</del> [console messages hidden] <!-- ``` TypeError: undefined is not a function at _showStatusChangeReason (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/main.js:145:21) at statusChange (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/main.js:185:13) at Object.trigger (file:///C:/code/Brackets/brackets-app/brackets/src/utils/EventDispatcher.js:222:40) at _setStatus (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/LiveDevelopment.js:328:17) at Object.init (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/LiveDevelopment.js:1497:9) at file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/main.js:272:25 at _callHandler (file:///C:/code/Brackets/brackets-app/brackets/src/utils/AppInit.js:93:13) at Object._dispatchReady (file:///C:/code/Brackets/brackets-app/brackets/src/utils/AppInit.js:113:13) at _documentReadyHandler (file:///C:/code/Brackets/brackets-app/brackets/test/SpecRunner.js:166:17) ``` --> <del>(because bootstrap-twipsy-mod is not loaded in the Jasmine window)</del> Fixed in #10285 **2)** Several LiveDevelopmentMultiBrowser tests fail for me with "timeout: timed out after 5000 msec waiting for livedevelopment.done.opened" (usually 2-3 tests fail per run). **3)** The console is littered with error messages after running this test suite: ``` [testWindow] [node-error 10:14:11 AM] nodeSocketTransport: Socket closed, but couldn't locate client SpecRunnerUtils.js:543 runs.forEach._testWindow.console.(anonymous function) SpecRunnerUtils.js:543 handleLogEvent src/extensions/default/DebugCommands/NodeDebugUtils.js:114 trigger EventDispatcher.js:222 triggerWithArray EventDispatcher.js:260 NodeConnection._receive NodeConnection.js:479 ``` (twice) ``` [testWindow] Exception in 'base:log' listener on NodeConnection {domains: Object, _registeredModules: Array[0], _pendingInterfaceRefreshDeferreds: Array[0], _pendingCommandDeferreds: Array[0], _autoReconnect: true…} TypeError: Cannot read property 'console' of null at Console.runs.forEach._testWindow.console.(anonymous function) [as log] (file:///C:/code/Brackets/brackets-app/brackets/test/spec/SpecRunnerUtils.js:544:53) at handleLogEvent (file:///C:/code/Brackets/brackets-app/brackets/src/extensions/default/DebugCommands/NodeDebugUtils.js:117:21) at NodeConnection.trigger (file:///C:/code/Brackets/brackets-app/brackets/src/utils/EventDispatcher.js:222:40) at Object.triggerWithArray (file:///C:/code/Brackets/brackets-app/brackets/src/utils/EventDispatcher.js:260:28) at NodeConnection._receive (file:///C:/code/Brackets/brackets-app/brackets/src/utils/NodeConnection.js:479:29) ``` (13 times - various listeners all with the same TypeError) Even if these are caused by the timeouts above, it suggests the code does not fail gracefully in the event of a connection timeout. **4)** One time this test suite seemed to get stuck with the Node process hanging onto file handles in the brackets/test/temp folder, which caused many other unit tests to fail since they were unable to clear that folder. Restating the Node process released the file handles. Depending on how easy it is to hit this failure case, we may want to do more to increase robustness...
priority
livedevelopmentmultibrowser unit test issues the jasmine test runner window always logs an exception when it is first loaded before any tests are run typeerror undefined is not a function at showstatuschangereason file c code brackets brackets app brackets src livedevelopment main js at statuschange file c code brackets brackets app brackets src livedevelopment main js at object trigger file c code brackets brackets app brackets src utils eventdispatcher js at setstatus file c code brackets brackets app brackets src livedevelopment livedevelopment js at object init file c code brackets brackets app brackets src livedevelopment livedevelopment js at file c code brackets brackets app brackets src livedevelopment main js at callhandler file c code brackets brackets app brackets src utils appinit js at object dispatchready file c code brackets brackets app brackets src utils appinit js at documentreadyhandler file c code brackets brackets app brackets test specrunner js because bootstrap twipsy mod is not loaded in the jasmine window fixed in several livedevelopmentmultibrowser tests fail for me with timeout timed out after msec waiting for livedevelopment done opened usually tests fail per run the console is littered with error messages after running this test suite nodesockettransport socket closed but couldn t locate client specrunnerutils js runs foreach testwindow console anonymous function specrunnerutils js handlelogevent src extensions default debugcommands nodedebugutils js trigger eventdispatcher js triggerwitharray eventdispatcher js nodeconnection receive nodeconnection js twice exception in base log listener on nodeconnection domains object registeredmodules array pendinginterfacerefreshdeferreds array pendingcommanddeferreds array autoreconnect true… typeerror cannot read property console of null at console runs foreach testwindow console anonymous function file c code brackets brackets app brackets test spec specrunnerutils js at handlelogevent file c code brackets brackets app brackets src extensions default debugcommands nodedebugutils js at nodeconnection trigger file c code brackets brackets app brackets src utils eventdispatcher js at object triggerwitharray file c code brackets brackets app brackets src utils eventdispatcher js at nodeconnection receive file c code brackets brackets app brackets src utils nodeconnection js times various listeners all with the same typeerror even if these are caused by the timeouts above it suggests the code does not fail gracefully in the event of a connection timeout one time this test suite seemed to get stuck with the node process hanging onto file handles in the brackets test temp folder which caused many other unit tests to fail since they were unable to clear that folder restating the node process released the file handles depending on how easy it is to hit this failure case we may want to do more to increase robustness
1
321,053
9,792,850,663
IssuesEvent
2019-06-10 18:27:02
dojot/dojot
https://api.github.com/repos/dojot/dojot
closed
[GUI] Template: Change metadata - does not validate fields populated with invalid values
Priority:Medium Status:Treated Team:Frontend Type:Bug
![screenshot_14](https://user-images.githubusercontent.com/37310063/50990900-7cd19400-14fa-11e9-9c7b-0102b8ce12c1.png) **Note:** validation is done by adding the metadata **Affected Version:** 0.3.0-nightly_20181227
1.0
[GUI] Template: Change metadata - does not validate fields populated with invalid values - ![screenshot_14](https://user-images.githubusercontent.com/37310063/50990900-7cd19400-14fa-11e9-9c7b-0102b8ce12c1.png) **Note:** validation is done by adding the metadata **Affected Version:** 0.3.0-nightly_20181227
priority
template change metadata does not validate fields populated with invalid values note validation is done by adding the metadata affected version nightly
1
75,196
3,456,583,548
IssuesEvent
2015-12-18 02:28:41
dkpro/dkpro-core
https://api.github.com/repos/dkpro/dkpro-core
closed
Use current directory as default target location for writers
enhancement Module-api.io Priority-Medium wontfix
``` For writers that write to directories, e.g. Conll2006Writer, we could maybe use the current directory as default output location. ``` Original issue reported on code.google.com by `richard.eckart` on 2014-06-24 17:20:17
1.0
Use current directory as default target location for writers - ``` For writers that write to directories, e.g. Conll2006Writer, we could maybe use the current directory as default output location. ``` Original issue reported on code.google.com by `richard.eckart` on 2014-06-24 17:20:17
priority
use current directory as default target location for writers for writers that write to directories e g we could maybe use the current directory as default output location original issue reported on code google com by richard eckart on
1
1,024
2,506,912,178
IssuesEvent
2015-01-12 14:54:17
ukwa/w3act
https://api.github.com/repos/ukwa/w3act
closed
CSV download from reports pages not working
bug Medium Priority
Hi @kinmanli : an attempt to download CSV from this just hangs up the browser: http://www.webarchive.org.uk/actdev/reportscreation/targets/?s=createdAt&o=desc&crawlFrequency=DAILY&tld=either
1.0
CSV download from reports pages not working - Hi @kinmanli : an attempt to download CSV from this just hangs up the browser: http://www.webarchive.org.uk/actdev/reportscreation/targets/?s=createdAt&o=desc&crawlFrequency=DAILY&tld=either
priority
csv download from reports pages not working hi kinmanli an attempt to download csv from this just hangs up the browser
1
369,603
10,915,304,423
IssuesEvent
2019-11-21 10:53:28
react-figma/react-figma
https://api.github.com/repos/react-figma/react-figma
closed
Different color formats support
complexity: medium priority: medium topic: components topic: primitives support type: feature or enhancement
ERROR: type should be string, got "https://facebook.github.io/react-native/docs/colors\r\n\r\n* 'rgb(255, 0, 255)'\r\n* 'rgba(255, 255, 255, 1.0)'\r\n* '#ff00ff00' \r\n* 'hsl(360, 100%, 100%)'\r\n* 'transparent'\r\n* Named colors: aliceblue, antiquewhite, etc."
1.0
Different color formats support - https://facebook.github.io/react-native/docs/colors * 'rgb(255, 0, 255)' * 'rgba(255, 255, 255, 1.0)' * '#ff00ff00' * 'hsl(360, 100%, 100%)' * 'transparent' * Named colors: aliceblue, antiquewhite, etc.
priority
different color formats support rgb rgba hsl transparent named colors aliceblue antiquewhite etc
1
4,374
2,550,873,794
IssuesEvent
2015-02-02 00:37:23
SO-Close-Vote-Reviewers/SOCVR-Chatbot
https://api.github.com/repos/SO-Close-Vote-Reviewers/SOCVR-Chatbot
closed
Add commands to get tags
enhancement help wanted medium-priority
The following commands need to get added: - current tag - next x tags - start event Details on those commands are int the setup.md file. Most of the work is in the sede branch.
1.0
Add commands to get tags - The following commands need to get added: - current tag - next x tags - start event Details on those commands are int the setup.md file. Most of the work is in the sede branch.
priority
add commands to get tags the following commands need to get added current tag next x tags start event details on those commands are int the setup md file most of the work is in the sede branch
1
826,164
31,559,398,126
IssuesEvent
2023-09-03 03:53:48
ubiquity/ubiquibot
https://api.github.com/repos/ubiquity/ubiquibot
opened
Linked Pull Request Not Closed On Expired Task
Priority: 2 (Medium) Time: <4 Hours
This should be diagnosed and fixed. > @wannacfuture - Releasing the bounty back to dev pool because the allocated duration already ended! Last activity time: Fri Aug 18 2023 23:07:15 GMT+0000 (Coordinated Universal Time) _Originally posted by @ubiquibot in https://github.com/ubiquity/ubiquibot/issues/431#issuecomment-1694037026_
1.0
Linked Pull Request Not Closed On Expired Task - This should be diagnosed and fixed. > @wannacfuture - Releasing the bounty back to dev pool because the allocated duration already ended! Last activity time: Fri Aug 18 2023 23:07:15 GMT+0000 (Coordinated Universal Time) _Originally posted by @ubiquibot in https://github.com/ubiquity/ubiquibot/issues/431#issuecomment-1694037026_
priority
linked pull request not closed on expired task this should be diagnosed and fixed wannacfuture releasing the bounty back to dev pool because the allocated duration already ended last activity time fri aug gmt coordinated universal time originally posted by ubiquibot in
1
733,473
25,307,451,687
IssuesEvent
2022-11-17 15:07:10
Fiserv/Support
https://api.github.com/repos/Fiserv/Support
closed
.docignore Articles Showing in Search
bug CommerceHub Priority - Medium Severity - Medium
# Reporting new issue for Commerce Hub **Region** (if applicable) Dev **Page** https://dev-developer.fiserv.com/search?q=test%20response&p=%5B%22CommerceHub%22%5D **Describe the bug** Search results contain articles that are in the .docignore **Expected behavior** .docignore articles should not show in search results they result in a 404 **Screenshots** ![image](https://user-images.githubusercontent.com/75692745/200655979-88ff12c9-a5a6-4648-bcb4-b0fe9fbe7f57.png) ![image](https://user-images.githubusercontent.com/75692745/200656057-fb87787f-c3a2-4c90-bd3a-8ba028b2950d.png)
1.0
.docignore Articles Showing in Search - # Reporting new issue for Commerce Hub **Region** (if applicable) Dev **Page** https://dev-developer.fiserv.com/search?q=test%20response&p=%5B%22CommerceHub%22%5D **Describe the bug** Search results contain articles that are in the .docignore **Expected behavior** .docignore articles should not show in search results they result in a 404 **Screenshots** ![image](https://user-images.githubusercontent.com/75692745/200655979-88ff12c9-a5a6-4648-bcb4-b0fe9fbe7f57.png) ![image](https://user-images.githubusercontent.com/75692745/200656057-fb87787f-c3a2-4c90-bd3a-8ba028b2950d.png)
priority
docignore articles showing in search reporting new issue for commerce hub region if applicable dev page describe the bug search results contain articles that are in the docignore expected behavior docignore articles should not show in search results they result in a screenshots
1
507,013
14,678,259,527
IssuesEvent
2020-12-31 02:36:33
rubyforgood/casa
https://api.github.com/repos/rubyforgood/casa
closed
Automate github issue weekly "is this still in progress?" checkins
:sparkles: :computer: Contributor Friendly / Devel Help Wanted Priority: Medium
Automate github issue weekly "is this still in progress?" checkins on "in progress" github issues without recent comments / commits. After two weeks (14 days) of checking-in and no comments or PRs, unassign the assignee and move the issue back to "to do" (automate as many parts of this as possible! Anything is better than nothing :) ) This is (hopefully) possible with github actions
1.0
Automate github issue weekly "is this still in progress?" checkins - Automate github issue weekly "is this still in progress?" checkins on "in progress" github issues without recent comments / commits. After two weeks (14 days) of checking-in and no comments or PRs, unassign the assignee and move the issue back to "to do" (automate as many parts of this as possible! Anything is better than nothing :) ) This is (hopefully) possible with github actions
priority
automate github issue weekly is this still in progress checkins automate github issue weekly is this still in progress checkins on in progress github issues without recent comments commits after two weeks days of checking in and no comments or prs unassign the assignee and move the issue back to to do automate as many parts of this as possible anything is better than nothing this is hopefully possible with github actions
1
55,593
3,073,806,877
IssuesEvent
2015-08-20 00:42:14
RobotiumTech/robotium
https://api.github.com/repos/RobotiumTech/robotium
closed
setDatePicker set wrong moth.
bug imported invalid Priority-Medium
_From [zoumy...@gmail.com](https://code.google.com/u/115552050981841755101/) on August 07, 2013 01:12:46_ What steps will reproduce the problem? 1.open DatePicker. 2.set Date 1986,7,21 but the date has been set Aug 21 1986 3. What is the expected output? What do you see instead? Jul 21 1986 What version of the product are you using? On what operating system? version :4.2.2 emulator Please provide any additional information below. _Original issue: http://code.google.com/p/robotium/issues/detail?id=496_
1.0
setDatePicker set wrong moth. - _From [zoumy...@gmail.com](https://code.google.com/u/115552050981841755101/) on August 07, 2013 01:12:46_ What steps will reproduce the problem? 1.open DatePicker. 2.set Date 1986,7,21 but the date has been set Aug 21 1986 3. What is the expected output? What do you see instead? Jul 21 1986 What version of the product are you using? On what operating system? version :4.2.2 emulator Please provide any additional information below. _Original issue: http://code.google.com/p/robotium/issues/detail?id=496_
priority
setdatepicker set wrong moth from on august what steps will reproduce the problem open datepicker set date but the date has been set aug what is the expected output what do you see instead jul what version of the product are you using on what operating system version emulator please provide any additional information below original issue
1
349,853
10,474,561,773
IssuesEvent
2019-09-23 14:43:32
LifeMC/LifeSkript
https://api.github.com/repos/LifeMC/LifeSkript
opened
Method 'loadScript' is too complex to analyze by data flow algorithm
priority: longtime goal priority: medium state: help wanted type: enhancement
**Describe the bug** IntelliJ gives warning "Method 'loadScript' is too complex to analyze by data flow algorithm" in the ScriptLoader#loadScript method. **To Reproduce** Open ScriptLoader and navigate to loadScript method in IntelliJ (tested on IntelliJ IDEA Ultimate Edition 2019.2) **Expected behavior** Expected no warning. Method should be splitted to other methods to reduce its code complexity. **Screenshots** N/A **Version information:** - Minecraft Version: N/A - Skript Version: 2.2.18 Development Build (master) **Additional information** N/A
2.0
Method 'loadScript' is too complex to analyze by data flow algorithm - **Describe the bug** IntelliJ gives warning "Method 'loadScript' is too complex to analyze by data flow algorithm" in the ScriptLoader#loadScript method. **To Reproduce** Open ScriptLoader and navigate to loadScript method in IntelliJ (tested on IntelliJ IDEA Ultimate Edition 2019.2) **Expected behavior** Expected no warning. Method should be splitted to other methods to reduce its code complexity. **Screenshots** N/A **Version information:** - Minecraft Version: N/A - Skript Version: 2.2.18 Development Build (master) **Additional information** N/A
priority
method loadscript is too complex to analyze by data flow algorithm describe the bug intellij gives warning method loadscript is too complex to analyze by data flow algorithm in the scriptloader loadscript method to reproduce open scriptloader and navigate to loadscript method in intellij tested on intellij idea ultimate edition expected behavior expected no warning method should be splitted to other methods to reduce its code complexity screenshots n a version information minecraft version n a skript version development build master additional information n a
1
789,765
27,804,979,962
IssuesEvent
2023-03-17 18:59:40
knative/docs
https://api.github.com/repos/knative/docs
closed
Fix tab formatting in code-samples folder
kind/bug kind/good-first-issue priority/medium kind/cleanup hacktoberfest good first issue help wanted
## Background The files in `code-samples` folder are no longer viewable on the website, only in GitHub. However, this has caused broken formatting because tabbed elements are not available in GitHub markdown. ## Task Remove the tabbed formatting, for example `=== "yaml"`, in the files in the `code-samples` folder. Make sure that if the tabbed formatting is under a step, you indent your new formatting within each step. Files that have this issue: - [ ] [./code-samples/serving/cloudevents/cloudevents-nodejs/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-nodejs/README.md) - [ ] [./code-samples/serving/cloudevents/cloudevents-rust/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-rust/README.md) - [ ] [./code-samples/serving/cloudevents/cloudevents-go/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-go/README.md) - [ ] [./code-samples/serving/cloudevents/cloudevents-vertx/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-vertx/README.md) - [ ] [./code-samples/serving/cloudevents/cloudevents-spring/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-spring/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-ruby/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-ruby/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-shell/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-shell/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-java-spring/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-java-spring/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-python/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-python/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-java-spark/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-java-spark/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-go/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-go/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-php/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-php/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-scala/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-scala/README.md)
1.0
Fix tab formatting in code-samples folder - ## Background The files in `code-samples` folder are no longer viewable on the website, only in GitHub. However, this has caused broken formatting because tabbed elements are not available in GitHub markdown. ## Task Remove the tabbed formatting, for example `=== "yaml"`, in the files in the `code-samples` folder. Make sure that if the tabbed formatting is under a step, you indent your new formatting within each step. Files that have this issue: - [ ] [./code-samples/serving/cloudevents/cloudevents-nodejs/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-nodejs/README.md) - [ ] [./code-samples/serving/cloudevents/cloudevents-rust/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-rust/README.md) - [ ] [./code-samples/serving/cloudevents/cloudevents-go/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-go/README.md) - [ ] [./code-samples/serving/cloudevents/cloudevents-vertx/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-vertx/README.md) - [ ] [./code-samples/serving/cloudevents/cloudevents-spring/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-spring/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-ruby/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-ruby/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-shell/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-shell/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-java-spring/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-java-spring/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-python/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-python/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-java-spark/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-java-spark/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-go/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-go/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-php/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-php/README.md) - [ ] [./code-samples/serving/hello-world/helloworld-scala/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-scala/README.md)
priority
fix tab formatting in code samples folder background the files in code samples folder are no longer viewable on the website only in github however this has caused broken formatting because tabbed elements are not available in github markdown task remove the tabbed formatting for example yaml in the files in the code samples folder make sure that if the tabbed formatting is under a step you indent your new formatting within each step files that have this issue
1
163,186
6,192,717,545
IssuesEvent
2017-07-05 03:28:44
start-jsk/jsk_apc
https://api.github.com/repos/start-jsk/jsk_apc
opened
Make additional fingers, finger bases and gears
enhancement issue/priority/medium
- [ ] Print additional fingers, finger bases and gears (fix magnet fitting) - [ ] Mold rubber to fingers - [ ] Print additional other parts
1.0
Make additional fingers, finger bases and gears - - [ ] Print additional fingers, finger bases and gears (fix magnet fitting) - [ ] Mold rubber to fingers - [ ] Print additional other parts
priority
make additional fingers finger bases and gears print additional fingers finger bases and gears fix magnet fitting mold rubber to fingers print additional other parts
1
350,723
10,500,868,163
IssuesEvent
2019-09-26 11:31:23
code4romania/monitorizare-vot-android
https://api.github.com/repos/code4romania/monitorizare-vot-android
closed
[Research] Investigate replacing realm with room
android enhancement help wanted medium priority research
We are planning a complete redo of the app, using kotlin. Please research the possibility of replacing Realm db with [Room](https://developer.android.com/topic/libraries/architecture/room).
1.0
[Research] Investigate replacing realm with room - We are planning a complete redo of the app, using kotlin. Please research the possibility of replacing Realm db with [Room](https://developer.android.com/topic/libraries/architecture/room).
priority
investigate replacing realm with room we are planning a complete redo of the app using kotlin please research the possibility of replacing realm db with
1
67,714
3,277,566,036
IssuesEvent
2015-10-27 01:40:51
saxifrage/caac-map
https://api.github.com/repos/saxifrage/caac-map
closed
Hover-over highlighting does not work for pathway colors
Medium Priority
The darker colors of the pathway blocks don't allow hover-overs to do anything. Seems like if we shift the color scheme at large to echo Chelsea's design, then the really dark blue could be used for all hover-overs. The ultra-light blue is the non-highlighted color, and the middle blue is the default color. ![screen shot 2015-09-04 at 11 41 13 am](https://cloud.githubusercontent.com/assets/1823300/9735216/2ee544c2-5607-11e5-93e3-9ab227ed479a.png)
1.0
Hover-over highlighting does not work for pathway colors - The darker colors of the pathway blocks don't allow hover-overs to do anything. Seems like if we shift the color scheme at large to echo Chelsea's design, then the really dark blue could be used for all hover-overs. The ultra-light blue is the non-highlighted color, and the middle blue is the default color. ![screen shot 2015-09-04 at 11 41 13 am](https://cloud.githubusercontent.com/assets/1823300/9735216/2ee544c2-5607-11e5-93e3-9ab227ed479a.png)
priority
hover over highlighting does not work for pathway colors the darker colors of the pathway blocks don t allow hover overs to do anything seems like if we shift the color scheme at large to echo chelsea s design then the really dark blue could be used for all hover overs the ultra light blue is the non highlighted color and the middle blue is the default color
1
30,550
2,724,021,608
IssuesEvent
2015-04-14 15:38:37
CruxFramework/crux-widgets
https://api.github.com/repos/CruxFramework/crux-widgets
closed
Create a method on JsWindow class to get the parent window
Core enhancement imported Milestone-4.0.0 Priority-Medium
_From [ge...@cruxframework.org](https://code.google.com/u/108728025643241132101/) on May 20, 2011 18:51:43_ Create a method on JsWindow class to get the parent window _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=249_
1.0
Create a method on JsWindow class to get the parent window - _From [ge...@cruxframework.org](https://code.google.com/u/108728025643241132101/) on May 20, 2011 18:51:43_ Create a method on JsWindow class to get the parent window _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=249_
priority
create a method on jswindow class to get the parent window from on may create a method on jswindow class to get the parent window original issue
1
160,956
6,106,076,560
IssuesEvent
2017-06-21 02:28:04
minio/minio-go
https://api.github.com/repos/minio/minio-go
closed
TestPutObjectStream is not passing with latest and master minio
priority: medium
--- FAIL: TestPutObjectStreaming (0.00s) api_functional_v4_test.go:406: Test 1 Error: The request signature we calculated does not match the signature you provided. Check your key and signing method. minio-go- testh1owqaacnr26h95qc test-object
1.0
TestPutObjectStream is not passing with latest and master minio - --- FAIL: TestPutObjectStreaming (0.00s) api_functional_v4_test.go:406: Test 1 Error: The request signature we calculated does not match the signature you provided. Check your key and signing method. minio-go- testh1owqaacnr26h95qc test-object
priority
testputobjectstream is not passing with latest and master minio fail testputobjectstreaming api functional test go test error the request signature we calculated does not match the signature you provided check your key and signing method minio go test object
1
522,088
15,148,714,826
IssuesEvent
2021-02-11 11:00:55
FraunhoferISST/IDS-Connector-Framework
https://api.github.com/repos/FraunhoferISST/IDS-Connector-Framework
opened
Forward received RejectionMessage to Connector
Priority: Medium Type: Enhancement
Currently a received RejectionMessage is not forwarded to the connector-developer upon receiving it at the endpoints of the IDS-Framework due to failed DAT-validation of the received RejectionMessage within the IDS-Framework. The received RejectionMessage should be forwarded to the connector developer so that the developer can evaluate it and the information contained like the correlationMessage, regardless of the status of the DAT-validation of the received message within the IDS-Framework.
1.0
Forward received RejectionMessage to Connector - Currently a received RejectionMessage is not forwarded to the connector-developer upon receiving it at the endpoints of the IDS-Framework due to failed DAT-validation of the received RejectionMessage within the IDS-Framework. The received RejectionMessage should be forwarded to the connector developer so that the developer can evaluate it and the information contained like the correlationMessage, regardless of the status of the DAT-validation of the received message within the IDS-Framework.
priority
forward received rejectionmessage to connector currently a received rejectionmessage is not forwarded to the connector developer upon receiving it at the endpoints of the ids framework due to failed dat validation of the received rejectionmessage within the ids framework the received rejectionmessage should be forwarded to the connector developer so that the developer can evaluate it and the information contained like the correlationmessage regardless of the status of the dat validation of the received message within the ids framework
1
710,668
24,427,238,851
IssuesEvent
2022-10-06 04:37:49
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
Renovate is stuck in an infinite loop overwriting a ci bot's commits
type:bug priority-3-medium status:in-progress regression
### How are you running Renovate? Mend Renovate hosted app on github.com ### If you're self-hosting Renovate, tell us what version of Renovate you run. _No response_ ### If you're self-hosting Renovate, select which platform you are using. _No response_ ### If you're self-hosting Renovate, tell us what version of the platform you run. _No response_ ### Was this something which used to work for you, and then stopped? It used to work, and then stopped ### Describe the bug Renovate is stuck in an infinite loop overwriting a ci bot's fixup commits. This started happening sometime in the past 2 months (the configuration has not been changed since). So far I have only observed this with python packages, but I do not have a similar procedure for other dependencies. The process is something like this: 1. Renovate updates package `x` and commits to the repository, which triggers the first ci run. 2. The repository uses lock files with `pip-compile`, so before running builds and tests, the ci agent runs a script to update the lockfiles. This takes care of resolving dependency changes of `x`. If any changes were made, the ci bot commits them as a fixup commit (`git commit --fixup HEAD`) and pushes it to GitHub. 3. The new commit triggers a second ci run. This time, the check should pass and move on. The ci bot is registered as a `gitIgnoredAuthor`, but in the past renovate would only overwrite the changes if the `main` branch was updated (I have `rebaseWhen` as `behind-base-branch`). Currently, every time the ci bot commits a change renovate seems to overwrite it the next time it runs. This means the repository update process is effectively stuck in an infinite loop. I'm not sure exactly what triggers this behavior and it's been a while since it worked, but it seems to have been around the same time renovate changed email names from `bot@renovateapp.com` to `29139614+renovate[bot]@users.noreply.github.com`. I recall having to manually rebase the existing branches as they were marked as modified. I've provided debug logs from a minimal reproduction repository, which is publicly available on [GitHub](https://github.com/corypaik/renovate-loop). The repository contains one example package that causes this issue, all others have been removed or disabled. There is also a public [Buildkite](https://buildkite.com/corypaik/renovate-loop) available for reference. ### Relevant debug logs <details><summary>Logs</summary> ``` DEBUG: No dangling containers to remove INFO: Repository started { "renovateVersion": "32.184.2" } DEBUG: Using localDir: /mnt/renovate/gh/corypaik/renovate-loop DEBUG: PackageFiles.clear() - Package files deleted { "baseBranches": [] } DEBUG: resetMemCache() DEBUG: initRepo("corypaik/renovate-loop") DEBUG: corypaik/renovate-loop default branch = main DEBUG: Using app token for git init DEBUG: Repository cache is restored from revision 13 DEBUG: Resetting npmrc DEBUG: detectSemanticCommits() DEBUG: checkOnboarding() DEBUG: isOnboarded() DEBUG: Checking cached config file name DEBUG: Existing config file confirmed DEBUG: Repository config { "fileName": "renovate.json", "config": { "$schema": "https://docs.renovatebot.com/renovate-schema.json", "extends": [ ":ignoreModulesAndTests", "group:monorepos", "group:recommended", "workarounds:all" ], "semanticCommits": "enabled", "semanticCommitType": "build", "semanticCommitScope": "", "prHourlyLimit": 2, "prConcurrentLimit": 10, "dependencyDashboard": true, "labels": [ "target: minor", "comp: build & ci", "action: review" ], "timezone": "America/Chihuahua", "schedule": [ "after 10pm every weekday", "before 4am every weekday" ], "baseBranches": [ "main" ], "rebaseWhen": "behind-base-branch", "pip_requirements": { "fileMatch": [ ".*requirements.in$" ] }, "prCreation": "not-pending", "stabilityDays": 3, "gitIgnoredAuthors": [ "buildkite-bot@buildkite.com" ], "pre-commit": { "enabled": true }, "packageRules": [ { "description": "Always pin direct python dependencies to ensure lockfile updates do not revert to a different version.", "matchFiles": [ "third_party/requirements.in" ], "rangeStrategy": "pin" }, { "description": "All python deps, except datasets (for minimal reproduction).", "matchPackagePatterns": [ "*" ], "matchUpdateTypes": [ "major", "minor", "patch" ], "matchLanguages": [ "python" ], "excludePackageNames": [ "datasets" ], "enabled": false } ] } } DEBUG: Repo is onboarded DEBUG: migrateAndValidate() DEBUG: No config migration necessary DEBUG: massaged config { "config": { "$schema": "https://docs.renovatebot.com/renovate-schema.json", "extends": [ "github>whitesource/merge-confidence:beta", ":ignoreModulesAndTests", "group:monorepos", "group:recommended", "workarounds:all" ], "semanticCommits": "enabled", "semanticCommitType": "build", "semanticCommitScope": "", "prHourlyLimit": 2, "prConcurrentLimit": 10, "dependencyDashboard": true, "labels": [ "target: minor", "comp: build & ci", "action: review" ], "timezone": "America/Chihuahua", "schedule": [ "after 10pm every weekday", "before 4am every weekday" ], "baseBranches": [ "main" ], "rebaseWhen": "behind-base-branch", "pip_requirements": { "fileMatch": [ ".*requirements.in$" ] }, "prCreation": "not-pending", "stabilityDays": 3, "gitIgnoredAuthors": [ "buildkite-bot@buildkite.com" ], "pre-commit": { "enabled": true }, "packageRules": [ { "description": [ "Always pin direct python dependencies to ensure lockfile updates do not revert to a different version." ], "matchFiles": [ "third_party/requirements.in" ], "rangeStrategy": "pin" }, { "description": [ "All python deps, except datasets (for minimal reproduction)." ], "matchPackagePatterns": [ "*" ], "matchUpdateTypes": [ "major", "minor", "patch" ], "matchLanguages": [ "python" ], "excludePackageNames": [ "datasets" ], "enabled": false } ] } } DEBUG: migrated config { "config": { "$schema": "https://docs.renovatebot.com/renovate-schema.json", "extends": [ "github>whitesource/merge-confidence:beta", ":ignoreModulesAndTests", "group:monorepos", "group:recommended", "workarounds:all" ], "semanticCommits": "enabled", "semanticCommitType": "build", "semanticCommitScope": "", "prHourlyLimit": 2, "prConcurrentLimit": 10, "dependencyDashboard": true, "labels": [ "target: minor", "comp: build & ci", "action: review" ], "timezone": "America/Chihuahua", "schedule": [ "after 10pm every weekday", "before 4am every weekday" ], "baseBranches": [ "main" ], "rebaseWhen": "behind-base-branch", "pip_requirements": { "fileMatch": [ ".*requirements.in$" ] }, "prCreation": "not-pending", "stabilityDays": 3, "gitIgnoredAuthors": [ "buildkite-bot@buildkite.com" ], "pre-commit": { "enabled": true }, "packageRules": [ { "description": [ "Always pin direct python dependencies to ensure lockfile updates do not revert to a different version." ], "matchFiles": [ "third_party/requirements.in" ], "rangeStrategy": "pin" }, { "description": [ "All python deps, except datasets (for minimal reproduction)." ], "matchPackagePatterns": [ "*" ], "matchUpdateTypes": [ "major", "minor", "patch" ], "matchLanguages": [ "python" ], "excludePackageNames": [ "datasets" ], "enabled": false } ] } } DEBUG: Setting hostRules from config DEBUG: Found repo ignorePaths { "ignorePaths": [ "**/node_modules/**", "**/bower_components/**", "**/vendor/**", "**/examples/**", "**/__tests__/**", "**/test/**", "**/tests/**", "**/__fixtures__/**" ] } DEBUG: No vulnerability alerts found DEBUG: No vulnerability alerts found DEBUG: findIssue(Dependency Dashboard) DEBUG: Retrieving issueList DEBUG: Retrieved 1 issues DEBUG: Found issue 1 DEBUG: baseBranches { "baseBranches": [ "main" ] } DEBUG: baseBranch: main { "baseBranch": "main" } DEBUG: extract() { "baseBranch": "main" } DEBUG: Found cached extract { "baseBranch": "main", "baseBranchSha": "5267848cc1c6b6dfb1d031f573b25d2ab36ebc82" } DEBUG: Deleted cached dep updates { "baseBranch": "main" } INFO: Dependency extraction complete { "baseBranch": "main", "stats": { "managers": { "bazel": { "fileCount": 1, "depCount": 1 }, "pip_requirements": { "fileCount": 2, "depCount": 24 } }, "total": { "fileCount": 3, "depCount": 25 } } } DEBUG: baseBranch: main { "baseBranch": "main" } DEBUG: PackageFiles.add() - Package file saved for branch { "baseBranch": "main" } DEBUG: Package releases lookups complete { "baseBranch": "main" } DEBUG: branchifyUpgrades { "baseBranch": "main" } DEBUG: 4 flattened updates found: datasets, datasets, datasets, datasets { "baseBranch": "main" } DEBUG: Returning 2 branch(es) { "baseBranch": "main" } DEBUG: config.repoIsOnboarded=true { "baseBranch": "main" } DEBUG: packageFiles with updates { "baseBranch": "main", "config": { "bazel": [ { "packageFile": "WORKSPACE", "deps": [ { "datasource": "github-tags", "depType": "http_archive", "depName": "rules_python", "packageName": "bazelbuild/rules_python", "currentValue": "refs", "managerData": { "def": "http_archive(\n name = \"rules_python\",\n sha256 = \"b593d13bb43c94ce94b483c2858e53a9b811f6f10e1e0eedc61073bd90e58d9c\",\n strip_prefix = \"rules_python-0.12.0\",\n url = \"https://github.com/bazelbuild/rules_python/archive/refs/tags/0.12.0.tar.gz\",\n)" }, "depIndex": 0, "warnings": [], "versioning": "semver", "skipReason": "invalid-value", "updates": [] } ] } ], "pip_requirements": [ { "packageFile": "third_party/requirements.in", "deps": [ { "depName": "datasets", "currentValue": "==1.11.0", "datasource": "pypi", "currentVersion": "1.11.0", "depIndex": 0, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/huggingface/datasets", "isSingleVersion": true, "fixedVersion": "1.11.0", "updates": [ { "bucket": "non-major", "newVersion": "1.18.4", "newValue": "==1.18.4", "releaseTimestamp": "2022-03-07T10:50:07.000Z", "newMajor": 1, "newMinor": 18, "updateType": "minor", "isRange": true, "branchName": "renovate/datasets-1.x" }, { "bucket": "major", "newVersion": "2.4.0", "newValue": "==2.4.0", "releaseTimestamp": "2022-07-25T13:21:43.000Z", "newMajor": 2, "newMinor": 4, "updateType": "major", "isRange": true, "branchName": "renovate/datasets-2.x" } ] } ] }, { "packageFile": "third_party/requirements.txt", "deps": [ { "depName": "certifi", "currentValue": "==2021.10.8", "datasource": "pypi", "currentVersion": "2021.10.8", "depIndex": 0, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/certifi/python-certifi", "isSingleVersion": true, "fixedVersion": "2021.10.8", "updates": [ { "bucket": "major", "newVersion": "2022.6.15", "newValue": "==2022.6.15", "releaseTimestamp": "2022-06-15T18:09:18.000Z", "newMajor": 2022, "newMinor": 6, "updateType": "major", "isRange": true, "branchName": "renovate/certifi-2022.x" } ] }, { "depName": "charset-normalizer", "currentValue": "==2.0.12", "datasource": "pypi", "currentVersion": "2.0.12", "depIndex": 1, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/ousret/charset_normalizer", "isSingleVersion": true, "fixedVersion": "2.0.12", "updates": [ { "bucket": "non-major", "newVersion": "2.1.1", "newValue": "==2.1.1", "releaseTimestamp": "2022-08-19T22:13:46.000Z", "newMajor": 2, "newMinor": 1, "updateType": "minor", "isRange": true, "branchName": "renovate/charset-normalizer-2.x" } ] }, { "depName": "datasets", "currentValue": "==1.11.0", "datasource": "pypi", "currentVersion": "1.11.0", "depIndex": 2, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/huggingface/datasets", "isSingleVersion": true, "fixedVersion": "1.11.0", "updates": [ { "bucket": "non-major", "newVersion": "1.18.4", "newValue": "==1.18.4", "releaseTimestamp": "2022-03-07T10:50:07.000Z", "newMajor": 1, "newMinor": 18, "updateType": "minor", "isRange": true, "branchName": "renovate/datasets-1.x" }, { "bucket": "major", "newVersion": "2.4.0", "newValue": "==2.4.0", "releaseTimestamp": "2022-07-25T13:21:43.000Z", "newMajor": 2, "newMinor": 4, "updateType": "major", "isRange": true, "branchName": "renovate/datasets-2.x" } ] }, { "depName": "dill", "currentValue": "==0.3.1.1", "datasource": "pypi", "currentVersion": "0.3.1.1", "depIndex": 3, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/uqfoundation/dill", "isSingleVersion": true, "fixedVersion": "0.3.1.1", "updates": [ { "bucket": "non-major", "newVersion": "0.3.5.1", "newValue": "==0.3.5.1", "releaseTimestamp": "2022-05-20T11:21:11.000Z", "newMajor": 0, "newMinor": 3, "updateType": "patch", "isRange": true, "branchName": "renovate/dill-0.x" } ] }, { "depName": "filelock", "currentValue": "==3.6.0", "datasource": "pypi", "currentVersion": "3.6.0", "depIndex": 4, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/tox-dev/py-filelock", "isSingleVersion": true, "fixedVersion": "3.6.0", "updates": [ { "bucket": "non-major", "newVersion": "3.8.0", "newValue": "==3.8.0", "releaseTimestamp": "2022-08-10T06:44:31.000Z", "newMajor": 3, "newMinor": 8, "updateType": "minor", "isRange": true, "branchName": "renovate/filelock-3.x" } ] }, { "depName": "fsspec", "currentValue": "==2021.11.1", "datasource": "pypi", "currentVersion": "2021.11.1", "depIndex": 5, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/fsspec/filesystem_spec", "changelogUrl": "https://filesystem-spec.readthedocs.io/en/latest/changelog.html", "isSingleVersion": true, "fixedVersion": "2021.11.1", "updates": [ { "bucket": "major", "newVersion": "2022.8.0", "newValue": "==2022.8.0", "releaseTimestamp": "2022-08-30T18:08:19.000Z", "newMajor": 2022, "newMinor": 8, "updateType": "major", "isRange": true, "branchName": "renovate/fsspec-2022.x" } ] }, { "depName": "huggingface-hub", "currentValue": "==0.0.19", "datasource": "pypi", "currentVersion": "0.0.19", "depIndex": 6, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/huggingface/huggingface_hub", "isSingleVersion": true, "fixedVersion": "0.0.19", "updates": [ { "bucket": "non-major", "newVersion": "0.9.1", "newValue": "==0.9.1", "releaseTimestamp": "2022-08-25T15:37:11.000Z", "newMajor": 0, "newMinor": 9, "updateType": "minor", "isRange": true, "branchName": "renovate/huggingface-hub-0.x" } ] }, { "depName": "idna", "currentValue": "==3.3", "datasource": "pypi", "currentVersion": "3.3", "depIndex": 7, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/kjd/idna", "fixedVersion": "3.3", "updates": [] }, { "depName": "multiprocess", "currentValue": "==0.70.9", "datasource": "pypi", "currentVersion": "0.70.9", "depIndex": 8, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/uqfoundation/multiprocess", "isSingleVersion": true, "fixedVersion": "0.70.9", "updates": [ { "bucket": "non-major", "newVersion": "0.70.13", "newValue": "==0.70.13", "releaseTimestamp": "2022-05-23T22:52:42.000Z", "newMajor": 0, "newMinor": 70, "updateType": "patch", "isRange": true, "branchName": "renovate/multiprocess-0.x" } ] }, { "depName": "numpy", "currentValue": "==1.21.3", "datasource": "pypi", "currentVersion": "1.21.3", "depIndex": 9, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/numpy/numpy", "homepage": "https://www.numpy.org", "isSingleVersion": true, "fixedVersion": "1.21.3", "updates": [ { "bucket": "non-major", "newVersion": "1.23.2", "newValue": "==1.23.2", "releaseTimestamp": "2022-08-14T00:14:09.000Z", "newMajor": 1, "newMinor": 23, "updateType": "minor", "isRange": true, "branchName": "renovate/numpy-1.x" } ] }, { "depName": "packaging", "currentValue": "==20.9", "datasource": "pypi", "currentVersion": "20.9", "depIndex": 10, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/pypa/packaging", "isSingleVersion": true, "fixedVersion": "20.9", "updates": [ { "bucket": "major", "newVersion": "21.3", "newValue": "==21.3", "releaseTimestamp": "2021-11-18T00:39:10.000Z", "newMajor": 21, "newMinor": 3, "updateType": "major", "isRange": true, "branchName": "renovate/packaging-21.x" } ] }, { "depName": "pandas", "currentValue": "==1.4.2", "datasource": "pypi", "currentVersion": "1.4.2", "depIndex": 11, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/pandas-dev/pandas", "homepage": "https://pandas.pydata.org", "isSingleVersion": true, "fixedVersion": "1.4.2", "updates": [ { "bucket": "non-major", "newVersion": "1.4.3", "newValue": "==1.4.3", "releaseTimestamp": "2022-06-23T13:30:09.000Z", "newMajor": 1, "newMinor": 4, "updateType": "patch", "isRange": true, "branchName": "renovate/pandas-1.x" } ] }, { "depName": "pyarrow", "currentValue": "==5.0.0", "datasource": "pypi", "currentVersion": "5.0.0", "depIndex": 12, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/apache/arrow", "homepage": "https://arrow.apache.org/", "isSingleVersion": true, "fixedVersion": "5.0.0", "updates": [ { "bucket": "major", "newVersion": "9.0.0", "newValue": "==9.0.0", "releaseTimestamp": "2022-08-03T10:45:55.000Z", "newMajor": 9, "newMinor": 0, "updateType": "major", "isRange": true, "branchName": "renovate/pyarrow-9.x" } ] }, { "depName": "pyparsing", "currentValue": "==2.4.7", "datasource": "pypi", "currentVersion": "2.4.7", "depIndex": 13, "warnings": [], "versioning": "pep440", "isSingleVersion": true, "fixedVersion": "2.4.7", "updates": [ { "bucket": "major", "newVersion": "3.0.9", "newValue": "==3.0.9", "releaseTimestamp": "2022-05-10T23:26:03.000Z", "newMajor": 3, "newMinor": 0, "updateType": "major", "isRange": true, "branchName": "renovate/pyparsing-3.x" } ] }, { "depName": "python-dateutil", "currentValue": "==2.8.2", "datasource": "pypi", "currentVersion": "2.8.2", "depIndex": 14, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/dateutil/dateutil", "fixedVersion": "2.8.2", "updates": [] }, { "depName": "pytz", "currentValue": "==2022.1", "datasource": "pypi", "currentVersion": "2022.1", "depIndex": 15, "warnings": [], "versioning": "pep440", "homepage": "http://pythonhosted.org/pytz", "isSingleVersion": true, "fixedVersion": "2022.1", "updates": [ { "bucket": "non-major", "newVersion": "2022.2.1", "newValue": "==2022.2.1", "releaseTimestamp": "2022-08-13T02:07:57.000Z", "newMajor": 2022, "newMinor": 2, "updateType": "minor", "isRange": true, "branchName": "renovate/pytz-2022.x" } ] }, { "depName": "pyyaml", "currentValue": "==5.4.1", "datasource": "pypi", "currentVersion": "5.4.1", "depIndex": 16, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/yaml/pyyaml", "homepage": "https://pyyaml.org/", "isSingleVersion": true, "fixedVersion": "5.4.1", "updates": [ { "bucket": "major", "newVersion": "6.0", "newValue": "==6.0", "releaseTimestamp": "2021-10-13T19:39:42.000Z", "newMajor": 6, "newMinor": 0, "updateType": "major", "isRange": true, "branchName": "renovate/pyyaml-6.x" } ] }, { "depName": "requests", "currentValue": "==2.27.1", "datasource": "pypi", "currentVersion": "2.27.1", "depIndex": 17, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/psf/requests", "homepage": "https://requests.readthedocs.io", "changelogUrl": "https://github.com/psf/requests/blob/master/HISTORY.md", "isSingleVersion": true, "fixedVersion": "2.27.1", "updates": [ { "bucket": "non-major", "newVersion": "2.28.1", "newValue": "==2.28.1", "releaseTimestamp": "2022-06-29T15:13:40.000Z", "newMajor": 2, "newMinor": 28, "updateType": "minor", "isRange": true, "branchName": "renovate/requests-2.x" } ] }, { "depName": "six", "currentValue": "==1.16.0", "datasource": "pypi", "currentVersion": "1.16.0", "depIndex": 18, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/benjaminp/six", "fixedVersion": "1.16.0", "updates": [] }, { "depName": "tqdm", "currentValue": "==4.63.0", "datasource": "pypi", "currentVersion": "4.63.0", "depIndex": 19, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/tqdm/tqdm", "homepage": "https://tqdm.github.io", "changelogUrl": "https://tqdm.github.io/releases", "isSingleVersion": true, "fixedVersion": "4.63.0", "updates": [ { "bucket": "non-major", "newVersion": "4.64.0", "newValue": "==4.64.0", "releaseTimestamp": "2022-04-04T01:48:46.000Z", "newMajor": 4, "newMinor": 64, "updateType": "minor", "isRange": true, "branchName": "renovate/tqdm-4.x" } ] }, { "depName": "typing-extensions", "currentValue": "==4.1.1", "datasource": "pypi", "currentVersion": "4.1.1", "depIndex": 20, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/python/typing_extensions", "changelogUrl": "https://github.com/python/typing_extensions/blob/main/CHANGELOG.md", "isSingleVersion": true, "fixedVersion": "4.1.1", "updates": [ { "bucket": "non-major", "newVersion": "4.3.0", "newValue": "==4.3.0", "releaseTimestamp": "2022-07-01T14:56:11.000Z", "newMajor": 4, "newMinor": 3, "updateType": "minor", "isRange": true, "branchName": "renovate/typing-extensions-4.x" } ] }, { "depName": "urllib3", "currentValue": "==1.26.8", "datasource": "pypi", "currentVersion": "1.26.8", "depIndex": 21, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/urllib3/urllib3", "homepage": "https://urllib3.readthedocs.io/", "isSingleVersion": true, "fixedVersion": "1.26.8", "updates": [ { "bucket": "non-major", "newVersion": "1.26.12", "newValue": "==1.26.12", "releaseTimestamp": "2022-08-22T13:27:19.000Z", "newMajor": 1, "newMinor": 26, "updateType": "patch", "isRange": true, "branchName": "renovate/urllib3-1.x" } ] }, { "depName": "xxhash", "currentValue": "==3.0.0", "datasource": "pypi", "currentVersion": "3.0.0", "depIndex": 22, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/ifduyue/python-xxhash", "fixedVersion": "3.0.0", "updates": [] } ] } ] } } DEBUG: processRepo() DEBUG: Processing 2 branches: renovate/datasets-1.x, renovate/datasets-2.x DEBUG: Calculating hourly PRs remaining DEBUG: getPrList success { "pullsTotal": 1, "requestsTotal": 1, "apiQuotaAffected": true } DEBUG: currentHourStart=2022-08-31T02:00:00.000+00:00 DEBUG: PR hourly limit remaining: 2 DEBUG: Calculating prConcurrentLimit (10) DEBUG: getBranchPr(renovate/datasets-1.x) DEBUG: findPr(renovate/datasets-1.x, undefined, open) DEBUG: Found PR #2 DEBUG: getBranchPr(renovate/datasets-2.x) DEBUG: findPr(renovate/datasets-2.x, undefined, open) DEBUG: findPr(renovate/datasets-2.x, undefined, closed) DEBUG: 1 PRs are currently open DEBUG: PR concurrent limit remaining: 9 DEBUG: Calculated maximum PRs remaining this run { "prsRemaining": 2 } DEBUG: PullRequests limit = 2 DEBUG: Calculating hourly PRs remaining DEBUG: currentHourStart=2022-08-31T02:00:00.000+00:00 DEBUG: PR hourly limit remaining: 2 DEBUG: Calculating branchConcurrentLimit (10) DEBUG: 1 already existing branches found: renovate/datasets-1.x DEBUG: Branch concurrent limit remaining: 9 DEBUG: Calculated maximum branches remaining this run { "branchesRemaining": 2 } DEBUG: Branches limit = 2 DEBUG: Branch fingerprint is unchanged, updates check can be skipped(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: getBranchPr(renovate/datasets-1.x)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: findPr(renovate/datasets-1.x, undefined, open)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Found PR #2(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: branchExists=true(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: dependencyDashboardCheck=undefined(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: PR rebase requested=false(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Checking if PR has been edited(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Found existing branch PR(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Checking schedule(after 10pm every weekday,before 4am every weekday, America/Chihuahua)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Found timezone(branch="renovate/datasets-1.x") { "baseBranch": "main", "timezone": "America/Chihuahua" } DEBUG: Adjusting now for timezone(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Checking 2 schedule(s)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Checking schedule "after 10pm every weekday"(branch="renovate/datasets-1.x") { "baseBranch": "main", "parsedSchedule": { "schedules": [ { "t_a": [ 79200 ], "d": [ 2, 3, 4, 5, 6 ] } ], "exceptions": [], "error": -1 } } DEBUG: Checking schedule "before 4am every weekday"(branch="renovate/datasets-1.x") { "baseBranch": "main", "parsedSchedule": { "schedules": [ { "t_b": [ 14400 ], "d": [ 2, 3, 4, 5, 6 ] } ], "exceptions": [], "error": -1 } } DEBUG: Package not scheduled(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Branch + PR exists but is not scheduled -- will update if necessary(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Branch already exists(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: getBranchPr(renovate/datasets-1.x)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: findPr(renovate/datasets-1.x, undefined, open)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Found PR #2(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Branch is behind base branch and needs rebasing(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Branch is unmodified, so can be rebased(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Using reuseExistingBranch: false(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Setting current branch to main(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Initializing git repository into /mnt/renovate/gh/corypaik/renovate-loop(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Performing blobless clone(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: git clone completed(branch="renovate/datasets-1.x") { "baseBranch": "main", "durationMs": 1449 } DEBUG: latest repository commit(branch="renovate/datasets-1.x") { "baseBranch": "main", "latestCommit": { "hash": "5267848cc1c6b6dfb1d031f573b25d2ab36ebc82", "date": "2022-08-30T20:06:50-06:00", "message": "ci: always run pip lockfile step", "refs": "HEAD -> main, origin/main, origin/HEAD", "body": "", "author_name": "Cory Paik", "author_email": "corypaik@gmail.com" } } DEBUG: latest commit(branch="renovate/datasets-1.x") { "baseBranch": "main", "branchName": "main", "latestCommitDate": "2022-08-30T20:06:50-06:00" } DEBUG: manager.getUpdatedPackageFiles() reuseExistingBranch=false(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Starting search at index 527(packageFile="third_party/requirements.txt", branch="renovate/datasets-1.x") { "baseBranch": "main", "depName": "datasets" } DEBUG: Found match at index 527(packageFile="third_party/requirements.txt", branch="renovate/datasets-1.x") { "baseBranch": "main", "depName": "datasets" } DEBUG: Contents updated(packageFile="third_party/requirements.txt", branch="renovate/datasets-1.x") { "baseBranch": "main", "depName": "datasets" } DEBUG: Starting search at index 8(packageFile="third_party/requirements.in", branch="renovate/datasets-1.x") { "baseBranch": "main", "depName": "datasets" } DEBUG: Found match at index 8(packageFile="third_party/requirements.in", branch="renovate/datasets-1.x") { "baseBranch": "main", "depName": "datasets" } DEBUG: Contents updated(packageFile="third_party/requirements.in", branch="renovate/datasets-1.x") { "baseBranch": "main", "depName": "datasets" } DEBUG: pip_requirements.updateArtifacts(third_party/requirements.txt)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Setting BUILDPACK_CACHE_DIR to /tmp/containerbase(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Using docker to execute(branch="renovate/datasets-1.x") { "baseBranch": "main", "image": "sidecar" } DEBUG: Resolved stable matching version(branch="renovate/datasets-1.x") { "baseBranch": "main", "toolName": "python", "resolvedVersion": "3.10.6" } DEBUG: containerbaseDir is separate from cacheDir(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: No tag or tagConstraint specified(branch="renovate/datasets-1.x") { "baseBranch": "main", "image": "docker.io/renovate/sidecar" } DEBUG: Fetching Docker image: docker.io/renovate/sidecar(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Finished fetching Docker image docker.io/renovate/sidecar@sha256:ee8108a5a17ef370f15da2c11bf7b2da9035153fde1587118ac9cc5c6c56cf0a(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Executing command(branch="renovate/datasets-1.x") { "baseBranch": "main", "command": "docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/corypaik/renovate-loop\":\"/mnt/renovate/gh/corypaik/renovate-loop\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -w \"/mnt/renovate/gh/corypaik/renovate-loop\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.10.6 && pip install --user hashin && hashin datasets==1.18.4 -r third_party/requirements.txt\"" } DEBUG: exec completed(branch="renovate/datasets-1.x") { "baseBranch": "main", "cmd": "docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/corypaik/renovate-loop\":\"/mnt/renovate/gh/corypaik/renovate-loop\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -w \"/mnt/renovate/gh/corypaik/renovate-loop\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.10.6 && pip install --user hashin && hashin datasets==1.18.4 -r third_party/requirements.txt\"", "durationMs": 27296, "stdout": "installing v2 tool python v3.10.6\nUsing prebuild python\nRequirement already satisfied: pip in /opt/buildpack/tools/python/3.10.6/lib/python3.10/site-packages (22.2.1)\nCollecting pip\n Downloading pip-22.2.2-py3-none-any.whl (2.0 MB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.0/2.0 MB 21.0 MB/s eta 0:00:00\nInstalling collected packages: pip\n Attempting uninstall: pip\n Found existing installation: pip 22.2.1\n Uninstalling pip-22.2.1:\n Successfully uninstalled pip-22.2.1\nSuccessfully installed pip-22.2.2\nFiles removed: 2\nlinking tool python v3.10.6\nPython 3.10.6\npip 22.2.2 from /opt/buildpack/tools/python/3.10.6/lib/python3.10/site-packages/pip (python 3.10)\nInstalled v2 /usr/local/buildpack/tools/v2/python.sh in 20 seconds\nCollecting hashin\n Downloading hashin-0.17.0-py2.py3-none-any.whl (16 kB)\nCollecting pip-api\n Downloading pip_api-0.0.30-py3-none-any.whl (111 kB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 111.6/111.6 kB 5.9 MB/s eta 0:00:00\nCollecting packaging\n Downloading packaging-21.3-py3-none-any.whl (40 kB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 40.8/40.8 kB 4.7 MB/s eta 0:00:00\nCollecting pyparsing!=3.0.5,>=2.0.2\n Downloading pyparsing-3.0.9-py3-none-any.whl (98 kB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 98.3/98.3 kB 11.1 MB/s eta 0:00:00\nRequirement already satisfied: pip in /opt/buildpack/tools/python/3.10.6/lib/python3.10/site-packages (from pip-api->hashin) (22.2.2)\nInstalling collected packages: pyparsing, pip-api, packaging, hashin\nSuccessfully installed hashin-0.17.0 packaging-21.3 pip-api-0.0.30 pyparsing-3.0.9\n", "stderr": "Found file in cache: /tmp/containerbase/f79e18f004c2278a4c76a549ca9e8beed6ee66d17e187e79a230ba06836df191/python-3.10.6-focal-x86_64.tar.xz\n" } DEBUG: Returning updated third_party/requirements.txt(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: pip_requirements.updateArtifacts(third_party/requirements.in)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: No hashin commands to run - returning(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Updated 2 package files(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Updated 1 lock files(branch="renovate/datasets-1.x") { "baseBranch": "main", "updatedArtifacts": [ "third_party/requirements.txt" ] } DEBUG: Getting comments for #2(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Found 0 comments(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: isBranchConflicted(main, renovate/datasets-1.x)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Using cached result false for isBranchConflicted(main, renovate/datasets-1.x)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: 2 file(s) to commit(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Preparing files for committing to branch renovate/datasets-1.x(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Setting git author name(branch="renovate/datasets-1.x") { "baseBranch": "main", "gitAuthorName": "Renovate Bot" } DEBUG: Setting git author email(branch="renovate/datasets-1.x") { "baseBranch": "main", "gitAuthorEmail": "bot@renovateapp.com" } DEBUG: git commit(branch="renovate/datasets-1.x") { "baseBranch": "main", "deletedFiles": [], "ignoredFiles": [], "result": { "author": null, "branch": "renovate/datasets-1.x", "commit": "210146775398f010baed48f2e6891d780406efa2", "root": false, "summary": { "changes": 2, "insertions": 4, "deletions": 4 } } } DEBUG: resetToCommit(5267848cc1c6b6dfb1d031f573b25d2ab36ebc82)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Fetching branch renovate/datasets-1.x(branch="renovate/datasets-1.x") { "baseBranch": "main" } INFO: Branch updated(branch="renovate/datasets-1.x") { "baseBranch": "main", "commitSha": "78dae2cf45f995142afb48597777d62fa9677a1f" } DEBUG: Updating renovate/stability-days status check state to green(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Setting branch status(branch="renovate/datasets-1.x") { "baseBranch": "main", "context": "renovate/stability-days", "state": "green" } DEBUG: Branch status pending(branch="renovate/datasets-1.x") { "baseBranch": "main", "commitSha": "78dae2cf45f995142afb48597777d62fa9677a1f" } DEBUG: getBranchPr(renovate/datasets-2.x)(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: findPr(renovate/datasets-2.x, undefined, open)(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: findPr(renovate/datasets-2.x, undefined, closed)(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: branchExists=false(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: dependencyDashboardCheck=undefined(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: recreateClosed is false(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: findPr(renovate/datasets-2.x, build: update dependency datasets to v2, !open)(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: prAlreadyExisted=false(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: Checking schedule(after 10pm every weekday,before 4am every weekday, America/Chihuahua)(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: Found timezone(branch="renovate/datasets-2.x") { "baseBranch": "main", "timezone": "America/Chihuahua" } DEBUG: Adjusting now for timezone(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: Checking 2 schedule(s)(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: Checking schedule "after 10pm every weekday"(branch="renovate/datasets-2.x") { "baseBranch": "main", "parsedSchedule": { "schedules": [ { "t_a": [ 79200 ], "d": [ 2, 3, 4, 5, 6 ] } ], "exceptions": [], "error": -1 } } DEBUG: Checking schedule "before 4am every weekday"(branch="renovate/datasets-2.x") { "baseBranch": "main", "parsedSchedule": { "schedules": [ { "t_b": [ 14400 ], "d": [ 2, 3, 4, 5, 6 ] } ], "exceptions": [], "error": -1 } } DEBUG: Package not scheduled(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: Skipping branch creation as not within schedule(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: getBranchPr(renovate/datasets-1.x) DEBUG: findPr(renovate/datasets-1.x, undefined, open) DEBUG: Found PR #2 DEBUG: Ensuring Dependency Dashboard DEBUG: ensureIssue(Dependency Dashboard) DEBUG: Patching issue DEBUG: Issue updated DEBUG: Removing any stale branches DEBUG: config.repoIsOnboarded=true DEBUG: Branch lists { "branchList": [ "renovate/datasets-1.x", "renovate/datasets-2.x" ], "renovateBranches": [ "renovate/datasets-1.x" ] } DEBUG: remainingBranches= DEBUG: No branches to clean up DEBUG: Cleaning up Renovate refs: refs/renovate/* DEBUG: PackageFiles.clear() - Package files deleted { "baseBranches": [ "main" ] } DEBUG: Renovate repository PR statistics { "stats": { "total": 1, "open": 1, "closed": 0, "merged": 0 } } DEBUG: Repository result: done, status: onboarded, enabled: true, onboarded: true DEBUG: Repository timing splits (milliseconds) { "splits": { "init": 3599, "extract": 494, "lookup": 2772, "onboarding": 0, "update": 35845 }, "total": 45597 } DEBUG: http statistics { "urls": { "https://api.github.com/graphql (POST,200)": 2, "https://api.github.com/repos/corypaik/renovate-loop/commits/78dae2cf45f995142afb48597777d62fa9677a1f/statuses (GET,200)": 2, "https://api.github.com/repos/corypaik/renovate-loop/commits/renovate/datasets-1.x/status (GET,200)": 1, "https://api.github.com/repos/corypaik/renovate-loop/contents/renovate.json (GET,200)": 1, "https://api.github.com/repos/corypaik/renovate-loop/git/commits (POST,201)": 1, "https://api.github.com/repos/corypaik/renovate-loop/git/refs/heads/renovate/datasets-1.x (PATCH,200)": 1, "https://api.github.com/repos/corypaik/renovate-loop/git/trees (POST,201)": 1, "https://api.github.com/repos/corypaik/renovate-loop/issues/1 (GET,200)": 2, "https://api.github.com/repos/corypaik/renovate-loop/issues/1 (PATCH,200)": 1, "https://api.github.com/repos/corypaik/renovate-loop/issues/2/comments (GET,200)": 1, "https://api.github.com/repos/corypaik/renovate-loop/pulls (GET,200)": 1, "https://api.github.com/repos/corypaik/renovate-loop/statuses/78dae2cf45f995142afb48597777d62fa9677a1f (POST,201)": 1, "https://api.github.com/repos/whitesource/merge-confidence/contents/beta.json (GET,200)": 1, "https://pypi.org/pypi/certifi/json (GET,200)": 1 }, "hostStats": { "api.github.com": { "requestCount": 16, "requestAvgMs": 269, "queueAvgMs": 0 }, "pypi.org": { "requestCount": 1, "requestAvgMs": 989, "queueAvgMs": 0 } }, "totalRequests": 17 } INFO: Repository finished { "durationMs": 45597 } ``` </details> ### Have you created a minimal reproduction repository? I have linked to a minimal reproduction repository in the bug description
1.0
Renovate is stuck in an infinite loop overwriting a ci bot's commits - ### How are you running Renovate? Mend Renovate hosted app on github.com ### If you're self-hosting Renovate, tell us what version of Renovate you run. _No response_ ### If you're self-hosting Renovate, select which platform you are using. _No response_ ### If you're self-hosting Renovate, tell us what version of the platform you run. _No response_ ### Was this something which used to work for you, and then stopped? It used to work, and then stopped ### Describe the bug Renovate is stuck in an infinite loop overwriting a ci bot's fixup commits. This started happening sometime in the past 2 months (the configuration has not been changed since). So far I have only observed this with python packages, but I do not have a similar procedure for other dependencies. The process is something like this: 1. Renovate updates package `x` and commits to the repository, which triggers the first ci run. 2. The repository uses lock files with `pip-compile`, so before running builds and tests, the ci agent runs a script to update the lockfiles. This takes care of resolving dependency changes of `x`. If any changes were made, the ci bot commits them as a fixup commit (`git commit --fixup HEAD`) and pushes it to GitHub. 3. The new commit triggers a second ci run. This time, the check should pass and move on. The ci bot is registered as a `gitIgnoredAuthor`, but in the past renovate would only overwrite the changes if the `main` branch was updated (I have `rebaseWhen` as `behind-base-branch`). Currently, every time the ci bot commits a change renovate seems to overwrite it the next time it runs. This means the repository update process is effectively stuck in an infinite loop. I'm not sure exactly what triggers this behavior and it's been a while since it worked, but it seems to have been around the same time renovate changed email names from `bot@renovateapp.com` to `29139614+renovate[bot]@users.noreply.github.com`. I recall having to manually rebase the existing branches as they were marked as modified. I've provided debug logs from a minimal reproduction repository, which is publicly available on [GitHub](https://github.com/corypaik/renovate-loop). The repository contains one example package that causes this issue, all others have been removed or disabled. There is also a public [Buildkite](https://buildkite.com/corypaik/renovate-loop) available for reference. ### Relevant debug logs <details><summary>Logs</summary> ``` DEBUG: No dangling containers to remove INFO: Repository started { "renovateVersion": "32.184.2" } DEBUG: Using localDir: /mnt/renovate/gh/corypaik/renovate-loop DEBUG: PackageFiles.clear() - Package files deleted { "baseBranches": [] } DEBUG: resetMemCache() DEBUG: initRepo("corypaik/renovate-loop") DEBUG: corypaik/renovate-loop default branch = main DEBUG: Using app token for git init DEBUG: Repository cache is restored from revision 13 DEBUG: Resetting npmrc DEBUG: detectSemanticCommits() DEBUG: checkOnboarding() DEBUG: isOnboarded() DEBUG: Checking cached config file name DEBUG: Existing config file confirmed DEBUG: Repository config { "fileName": "renovate.json", "config": { "$schema": "https://docs.renovatebot.com/renovate-schema.json", "extends": [ ":ignoreModulesAndTests", "group:monorepos", "group:recommended", "workarounds:all" ], "semanticCommits": "enabled", "semanticCommitType": "build", "semanticCommitScope": "", "prHourlyLimit": 2, "prConcurrentLimit": 10, "dependencyDashboard": true, "labels": [ "target: minor", "comp: build & ci", "action: review" ], "timezone": "America/Chihuahua", "schedule": [ "after 10pm every weekday", "before 4am every weekday" ], "baseBranches": [ "main" ], "rebaseWhen": "behind-base-branch", "pip_requirements": { "fileMatch": [ ".*requirements.in$" ] }, "prCreation": "not-pending", "stabilityDays": 3, "gitIgnoredAuthors": [ "buildkite-bot@buildkite.com" ], "pre-commit": { "enabled": true }, "packageRules": [ { "description": "Always pin direct python dependencies to ensure lockfile updates do not revert to a different version.", "matchFiles": [ "third_party/requirements.in" ], "rangeStrategy": "pin" }, { "description": "All python deps, except datasets (for minimal reproduction).", "matchPackagePatterns": [ "*" ], "matchUpdateTypes": [ "major", "minor", "patch" ], "matchLanguages": [ "python" ], "excludePackageNames": [ "datasets" ], "enabled": false } ] } } DEBUG: Repo is onboarded DEBUG: migrateAndValidate() DEBUG: No config migration necessary DEBUG: massaged config { "config": { "$schema": "https://docs.renovatebot.com/renovate-schema.json", "extends": [ "github>whitesource/merge-confidence:beta", ":ignoreModulesAndTests", "group:monorepos", "group:recommended", "workarounds:all" ], "semanticCommits": "enabled", "semanticCommitType": "build", "semanticCommitScope": "", "prHourlyLimit": 2, "prConcurrentLimit": 10, "dependencyDashboard": true, "labels": [ "target: minor", "comp: build & ci", "action: review" ], "timezone": "America/Chihuahua", "schedule": [ "after 10pm every weekday", "before 4am every weekday" ], "baseBranches": [ "main" ], "rebaseWhen": "behind-base-branch", "pip_requirements": { "fileMatch": [ ".*requirements.in$" ] }, "prCreation": "not-pending", "stabilityDays": 3, "gitIgnoredAuthors": [ "buildkite-bot@buildkite.com" ], "pre-commit": { "enabled": true }, "packageRules": [ { "description": [ "Always pin direct python dependencies to ensure lockfile updates do not revert to a different version." ], "matchFiles": [ "third_party/requirements.in" ], "rangeStrategy": "pin" }, { "description": [ "All python deps, except datasets (for minimal reproduction)." ], "matchPackagePatterns": [ "*" ], "matchUpdateTypes": [ "major", "minor", "patch" ], "matchLanguages": [ "python" ], "excludePackageNames": [ "datasets" ], "enabled": false } ] } } DEBUG: migrated config { "config": { "$schema": "https://docs.renovatebot.com/renovate-schema.json", "extends": [ "github>whitesource/merge-confidence:beta", ":ignoreModulesAndTests", "group:monorepos", "group:recommended", "workarounds:all" ], "semanticCommits": "enabled", "semanticCommitType": "build", "semanticCommitScope": "", "prHourlyLimit": 2, "prConcurrentLimit": 10, "dependencyDashboard": true, "labels": [ "target: minor", "comp: build & ci", "action: review" ], "timezone": "America/Chihuahua", "schedule": [ "after 10pm every weekday", "before 4am every weekday" ], "baseBranches": [ "main" ], "rebaseWhen": "behind-base-branch", "pip_requirements": { "fileMatch": [ ".*requirements.in$" ] }, "prCreation": "not-pending", "stabilityDays": 3, "gitIgnoredAuthors": [ "buildkite-bot@buildkite.com" ], "pre-commit": { "enabled": true }, "packageRules": [ { "description": [ "Always pin direct python dependencies to ensure lockfile updates do not revert to a different version." ], "matchFiles": [ "third_party/requirements.in" ], "rangeStrategy": "pin" }, { "description": [ "All python deps, except datasets (for minimal reproduction)." ], "matchPackagePatterns": [ "*" ], "matchUpdateTypes": [ "major", "minor", "patch" ], "matchLanguages": [ "python" ], "excludePackageNames": [ "datasets" ], "enabled": false } ] } } DEBUG: Setting hostRules from config DEBUG: Found repo ignorePaths { "ignorePaths": [ "**/node_modules/**", "**/bower_components/**", "**/vendor/**", "**/examples/**", "**/__tests__/**", "**/test/**", "**/tests/**", "**/__fixtures__/**" ] } DEBUG: No vulnerability alerts found DEBUG: No vulnerability alerts found DEBUG: findIssue(Dependency Dashboard) DEBUG: Retrieving issueList DEBUG: Retrieved 1 issues DEBUG: Found issue 1 DEBUG: baseBranches { "baseBranches": [ "main" ] } DEBUG: baseBranch: main { "baseBranch": "main" } DEBUG: extract() { "baseBranch": "main" } DEBUG: Found cached extract { "baseBranch": "main", "baseBranchSha": "5267848cc1c6b6dfb1d031f573b25d2ab36ebc82" } DEBUG: Deleted cached dep updates { "baseBranch": "main" } INFO: Dependency extraction complete { "baseBranch": "main", "stats": { "managers": { "bazel": { "fileCount": 1, "depCount": 1 }, "pip_requirements": { "fileCount": 2, "depCount": 24 } }, "total": { "fileCount": 3, "depCount": 25 } } } DEBUG: baseBranch: main { "baseBranch": "main" } DEBUG: PackageFiles.add() - Package file saved for branch { "baseBranch": "main" } DEBUG: Package releases lookups complete { "baseBranch": "main" } DEBUG: branchifyUpgrades { "baseBranch": "main" } DEBUG: 4 flattened updates found: datasets, datasets, datasets, datasets { "baseBranch": "main" } DEBUG: Returning 2 branch(es) { "baseBranch": "main" } DEBUG: config.repoIsOnboarded=true { "baseBranch": "main" } DEBUG: packageFiles with updates { "baseBranch": "main", "config": { "bazel": [ { "packageFile": "WORKSPACE", "deps": [ { "datasource": "github-tags", "depType": "http_archive", "depName": "rules_python", "packageName": "bazelbuild/rules_python", "currentValue": "refs", "managerData": { "def": "http_archive(\n name = \"rules_python\",\n sha256 = \"b593d13bb43c94ce94b483c2858e53a9b811f6f10e1e0eedc61073bd90e58d9c\",\n strip_prefix = \"rules_python-0.12.0\",\n url = \"https://github.com/bazelbuild/rules_python/archive/refs/tags/0.12.0.tar.gz\",\n)" }, "depIndex": 0, "warnings": [], "versioning": "semver", "skipReason": "invalid-value", "updates": [] } ] } ], "pip_requirements": [ { "packageFile": "third_party/requirements.in", "deps": [ { "depName": "datasets", "currentValue": "==1.11.0", "datasource": "pypi", "currentVersion": "1.11.0", "depIndex": 0, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/huggingface/datasets", "isSingleVersion": true, "fixedVersion": "1.11.0", "updates": [ { "bucket": "non-major", "newVersion": "1.18.4", "newValue": "==1.18.4", "releaseTimestamp": "2022-03-07T10:50:07.000Z", "newMajor": 1, "newMinor": 18, "updateType": "minor", "isRange": true, "branchName": "renovate/datasets-1.x" }, { "bucket": "major", "newVersion": "2.4.0", "newValue": "==2.4.0", "releaseTimestamp": "2022-07-25T13:21:43.000Z", "newMajor": 2, "newMinor": 4, "updateType": "major", "isRange": true, "branchName": "renovate/datasets-2.x" } ] } ] }, { "packageFile": "third_party/requirements.txt", "deps": [ { "depName": "certifi", "currentValue": "==2021.10.8", "datasource": "pypi", "currentVersion": "2021.10.8", "depIndex": 0, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/certifi/python-certifi", "isSingleVersion": true, "fixedVersion": "2021.10.8", "updates": [ { "bucket": "major", "newVersion": "2022.6.15", "newValue": "==2022.6.15", "releaseTimestamp": "2022-06-15T18:09:18.000Z", "newMajor": 2022, "newMinor": 6, "updateType": "major", "isRange": true, "branchName": "renovate/certifi-2022.x" } ] }, { "depName": "charset-normalizer", "currentValue": "==2.0.12", "datasource": "pypi", "currentVersion": "2.0.12", "depIndex": 1, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/ousret/charset_normalizer", "isSingleVersion": true, "fixedVersion": "2.0.12", "updates": [ { "bucket": "non-major", "newVersion": "2.1.1", "newValue": "==2.1.1", "releaseTimestamp": "2022-08-19T22:13:46.000Z", "newMajor": 2, "newMinor": 1, "updateType": "minor", "isRange": true, "branchName": "renovate/charset-normalizer-2.x" } ] }, { "depName": "datasets", "currentValue": "==1.11.0", "datasource": "pypi", "currentVersion": "1.11.0", "depIndex": 2, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/huggingface/datasets", "isSingleVersion": true, "fixedVersion": "1.11.0", "updates": [ { "bucket": "non-major", "newVersion": "1.18.4", "newValue": "==1.18.4", "releaseTimestamp": "2022-03-07T10:50:07.000Z", "newMajor": 1, "newMinor": 18, "updateType": "minor", "isRange": true, "branchName": "renovate/datasets-1.x" }, { "bucket": "major", "newVersion": "2.4.0", "newValue": "==2.4.0", "releaseTimestamp": "2022-07-25T13:21:43.000Z", "newMajor": 2, "newMinor": 4, "updateType": "major", "isRange": true, "branchName": "renovate/datasets-2.x" } ] }, { "depName": "dill", "currentValue": "==0.3.1.1", "datasource": "pypi", "currentVersion": "0.3.1.1", "depIndex": 3, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/uqfoundation/dill", "isSingleVersion": true, "fixedVersion": "0.3.1.1", "updates": [ { "bucket": "non-major", "newVersion": "0.3.5.1", "newValue": "==0.3.5.1", "releaseTimestamp": "2022-05-20T11:21:11.000Z", "newMajor": 0, "newMinor": 3, "updateType": "patch", "isRange": true, "branchName": "renovate/dill-0.x" } ] }, { "depName": "filelock", "currentValue": "==3.6.0", "datasource": "pypi", "currentVersion": "3.6.0", "depIndex": 4, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/tox-dev/py-filelock", "isSingleVersion": true, "fixedVersion": "3.6.0", "updates": [ { "bucket": "non-major", "newVersion": "3.8.0", "newValue": "==3.8.0", "releaseTimestamp": "2022-08-10T06:44:31.000Z", "newMajor": 3, "newMinor": 8, "updateType": "minor", "isRange": true, "branchName": "renovate/filelock-3.x" } ] }, { "depName": "fsspec", "currentValue": "==2021.11.1", "datasource": "pypi", "currentVersion": "2021.11.1", "depIndex": 5, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/fsspec/filesystem_spec", "changelogUrl": "https://filesystem-spec.readthedocs.io/en/latest/changelog.html", "isSingleVersion": true, "fixedVersion": "2021.11.1", "updates": [ { "bucket": "major", "newVersion": "2022.8.0", "newValue": "==2022.8.0", "releaseTimestamp": "2022-08-30T18:08:19.000Z", "newMajor": 2022, "newMinor": 8, "updateType": "major", "isRange": true, "branchName": "renovate/fsspec-2022.x" } ] }, { "depName": "huggingface-hub", "currentValue": "==0.0.19", "datasource": "pypi", "currentVersion": "0.0.19", "depIndex": 6, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/huggingface/huggingface_hub", "isSingleVersion": true, "fixedVersion": "0.0.19", "updates": [ { "bucket": "non-major", "newVersion": "0.9.1", "newValue": "==0.9.1", "releaseTimestamp": "2022-08-25T15:37:11.000Z", "newMajor": 0, "newMinor": 9, "updateType": "minor", "isRange": true, "branchName": "renovate/huggingface-hub-0.x" } ] }, { "depName": "idna", "currentValue": "==3.3", "datasource": "pypi", "currentVersion": "3.3", "depIndex": 7, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/kjd/idna", "fixedVersion": "3.3", "updates": [] }, { "depName": "multiprocess", "currentValue": "==0.70.9", "datasource": "pypi", "currentVersion": "0.70.9", "depIndex": 8, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/uqfoundation/multiprocess", "isSingleVersion": true, "fixedVersion": "0.70.9", "updates": [ { "bucket": "non-major", "newVersion": "0.70.13", "newValue": "==0.70.13", "releaseTimestamp": "2022-05-23T22:52:42.000Z", "newMajor": 0, "newMinor": 70, "updateType": "patch", "isRange": true, "branchName": "renovate/multiprocess-0.x" } ] }, { "depName": "numpy", "currentValue": "==1.21.3", "datasource": "pypi", "currentVersion": "1.21.3", "depIndex": 9, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/numpy/numpy", "homepage": "https://www.numpy.org", "isSingleVersion": true, "fixedVersion": "1.21.3", "updates": [ { "bucket": "non-major", "newVersion": "1.23.2", "newValue": "==1.23.2", "releaseTimestamp": "2022-08-14T00:14:09.000Z", "newMajor": 1, "newMinor": 23, "updateType": "minor", "isRange": true, "branchName": "renovate/numpy-1.x" } ] }, { "depName": "packaging", "currentValue": "==20.9", "datasource": "pypi", "currentVersion": "20.9", "depIndex": 10, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/pypa/packaging", "isSingleVersion": true, "fixedVersion": "20.9", "updates": [ { "bucket": "major", "newVersion": "21.3", "newValue": "==21.3", "releaseTimestamp": "2021-11-18T00:39:10.000Z", "newMajor": 21, "newMinor": 3, "updateType": "major", "isRange": true, "branchName": "renovate/packaging-21.x" } ] }, { "depName": "pandas", "currentValue": "==1.4.2", "datasource": "pypi", "currentVersion": "1.4.2", "depIndex": 11, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/pandas-dev/pandas", "homepage": "https://pandas.pydata.org", "isSingleVersion": true, "fixedVersion": "1.4.2", "updates": [ { "bucket": "non-major", "newVersion": "1.4.3", "newValue": "==1.4.3", "releaseTimestamp": "2022-06-23T13:30:09.000Z", "newMajor": 1, "newMinor": 4, "updateType": "patch", "isRange": true, "branchName": "renovate/pandas-1.x" } ] }, { "depName": "pyarrow", "currentValue": "==5.0.0", "datasource": "pypi", "currentVersion": "5.0.0", "depIndex": 12, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/apache/arrow", "homepage": "https://arrow.apache.org/", "isSingleVersion": true, "fixedVersion": "5.0.0", "updates": [ { "bucket": "major", "newVersion": "9.0.0", "newValue": "==9.0.0", "releaseTimestamp": "2022-08-03T10:45:55.000Z", "newMajor": 9, "newMinor": 0, "updateType": "major", "isRange": true, "branchName": "renovate/pyarrow-9.x" } ] }, { "depName": "pyparsing", "currentValue": "==2.4.7", "datasource": "pypi", "currentVersion": "2.4.7", "depIndex": 13, "warnings": [], "versioning": "pep440", "isSingleVersion": true, "fixedVersion": "2.4.7", "updates": [ { "bucket": "major", "newVersion": "3.0.9", "newValue": "==3.0.9", "releaseTimestamp": "2022-05-10T23:26:03.000Z", "newMajor": 3, "newMinor": 0, "updateType": "major", "isRange": true, "branchName": "renovate/pyparsing-3.x" } ] }, { "depName": "python-dateutil", "currentValue": "==2.8.2", "datasource": "pypi", "currentVersion": "2.8.2", "depIndex": 14, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/dateutil/dateutil", "fixedVersion": "2.8.2", "updates": [] }, { "depName": "pytz", "currentValue": "==2022.1", "datasource": "pypi", "currentVersion": "2022.1", "depIndex": 15, "warnings": [], "versioning": "pep440", "homepage": "http://pythonhosted.org/pytz", "isSingleVersion": true, "fixedVersion": "2022.1", "updates": [ { "bucket": "non-major", "newVersion": "2022.2.1", "newValue": "==2022.2.1", "releaseTimestamp": "2022-08-13T02:07:57.000Z", "newMajor": 2022, "newMinor": 2, "updateType": "minor", "isRange": true, "branchName": "renovate/pytz-2022.x" } ] }, { "depName": "pyyaml", "currentValue": "==5.4.1", "datasource": "pypi", "currentVersion": "5.4.1", "depIndex": 16, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/yaml/pyyaml", "homepage": "https://pyyaml.org/", "isSingleVersion": true, "fixedVersion": "5.4.1", "updates": [ { "bucket": "major", "newVersion": "6.0", "newValue": "==6.0", "releaseTimestamp": "2021-10-13T19:39:42.000Z", "newMajor": 6, "newMinor": 0, "updateType": "major", "isRange": true, "branchName": "renovate/pyyaml-6.x" } ] }, { "depName": "requests", "currentValue": "==2.27.1", "datasource": "pypi", "currentVersion": "2.27.1", "depIndex": 17, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/psf/requests", "homepage": "https://requests.readthedocs.io", "changelogUrl": "https://github.com/psf/requests/blob/master/HISTORY.md", "isSingleVersion": true, "fixedVersion": "2.27.1", "updates": [ { "bucket": "non-major", "newVersion": "2.28.1", "newValue": "==2.28.1", "releaseTimestamp": "2022-06-29T15:13:40.000Z", "newMajor": 2, "newMinor": 28, "updateType": "minor", "isRange": true, "branchName": "renovate/requests-2.x" } ] }, { "depName": "six", "currentValue": "==1.16.0", "datasource": "pypi", "currentVersion": "1.16.0", "depIndex": 18, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/benjaminp/six", "fixedVersion": "1.16.0", "updates": [] }, { "depName": "tqdm", "currentValue": "==4.63.0", "datasource": "pypi", "currentVersion": "4.63.0", "depIndex": 19, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/tqdm/tqdm", "homepage": "https://tqdm.github.io", "changelogUrl": "https://tqdm.github.io/releases", "isSingleVersion": true, "fixedVersion": "4.63.0", "updates": [ { "bucket": "non-major", "newVersion": "4.64.0", "newValue": "==4.64.0", "releaseTimestamp": "2022-04-04T01:48:46.000Z", "newMajor": 4, "newMinor": 64, "updateType": "minor", "isRange": true, "branchName": "renovate/tqdm-4.x" } ] }, { "depName": "typing-extensions", "currentValue": "==4.1.1", "datasource": "pypi", "currentVersion": "4.1.1", "depIndex": 20, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/python/typing_extensions", "changelogUrl": "https://github.com/python/typing_extensions/blob/main/CHANGELOG.md", "isSingleVersion": true, "fixedVersion": "4.1.1", "updates": [ { "bucket": "non-major", "newVersion": "4.3.0", "newValue": "==4.3.0", "releaseTimestamp": "2022-07-01T14:56:11.000Z", "newMajor": 4, "newMinor": 3, "updateType": "minor", "isRange": true, "branchName": "renovate/typing-extensions-4.x" } ] }, { "depName": "urllib3", "currentValue": "==1.26.8", "datasource": "pypi", "currentVersion": "1.26.8", "depIndex": 21, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/urllib3/urllib3", "homepage": "https://urllib3.readthedocs.io/", "isSingleVersion": true, "fixedVersion": "1.26.8", "updates": [ { "bucket": "non-major", "newVersion": "1.26.12", "newValue": "==1.26.12", "releaseTimestamp": "2022-08-22T13:27:19.000Z", "newMajor": 1, "newMinor": 26, "updateType": "patch", "isRange": true, "branchName": "renovate/urllib3-1.x" } ] }, { "depName": "xxhash", "currentValue": "==3.0.0", "datasource": "pypi", "currentVersion": "3.0.0", "depIndex": 22, "warnings": [], "versioning": "pep440", "sourceUrl": "https://github.com/ifduyue/python-xxhash", "fixedVersion": "3.0.0", "updates": [] } ] } ] } } DEBUG: processRepo() DEBUG: Processing 2 branches: renovate/datasets-1.x, renovate/datasets-2.x DEBUG: Calculating hourly PRs remaining DEBUG: getPrList success { "pullsTotal": 1, "requestsTotal": 1, "apiQuotaAffected": true } DEBUG: currentHourStart=2022-08-31T02:00:00.000+00:00 DEBUG: PR hourly limit remaining: 2 DEBUG: Calculating prConcurrentLimit (10) DEBUG: getBranchPr(renovate/datasets-1.x) DEBUG: findPr(renovate/datasets-1.x, undefined, open) DEBUG: Found PR #2 DEBUG: getBranchPr(renovate/datasets-2.x) DEBUG: findPr(renovate/datasets-2.x, undefined, open) DEBUG: findPr(renovate/datasets-2.x, undefined, closed) DEBUG: 1 PRs are currently open DEBUG: PR concurrent limit remaining: 9 DEBUG: Calculated maximum PRs remaining this run { "prsRemaining": 2 } DEBUG: PullRequests limit = 2 DEBUG: Calculating hourly PRs remaining DEBUG: currentHourStart=2022-08-31T02:00:00.000+00:00 DEBUG: PR hourly limit remaining: 2 DEBUG: Calculating branchConcurrentLimit (10) DEBUG: 1 already existing branches found: renovate/datasets-1.x DEBUG: Branch concurrent limit remaining: 9 DEBUG: Calculated maximum branches remaining this run { "branchesRemaining": 2 } DEBUG: Branches limit = 2 DEBUG: Branch fingerprint is unchanged, updates check can be skipped(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: getBranchPr(renovate/datasets-1.x)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: findPr(renovate/datasets-1.x, undefined, open)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Found PR #2(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: branchExists=true(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: dependencyDashboardCheck=undefined(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: PR rebase requested=false(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Checking if PR has been edited(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Found existing branch PR(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Checking schedule(after 10pm every weekday,before 4am every weekday, America/Chihuahua)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Found timezone(branch="renovate/datasets-1.x") { "baseBranch": "main", "timezone": "America/Chihuahua" } DEBUG: Adjusting now for timezone(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Checking 2 schedule(s)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Checking schedule "after 10pm every weekday"(branch="renovate/datasets-1.x") { "baseBranch": "main", "parsedSchedule": { "schedules": [ { "t_a": [ 79200 ], "d": [ 2, 3, 4, 5, 6 ] } ], "exceptions": [], "error": -1 } } DEBUG: Checking schedule "before 4am every weekday"(branch="renovate/datasets-1.x") { "baseBranch": "main", "parsedSchedule": { "schedules": [ { "t_b": [ 14400 ], "d": [ 2, 3, 4, 5, 6 ] } ], "exceptions": [], "error": -1 } } DEBUG: Package not scheduled(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Branch + PR exists but is not scheduled -- will update if necessary(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Branch already exists(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: getBranchPr(renovate/datasets-1.x)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: findPr(renovate/datasets-1.x, undefined, open)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Found PR #2(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Branch is behind base branch and needs rebasing(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Branch is unmodified, so can be rebased(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Using reuseExistingBranch: false(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Setting current branch to main(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Initializing git repository into /mnt/renovate/gh/corypaik/renovate-loop(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Performing blobless clone(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: git clone completed(branch="renovate/datasets-1.x") { "baseBranch": "main", "durationMs": 1449 } DEBUG: latest repository commit(branch="renovate/datasets-1.x") { "baseBranch": "main", "latestCommit": { "hash": "5267848cc1c6b6dfb1d031f573b25d2ab36ebc82", "date": "2022-08-30T20:06:50-06:00", "message": "ci: always run pip lockfile step", "refs": "HEAD -> main, origin/main, origin/HEAD", "body": "", "author_name": "Cory Paik", "author_email": "corypaik@gmail.com" } } DEBUG: latest commit(branch="renovate/datasets-1.x") { "baseBranch": "main", "branchName": "main", "latestCommitDate": "2022-08-30T20:06:50-06:00" } DEBUG: manager.getUpdatedPackageFiles() reuseExistingBranch=false(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Starting search at index 527(packageFile="third_party/requirements.txt", branch="renovate/datasets-1.x") { "baseBranch": "main", "depName": "datasets" } DEBUG: Found match at index 527(packageFile="third_party/requirements.txt", branch="renovate/datasets-1.x") { "baseBranch": "main", "depName": "datasets" } DEBUG: Contents updated(packageFile="third_party/requirements.txt", branch="renovate/datasets-1.x") { "baseBranch": "main", "depName": "datasets" } DEBUG: Starting search at index 8(packageFile="third_party/requirements.in", branch="renovate/datasets-1.x") { "baseBranch": "main", "depName": "datasets" } DEBUG: Found match at index 8(packageFile="third_party/requirements.in", branch="renovate/datasets-1.x") { "baseBranch": "main", "depName": "datasets" } DEBUG: Contents updated(packageFile="third_party/requirements.in", branch="renovate/datasets-1.x") { "baseBranch": "main", "depName": "datasets" } DEBUG: pip_requirements.updateArtifacts(third_party/requirements.txt)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Setting BUILDPACK_CACHE_DIR to /tmp/containerbase(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Using docker to execute(branch="renovate/datasets-1.x") { "baseBranch": "main", "image": "sidecar" } DEBUG: Resolved stable matching version(branch="renovate/datasets-1.x") { "baseBranch": "main", "toolName": "python", "resolvedVersion": "3.10.6" } DEBUG: containerbaseDir is separate from cacheDir(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: No tag or tagConstraint specified(branch="renovate/datasets-1.x") { "baseBranch": "main", "image": "docker.io/renovate/sidecar" } DEBUG: Fetching Docker image: docker.io/renovate/sidecar(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Finished fetching Docker image docker.io/renovate/sidecar@sha256:ee8108a5a17ef370f15da2c11bf7b2da9035153fde1587118ac9cc5c6c56cf0a(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Executing command(branch="renovate/datasets-1.x") { "baseBranch": "main", "command": "docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/corypaik/renovate-loop\":\"/mnt/renovate/gh/corypaik/renovate-loop\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -w \"/mnt/renovate/gh/corypaik/renovate-loop\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.10.6 && pip install --user hashin && hashin datasets==1.18.4 -r third_party/requirements.txt\"" } DEBUG: exec completed(branch="renovate/datasets-1.x") { "baseBranch": "main", "cmd": "docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/corypaik/renovate-loop\":\"/mnt/renovate/gh/corypaik/renovate-loop\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -w \"/mnt/renovate/gh/corypaik/renovate-loop\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.10.6 && pip install --user hashin && hashin datasets==1.18.4 -r third_party/requirements.txt\"", "durationMs": 27296, "stdout": "installing v2 tool python v3.10.6\nUsing prebuild python\nRequirement already satisfied: pip in /opt/buildpack/tools/python/3.10.6/lib/python3.10/site-packages (22.2.1)\nCollecting pip\n Downloading pip-22.2.2-py3-none-any.whl (2.0 MB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.0/2.0 MB 21.0 MB/s eta 0:00:00\nInstalling collected packages: pip\n Attempting uninstall: pip\n Found existing installation: pip 22.2.1\n Uninstalling pip-22.2.1:\n Successfully uninstalled pip-22.2.1\nSuccessfully installed pip-22.2.2\nFiles removed: 2\nlinking tool python v3.10.6\nPython 3.10.6\npip 22.2.2 from /opt/buildpack/tools/python/3.10.6/lib/python3.10/site-packages/pip (python 3.10)\nInstalled v2 /usr/local/buildpack/tools/v2/python.sh in 20 seconds\nCollecting hashin\n Downloading hashin-0.17.0-py2.py3-none-any.whl (16 kB)\nCollecting pip-api\n Downloading pip_api-0.0.30-py3-none-any.whl (111 kB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 111.6/111.6 kB 5.9 MB/s eta 0:00:00\nCollecting packaging\n Downloading packaging-21.3-py3-none-any.whl (40 kB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 40.8/40.8 kB 4.7 MB/s eta 0:00:00\nCollecting pyparsing!=3.0.5,>=2.0.2\n Downloading pyparsing-3.0.9-py3-none-any.whl (98 kB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 98.3/98.3 kB 11.1 MB/s eta 0:00:00\nRequirement already satisfied: pip in /opt/buildpack/tools/python/3.10.6/lib/python3.10/site-packages (from pip-api->hashin) (22.2.2)\nInstalling collected packages: pyparsing, pip-api, packaging, hashin\nSuccessfully installed hashin-0.17.0 packaging-21.3 pip-api-0.0.30 pyparsing-3.0.9\n", "stderr": "Found file in cache: /tmp/containerbase/f79e18f004c2278a4c76a549ca9e8beed6ee66d17e187e79a230ba06836df191/python-3.10.6-focal-x86_64.tar.xz\n" } DEBUG: Returning updated third_party/requirements.txt(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: pip_requirements.updateArtifacts(third_party/requirements.in)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: No hashin commands to run - returning(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Updated 2 package files(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Updated 1 lock files(branch="renovate/datasets-1.x") { "baseBranch": "main", "updatedArtifacts": [ "third_party/requirements.txt" ] } DEBUG: Getting comments for #2(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Found 0 comments(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: isBranchConflicted(main, renovate/datasets-1.x)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Using cached result false for isBranchConflicted(main, renovate/datasets-1.x)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: 2 file(s) to commit(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Preparing files for committing to branch renovate/datasets-1.x(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Setting git author name(branch="renovate/datasets-1.x") { "baseBranch": "main", "gitAuthorName": "Renovate Bot" } DEBUG: Setting git author email(branch="renovate/datasets-1.x") { "baseBranch": "main", "gitAuthorEmail": "bot@renovateapp.com" } DEBUG: git commit(branch="renovate/datasets-1.x") { "baseBranch": "main", "deletedFiles": [], "ignoredFiles": [], "result": { "author": null, "branch": "renovate/datasets-1.x", "commit": "210146775398f010baed48f2e6891d780406efa2", "root": false, "summary": { "changes": 2, "insertions": 4, "deletions": 4 } } } DEBUG: resetToCommit(5267848cc1c6b6dfb1d031f573b25d2ab36ebc82)(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Fetching branch renovate/datasets-1.x(branch="renovate/datasets-1.x") { "baseBranch": "main" } INFO: Branch updated(branch="renovate/datasets-1.x") { "baseBranch": "main", "commitSha": "78dae2cf45f995142afb48597777d62fa9677a1f" } DEBUG: Updating renovate/stability-days status check state to green(branch="renovate/datasets-1.x") { "baseBranch": "main" } DEBUG: Setting branch status(branch="renovate/datasets-1.x") { "baseBranch": "main", "context": "renovate/stability-days", "state": "green" } DEBUG: Branch status pending(branch="renovate/datasets-1.x") { "baseBranch": "main", "commitSha": "78dae2cf45f995142afb48597777d62fa9677a1f" } DEBUG: getBranchPr(renovate/datasets-2.x)(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: findPr(renovate/datasets-2.x, undefined, open)(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: findPr(renovate/datasets-2.x, undefined, closed)(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: branchExists=false(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: dependencyDashboardCheck=undefined(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: recreateClosed is false(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: findPr(renovate/datasets-2.x, build: update dependency datasets to v2, !open)(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: prAlreadyExisted=false(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: Checking schedule(after 10pm every weekday,before 4am every weekday, America/Chihuahua)(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: Found timezone(branch="renovate/datasets-2.x") { "baseBranch": "main", "timezone": "America/Chihuahua" } DEBUG: Adjusting now for timezone(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: Checking 2 schedule(s)(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: Checking schedule "after 10pm every weekday"(branch="renovate/datasets-2.x") { "baseBranch": "main", "parsedSchedule": { "schedules": [ { "t_a": [ 79200 ], "d": [ 2, 3, 4, 5, 6 ] } ], "exceptions": [], "error": -1 } } DEBUG: Checking schedule "before 4am every weekday"(branch="renovate/datasets-2.x") { "baseBranch": "main", "parsedSchedule": { "schedules": [ { "t_b": [ 14400 ], "d": [ 2, 3, 4, 5, 6 ] } ], "exceptions": [], "error": -1 } } DEBUG: Package not scheduled(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: Skipping branch creation as not within schedule(branch="renovate/datasets-2.x") { "baseBranch": "main" } DEBUG: getBranchPr(renovate/datasets-1.x) DEBUG: findPr(renovate/datasets-1.x, undefined, open) DEBUG: Found PR #2 DEBUG: Ensuring Dependency Dashboard DEBUG: ensureIssue(Dependency Dashboard) DEBUG: Patching issue DEBUG: Issue updated DEBUG: Removing any stale branches DEBUG: config.repoIsOnboarded=true DEBUG: Branch lists { "branchList": [ "renovate/datasets-1.x", "renovate/datasets-2.x" ], "renovateBranches": [ "renovate/datasets-1.x" ] } DEBUG: remainingBranches= DEBUG: No branches to clean up DEBUG: Cleaning up Renovate refs: refs/renovate/* DEBUG: PackageFiles.clear() - Package files deleted { "baseBranches": [ "main" ] } DEBUG: Renovate repository PR statistics { "stats": { "total": 1, "open": 1, "closed": 0, "merged": 0 } } DEBUG: Repository result: done, status: onboarded, enabled: true, onboarded: true DEBUG: Repository timing splits (milliseconds) { "splits": { "init": 3599, "extract": 494, "lookup": 2772, "onboarding": 0, "update": 35845 }, "total": 45597 } DEBUG: http statistics { "urls": { "https://api.github.com/graphql (POST,200)": 2, "https://api.github.com/repos/corypaik/renovate-loop/commits/78dae2cf45f995142afb48597777d62fa9677a1f/statuses (GET,200)": 2, "https://api.github.com/repos/corypaik/renovate-loop/commits/renovate/datasets-1.x/status (GET,200)": 1, "https://api.github.com/repos/corypaik/renovate-loop/contents/renovate.json (GET,200)": 1, "https://api.github.com/repos/corypaik/renovate-loop/git/commits (POST,201)": 1, "https://api.github.com/repos/corypaik/renovate-loop/git/refs/heads/renovate/datasets-1.x (PATCH,200)": 1, "https://api.github.com/repos/corypaik/renovate-loop/git/trees (POST,201)": 1, "https://api.github.com/repos/corypaik/renovate-loop/issues/1 (GET,200)": 2, "https://api.github.com/repos/corypaik/renovate-loop/issues/1 (PATCH,200)": 1, "https://api.github.com/repos/corypaik/renovate-loop/issues/2/comments (GET,200)": 1, "https://api.github.com/repos/corypaik/renovate-loop/pulls (GET,200)": 1, "https://api.github.com/repos/corypaik/renovate-loop/statuses/78dae2cf45f995142afb48597777d62fa9677a1f (POST,201)": 1, "https://api.github.com/repos/whitesource/merge-confidence/contents/beta.json (GET,200)": 1, "https://pypi.org/pypi/certifi/json (GET,200)": 1 }, "hostStats": { "api.github.com": { "requestCount": 16, "requestAvgMs": 269, "queueAvgMs": 0 }, "pypi.org": { "requestCount": 1, "requestAvgMs": 989, "queueAvgMs": 0 } }, "totalRequests": 17 } INFO: Repository finished { "durationMs": 45597 } ``` </details> ### Have you created a minimal reproduction repository? I have linked to a minimal reproduction repository in the bug description
priority
renovate is stuck in an infinite loop overwriting a ci bot s commits how are you running renovate mend renovate hosted app on github com if you re self hosting renovate tell us what version of renovate you run no response if you re self hosting renovate select which platform you are using no response if you re self hosting renovate tell us what version of the platform you run no response was this something which used to work for you and then stopped it used to work and then stopped describe the bug renovate is stuck in an infinite loop overwriting a ci bot s fixup commits this started happening sometime in the past months the configuration has not been changed since so far i have only observed this with python packages but i do not have a similar procedure for other dependencies the process is something like this renovate updates package x and commits to the repository which triggers the first ci run the repository uses lock files with pip compile so before running builds and tests the ci agent runs a script to update the lockfiles this takes care of resolving dependency changes of x if any changes were made the ci bot commits them as a fixup commit git commit fixup head and pushes it to github the new commit triggers a second ci run this time the check should pass and move on the ci bot is registered as a gitignoredauthor but in the past renovate would only overwrite the changes if the main branch was updated i have rebasewhen as behind base branch currently every time the ci bot commits a change renovate seems to overwrite it the next time it runs this means the repository update process is effectively stuck in an infinite loop i m not sure exactly what triggers this behavior and it s been a while since it worked but it seems to have been around the same time renovate changed email names from bot renovateapp com to renovate users noreply github com i recall having to manually rebase the existing branches as they were marked as modified i ve provided debug logs from a minimal reproduction repository which is publicly available on the repository contains one example package that causes this issue all others have been removed or disabled there is also a public available for reference relevant debug logs logs debug no dangling containers to remove info repository started renovateversion debug using localdir mnt renovate gh corypaik renovate loop debug packagefiles clear package files deleted basebranches debug resetmemcache debug initrepo corypaik renovate loop debug corypaik renovate loop default branch main debug using app token for git init debug repository cache is restored from revision debug resetting npmrc debug detectsemanticcommits debug checkonboarding debug isonboarded debug checking cached config file name debug existing config file confirmed debug repository config filename renovate json config schema extends ignoremodulesandtests group monorepos group recommended workarounds all semanticcommits enabled semanticcommittype build semanticcommitscope prhourlylimit prconcurrentlimit dependencydashboard true labels target minor comp build ci action review timezone america chihuahua schedule after every weekday before every weekday basebranches main rebasewhen behind base branch pip requirements filematch requirements in prcreation not pending stabilitydays gitignoredauthors buildkite bot buildkite com pre commit enabled true packagerules description always pin direct python dependencies to ensure lockfile updates do not revert to a different version matchfiles third party requirements in rangestrategy pin description all python deps except datasets for minimal reproduction matchpackagepatterns matchupdatetypes major minor patch matchlanguages python excludepackagenames datasets enabled false debug repo is onboarded debug migrateandvalidate debug no config migration necessary debug massaged config config schema extends github whitesource merge confidence beta ignoremodulesandtests group monorepos group recommended workarounds all semanticcommits enabled semanticcommittype build semanticcommitscope prhourlylimit prconcurrentlimit dependencydashboard true labels target minor comp build ci action review timezone america chihuahua schedule after every weekday before every weekday basebranches main rebasewhen behind base branch pip requirements filematch requirements in prcreation not pending stabilitydays gitignoredauthors buildkite bot buildkite com pre commit enabled true packagerules description always pin direct python dependencies to ensure lockfile updates do not revert to a different version matchfiles third party requirements in rangestrategy pin description all python deps except datasets for minimal reproduction matchpackagepatterns matchupdatetypes major minor patch matchlanguages python excludepackagenames datasets enabled false debug migrated config config schema extends github whitesource merge confidence beta ignoremodulesandtests group monorepos group recommended workarounds all semanticcommits enabled semanticcommittype build semanticcommitscope prhourlylimit prconcurrentlimit dependencydashboard true labels target minor comp build ci action review timezone america chihuahua schedule after every weekday before every weekday basebranches main rebasewhen behind base branch pip requirements filematch requirements in prcreation not pending stabilitydays gitignoredauthors buildkite bot buildkite com pre commit enabled true packagerules description always pin direct python dependencies to ensure lockfile updates do not revert to a different version matchfiles third party requirements in rangestrategy pin description all python deps except datasets for minimal reproduction matchpackagepatterns matchupdatetypes major minor patch matchlanguages python excludepackagenames datasets enabled false debug setting hostrules from config debug found repo ignorepaths ignorepaths node modules bower components vendor examples tests test tests fixtures debug no vulnerability alerts found debug no vulnerability alerts found debug findissue dependency dashboard debug retrieving issuelist debug retrieved issues debug found issue debug basebranches basebranches main debug basebranch main basebranch main debug extract basebranch main debug found cached extract basebranch main basebranchsha debug deleted cached dep updates basebranch main info dependency extraction complete basebranch main stats managers bazel filecount depcount pip requirements filecount depcount total filecount depcount debug basebranch main basebranch main debug packagefiles add package file saved for branch basebranch main debug package releases lookups complete basebranch main debug branchifyupgrades basebranch main debug flattened updates found datasets datasets datasets datasets basebranch main debug returning branch es basebranch main debug config repoisonboarded true basebranch main debug packagefiles with updates basebranch main config bazel packagefile workspace deps datasource github tags deptype http archive depname rules python packagename bazelbuild rules python currentvalue refs managerdata def http archive n name rules python n n strip prefix rules python n url depindex warnings versioning semver skipreason invalid value updates pip requirements packagefile third party requirements in deps depname datasets currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate datasets x bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate datasets x packagefile third party requirements txt deps depname certifi currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate certifi x depname charset normalizer currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate charset normalizer x depname datasets currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate datasets x bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate datasets x depname dill currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype patch isrange true branchname renovate dill x depname filelock currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate filelock x depname fsspec currentvalue datasource pypi currentversion depindex warnings versioning sourceurl changelogurl issingleversion true fixedversion updates bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate fsspec x depname huggingface hub currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate huggingface hub x depname idna currentvalue datasource pypi currentversion depindex warnings versioning sourceurl fixedversion updates depname multiprocess currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype patch isrange true branchname renovate multiprocess x depname numpy currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate numpy x depname packaging currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate packaging x depname pandas currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype patch isrange true branchname renovate pandas x depname pyarrow currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage issingleversion true fixedversion updates bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate pyarrow x depname pyparsing currentvalue datasource pypi currentversion depindex warnings versioning issingleversion true fixedversion updates bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate pyparsing x depname python dateutil currentvalue datasource pypi currentversion depindex warnings versioning sourceurl fixedversion updates depname pytz currentvalue datasource pypi currentversion depindex warnings versioning homepage issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate pytz x depname pyyaml currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage issingleversion true fixedversion updates bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate pyyaml x depname requests currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage changelogurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate requests x depname six currentvalue datasource pypi currentversion depindex warnings versioning sourceurl fixedversion updates depname tqdm currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage changelogurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate tqdm x depname typing extensions currentvalue datasource pypi currentversion depindex warnings versioning sourceurl changelogurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate typing extensions x depname currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype patch isrange true branchname renovate x depname xxhash currentvalue datasource pypi currentversion depindex warnings versioning sourceurl fixedversion updates debug processrepo debug processing branches renovate datasets x renovate datasets x debug calculating hourly prs remaining debug getprlist success pullstotal requeststotal apiquotaaffected true debug currenthourstart debug pr hourly limit remaining debug calculating prconcurrentlimit debug getbranchpr renovate datasets x debug findpr renovate datasets x undefined open debug found pr debug getbranchpr renovate datasets x debug findpr renovate datasets x undefined open debug findpr renovate datasets x undefined closed debug prs are currently open debug pr concurrent limit remaining debug calculated maximum prs remaining this run prsremaining debug pullrequests limit debug calculating hourly prs remaining debug currenthourstart debug pr hourly limit remaining debug calculating branchconcurrentlimit debug already existing branches found renovate datasets x debug branch concurrent limit remaining debug calculated maximum branches remaining this run branchesremaining debug branches limit debug branch fingerprint is unchanged updates check can be skipped branch renovate datasets x basebranch main debug getbranchpr renovate datasets x branch renovate datasets x basebranch main debug findpr renovate datasets x undefined open branch renovate datasets x basebranch main debug found pr branch renovate datasets x basebranch main debug branchexists true branch renovate datasets x basebranch main debug dependencydashboardcheck undefined branch renovate datasets x basebranch main debug pr rebase requested false branch renovate datasets x basebranch main debug checking if pr has been edited branch renovate datasets x basebranch main debug found existing branch pr branch renovate datasets x basebranch main debug checking schedule after every weekday before every weekday america chihuahua branch renovate datasets x basebranch main debug found timezone branch renovate datasets x basebranch main timezone america chihuahua debug adjusting now for timezone branch renovate datasets x basebranch main debug checking schedule s branch renovate datasets x basebranch main debug checking schedule after every weekday branch renovate datasets x basebranch main parsedschedule schedules t a d exceptions error debug checking schedule before every weekday branch renovate datasets x basebranch main parsedschedule schedules t b d exceptions error debug package not scheduled branch renovate datasets x basebranch main debug branch pr exists but is not scheduled will update if necessary branch renovate datasets x basebranch main debug branch already exists branch renovate datasets x basebranch main debug getbranchpr renovate datasets x branch renovate datasets x basebranch main debug findpr renovate datasets x undefined open branch renovate datasets x basebranch main debug found pr branch renovate datasets x basebranch main debug branch is behind base branch and needs rebasing branch renovate datasets x basebranch main debug branch is unmodified so can be rebased branch renovate datasets x basebranch main debug using reuseexistingbranch false branch renovate datasets x basebranch main debug setting current branch to main branch renovate datasets x basebranch main debug initializing git repository into mnt renovate gh corypaik renovate loop branch renovate datasets x basebranch main debug performing blobless clone branch renovate datasets x basebranch main debug git clone completed branch renovate datasets x basebranch main durationms debug latest repository commit branch renovate datasets x basebranch main latestcommit hash date message ci always run pip lockfile step refs head main origin main origin head body author name cory paik author email corypaik gmail com debug latest commit branch renovate datasets x basebranch main branchname main latestcommitdate debug manager getupdatedpackagefiles reuseexistingbranch false branch renovate datasets x basebranch main debug starting search at index packagefile third party requirements txt branch renovate datasets x basebranch main depname datasets debug found match at index packagefile third party requirements txt branch renovate datasets x basebranch main depname datasets debug contents updated packagefile third party requirements txt branch renovate datasets x basebranch main depname datasets debug starting search at index packagefile third party requirements in branch renovate datasets x basebranch main depname datasets debug found match at index packagefile third party requirements in branch renovate datasets x basebranch main depname datasets debug contents updated packagefile third party requirements in branch renovate datasets x basebranch main depname datasets debug pip requirements updateartifacts third party requirements txt branch renovate datasets x basebranch main debug setting buildpack cache dir to tmp containerbase branch renovate datasets x basebranch main debug using docker to execute branch renovate datasets x basebranch main image sidecar debug resolved stable matching version branch renovate datasets x basebranch main toolname python resolvedversion debug containerbasedir is separate from cachedir branch renovate datasets x basebranch main debug no tag or tagconstraint specified branch renovate datasets x basebranch main image docker io renovate sidecar debug fetching docker image docker io renovate sidecar branch renovate datasets x basebranch main debug finished fetching docker image docker io renovate sidecar branch renovate datasets x basebranch main debug executing command branch renovate datasets x basebranch main command docker run rm name renovate sidecar label renovate child v mnt renovate gh corypaik renovate loop mnt renovate gh corypaik renovate loop v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e pip cache dir e buildpack cache dir w mnt renovate gh corypaik renovate loop docker io renovate sidecar bash l c install tool python pip install user hashin hashin datasets r third party requirements txt debug exec completed branch renovate datasets x basebranch main cmd docker run rm name renovate sidecar label renovate child v mnt renovate gh corypaik renovate loop mnt renovate gh corypaik renovate loop v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e pip cache dir e buildpack cache dir w mnt renovate gh corypaik renovate loop docker io renovate sidecar bash l c install tool python pip install user hashin hashin datasets r third party requirements txt durationms stdout installing tool python nusing prebuild python nrequirement already satisfied pip in opt buildpack tools python lib site packages ncollecting pip n downloading pip none any whl mb n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ mb mb s eta ninstalling collected packages pip n attempting uninstall pip n found existing installation pip n uninstalling pip n successfully uninstalled pip nsuccessfully installed pip nfiles removed nlinking tool python npython npip from opt buildpack tools python lib site packages pip python ninstalled usr local buildpack tools python sh in seconds ncollecting hashin n downloading hashin none any whl kb ncollecting pip api n downloading pip api none any whl kb n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ kb mb s eta ncollecting packaging n downloading packaging none any whl kb n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ kb mb s eta ncollecting pyparsing n downloading pyparsing none any whl kb n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ kb mb s eta nrequirement already satisfied pip in opt buildpack tools python lib site packages from pip api hashin ninstalling collected packages pyparsing pip api packaging hashin nsuccessfully installed hashin packaging pip api pyparsing n stderr found file in cache tmp containerbase python focal tar xz n debug returning updated third party requirements txt branch renovate datasets x basebranch main debug pip requirements updateartifacts third party requirements in branch renovate datasets x basebranch main debug no hashin commands to run returning branch renovate datasets x basebranch main debug updated package files branch renovate datasets x basebranch main debug updated lock files branch renovate datasets x basebranch main updatedartifacts third party requirements txt debug getting comments for branch renovate datasets x basebranch main debug found comments branch renovate datasets x basebranch main debug isbranchconflicted main renovate datasets x branch renovate datasets x basebranch main debug using cached result false for isbranchconflicted main renovate datasets x branch renovate datasets x basebranch main debug file s to commit branch renovate datasets x basebranch main debug preparing files for committing to branch renovate datasets x branch renovate datasets x basebranch main debug setting git author name branch renovate datasets x basebranch main gitauthorname renovate bot debug setting git author email branch renovate datasets x basebranch main gitauthoremail bot renovateapp com debug git commit branch renovate datasets x basebranch main deletedfiles ignoredfiles result author null branch renovate datasets x commit root false summary changes insertions deletions debug resettocommit branch renovate datasets x basebranch main debug fetching branch renovate datasets x branch renovate datasets x basebranch main info branch updated branch renovate datasets x basebranch main commitsha debug updating renovate stability days status check state to green branch renovate datasets x basebranch main debug setting branch status branch renovate datasets x basebranch main context renovate stability days state green debug branch status pending branch renovate datasets x basebranch main commitsha debug getbranchpr renovate datasets x branch renovate datasets x basebranch main debug findpr renovate datasets x undefined open branch renovate datasets x basebranch main debug findpr renovate datasets x undefined closed branch renovate datasets x basebranch main debug branchexists false branch renovate datasets x basebranch main debug dependencydashboardcheck undefined branch renovate datasets x basebranch main debug recreateclosed is false branch renovate datasets x basebranch main debug findpr renovate datasets x build update dependency datasets to open branch renovate datasets x basebranch main debug pralreadyexisted false branch renovate datasets x basebranch main debug checking schedule after every weekday before every weekday america chihuahua branch renovate datasets x basebranch main debug found timezone branch renovate datasets x basebranch main timezone america chihuahua debug adjusting now for timezone branch renovate datasets x basebranch main debug checking schedule s branch renovate datasets x basebranch main debug checking schedule after every weekday branch renovate datasets x basebranch main parsedschedule schedules t a d exceptions error debug checking schedule before every weekday branch renovate datasets x basebranch main parsedschedule schedules t b d exceptions error debug package not scheduled branch renovate datasets x basebranch main debug skipping branch creation as not within schedule branch renovate datasets x basebranch main debug getbranchpr renovate datasets x debug findpr renovate datasets x undefined open debug found pr debug ensuring dependency dashboard debug ensureissue dependency dashboard debug patching issue debug issue updated debug removing any stale branches debug config repoisonboarded true debug branch lists branchlist renovate datasets x renovate datasets x renovatebranches renovate datasets x debug remainingbranches debug no branches to clean up debug cleaning up renovate refs refs renovate debug packagefiles clear package files deleted basebranches main debug renovate repository pr statistics stats total open closed merged debug repository result done status onboarded enabled true onboarded true debug repository timing splits milliseconds splits init extract lookup onboarding update total debug http statistics urls post get get get post patch post get patch get get post get get hoststats api github com requestcount requestavgms queueavgms pypi org requestcount requestavgms queueavgms totalrequests info repository finished durationms have you created a minimal reproduction repository i have linked to a minimal reproduction repository in the bug description
1
180,393
6,649,277,897
IssuesEvent
2017-09-28 12:41:23
herbiehp/unicenta
https://api.github.com/repos/herbiehp/unicenta
closed
Modify Customer Form
3-Medium Priority enhancement help wanted
Modify customer form to add additional attributes that are check boxes (like visible) and add columns to the customers table to store them. To make it generic and usable by any business the names displayed in the GUI (and field names) should be something like CustAttr1, CustAttr2, etc with the ability to change the name displayed in the GUI with a Resources file. A use case is for an organization that offers memberships to its customers. Example A country club that runs golf, curling and tennis operations. It would be useful for them to track customer memberships and the types in uniCenta. Some types could be: - Golf Membership - Full - Golf Membership - Weekdays - Curling Membership - Full - Curling Membership - 1 night per week - Tennis Membership - Full I am willing to take this on but don't know where to start on this yet so I may need some help. Thx, TekMason
1.0
Modify Customer Form - Modify customer form to add additional attributes that are check boxes (like visible) and add columns to the customers table to store them. To make it generic and usable by any business the names displayed in the GUI (and field names) should be something like CustAttr1, CustAttr2, etc with the ability to change the name displayed in the GUI with a Resources file. A use case is for an organization that offers memberships to its customers. Example A country club that runs golf, curling and tennis operations. It would be useful for them to track customer memberships and the types in uniCenta. Some types could be: - Golf Membership - Full - Golf Membership - Weekdays - Curling Membership - Full - Curling Membership - 1 night per week - Tennis Membership - Full I am willing to take this on but don't know where to start on this yet so I may need some help. Thx, TekMason
priority
modify customer form modify customer form to add additional attributes that are check boxes like visible and add columns to the customers table to store them to make it generic and usable by any business the names displayed in the gui and field names should be something like etc with the ability to change the name displayed in the gui with a resources file a use case is for an organization that offers memberships to its customers example a country club that runs golf curling and tennis operations it would be useful for them to track customer memberships and the types in unicenta some types could be golf membership full golf membership weekdays curling membership full curling membership night per week tennis membership full i am willing to take this on but don t know where to start on this yet so i may need some help thx tekmason
1
559,360
16,557,057,288
IssuesEvent
2021-05-28 15:02:32
guardicore/monkey
https://api.github.com/repos/guardicore/monkey
closed
Configure MongoDB on Monkey Island initialization
Complexity: Medium Enhancement Priority: High python
Monkey Island needs to write runtime artifacts to a writable location and assume that the source code directory is read-only. Currently, MongoDB is started by the `linux/run.sh`, `windows\run_mongodb.bat`, and `appimage/run_appimage.sh`. These scripts do not have access to the `data_dir` property in `server_config.json`, and giving them access would add even more redundant functionality to all of these scripts. - [x] Add a `"mongodb"` section to the `server_config.json` with a `"start_mongodb"` boolean. (0) - @VakarisZ ```json { "log_level": "debug", "data_dir": "$HOME/.monkey_island", "environment": { "server_config": "password", "deployment": "development", "user": "my_user", "password_hash": "DEADBEEF" } "mongodb" { "start_mongodb": true } } ``` - Make sure this section exists and is correct for the default `server_config.json`, the docker `server_config.json`, the windows installer `server_config.json`, and the AppImage `server_config.json` - [x] If `"start_mongodb": true`, then monkey_island.py will start the mongodb process and set it's `--dbpath` argument to `data_dir/db`. (0d) - @VakarisZ - [x] Refactor the data dir folder creation and permission setting into a separate function and re-use it to create db dir with relevant permissions as well (0d). - [x] Refactor mongodb initialization to be decoupled from environment config (0d) - @VakarisZ - [x] Remove the code that starts mongodb from `appimage/run_appimage.sh`. If this script no longer serves any purpose, remove it and modify the AppImage to run without it. (0d) - @VakarisZ - [x] Remove `linux/run.sh`. Modify documentation with new instructions on how to start monkey island. (0d) - @VakarisZ - [x] Remove code that creates db dir from deployment scripts, mongodb install scripts, docs, etc. (0) - @VakarisZ - [x] Remove the code that starts mongodb from `monkey_island/windows/*`. If any of these scripts no longer serves a purpose, remove them.(0d) - @VakarisZ - [x] Ensure the mongodb process is properly cleaned up when Monkey Island exits (0d) - [x] Review PR and fix (0d)
1.0
Configure MongoDB on Monkey Island initialization - Monkey Island needs to write runtime artifacts to a writable location and assume that the source code directory is read-only. Currently, MongoDB is started by the `linux/run.sh`, `windows\run_mongodb.bat`, and `appimage/run_appimage.sh`. These scripts do not have access to the `data_dir` property in `server_config.json`, and giving them access would add even more redundant functionality to all of these scripts. - [x] Add a `"mongodb"` section to the `server_config.json` with a `"start_mongodb"` boolean. (0) - @VakarisZ ```json { "log_level": "debug", "data_dir": "$HOME/.monkey_island", "environment": { "server_config": "password", "deployment": "development", "user": "my_user", "password_hash": "DEADBEEF" } "mongodb" { "start_mongodb": true } } ``` - Make sure this section exists and is correct for the default `server_config.json`, the docker `server_config.json`, the windows installer `server_config.json`, and the AppImage `server_config.json` - [x] If `"start_mongodb": true`, then monkey_island.py will start the mongodb process and set it's `--dbpath` argument to `data_dir/db`. (0d) - @VakarisZ - [x] Refactor the data dir folder creation and permission setting into a separate function and re-use it to create db dir with relevant permissions as well (0d). - [x] Refactor mongodb initialization to be decoupled from environment config (0d) - @VakarisZ - [x] Remove the code that starts mongodb from `appimage/run_appimage.sh`. If this script no longer serves any purpose, remove it and modify the AppImage to run without it. (0d) - @VakarisZ - [x] Remove `linux/run.sh`. Modify documentation with new instructions on how to start monkey island. (0d) - @VakarisZ - [x] Remove code that creates db dir from deployment scripts, mongodb install scripts, docs, etc. (0) - @VakarisZ - [x] Remove the code that starts mongodb from `monkey_island/windows/*`. If any of these scripts no longer serves a purpose, remove them.(0d) - @VakarisZ - [x] Ensure the mongodb process is properly cleaned up when Monkey Island exits (0d) - [x] Review PR and fix (0d)
priority
configure mongodb on monkey island initialization monkey island needs to write runtime artifacts to a writable location and assume that the source code directory is read only currently mongodb is started by the linux run sh windows run mongodb bat and appimage run appimage sh these scripts do not have access to the data dir property in server config json and giving them access would add even more redundant functionality to all of these scripts add a mongodb section to the server config json with a start mongodb boolean vakarisz json log level debug data dir home monkey island environment server config password deployment development user my user password hash deadbeef mongodb start mongodb true make sure this section exists and is correct for the default server config json the docker server config json the windows installer server config json and the appimage server config json if start mongodb true then monkey island py will start the mongodb process and set it s dbpath argument to data dir db vakarisz refactor the data dir folder creation and permission setting into a separate function and re use it to create db dir with relevant permissions as well refactor mongodb initialization to be decoupled from environment config vakarisz remove the code that starts mongodb from appimage run appimage sh if this script no longer serves any purpose remove it and modify the appimage to run without it vakarisz remove linux run sh modify documentation with new instructions on how to start monkey island vakarisz remove code that creates db dir from deployment scripts mongodb install scripts docs etc vakarisz remove the code that starts mongodb from monkey island windows if any of these scripts no longer serves a purpose remove them vakarisz ensure the mongodb process is properly cleaned up when monkey island exits review pr and fix
1
247,239
7,915,576,371
IssuesEvent
2018-07-04 00:08:31
facelessuser/pymdown-extensions
https://api.github.com/repos/facelessuser/pymdown-extensions
closed
SuperFences: Preserve Tabs and \r
Bug Priority - Medium Severity - Major
The preserve tabs feature is useful for preserving the tab character in code blocks, but it makes fences get processed before whitespace normalization. This means before `\r\n` is transformed to `\n`. On a Windows system, this can cause an issue as the content is scanned assuming normalization. We should strip trailing '\r' from lines to be safe.
1.0
SuperFences: Preserve Tabs and \r - The preserve tabs feature is useful for preserving the tab character in code blocks, but it makes fences get processed before whitespace normalization. This means before `\r\n` is transformed to `\n`. On a Windows system, this can cause an issue as the content is scanned assuming normalization. We should strip trailing '\r' from lines to be safe.
priority
superfences preserve tabs and r the preserve tabs feature is useful for preserving the tab character in code blocks but it makes fences get processed before whitespace normalization this means before r n is transformed to n on a windows system this can cause an issue as the content is scanned assuming normalization we should strip trailing r from lines to be safe
1
476,723
13,749,106,737
IssuesEvent
2020-10-06 10:00:03
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
closed
User group invitation tab invited message not showing
bug hacktoberfest priority: medium
**Describe the bug** When you go into the user profile groups tab and under that Invitation you will not show the invitation message. E.g https://learndash/members/mateo/groups/invites **To Reproduce** Steps to reproduce the behavior: 1. Invite member into your group with some invitation text. 2. Now login into the user you have invited. 3. Go to https://learndash/members/<member>/groups/invites 4. You will not show the invitation message. https://prnt.sc/unnnl7 **Expected behavior** We should show the invitation message. https://prnt.sc/unno80 **Screenshots** https://prnt.sc/unnnl7 **Support ticket links** If applicable, add HelpScout link or ticket number where the issue was originally reported.
1.0
User group invitation tab invited message not showing - **Describe the bug** When you go into the user profile groups tab and under that Invitation you will not show the invitation message. E.g https://learndash/members/mateo/groups/invites **To Reproduce** Steps to reproduce the behavior: 1. Invite member into your group with some invitation text. 2. Now login into the user you have invited. 3. Go to https://learndash/members/<member>/groups/invites 4. You will not show the invitation message. https://prnt.sc/unnnl7 **Expected behavior** We should show the invitation message. https://prnt.sc/unno80 **Screenshots** https://prnt.sc/unnnl7 **Support ticket links** If applicable, add HelpScout link or ticket number where the issue was originally reported.
priority
user group invitation tab invited message not showing describe the bug when you go into the user profile groups tab and under that invitation you will not show the invitation message e g to reproduce steps to reproduce the behavior invite member into your group with some invitation text now login into the user you have invited go to you will not show the invitation message expected behavior we should show the invitation message screenshots support ticket links if applicable add helpscout link or ticket number where the issue was originally reported
1
289,431
8,870,684,399
IssuesEvent
2019-01-11 10:13:44
georchestra/georchestra
https://api.github.com/repos/georchestra/georchestra
closed
CAS - Relicates of config.jar in the JSP files
2018 bug priority-medium
**Bug description** We can find some traces of the shared config here: https://github.com/georchestra/georchestra/blob/master/cas-server-webapp/src/main/webapp/WEB-INF/view/jsp/default/ui/includes/top.jsp#L38-L42 **geOrchestra version or branch** since 18.06, also present on master currently **Expected behavior** We should get rid of these shared.* variables previously defined in the old config.jar process. Maybe the concerned variables should end in the default.properties.
1.0
CAS - Relicates of config.jar in the JSP files - **Bug description** We can find some traces of the shared config here: https://github.com/georchestra/georchestra/blob/master/cas-server-webapp/src/main/webapp/WEB-INF/view/jsp/default/ui/includes/top.jsp#L38-L42 **geOrchestra version or branch** since 18.06, also present on master currently **Expected behavior** We should get rid of these shared.* variables previously defined in the old config.jar process. Maybe the concerned variables should end in the default.properties.
priority
cas relicates of config jar in the jsp files bug description we can find some traces of the shared config here georchestra version or branch since also present on master currently expected behavior we should get rid of these shared variables previously defined in the old config jar process maybe the concerned variables should end in the default properties
1
100,528
4,097,848,842
IssuesEvent
2016-06-03 04:41:54
Putaitu/mondai
https://api.github.com/repos/Putaitu/mondai
closed
Reloading resources doesn't work immediately
estimate:1h priority:medium type:bug version:0.1.3
This appears to be a GitHub issue, that there is an extended output cache on the API. Is there anything that can be done to circumvent it?
1.0
Reloading resources doesn't work immediately - This appears to be a GitHub issue, that there is an extended output cache on the API. Is there anything that can be done to circumvent it?
priority
reloading resources doesn t work immediately this appears to be a github issue that there is an extended output cache on the api is there anything that can be done to circumvent it
1
438,820
12,651,991,474
IssuesEvent
2020-06-17 02:09:13
minio/minio
https://api.github.com/repos/minio/minio
closed
"Invalid argument" errors due to setrlimit calls on macOS
community priority: medium
Hello! When using MinIO on macOS, we frequently see log lines that look like the following: ``` API: SYSTEM() Time: <time> Error: invalid argument 1: /Users/user/go/pkg/mod/github.com/minio/minio@v0.0.0-20180508161510-54cd29b51c38/cmd/gateway-main.go:166:cmd.StartGateway() ``` These are ignorable and everything seems to work, but still annoying. After a bunch of debugging, we tracked this down to how MinIO is trying to increase rlimits in https://github.com/minio/minio/blob/release/cmd/server-rlimit.go#L25-L55, and in https://github.com/minio/minio/tree/release/pkg/sys There is an open Go bug about how the Go syscall for getting the max limit is lying. https://github.com/golang/go/issues/30401 This changed between Go releases on macOS, and we believe MinIO is hitting this bug. Would it be useful to push up a code review to fix this? I think the solution is to extend github.com/minio/minio/pkg/sys to have special darwin-specific handling for not letting the max limit above 10240, as described in the Go issue. ## Expected Behavior No inscrutable logs about invalid arguments while MinIO starts and tries to set limits. ## Current Behavior MinIO logs "invalid argument" on macOS ## Possible Solution Extend github.com/minio/minio/pkg/sys to specifically not allow the max limit above 10240, as described in the Go issue about this (Go's syscall behavior on Darwin changed): https://github.com/golang/go/issues/30401 ## Steps to Reproduce (for bugs) 1. Run on macOS, look at logs ## Regression This is a problem introduced by changing Go standard libraries. ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used (`minio version`): latest * Environment name and version (e.g. nginx 1.9.1): Go 1.14 * Server type and version: Mac Mini * Operating System and version (`uname -a`): macOS High Sierra * Link to your project: https://www.storj.io :wave: Thanks so much!
1.0
"Invalid argument" errors due to setrlimit calls on macOS - Hello! When using MinIO on macOS, we frequently see log lines that look like the following: ``` API: SYSTEM() Time: <time> Error: invalid argument 1: /Users/user/go/pkg/mod/github.com/minio/minio@v0.0.0-20180508161510-54cd29b51c38/cmd/gateway-main.go:166:cmd.StartGateway() ``` These are ignorable and everything seems to work, but still annoying. After a bunch of debugging, we tracked this down to how MinIO is trying to increase rlimits in https://github.com/minio/minio/blob/release/cmd/server-rlimit.go#L25-L55, and in https://github.com/minio/minio/tree/release/pkg/sys There is an open Go bug about how the Go syscall for getting the max limit is lying. https://github.com/golang/go/issues/30401 This changed between Go releases on macOS, and we believe MinIO is hitting this bug. Would it be useful to push up a code review to fix this? I think the solution is to extend github.com/minio/minio/pkg/sys to have special darwin-specific handling for not letting the max limit above 10240, as described in the Go issue. ## Expected Behavior No inscrutable logs about invalid arguments while MinIO starts and tries to set limits. ## Current Behavior MinIO logs "invalid argument" on macOS ## Possible Solution Extend github.com/minio/minio/pkg/sys to specifically not allow the max limit above 10240, as described in the Go issue about this (Go's syscall behavior on Darwin changed): https://github.com/golang/go/issues/30401 ## Steps to Reproduce (for bugs) 1. Run on macOS, look at logs ## Regression This is a problem introduced by changing Go standard libraries. ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used (`minio version`): latest * Environment name and version (e.g. nginx 1.9.1): Go 1.14 * Server type and version: Mac Mini * Operating System and version (`uname -a`): macOS High Sierra * Link to your project: https://www.storj.io :wave: Thanks so much!
priority
invalid argument errors due to setrlimit calls on macos hello when using minio on macos we frequently see log lines that look like the following api system time error invalid argument users user go pkg mod github com minio minio cmd gateway main go cmd startgateway these are ignorable and everything seems to work but still annoying after a bunch of debugging we tracked this down to how minio is trying to increase rlimits in and in there is an open go bug about how the go syscall for getting the max limit is lying this changed between go releases on macos and we believe minio is hitting this bug would it be useful to push up a code review to fix this i think the solution is to extend github com minio minio pkg sys to have special darwin specific handling for not letting the max limit above as described in the go issue expected behavior no inscrutable logs about invalid arguments while minio starts and tries to set limits current behavior minio logs invalid argument on macos possible solution extend github com minio minio pkg sys to specifically not allow the max limit above as described in the go issue about this go s syscall behavior on darwin changed steps to reproduce for bugs run on macos look at logs regression this is a problem introduced by changing go standard libraries your environment version used minio version latest environment name and version e g nginx go server type and version mac mini operating system and version uname a macos high sierra link to your project wave thanks so much
1
533,130
15,577,446,931
IssuesEvent
2021-03-17 13:33:01
Proof-Of-Humanity/proof-of-humanity-web
https://api.github.com/repos/Proof-Of-Humanity/proof-of-humanity-web
opened
Submit Profile Page - Reword Primary Document Button Text
priority: medium status: available type: enhancement :sparkles:
Instead of "primary document", we should say registry rules or something like that.
1.0
Submit Profile Page - Reword Primary Document Button Text - Instead of "primary document", we should say registry rules or something like that.
priority
submit profile page reword primary document button text instead of primary document we should say registry rules or something like that
1
368,912
10,886,082,145
IssuesEvent
2019-11-18 11:46:39
canonical-web-and-design/vanilla-framework
https://api.github.com/repos/canonical-web-and-design/vanilla-framework
closed
Contextual menu (left/center) doesn't grow to fit item's size
Priority: Medium
Contextual menu pattern is defined to have a with between 10rem - 21rem and should adapt to contents (menu items) width. This only works for default (right aligned) menu: <img width="631" alt="screen shot 2018-11-19 at 17 10 16" src="https://user-images.githubusercontent.com/83575/48719886-cb5f6200-ec1e-11e8-8c66-6c8baacff434.png"> Left aligned or centred menus don't grow to fit size of the content and always stay at min-width of 10rem: <img width="642" alt="screen shot 2018-11-19 at 17 13 27" src="https://user-images.githubusercontent.com/83575/48719898-d31f0680-ec1e-11e8-834f-6cb4bc06d187.png"> <img width="646" alt="screen shot 2018-11-19 at 17 10 36" src="https://user-images.githubusercontent.com/83575/48719900-d31f0680-ec1e-11e8-9692-8b5d706e6f83.png"> Found in snapcraft.io: https://github.com/canonical-websites/snapcraft.io/issues/1332 **To Reproduce** Steps to reproduce the behavior: 1. Use `p-contextual-menu--left` or `p-contextual-menu--center` 2. Have a menu item longer then 10 rem 3. See that menu doesn't grow to fit the item. **Desktop:** - OS: Mac OS - Browser: Chrome, Firefox - Version: latest
1.0
Contextual menu (left/center) doesn't grow to fit item's size - Contextual menu pattern is defined to have a with between 10rem - 21rem and should adapt to contents (menu items) width. This only works for default (right aligned) menu: <img width="631" alt="screen shot 2018-11-19 at 17 10 16" src="https://user-images.githubusercontent.com/83575/48719886-cb5f6200-ec1e-11e8-8c66-6c8baacff434.png"> Left aligned or centred menus don't grow to fit size of the content and always stay at min-width of 10rem: <img width="642" alt="screen shot 2018-11-19 at 17 13 27" src="https://user-images.githubusercontent.com/83575/48719898-d31f0680-ec1e-11e8-834f-6cb4bc06d187.png"> <img width="646" alt="screen shot 2018-11-19 at 17 10 36" src="https://user-images.githubusercontent.com/83575/48719900-d31f0680-ec1e-11e8-9692-8b5d706e6f83.png"> Found in snapcraft.io: https://github.com/canonical-websites/snapcraft.io/issues/1332 **To Reproduce** Steps to reproduce the behavior: 1. Use `p-contextual-menu--left` or `p-contextual-menu--center` 2. Have a menu item longer then 10 rem 3. See that menu doesn't grow to fit the item. **Desktop:** - OS: Mac OS - Browser: Chrome, Firefox - Version: latest
priority
contextual menu left center doesn t grow to fit item s size contextual menu pattern is defined to have a with between and should adapt to contents menu items width this only works for default right aligned menu img width alt screen shot at src left aligned or centred menus don t grow to fit size of the content and always stay at min width of img width alt screen shot at src img width alt screen shot at src found in snapcraft io to reproduce steps to reproduce the behavior use p contextual menu left or p contextual menu center have a menu item longer then rem see that menu doesn t grow to fit the item desktop os mac os browser chrome firefox version latest
1
804,612
29,495,147,921
IssuesEvent
2023-06-02 16:17:54
SolarWindss/Hearthstone.js
https://api.github.com/repos/SolarWindss/Hearthstone.js
closed
Move `validateCard` from interact to functions
priority: medium time: short improvement
This doesn't belong in interact - [x] Move - [x] Update src & tests
1.0
Move `validateCard` from interact to functions - This doesn't belong in interact - [x] Move - [x] Update src & tests
priority
move validatecard from interact to functions this doesn t belong in interact move update src tests
1
99,940
4,074,675,549
IssuesEvent
2016-05-28 16:29:15
BugBusterSWE/documentation
https://api.github.com/repos/BugBusterSWE/documentation
reopened
Aggiungere automazione APIDoc
priority:medium
*Documento in cui si trova il problema*: Norme di Progetto Activity #494 *Descrizione del problema*: Aggiungere automazione APIDoc Link task: [https://bugbusters.teamwork.com/tasks/6938623](https://bugbusters.teamwork.com/tasks/6938623)
1.0
Aggiungere automazione APIDoc - *Documento in cui si trova il problema*: Norme di Progetto Activity #494 *Descrizione del problema*: Aggiungere automazione APIDoc Link task: [https://bugbusters.teamwork.com/tasks/6938623](https://bugbusters.teamwork.com/tasks/6938623)
priority
aggiungere automazione apidoc documento in cui si trova il problema norme di progetto activity descrizione del problema aggiungere automazione apidoc link task
1
481,788
13,891,889,024
IssuesEvent
2020-10-19 11:22:33
sunpy/sunpy
https://api.github.com/repos/sunpy/sunpy
closed
Maps using the CD matrix are not correctly modified by resample
Bug(?) Close? Effort High Package Intermediate Priority Medium map
resample only changes the CDELT flags not CD if present.
1.0
Maps using the CD matrix are not correctly modified by resample - resample only changes the CDELT flags not CD if present.
priority
maps using the cd matrix are not correctly modified by resample resample only changes the cdelt flags not cd if present
1
747,678
26,095,245,726
IssuesEvent
2022-12-26 18:18:08
canaltin-byte/SWE573-SDP-Can
https://api.github.com/repos/canaltin-byte/SWE573-SDP-Can
closed
Home Page Name and surname
enhancement priority : Low Front-end Effort: Medium Home Page
User id should not be accessible. User Name and Surname should be there
1.0
Home Page Name and surname - User id should not be accessible. User Name and Surname should be there
priority
home page name and surname user id should not be accessible user name and surname should be there
1
696,650
23,909,701,924
IssuesEvent
2022-09-09 06:54:44
dmwm/CRABServer
https://api.github.com/repos/dmwm/CRABServer
opened
If extraJDL contains DESIRED_Sites, then submission fails after 25 tries
Priority: Medium
I submitted the task `220907_161045:dmapelli_crab_20220907_181041` that contained an `extraJDL` similar to ```python config.Debug.extraJDL = [ '+DESIRED_Sites = "T1_DE_KIT,T2_US_Purdue,T3_UK_London_QMUL"' ] ``` and the submission failed with the error ```plaintext Failure message from server: The CRAB server backend was not able to submit the jobs to the Grid schedulers. This could be a temporary glitch. Please try again later. If the error persists send an e-mail to cmstalk+computing-tools@dovecotmta.cern.ch. The submission was retried 25 times on 5 schedulers. These are the failures per Grid scheduler: {'crab3@vocms059.cern.ch': ["Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cer [... message truncated to the first 1000 chars ...] ``` Looking at the log `/data/container/TaskWorker/logs/tasks/dmapelli/220907_161045:dmapelli_crab_20220907_181041.log` (on `crab-preprod-tw01`) I noticed that the TW tried submitting the task from 18:11 to 18:39, and gave up after 25 tries on 5 different schedd. I know that I made a mistake and that I should have used the `whiteList` instead, but maybe we can add a simple check that when we find `DESIRED_Sites` in the extraJDL we fail the task submission straightaway.
1.0
If extraJDL contains DESIRED_Sites, then submission fails after 25 tries - I submitted the task `220907_161045:dmapelli_crab_20220907_181041` that contained an `extraJDL` similar to ```python config.Debug.extraJDL = [ '+DESIRED_Sites = "T1_DE_KIT,T2_US_Purdue,T3_UK_London_QMUL"' ] ``` and the submission failed with the error ```plaintext Failure message from server: The CRAB server backend was not able to submit the jobs to the Grid schedulers. This could be a temporary glitch. Please try again later. If the error persists send an e-mail to cmstalk+computing-tools@dovecotmta.cern.ch. The submission was retried 25 times on 5 schedulers. These are the failures per Grid scheduler: {'crab3@vocms059.cern.ch': ["Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cer [... message truncated to the first 1000 chars ...] ``` Looking at the log `/data/container/TaskWorker/logs/tasks/dmapelli/220907_161045:dmapelli_crab_20220907_181041.log` (on `crab-preprod-tw01`) I noticed that the TW tried submitting the task from 18:11 to 18:39, and gave up after 25 tries on 5 different schedd. I know that I made a mistake and that I should have used the `whiteList` instead, but maybe we can add a simple check that when we find `DESIRED_Sites` in the extraJDL we fail the task submission straightaway.
priority
if extrajdl contains desired sites then submission fails after tries i submitted the task dmapelli crab that contained an extrajdl similar to python config debug extrajdl and the submission failed with the error plaintext failure message from server the crab server backend was not able to submit the jobs to the grid schedulers this could be a temporary glitch please try again later if the error persists send an e mail to cmstalk computing tools dovecotmta cern ch the submission was retried times on schedulers these are the failures per grid scheduler cern ch failed to submit task to cern ch task dmapelli crab n error when performing action on the schedd failed to submit task to cern ch task dmapelli crab n error when performing action on the schedd failed to submit task to cern ch task dmapelli crab n error when performing action on the schedd failed to submit task to cern ch task dmapelli crab n error when performing action on the schedd failed to submit task to cer looking at the log data container taskworker logs tasks dmapelli dmapelli crab log on crab preprod i noticed that the tw tried submitting the task from to and gave up after tries on different schedd i know that i made a mistake and that i should have used the whitelist instead but maybe we can add a simple check that when we find desired sites in the extrajdl we fail the task submission straightaway
1
556,151
16,476,186,850
IssuesEvent
2021-05-24 05:47:30
Repair-DeskPOS/RepairDesk-BUGS-IMPROVEMENTS
https://api.github.com/repos/Repair-DeskPOS/RepairDesk-BUGS-IMPROVEMENTS
closed
iPAD - Deposit Feature
Cellular Surgeons Medium Priority enhancement iPAD POS Register
**Suggestion # 1** I've created a product under iPhone 7 >> deposit / bench fee which is not refundable. Lets say customer comes to our store and using the iPAD I check in this customer >> tap on checkout and collect the deposit amount. This works fine However when I open the device to find device issue and edit the ticket with new device issue, as soon I click on checkout it says invoice is already paid. Just like web platform it would be nice to edit and existing ticket that has been completely paid. **Suggestion # 2** Also once I ring the item it would be cool to display paid amount along with balance on the right hand side. **Brooks Johnson | Cellular Surgeon**
1.0
iPAD - Deposit Feature - **Suggestion # 1** I've created a product under iPhone 7 >> deposit / bench fee which is not refundable. Lets say customer comes to our store and using the iPAD I check in this customer >> tap on checkout and collect the deposit amount. This works fine However when I open the device to find device issue and edit the ticket with new device issue, as soon I click on checkout it says invoice is already paid. Just like web platform it would be nice to edit and existing ticket that has been completely paid. **Suggestion # 2** Also once I ring the item it would be cool to display paid amount along with balance on the right hand side. **Brooks Johnson | Cellular Surgeon**
priority
ipad deposit feature suggestion i ve created a product under iphone deposit bench fee which is not refundable lets say customer comes to our store and using the ipad i check in this customer tap on checkout and collect the deposit amount this works fine however when i open the device to find device issue and edit the ticket with new device issue as soon i click on checkout it says invoice is already paid just like web platform it would be nice to edit and existing ticket that has been completely paid suggestion also once i ring the item it would be cool to display paid amount along with balance on the right hand side brooks johnson cellular surgeon
1
252,702
8,039,294,016
IssuesEvent
2018-07-30 17:55:36
systers/communities
https://api.github.com/repos/systers/communities
closed
Set up the repository with basic angular files
Category: Coding Difficulty: MEDIUM Priority: HIGH Program: GSoC Type: Enhancement
## Description As a user, I need set up the repository, so that I can restart the project with angular framework ## Acceptance Criteria - Basic working Angular App - Basic files and modules established ### Update [Required] - README - Create multiple new files ## Definition of Done - [ ] All of the required items are completed. - [ ] Approval by 1 mentor. ## Estimation 1 hour Can I be assigned this issue @Tharangi @divyanshu-rawat @Janiceilene @MeepyMay ?
1.0
Set up the repository with basic angular files - ## Description As a user, I need set up the repository, so that I can restart the project with angular framework ## Acceptance Criteria - Basic working Angular App - Basic files and modules established ### Update [Required] - README - Create multiple new files ## Definition of Done - [ ] All of the required items are completed. - [ ] Approval by 1 mentor. ## Estimation 1 hour Can I be assigned this issue @Tharangi @divyanshu-rawat @Janiceilene @MeepyMay ?
priority
set up the repository with basic angular files description as a user i need set up the repository so that i can restart the project with angular framework acceptance criteria basic working angular app basic files and modules established update readme create multiple new files definition of done all of the required items are completed approval by mentor estimation hour can i be assigned this issue tharangi divyanshu rawat janiceilene meepymay
1
68,233
3,285,102,990
IssuesEvent
2015-10-28 19:07:44
pantheon-systems/WordPress
https://api.github.com/repos/pantheon-systems/WordPress
closed
Allow for cache TTL of zero in cache plugin
priority:medium
In some instances such as development users dont want to have the minimum TTL of 600s.
1.0
Allow for cache TTL of zero in cache plugin - In some instances such as development users dont want to have the minimum TTL of 600s.
priority
allow for cache ttl of zero in cache plugin in some instances such as development users dont want to have the minimum ttl of
1
725,678
24,971,367,564
IssuesEvent
2022-11-02 01:42:23
aws-samples/aws-last-mile-delivery-hyperlocal
https://api.github.com/repos/aws-samples/aws-last-mile-delivery-hyperlocal
closed
Setup yarn dependency checks for all packages
enhancement dependencies priority:medium component:all effort:low
* first candidate would be [depcheck](https://github.com/depcheck/depcheck). * to make sure that there are no missing dependencies in any packages
1.0
Setup yarn dependency checks for all packages - * first candidate would be [depcheck](https://github.com/depcheck/depcheck). * to make sure that there are no missing dependencies in any packages
priority
setup yarn dependency checks for all packages first candidate would be to make sure that there are no missing dependencies in any packages
1
94,280
3,923,997,027
IssuesEvent
2016-04-22 13:46:00
EnvironmentAgency/pafs-user
https://api.github.com/repos/EnvironmentAgency/pafs-user
opened
CR:: User Requirements::Funding Sources:: Change the word 'expected'
Change Priority - 3 Medium Sprint 3
**Issue:** The word 'expected' caused participants to assume the funding types listed on this page, represent funding that is yet to be secured. ![funding source](https://cloud.githubusercontent.com/assets/17175769/14743271/ad94de56-0898-11e6-9baf-03aa577eb210.PNG) For example, if you have already secured some private funding. This was not seen as required to be entered here as it is not expected -you have already secured it. **Recommendation** To Be Discussed
1.0
CR:: User Requirements::Funding Sources:: Change the word 'expected' - **Issue:** The word 'expected' caused participants to assume the funding types listed on this page, represent funding that is yet to be secured. ![funding source](https://cloud.githubusercontent.com/assets/17175769/14743271/ad94de56-0898-11e6-9baf-03aa577eb210.PNG) For example, if you have already secured some private funding. This was not seen as required to be entered here as it is not expected -you have already secured it. **Recommendation** To Be Discussed
priority
cr user requirements funding sources change the word expected issue the word expected caused participants to assume the funding types listed on this page represent funding that is yet to be secured for example if you have already secured some private funding this was not seen as required to be entered here as it is not expected you have already secured it recommendation to be discussed
1
445,052
12,825,477,672
IssuesEvent
2020-07-06 15:01:28
radcampaign/allaboveall2020
https://api.github.com/repos/radcampaign/allaboveall2020
closed
Tweak menu font size/icons
Priority: Medium
More QC from Allyson: she'd like the main menu to stand out a bit more. In comparing with the comps, I know the font size is dead on to what Kien provided. Possibly we could add more weight? How much bigger could the menu font go? I know there's a challenge to get the current items to fit. Maybe the social and search icons could decrease in size a notch or two? Or could we reduce the whitespace between the logo and the start of the black bar and menu? Again, no rush, just logging.
1.0
Tweak menu font size/icons - More QC from Allyson: she'd like the main menu to stand out a bit more. In comparing with the comps, I know the font size is dead on to what Kien provided. Possibly we could add more weight? How much bigger could the menu font go? I know there's a challenge to get the current items to fit. Maybe the social and search icons could decrease in size a notch or two? Or could we reduce the whitespace between the logo and the start of the black bar and menu? Again, no rush, just logging.
priority
tweak menu font size icons more qc from allyson she d like the main menu to stand out a bit more in comparing with the comps i know the font size is dead on to what kien provided possibly we could add more weight how much bigger could the menu font go i know there s a challenge to get the current items to fit maybe the social and search icons could decrease in size a notch or two or could we reduce the whitespace between the logo and the start of the black bar and menu again no rush just logging
1
471,331
13,565,237,770
IssuesEvent
2020-09-18 11:21:56
DXHeroes/dx-scanner
https://api.github.com/repos/DXHeroes/dx-scanner
opened
Create a language detector for Swift (~100 new lines of code)
Difficulty: Easy Priority: Medium Status: Available Type: Enhancement good first issue
We'd like to measure the quality of apps written in Swift. There are "language detectors" to determine what language is on which path. It's necessary to implement **SwiftLanguageDetector** the same way as is [JavaScriptLanguageDetector](https://github.com/DXHeroes/dx-scanner/blob/master/src/detectors/JavaScript/JavaScriptLanguageDetector.ts) implemented. It's around 100 new lines of code including tests. More info in [our Contribution guide](https://github.com/DXHeroes/dx-scanner/blob/master/CONTRIBUTING.md).
1.0
Create a language detector for Swift (~100 new lines of code) - We'd like to measure the quality of apps written in Swift. There are "language detectors" to determine what language is on which path. It's necessary to implement **SwiftLanguageDetector** the same way as is [JavaScriptLanguageDetector](https://github.com/DXHeroes/dx-scanner/blob/master/src/detectors/JavaScript/JavaScriptLanguageDetector.ts) implemented. It's around 100 new lines of code including tests. More info in [our Contribution guide](https://github.com/DXHeroes/dx-scanner/blob/master/CONTRIBUTING.md).
priority
create a language detector for swift new lines of code we d like to measure the quality of apps written in swift there are language detectors to determine what language is on which path it s necessary to implement swiftlanguagedetector the same way as is implemented it s around new lines of code including tests more info in
1
650,780
21,416,982,859
IssuesEvent
2022-04-22 11:54:59
sahar-avsh/SWE-599
https://api.github.com/repos/sahar-avsh/SWE-599
closed
Q&A - Creating a question UI
enhancement Show stopper Hard medium priority Q&A
For creating question: - [x] A **form** shall open - [x] User shall enter **title** - [x] User shall be able to enter **description** - [x] User shall be able to **attach** her/his **decks** - [x] User shall be able to **attach** her/his **resources**
1.0
Q&A - Creating a question UI - For creating question: - [x] A **form** shall open - [x] User shall enter **title** - [x] User shall be able to enter **description** - [x] User shall be able to **attach** her/his **decks** - [x] User shall be able to **attach** her/his **resources**
priority
q a creating a question ui for creating question a form shall open user shall enter title user shall be able to enter description user shall be able to attach her his decks user shall be able to attach her his resources
1
224,225
7,467,858,395
IssuesEvent
2018-04-02 16:52:53
enforcer574/smashclub
https://api.github.com/repos/enforcer574/smashclub
opened
Photo Slideshow on Home Page
Complexity: Medium Priority: 3 - Medium Type: User Request
Replace the static image on the "about" portlet of the home page with a timed slideshow. Admins can upload photos to a directory in the assets folder and they will be displayed in the rotation. A setting in the site_settings table controls the slideshow speed.
1.0
Photo Slideshow on Home Page - Replace the static image on the "about" portlet of the home page with a timed slideshow. Admins can upload photos to a directory in the assets folder and they will be displayed in the rotation. A setting in the site_settings table controls the slideshow speed.
priority
photo slideshow on home page replace the static image on the about portlet of the home page with a timed slideshow admins can upload photos to a directory in the assets folder and they will be displayed in the rotation a setting in the site settings table controls the slideshow speed
1
406,501
11,894,049,716
IssuesEvent
2020-03-29 14:14:33
robotframework/robotframework
https://api.github.com/repos/robotframework/robotframework
closed
Dynamic API: Add new `get_keyword_source` method
enhancement priority: medium rc 1
This is needed to make it possible to add source information to Libdoc spec files (#3507) as well as to Libdoc's model objects (#3448). External tools like editors can then use this information to implement "go to definition" functionality. This method needs to be able to return both the path to the source file and the line number, and we need to decide which format to use. Good alternatives include a tuple `(source, lineno)` as well as a string `source:lineno`. Not sure why, but I like the latter approach better.
1.0
Dynamic API: Add new `get_keyword_source` method - This is needed to make it possible to add source information to Libdoc spec files (#3507) as well as to Libdoc's model objects (#3448). External tools like editors can then use this information to implement "go to definition" functionality. This method needs to be able to return both the path to the source file and the line number, and we need to decide which format to use. Good alternatives include a tuple `(source, lineno)` as well as a string `source:lineno`. Not sure why, but I like the latter approach better.
priority
dynamic api add new get keyword source method this is needed to make it possible to add source information to libdoc spec files as well as to libdoc s model objects external tools like editors can then use this information to implement go to definition functionality this method needs to be able to return both the path to the source file and the line number and we need to decide which format to use good alternatives include a tuple source lineno as well as a string source lineno not sure why but i like the latter approach better
1
707,107
24,295,583,272
IssuesEvent
2022-09-29 09:46:52
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
Upgrade versions in .tool-versions
type:feature help wanted priority-3-medium new package manager status:in-progress
**What would you like Renovate to be able to do?** When upgrading node, ruby or the like, it should also change the version in `.tool-versions`, used by [asdf](https://github.com/asdf-vm/asdf). **Describe the solution you'd like** See above. **Describe alternatives you've considered** Doing it manually. **Additional context** I'd imagine other version managers have similar files. Maybe renovate already supports them, but if not, it'd be great if it did as well.
1.0
Upgrade versions in .tool-versions - **What would you like Renovate to be able to do?** When upgrading node, ruby or the like, it should also change the version in `.tool-versions`, used by [asdf](https://github.com/asdf-vm/asdf). **Describe the solution you'd like** See above. **Describe alternatives you've considered** Doing it manually. **Additional context** I'd imagine other version managers have similar files. Maybe renovate already supports them, but if not, it'd be great if it did as well.
priority
upgrade versions in tool versions what would you like renovate to be able to do when upgrading node ruby or the like it should also change the version in tool versions used by describe the solution you d like see above describe alternatives you ve considered doing it manually additional context i d imagine other version managers have similar files maybe renovate already supports them but if not it d be great if it did as well
1
457,584
13,158,552,952
IssuesEvent
2020-08-10 14:30:06
canonical-web-and-design/jaas-dashboard
https://api.github.com/repos/canonical-web-and-design/jaas-dashboard
opened
Leader information not shown in unit list
Model Details Priority: Medium
When a unit is in HA and a leader has been elected we should indicate as such. I'd expect to see the leader status in the unit list but also in the unit details. @ziheliu214 can you add this to the unit list/details designs please. related: https://discourse.juju.is/t/leadership-in-juju-operations-perspective/3406
1.0
Leader information not shown in unit list - When a unit is in HA and a leader has been elected we should indicate as such. I'd expect to see the leader status in the unit list but also in the unit details. @ziheliu214 can you add this to the unit list/details designs please. related: https://discourse.juju.is/t/leadership-in-juju-operations-perspective/3406
priority
leader information not shown in unit list when a unit is in ha and a leader has been elected we should indicate as such i d expect to see the leader status in the unit list but also in the unit details can you add this to the unit list details designs please related
1
375,306
11,102,411,831
IssuesEvent
2019-12-16 23:58:12
AugurProject/augur
https://api.github.com/repos/AugurProject/augur
closed
ZeroXTrade does not enforce ERC1155_PROXY_ID in AssetData
Priority: Medium V2 Audit
While the off-chain calls to ZeroXTrade (createZeroXOrderFor) will encode the ERC1155_PROXY_ID (`0xa7cb5fb7`) into the first four bytes of the order that is submitted to 0x.fillOrder, there is no check inside the actual trade function to ensure that is what is being passed along to the 0x Exchange. Alternate PROXY_IDs could behave differently and might even look in different places for actual asset being transferred, which could lead to unexpected fill amounts being returned, and potentially even unexpected trades occurring (although the tokenAddress being immediately after proxy_id is likely to limit severity). New proxies could be added in the future with unexpected functionality.
1.0
ZeroXTrade does not enforce ERC1155_PROXY_ID in AssetData - While the off-chain calls to ZeroXTrade (createZeroXOrderFor) will encode the ERC1155_PROXY_ID (`0xa7cb5fb7`) into the first four bytes of the order that is submitted to 0x.fillOrder, there is no check inside the actual trade function to ensure that is what is being passed along to the 0x Exchange. Alternate PROXY_IDs could behave differently and might even look in different places for actual asset being transferred, which could lead to unexpected fill amounts being returned, and potentially even unexpected trades occurring (although the tokenAddress being immediately after proxy_id is likely to limit severity). New proxies could be added in the future with unexpected functionality.
priority
zeroxtrade does not enforce proxy id in assetdata while the off chain calls to zeroxtrade createzeroxorderfor will encode the proxy id into the first four bytes of the order that is submitted to fillorder there is no check inside the actual trade function to ensure that is what is being passed along to the exchange alternate proxy ids could behave differently and might even look in different places for actual asset being transferred which could lead to unexpected fill amounts being returned and potentially even unexpected trades occurring although the tokenaddress being immediately after proxy id is likely to limit severity new proxies could be added in the future with unexpected functionality
1
204,028
7,079,438,127
IssuesEvent
2018-01-10 09:34:02
Automattic/liveblog
https://api.github.com/repos/Automattic/liveblog
closed
Remove reliance on GET params for single entry ajax requests
Priority::Medium enhancement
Scenario: * A user loads a liveblog * There are several key events shown in the key events widget * Some of the key events have not loaded in the initial set of entries that are loaded by default * The user clicks on one of the entries that isn't visible * The plugin fires off an ajax request [appending an `index` GET param to the endpoint URL](https://github.com/Automattic/liveblog/blob/e50c8dad22e7519f42a0045d5b564d3e60d9c5e4/js/liveblog-lazyloader.js#L82-L85) * The single entry is returned and displayed See where the [ajax action is set up](https://github.com/Automattic/liveblog/blob/f82842828cad0c7b921459b10a6e534f9cbe5355/liveblog.php#L299) and the [request is processed](https://github.com/Automattic/liveblog/blob/f82842828cad0c7b921459b10a6e534f9cbe5355/liveblog.php#L519-L575). We should refactor this to append an ID or some such to the endpoint URL, rather than a GET param, to avoid bypassing caching mechanisms like Batcache.
1.0
Remove reliance on GET params for single entry ajax requests - Scenario: * A user loads a liveblog * There are several key events shown in the key events widget * Some of the key events have not loaded in the initial set of entries that are loaded by default * The user clicks on one of the entries that isn't visible * The plugin fires off an ajax request [appending an `index` GET param to the endpoint URL](https://github.com/Automattic/liveblog/blob/e50c8dad22e7519f42a0045d5b564d3e60d9c5e4/js/liveblog-lazyloader.js#L82-L85) * The single entry is returned and displayed See where the [ajax action is set up](https://github.com/Automattic/liveblog/blob/f82842828cad0c7b921459b10a6e534f9cbe5355/liveblog.php#L299) and the [request is processed](https://github.com/Automattic/liveblog/blob/f82842828cad0c7b921459b10a6e534f9cbe5355/liveblog.php#L519-L575). We should refactor this to append an ID or some such to the endpoint URL, rather than a GET param, to avoid bypassing caching mechanisms like Batcache.
priority
remove reliance on get params for single entry ajax requests scenario a user loads a liveblog there are several key events shown in the key events widget some of the key events have not loaded in the initial set of entries that are loaded by default the user clicks on one of the entries that isn t visible the plugin fires off an ajax request the single entry is returned and displayed see where the and the we should refactor this to append an id or some such to the endpoint url rather than a get param to avoid bypassing caching mechanisms like batcache
1
828,734
31,840,844,496
IssuesEvent
2023-09-14 16:12:04
bcgov/foi-flow
https://api.github.com/repos/bcgov/foi-flow
closed
Single Source of Truth - Divisions
Task dev medium priority
Title of ticket: #### Description This task has been created to Remove existing `Program Area Divisions ` Tables and related dependencies like `PROGRAM AREAS ?? ` inside DOC Reviewer DB , in order to keep single source of truth of Divisions across FOI FLOW app. At present DOC REVIEWER Web App (only?) using this table(s) #### Dependencies Are there any dependencies? #### DOD - [x] Analyze the repercussions of removing Divisions, Program areas from Doc Reviewer DB - On initial analysis, its only used for Left panel on Doc Reviewer Web app, so Idea is to Create a new endpoint for Divisions by Program area from FOI FLOW - [x] Once above Analyze DOD is over and correct, Develop an API endpoint or use/check existing APIs for Divisions - [x] Update Left Panel Front End component to use FOI FLOW API endpoint for divisions. - [x] Integration test - [ ]
1.0
Single Source of Truth - Divisions - Title of ticket: #### Description This task has been created to Remove existing `Program Area Divisions ` Tables and related dependencies like `PROGRAM AREAS ?? ` inside DOC Reviewer DB , in order to keep single source of truth of Divisions across FOI FLOW app. At present DOC REVIEWER Web App (only?) using this table(s) #### Dependencies Are there any dependencies? #### DOD - [x] Analyze the repercussions of removing Divisions, Program areas from Doc Reviewer DB - On initial analysis, its only used for Left panel on Doc Reviewer Web app, so Idea is to Create a new endpoint for Divisions by Program area from FOI FLOW - [x] Once above Analyze DOD is over and correct, Develop an API endpoint or use/check existing APIs for Divisions - [x] Update Left Panel Front End component to use FOI FLOW API endpoint for divisions. - [x] Integration test - [ ]
priority
single source of truth divisions title of ticket description this task has been created to remove existing program area divisions tables and related dependencies like program areas inside doc reviewer db in order to keep single source of truth of divisions across foi flow app at present doc reviewer web app only using this table s dependencies are there any dependencies dod analyze the repercussions of removing divisions program areas from doc reviewer db on initial analysis its only used for left panel on doc reviewer web app so idea is to create a new endpoint for divisions by program area from foi flow once above analyze dod is over and correct develop an api endpoint or use check existing apis for divisions update left panel front end component to use foi flow api endpoint for divisions integration test
1
706,036
24,258,019,046
IssuesEvent
2022-09-27 19:37:18
gammapy/gammapy
https://api.github.com/repos/gammapy/gammapy
closed
Gammapy validation: HESS DL3 DR1
effort-medium package-novice priority-high
As part of the Gammapy validation effort while preparing v1.0, we should script some of the H.E.S.S. data level 3, data release 1 analyses for validation. Note that large parts of Gammapy were rewritten since Gammapy v0.12 (May 2019) which was used for https://arxiv.org/pdf/1910.08088.pdf requiring the validation exercise to be re-done with the latest version of Gammapy, and from now on moving forward we'll maintain the validation scripts in the Gammapy team to avoid regressions. One concrete case of a regression that we already noticed and fixed in the meantime is https://github.com/gammapy/gammapy/pull/2367 . A task description and references are here: - https://github.com/gammapy/gammapy-benchmarks - https://github.com/gammapy/gammapy-benchmarks/blob/master/validation - https://github.com/gammapy/gammapy-benchmarks/tree/master/validation/hess-dl3-dr1 We're looking for help! If you can contribute, please leave a comment here, or contact me on Slack. For this, you don't have to be a Gammapy developer, what's required is some Python & Gammapy & IACT analysis experience, and at least ~ 2 full days to contribute in Nov 2019, although ~ 1 week in Nov & Dec 2019 is more realistic if you're willing to script the analysis for all targets.
1.0
Gammapy validation: HESS DL3 DR1 - As part of the Gammapy validation effort while preparing v1.0, we should script some of the H.E.S.S. data level 3, data release 1 analyses for validation. Note that large parts of Gammapy were rewritten since Gammapy v0.12 (May 2019) which was used for https://arxiv.org/pdf/1910.08088.pdf requiring the validation exercise to be re-done with the latest version of Gammapy, and from now on moving forward we'll maintain the validation scripts in the Gammapy team to avoid regressions. One concrete case of a regression that we already noticed and fixed in the meantime is https://github.com/gammapy/gammapy/pull/2367 . A task description and references are here: - https://github.com/gammapy/gammapy-benchmarks - https://github.com/gammapy/gammapy-benchmarks/blob/master/validation - https://github.com/gammapy/gammapy-benchmarks/tree/master/validation/hess-dl3-dr1 We're looking for help! If you can contribute, please leave a comment here, or contact me on Slack. For this, you don't have to be a Gammapy developer, what's required is some Python & Gammapy & IACT analysis experience, and at least ~ 2 full days to contribute in Nov 2019, although ~ 1 week in Nov & Dec 2019 is more realistic if you're willing to script the analysis for all targets.
priority
gammapy validation hess as part of the gammapy validation effort while preparing we should script some of the h e s s data level data release analyses for validation note that large parts of gammapy were rewritten since gammapy may which was used for requiring the validation exercise to be re done with the latest version of gammapy and from now on moving forward we ll maintain the validation scripts in the gammapy team to avoid regressions one concrete case of a regression that we already noticed and fixed in the meantime is a task description and references are here we re looking for help if you can contribute please leave a comment here or contact me on slack for this you don t have to be a gammapy developer what s required is some python gammapy iact analysis experience and at least full days to contribute in nov although week in nov dec is more realistic if you re willing to script the analysis for all targets
1
236,074
7,745,930,033
IssuesEvent
2018-05-29 19:58:15
strapi/strapi
https://api.github.com/repos/strapi/strapi
closed
Overwriting GraphQL schema
priority: medium type: bug 🐛
<!-- ⚠️ If you do not respect these two points, your issue will be closed. - Don't forget the checkbox at the end of your issue. - Respect this template. --> <!-- ⚠️ Before writing your issue make sure you are using:--> <!-- Node 9.x.x --> <!-- npm 5.x.x --> <!-- The latest version of Strapi. --> **Informations** - **Node.js version**: 9.11.1 - **npm version**: 5.6.0 - **Strapi version**: 3.0.0-alpha.12.2 - **Database**: mongo - **Operating system**: macOS **What is the current behavior?** I'm trying to overwrite the GraphQL query from ```graphql type Query { practice(id: String): Practice type: { Practice: false }, resolver: { Query: { practice: { description: 'Return the practice of the authenticated user', resolver: 'Practice.findMine' }, practices: false } } } ``` to ```graphql type Query { practice: Practice type: { Practice: false }, resolver: { Query: { practice: { description: 'Return the practice of the authenticated user', resolver: 'Practice.findMine' }, practices: false } } } ``` But running strapi will error in `Error: Resolver Query.practice must be object or function` **Steps to reproduce the problem** Try to overwrite the default `findOne` route of any model like the example above **What is the expected behavior?** It should overwrite the schema without a problem, even the generated code looks correct ```graphql # generated schema file type Query { patient(id: String!): Patient patients(sort: String, limit: Int, start: Int, where: JSON): [Patient] practice: Practice # this looks correct, it doesn't have the id: String, argument ``` **Suggested solutions** I don't know any, I'm trying to find a solution myself without success --- <!-- ⚠️ Make sure to browse the opened and closed issues. --> - [x] I'm sure that this feature hasn't already been referenced. <!-- ⬆️ Don't delete this checkbox from your issue and approve it. Add an `x` -> [x] or click on the checkbox when your issue is created. -->
1.0
Overwriting GraphQL schema - <!-- ⚠️ If you do not respect these two points, your issue will be closed. - Don't forget the checkbox at the end of your issue. - Respect this template. --> <!-- ⚠️ Before writing your issue make sure you are using:--> <!-- Node 9.x.x --> <!-- npm 5.x.x --> <!-- The latest version of Strapi. --> **Informations** - **Node.js version**: 9.11.1 - **npm version**: 5.6.0 - **Strapi version**: 3.0.0-alpha.12.2 - **Database**: mongo - **Operating system**: macOS **What is the current behavior?** I'm trying to overwrite the GraphQL query from ```graphql type Query { practice(id: String): Practice type: { Practice: false }, resolver: { Query: { practice: { description: 'Return the practice of the authenticated user', resolver: 'Practice.findMine' }, practices: false } } } ``` to ```graphql type Query { practice: Practice type: { Practice: false }, resolver: { Query: { practice: { description: 'Return the practice of the authenticated user', resolver: 'Practice.findMine' }, practices: false } } } ``` But running strapi will error in `Error: Resolver Query.practice must be object or function` **Steps to reproduce the problem** Try to overwrite the default `findOne` route of any model like the example above **What is the expected behavior?** It should overwrite the schema without a problem, even the generated code looks correct ```graphql # generated schema file type Query { patient(id: String!): Patient patients(sort: String, limit: Int, start: Int, where: JSON): [Patient] practice: Practice # this looks correct, it doesn't have the id: String, argument ``` **Suggested solutions** I don't know any, I'm trying to find a solution myself without success --- <!-- ⚠️ Make sure to browse the opened and closed issues. --> - [x] I'm sure that this feature hasn't already been referenced. <!-- ⬆️ Don't delete this checkbox from your issue and approve it. Add an `x` -> [x] or click on the checkbox when your issue is created. -->
priority
overwriting graphql schema ⚠️ if you do not respect these two points your issue will be closed don t forget the checkbox at the end of your issue respect this template informations node js version npm version strapi version alpha database mongo operating system macos what is the current behavior i m trying to overwrite the graphql query from graphql type query practice id string practice type practice false resolver query practice description return the practice of the authenticated user resolver practice findmine practices false to graphql type query practice practice type practice false resolver query practice description return the practice of the authenticated user resolver practice findmine practices false but running strapi will error in error resolver query practice must be object or function steps to reproduce the problem try to overwrite the default findone route of any model like the example above what is the expected behavior it should overwrite the schema without a problem even the generated code looks correct graphql generated schema file type query patient id string patient patients sort string limit int start int where json practice practice this looks correct it doesn t have the id string argument suggested solutions i don t know any i m trying to find a solution myself without success i m sure that this feature hasn t already been referenced ⬆️ don t delete this checkbox from your issue and approve it add an x or click on the checkbox when your issue is created
1
803,510
29,180,221,691
IssuesEvent
2023-05-19 11:14:37
Vatsim-Scandinavia/controlcenter
https://api.github.com/repos/Vatsim-Scandinavia/controlcenter
closed
Visiting Controllers can't book positions
back-end bug priority: medium
We missed them out in the logic checks. Users who have an active visiting endorsement should also be able to book.
1.0
Visiting Controllers can't book positions - We missed them out in the logic checks. Users who have an active visiting endorsement should also be able to book.
priority
visiting controllers can t book positions we missed them out in the logic checks users who have an active visiting endorsement should also be able to book
1
39,837
2,860,368,169
IssuesEvent
2015-06-03 15:32:09
CUL-DigitalServices/grasshopper-ui
https://api.github.com/repos/CUL-DigitalServices/grasshopper-ui
opened
Switching from batch edit to calendar view binds event handlers again
Medium Priority To Fix
Instead of binding them all over again, binding them once when switching to the view initially should be enough.
1.0
Switching from batch edit to calendar view binds event handlers again - Instead of binding them all over again, binding them once when switching to the view initially should be enough.
priority
switching from batch edit to calendar view binds event handlers again instead of binding them all over again binding them once when switching to the view initially should be enough
1
781,917
27,454,426,190
IssuesEvent
2023-03-02 20:03:42
phetsims/scenery
https://api.github.com/repos/phetsims/scenery
closed
Moving to phet.{{X}} namespaces instead of multiple globals
dev:enhancement priority:3-medium status:ready-for-review dev:chip-away
During the namespace discussion, it was noted that there are a number of advantages to move to a single global variable (`phet`). The consistency will help with things like https://github.com/phetsims/scenery/issues/965, and it will override fewer names. Those seem like good arguments to me (that somewhat override the verbosity concerns, since you can always create local variables if things are too verbose). This seems like a few core changes (changing the available namespacing), and then a lot of refactoring to point everything to the new namespace (phetCore/axon/dot/kite/scenery at a minimum). I'm curious if there are objections to this approach
1.0
Moving to phet.{{X}} namespaces instead of multiple globals - During the namespace discussion, it was noted that there are a number of advantages to move to a single global variable (`phet`). The consistency will help with things like https://github.com/phetsims/scenery/issues/965, and it will override fewer names. Those seem like good arguments to me (that somewhat override the verbosity concerns, since you can always create local variables if things are too verbose). This seems like a few core changes (changing the available namespacing), and then a lot of refactoring to point everything to the new namespace (phetCore/axon/dot/kite/scenery at a minimum). I'm curious if there are objections to this approach
priority
moving to phet x namespaces instead of multiple globals during the namespace discussion it was noted that there are a number of advantages to move to a single global variable phet the consistency will help with things like and it will override fewer names those seem like good arguments to me that somewhat override the verbosity concerns since you can always create local variables if things are too verbose this seems like a few core changes changing the available namespacing and then a lot of refactoring to point everything to the new namespace phetcore axon dot kite scenery at a minimum i m curious if there are objections to this approach
1
546,041
15,982,997,889
IssuesEvent
2021-04-18 07:10:25
r-lib/styler
https://api.github.com/repos/r-lib/styler
closed
Indenting leading spaces is inconsistent?
Complexity: Medium Priority: High Status: Unassigned Type: Bug
Hi, I noticed that when there is a certain number of leading spaces, the line does not indent to the correct number. For example, with `indent_by = 2` (though this issue appears for more than just `2` spaces), ### 4 leading spaces (for example): ```R function() { print("hi") } ``` does not indent to 2 spaces. However, ### 17 leading spaces (for example): ```R function() { print("hi") } ``` does indent to 2. Can anyone reproduce this, or does anyone know why this is? ### System info styler: 1.3.2 R: 4.0.4
1.0
Indenting leading spaces is inconsistent? - Hi, I noticed that when there is a certain number of leading spaces, the line does not indent to the correct number. For example, with `indent_by = 2` (though this issue appears for more than just `2` spaces), ### 4 leading spaces (for example): ```R function() { print("hi") } ``` does not indent to 2 spaces. However, ### 17 leading spaces (for example): ```R function() { print("hi") } ``` does indent to 2. Can anyone reproduce this, or does anyone know why this is? ### System info styler: 1.3.2 R: 4.0.4
priority
indenting leading spaces is inconsistent hi i noticed that when there is a certain number of leading spaces the line does not indent to the correct number for example with indent by though this issue appears for more than just spaces leading spaces for example r function print hi does not indent to spaces however leading spaces for example r function print hi does indent to can anyone reproduce this or does anyone know why this is system info styler r
1
116,944
4,709,096,787
IssuesEvent
2016-10-14 03:23:24
honestbleeps/Reddit-Enhancement-Suite
https://api.github.com/repos/honestbleeps/Reddit-Enhancement-Suite
closed
Add a class to notifications with module id and notification ID
Difficulty-­2_Medium Difficulty-­­1_Easy hacktoberfest Priority-4_Some Interest RE-Enhancement
This might be a duplicate https://www.reddit.com/r/Enhancement/comments/4c9c5f/feature_request_add_a_class_or_id_to_the/
1.0
Add a class to notifications with module id and notification ID - This might be a duplicate https://www.reddit.com/r/Enhancement/comments/4c9c5f/feature_request_add_a_class_or_id_to_the/
priority
add a class to notifications with module id and notification id this might be a duplicate
1
809,125
30,175,401,407
IssuesEvent
2023-07-04 03:48:36
Laserwolve-Games/DaggerQuest
https://api.github.com/repos/Laserwolve-Games/DaggerQuest
opened
Spawning Loot Occasionally Instead Rotates Other Loot
Bug Hard Difficulty Medium Priority
This is very hard to reproduce. Occasionally when spawning loot, no loot is a spawned, and a previously spawned piece of loot changes it's direction (animation). Need to find out if the loot that enemies drop also can cause this issue.
1.0
Spawning Loot Occasionally Instead Rotates Other Loot - This is very hard to reproduce. Occasionally when spawning loot, no loot is a spawned, and a previously spawned piece of loot changes it's direction (animation). Need to find out if the loot that enemies drop also can cause this issue.
priority
spawning loot occasionally instead rotates other loot this is very hard to reproduce occasionally when spawning loot no loot is a spawned and a previously spawned piece of loot changes it s direction animation need to find out if the loot that enemies drop also can cause this issue
1
199,741
6,993,824,286
IssuesEvent
2017-12-15 13:04:19
emory-libraries/ezpaarse-platforms
https://api.github.com/repos/emory-libraries/ezpaarse-platforms
closed
Policy Map
Additional Parser Medium Priority review
### Example:star::star: : www.policymap.com ### Priority: Medium ### Subscriber (Library): Woodruff
1.0
Policy Map - ### Example:star::star: : www.policymap.com ### Priority: Medium ### Subscriber (Library): Woodruff
priority
policy map example star star priority medium subscriber library woodruff
1
523,321
15,178,210,461
IssuesEvent
2021-02-14 14:33:19
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
opened
Adapt Religious Clothes
:art: 3D graphics :triangular_ruler: :grey_exclamation: priority medium :question: suggestion :question:
<!-- DO NOT REMOVE PRE-EXISTING LINES IF YOU WANT TO SUGGEST A FEW THINGS, OPEN A NEW ISSUE PER EVERY SUGGESTION ---------------------------------------------------------------------------------------------------------- --> **Describe your suggestion in full detail below:** We should adapt religious clothes making sure every religion will have them. I started to write their triggers... <details> <summary>Click to expand</summary> ![image](https://user-images.githubusercontent.com/46384992/107879542-dcc57080-6ef2-11eb-9634-aaa7b911413e.png) </details> ...but it turned out a too big project, so here I'm. If the religion can't get any special clothes, it should be using Catholic clothes instead.
1.0
Adapt Religious Clothes - <!-- DO NOT REMOVE PRE-EXISTING LINES IF YOU WANT TO SUGGEST A FEW THINGS, OPEN A NEW ISSUE PER EVERY SUGGESTION ---------------------------------------------------------------------------------------------------------- --> **Describe your suggestion in full detail below:** We should adapt religious clothes making sure every religion will have them. I started to write their triggers... <details> <summary>Click to expand</summary> ![image](https://user-images.githubusercontent.com/46384992/107879542-dcc57080-6ef2-11eb-9634-aaa7b911413e.png) </details> ...but it turned out a too big project, so here I'm. If the religion can't get any special clothes, it should be using Catholic clothes instead.
priority
adapt religious clothes do not remove pre existing lines if you want to suggest a few things open a new issue per every suggestion describe your suggestion in full detail below we should adapt religious clothes making sure every religion will have them i started to write their triggers click to expand but it turned out a too big project so here i m if the religion can t get any special clothes it should be using catholic clothes instead
1
30,501
2,723,981,498
IssuesEvent
2015-04-14 15:27:21
CruxFramework/crux-widgets
https://api.github.com/repos/CruxFramework/crux-widgets
closed
FileProtocolIteratorFactory pass file relative names for its filter, when scanning for resources
bug CruxScannotation imported Milestone-3.0.1 Priority-Medium
_From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on October 08, 2010 11:49:50_ FileProtocolIteratorFactory pass file relative names for its filter, when scanning for resources _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=204_
1.0
FileProtocolIteratorFactory pass file relative names for its filter, when scanning for resources - _From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on October 08, 2010 11:49:50_ FileProtocolIteratorFactory pass file relative names for its filter, when scanning for resources _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=204_
priority
fileprotocoliteratorfactory pass file relative names for its filter when scanning for resources from on october fileprotocoliteratorfactory pass file relative names for its filter when scanning for resources original issue
1
466,771
13,433,366,639
IssuesEvent
2020-09-07 09:43:44
enso-org/ide
https://api.github.com/repos/enso-org/ide
closed
Lexer: Implement parsing of literals.
Category: IDE Change: Non-Breaking Difficulty: Core Contributor Priority: Medium Status: Invalid Type: Enhancement
### Summary Reimplement parsing of literals from scala in rust. ### Value Literals are another primitive of our AST. ### Specification Reimplement parsing of literals from scala in rust. ### Acceptance Criteria & Test Implement and test the specified task above.
1.0
Lexer: Implement parsing of literals. - ### Summary Reimplement parsing of literals from scala in rust. ### Value Literals are another primitive of our AST. ### Specification Reimplement parsing of literals from scala in rust. ### Acceptance Criteria & Test Implement and test the specified task above.
priority
lexer implement parsing of literals summary reimplement parsing of literals from scala in rust value literals are another primitive of our ast specification reimplement parsing of literals from scala in rust acceptance criteria test implement and test the specified task above
1
612,417
19,012,081,779
IssuesEvent
2021-11-23 10:24:27
containrrr/watchtower
https://api.github.com/repos/containrrr/watchtower
closed
alibaba Cloud private warehouse cannot pull the image and automatically upgrade
Type: Bug Priority: Medium Status: Available
my docker-compose.yml ```yaml watchtower: container_name: watchtower image: containrrr/watchtower restart: always environment: - TZ=Asia/Shanghai privileged: true depends_on: - mj-admin-service - mj-admin-portal volumes: - /var/run/docker.sock:/var/run/docker.sock - /root/.docker/config.json:/config.json command: mj-admin-service --cleanup --interval 10 logging: driver: "json-file" options: max-size: "1m" max-file: "1" ``` the log shows that it is always ready, and there is no more log output: ![image](https://user-images.githubusercontent.com/12794749/142966581-3095f521-3754-4491-8350-68331ec6636e.png)
1.0
alibaba Cloud private warehouse cannot pull the image and automatically upgrade - my docker-compose.yml ```yaml watchtower: container_name: watchtower image: containrrr/watchtower restart: always environment: - TZ=Asia/Shanghai privileged: true depends_on: - mj-admin-service - mj-admin-portal volumes: - /var/run/docker.sock:/var/run/docker.sock - /root/.docker/config.json:/config.json command: mj-admin-service --cleanup --interval 10 logging: driver: "json-file" options: max-size: "1m" max-file: "1" ``` the log shows that it is always ready, and there is no more log output: ![image](https://user-images.githubusercontent.com/12794749/142966581-3095f521-3754-4491-8350-68331ec6636e.png)
priority
alibaba cloud private warehouse cannot pull the image and automatically upgrade my docker compose yml yaml watchtower container name watchtower image containrrr watchtower restart always environment tz asia shanghai privileged true depends on mj admin service mj admin portal volumes var run docker sock var run docker sock root docker config json config json command mj admin service cleanup interval logging driver json file options max size max file the log shows that it is always ready and there is no more log output:
1
711,280
24,456,551,366
IssuesEvent
2022-10-07 07:20:09
HiAvatar/backend
https://api.github.com/repos/HiAvatar/backend
closed
Project API 관련 예외처리
Priority: Medium
### Description 프로젝트와 관련된 예외 처리를 수행한다. `@RestControllerAdvice`를 적용한 `ProjectControllerAdvice.class`에서 `@ExceptionHandler`를 사용하여 `ProjectController.class`의 API 로직으로부터 예외 처리 코드를 분리시키고, 필요한 예외 로직을 추가한다. <br> ### Todo List 공통 - [ ] 존재하지 않는 프로젝트를 조회하는 요청이 왔을 때 - [ ] 특정 파일을 찾을 수 없을 때 음성 관련 - [ ] base64 -> file 변환에 실패했을 때 - [ ] file -> base64 변환에 실패했을 때 - [ ] 음성 파일 생성이 실패 했을 때 - [ ] 음성 파일 URL을 찾을 수 없을 때 영상 관련 - [ ] 영상 파일 생성이 실패 했을 때 <br> ### e.t.c (Optional) (참조 및 링크 첨부하기)
1.0
Project API 관련 예외처리 - ### Description 프로젝트와 관련된 예외 처리를 수행한다. `@RestControllerAdvice`를 적용한 `ProjectControllerAdvice.class`에서 `@ExceptionHandler`를 사용하여 `ProjectController.class`의 API 로직으로부터 예외 처리 코드를 분리시키고, 필요한 예외 로직을 추가한다. <br> ### Todo List 공통 - [ ] 존재하지 않는 프로젝트를 조회하는 요청이 왔을 때 - [ ] 특정 파일을 찾을 수 없을 때 음성 관련 - [ ] base64 -> file 변환에 실패했을 때 - [ ] file -> base64 변환에 실패했을 때 - [ ] 음성 파일 생성이 실패 했을 때 - [ ] 음성 파일 URL을 찾을 수 없을 때 영상 관련 - [ ] 영상 파일 생성이 실패 했을 때 <br> ### e.t.c (Optional) (참조 및 링크 첨부하기)
priority
project api 관련 예외처리 description 프로젝트와 관련된 예외 처리를 수행한다 restcontrolleradvice 를 적용한 projectcontrolleradvice class 에서 exceptionhandler 를 사용하여 projectcontroller class 의 api 로직으로부터 예외 처리 코드를 분리시키고 필요한 예외 로직을 추가한다 todo list 공통 존재하지 않는 프로젝트를 조회하는 요청이 왔을 때 특정 파일을 찾을 수 없을 때 음성 관련 file 변환에 실패했을 때 file 변환에 실패했을 때 음성 파일 생성이 실패 했을 때 음성 파일 url을 찾을 수 없을 때 영상 관련 영상 파일 생성이 실패 했을 때 e t c optional 참조 및 링크 첨부하기
1
752,736
26,323,118,746
IssuesEvent
2023-01-10 02:36:42
LuanRT/YouTube.js
https://api.github.com/repos/LuanRT/YouTube.js
closed
V2.0.0 --- InnertubeError: Something went wrong at MusicDetailHeader!
bug good first issue Stale priority: medium
### Steps to reproduce I'm no sure why but calling `client.music.getAlbum();` with the id `"MPREb_Ml86XqtmqaJ"` crashes. ### Failure Logs ```shell InnertubeError: Something went wrong at MusicDetailHeader! This is a bug, please report it at https://github.com/LuanRT/YouTube.js/issues at Function.formatError (/app/node_modules/youtubei.js/dist/src/parser/index.js:317:22) at Function.parseItem (/app/node_modules/youtubei.js/dist/src/parser/index.js:268:22) at Function.parse (/app/node_modules/youtubei.js/dist/src/parser/index.js:310:53) at Function.parseResponse (/app/node_modules/youtubei.js/dist/src/parser/index.js:178:28) at new Album (/app/node_modules/youtubei.js/dist/src/parser/ytmusic/Album.js:27:67) at Music.<anonymous> (/app/node_modules/youtubei.js/dist/src/core/Music.js:121:20) at Generator.next (<anonymous>) at fulfilled (/app/node_modules/youtubei.js/dist/src/core/Music.js:5:58) at processTicksAndRejections (node:internal/process/task_queues:96:5) { info: { stack: "TypeError: Cannot read properties of undefined (reading 'text')\n" + ' at new MusicDetailHeader (/app/node_modules/youtubei.js/dist/src/parser/classes/MusicDetailHeader.js:17:88)\n' + ' at Function.parseItem (/app/node_modules/youtubei.js/dist/src/parser/index.js:263:32)\n' + ' at Function.parse (/app/node_modules/youtubei.js/dist/src/parser/index.js:310:53)\n' + ' at Function.parseResponse (/app/node_modules/youtubei.js/dist/src/parser/index.js:178:28)\n' + ' at new Album (/app/node_modules/youtubei.js/dist/src/parser/ytmusic/Album.js:27:67)\n' + ' at Music.<anonymous> (/app/node_modules/youtubei.js/dist/src/core/Music.js:121:20)\n' + ' at Generator.next (<anonymous>)\n' + ' at fulfilled (/app/node_modules/youtubei.js/dist/src/core/Music.js:5:58)\n' + ' at processTicksAndRejections (node:internal/process/task_queues:96:5)' }, date: 2022-10-06T17:47:24.863Z, version: '2.0.0' } ``` ### Expected behavior not crash ### Current behavior crashes ### Version Default ### Anything else? _No response_ ### Checklist - [X] I am running the latest version. - [X] I checked the documentation and found no answer. - [X] I have searched the existing issues and made sure this is not a duplicate. - [X] I have provided sufficient information.
1.0
V2.0.0 --- InnertubeError: Something went wrong at MusicDetailHeader! - ### Steps to reproduce I'm no sure why but calling `client.music.getAlbum();` with the id `"MPREb_Ml86XqtmqaJ"` crashes. ### Failure Logs ```shell InnertubeError: Something went wrong at MusicDetailHeader! This is a bug, please report it at https://github.com/LuanRT/YouTube.js/issues at Function.formatError (/app/node_modules/youtubei.js/dist/src/parser/index.js:317:22) at Function.parseItem (/app/node_modules/youtubei.js/dist/src/parser/index.js:268:22) at Function.parse (/app/node_modules/youtubei.js/dist/src/parser/index.js:310:53) at Function.parseResponse (/app/node_modules/youtubei.js/dist/src/parser/index.js:178:28) at new Album (/app/node_modules/youtubei.js/dist/src/parser/ytmusic/Album.js:27:67) at Music.<anonymous> (/app/node_modules/youtubei.js/dist/src/core/Music.js:121:20) at Generator.next (<anonymous>) at fulfilled (/app/node_modules/youtubei.js/dist/src/core/Music.js:5:58) at processTicksAndRejections (node:internal/process/task_queues:96:5) { info: { stack: "TypeError: Cannot read properties of undefined (reading 'text')\n" + ' at new MusicDetailHeader (/app/node_modules/youtubei.js/dist/src/parser/classes/MusicDetailHeader.js:17:88)\n' + ' at Function.parseItem (/app/node_modules/youtubei.js/dist/src/parser/index.js:263:32)\n' + ' at Function.parse (/app/node_modules/youtubei.js/dist/src/parser/index.js:310:53)\n' + ' at Function.parseResponse (/app/node_modules/youtubei.js/dist/src/parser/index.js:178:28)\n' + ' at new Album (/app/node_modules/youtubei.js/dist/src/parser/ytmusic/Album.js:27:67)\n' + ' at Music.<anonymous> (/app/node_modules/youtubei.js/dist/src/core/Music.js:121:20)\n' + ' at Generator.next (<anonymous>)\n' + ' at fulfilled (/app/node_modules/youtubei.js/dist/src/core/Music.js:5:58)\n' + ' at processTicksAndRejections (node:internal/process/task_queues:96:5)' }, date: 2022-10-06T17:47:24.863Z, version: '2.0.0' } ``` ### Expected behavior not crash ### Current behavior crashes ### Version Default ### Anything else? _No response_ ### Checklist - [X] I am running the latest version. - [X] I checked the documentation and found no answer. - [X] I have searched the existing issues and made sure this is not a duplicate. - [X] I have provided sufficient information.
priority
innertubeerror something went wrong at musicdetailheader steps to reproduce i m no sure why but calling client music getalbum with the id mpreb crashes failure logs shell innertubeerror something went wrong at musicdetailheader this is a bug please report it at at function formaterror app node modules youtubei js dist src parser index js at function parseitem app node modules youtubei js dist src parser index js at function parse app node modules youtubei js dist src parser index js at function parseresponse app node modules youtubei js dist src parser index js at new album app node modules youtubei js dist src parser ytmusic album js at music app node modules youtubei js dist src core music js at generator next at fulfilled app node modules youtubei js dist src core music js at processticksandrejections node internal process task queues info stack typeerror cannot read properties of undefined reading text n at new musicdetailheader app node modules youtubei js dist src parser classes musicdetailheader js n at function parseitem app node modules youtubei js dist src parser index js n at function parse app node modules youtubei js dist src parser index js n at function parseresponse app node modules youtubei js dist src parser index js n at new album app node modules youtubei js dist src parser ytmusic album js n at music app node modules youtubei js dist src core music js n at generator next n at fulfilled app node modules youtubei js dist src core music js n at processticksandrejections node internal process task queues date version expected behavior not crash current behavior crashes version default anything else no response checklist i am running the latest version i checked the documentation and found no answer i have searched the existing issues and made sure this is not a duplicate i have provided sufficient information
1
696,467
23,902,227,886
IssuesEvent
2022-09-08 19:56:34
PrefectHQ/prefect
https://api.github.com/repos/PrefectHQ/prefect
closed
Display intended run date time on flow run page
enhancement from:slack ui v2 priority:medium
### Prefect Version 2.x ### Describe the proposed behavior Display/expose the intended flow run date time in the table on the flow run details page. ### Describe the current behavior Currently only the Created and Updated date times are displayed on this table. ### Additional context From community Slack: When a Flow is running on a schedule and it runs late or is delayed for any reason, the date time recorded on the flow run instance in the Flow Run tab is the date time at which it was run not the actual intended time it was scheduled to run. Is there a way to find this information? I found a date time in the idempotency key in the flow instance information, however, i am not sure if this is its intended purpose
1.0
Display intended run date time on flow run page - ### Prefect Version 2.x ### Describe the proposed behavior Display/expose the intended flow run date time in the table on the flow run details page. ### Describe the current behavior Currently only the Created and Updated date times are displayed on this table. ### Additional context From community Slack: When a Flow is running on a schedule and it runs late or is delayed for any reason, the date time recorded on the flow run instance in the Flow Run tab is the date time at which it was run not the actual intended time it was scheduled to run. Is there a way to find this information? I found a date time in the idempotency key in the flow instance information, however, i am not sure if this is its intended purpose
priority
display intended run date time on flow run page prefect version x describe the proposed behavior display expose the intended flow run date time in the table on the flow run details page describe the current behavior currently only the created and updated date times are displayed on this table additional context from community slack when a flow is running on a schedule and it runs late or is delayed for any reason the date time recorded on the flow run instance in the flow run tab is the date time at which it was run not the actual intended time it was scheduled to run is there a way to find this information i found a date time in the idempotency key in the flow instance information however i am not sure if this is its intended purpose
1
608,381
18,837,044,628
IssuesEvent
2021-11-11 03:05:10
dhowe/AdNauseam
https://api.github.com/repos/dhowe/AdNauseam
closed
[3.11.5b1 ] Any changes in filter list activates "Activate the AdNauseam filter ?" warning
PRIORITY: Medium Bug
Whenever you disactivate/activate some filter in the filter list (for example, uBlock filters), after applying changes it displays: `Activate the AdNauseam filter ?` warning. When it should not. ![Screen Shot 2021-11-08 at 17 00 52](https://user-images.githubusercontent.com/4967860/140775562-f398296a-dba6-4724-9349-fc2ab0c0ea60.png)
1.0
[3.11.5b1 ] Any changes in filter list activates "Activate the AdNauseam filter ?" warning - Whenever you disactivate/activate some filter in the filter list (for example, uBlock filters), after applying changes it displays: `Activate the AdNauseam filter ?` warning. When it should not. ![Screen Shot 2021-11-08 at 17 00 52](https://user-images.githubusercontent.com/4967860/140775562-f398296a-dba6-4724-9349-fc2ab0c0ea60.png)
priority
any changes in filter list activates activate the adnauseam filter warning whenever you disactivate activate some filter in the filter list for example ublock filters after applying changes it displays activate the adnauseam filter warning when it should not
1
217,944
7,329,325,442
IssuesEvent
2018-03-05 04:09:36
tootsuite/mastodon
https://api.github.com/repos/tootsuite/mastodon
closed
Add badge for muted/blocked of user profile in last column
enhancement fit for beginners priority - medium ui
It would be great to add a badge (like the ```FOLLOWS YOU``` one) if the user is currently muted and/or blocked when displaying an user profile in the last column. * * * * - [x] I searched or browsed the repo’s other issues to ensure this is not a duplicate. - [ ] This bug happens on a [tagged release](https://github.com/tootsuite/mastodon/releases) and not on `master` (If you're a user, don't worry about this).
1.0
Add badge for muted/blocked of user profile in last column - It would be great to add a badge (like the ```FOLLOWS YOU``` one) if the user is currently muted and/or blocked when displaying an user profile in the last column. * * * * - [x] I searched or browsed the repo’s other issues to ensure this is not a duplicate. - [ ] This bug happens on a [tagged release](https://github.com/tootsuite/mastodon/releases) and not on `master` (If you're a user, don't worry about this).
priority
add badge for muted blocked of user profile in last column it would be great to add a badge like the follows you one if the user is currently muted and or blocked when displaying an user profile in the last column i searched or browsed the repo’s other issues to ensure this is not a duplicate this bug happens on a and not on master if you re a user don t worry about this
1
617,735
19,403,360,892
IssuesEvent
2021-12-19 15:28:26
ibissource/frank-flow
https://api.github.com/repos/ibissource/frank-flow
closed
Highlight the element in the canvas at which the user is editing in the Editor.
feature priority:medium work:medium
**Is your feature request related to a problem? Please describe.** Clicking on an element shows the code. But clicking on the code doesn't show the element. **Describe the solution you'd like** Do the same but the other way around. **Additional context** Panning to that position would be even better, but may be confusing.
1.0
Highlight the element in the canvas at which the user is editing in the Editor. - **Is your feature request related to a problem? Please describe.** Clicking on an element shows the code. But clicking on the code doesn't show the element. **Describe the solution you'd like** Do the same but the other way around. **Additional context** Panning to that position would be even better, but may be confusing.
priority
highlight the element in the canvas at which the user is editing in the editor is your feature request related to a problem please describe clicking on an element shows the code but clicking on the code doesn t show the element describe the solution you d like do the same but the other way around additional context panning to that position would be even better but may be confusing
1
514,281
14,936,523,807
IssuesEvent
2021-01-25 13:30:15
bounswe/bounswe2020group4
https://api.github.com/repos/bounswe/bounswe2020group4
opened
(BKND) Unit tests
Backend Coding Effort: Medium Priority: Medium
Unit tests will be added for email sending and report endpoints. Deadline 25.01.2021
1.0
(BKND) Unit tests - Unit tests will be added for email sending and report endpoints. Deadline 25.01.2021
priority
bknd unit tests unit tests will be added for email sending and report endpoints deadline
1
806,033
29,797,987,007
IssuesEvent
2023-06-16 05:20:28
SubnauticaModding/Nautilus
https://api.github.com/repos/SubnauticaModding/Nautilus
closed
Docs page on asynchronous prefab construction
Priority: Medium Type: Feature Request
**Describe the feature** Does not need to be elaborate, but a beginner's guide with examples and tips and tricks would be nice. We should also warn beginners that asynchronous prefabs are visible in the world at the origin while being constructed, so you need to set them inactive.
1.0
Docs page on asynchronous prefab construction - **Describe the feature** Does not need to be elaborate, but a beginner's guide with examples and tips and tricks would be nice. We should also warn beginners that asynchronous prefabs are visible in the world at the origin while being constructed, so you need to set them inactive.
priority
docs page on asynchronous prefab construction describe the feature does not need to be elaborate but a beginner s guide with examples and tips and tricks would be nice we should also warn beginners that asynchronous prefabs are visible in the world at the origin while being constructed so you need to set them inactive
1
813,159
30,447,060,693
IssuesEvent
2023-07-15 20:17:23
berkeli/My-Coursework-Planner
https://api.github.com/repos/berkeli/My-Coursework-Planner
opened
[TECH ED] What is an algorithm
🐂 Size Medium 🏕 Priority Mandatory 📅 Week 2 🎯 Topic Communication 📅 Fundamentals 🎯 Topic Problem-Solving
### Link to the coursework https://www.youtube.com/watch?v=cDA3_5982h8 ### Why are we doing this? Programming is all about algorithms. But what is an algorithm? Read [Introduction to Algorithms](https://github.com/thejsway/thejsway/blob/master/manuscript/intro02.md#introduction-to-algorithms) in the book the JS Way Watch [The Exact Instructions Challenge😄](https://www.youtube.com/watch?v=cDA3_5982h8) These show "algorithm recipes", and the video shows that you have to think carefully when building them. How does CYF Blocks help you build algorithms? Discuss this together in a Slack thread. ### Maximum time in hours 2 ### How to submit Discuss this together in a Slack thread
1.0
[TECH ED] What is an algorithm - ### Link to the coursework https://www.youtube.com/watch?v=cDA3_5982h8 ### Why are we doing this? Programming is all about algorithms. But what is an algorithm? Read [Introduction to Algorithms](https://github.com/thejsway/thejsway/blob/master/manuscript/intro02.md#introduction-to-algorithms) in the book the JS Way Watch [The Exact Instructions Challenge😄](https://www.youtube.com/watch?v=cDA3_5982h8) These show "algorithm recipes", and the video shows that you have to think carefully when building them. How does CYF Blocks help you build algorithms? Discuss this together in a Slack thread. ### Maximum time in hours 2 ### How to submit Discuss this together in a Slack thread
priority
what is an algorithm link to the coursework why are we doing this programming is all about algorithms but what is an algorithm read in the book the js way watch these show algorithm recipes and the video shows that you have to think carefully when building them how does cyf blocks help you build algorithms discuss this together in a slack thread maximum time in hours how to submit discuss this together in a slack thread
1