Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
438,071 | 12,610,637,765 | IssuesEvent | 2020-06-12 05:37:33 | confidantstation/Confidant-Station | https://api.github.com/repos/confidantstation/Confidant-Station | closed | Optimize confidant friends, file transfer data storage structure | Priority: Medium Status: In Progress Type: Enhancement | Currently, a user reserves n friends. The number of file transfers is changed to a global pool | 1.0 | Optimize confidant friends, file transfer data storage structure - Currently, a user reserves n friends. The number of file transfers is changed to a global pool | priority | optimize confidant friends file transfer data storage structure currently a user reserves n friends the number of file transfers is changed to a global pool | 1 |
554,487 | 16,430,524,636 | IssuesEvent | 2021-05-20 00:28:23 | MeAmAnUsername/pie | https://api.github.com/repos/MeAmAnUsername/pie | opened | Remove warning on single element in multi-import if element is renamed | Component: editor Priority: medium Status: specified Type: bug | Remove warning on single element in multi-import if element is renamed.
Could also add a note suggesting a rewrite if the renaming happens for a single import.
Example of how it should work
```
import a:b:{c}:d:{someFunc, someOtherFunc} // warning: single element multi-import (quick fix: remove braces)
import a:b:{c as something}:d:{someFunc, someOtherFunc} // no warning, currently there is a warning
import a:b:{c as something}:d:someFunc // optional note: could use a:b:c:d:someFunc as something:someFunc
```
| 1.0 | Remove warning on single element in multi-import if element is renamed - Remove warning on single element in multi-import if element is renamed.
Could also add a note suggesting a rewrite if the renaming happens for a single import.
Example of how it should work
```
import a:b:{c}:d:{someFunc, someOtherFunc} // warning: single element multi-import (quick fix: remove braces)
import a:b:{c as something}:d:{someFunc, someOtherFunc} // no warning, currently there is a warning
import a:b:{c as something}:d:someFunc // optional note: could use a:b:c:d:someFunc as something:someFunc
```
| priority | remove warning on single element in multi import if element is renamed remove warning on single element in multi import if element is renamed could also add a note suggesting a rewrite if the renaming happens for a single import example of how it should work import a b c d somefunc someotherfunc warning single element multi import quick fix remove braces import a b c as something d somefunc someotherfunc no warning currently there is a warning import a b c as something d somefunc optional note could use a b c d somefunc as something somefunc | 1 |
427,511 | 12,396,183,081 | IssuesEvent | 2020-05-20 20:03:02 | react-figma/react-figma | https://api.github.com/repos/react-figma/react-figma | closed | Create polyfills for functions as ’fetch’ in main thread | complexity: hard priority: medium type: feature or enhancement | Currently things such as `fetch` are undefined in the main execution thread where all components are rendered. We need to polyfill these functions using a bridge between UI and main threads | 1.0 | Create polyfills for functions as ’fetch’ in main thread - Currently things such as `fetch` are undefined in the main execution thread where all components are rendered. We need to polyfill these functions using a bridge between UI and main threads | priority | create polyfills for functions as ’fetch’ in main thread currently things such as fetch are undefined in the main execution thread where all components are rendered we need to polyfill these functions using a bridge between ui and main threads | 1 |
791,994 | 27,884,351,923 | IssuesEvent | 2023-03-21 22:13:25 | agrc/electrofishing | https://api.github.com/repos/agrc/electrofishing | closed | Gut Check Metrics Before Submission | waiting medium priority | >When submitting a report create a summary that allows for easy QA/QC using metrics like Condition Factor, Average length, Maximum/Minimum length by species that could serve as red flags. Prior to submitting report the verification of the data includes some measure of condition based on length and weight relationship for fish collected.
Ref 54) from https://docs.google.com/document/d/1pTsCfe0-Kmp20vSLy9DJtVIEFB_PwfoabOojmtVQXYM/edit
@ToddVolkening: Can you get us specifics on what they want? | 1.0 | Gut Check Metrics Before Submission - >When submitting a report create a summary that allows for easy QA/QC using metrics like Condition Factor, Average length, Maximum/Minimum length by species that could serve as red flags. Prior to submitting report the verification of the data includes some measure of condition based on length and weight relationship for fish collected.
Ref 54) from https://docs.google.com/document/d/1pTsCfe0-Kmp20vSLy9DJtVIEFB_PwfoabOojmtVQXYM/edit
@ToddVolkening: Can you get us specifics on what they want? | priority | gut check metrics before submission when submitting a report create a summary that allows for easy qa qc using metrics like condition factor average length maximum minimum length by species that could serve as red flags prior to submitting report the verification of the data includes some measure of condition based on length and weight relationship for fish collected ref from toddvolkening can you get us specifics on what they want | 1 |
57,809 | 3,083,990,705 | IssuesEvent | 2015-08-24 12:45:07 | StefanIsidorovic/salira | https://api.github.com/repos/StefanIsidorovic/salira | closed | Unary minus | auto-migrated Priority-Medium Type-Other | ```
Look at PPJ->ispit->aritmeticki izrazi
UNARY MINUS
(-1) have to be interpreted like negative number and (n-1) like
functor("minus", ...)
```
Original issue reported on code.google.com by `missuchi...@gmail.com` on 7 Jun 2015 at 2:13 | 1.0 | Unary minus - ```
Look at PPJ->ispit->aritmeticki izrazi
UNARY MINUS
(-1) have to be interpreted like negative number and (n-1) like
functor("minus", ...)
```
Original issue reported on code.google.com by `missuchi...@gmail.com` on 7 Jun 2015 at 2:13 | priority | unary minus look at ppj ispit aritmeticki izrazi unary minus have to be interpreted like negative number and n like functor minus original issue reported on code google com by missuchi gmail com on jun at | 1 |
339,862 | 10,263,332,877 | IssuesEvent | 2019-08-22 14:10:23 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | Separate MSTransferor and MSMonitor in different process | Enhancement Medium Priority New Feature ReqMgr2MS | **Impact of the new feature**
ReqMgr2MS
**Is your feature request related to a problem? Please describe.**
Once ReqMgr2MS hit production, it will be under heavy load and we better allocate a different CPU for each of those threads.
With the thread/service separation, we also need to take into consideration the log file that each of them is going to write to.
**Describe the solution you'd like**
Start two different process, one for MSTransferor and one for MSMonitor.
**Describe alternatives you've considered**
**Additional context**
Changes will also be required to the deployment repository. | 1.0 | Separate MSTransferor and MSMonitor in different process - **Impact of the new feature**
ReqMgr2MS
**Is your feature request related to a problem? Please describe.**
Once ReqMgr2MS hit production, it will be under heavy load and we better allocate a different CPU for each of those threads.
With the thread/service separation, we also need to take into consideration the log file that each of them is going to write to.
**Describe the solution you'd like**
Start two different process, one for MSTransferor and one for MSMonitor.
**Describe alternatives you've considered**
**Additional context**
Changes will also be required to the deployment repository. | priority | separate mstransferor and msmonitor in different process impact of the new feature is your feature request related to a problem please describe once hit production it will be under heavy load and we better allocate a different cpu for each of those threads with the thread service separation we also need to take into consideration the log file that each of them is going to write to describe the solution you d like start two different process one for mstransferor and one for msmonitor describe alternatives you ve considered additional context changes will also be required to the deployment repository | 1 |
517,384 | 15,008,360,404 | IssuesEvent | 2021-01-31 09:44:22 | bounswe/bounswe2020group9 | https://api.github.com/repos/bounswe/bounswe2020group9 | closed | iOS - Customer/ Add Review | Estimation - Medium Mobile Priority - High Status - Completed | Implement "adding reviews" feature.
As discussed in the meeting, customer shall be able to add review only for the products they have already purchased. Therefore, it should be carried out on the Orders Page.
**Deadline: 25.01.2021** | 1.0 | iOS - Customer/ Add Review - Implement "adding reviews" feature.
As discussed in the meeting, customer shall be able to add review only for the products they have already purchased. Therefore, it should be carried out on the Orders Page.
**Deadline: 25.01.2021** | priority | ios customer add review implement adding reviews feature as discussed in the meeting customer shall be able to add review only for the products they have already purchased therefore it should be carried out on the orders page deadline | 1 |
642,057 | 20,866,178,239 | IssuesEvent | 2022-03-22 07:26:31 | AY2122S2-CS2103T-T17-4/tp | https://api.github.com/repos/AY2122S2-CS2103T-T17-4/tp | closed | Add status for each person | type.Story priority.Medium | As an advanced user, I can check the status of a person so that I can focus on contacting people that have not yet been contacted. | 1.0 | Add status for each person - As an advanced user, I can check the status of a person so that I can focus on contacting people that have not yet been contacted. | priority | add status for each person as an advanced user i can check the status of a person so that i can focus on contacting people that have not yet been contacted | 1 |
246,493 | 7,895,376,819 | IssuesEvent | 2018-06-29 02:52:56 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Make IceT the default when available. | Expected Use: 3 - Occasional Feature Impact: 3 - Medium OS: All Priority: Normal Support Group: Any | Several developers (Mark Miller, Cyrus Harrison, Brad Whitlock, Tom Fogal and Eric Brugger) have been discussing making IceT the default when available. The consensus was that we should. Mark and Tom suggested it would still be nice to have a way to disable it for debugging and other purposes. Mark proposed adding a flag to disable it on the command line as well as control it through the GUI and CLI.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Eric Brugger
Original creation: 01/11/2016 10:44 am
Original update: 02/18/2016 12:45 pm
Ticket number: 2506 | 1.0 | Make IceT the default when available. - Several developers (Mark Miller, Cyrus Harrison, Brad Whitlock, Tom Fogal and Eric Brugger) have been discussing making IceT the default when available. The consensus was that we should. Mark and Tom suggested it would still be nice to have a way to disable it for debugging and other purposes. Mark proposed adding a flag to disable it on the command line as well as control it through the GUI and CLI.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Eric Brugger
Original creation: 01/11/2016 10:44 am
Original update: 02/18/2016 12:45 pm
Ticket number: 2506 | priority | make icet the default when available several developers mark miller cyrus harrison brad whitlock tom fogal and eric brugger have been discussing making icet the default when available the consensus was that we should mark and tom suggested it would still be nice to have a way to disable it for debugging and other purposes mark proposed adding a flag to disable it on the command line as well as control it through the gui and cli redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author eric brugger original creation am original update pm ticket number | 1 |
48,013 | 2,990,117,627 | IssuesEvent | 2015-07-21 07:02:47 | jayway/rest-assured | https://api.github.com/repos/jayway/rest-assured | closed | Lazily merge path arguments | bug imported invalid Priority-Medium | _From [johan.ha...@gmail.com](https://code.google.com/u/105676376875942041029/) on May 02, 2013 08:45:15_
Make this work:
... rootPath("x.y.%s.z").body("w", withArguments("u"), .. ).
_Original issue: http://code.google.com/p/rest-assured/issues/detail?id=232_ | 1.0 | Lazily merge path arguments - _From [johan.ha...@gmail.com](https://code.google.com/u/105676376875942041029/) on May 02, 2013 08:45:15_
Make this work:
... rootPath("x.y.%s.z").body("w", withArguments("u"), .. ).
_Original issue: http://code.google.com/p/rest-assured/issues/detail?id=232_ | priority | lazily merge path arguments from on may make this work rootpath x y s z body w witharguments u original issue | 1 |
525,033 | 15,228,015,416 | IssuesEvent | 2021-02-18 10:56:03 | ivpn/ios-app | https://api.github.com/repos/ivpn/ios-app | closed | UI issue regarding the circles animation when changing networks | Network Protection priority: medium type: bug | **Description:**
In the App Store build 2.0.4, as well as in the latest beta 2.1.0 (23), there is an UI issue with the connecting circle animation while changing networks and having the following network trust settings (Mobile data: Untrusted, WIFI: Trusted or vide versa).
When changing networks, it is observed two circles at the same time, one for the disconnected server and another one for the server connected to.
**Note:**
See attached video for further details.
**Actual Result:**
Two connecting circles when changing networks while having different trust status set from every network.
**Expected Result:**
Only one circle animation should be shown when connecting or disconnecting.
**Steps to reproduce:**
1. Over WIFI, install App Store version 2.0.4.
2. Login.
3. Select e.g. OpenVPN as protocol and connect to any server.
4. Enable Network Protection.
5. Set Mobile Data to Untrusted and WIFI network to Trusted.
6. Proceed to the main screen.
7. Change network from WIFI to Mobile Data.
8. Observe that two circles are displayed when connecting to the server.
9. Change network from Mobile Data to WIFI.
10. Observe that two circles are displayed when disconnecting from the server.
**Extra Info:**
* Device: iPhone XR
* OS name and version: iOS 14.3
* IVPN app version: App Store 2.0.4, Beta 2.1.0 (23)
https://user-images.githubusercontent.com/16647612/105837005-56420100-5fce-11eb-9f62-c9981cf373f6.MP4
| 1.0 | UI issue regarding the circles animation when changing networks - **Description:**
In the App Store build 2.0.4, as well as in the latest beta 2.1.0 (23), there is an UI issue with the connecting circle animation while changing networks and having the following network trust settings (Mobile data: Untrusted, WIFI: Trusted or vide versa).
When changing networks, it is observed two circles at the same time, one for the disconnected server and another one for the server connected to.
**Note:**
See attached video for further details.
**Actual Result:**
Two connecting circles when changing networks while having different trust status set from every network.
**Expected Result:**
Only one circle animation should be shown when connecting or disconnecting.
**Steps to reproduce:**
1. Over WIFI, install App Store version 2.0.4.
2. Login.
3. Select e.g. OpenVPN as protocol and connect to any server.
4. Enable Network Protection.
5. Set Mobile Data to Untrusted and WIFI network to Trusted.
6. Proceed to the main screen.
7. Change network from WIFI to Mobile Data.
8. Observe that two circles are displayed when connecting to the server.
9. Change network from Mobile Data to WIFI.
10. Observe that two circles are displayed when disconnecting from the server.
**Extra Info:**
* Device: iPhone XR
* OS name and version: iOS 14.3
* IVPN app version: App Store 2.0.4, Beta 2.1.0 (23)
https://user-images.githubusercontent.com/16647612/105837005-56420100-5fce-11eb-9f62-c9981cf373f6.MP4
| priority | ui issue regarding the circles animation when changing networks description in the app store build as well as in the latest beta there is an ui issue with the connecting circle animation while changing networks and having the following network trust settings mobile data untrusted wifi trusted or vide versa when changing networks it is observed two circles at the same time one for the disconnected server and another one for the server connected to note see attached video for further details actual result two connecting circles when changing networks while having different trust status set from every network expected result only one circle animation should be shown when connecting or disconnecting steps to reproduce over wifi install app store version login select e g openvpn as protocol and connect to any server enable network protection set mobile data to untrusted and wifi network to trusted proceed to the main screen change network from wifi to mobile data observe that two circles are displayed when connecting to the server change network from mobile data to wifi observe that two circles are displayed when disconnecting from the server extra info device iphone xr os name and version ios ivpn app version app store beta | 1 |
819,070 | 30,718,765,989 | IssuesEvent | 2023-07-27 14:36:13 | containrrr/watchtower | https://api.github.com/repos/containrrr/watchtower | opened | ghcr Private Repo Issue | Type: Bug Priority: Medium Status: Available | ### Describe the bug
I become a 404 Error because the URL what Watchtower tries to reach is unavailable.
### Steps to reproduce
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
### Expected behavior
That WatchTower checks the repo and loads updates.
### Screenshots
_No response_
### Environment
- Platform Synology DSM
- Architecture x64
- Docker Version 20.10.23
### Your logs
```text
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="No new images found for /ghcr-io-derrobin154-lurchbot-1"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="No pull needed. Skipping image."
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Found a match"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg=Comparing local="sha256:81c94cc8d71a4763316bd6623ee89e981d37936bdb7cc9eb76485a31bd44ee81" remote="sha256:81c94cc8d71a4763316bd6623ee89e981d37936bdb7cc9eb76485a31bd44ee81"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Found a remote digest to compare with" remote="sha256:81c94cc8d71a4763316bd6623ee89e981d37936bdb7cc9eb76485a31bd44ee81"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Doing a HEAD request to fetch a digest" url="https://ghcr.io/v2/derrobin154/lurchbot/manifests/master"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Parsing image ref" host=ghcr.io image=derrobin154/lurchbot normalized="ghcr.io/derrobin154/lurchbot:master" tag=master
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Credentials found."
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Setting scope for auth token" image=ghcr.io/derrobin154/lurchbot scope="repository:ghcr.io/derrobin154/lurchbot:pull"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Checking challenge header content" realm="https://ghcr.io/token" service=ghcr.io
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Got response to challenge request" header="Bearer realm=\"https://ghcr.io/token\",service=\"ghcr.io\",scope=\"repository:user/image:pull\"" status="401 Unauthorized"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Building challenge URL" URL="https://ghcr.io/v2/"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Checking if pull is needed" container=/ghcr-io-derrobin154-lurchbot-1 image="ghcr.io/derrobin154/lurchbot:master"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Credentials loaded"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Got image name: ghcr.io/derrobin154/lurchbot:master"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Loaded auth credentials for user derrobin154, on registry ghcr.io/derrobin154/lurchbot:master, from file /config.json"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Trying to load authentication credentials." container=/ghcr-io-derrobin154-lurchbot-1 image="ghcr.io/derrobin154/lurchbot:master"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Retrieving running, stopped and exited containers"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Checking containers for updated images"
```
### Additional context
I take the Auth Config from docker login and put it in config.json.
I also have the Problem, i download ex. the Watchtower Image from DockerHub over DSM and in Watchtower i become this error:
2023/07/27 16:32:27 | stderr | time="2023-07-27T14:32:27Z" level=debug msg="No credentials for containrrr found" config_file=/config.json
2023/07/27 16:32:27 | stderr | time="2023-07-27T14:32:27Z" level=debug msg="Trying to load authentication credentials." container=/watchtower image="containrrr/watchtower:latest"
because the Image name is containerrr/watchtower and in the config.json containerrr is not set because it is from dockerhub. | 1.0 | ghcr Private Repo Issue - ### Describe the bug
I become a 404 Error because the URL what Watchtower tries to reach is unavailable.
### Steps to reproduce
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
### Expected behavior
That WatchTower checks the repo and loads updates.
### Screenshots
_No response_
### Environment
- Platform Synology DSM
- Architecture x64
- Docker Version 20.10.23
### Your logs
```text
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="No new images found for /ghcr-io-derrobin154-lurchbot-1"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="No pull needed. Skipping image."
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Found a match"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg=Comparing local="sha256:81c94cc8d71a4763316bd6623ee89e981d37936bdb7cc9eb76485a31bd44ee81" remote="sha256:81c94cc8d71a4763316bd6623ee89e981d37936bdb7cc9eb76485a31bd44ee81"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Found a remote digest to compare with" remote="sha256:81c94cc8d71a4763316bd6623ee89e981d37936bdb7cc9eb76485a31bd44ee81"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Doing a HEAD request to fetch a digest" url="https://ghcr.io/v2/derrobin154/lurchbot/manifests/master"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Parsing image ref" host=ghcr.io image=derrobin154/lurchbot normalized="ghcr.io/derrobin154/lurchbot:master" tag=master
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Credentials found."
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Setting scope for auth token" image=ghcr.io/derrobin154/lurchbot scope="repository:ghcr.io/derrobin154/lurchbot:pull"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Checking challenge header content" realm="https://ghcr.io/token" service=ghcr.io
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Got response to challenge request" header="Bearer realm=\"https://ghcr.io/token\",service=\"ghcr.io\",scope=\"repository:user/image:pull\"" status="401 Unauthorized"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Building challenge URL" URL="https://ghcr.io/v2/"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Checking if pull is needed" container=/ghcr-io-derrobin154-lurchbot-1 image="ghcr.io/derrobin154/lurchbot:master"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Credentials loaded"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Got image name: ghcr.io/derrobin154/lurchbot:master"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Loaded auth credentials for user derrobin154, on registry ghcr.io/derrobin154/lurchbot:master, from file /config.json"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Trying to load authentication credentials." container=/ghcr-io-derrobin154-lurchbot-1 image="ghcr.io/derrobin154/lurchbot:master"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Retrieving running, stopped and exited containers"
2023/07/27 16:32:23 stderr time="2023-07-27T14:32:23Z" level=debug msg="Checking containers for updated images"
```
### Additional context
I take the Auth Config from docker login and put it in config.json.
I also have the Problem, i download ex. the Watchtower Image from DockerHub over DSM and in Watchtower i become this error:
2023/07/27 16:32:27 | stderr | time="2023-07-27T14:32:27Z" level=debug msg="No credentials for containrrr found" config_file=/config.json
2023/07/27 16:32:27 | stderr | time="2023-07-27T14:32:27Z" level=debug msg="Trying to load authentication credentials." container=/watchtower image="containrrr/watchtower:latest"
because the Image name is containerrr/watchtower and in the config.json containerrr is not set because it is from dockerhub. | priority | ghcr private repo issue describe the bug i become a error because the url what watchtower tries to reach is unavailable steps to reproduce go to click on scroll down to see error expected behavior that watchtower checks the repo and loads updates screenshots no response environment platform synology dsm architecture docker version your logs text stderr time level debug msg no new images found for ghcr io lurchbot stderr time level debug msg no pull needed skipping image stderr time level debug msg found a match stderr time level debug msg comparing local remote stderr time level debug msg found a remote digest to compare with remote stderr time level debug msg doing a head request to fetch a digest url stderr time level debug msg parsing image ref host ghcr io image lurchbot normalized ghcr io lurchbot master tag master stderr time level debug msg credentials found stderr time level debug msg setting scope for auth token image ghcr io lurchbot scope repository ghcr io lurchbot pull stderr time level debug msg checking challenge header content realm service ghcr io stderr time level debug msg got response to challenge request header bearer realm status unauthorized stderr time level debug msg building challenge url url stderr time level debug msg checking if pull is needed container ghcr io lurchbot image ghcr io lurchbot master stderr time level debug msg credentials loaded stderr time level debug msg got image name ghcr io lurchbot master stderr time level debug msg loaded auth credentials for user on registry ghcr io lurchbot master from file config json stderr time level debug msg trying to load authentication credentials container ghcr io lurchbot image ghcr io lurchbot master stderr time level debug msg retrieving running stopped and exited containers stderr time level debug msg checking containers for updated images additional context i take the auth config from docker login and put it in config json i also have the problem i download ex the watchtower image from dockerhub over dsm and in watchtower i become this error stderr time level debug msg no credentials for containrrr found config file config json stderr time level debug msg trying to load authentication credentials container watchtower image containrrr watchtower latest because the image name is containerrr watchtower and in the config json containerrr is not set because it is from dockerhub | 1 |
69,965 | 3,316,353,453 | IssuesEvent | 2015-11-06 16:33:13 | TeselaGen/Peony-Issue-Tracking | https://api.github.com/repos/TeselaGen/Peony-Issue-Tracking | opened | Rationalize right-click menu for all library views | Customer: DAS Phase I Priority: Medium Status: In Progress Type: Enhancement | _From @mfero on September 24, 2015 20:37_
We should rationalize the right-click menu across all library views. Menu items specific to a particular library can sit below a separator.
My Protocols: Rename, Edit, Delete, Create Copy, Export
My Strains: Rename, Edit, Delete, Create Copy, Export
My Sequence: Rename, Edit, Delete, Create Copy, Export || Create Parts from Sequences, Create Parts from Sequence Features
My Amino Acid Sequence: Rename, Edit, Delete, Create Copy, Export
My Parts: Rename, Edit, Delete, Create Copy, Export
My Projects: Rename, Edit, Delete, Create Copy, Export
_Copied from original issue: TeselaGen/ve#1390_ | 1.0 | Rationalize right-click menu for all library views - _From @mfero on September 24, 2015 20:37_
We should rationalize the right-click menu across all library views. Menu items specific to a particular library can sit below a separator.
My Protocols: Rename, Edit, Delete, Create Copy, Export
My Strains: Rename, Edit, Delete, Create Copy, Export
My Sequence: Rename, Edit, Delete, Create Copy, Export || Create Parts from Sequences, Create Parts from Sequence Features
My Amino Acid Sequence: Rename, Edit, Delete, Create Copy, Export
My Parts: Rename, Edit, Delete, Create Copy, Export
My Projects: Rename, Edit, Delete, Create Copy, Export
_Copied from original issue: TeselaGen/ve#1390_ | priority | rationalize right click menu for all library views from mfero on september we should rationalize the right click menu across all library views menu items specific to a particular library can sit below a separator my protocols rename edit delete create copy export my strains rename edit delete create copy export my sequence rename edit delete create copy export create parts from sequences create parts from sequence features my amino acid sequence rename edit delete create copy export my parts rename edit delete create copy export my projects rename edit delete create copy export copied from original issue teselagen ve | 1 |
685,290 | 23,451,453,556 | IssuesEvent | 2022-08-16 03:31:39 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YSQL] Inequality is shown to be getting pushed down to docDB when yb_enable_optimizer_statistics is set to true | kind/bug area/ysql priority/medium | Jira Link: [DB-2897](https://yugabyte.atlassian.net/browse/DB-2897)
### Description
When `yb_enable_optimizer_statistics` is set to true, an inequality filter on the hash indexed primary key seems to be pushed down to docDB. The query plan shows an Index Scan. A range index scan cannot be performed on the hash index and so this is not correct.
```sql
CREATE TABLE test(k int, v int, primary key (k HASH)) SPLIT INTO 10 TABLETS;
INSERT INTO test SELECT * FROM generate_series(1,1000);
-- yb_enable_optimizer_statistics is false by default
SET yb_enable_optimizer_statistics = false;
EXPLAIN SELECT * FROM test where k > 500;
```
```
QUERY PLAN
----------------------------------------------------
Seq Scan on test (cost=0.00..0.00 rows=1 width=8)
Filter: (k > 500)
```
^ This is the correct plan.
```sql
SET yb_enable_optimizer_statistics = true;
ANALYSE test;
EXPLAIN SELECT * FROM test where k > 500;
```
```
QUERY PLAN
-------------------------------------------------------------------------
Index Scan using test_pkey on test (cost=0.00..60.25 rows=500 width=8)
Index Cond: (k > 500)
```
^ After enabling the feature and calling ANALYSE to compute statistics, the query planner proposes an Index Scan which is not possible.
However, the results of the query seem correct. The request sent over to docDB also shows that we do not pursue the Index Scan and indeed perform a Sequential Scan. | 1.0 | [YSQL] Inequality is shown to be getting pushed down to docDB when yb_enable_optimizer_statistics is set to true - Jira Link: [DB-2897](https://yugabyte.atlassian.net/browse/DB-2897)
### Description
When `yb_enable_optimizer_statistics` is set to true, an inequality filter on the hash indexed primary key seems to be pushed down to docDB. The query plan shows an Index Scan. A range index scan cannot be performed on the hash index and so this is not correct.
```sql
CREATE TABLE test(k int, v int, primary key (k HASH)) SPLIT INTO 10 TABLETS;
INSERT INTO test SELECT * FROM generate_series(1,1000);
-- yb_enable_optimizer_statistics is false by default
SET yb_enable_optimizer_statistics = false;
EXPLAIN SELECT * FROM test where k > 500;
```
```
QUERY PLAN
----------------------------------------------------
Seq Scan on test (cost=0.00..0.00 rows=1 width=8)
Filter: (k > 500)
```
^ This is the correct plan.
```sql
SET yb_enable_optimizer_statistics = true;
ANALYSE test;
EXPLAIN SELECT * FROM test where k > 500;
```
```
QUERY PLAN
-------------------------------------------------------------------------
Index Scan using test_pkey on test (cost=0.00..60.25 rows=500 width=8)
Index Cond: (k > 500)
```
^ After enabling the feature and calling ANALYSE to compute statistics, the query planner proposes an Index Scan which is not possible.
However, the results of the query seem correct. The request sent over to docDB also shows that we do not pursue the Index Scan and indeed perform a Sequential Scan. | priority | inequality is shown to be getting pushed down to docdb when yb enable optimizer statistics is set to true jira link description when yb enable optimizer statistics is set to true an inequality filter on the hash indexed primary key seems to be pushed down to docdb the query plan shows an index scan a range index scan cannot be performed on the hash index and so this is not correct sql create table test k int v int primary key k hash split into tablets insert into test select from generate series yb enable optimizer statistics is false by default set yb enable optimizer statistics false explain select from test where k query plan seq scan on test cost rows width filter k this is the correct plan sql set yb enable optimizer statistics true analyse test explain select from test where k query plan index scan using test pkey on test cost rows width index cond k after enabling the feature and calling analyse to compute statistics the query planner proposes an index scan which is not possible however the results of the query seem correct the request sent over to docdb also shows that we do not pursue the index scan and indeed perform a sequential scan | 1 |
213,178 | 7,246,466,448 | IssuesEvent | 2018-02-14 21:45:59 | Motoxpro/WorldCupStatsSite | https://api.github.com/repos/Motoxpro/WorldCupStatsSite | closed | Need to get track length for world championships | Medium Priority Data Issue | need to download and parse the worlds pdfs for track length | 1.0 | Need to get track length for world championships - need to download and parse the worlds pdfs for track length | priority | need to get track length for world championships need to download and parse the worlds pdfs for track length | 1 |
160,765 | 6,102,038,199 | IssuesEvent | 2017-06-20 15:41:47 | OperationCode/operationcode_frontend | https://api.github.com/repos/OperationCode/operationcode_frontend | closed | Add sentry | beginner friendly Priority: Medium Status: Available Type: Feature | <!-- Please fill out one of the sections below based on the type of issue you're creating -->
# Feature
## Why is this feature being added?
<!-- What problem is it solving? What value does it add? -->
Sentry helps track errors.
## What should your feature do?
Follow https://sentry.io/for/react/
Our public DN is `https://23e9c0c37a3841b8a692c98b8978f515@sentry.io/147247` | 1.0 | Add sentry - <!-- Please fill out one of the sections below based on the type of issue you're creating -->
# Feature
## Why is this feature being added?
<!-- What problem is it solving? What value does it add? -->
Sentry helps track errors.
## What should your feature do?
Follow https://sentry.io/for/react/
Our public DN is `https://23e9c0c37a3841b8a692c98b8978f515@sentry.io/147247` | priority | add sentry feature why is this feature being added sentry helps track errors what should your feature do follow our public dn is | 1 |
659,086 | 21,916,230,207 | IssuesEvent | 2022-05-21 21:42:06 | SkriptLang/Skript | https://api.github.com/repos/SkriptLang/Skript | closed | Cant remove saturation effect | bug priority: medium completed | ### Skript/Server Version
```
[23:45:09 INFO]: [Skript] Skript's aliases can be found here: https://github.com/SkriptLang/skript-aliases
[23:45:09 INFO]: [Skript] Skript's documentation can be found here: https://skriptlang.github.io/Skript
[23:45:09 INFO]: [Skript] Server Version: git-Purpur-"4ff0630" (MC: 1.17.1)
[23:45:09 INFO]: [Skript] Skript Version: 2.6.1
[23:45:09 INFO]: [Skript] Installed Skript Addons:
[23:45:09 INFO]: [Skript] - skript-yaml v1.4
[23:45:09 INFO]: [Skript] - skript-reflect v2.2.3 (https://github.com/TPGamesNL/skript-reflect)
[23:45:09 INFO]: [Skript] - Skore v2.0.2-BETA (1.17 only)
[23:45:09 INFO]: [Skript] Installed dependencies: None
```
### Bug Description
i can give/apply the saturation without any problems.
but I can't remove the effect from the player.
it just throws a `Can't understand this condition/effect: remove saturation from player (test.sk, line 2: remove saturation from player')`
it should be fixed in https://github.com/SkriptLang/Skript/pull/1142 but its not. (see issue https://github.com/SkriptLang/Skript/issues/692)
The Expession below probably conflicts with the effect name.
https://github.com/SkriptLang/Skript/blob/d09320866fa15bd6f2fc77b13f3e0cfd2c46ef36/src/main/java/ch/njol/skript/expressions/ExprSaturation.java
also the potion effect isnt listed in the aliases https://github.com/SkriptLang/skript-aliases/blob/master/brewing.sk
Potion Effect Types: https://skripthub.net/docs/?id=2161
Saturation Expression: https://skripthub.net/docs/?id=876
### Expected Behavior
remove the potion effect from the player
### Steps to Reproduce
```
on rightclick with any pickaxe:
set {_p} to player
remove saturation from {_p} # gives no errors, but none of them are working
remove saturation from player
```
### Errors or Screenshots
```
[23:44:37 ERROR]: [Skript] Line 20: (Friends.sk)
Can't understand this condition/effect: remove saturation from player
Line: remove saturation from player
```
### Other
_No response_
### Agreement
- [x] I have read the guidelines above and confirm I am following them with this report. | 1.0 | Cant remove saturation effect - ### Skript/Server Version
```
[23:45:09 INFO]: [Skript] Skript's aliases can be found here: https://github.com/SkriptLang/skript-aliases
[23:45:09 INFO]: [Skript] Skript's documentation can be found here: https://skriptlang.github.io/Skript
[23:45:09 INFO]: [Skript] Server Version: git-Purpur-"4ff0630" (MC: 1.17.1)
[23:45:09 INFO]: [Skript] Skript Version: 2.6.1
[23:45:09 INFO]: [Skript] Installed Skript Addons:
[23:45:09 INFO]: [Skript] - skript-yaml v1.4
[23:45:09 INFO]: [Skript] - skript-reflect v2.2.3 (https://github.com/TPGamesNL/skript-reflect)
[23:45:09 INFO]: [Skript] - Skore v2.0.2-BETA (1.17 only)
[23:45:09 INFO]: [Skript] Installed dependencies: None
```
### Bug Description
i can give/apply the saturation without any problems.
but I can't remove the effect from the player.
it just throws a `Can't understand this condition/effect: remove saturation from player (test.sk, line 2: remove saturation from player')`
it should be fixed in https://github.com/SkriptLang/Skript/pull/1142 but its not. (see issue https://github.com/SkriptLang/Skript/issues/692)
The Expession below probably conflicts with the effect name.
https://github.com/SkriptLang/Skript/blob/d09320866fa15bd6f2fc77b13f3e0cfd2c46ef36/src/main/java/ch/njol/skript/expressions/ExprSaturation.java
also the potion effect isnt listed in the aliases https://github.com/SkriptLang/skript-aliases/blob/master/brewing.sk
Potion Effect Types: https://skripthub.net/docs/?id=2161
Saturation Expression: https://skripthub.net/docs/?id=876
### Expected Behavior
remove the potion effect from the player
### Steps to Reproduce
```
on rightclick with any pickaxe:
set {_p} to player
remove saturation from {_p} # gives no errors, but none of them are working
remove saturation from player
```
### Errors or Screenshots
```
[23:44:37 ERROR]: [Skript] Line 20: (Friends.sk)
Can't understand this condition/effect: remove saturation from player
Line: remove saturation from player
```
### Other
_No response_
### Agreement
- [x] I have read the guidelines above and confirm I am following them with this report. | priority | cant remove saturation effect skript server version skript s aliases can be found here skript s documentation can be found here server version git purpur mc skript version installed skript addons skript yaml skript reflect skore beta only installed dependencies none bug description i can give apply the saturation without any problems but i can t remove the effect from the player it just throws a can t understand this condition effect remove saturation from player test sk line remove saturation from player it should be fixed in but its not see issue the expession below probably conflicts with the effect name also the potion effect isnt listed in the aliases potion effect types saturation expression expected behavior remove the potion effect from the player steps to reproduce on rightclick with any pickaxe set p to player remove saturation from p gives no errors but none of them are working remove saturation from player errors or screenshots line friends sk can t understand this condition effect remove saturation from player line remove saturation from player other no response agreement i have read the guidelines above and confirm i am following them with this report | 1 |
749,535 | 26,166,800,480 | IssuesEvent | 2023-01-01 11:45:47 | docker-mailserver/docker-mailserver | https://api.github.com/repos/docker-mailserver/docker-mailserver | opened | [BUG] postfix: reject_unknown_client_hostname prevents legitimate mail from being received | kind/bug meta/needs triage priority/medium | ### Miscellaneous first checks
- [X] I checked that all ports are open and not blocked by my ISP / hosting provider.
- [X] I know that SSL errors are likely the result of a wrong setup on the user side and not caused by DMS itself. I'm confident my setup is correct.
### Affected Component(s)
postfix
### What happened and when does this occur?
```Markdown
mailserver | Jan 1 12:35:19 mail postfix/smtpd[1556]: NOQUEUE: reject: RCPT from unknown[54.240.85.218]: 450 4.7.25 Client host rejected: cannot find your hostname, [54.240.85.218]; from=<010101856d1d37d3-cb9c308b-0000-0000-0000-a21593c7994e-000000@bounce.stripe.com> to=<my-localpart+stripe@example.com> proto=ESMTP helo=<a85-218.smtp-out.us-west-2.amazonses.com>
```
```
### What did you expect to happen?
```Markdown
I expected an option to turn `reject_unknown_client_hostname` off, if it is not considered necessary. I can overwrite the config in the docker-compose, but that feels hacky.
```
### How do we replicate the issue?
```Markdown
I have absolutely no idea. I don't think this happens for all SES mails.
```
### DMS version
latest (d5db200744f1)
### What operating system is DMS running on?
Linux
### Which operating system version?
Archlinux
### What instruction set architecture is DMS running on?
AMD64 / x86_64
### What container orchestration tool are you using?
Docker Compose
### docker-compose.yml
```yml
moritz@pan ❯ cat docker-compose.yml docker-mailserver -> master + ? ! $
services:
mailserver:
image: docker.io/mailserver/docker-mailserver:latest
#entrypoint: ["/bin/sh", "-c", "while true; do sleep 10; done"]
container_name: mailserver
# If the FQDN for your mail-server is only two labels (eg: example.com),
# you can assign this entirely to `hostname` and remove `domainname`.
hostname: mail
domainname: moritz.sh
env_file: mailserver.env
network_mode: host
# More information about the mail-server ports:
# https://docker-mailserver.github.io/docker-mailserver/edge/config/security/understanding-the-ports/
# To avoid conflicts with yaml base-60 float, DO NOT remove the quotation marks.
ports:
- "25:25" # SMTP (explicit TLS => STARTTLS)
- "143:143" # IMAP4 (explicit TLS => STARTTLS)
- "465:465" # ESMTP (implicit TLS)
- "587:587" # ESMTP (explicit TLS => STARTTLS)
- "993:993" # IMAP4 (implicit TLS)
- "4190:4190" # MANAGE SIEVE
volumes:
- ./docker-data/dms/mail-data/:/var/mail/
- ./docker-data/dms/mail-state/:/var/mail-state/
- ./docker-data/dms/mail-logs/:/var/log/mail/
- ./config/:/tmp/docker-mailserver/
- ./docker-data/dms/config/dovecot/fts-xapian-plugin.conf:/etc/dovecot/conf.d/10-plugin.conf:ro
- /etc/localtime:/etc/localtime:ro
- /etc/letsencrypt:/etc/letsencrypt
restart: always
stop_grace_period: 1m
cap_add:
- NET_ADMIN
healthcheck:
test: "ss --listening --tcp | grep -P 'LISTEN.+:smtp' || exit 1"
timeout: 3s
retries: 0
```
### Relevant log output
```bash
mailserver | Jan 1 12:35:19 mail postfix/smtpd[1556]: NOQUEUE: reject: RCPT from unknown[54.240.85.218]: 450 4.7.25 Client host rejected: cannot find your hostname, [54.240.85.218]; from=<010101856d1d37d3-cb9c308b-0000-0000-0000-a21593c7994e-000000@bounce.stripe.com> to=<my-localpart+stripe@example.com> proto=ESMTP helo=<a85-218.smtp-out.us-west-2.amazonses.com>
```
```
### Other relevant information
_No response_
### What level of experience do you have with Docker and mail servers?
- [ ] I am inexperienced with docker
- [X] I am rather experienced with docker
- [ ] I am inexperienced with mail servers
- [ ] I am rather experienced with mail servers
- [ ] I am uncomfortable with the CLI
- [X] I am rather comfortable with the CLI
### Code of conduct
- [X] I have read this project's [Code of Conduct](https://github.com/docker-mailserver/docker-mailserver/blob/master/CODE_OF_CONDUCT.md) and I agree
- [X] I have read the [README](https://github.com/docker-mailserver/docker-mailserver/blob/master/README.md) and the [documentation](https://docker-mailserver.github.io/docker-mailserver/edge/) and I searched the [issue tracker](https://github.com/docker-mailserver/docker-mailserver/issues?q=is%3Aissue) but could not find a solution
### Improvements to this form?
_No response_ | 1.0 | [BUG] postfix: reject_unknown_client_hostname prevents legitimate mail from being received - ### Miscellaneous first checks
- [X] I checked that all ports are open and not blocked by my ISP / hosting provider.
- [X] I know that SSL errors are likely the result of a wrong setup on the user side and not caused by DMS itself. I'm confident my setup is correct.
### Affected Component(s)
postfix
### What happened and when does this occur?
```Markdown
mailserver | Jan 1 12:35:19 mail postfix/smtpd[1556]: NOQUEUE: reject: RCPT from unknown[54.240.85.218]: 450 4.7.25 Client host rejected: cannot find your hostname, [54.240.85.218]; from=<010101856d1d37d3-cb9c308b-0000-0000-0000-a21593c7994e-000000@bounce.stripe.com> to=<my-localpart+stripe@example.com> proto=ESMTP helo=<a85-218.smtp-out.us-west-2.amazonses.com>
```
```
### What did you expect to happen?
```Markdown
I expected an option to turn `reject_unknown_client_hostname` off, if it is not considered necessary. I can overwrite the config in the docker-compose, but that feels hacky.
```
### How do we replicate the issue?
```Markdown
I have absolutely no idea. I don't think this happens for all SES mails.
```
### DMS version
latest (d5db200744f1)
### What operating system is DMS running on?
Linux
### Which operating system version?
Archlinux
### What instruction set architecture is DMS running on?
AMD64 / x86_64
### What container orchestration tool are you using?
Docker Compose
### docker-compose.yml
```yml
moritz@pan ❯ cat docker-compose.yml docker-mailserver -> master + ? ! $
services:
mailserver:
image: docker.io/mailserver/docker-mailserver:latest
#entrypoint: ["/bin/sh", "-c", "while true; do sleep 10; done"]
container_name: mailserver
# If the FQDN for your mail-server is only two labels (eg: example.com),
# you can assign this entirely to `hostname` and remove `domainname`.
hostname: mail
domainname: moritz.sh
env_file: mailserver.env
network_mode: host
# More information about the mail-server ports:
# https://docker-mailserver.github.io/docker-mailserver/edge/config/security/understanding-the-ports/
# To avoid conflicts with yaml base-60 float, DO NOT remove the quotation marks.
ports:
- "25:25" # SMTP (explicit TLS => STARTTLS)
- "143:143" # IMAP4 (explicit TLS => STARTTLS)
- "465:465" # ESMTP (implicit TLS)
- "587:587" # ESMTP (explicit TLS => STARTTLS)
- "993:993" # IMAP4 (implicit TLS)
- "4190:4190" # MANAGE SIEVE
volumes:
- ./docker-data/dms/mail-data/:/var/mail/
- ./docker-data/dms/mail-state/:/var/mail-state/
- ./docker-data/dms/mail-logs/:/var/log/mail/
- ./config/:/tmp/docker-mailserver/
- ./docker-data/dms/config/dovecot/fts-xapian-plugin.conf:/etc/dovecot/conf.d/10-plugin.conf:ro
- /etc/localtime:/etc/localtime:ro
- /etc/letsencrypt:/etc/letsencrypt
restart: always
stop_grace_period: 1m
cap_add:
- NET_ADMIN
healthcheck:
test: "ss --listening --tcp | grep -P 'LISTEN.+:smtp' || exit 1"
timeout: 3s
retries: 0
```
### Relevant log output
```bash
mailserver | Jan 1 12:35:19 mail postfix/smtpd[1556]: NOQUEUE: reject: RCPT from unknown[54.240.85.218]: 450 4.7.25 Client host rejected: cannot find your hostname, [54.240.85.218]; from=<010101856d1d37d3-cb9c308b-0000-0000-0000-a21593c7994e-000000@bounce.stripe.com> to=<my-localpart+stripe@example.com> proto=ESMTP helo=<a85-218.smtp-out.us-west-2.amazonses.com>
```
```
### Other relevant information
_No response_
### What level of experience do you have with Docker and mail servers?
- [ ] I am inexperienced with docker
- [X] I am rather experienced with docker
- [ ] I am inexperienced with mail servers
- [ ] I am rather experienced with mail servers
- [ ] I am uncomfortable with the CLI
- [X] I am rather comfortable with the CLI
### Code of conduct
- [X] I have read this project's [Code of Conduct](https://github.com/docker-mailserver/docker-mailserver/blob/master/CODE_OF_CONDUCT.md) and I agree
- [X] I have read the [README](https://github.com/docker-mailserver/docker-mailserver/blob/master/README.md) and the [documentation](https://docker-mailserver.github.io/docker-mailserver/edge/) and I searched the [issue tracker](https://github.com/docker-mailserver/docker-mailserver/issues?q=is%3Aissue) but could not find a solution
### Improvements to this form?
_No response_ | priority | postfix reject unknown client hostname prevents legitimate mail from being received miscellaneous first checks i checked that all ports are open and not blocked by my isp hosting provider i know that ssl errors are likely the result of a wrong setup on the user side and not caused by dms itself i m confident my setup is correct affected component s postfix what happened and when does this occur markdown mailserver jan mail postfix smtpd noqueue reject rcpt from unknown client host rejected cannot find your hostname from to proto esmtp helo what did you expect to happen markdown i expected an option to turn reject unknown client hostname off if it is not considered necessary i can overwrite the config in the docker compose but that feels hacky how do we replicate the issue markdown i have absolutely no idea i don t think this happens for all ses mails dms version latest what operating system is dms running on linux which operating system version archlinux what instruction set architecture is dms running on what container orchestration tool are you using docker compose docker compose yml yml moritz pan ❯ cat docker compose yml docker mailserver master services mailserver image docker io mailserver docker mailserver latest entrypoint container name mailserver if the fqdn for your mail server is only two labels eg example com you can assign this entirely to hostname and remove domainname hostname mail domainname moritz sh env file mailserver env network mode host more information about the mail server ports to avoid conflicts with yaml base float do not remove the quotation marks ports smtp explicit tls starttls explicit tls starttls esmtp implicit tls esmtp explicit tls starttls implicit tls manage sieve volumes docker data dms mail data var mail docker data dms mail state var mail state docker data dms mail logs var log mail config tmp docker mailserver docker data dms config dovecot fts xapian plugin conf etc dovecot conf d plugin conf ro etc localtime etc localtime ro etc letsencrypt etc letsencrypt restart always stop grace period cap add net admin healthcheck test ss listening tcp grep p listen smtp exit timeout retries relevant log output bash mailserver jan mail postfix smtpd noqueue reject rcpt from unknown client host rejected cannot find your hostname from to proto esmtp helo other relevant information no response what level of experience do you have with docker and mail servers i am inexperienced with docker i am rather experienced with docker i am inexperienced with mail servers i am rather experienced with mail servers i am uncomfortable with the cli i am rather comfortable with the cli code of conduct i have read this project s and i agree i have read the and the and i searched the but could not find a solution improvements to this form no response | 1 |
617,243 | 19,345,995,265 | IssuesEvent | 2021-12-15 10:51:46 | google/android-fhir | https://api.github.com/repos/google/android-fhir | closed | Support evaluation of FHIRPath expressions and calculation within Questionnaire | enhancement medium priority Q4 2021 | **Is your feature request related to a problem? Please describe.**
There is a need to be able to evaluate expressions within a Questionnaire. This is currently not supported by the data-capture library
**Describe the solution you'd like**
Support for FHIRPath expressions
**Describe alternatives you've considered**
Other options include CQL.
**Additional context**
@maimoonak - can you provide some example use cases for the types of expressions you are trying to evaluate | 1.0 | Support evaluation of FHIRPath expressions and calculation within Questionnaire - **Is your feature request related to a problem? Please describe.**
There is a need to be able to evaluate expressions within a Questionnaire. This is currently not supported by the data-capture library
**Describe the solution you'd like**
Support for FHIRPath expressions
**Describe alternatives you've considered**
Other options include CQL.
**Additional context**
@maimoonak - can you provide some example use cases for the types of expressions you are trying to evaluate | priority | support evaluation of fhirpath expressions and calculation within questionnaire is your feature request related to a problem please describe there is a need to be able to evaluate expressions within a questionnaire this is currently not supported by the data capture library describe the solution you d like support for fhirpath expressions describe alternatives you ve considered other options include cql additional context maimoonak can you provide some example use cases for the types of expressions you are trying to evaluate | 1 |
684,158 | 23,409,453,335 | IssuesEvent | 2022-08-12 15:53:17 | Kong/kubernetes-ingress-controller | https://api.github.com/repos/Kong/kubernetes-ingress-controller | closed | Make Gateway API enabled by default | priority/medium area/gateway-api | ### Problem Statement
Now that [Gateway API](https://github.com/kubernetes-sigs/gateway-api) has APIs in `v1beta1` with the release of `v0.5.0` we are ready to call our Gateway API implementation beta as well. The purpose of this task is to mark the beta APIs as beta and enable them by default for future KIC releases. Alpha APIs should remain in their current alpha state behind a feature gate that is off.
### Proposed Solution
Create a new feature gate called `GatewayAlpha` which implicates support for the alpha apis. Put anything that's pre-beta under that gate, and turn it off by default. Mark `Gateway` feature gate as beta and turn it on by default. Update relevant documentation.
### Acceptance Criteria
- [x] beta gateway APIs are supported and on by default in KIC https://github.com/Kong/kubernetes-ingress-controller/pull/2781
- [x] alpha gateway APIs are off by default in KIC https://github.com/Kong/kubernetes-ingress-controller/pull/2781
- [x] documentation is updated to adjust for these changes | 1.0 | Make Gateway API enabled by default - ### Problem Statement
Now that [Gateway API](https://github.com/kubernetes-sigs/gateway-api) has APIs in `v1beta1` with the release of `v0.5.0` we are ready to call our Gateway API implementation beta as well. The purpose of this task is to mark the beta APIs as beta and enable them by default for future KIC releases. Alpha APIs should remain in their current alpha state behind a feature gate that is off.
### Proposed Solution
Create a new feature gate called `GatewayAlpha` which implicates support for the alpha apis. Put anything that's pre-beta under that gate, and turn it off by default. Mark `Gateway` feature gate as beta and turn it on by default. Update relevant documentation.
### Acceptance Criteria
- [x] beta gateway APIs are supported and on by default in KIC https://github.com/Kong/kubernetes-ingress-controller/pull/2781
- [x] alpha gateway APIs are off by default in KIC https://github.com/Kong/kubernetes-ingress-controller/pull/2781
- [x] documentation is updated to adjust for these changes | priority | make gateway api enabled by default problem statement now that has apis in with the release of we are ready to call our gateway api implementation beta as well the purpose of this task is to mark the beta apis as beta and enable them by default for future kic releases alpha apis should remain in their current alpha state behind a feature gate that is off proposed solution create a new feature gate called gatewayalpha which implicates support for the alpha apis put anything that s pre beta under that gate and turn it off by default mark gateway feature gate as beta and turn it on by default update relevant documentation acceptance criteria beta gateway apis are supported and on by default in kic alpha gateway apis are off by default in kic documentation is updated to adjust for these changes | 1 |
671,620 | 22,769,284,204 | IssuesEvent | 2022-07-08 08:27:39 | canonical-web-and-design/ubuntu.com | https://api.github.com/repos/canonical-web-and-design/ubuntu.com | closed | Blog cards are missing the coloured strip at the top | Priority: Medium | There is a fallback class for cards that don't have a specific colour assigned that is over riding everything https://github.com/canonical-web-and-design/ubuntu.com/blob/main/static/sass/_pattern_blog-card.scss#L119
 | 1.0 | Blog cards are missing the coloured strip at the top - There is a fallback class for cards that don't have a specific colour assigned that is over riding everything https://github.com/canonical-web-and-design/ubuntu.com/blob/main/static/sass/_pattern_blog-card.scss#L119
 | priority | blog cards are missing the coloured strip at the top there is a fallback class for cards that don t have a specific colour assigned that is over riding everything | 1 |
528,102 | 15,360,136,941 | IssuesEvent | 2021-03-01 16:36:12 | ChainSafe/gossamer | https://api.github.com/repos/ChainSafe/gossamer | closed | update runtime keystore to use GlobalKeystore, use key type IDs in ext_crypto funcs | Priority: 3 - Medium approved wasm | ## Task summary
<!-- A clear and concise description of what the task is. -->
- currently the runtime uses only the `Acco` keystore
- it should be using the `GlobalKeystore` instead (see dot/node.go `createRuntime`)
- the runtime should then use the correct keystore based on the `keyTypeID` that's passed into ext_crypto_... funcs
- the `keyTypeID` corresponds to the keystore's `Name` field
## Specification reference
<!-- Provide a reference to the specification as to what is being implemented. -->
- spec appendix E ext_crypto funcs
## Other information and links
<!-- Add any other context, existing implementation reference or screenshots about the task here. -->
-
<!-- Thank you 💪 --> | 1.0 | update runtime keystore to use GlobalKeystore, use key type IDs in ext_crypto funcs - ## Task summary
<!-- A clear and concise description of what the task is. -->
- currently the runtime uses only the `Acco` keystore
- it should be using the `GlobalKeystore` instead (see dot/node.go `createRuntime`)
- the runtime should then use the correct keystore based on the `keyTypeID` that's passed into ext_crypto_... funcs
- the `keyTypeID` corresponds to the keystore's `Name` field
## Specification reference
<!-- Provide a reference to the specification as to what is being implemented. -->
- spec appendix E ext_crypto funcs
## Other information and links
<!-- Add any other context, existing implementation reference or screenshots about the task here. -->
-
<!-- Thank you 💪 --> | priority | update runtime keystore to use globalkeystore use key type ids in ext crypto funcs task summary currently the runtime uses only the acco keystore it should be using the globalkeystore instead see dot node go createruntime the runtime should then use the correct keystore based on the keytypeid that s passed into ext crypto funcs the keytypeid corresponds to the keystore s name field specification reference spec appendix e ext crypto funcs other information and links | 1 |
707,528 | 24,309,123,912 | IssuesEvent | 2022-09-29 20:19:46 | georchestra/georchestra | https://api.github.com/repos/georchestra/georchestra | closed | mapfishapp: produce SLD v1.1.0 compliant documents | feature 0 - Backlog priority-medium | ... targeted to WMS >= 1.3.0 servers.
Keep current SLD 1.0 service for WMS version <= 1.1.1 servers.
| 1.0 | mapfishapp: produce SLD v1.1.0 compliant documents - ... targeted to WMS >= 1.3.0 servers.
Keep current SLD 1.0 service for WMS version <= 1.1.1 servers.
| priority | mapfishapp produce sld compliant documents targeted to wms servers keep current sld service for wms version servers | 1 |
509,570 | 14,739,817,832 | IssuesEvent | 2021-01-07 07:59:23 | konveyor/forklift-ui | https://api.github.com/repos/konveyor/forklift-ui | closed | Non-ready providers should be excluded from any provider selections in forms | medium-priority | If a provider does not have the Ready condition, inventory API requests for resources in that provider will fail. We should either filter out or disable those options in Select fields for providers (and if filtering out, show a message when no providers remain instead of showing an empty dropdown). This might also be a good opportunity to factor out a `ProviderSelect` common component, since we have some duplicated code for getting the available provider options and setting up that Select. | 1.0 | Non-ready providers should be excluded from any provider selections in forms - If a provider does not have the Ready condition, inventory API requests for resources in that provider will fail. We should either filter out or disable those options in Select fields for providers (and if filtering out, show a message when no providers remain instead of showing an empty dropdown). This might also be a good opportunity to factor out a `ProviderSelect` common component, since we have some duplicated code for getting the available provider options and setting up that Select. | priority | non ready providers should be excluded from any provider selections in forms if a provider does not have the ready condition inventory api requests for resources in that provider will fail we should either filter out or disable those options in select fields for providers and if filtering out show a message when no providers remain instead of showing an empty dropdown this might also be a good opportunity to factor out a providerselect common component since we have some duplicated code for getting the available provider options and setting up that select | 1 |
642,165 | 20,868,857,197 | IssuesEvent | 2022-03-22 10:01:23 | LiskHQ/lisk-desktop | https://api.github.com/repos/LiskHQ/lisk-desktop | closed | Distorted UI on wallet balance card when in discrete mode | type: bug unplanned priority: medium | ### Expected behavior
There should be no distorted view on the balance card when discrete mode is toggled on.
### Actual behavior
Switching to discrete mode cause a misplacement in the content of the balance card as seen in the screenshot.

### Steps to reproduce
- Navigate to the wallets page
- Toggle to discrete mode ( It would be observed that the balance card content UI is distorted)
### Which version(s) does this affect? (Environment, OS, etc...)
v2.3.0-rc-1
| 1.0 | Distorted UI on wallet balance card when in discrete mode - ### Expected behavior
There should be no distorted view on the balance card when discrete mode is toggled on.
### Actual behavior
Switching to discrete mode cause a misplacement in the content of the balance card as seen in the screenshot.

### Steps to reproduce
- Navigate to the wallets page
- Toggle to discrete mode ( It would be observed that the balance card content UI is distorted)
### Which version(s) does this affect? (Environment, OS, etc...)
v2.3.0-rc-1
| priority | distorted ui on wallet balance card when in discrete mode expected behavior there should be no distorted view on the balance card when discrete mode is toggled on actual behavior switching to discrete mode cause a misplacement in the content of the balance card as seen in the screenshot steps to reproduce navigate to the wallets page toggle to discrete mode it would be observed that the balance card content ui is distorted which version s does this affect environment os etc rc | 1 |
473,705 | 13,646,320,093 | IssuesEvent | 2020-09-25 22:50:21 | shahednasser/sbuttons | https://api.github.com/repos/shahednasser/sbuttons | opened | Change font for buttons in header | Hacktoberfest Priority: Medium good first issue help wanted up-for-grabs website | Change the font for the "Download" and "Documentation" links in the header to "Syne" | 1.0 | Change font for buttons in header - Change the font for the "Download" and "Documentation" links in the header to "Syne" | priority | change font for buttons in header change the font for the download and documentation links in the header to syne | 1 |
503,985 | 14,610,588,559 | IssuesEvent | 2020-12-22 00:49:16 | ableplayer/ableplayer | https://api.github.com/repos/ableplayer/ableplayer | closed | Eliminate options related to description | enhancement - medium priority | User testing with screen reader users has led to some insights related to description:
- If description is available as _both_ a separate described version of the video and text-based description (i.e., a WebVTT description track) nearly 100% of users would choose the separate described version.
- It is extremely rare for videos to be made available with description at all, and especially rare and unlikely that videos will be made available with description in both formats supported by Able Player.
- The Audio Description Preferences dialog is verbose and confusing, and arguably unnecessary given the preceding two items.
Therefore, the Preferred Format option should be eliminated, and should always default to "Alternative described version of the video" if both formats are available. With that option gone, the explanatory text in the dialog can be scaled back significantly. | 1.0 | Eliminate options related to description - User testing with screen reader users has led to some insights related to description:
- If description is available as _both_ a separate described version of the video and text-based description (i.e., a WebVTT description track) nearly 100% of users would choose the separate described version.
- It is extremely rare for videos to be made available with description at all, and especially rare and unlikely that videos will be made available with description in both formats supported by Able Player.
- The Audio Description Preferences dialog is verbose and confusing, and arguably unnecessary given the preceding two items.
Therefore, the Preferred Format option should be eliminated, and should always default to "Alternative described version of the video" if both formats are available. With that option gone, the explanatory text in the dialog can be scaled back significantly. | priority | eliminate options related to description user testing with screen reader users has led to some insights related to description if description is available as both a separate described version of the video and text based description i e a webvtt description track nearly of users would choose the separate described version it is extremely rare for videos to be made available with description at all and especially rare and unlikely that videos will be made available with description in both formats supported by able player the audio description preferences dialog is verbose and confusing and arguably unnecessary given the preceding two items therefore the preferred format option should be eliminated and should always default to alternative described version of the video if both formats are available with that option gone the explanatory text in the dialog can be scaled back significantly | 1 |
548,352 | 16,062,580,992 | IssuesEvent | 2021-04-23 14:27:49 | enso-org/ide | https://api.github.com/repos/enso-org/ide | opened | Strange things happens when putting colon as a separator in list. | Category: Controllers Priority: Medium Type: Bug | <!--
Please ensure that you are using the latest version of Enso IDE before reporting
the bug! It may have been fixed since.
-->
### What did you do?
I put `:` as a separator list by mistake, trying to make node `["x": [4.0]]`
### What did you expect to see?
The node with that exact expression, probably with syntax error message
### What did you see instead?
An expression without last ]
### Enso Version
alpha-3
### Additional notes
Trying to put such expressions in code editor also works strangely. | 1.0 | Strange things happens when putting colon as a separator in list. - <!--
Please ensure that you are using the latest version of Enso IDE before reporting
the bug! It may have been fixed since.
-->
### What did you do?
I put `:` as a separator list by mistake, trying to make node `["x": [4.0]]`
### What did you expect to see?
The node with that exact expression, probably with syntax error message
### What did you see instead?
An expression without last ]
### Enso Version
alpha-3
### Additional notes
Trying to put such expressions in code editor also works strangely. | priority | strange things happens when putting colon as a separator in list please ensure that you are using the latest version of enso ide before reporting the bug it may have been fixed since what did you do i put as a separator list by mistake trying to make node what did you expect to see the node with that exact expression probably with syntax error message what did you see instead an expression without last enso version alpha additional notes trying to put such expressions in code editor also works strangely | 1 |
738,649 | 25,570,643,406 | IssuesEvent | 2022-11-30 17:23:43 | uhh-cms/columnflow | https://api.github.com/repos/uhh-cms/columnflow | opened | Allow selection masks being passed to muon / electron weight producers | enhancement medium-priority | As it is right now, the muon and electron weight producers in https://github.com/uhh-cms/columnflow/blob/master/columnflow/production/muon.py and https://github.com/uhh-cms/columnflow/blob/master/columnflow/production/electron.py are very generic, but still they automatically read all leptons available.
In some use cases, where the unwanted leptons were already removed in an upstream reduction step, this might work, but in case some (e.g.) veto leptons are kept, we need to be able to select those we want to obtain the weights with.
A similar approach has been already followed in the [btag weight producer](https://github.com/uhh-cms/columnflow/blob/16c8a35f601de294a0ac3dd7501d4bf3f2e70080/columnflow/production/btag.py#L26), so adding this to the lepton weights as well should be straight forward. | 1.0 | Allow selection masks being passed to muon / electron weight producers - As it is right now, the muon and electron weight producers in https://github.com/uhh-cms/columnflow/blob/master/columnflow/production/muon.py and https://github.com/uhh-cms/columnflow/blob/master/columnflow/production/electron.py are very generic, but still they automatically read all leptons available.
In some use cases, where the unwanted leptons were already removed in an upstream reduction step, this might work, but in case some (e.g.) veto leptons are kept, we need to be able to select those we want to obtain the weights with.
A similar approach has been already followed in the [btag weight producer](https://github.com/uhh-cms/columnflow/blob/16c8a35f601de294a0ac3dd7501d4bf3f2e70080/columnflow/production/btag.py#L26), so adding this to the lepton weights as well should be straight forward. | priority | allow selection masks being passed to muon electron weight producers as it is right now the muon and electron weight producers in and are very generic but still they automatically read all leptons available in some use cases where the unwanted leptons were already removed in an upstream reduction step this might work but in case some e g veto leptons are kept we need to be able to select those we want to obtain the weights with a similar approach has been already followed in the so adding this to the lepton weights as well should be straight forward | 1 |
713,679 | 24,535,334,769 | IssuesEvent | 2022-10-11 20:11:32 | RobotLocomotion/drake | https://api.github.com/repos/RobotLocomotion/drake | closed | Install Python auto-complete helper files (`*.pyi`) | component: distribution priority: medium | See https://github.com/pybind/pybind11/issues/2350#issuecomment-668879301.
As I understand it, some IDEs will not help the user auto-complete their function-call arguments unless we also install `*.pyi` files for our native modules.
As part of our build process, we should import `mypy` (maybe `python3-mypy` if it's new enough, otherwise the wheel) and then invoke it to create pyi files, and then install them alongside our modules.
| 1.0 | Install Python auto-complete helper files (`*.pyi`) - See https://github.com/pybind/pybind11/issues/2350#issuecomment-668879301.
As I understand it, some IDEs will not help the user auto-complete their function-call arguments unless we also install `*.pyi` files for our native modules.
As part of our build process, we should import `mypy` (maybe `python3-mypy` if it's new enough, otherwise the wheel) and then invoke it to create pyi files, and then install them alongside our modules.
| priority | install python auto complete helper files pyi see as i understand it some ides will not help the user auto complete their function call arguments unless we also install pyi files for our native modules as part of our build process we should import mypy maybe mypy if it s new enough otherwise the wheel and then invoke it to create pyi files and then install them alongside our modules | 1 |
1,566 | 2,515,613,855 | IssuesEvent | 2015-01-15 19:50:18 | adobe/brackets | https://api.github.com/repos/adobe/brackets | reopened | LiveDevelopmentMultiBrowser unit test issues | F Live Preview MultiBrowser medium priority | **1)** <del>The Jasmine test-runner window always logs an exception when it is first loaded, before any tests are run:</del> [console messages hidden]
<!--
```
TypeError: undefined is not a function
at _showStatusChangeReason (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/main.js:145:21)
at statusChange (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/main.js:185:13)
at Object.trigger (file:///C:/code/Brackets/brackets-app/brackets/src/utils/EventDispatcher.js:222:40)
at _setStatus (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/LiveDevelopment.js:328:17)
at Object.init (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/LiveDevelopment.js:1497:9)
at file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/main.js:272:25
at _callHandler (file:///C:/code/Brackets/brackets-app/brackets/src/utils/AppInit.js:93:13)
at Object._dispatchReady (file:///C:/code/Brackets/brackets-app/brackets/src/utils/AppInit.js:113:13)
at _documentReadyHandler (file:///C:/code/Brackets/brackets-app/brackets/test/SpecRunner.js:166:17)
```
-->
<del>(because bootstrap-twipsy-mod is not loaded in the Jasmine window)</del> Fixed in #10285
**2)** Several LiveDevelopmentMultiBrowser tests fail for me with "timeout: timed out after 5000 msec waiting for livedevelopment.done.opened" (usually 2-3 tests fail per run).
**3)** The console is littered with error messages after running this test suite:
```
[testWindow] [node-error 10:14:11 AM] nodeSocketTransport: Socket closed, but couldn't locate client SpecRunnerUtils.js:543
runs.forEach._testWindow.console.(anonymous function) SpecRunnerUtils.js:543
handleLogEvent src/extensions/default/DebugCommands/NodeDebugUtils.js:114
trigger EventDispatcher.js:222
triggerWithArray EventDispatcher.js:260
NodeConnection._receive NodeConnection.js:479
```
(twice)
```
[testWindow] Exception in 'base:log' listener on NodeConnection {domains: Object, _registeredModules: Array[0], _pendingInterfaceRefreshDeferreds: Array[0], _pendingCommandDeferreds: Array[0], _autoReconnect: true…}
TypeError: Cannot read property 'console' of null
at Console.runs.forEach._testWindow.console.(anonymous function) [as log] (file:///C:/code/Brackets/brackets-app/brackets/test/spec/SpecRunnerUtils.js:544:53)
at handleLogEvent (file:///C:/code/Brackets/brackets-app/brackets/src/extensions/default/DebugCommands/NodeDebugUtils.js:117:21)
at NodeConnection.trigger (file:///C:/code/Brackets/brackets-app/brackets/src/utils/EventDispatcher.js:222:40)
at Object.triggerWithArray (file:///C:/code/Brackets/brackets-app/brackets/src/utils/EventDispatcher.js:260:28)
at NodeConnection._receive (file:///C:/code/Brackets/brackets-app/brackets/src/utils/NodeConnection.js:479:29)
```
(13 times - various listeners all with the same TypeError)
Even if these are caused by the timeouts above, it suggests the code does not fail gracefully in the event of a connection timeout.
**4)** One time this test suite seemed to get stuck with the Node process hanging onto file handles in the brackets/test/temp folder, which caused many other unit tests to fail since they were unable to clear that folder. Restating the Node process released the file handles. Depending on how easy it is to hit this failure case, we may want to do more to increase robustness... | 1.0 | LiveDevelopmentMultiBrowser unit test issues - **1)** <del>The Jasmine test-runner window always logs an exception when it is first loaded, before any tests are run:</del> [console messages hidden]
<!--
```
TypeError: undefined is not a function
at _showStatusChangeReason (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/main.js:145:21)
at statusChange (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/main.js:185:13)
at Object.trigger (file:///C:/code/Brackets/brackets-app/brackets/src/utils/EventDispatcher.js:222:40)
at _setStatus (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/LiveDevelopment.js:328:17)
at Object.init (file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/LiveDevelopment.js:1497:9)
at file:///C:/code/Brackets/brackets-app/brackets/src/LiveDevelopment/main.js:272:25
at _callHandler (file:///C:/code/Brackets/brackets-app/brackets/src/utils/AppInit.js:93:13)
at Object._dispatchReady (file:///C:/code/Brackets/brackets-app/brackets/src/utils/AppInit.js:113:13)
at _documentReadyHandler (file:///C:/code/Brackets/brackets-app/brackets/test/SpecRunner.js:166:17)
```
-->
<del>(because bootstrap-twipsy-mod is not loaded in the Jasmine window)</del> Fixed in #10285
**2)** Several LiveDevelopmentMultiBrowser tests fail for me with "timeout: timed out after 5000 msec waiting for livedevelopment.done.opened" (usually 2-3 tests fail per run).
**3)** The console is littered with error messages after running this test suite:
```
[testWindow] [node-error 10:14:11 AM] nodeSocketTransport: Socket closed, but couldn't locate client SpecRunnerUtils.js:543
runs.forEach._testWindow.console.(anonymous function) SpecRunnerUtils.js:543
handleLogEvent src/extensions/default/DebugCommands/NodeDebugUtils.js:114
trigger EventDispatcher.js:222
triggerWithArray EventDispatcher.js:260
NodeConnection._receive NodeConnection.js:479
```
(twice)
```
[testWindow] Exception in 'base:log' listener on NodeConnection {domains: Object, _registeredModules: Array[0], _pendingInterfaceRefreshDeferreds: Array[0], _pendingCommandDeferreds: Array[0], _autoReconnect: true…}
TypeError: Cannot read property 'console' of null
at Console.runs.forEach._testWindow.console.(anonymous function) [as log] (file:///C:/code/Brackets/brackets-app/brackets/test/spec/SpecRunnerUtils.js:544:53)
at handleLogEvent (file:///C:/code/Brackets/brackets-app/brackets/src/extensions/default/DebugCommands/NodeDebugUtils.js:117:21)
at NodeConnection.trigger (file:///C:/code/Brackets/brackets-app/brackets/src/utils/EventDispatcher.js:222:40)
at Object.triggerWithArray (file:///C:/code/Brackets/brackets-app/brackets/src/utils/EventDispatcher.js:260:28)
at NodeConnection._receive (file:///C:/code/Brackets/brackets-app/brackets/src/utils/NodeConnection.js:479:29)
```
(13 times - various listeners all with the same TypeError)
Even if these are caused by the timeouts above, it suggests the code does not fail gracefully in the event of a connection timeout.
**4)** One time this test suite seemed to get stuck with the Node process hanging onto file handles in the brackets/test/temp folder, which caused many other unit tests to fail since they were unable to clear that folder. Restating the Node process released the file handles. Depending on how easy it is to hit this failure case, we may want to do more to increase robustness... | priority | livedevelopmentmultibrowser unit test issues the jasmine test runner window always logs an exception when it is first loaded before any tests are run typeerror undefined is not a function at showstatuschangereason file c code brackets brackets app brackets src livedevelopment main js at statuschange file c code brackets brackets app brackets src livedevelopment main js at object trigger file c code brackets brackets app brackets src utils eventdispatcher js at setstatus file c code brackets brackets app brackets src livedevelopment livedevelopment js at object init file c code brackets brackets app brackets src livedevelopment livedevelopment js at file c code brackets brackets app brackets src livedevelopment main js at callhandler file c code brackets brackets app brackets src utils appinit js at object dispatchready file c code brackets brackets app brackets src utils appinit js at documentreadyhandler file c code brackets brackets app brackets test specrunner js because bootstrap twipsy mod is not loaded in the jasmine window fixed in several livedevelopmentmultibrowser tests fail for me with timeout timed out after msec waiting for livedevelopment done opened usually tests fail per run the console is littered with error messages after running this test suite nodesockettransport socket closed but couldn t locate client specrunnerutils js runs foreach testwindow console anonymous function specrunnerutils js handlelogevent src extensions default debugcommands nodedebugutils js trigger eventdispatcher js triggerwitharray eventdispatcher js nodeconnection receive nodeconnection js twice exception in base log listener on nodeconnection domains object registeredmodules array pendinginterfacerefreshdeferreds array pendingcommanddeferreds array autoreconnect true… typeerror cannot read property console of null at console runs foreach testwindow console anonymous function file c code brackets brackets app brackets test spec specrunnerutils js at handlelogevent file c code brackets brackets app brackets src extensions default debugcommands nodedebugutils js at nodeconnection trigger file c code brackets brackets app brackets src utils eventdispatcher js at object triggerwitharray file c code brackets brackets app brackets src utils eventdispatcher js at nodeconnection receive file c code brackets brackets app brackets src utils nodeconnection js times various listeners all with the same typeerror even if these are caused by the timeouts above it suggests the code does not fail gracefully in the event of a connection timeout one time this test suite seemed to get stuck with the node process hanging onto file handles in the brackets test temp folder which caused many other unit tests to fail since they were unable to clear that folder restating the node process released the file handles depending on how easy it is to hit this failure case we may want to do more to increase robustness | 1 |
321,053 | 9,792,850,663 | IssuesEvent | 2019-06-10 18:27:02 | dojot/dojot | https://api.github.com/repos/dojot/dojot | closed | [GUI] Template: Change metadata - does not validate fields populated with invalid values | Priority:Medium Status:Treated Team:Frontend Type:Bug |

**Note:** validation is done by adding the metadata
**Affected Version:** 0.3.0-nightly_20181227 | 1.0 | [GUI] Template: Change metadata - does not validate fields populated with invalid values -

**Note:** validation is done by adding the metadata
**Affected Version:** 0.3.0-nightly_20181227 | priority | template change metadata does not validate fields populated with invalid values note validation is done by adding the metadata affected version nightly | 1 |
75,196 | 3,456,583,548 | IssuesEvent | 2015-12-18 02:28:41 | dkpro/dkpro-core | https://api.github.com/repos/dkpro/dkpro-core | closed | Use current directory as default target location for writers | enhancement Module-api.io Priority-Medium wontfix | ```
For writers that write to directories, e.g. Conll2006Writer, we could maybe use the
current directory as default output location.
```
Original issue reported on code.google.com by `richard.eckart` on 2014-06-24 17:20:17 | 1.0 | Use current directory as default target location for writers - ```
For writers that write to directories, e.g. Conll2006Writer, we could maybe use the
current directory as default output location.
```
Original issue reported on code.google.com by `richard.eckart` on 2014-06-24 17:20:17 | priority | use current directory as default target location for writers for writers that write to directories e g we could maybe use the current directory as default output location original issue reported on code google com by richard eckart on | 1 |
1,024 | 2,506,912,178 | IssuesEvent | 2015-01-12 14:54:17 | ukwa/w3act | https://api.github.com/repos/ukwa/w3act | closed | CSV download from reports pages not working | bug Medium Priority | Hi @kinmanli : an attempt to download CSV from this just hangs up the browser:
http://www.webarchive.org.uk/actdev/reportscreation/targets/?s=createdAt&o=desc&crawlFrequency=DAILY&tld=either
| 1.0 | CSV download from reports pages not working - Hi @kinmanli : an attempt to download CSV from this just hangs up the browser:
http://www.webarchive.org.uk/actdev/reportscreation/targets/?s=createdAt&o=desc&crawlFrequency=DAILY&tld=either
| priority | csv download from reports pages not working hi kinmanli an attempt to download csv from this just hangs up the browser | 1 |
369,603 | 10,915,304,423 | IssuesEvent | 2019-11-21 10:53:28 | react-figma/react-figma | https://api.github.com/repos/react-figma/react-figma | closed | Different color formats support | complexity: medium priority: medium topic: components topic: primitives support type: feature or enhancement | ERROR: type should be string, got "https://facebook.github.io/react-native/docs/colors\r\n\r\n* 'rgb(255, 0, 255)'\r\n* 'rgba(255, 255, 255, 1.0)'\r\n* '#ff00ff00' \r\n* 'hsl(360, 100%, 100%)'\r\n* 'transparent'\r\n* Named colors: aliceblue, antiquewhite, etc." | 1.0 | Different color formats support - https://facebook.github.io/react-native/docs/colors
* 'rgb(255, 0, 255)'
* 'rgba(255, 255, 255, 1.0)'
* '#ff00ff00'
* 'hsl(360, 100%, 100%)'
* 'transparent'
* Named colors: aliceblue, antiquewhite, etc. | priority | different color formats support rgb rgba hsl transparent named colors aliceblue antiquewhite etc | 1 |
4,374 | 2,550,873,794 | IssuesEvent | 2015-02-02 00:37:23 | SO-Close-Vote-Reviewers/SOCVR-Chatbot | https://api.github.com/repos/SO-Close-Vote-Reviewers/SOCVR-Chatbot | closed | Add commands to get tags | enhancement help wanted medium-priority | The following commands need to get added:
- current tag
- next x tags
- start event
Details on those commands are int the setup.md file.
Most of the work is in the sede branch. | 1.0 | Add commands to get tags - The following commands need to get added:
- current tag
- next x tags
- start event
Details on those commands are int the setup.md file.
Most of the work is in the sede branch. | priority | add commands to get tags the following commands need to get added current tag next x tags start event details on those commands are int the setup md file most of the work is in the sede branch | 1 |
826,164 | 31,559,398,126 | IssuesEvent | 2023-09-03 03:53:48 | ubiquity/ubiquibot | https://api.github.com/repos/ubiquity/ubiquibot | opened | Linked Pull Request Not Closed On Expired Task | Priority: 2 (Medium) Time: <4 Hours | This should be diagnosed and fixed.
> @wannacfuture - Releasing the bounty back to dev pool because the allocated duration already ended!
Last activity time: Fri Aug 18 2023 23:07:15 GMT+0000 (Coordinated Universal Time)
_Originally posted by @ubiquibot in https://github.com/ubiquity/ubiquibot/issues/431#issuecomment-1694037026_ | 1.0 | Linked Pull Request Not Closed On Expired Task - This should be diagnosed and fixed.
> @wannacfuture - Releasing the bounty back to dev pool because the allocated duration already ended!
Last activity time: Fri Aug 18 2023 23:07:15 GMT+0000 (Coordinated Universal Time)
_Originally posted by @ubiquibot in https://github.com/ubiquity/ubiquibot/issues/431#issuecomment-1694037026_ | priority | linked pull request not closed on expired task this should be diagnosed and fixed wannacfuture releasing the bounty back to dev pool because the allocated duration already ended last activity time fri aug gmt coordinated universal time originally posted by ubiquibot in | 1 |
733,473 | 25,307,451,687 | IssuesEvent | 2022-11-17 15:07:10 | Fiserv/Support | https://api.github.com/repos/Fiserv/Support | closed | .docignore Articles Showing in Search | bug CommerceHub Priority - Medium Severity - Medium | # Reporting new issue for Commerce Hub
**Region** (if applicable)
Dev
**Page**
https://dev-developer.fiserv.com/search?q=test%20response&p=%5B%22CommerceHub%22%5D
**Describe the bug**
Search results contain articles that are in the .docignore
**Expected behavior**
.docignore articles should not show in search results they result in a 404
**Screenshots**


| 1.0 | .docignore Articles Showing in Search - # Reporting new issue for Commerce Hub
**Region** (if applicable)
Dev
**Page**
https://dev-developer.fiserv.com/search?q=test%20response&p=%5B%22CommerceHub%22%5D
**Describe the bug**
Search results contain articles that are in the .docignore
**Expected behavior**
.docignore articles should not show in search results they result in a 404
**Screenshots**


| priority | docignore articles showing in search reporting new issue for commerce hub region if applicable dev page describe the bug search results contain articles that are in the docignore expected behavior docignore articles should not show in search results they result in a screenshots | 1 |
507,013 | 14,678,259,527 | IssuesEvent | 2020-12-31 02:36:33 | rubyforgood/casa | https://api.github.com/repos/rubyforgood/casa | closed | Automate github issue weekly "is this still in progress?" checkins | :sparkles: :computer: Contributor Friendly / Devel Help Wanted Priority: Medium | Automate github issue weekly "is this still in progress?" checkins on "in progress" github issues without recent comments / commits. After two weeks (14 days) of checking-in and no comments or PRs, unassign the assignee and move the issue back to "to do" (automate as many parts of this as possible! Anything is better than nothing :) )
This is (hopefully) possible with github actions | 1.0 | Automate github issue weekly "is this still in progress?" checkins - Automate github issue weekly "is this still in progress?" checkins on "in progress" github issues without recent comments / commits. After two weeks (14 days) of checking-in and no comments or PRs, unassign the assignee and move the issue back to "to do" (automate as many parts of this as possible! Anything is better than nothing :) )
This is (hopefully) possible with github actions | priority | automate github issue weekly is this still in progress checkins automate github issue weekly is this still in progress checkins on in progress github issues without recent comments commits after two weeks days of checking in and no comments or prs unassign the assignee and move the issue back to to do automate as many parts of this as possible anything is better than nothing this is hopefully possible with github actions | 1 |
55,593 | 3,073,806,877 | IssuesEvent | 2015-08-20 00:42:14 | RobotiumTech/robotium | https://api.github.com/repos/RobotiumTech/robotium | closed | setDatePicker set wrong moth. | bug imported invalid Priority-Medium | _From [zoumy...@gmail.com](https://code.google.com/u/115552050981841755101/) on August 07, 2013 01:12:46_
What steps will reproduce the problem? 1.open DatePicker.
2.set Date 1986,7,21 but the date has been set Aug 21 1986 3. What is the expected output? What do you see instead? Jul 21 1986 What version of the product are you using? On what operating system? version :4.2.2 emulator Please provide any additional information below.
_Original issue: http://code.google.com/p/robotium/issues/detail?id=496_ | 1.0 | setDatePicker set wrong moth. - _From [zoumy...@gmail.com](https://code.google.com/u/115552050981841755101/) on August 07, 2013 01:12:46_
What steps will reproduce the problem? 1.open DatePicker.
2.set Date 1986,7,21 but the date has been set Aug 21 1986 3. What is the expected output? What do you see instead? Jul 21 1986 What version of the product are you using? On what operating system? version :4.2.2 emulator Please provide any additional information below.
_Original issue: http://code.google.com/p/robotium/issues/detail?id=496_ | priority | setdatepicker set wrong moth from on august what steps will reproduce the problem open datepicker set date but the date has been set aug what is the expected output what do you see instead jul what version of the product are you using on what operating system version emulator please provide any additional information below original issue | 1 |
349,853 | 10,474,561,773 | IssuesEvent | 2019-09-23 14:43:32 | LifeMC/LifeSkript | https://api.github.com/repos/LifeMC/LifeSkript | opened | Method 'loadScript' is too complex to analyze by data flow algorithm | priority: longtime goal priority: medium state: help wanted type: enhancement | **Describe the bug**
IntelliJ gives warning "Method 'loadScript' is too complex to analyze by data flow algorithm" in the ScriptLoader#loadScript method.
**To Reproduce**
Open ScriptLoader and navigate to loadScript method in IntelliJ (tested on IntelliJ IDEA Ultimate Edition 2019.2)
**Expected behavior**
Expected no warning. Method should be splitted to other methods to reduce its code complexity.
**Screenshots**
N/A
**Version information:**
- Minecraft Version: N/A
- Skript Version: 2.2.18 Development Build (master)
**Additional information**
N/A | 2.0 | Method 'loadScript' is too complex to analyze by data flow algorithm - **Describe the bug**
IntelliJ gives warning "Method 'loadScript' is too complex to analyze by data flow algorithm" in the ScriptLoader#loadScript method.
**To Reproduce**
Open ScriptLoader and navigate to loadScript method in IntelliJ (tested on IntelliJ IDEA Ultimate Edition 2019.2)
**Expected behavior**
Expected no warning. Method should be splitted to other methods to reduce its code complexity.
**Screenshots**
N/A
**Version information:**
- Minecraft Version: N/A
- Skript Version: 2.2.18 Development Build (master)
**Additional information**
N/A | priority | method loadscript is too complex to analyze by data flow algorithm describe the bug intellij gives warning method loadscript is too complex to analyze by data flow algorithm in the scriptloader loadscript method to reproduce open scriptloader and navigate to loadscript method in intellij tested on intellij idea ultimate edition expected behavior expected no warning method should be splitted to other methods to reduce its code complexity screenshots n a version information minecraft version n a skript version development build master additional information n a | 1 |
789,765 | 27,804,979,962 | IssuesEvent | 2023-03-17 18:59:40 | knative/docs | https://api.github.com/repos/knative/docs | closed | Fix tab formatting in code-samples folder | kind/bug kind/good-first-issue priority/medium kind/cleanup hacktoberfest good first issue help wanted | ## Background
The files in `code-samples` folder are no longer viewable on the website, only in GitHub. However, this has caused broken formatting because tabbed elements are not available in GitHub markdown.
## Task
Remove the tabbed formatting, for example `=== "yaml"`, in the files in the `code-samples` folder. Make sure that if the tabbed formatting is under a step, you indent your new formatting within each step.
Files that have this issue:
- [ ] [./code-samples/serving/cloudevents/cloudevents-nodejs/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-nodejs/README.md)
- [ ] [./code-samples/serving/cloudevents/cloudevents-rust/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-rust/README.md)
- [ ] [./code-samples/serving/cloudevents/cloudevents-go/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-go/README.md)
- [ ] [./code-samples/serving/cloudevents/cloudevents-vertx/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-vertx/README.md)
- [ ] [./code-samples/serving/cloudevents/cloudevents-spring/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-spring/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-ruby/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-ruby/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-shell/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-shell/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-java-spring/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-java-spring/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-python/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-python/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-java-spark/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-java-spark/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-go/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-go/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-php/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-php/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-scala/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-scala/README.md)
| 1.0 | Fix tab formatting in code-samples folder - ## Background
The files in `code-samples` folder are no longer viewable on the website, only in GitHub. However, this has caused broken formatting because tabbed elements are not available in GitHub markdown.
## Task
Remove the tabbed formatting, for example `=== "yaml"`, in the files in the `code-samples` folder. Make sure that if the tabbed formatting is under a step, you indent your new formatting within each step.
Files that have this issue:
- [ ] [./code-samples/serving/cloudevents/cloudevents-nodejs/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-nodejs/README.md)
- [ ] [./code-samples/serving/cloudevents/cloudevents-rust/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-rust/README.md)
- [ ] [./code-samples/serving/cloudevents/cloudevents-go/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-go/README.md)
- [ ] [./code-samples/serving/cloudevents/cloudevents-vertx/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-vertx/README.md)
- [ ] [./code-samples/serving/cloudevents/cloudevents-spring/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/cloudevents/cloudevents-spring/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-ruby/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-ruby/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-shell/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-shell/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-java-spring/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-java-spring/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-python/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-python/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-java-spark/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-java-spark/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-go/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-go/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-php/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-php/README.md)
- [ ] [./code-samples/serving/hello-world/helloworld-scala/README.md](https://github.com/knative/docs/tree/main/code-samples/serving/hello-world/helloworld-scala/README.md)
| priority | fix tab formatting in code samples folder background the files in code samples folder are no longer viewable on the website only in github however this has caused broken formatting because tabbed elements are not available in github markdown task remove the tabbed formatting for example yaml in the files in the code samples folder make sure that if the tabbed formatting is under a step you indent your new formatting within each step files that have this issue | 1 |
163,186 | 6,192,717,545 | IssuesEvent | 2017-07-05 03:28:44 | start-jsk/jsk_apc | https://api.github.com/repos/start-jsk/jsk_apc | opened | Make additional fingers, finger bases and gears | enhancement issue/priority/medium | - [ ] Print additional fingers, finger bases and gears (fix magnet fitting)
- [ ] Mold rubber to fingers
- [ ] Print additional other parts | 1.0 | Make additional fingers, finger bases and gears - - [ ] Print additional fingers, finger bases and gears (fix magnet fitting)
- [ ] Mold rubber to fingers
- [ ] Print additional other parts | priority | make additional fingers finger bases and gears print additional fingers finger bases and gears fix magnet fitting mold rubber to fingers print additional other parts | 1 |
350,723 | 10,500,868,163 | IssuesEvent | 2019-09-26 11:31:23 | code4romania/monitorizare-vot-android | https://api.github.com/repos/code4romania/monitorizare-vot-android | closed | [Research] Investigate replacing realm with room | android enhancement help wanted medium priority research | We are planning a complete redo of the app, using kotlin.
Please research the possibility of replacing Realm db with [Room](https://developer.android.com/topic/libraries/architecture/room). | 1.0 | [Research] Investigate replacing realm with room - We are planning a complete redo of the app, using kotlin.
Please research the possibility of replacing Realm db with [Room](https://developer.android.com/topic/libraries/architecture/room). | priority | investigate replacing realm with room we are planning a complete redo of the app using kotlin please research the possibility of replacing realm db with | 1 |
67,714 | 3,277,566,036 | IssuesEvent | 2015-10-27 01:40:51 | saxifrage/caac-map | https://api.github.com/repos/saxifrage/caac-map | closed | Hover-over highlighting does not work for pathway colors | Medium Priority | The darker colors of the pathway blocks don't allow hover-overs to do anything. Seems like if we shift the color scheme at large to echo Chelsea's design, then the really dark blue could be used for all hover-overs.
The ultra-light blue is the non-highlighted color, and the middle blue is the default color.

| 1.0 | Hover-over highlighting does not work for pathway colors - The darker colors of the pathway blocks don't allow hover-overs to do anything. Seems like if we shift the color scheme at large to echo Chelsea's design, then the really dark blue could be used for all hover-overs.
The ultra-light blue is the non-highlighted color, and the middle blue is the default color.

| priority | hover over highlighting does not work for pathway colors the darker colors of the pathway blocks don t allow hover overs to do anything seems like if we shift the color scheme at large to echo chelsea s design then the really dark blue could be used for all hover overs the ultra light blue is the non highlighted color and the middle blue is the default color | 1 |
30,550 | 2,724,021,608 | IssuesEvent | 2015-04-14 15:38:37 | CruxFramework/crux-widgets | https://api.github.com/repos/CruxFramework/crux-widgets | closed | Create a method on JsWindow class to get the parent window | Core enhancement imported Milestone-4.0.0 Priority-Medium | _From [ge...@cruxframework.org](https://code.google.com/u/108728025643241132101/) on May 20, 2011 18:51:43_
Create a method on JsWindow class to get the parent window
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=249_ | 1.0 | Create a method on JsWindow class to get the parent window - _From [ge...@cruxframework.org](https://code.google.com/u/108728025643241132101/) on May 20, 2011 18:51:43_
Create a method on JsWindow class to get the parent window
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=249_ | priority | create a method on jswindow class to get the parent window from on may create a method on jswindow class to get the parent window original issue | 1 |
160,956 | 6,106,076,560 | IssuesEvent | 2017-06-21 02:28:04 | minio/minio-go | https://api.github.com/repos/minio/minio-go | closed | TestPutObjectStream is not passing with latest and master minio | priority: medium | --- FAIL: TestPutObjectStreaming (0.00s)
api_functional_v4_test.go:406: Test 1 Error: The request signature we calculated does not match the signature you provided. Check your key and signing method. minio-go-
testh1owqaacnr26h95qc test-object
| 1.0 | TestPutObjectStream is not passing with latest and master minio - --- FAIL: TestPutObjectStreaming (0.00s)
api_functional_v4_test.go:406: Test 1 Error: The request signature we calculated does not match the signature you provided. Check your key and signing method. minio-go-
testh1owqaacnr26h95qc test-object
| priority | testputobjectstream is not passing with latest and master minio fail testputobjectstreaming api functional test go test error the request signature we calculated does not match the signature you provided check your key and signing method minio go test object | 1 |
522,088 | 15,148,714,826 | IssuesEvent | 2021-02-11 11:00:55 | FraunhoferISST/IDS-Connector-Framework | https://api.github.com/repos/FraunhoferISST/IDS-Connector-Framework | opened | Forward received RejectionMessage to Connector | Priority: Medium Type: Enhancement | Currently a received RejectionMessage is not forwarded to the connector-developer upon receiving it at the endpoints of the IDS-Framework due to failed DAT-validation of the received RejectionMessage within the IDS-Framework.
The received RejectionMessage should be forwarded to the connector developer so that the developer can evaluate it and the information contained like the correlationMessage, regardless of the status of the DAT-validation of the received message within the IDS-Framework. | 1.0 | Forward received RejectionMessage to Connector - Currently a received RejectionMessage is not forwarded to the connector-developer upon receiving it at the endpoints of the IDS-Framework due to failed DAT-validation of the received RejectionMessage within the IDS-Framework.
The received RejectionMessage should be forwarded to the connector developer so that the developer can evaluate it and the information contained like the correlationMessage, regardless of the status of the DAT-validation of the received message within the IDS-Framework. | priority | forward received rejectionmessage to connector currently a received rejectionmessage is not forwarded to the connector developer upon receiving it at the endpoints of the ids framework due to failed dat validation of the received rejectionmessage within the ids framework the received rejectionmessage should be forwarded to the connector developer so that the developer can evaluate it and the information contained like the correlationmessage regardless of the status of the dat validation of the received message within the ids framework | 1 |
710,668 | 24,427,238,851 | IssuesEvent | 2022-10-06 04:37:49 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | Renovate is stuck in an infinite loop overwriting a ci bot's commits | type:bug priority-3-medium status:in-progress regression | ### How are you running Renovate?
Mend Renovate hosted app on github.com
### If you're self-hosting Renovate, tell us what version of Renovate you run.
_No response_
### If you're self-hosting Renovate, select which platform you are using.
_No response_
### If you're self-hosting Renovate, tell us what version of the platform you run.
_No response_
### Was this something which used to work for you, and then stopped?
It used to work, and then stopped
### Describe the bug
Renovate is stuck in an infinite loop overwriting a ci bot's fixup commits. This started happening sometime in the past 2 months (the configuration has not been changed since). So far I have only observed this with python packages, but I do not have a similar procedure for other dependencies.
The process is something like this:
1. Renovate updates package `x` and commits to the repository, which triggers the first ci run.
2. The repository uses lock files with `pip-compile`, so before running builds and tests, the ci agent runs a script to update the lockfiles. This takes care of resolving dependency changes of `x`. If any changes were made, the ci bot commits them as a fixup commit (`git commit --fixup HEAD`) and pushes it to GitHub.
3. The new commit triggers a second ci run. This time, the check should pass and move on.
The ci bot is registered as a `gitIgnoredAuthor`, but in the past renovate would only overwrite the changes if the `main` branch was updated (I have `rebaseWhen` as `behind-base-branch`). Currently, every time the ci bot commits a change renovate seems to overwrite it the next time it runs. This means the repository update process is effectively stuck in an infinite loop. I'm not sure exactly what triggers this behavior and it's been a while since it worked, but it seems to have been around the same time renovate changed email names from `bot@renovateapp.com` to `29139614+renovate[bot]@users.noreply.github.com`. I recall having to manually rebase the existing branches as they were marked as modified.
I've provided debug logs from a minimal reproduction repository, which is publicly available on [GitHub](https://github.com/corypaik/renovate-loop). The repository contains one example package that causes this issue, all others have been removed or disabled. There is also a public [Buildkite](https://buildkite.com/corypaik/renovate-loop) available for reference.
### Relevant debug logs
<details><summary>Logs</summary>
```
DEBUG: No dangling containers to remove
INFO: Repository started
{
"renovateVersion": "32.184.2"
}
DEBUG: Using localDir: /mnt/renovate/gh/corypaik/renovate-loop
DEBUG: PackageFiles.clear() - Package files deleted
{
"baseBranches": []
}
DEBUG: resetMemCache()
DEBUG: initRepo("corypaik/renovate-loop")
DEBUG: corypaik/renovate-loop default branch = main
DEBUG: Using app token for git init
DEBUG: Repository cache is restored from revision 13
DEBUG: Resetting npmrc
DEBUG: detectSemanticCommits()
DEBUG: checkOnboarding()
DEBUG: isOnboarded()
DEBUG: Checking cached config file name
DEBUG: Existing config file confirmed
DEBUG: Repository config
{
"fileName": "renovate.json",
"config": {
"$schema": "https://docs.renovatebot.com/renovate-schema.json",
"extends": [
":ignoreModulesAndTests",
"group:monorepos",
"group:recommended",
"workarounds:all"
],
"semanticCommits": "enabled",
"semanticCommitType": "build",
"semanticCommitScope": "",
"prHourlyLimit": 2,
"prConcurrentLimit": 10,
"dependencyDashboard": true,
"labels": [
"target: minor",
"comp: build & ci",
"action: review"
],
"timezone": "America/Chihuahua",
"schedule": [
"after 10pm every weekday",
"before 4am every weekday"
],
"baseBranches": [
"main"
],
"rebaseWhen": "behind-base-branch",
"pip_requirements": {
"fileMatch": [
".*requirements.in$"
]
},
"prCreation": "not-pending",
"stabilityDays": 3,
"gitIgnoredAuthors": [
"buildkite-bot@buildkite.com"
],
"pre-commit": {
"enabled": true
},
"packageRules": [
{
"description": "Always pin direct python dependencies to ensure lockfile updates do not revert to a different version.",
"matchFiles": [
"third_party/requirements.in"
],
"rangeStrategy": "pin"
},
{
"description": "All python deps, except datasets (for minimal reproduction).",
"matchPackagePatterns": [
"*"
],
"matchUpdateTypes": [
"major",
"minor",
"patch"
],
"matchLanguages": [
"python"
],
"excludePackageNames": [
"datasets"
],
"enabled": false
}
]
}
}
DEBUG: Repo is onboarded
DEBUG: migrateAndValidate()
DEBUG: No config migration necessary
DEBUG: massaged config
{
"config": {
"$schema": "https://docs.renovatebot.com/renovate-schema.json",
"extends": [
"github>whitesource/merge-confidence:beta",
":ignoreModulesAndTests",
"group:monorepos",
"group:recommended",
"workarounds:all"
],
"semanticCommits": "enabled",
"semanticCommitType": "build",
"semanticCommitScope": "",
"prHourlyLimit": 2,
"prConcurrentLimit": 10,
"dependencyDashboard": true,
"labels": [
"target: minor",
"comp: build & ci",
"action: review"
],
"timezone": "America/Chihuahua",
"schedule": [
"after 10pm every weekday",
"before 4am every weekday"
],
"baseBranches": [
"main"
],
"rebaseWhen": "behind-base-branch",
"pip_requirements": {
"fileMatch": [
".*requirements.in$"
]
},
"prCreation": "not-pending",
"stabilityDays": 3,
"gitIgnoredAuthors": [
"buildkite-bot@buildkite.com"
],
"pre-commit": {
"enabled": true
},
"packageRules": [
{
"description": [
"Always pin direct python dependencies to ensure lockfile updates do not revert to a different version."
],
"matchFiles": [
"third_party/requirements.in"
],
"rangeStrategy": "pin"
},
{
"description": [
"All python deps, except datasets (for minimal reproduction)."
],
"matchPackagePatterns": [
"*"
],
"matchUpdateTypes": [
"major",
"minor",
"patch"
],
"matchLanguages": [
"python"
],
"excludePackageNames": [
"datasets"
],
"enabled": false
}
]
}
}
DEBUG: migrated config
{
"config": {
"$schema": "https://docs.renovatebot.com/renovate-schema.json",
"extends": [
"github>whitesource/merge-confidence:beta",
":ignoreModulesAndTests",
"group:monorepos",
"group:recommended",
"workarounds:all"
],
"semanticCommits": "enabled",
"semanticCommitType": "build",
"semanticCommitScope": "",
"prHourlyLimit": 2,
"prConcurrentLimit": 10,
"dependencyDashboard": true,
"labels": [
"target: minor",
"comp: build & ci",
"action: review"
],
"timezone": "America/Chihuahua",
"schedule": [
"after 10pm every weekday",
"before 4am every weekday"
],
"baseBranches": [
"main"
],
"rebaseWhen": "behind-base-branch",
"pip_requirements": {
"fileMatch": [
".*requirements.in$"
]
},
"prCreation": "not-pending",
"stabilityDays": 3,
"gitIgnoredAuthors": [
"buildkite-bot@buildkite.com"
],
"pre-commit": {
"enabled": true
},
"packageRules": [
{
"description": [
"Always pin direct python dependencies to ensure lockfile updates do not revert to a different version."
],
"matchFiles": [
"third_party/requirements.in"
],
"rangeStrategy": "pin"
},
{
"description": [
"All python deps, except datasets (for minimal reproduction)."
],
"matchPackagePatterns": [
"*"
],
"matchUpdateTypes": [
"major",
"minor",
"patch"
],
"matchLanguages": [
"python"
],
"excludePackageNames": [
"datasets"
],
"enabled": false
}
]
}
}
DEBUG: Setting hostRules from config
DEBUG: Found repo ignorePaths
{
"ignorePaths": [
"**/node_modules/**",
"**/bower_components/**",
"**/vendor/**",
"**/examples/**",
"**/__tests__/**",
"**/test/**",
"**/tests/**",
"**/__fixtures__/**"
]
}
DEBUG: No vulnerability alerts found
DEBUG: No vulnerability alerts found
DEBUG: findIssue(Dependency Dashboard)
DEBUG: Retrieving issueList
DEBUG: Retrieved 1 issues
DEBUG: Found issue 1
DEBUG: baseBranches
{
"baseBranches": [
"main"
]
}
DEBUG: baseBranch: main
{
"baseBranch": "main"
}
DEBUG: extract()
{
"baseBranch": "main"
}
DEBUG: Found cached extract
{
"baseBranch": "main",
"baseBranchSha": "5267848cc1c6b6dfb1d031f573b25d2ab36ebc82"
}
DEBUG: Deleted cached dep updates
{
"baseBranch": "main"
}
INFO: Dependency extraction complete
{
"baseBranch": "main",
"stats": {
"managers": {
"bazel": {
"fileCount": 1,
"depCount": 1
},
"pip_requirements": {
"fileCount": 2,
"depCount": 24
}
},
"total": {
"fileCount": 3,
"depCount": 25
}
}
}
DEBUG: baseBranch: main
{
"baseBranch": "main"
}
DEBUG: PackageFiles.add() - Package file saved for branch
{
"baseBranch": "main"
}
DEBUG: Package releases lookups complete
{
"baseBranch": "main"
}
DEBUG: branchifyUpgrades
{
"baseBranch": "main"
}
DEBUG: 4 flattened updates found: datasets, datasets, datasets, datasets
{
"baseBranch": "main"
}
DEBUG: Returning 2 branch(es)
{
"baseBranch": "main"
}
DEBUG: config.repoIsOnboarded=true
{
"baseBranch": "main"
}
DEBUG: packageFiles with updates
{
"baseBranch": "main",
"config": {
"bazel": [
{
"packageFile": "WORKSPACE",
"deps": [
{
"datasource": "github-tags",
"depType": "http_archive",
"depName": "rules_python",
"packageName": "bazelbuild/rules_python",
"currentValue": "refs",
"managerData": {
"def": "http_archive(\n name = \"rules_python\",\n sha256 = \"b593d13bb43c94ce94b483c2858e53a9b811f6f10e1e0eedc61073bd90e58d9c\",\n strip_prefix = \"rules_python-0.12.0\",\n url = \"https://github.com/bazelbuild/rules_python/archive/refs/tags/0.12.0.tar.gz\",\n)"
},
"depIndex": 0,
"warnings": [],
"versioning": "semver",
"skipReason": "invalid-value",
"updates": []
}
]
}
],
"pip_requirements": [
{
"packageFile": "third_party/requirements.in",
"deps": [
{
"depName": "datasets",
"currentValue": "==1.11.0",
"datasource": "pypi",
"currentVersion": "1.11.0",
"depIndex": 0,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/huggingface/datasets",
"isSingleVersion": true,
"fixedVersion": "1.11.0",
"updates": [
{
"bucket": "non-major",
"newVersion": "1.18.4",
"newValue": "==1.18.4",
"releaseTimestamp": "2022-03-07T10:50:07.000Z",
"newMajor": 1,
"newMinor": 18,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/datasets-1.x"
},
{
"bucket": "major",
"newVersion": "2.4.0",
"newValue": "==2.4.0",
"releaseTimestamp": "2022-07-25T13:21:43.000Z",
"newMajor": 2,
"newMinor": 4,
"updateType": "major",
"isRange": true,
"branchName": "renovate/datasets-2.x"
}
]
}
]
},
{
"packageFile": "third_party/requirements.txt",
"deps": [
{
"depName": "certifi",
"currentValue": "==2021.10.8",
"datasource": "pypi",
"currentVersion": "2021.10.8",
"depIndex": 0,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/certifi/python-certifi",
"isSingleVersion": true,
"fixedVersion": "2021.10.8",
"updates": [
{
"bucket": "major",
"newVersion": "2022.6.15",
"newValue": "==2022.6.15",
"releaseTimestamp": "2022-06-15T18:09:18.000Z",
"newMajor": 2022,
"newMinor": 6,
"updateType": "major",
"isRange": true,
"branchName": "renovate/certifi-2022.x"
}
]
},
{
"depName": "charset-normalizer",
"currentValue": "==2.0.12",
"datasource": "pypi",
"currentVersion": "2.0.12",
"depIndex": 1,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/ousret/charset_normalizer",
"isSingleVersion": true,
"fixedVersion": "2.0.12",
"updates": [
{
"bucket": "non-major",
"newVersion": "2.1.1",
"newValue": "==2.1.1",
"releaseTimestamp": "2022-08-19T22:13:46.000Z",
"newMajor": 2,
"newMinor": 1,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/charset-normalizer-2.x"
}
]
},
{
"depName": "datasets",
"currentValue": "==1.11.0",
"datasource": "pypi",
"currentVersion": "1.11.0",
"depIndex": 2,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/huggingface/datasets",
"isSingleVersion": true,
"fixedVersion": "1.11.0",
"updates": [
{
"bucket": "non-major",
"newVersion": "1.18.4",
"newValue": "==1.18.4",
"releaseTimestamp": "2022-03-07T10:50:07.000Z",
"newMajor": 1,
"newMinor": 18,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/datasets-1.x"
},
{
"bucket": "major",
"newVersion": "2.4.0",
"newValue": "==2.4.0",
"releaseTimestamp": "2022-07-25T13:21:43.000Z",
"newMajor": 2,
"newMinor": 4,
"updateType": "major",
"isRange": true,
"branchName": "renovate/datasets-2.x"
}
]
},
{
"depName": "dill",
"currentValue": "==0.3.1.1",
"datasource": "pypi",
"currentVersion": "0.3.1.1",
"depIndex": 3,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/uqfoundation/dill",
"isSingleVersion": true,
"fixedVersion": "0.3.1.1",
"updates": [
{
"bucket": "non-major",
"newVersion": "0.3.5.1",
"newValue": "==0.3.5.1",
"releaseTimestamp": "2022-05-20T11:21:11.000Z",
"newMajor": 0,
"newMinor": 3,
"updateType": "patch",
"isRange": true,
"branchName": "renovate/dill-0.x"
}
]
},
{
"depName": "filelock",
"currentValue": "==3.6.0",
"datasource": "pypi",
"currentVersion": "3.6.0",
"depIndex": 4,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/tox-dev/py-filelock",
"isSingleVersion": true,
"fixedVersion": "3.6.0",
"updates": [
{
"bucket": "non-major",
"newVersion": "3.8.0",
"newValue": "==3.8.0",
"releaseTimestamp": "2022-08-10T06:44:31.000Z",
"newMajor": 3,
"newMinor": 8,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/filelock-3.x"
}
]
},
{
"depName": "fsspec",
"currentValue": "==2021.11.1",
"datasource": "pypi",
"currentVersion": "2021.11.1",
"depIndex": 5,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/fsspec/filesystem_spec",
"changelogUrl": "https://filesystem-spec.readthedocs.io/en/latest/changelog.html",
"isSingleVersion": true,
"fixedVersion": "2021.11.1",
"updates": [
{
"bucket": "major",
"newVersion": "2022.8.0",
"newValue": "==2022.8.0",
"releaseTimestamp": "2022-08-30T18:08:19.000Z",
"newMajor": 2022,
"newMinor": 8,
"updateType": "major",
"isRange": true,
"branchName": "renovate/fsspec-2022.x"
}
]
},
{
"depName": "huggingface-hub",
"currentValue": "==0.0.19",
"datasource": "pypi",
"currentVersion": "0.0.19",
"depIndex": 6,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/huggingface/huggingface_hub",
"isSingleVersion": true,
"fixedVersion": "0.0.19",
"updates": [
{
"bucket": "non-major",
"newVersion": "0.9.1",
"newValue": "==0.9.1",
"releaseTimestamp": "2022-08-25T15:37:11.000Z",
"newMajor": 0,
"newMinor": 9,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/huggingface-hub-0.x"
}
]
},
{
"depName": "idna",
"currentValue": "==3.3",
"datasource": "pypi",
"currentVersion": "3.3",
"depIndex": 7,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/kjd/idna",
"fixedVersion": "3.3",
"updates": []
},
{
"depName": "multiprocess",
"currentValue": "==0.70.9",
"datasource": "pypi",
"currentVersion": "0.70.9",
"depIndex": 8,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/uqfoundation/multiprocess",
"isSingleVersion": true,
"fixedVersion": "0.70.9",
"updates": [
{
"bucket": "non-major",
"newVersion": "0.70.13",
"newValue": "==0.70.13",
"releaseTimestamp": "2022-05-23T22:52:42.000Z",
"newMajor": 0,
"newMinor": 70,
"updateType": "patch",
"isRange": true,
"branchName": "renovate/multiprocess-0.x"
}
]
},
{
"depName": "numpy",
"currentValue": "==1.21.3",
"datasource": "pypi",
"currentVersion": "1.21.3",
"depIndex": 9,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/numpy/numpy",
"homepage": "https://www.numpy.org",
"isSingleVersion": true,
"fixedVersion": "1.21.3",
"updates": [
{
"bucket": "non-major",
"newVersion": "1.23.2",
"newValue": "==1.23.2",
"releaseTimestamp": "2022-08-14T00:14:09.000Z",
"newMajor": 1,
"newMinor": 23,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/numpy-1.x"
}
]
},
{
"depName": "packaging",
"currentValue": "==20.9",
"datasource": "pypi",
"currentVersion": "20.9",
"depIndex": 10,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/pypa/packaging",
"isSingleVersion": true,
"fixedVersion": "20.9",
"updates": [
{
"bucket": "major",
"newVersion": "21.3",
"newValue": "==21.3",
"releaseTimestamp": "2021-11-18T00:39:10.000Z",
"newMajor": 21,
"newMinor": 3,
"updateType": "major",
"isRange": true,
"branchName": "renovate/packaging-21.x"
}
]
},
{
"depName": "pandas",
"currentValue": "==1.4.2",
"datasource": "pypi",
"currentVersion": "1.4.2",
"depIndex": 11,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/pandas-dev/pandas",
"homepage": "https://pandas.pydata.org",
"isSingleVersion": true,
"fixedVersion": "1.4.2",
"updates": [
{
"bucket": "non-major",
"newVersion": "1.4.3",
"newValue": "==1.4.3",
"releaseTimestamp": "2022-06-23T13:30:09.000Z",
"newMajor": 1,
"newMinor": 4,
"updateType": "patch",
"isRange": true,
"branchName": "renovate/pandas-1.x"
}
]
},
{
"depName": "pyarrow",
"currentValue": "==5.0.0",
"datasource": "pypi",
"currentVersion": "5.0.0",
"depIndex": 12,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/apache/arrow",
"homepage": "https://arrow.apache.org/",
"isSingleVersion": true,
"fixedVersion": "5.0.0",
"updates": [
{
"bucket": "major",
"newVersion": "9.0.0",
"newValue": "==9.0.0",
"releaseTimestamp": "2022-08-03T10:45:55.000Z",
"newMajor": 9,
"newMinor": 0,
"updateType": "major",
"isRange": true,
"branchName": "renovate/pyarrow-9.x"
}
]
},
{
"depName": "pyparsing",
"currentValue": "==2.4.7",
"datasource": "pypi",
"currentVersion": "2.4.7",
"depIndex": 13,
"warnings": [],
"versioning": "pep440",
"isSingleVersion": true,
"fixedVersion": "2.4.7",
"updates": [
{
"bucket": "major",
"newVersion": "3.0.9",
"newValue": "==3.0.9",
"releaseTimestamp": "2022-05-10T23:26:03.000Z",
"newMajor": 3,
"newMinor": 0,
"updateType": "major",
"isRange": true,
"branchName": "renovate/pyparsing-3.x"
}
]
},
{
"depName": "python-dateutil",
"currentValue": "==2.8.2",
"datasource": "pypi",
"currentVersion": "2.8.2",
"depIndex": 14,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/dateutil/dateutil",
"fixedVersion": "2.8.2",
"updates": []
},
{
"depName": "pytz",
"currentValue": "==2022.1",
"datasource": "pypi",
"currentVersion": "2022.1",
"depIndex": 15,
"warnings": [],
"versioning": "pep440",
"homepage": "http://pythonhosted.org/pytz",
"isSingleVersion": true,
"fixedVersion": "2022.1",
"updates": [
{
"bucket": "non-major",
"newVersion": "2022.2.1",
"newValue": "==2022.2.1",
"releaseTimestamp": "2022-08-13T02:07:57.000Z",
"newMajor": 2022,
"newMinor": 2,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/pytz-2022.x"
}
]
},
{
"depName": "pyyaml",
"currentValue": "==5.4.1",
"datasource": "pypi",
"currentVersion": "5.4.1",
"depIndex": 16,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/yaml/pyyaml",
"homepage": "https://pyyaml.org/",
"isSingleVersion": true,
"fixedVersion": "5.4.1",
"updates": [
{
"bucket": "major",
"newVersion": "6.0",
"newValue": "==6.0",
"releaseTimestamp": "2021-10-13T19:39:42.000Z",
"newMajor": 6,
"newMinor": 0,
"updateType": "major",
"isRange": true,
"branchName": "renovate/pyyaml-6.x"
}
]
},
{
"depName": "requests",
"currentValue": "==2.27.1",
"datasource": "pypi",
"currentVersion": "2.27.1",
"depIndex": 17,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/psf/requests",
"homepage": "https://requests.readthedocs.io",
"changelogUrl": "https://github.com/psf/requests/blob/master/HISTORY.md",
"isSingleVersion": true,
"fixedVersion": "2.27.1",
"updates": [
{
"bucket": "non-major",
"newVersion": "2.28.1",
"newValue": "==2.28.1",
"releaseTimestamp": "2022-06-29T15:13:40.000Z",
"newMajor": 2,
"newMinor": 28,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/requests-2.x"
}
]
},
{
"depName": "six",
"currentValue": "==1.16.0",
"datasource": "pypi",
"currentVersion": "1.16.0",
"depIndex": 18,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/benjaminp/six",
"fixedVersion": "1.16.0",
"updates": []
},
{
"depName": "tqdm",
"currentValue": "==4.63.0",
"datasource": "pypi",
"currentVersion": "4.63.0",
"depIndex": 19,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/tqdm/tqdm",
"homepage": "https://tqdm.github.io",
"changelogUrl": "https://tqdm.github.io/releases",
"isSingleVersion": true,
"fixedVersion": "4.63.0",
"updates": [
{
"bucket": "non-major",
"newVersion": "4.64.0",
"newValue": "==4.64.0",
"releaseTimestamp": "2022-04-04T01:48:46.000Z",
"newMajor": 4,
"newMinor": 64,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/tqdm-4.x"
}
]
},
{
"depName": "typing-extensions",
"currentValue": "==4.1.1",
"datasource": "pypi",
"currentVersion": "4.1.1",
"depIndex": 20,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/python/typing_extensions",
"changelogUrl": "https://github.com/python/typing_extensions/blob/main/CHANGELOG.md",
"isSingleVersion": true,
"fixedVersion": "4.1.1",
"updates": [
{
"bucket": "non-major",
"newVersion": "4.3.0",
"newValue": "==4.3.0",
"releaseTimestamp": "2022-07-01T14:56:11.000Z",
"newMajor": 4,
"newMinor": 3,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/typing-extensions-4.x"
}
]
},
{
"depName": "urllib3",
"currentValue": "==1.26.8",
"datasource": "pypi",
"currentVersion": "1.26.8",
"depIndex": 21,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/urllib3/urllib3",
"homepage": "https://urllib3.readthedocs.io/",
"isSingleVersion": true,
"fixedVersion": "1.26.8",
"updates": [
{
"bucket": "non-major",
"newVersion": "1.26.12",
"newValue": "==1.26.12",
"releaseTimestamp": "2022-08-22T13:27:19.000Z",
"newMajor": 1,
"newMinor": 26,
"updateType": "patch",
"isRange": true,
"branchName": "renovate/urllib3-1.x"
}
]
},
{
"depName": "xxhash",
"currentValue": "==3.0.0",
"datasource": "pypi",
"currentVersion": "3.0.0",
"depIndex": 22,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/ifduyue/python-xxhash",
"fixedVersion": "3.0.0",
"updates": []
}
]
}
]
}
}
DEBUG: processRepo()
DEBUG: Processing 2 branches: renovate/datasets-1.x, renovate/datasets-2.x
DEBUG: Calculating hourly PRs remaining
DEBUG: getPrList success
{
"pullsTotal": 1,
"requestsTotal": 1,
"apiQuotaAffected": true
}
DEBUG: currentHourStart=2022-08-31T02:00:00.000+00:00
DEBUG: PR hourly limit remaining: 2
DEBUG: Calculating prConcurrentLimit (10)
DEBUG: getBranchPr(renovate/datasets-1.x)
DEBUG: findPr(renovate/datasets-1.x, undefined, open)
DEBUG: Found PR #2
DEBUG: getBranchPr(renovate/datasets-2.x)
DEBUG: findPr(renovate/datasets-2.x, undefined, open)
DEBUG: findPr(renovate/datasets-2.x, undefined, closed)
DEBUG: 1 PRs are currently open
DEBUG: PR concurrent limit remaining: 9
DEBUG: Calculated maximum PRs remaining this run
{
"prsRemaining": 2
}
DEBUG: PullRequests limit = 2
DEBUG: Calculating hourly PRs remaining
DEBUG: currentHourStart=2022-08-31T02:00:00.000+00:00
DEBUG: PR hourly limit remaining: 2
DEBUG: Calculating branchConcurrentLimit (10)
DEBUG: 1 already existing branches found: renovate/datasets-1.x
DEBUG: Branch concurrent limit remaining: 9
DEBUG: Calculated maximum branches remaining this run
{
"branchesRemaining": 2
}
DEBUG: Branches limit = 2
DEBUG: Branch fingerprint is unchanged, updates check can be skipped(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: getBranchPr(renovate/datasets-1.x)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: findPr(renovate/datasets-1.x, undefined, open)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Found PR #2(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: branchExists=true(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: dependencyDashboardCheck=undefined(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: PR rebase requested=false(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Checking if PR has been edited(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Found existing branch PR(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Checking schedule(after 10pm every weekday,before 4am every weekday, America/Chihuahua)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Found timezone(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"timezone": "America/Chihuahua"
}
DEBUG: Adjusting now for timezone(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Checking 2 schedule(s)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Checking schedule "after 10pm every weekday"(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"parsedSchedule": {
"schedules": [
{
"t_a": [
79200
],
"d": [
2,
3,
4,
5,
6
]
}
],
"exceptions": [],
"error": -1
}
}
DEBUG: Checking schedule "before 4am every weekday"(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"parsedSchedule": {
"schedules": [
{
"t_b": [
14400
],
"d": [
2,
3,
4,
5,
6
]
}
],
"exceptions": [],
"error": -1
}
}
DEBUG: Package not scheduled(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Branch + PR exists but is not scheduled -- will update if necessary(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Branch already exists(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: getBranchPr(renovate/datasets-1.x)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: findPr(renovate/datasets-1.x, undefined, open)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Found PR #2(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Branch is behind base branch and needs rebasing(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Branch is unmodified, so can be rebased(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Using reuseExistingBranch: false(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Setting current branch to main(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Initializing git repository into /mnt/renovate/gh/corypaik/renovate-loop(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Performing blobless clone(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: git clone completed(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"durationMs": 1449
}
DEBUG: latest repository commit(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"latestCommit": {
"hash": "5267848cc1c6b6dfb1d031f573b25d2ab36ebc82",
"date": "2022-08-30T20:06:50-06:00",
"message": "ci: always run pip lockfile step",
"refs": "HEAD -> main, origin/main, origin/HEAD",
"body": "",
"author_name": "Cory Paik",
"author_email": "corypaik@gmail.com"
}
}
DEBUG: latest commit(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"branchName": "main",
"latestCommitDate": "2022-08-30T20:06:50-06:00"
}
DEBUG: manager.getUpdatedPackageFiles() reuseExistingBranch=false(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Starting search at index 527(packageFile="third_party/requirements.txt", branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"depName": "datasets"
}
DEBUG: Found match at index 527(packageFile="third_party/requirements.txt", branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"depName": "datasets"
}
DEBUG: Contents updated(packageFile="third_party/requirements.txt", branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"depName": "datasets"
}
DEBUG: Starting search at index 8(packageFile="third_party/requirements.in", branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"depName": "datasets"
}
DEBUG: Found match at index 8(packageFile="third_party/requirements.in", branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"depName": "datasets"
}
DEBUG: Contents updated(packageFile="third_party/requirements.in", branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"depName": "datasets"
}
DEBUG: pip_requirements.updateArtifacts(third_party/requirements.txt)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Setting BUILDPACK_CACHE_DIR to /tmp/containerbase(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Using docker to execute(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"image": "sidecar"
}
DEBUG: Resolved stable matching version(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"toolName": "python",
"resolvedVersion": "3.10.6"
}
DEBUG: containerbaseDir is separate from cacheDir(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: No tag or tagConstraint specified(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"image": "docker.io/renovate/sidecar"
}
DEBUG: Fetching Docker image: docker.io/renovate/sidecar(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Finished fetching Docker image docker.io/renovate/sidecar@sha256:ee8108a5a17ef370f15da2c11bf7b2da9035153fde1587118ac9cc5c6c56cf0a(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Executing command(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"command": "docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/corypaik/renovate-loop\":\"/mnt/renovate/gh/corypaik/renovate-loop\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -w \"/mnt/renovate/gh/corypaik/renovate-loop\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.10.6 && pip install --user hashin && hashin datasets==1.18.4 -r third_party/requirements.txt\""
}
DEBUG: exec completed(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"cmd": "docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/corypaik/renovate-loop\":\"/mnt/renovate/gh/corypaik/renovate-loop\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -w \"/mnt/renovate/gh/corypaik/renovate-loop\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.10.6 && pip install --user hashin && hashin datasets==1.18.4 -r third_party/requirements.txt\"",
"durationMs": 27296,
"stdout": "installing v2 tool python v3.10.6\nUsing prebuild python\nRequirement already satisfied: pip in /opt/buildpack/tools/python/3.10.6/lib/python3.10/site-packages (22.2.1)\nCollecting pip\n Downloading pip-22.2.2-py3-none-any.whl (2.0 MB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.0/2.0 MB 21.0 MB/s eta 0:00:00\nInstalling collected packages: pip\n Attempting uninstall: pip\n Found existing installation: pip 22.2.1\n Uninstalling pip-22.2.1:\n Successfully uninstalled pip-22.2.1\nSuccessfully installed pip-22.2.2\nFiles removed: 2\nlinking tool python v3.10.6\nPython 3.10.6\npip 22.2.2 from /opt/buildpack/tools/python/3.10.6/lib/python3.10/site-packages/pip (python 3.10)\nInstalled v2 /usr/local/buildpack/tools/v2/python.sh in 20 seconds\nCollecting hashin\n Downloading hashin-0.17.0-py2.py3-none-any.whl (16 kB)\nCollecting pip-api\n Downloading pip_api-0.0.30-py3-none-any.whl (111 kB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 111.6/111.6 kB 5.9 MB/s eta 0:00:00\nCollecting packaging\n Downloading packaging-21.3-py3-none-any.whl (40 kB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 40.8/40.8 kB 4.7 MB/s eta 0:00:00\nCollecting pyparsing!=3.0.5,>=2.0.2\n Downloading pyparsing-3.0.9-py3-none-any.whl (98 kB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 98.3/98.3 kB 11.1 MB/s eta 0:00:00\nRequirement already satisfied: pip in /opt/buildpack/tools/python/3.10.6/lib/python3.10/site-packages (from pip-api->hashin) (22.2.2)\nInstalling collected packages: pyparsing, pip-api, packaging, hashin\nSuccessfully installed hashin-0.17.0 packaging-21.3 pip-api-0.0.30 pyparsing-3.0.9\n",
"stderr": "Found file in cache: /tmp/containerbase/f79e18f004c2278a4c76a549ca9e8beed6ee66d17e187e79a230ba06836df191/python-3.10.6-focal-x86_64.tar.xz\n"
}
DEBUG: Returning updated third_party/requirements.txt(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: pip_requirements.updateArtifacts(third_party/requirements.in)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: No hashin commands to run - returning(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Updated 2 package files(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Updated 1 lock files(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"updatedArtifacts": [
"third_party/requirements.txt"
]
}
DEBUG: Getting comments for #2(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Found 0 comments(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: isBranchConflicted(main, renovate/datasets-1.x)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Using cached result false for isBranchConflicted(main, renovate/datasets-1.x)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: 2 file(s) to commit(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Preparing files for committing to branch renovate/datasets-1.x(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Setting git author name(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"gitAuthorName": "Renovate Bot"
}
DEBUG: Setting git author email(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"gitAuthorEmail": "bot@renovateapp.com"
}
DEBUG: git commit(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"deletedFiles": [],
"ignoredFiles": [],
"result": {
"author": null,
"branch": "renovate/datasets-1.x",
"commit": "210146775398f010baed48f2e6891d780406efa2",
"root": false,
"summary": {
"changes": 2,
"insertions": 4,
"deletions": 4
}
}
}
DEBUG: resetToCommit(5267848cc1c6b6dfb1d031f573b25d2ab36ebc82)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Fetching branch renovate/datasets-1.x(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
INFO: Branch updated(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"commitSha": "78dae2cf45f995142afb48597777d62fa9677a1f"
}
DEBUG: Updating renovate/stability-days status check state to green(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Setting branch status(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"context": "renovate/stability-days",
"state": "green"
}
DEBUG: Branch status pending(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"commitSha": "78dae2cf45f995142afb48597777d62fa9677a1f"
}
DEBUG: getBranchPr(renovate/datasets-2.x)(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: findPr(renovate/datasets-2.x, undefined, open)(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: findPr(renovate/datasets-2.x, undefined, closed)(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: branchExists=false(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: dependencyDashboardCheck=undefined(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: recreateClosed is false(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: findPr(renovate/datasets-2.x, build: update dependency datasets to v2, !open)(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: prAlreadyExisted=false(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: Checking schedule(after 10pm every weekday,before 4am every weekday, America/Chihuahua)(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: Found timezone(branch="renovate/datasets-2.x")
{
"baseBranch": "main",
"timezone": "America/Chihuahua"
}
DEBUG: Adjusting now for timezone(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: Checking 2 schedule(s)(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: Checking schedule "after 10pm every weekday"(branch="renovate/datasets-2.x")
{
"baseBranch": "main",
"parsedSchedule": {
"schedules": [
{
"t_a": [
79200
],
"d": [
2,
3,
4,
5,
6
]
}
],
"exceptions": [],
"error": -1
}
}
DEBUG: Checking schedule "before 4am every weekday"(branch="renovate/datasets-2.x")
{
"baseBranch": "main",
"parsedSchedule": {
"schedules": [
{
"t_b": [
14400
],
"d": [
2,
3,
4,
5,
6
]
}
],
"exceptions": [],
"error": -1
}
}
DEBUG: Package not scheduled(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: Skipping branch creation as not within schedule(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: getBranchPr(renovate/datasets-1.x)
DEBUG: findPr(renovate/datasets-1.x, undefined, open)
DEBUG: Found PR #2
DEBUG: Ensuring Dependency Dashboard
DEBUG: ensureIssue(Dependency Dashboard)
DEBUG: Patching issue
DEBUG: Issue updated
DEBUG: Removing any stale branches
DEBUG: config.repoIsOnboarded=true
DEBUG: Branch lists
{
"branchList": [
"renovate/datasets-1.x",
"renovate/datasets-2.x"
],
"renovateBranches": [
"renovate/datasets-1.x"
]
}
DEBUG: remainingBranches=
DEBUG: No branches to clean up
DEBUG: Cleaning up Renovate refs: refs/renovate/*
DEBUG: PackageFiles.clear() - Package files deleted
{
"baseBranches": [
"main"
]
}
DEBUG: Renovate repository PR statistics
{
"stats": {
"total": 1,
"open": 1,
"closed": 0,
"merged": 0
}
}
DEBUG: Repository result: done, status: onboarded, enabled: true, onboarded: true
DEBUG: Repository timing splits (milliseconds)
{
"splits": {
"init": 3599,
"extract": 494,
"lookup": 2772,
"onboarding": 0,
"update": 35845
},
"total": 45597
}
DEBUG: http statistics
{
"urls": {
"https://api.github.com/graphql (POST,200)": 2,
"https://api.github.com/repos/corypaik/renovate-loop/commits/78dae2cf45f995142afb48597777d62fa9677a1f/statuses (GET,200)": 2,
"https://api.github.com/repos/corypaik/renovate-loop/commits/renovate/datasets-1.x/status (GET,200)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/contents/renovate.json (GET,200)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/git/commits (POST,201)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/git/refs/heads/renovate/datasets-1.x (PATCH,200)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/git/trees (POST,201)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/issues/1 (GET,200)": 2,
"https://api.github.com/repos/corypaik/renovate-loop/issues/1 (PATCH,200)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/issues/2/comments (GET,200)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/pulls (GET,200)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/statuses/78dae2cf45f995142afb48597777d62fa9677a1f (POST,201)": 1,
"https://api.github.com/repos/whitesource/merge-confidence/contents/beta.json (GET,200)": 1,
"https://pypi.org/pypi/certifi/json (GET,200)": 1
},
"hostStats": {
"api.github.com": {
"requestCount": 16,
"requestAvgMs": 269,
"queueAvgMs": 0
},
"pypi.org": {
"requestCount": 1,
"requestAvgMs": 989,
"queueAvgMs": 0
}
},
"totalRequests": 17
}
INFO: Repository finished
{
"durationMs": 45597
}
```
</details>
### Have you created a minimal reproduction repository?
I have linked to a minimal reproduction repository in the bug description | 1.0 | Renovate is stuck in an infinite loop overwriting a ci bot's commits - ### How are you running Renovate?
Mend Renovate hosted app on github.com
### If you're self-hosting Renovate, tell us what version of Renovate you run.
_No response_
### If you're self-hosting Renovate, select which platform you are using.
_No response_
### If you're self-hosting Renovate, tell us what version of the platform you run.
_No response_
### Was this something which used to work for you, and then stopped?
It used to work, and then stopped
### Describe the bug
Renovate is stuck in an infinite loop overwriting a ci bot's fixup commits. This started happening sometime in the past 2 months (the configuration has not been changed since). So far I have only observed this with python packages, but I do not have a similar procedure for other dependencies.
The process is something like this:
1. Renovate updates package `x` and commits to the repository, which triggers the first ci run.
2. The repository uses lock files with `pip-compile`, so before running builds and tests, the ci agent runs a script to update the lockfiles. This takes care of resolving dependency changes of `x`. If any changes were made, the ci bot commits them as a fixup commit (`git commit --fixup HEAD`) and pushes it to GitHub.
3. The new commit triggers a second ci run. This time, the check should pass and move on.
The ci bot is registered as a `gitIgnoredAuthor`, but in the past renovate would only overwrite the changes if the `main` branch was updated (I have `rebaseWhen` as `behind-base-branch`). Currently, every time the ci bot commits a change renovate seems to overwrite it the next time it runs. This means the repository update process is effectively stuck in an infinite loop. I'm not sure exactly what triggers this behavior and it's been a while since it worked, but it seems to have been around the same time renovate changed email names from `bot@renovateapp.com` to `29139614+renovate[bot]@users.noreply.github.com`. I recall having to manually rebase the existing branches as they were marked as modified.
I've provided debug logs from a minimal reproduction repository, which is publicly available on [GitHub](https://github.com/corypaik/renovate-loop). The repository contains one example package that causes this issue, all others have been removed or disabled. There is also a public [Buildkite](https://buildkite.com/corypaik/renovate-loop) available for reference.
### Relevant debug logs
<details><summary>Logs</summary>
```
DEBUG: No dangling containers to remove
INFO: Repository started
{
"renovateVersion": "32.184.2"
}
DEBUG: Using localDir: /mnt/renovate/gh/corypaik/renovate-loop
DEBUG: PackageFiles.clear() - Package files deleted
{
"baseBranches": []
}
DEBUG: resetMemCache()
DEBUG: initRepo("corypaik/renovate-loop")
DEBUG: corypaik/renovate-loop default branch = main
DEBUG: Using app token for git init
DEBUG: Repository cache is restored from revision 13
DEBUG: Resetting npmrc
DEBUG: detectSemanticCommits()
DEBUG: checkOnboarding()
DEBUG: isOnboarded()
DEBUG: Checking cached config file name
DEBUG: Existing config file confirmed
DEBUG: Repository config
{
"fileName": "renovate.json",
"config": {
"$schema": "https://docs.renovatebot.com/renovate-schema.json",
"extends": [
":ignoreModulesAndTests",
"group:monorepos",
"group:recommended",
"workarounds:all"
],
"semanticCommits": "enabled",
"semanticCommitType": "build",
"semanticCommitScope": "",
"prHourlyLimit": 2,
"prConcurrentLimit": 10,
"dependencyDashboard": true,
"labels": [
"target: minor",
"comp: build & ci",
"action: review"
],
"timezone": "America/Chihuahua",
"schedule": [
"after 10pm every weekday",
"before 4am every weekday"
],
"baseBranches": [
"main"
],
"rebaseWhen": "behind-base-branch",
"pip_requirements": {
"fileMatch": [
".*requirements.in$"
]
},
"prCreation": "not-pending",
"stabilityDays": 3,
"gitIgnoredAuthors": [
"buildkite-bot@buildkite.com"
],
"pre-commit": {
"enabled": true
},
"packageRules": [
{
"description": "Always pin direct python dependencies to ensure lockfile updates do not revert to a different version.",
"matchFiles": [
"third_party/requirements.in"
],
"rangeStrategy": "pin"
},
{
"description": "All python deps, except datasets (for minimal reproduction).",
"matchPackagePatterns": [
"*"
],
"matchUpdateTypes": [
"major",
"minor",
"patch"
],
"matchLanguages": [
"python"
],
"excludePackageNames": [
"datasets"
],
"enabled": false
}
]
}
}
DEBUG: Repo is onboarded
DEBUG: migrateAndValidate()
DEBUG: No config migration necessary
DEBUG: massaged config
{
"config": {
"$schema": "https://docs.renovatebot.com/renovate-schema.json",
"extends": [
"github>whitesource/merge-confidence:beta",
":ignoreModulesAndTests",
"group:monorepos",
"group:recommended",
"workarounds:all"
],
"semanticCommits": "enabled",
"semanticCommitType": "build",
"semanticCommitScope": "",
"prHourlyLimit": 2,
"prConcurrentLimit": 10,
"dependencyDashboard": true,
"labels": [
"target: minor",
"comp: build & ci",
"action: review"
],
"timezone": "America/Chihuahua",
"schedule": [
"after 10pm every weekday",
"before 4am every weekday"
],
"baseBranches": [
"main"
],
"rebaseWhen": "behind-base-branch",
"pip_requirements": {
"fileMatch": [
".*requirements.in$"
]
},
"prCreation": "not-pending",
"stabilityDays": 3,
"gitIgnoredAuthors": [
"buildkite-bot@buildkite.com"
],
"pre-commit": {
"enabled": true
},
"packageRules": [
{
"description": [
"Always pin direct python dependencies to ensure lockfile updates do not revert to a different version."
],
"matchFiles": [
"third_party/requirements.in"
],
"rangeStrategy": "pin"
},
{
"description": [
"All python deps, except datasets (for minimal reproduction)."
],
"matchPackagePatterns": [
"*"
],
"matchUpdateTypes": [
"major",
"minor",
"patch"
],
"matchLanguages": [
"python"
],
"excludePackageNames": [
"datasets"
],
"enabled": false
}
]
}
}
DEBUG: migrated config
{
"config": {
"$schema": "https://docs.renovatebot.com/renovate-schema.json",
"extends": [
"github>whitesource/merge-confidence:beta",
":ignoreModulesAndTests",
"group:monorepos",
"group:recommended",
"workarounds:all"
],
"semanticCommits": "enabled",
"semanticCommitType": "build",
"semanticCommitScope": "",
"prHourlyLimit": 2,
"prConcurrentLimit": 10,
"dependencyDashboard": true,
"labels": [
"target: minor",
"comp: build & ci",
"action: review"
],
"timezone": "America/Chihuahua",
"schedule": [
"after 10pm every weekday",
"before 4am every weekday"
],
"baseBranches": [
"main"
],
"rebaseWhen": "behind-base-branch",
"pip_requirements": {
"fileMatch": [
".*requirements.in$"
]
},
"prCreation": "not-pending",
"stabilityDays": 3,
"gitIgnoredAuthors": [
"buildkite-bot@buildkite.com"
],
"pre-commit": {
"enabled": true
},
"packageRules": [
{
"description": [
"Always pin direct python dependencies to ensure lockfile updates do not revert to a different version."
],
"matchFiles": [
"third_party/requirements.in"
],
"rangeStrategy": "pin"
},
{
"description": [
"All python deps, except datasets (for minimal reproduction)."
],
"matchPackagePatterns": [
"*"
],
"matchUpdateTypes": [
"major",
"minor",
"patch"
],
"matchLanguages": [
"python"
],
"excludePackageNames": [
"datasets"
],
"enabled": false
}
]
}
}
DEBUG: Setting hostRules from config
DEBUG: Found repo ignorePaths
{
"ignorePaths": [
"**/node_modules/**",
"**/bower_components/**",
"**/vendor/**",
"**/examples/**",
"**/__tests__/**",
"**/test/**",
"**/tests/**",
"**/__fixtures__/**"
]
}
DEBUG: No vulnerability alerts found
DEBUG: No vulnerability alerts found
DEBUG: findIssue(Dependency Dashboard)
DEBUG: Retrieving issueList
DEBUG: Retrieved 1 issues
DEBUG: Found issue 1
DEBUG: baseBranches
{
"baseBranches": [
"main"
]
}
DEBUG: baseBranch: main
{
"baseBranch": "main"
}
DEBUG: extract()
{
"baseBranch": "main"
}
DEBUG: Found cached extract
{
"baseBranch": "main",
"baseBranchSha": "5267848cc1c6b6dfb1d031f573b25d2ab36ebc82"
}
DEBUG: Deleted cached dep updates
{
"baseBranch": "main"
}
INFO: Dependency extraction complete
{
"baseBranch": "main",
"stats": {
"managers": {
"bazel": {
"fileCount": 1,
"depCount": 1
},
"pip_requirements": {
"fileCount": 2,
"depCount": 24
}
},
"total": {
"fileCount": 3,
"depCount": 25
}
}
}
DEBUG: baseBranch: main
{
"baseBranch": "main"
}
DEBUG: PackageFiles.add() - Package file saved for branch
{
"baseBranch": "main"
}
DEBUG: Package releases lookups complete
{
"baseBranch": "main"
}
DEBUG: branchifyUpgrades
{
"baseBranch": "main"
}
DEBUG: 4 flattened updates found: datasets, datasets, datasets, datasets
{
"baseBranch": "main"
}
DEBUG: Returning 2 branch(es)
{
"baseBranch": "main"
}
DEBUG: config.repoIsOnboarded=true
{
"baseBranch": "main"
}
DEBUG: packageFiles with updates
{
"baseBranch": "main",
"config": {
"bazel": [
{
"packageFile": "WORKSPACE",
"deps": [
{
"datasource": "github-tags",
"depType": "http_archive",
"depName": "rules_python",
"packageName": "bazelbuild/rules_python",
"currentValue": "refs",
"managerData": {
"def": "http_archive(\n name = \"rules_python\",\n sha256 = \"b593d13bb43c94ce94b483c2858e53a9b811f6f10e1e0eedc61073bd90e58d9c\",\n strip_prefix = \"rules_python-0.12.0\",\n url = \"https://github.com/bazelbuild/rules_python/archive/refs/tags/0.12.0.tar.gz\",\n)"
},
"depIndex": 0,
"warnings": [],
"versioning": "semver",
"skipReason": "invalid-value",
"updates": []
}
]
}
],
"pip_requirements": [
{
"packageFile": "third_party/requirements.in",
"deps": [
{
"depName": "datasets",
"currentValue": "==1.11.0",
"datasource": "pypi",
"currentVersion": "1.11.0",
"depIndex": 0,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/huggingface/datasets",
"isSingleVersion": true,
"fixedVersion": "1.11.0",
"updates": [
{
"bucket": "non-major",
"newVersion": "1.18.4",
"newValue": "==1.18.4",
"releaseTimestamp": "2022-03-07T10:50:07.000Z",
"newMajor": 1,
"newMinor": 18,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/datasets-1.x"
},
{
"bucket": "major",
"newVersion": "2.4.0",
"newValue": "==2.4.0",
"releaseTimestamp": "2022-07-25T13:21:43.000Z",
"newMajor": 2,
"newMinor": 4,
"updateType": "major",
"isRange": true,
"branchName": "renovate/datasets-2.x"
}
]
}
]
},
{
"packageFile": "third_party/requirements.txt",
"deps": [
{
"depName": "certifi",
"currentValue": "==2021.10.8",
"datasource": "pypi",
"currentVersion": "2021.10.8",
"depIndex": 0,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/certifi/python-certifi",
"isSingleVersion": true,
"fixedVersion": "2021.10.8",
"updates": [
{
"bucket": "major",
"newVersion": "2022.6.15",
"newValue": "==2022.6.15",
"releaseTimestamp": "2022-06-15T18:09:18.000Z",
"newMajor": 2022,
"newMinor": 6,
"updateType": "major",
"isRange": true,
"branchName": "renovate/certifi-2022.x"
}
]
},
{
"depName": "charset-normalizer",
"currentValue": "==2.0.12",
"datasource": "pypi",
"currentVersion": "2.0.12",
"depIndex": 1,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/ousret/charset_normalizer",
"isSingleVersion": true,
"fixedVersion": "2.0.12",
"updates": [
{
"bucket": "non-major",
"newVersion": "2.1.1",
"newValue": "==2.1.1",
"releaseTimestamp": "2022-08-19T22:13:46.000Z",
"newMajor": 2,
"newMinor": 1,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/charset-normalizer-2.x"
}
]
},
{
"depName": "datasets",
"currentValue": "==1.11.0",
"datasource": "pypi",
"currentVersion": "1.11.0",
"depIndex": 2,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/huggingface/datasets",
"isSingleVersion": true,
"fixedVersion": "1.11.0",
"updates": [
{
"bucket": "non-major",
"newVersion": "1.18.4",
"newValue": "==1.18.4",
"releaseTimestamp": "2022-03-07T10:50:07.000Z",
"newMajor": 1,
"newMinor": 18,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/datasets-1.x"
},
{
"bucket": "major",
"newVersion": "2.4.0",
"newValue": "==2.4.0",
"releaseTimestamp": "2022-07-25T13:21:43.000Z",
"newMajor": 2,
"newMinor": 4,
"updateType": "major",
"isRange": true,
"branchName": "renovate/datasets-2.x"
}
]
},
{
"depName": "dill",
"currentValue": "==0.3.1.1",
"datasource": "pypi",
"currentVersion": "0.3.1.1",
"depIndex": 3,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/uqfoundation/dill",
"isSingleVersion": true,
"fixedVersion": "0.3.1.1",
"updates": [
{
"bucket": "non-major",
"newVersion": "0.3.5.1",
"newValue": "==0.3.5.1",
"releaseTimestamp": "2022-05-20T11:21:11.000Z",
"newMajor": 0,
"newMinor": 3,
"updateType": "patch",
"isRange": true,
"branchName": "renovate/dill-0.x"
}
]
},
{
"depName": "filelock",
"currentValue": "==3.6.0",
"datasource": "pypi",
"currentVersion": "3.6.0",
"depIndex": 4,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/tox-dev/py-filelock",
"isSingleVersion": true,
"fixedVersion": "3.6.0",
"updates": [
{
"bucket": "non-major",
"newVersion": "3.8.0",
"newValue": "==3.8.0",
"releaseTimestamp": "2022-08-10T06:44:31.000Z",
"newMajor": 3,
"newMinor": 8,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/filelock-3.x"
}
]
},
{
"depName": "fsspec",
"currentValue": "==2021.11.1",
"datasource": "pypi",
"currentVersion": "2021.11.1",
"depIndex": 5,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/fsspec/filesystem_spec",
"changelogUrl": "https://filesystem-spec.readthedocs.io/en/latest/changelog.html",
"isSingleVersion": true,
"fixedVersion": "2021.11.1",
"updates": [
{
"bucket": "major",
"newVersion": "2022.8.0",
"newValue": "==2022.8.0",
"releaseTimestamp": "2022-08-30T18:08:19.000Z",
"newMajor": 2022,
"newMinor": 8,
"updateType": "major",
"isRange": true,
"branchName": "renovate/fsspec-2022.x"
}
]
},
{
"depName": "huggingface-hub",
"currentValue": "==0.0.19",
"datasource": "pypi",
"currentVersion": "0.0.19",
"depIndex": 6,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/huggingface/huggingface_hub",
"isSingleVersion": true,
"fixedVersion": "0.0.19",
"updates": [
{
"bucket": "non-major",
"newVersion": "0.9.1",
"newValue": "==0.9.1",
"releaseTimestamp": "2022-08-25T15:37:11.000Z",
"newMajor": 0,
"newMinor": 9,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/huggingface-hub-0.x"
}
]
},
{
"depName": "idna",
"currentValue": "==3.3",
"datasource": "pypi",
"currentVersion": "3.3",
"depIndex": 7,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/kjd/idna",
"fixedVersion": "3.3",
"updates": []
},
{
"depName": "multiprocess",
"currentValue": "==0.70.9",
"datasource": "pypi",
"currentVersion": "0.70.9",
"depIndex": 8,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/uqfoundation/multiprocess",
"isSingleVersion": true,
"fixedVersion": "0.70.9",
"updates": [
{
"bucket": "non-major",
"newVersion": "0.70.13",
"newValue": "==0.70.13",
"releaseTimestamp": "2022-05-23T22:52:42.000Z",
"newMajor": 0,
"newMinor": 70,
"updateType": "patch",
"isRange": true,
"branchName": "renovate/multiprocess-0.x"
}
]
},
{
"depName": "numpy",
"currentValue": "==1.21.3",
"datasource": "pypi",
"currentVersion": "1.21.3",
"depIndex": 9,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/numpy/numpy",
"homepage": "https://www.numpy.org",
"isSingleVersion": true,
"fixedVersion": "1.21.3",
"updates": [
{
"bucket": "non-major",
"newVersion": "1.23.2",
"newValue": "==1.23.2",
"releaseTimestamp": "2022-08-14T00:14:09.000Z",
"newMajor": 1,
"newMinor": 23,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/numpy-1.x"
}
]
},
{
"depName": "packaging",
"currentValue": "==20.9",
"datasource": "pypi",
"currentVersion": "20.9",
"depIndex": 10,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/pypa/packaging",
"isSingleVersion": true,
"fixedVersion": "20.9",
"updates": [
{
"bucket": "major",
"newVersion": "21.3",
"newValue": "==21.3",
"releaseTimestamp": "2021-11-18T00:39:10.000Z",
"newMajor": 21,
"newMinor": 3,
"updateType": "major",
"isRange": true,
"branchName": "renovate/packaging-21.x"
}
]
},
{
"depName": "pandas",
"currentValue": "==1.4.2",
"datasource": "pypi",
"currentVersion": "1.4.2",
"depIndex": 11,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/pandas-dev/pandas",
"homepage": "https://pandas.pydata.org",
"isSingleVersion": true,
"fixedVersion": "1.4.2",
"updates": [
{
"bucket": "non-major",
"newVersion": "1.4.3",
"newValue": "==1.4.3",
"releaseTimestamp": "2022-06-23T13:30:09.000Z",
"newMajor": 1,
"newMinor": 4,
"updateType": "patch",
"isRange": true,
"branchName": "renovate/pandas-1.x"
}
]
},
{
"depName": "pyarrow",
"currentValue": "==5.0.0",
"datasource": "pypi",
"currentVersion": "5.0.0",
"depIndex": 12,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/apache/arrow",
"homepage": "https://arrow.apache.org/",
"isSingleVersion": true,
"fixedVersion": "5.0.0",
"updates": [
{
"bucket": "major",
"newVersion": "9.0.0",
"newValue": "==9.0.0",
"releaseTimestamp": "2022-08-03T10:45:55.000Z",
"newMajor": 9,
"newMinor": 0,
"updateType": "major",
"isRange": true,
"branchName": "renovate/pyarrow-9.x"
}
]
},
{
"depName": "pyparsing",
"currentValue": "==2.4.7",
"datasource": "pypi",
"currentVersion": "2.4.7",
"depIndex": 13,
"warnings": [],
"versioning": "pep440",
"isSingleVersion": true,
"fixedVersion": "2.4.7",
"updates": [
{
"bucket": "major",
"newVersion": "3.0.9",
"newValue": "==3.0.9",
"releaseTimestamp": "2022-05-10T23:26:03.000Z",
"newMajor": 3,
"newMinor": 0,
"updateType": "major",
"isRange": true,
"branchName": "renovate/pyparsing-3.x"
}
]
},
{
"depName": "python-dateutil",
"currentValue": "==2.8.2",
"datasource": "pypi",
"currentVersion": "2.8.2",
"depIndex": 14,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/dateutil/dateutil",
"fixedVersion": "2.8.2",
"updates": []
},
{
"depName": "pytz",
"currentValue": "==2022.1",
"datasource": "pypi",
"currentVersion": "2022.1",
"depIndex": 15,
"warnings": [],
"versioning": "pep440",
"homepage": "http://pythonhosted.org/pytz",
"isSingleVersion": true,
"fixedVersion": "2022.1",
"updates": [
{
"bucket": "non-major",
"newVersion": "2022.2.1",
"newValue": "==2022.2.1",
"releaseTimestamp": "2022-08-13T02:07:57.000Z",
"newMajor": 2022,
"newMinor": 2,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/pytz-2022.x"
}
]
},
{
"depName": "pyyaml",
"currentValue": "==5.4.1",
"datasource": "pypi",
"currentVersion": "5.4.1",
"depIndex": 16,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/yaml/pyyaml",
"homepage": "https://pyyaml.org/",
"isSingleVersion": true,
"fixedVersion": "5.4.1",
"updates": [
{
"bucket": "major",
"newVersion": "6.0",
"newValue": "==6.0",
"releaseTimestamp": "2021-10-13T19:39:42.000Z",
"newMajor": 6,
"newMinor": 0,
"updateType": "major",
"isRange": true,
"branchName": "renovate/pyyaml-6.x"
}
]
},
{
"depName": "requests",
"currentValue": "==2.27.1",
"datasource": "pypi",
"currentVersion": "2.27.1",
"depIndex": 17,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/psf/requests",
"homepage": "https://requests.readthedocs.io",
"changelogUrl": "https://github.com/psf/requests/blob/master/HISTORY.md",
"isSingleVersion": true,
"fixedVersion": "2.27.1",
"updates": [
{
"bucket": "non-major",
"newVersion": "2.28.1",
"newValue": "==2.28.1",
"releaseTimestamp": "2022-06-29T15:13:40.000Z",
"newMajor": 2,
"newMinor": 28,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/requests-2.x"
}
]
},
{
"depName": "six",
"currentValue": "==1.16.0",
"datasource": "pypi",
"currentVersion": "1.16.0",
"depIndex": 18,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/benjaminp/six",
"fixedVersion": "1.16.0",
"updates": []
},
{
"depName": "tqdm",
"currentValue": "==4.63.0",
"datasource": "pypi",
"currentVersion": "4.63.0",
"depIndex": 19,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/tqdm/tqdm",
"homepage": "https://tqdm.github.io",
"changelogUrl": "https://tqdm.github.io/releases",
"isSingleVersion": true,
"fixedVersion": "4.63.0",
"updates": [
{
"bucket": "non-major",
"newVersion": "4.64.0",
"newValue": "==4.64.0",
"releaseTimestamp": "2022-04-04T01:48:46.000Z",
"newMajor": 4,
"newMinor": 64,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/tqdm-4.x"
}
]
},
{
"depName": "typing-extensions",
"currentValue": "==4.1.1",
"datasource": "pypi",
"currentVersion": "4.1.1",
"depIndex": 20,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/python/typing_extensions",
"changelogUrl": "https://github.com/python/typing_extensions/blob/main/CHANGELOG.md",
"isSingleVersion": true,
"fixedVersion": "4.1.1",
"updates": [
{
"bucket": "non-major",
"newVersion": "4.3.0",
"newValue": "==4.3.0",
"releaseTimestamp": "2022-07-01T14:56:11.000Z",
"newMajor": 4,
"newMinor": 3,
"updateType": "minor",
"isRange": true,
"branchName": "renovate/typing-extensions-4.x"
}
]
},
{
"depName": "urllib3",
"currentValue": "==1.26.8",
"datasource": "pypi",
"currentVersion": "1.26.8",
"depIndex": 21,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/urllib3/urllib3",
"homepage": "https://urllib3.readthedocs.io/",
"isSingleVersion": true,
"fixedVersion": "1.26.8",
"updates": [
{
"bucket": "non-major",
"newVersion": "1.26.12",
"newValue": "==1.26.12",
"releaseTimestamp": "2022-08-22T13:27:19.000Z",
"newMajor": 1,
"newMinor": 26,
"updateType": "patch",
"isRange": true,
"branchName": "renovate/urllib3-1.x"
}
]
},
{
"depName": "xxhash",
"currentValue": "==3.0.0",
"datasource": "pypi",
"currentVersion": "3.0.0",
"depIndex": 22,
"warnings": [],
"versioning": "pep440",
"sourceUrl": "https://github.com/ifduyue/python-xxhash",
"fixedVersion": "3.0.0",
"updates": []
}
]
}
]
}
}
DEBUG: processRepo()
DEBUG: Processing 2 branches: renovate/datasets-1.x, renovate/datasets-2.x
DEBUG: Calculating hourly PRs remaining
DEBUG: getPrList success
{
"pullsTotal": 1,
"requestsTotal": 1,
"apiQuotaAffected": true
}
DEBUG: currentHourStart=2022-08-31T02:00:00.000+00:00
DEBUG: PR hourly limit remaining: 2
DEBUG: Calculating prConcurrentLimit (10)
DEBUG: getBranchPr(renovate/datasets-1.x)
DEBUG: findPr(renovate/datasets-1.x, undefined, open)
DEBUG: Found PR #2
DEBUG: getBranchPr(renovate/datasets-2.x)
DEBUG: findPr(renovate/datasets-2.x, undefined, open)
DEBUG: findPr(renovate/datasets-2.x, undefined, closed)
DEBUG: 1 PRs are currently open
DEBUG: PR concurrent limit remaining: 9
DEBUG: Calculated maximum PRs remaining this run
{
"prsRemaining": 2
}
DEBUG: PullRequests limit = 2
DEBUG: Calculating hourly PRs remaining
DEBUG: currentHourStart=2022-08-31T02:00:00.000+00:00
DEBUG: PR hourly limit remaining: 2
DEBUG: Calculating branchConcurrentLimit (10)
DEBUG: 1 already existing branches found: renovate/datasets-1.x
DEBUG: Branch concurrent limit remaining: 9
DEBUG: Calculated maximum branches remaining this run
{
"branchesRemaining": 2
}
DEBUG: Branches limit = 2
DEBUG: Branch fingerprint is unchanged, updates check can be skipped(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: getBranchPr(renovate/datasets-1.x)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: findPr(renovate/datasets-1.x, undefined, open)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Found PR #2(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: branchExists=true(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: dependencyDashboardCheck=undefined(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: PR rebase requested=false(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Checking if PR has been edited(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Found existing branch PR(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Checking schedule(after 10pm every weekday,before 4am every weekday, America/Chihuahua)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Found timezone(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"timezone": "America/Chihuahua"
}
DEBUG: Adjusting now for timezone(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Checking 2 schedule(s)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Checking schedule "after 10pm every weekday"(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"parsedSchedule": {
"schedules": [
{
"t_a": [
79200
],
"d": [
2,
3,
4,
5,
6
]
}
],
"exceptions": [],
"error": -1
}
}
DEBUG: Checking schedule "before 4am every weekday"(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"parsedSchedule": {
"schedules": [
{
"t_b": [
14400
],
"d": [
2,
3,
4,
5,
6
]
}
],
"exceptions": [],
"error": -1
}
}
DEBUG: Package not scheduled(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Branch + PR exists but is not scheduled -- will update if necessary(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Branch already exists(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: getBranchPr(renovate/datasets-1.x)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: findPr(renovate/datasets-1.x, undefined, open)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Found PR #2(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Branch is behind base branch and needs rebasing(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Branch is unmodified, so can be rebased(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Using reuseExistingBranch: false(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Setting current branch to main(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Initializing git repository into /mnt/renovate/gh/corypaik/renovate-loop(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Performing blobless clone(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: git clone completed(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"durationMs": 1449
}
DEBUG: latest repository commit(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"latestCommit": {
"hash": "5267848cc1c6b6dfb1d031f573b25d2ab36ebc82",
"date": "2022-08-30T20:06:50-06:00",
"message": "ci: always run pip lockfile step",
"refs": "HEAD -> main, origin/main, origin/HEAD",
"body": "",
"author_name": "Cory Paik",
"author_email": "corypaik@gmail.com"
}
}
DEBUG: latest commit(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"branchName": "main",
"latestCommitDate": "2022-08-30T20:06:50-06:00"
}
DEBUG: manager.getUpdatedPackageFiles() reuseExistingBranch=false(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Starting search at index 527(packageFile="third_party/requirements.txt", branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"depName": "datasets"
}
DEBUG: Found match at index 527(packageFile="third_party/requirements.txt", branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"depName": "datasets"
}
DEBUG: Contents updated(packageFile="third_party/requirements.txt", branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"depName": "datasets"
}
DEBUG: Starting search at index 8(packageFile="third_party/requirements.in", branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"depName": "datasets"
}
DEBUG: Found match at index 8(packageFile="third_party/requirements.in", branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"depName": "datasets"
}
DEBUG: Contents updated(packageFile="third_party/requirements.in", branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"depName": "datasets"
}
DEBUG: pip_requirements.updateArtifacts(third_party/requirements.txt)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Setting BUILDPACK_CACHE_DIR to /tmp/containerbase(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Using docker to execute(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"image": "sidecar"
}
DEBUG: Resolved stable matching version(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"toolName": "python",
"resolvedVersion": "3.10.6"
}
DEBUG: containerbaseDir is separate from cacheDir(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: No tag or tagConstraint specified(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"image": "docker.io/renovate/sidecar"
}
DEBUG: Fetching Docker image: docker.io/renovate/sidecar(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Finished fetching Docker image docker.io/renovate/sidecar@sha256:ee8108a5a17ef370f15da2c11bf7b2da9035153fde1587118ac9cc5c6c56cf0a(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Executing command(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"command": "docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/corypaik/renovate-loop\":\"/mnt/renovate/gh/corypaik/renovate-loop\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -w \"/mnt/renovate/gh/corypaik/renovate-loop\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.10.6 && pip install --user hashin && hashin datasets==1.18.4 -r third_party/requirements.txt\""
}
DEBUG: exec completed(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"cmd": "docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/corypaik/renovate-loop\":\"/mnt/renovate/gh/corypaik/renovate-loop\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e PIP_CACHE_DIR -e BUILDPACK_CACHE_DIR -w \"/mnt/renovate/gh/corypaik/renovate-loop\" docker.io/renovate/sidecar bash -l -c \"install-tool python 3.10.6 && pip install --user hashin && hashin datasets==1.18.4 -r third_party/requirements.txt\"",
"durationMs": 27296,
"stdout": "installing v2 tool python v3.10.6\nUsing prebuild python\nRequirement already satisfied: pip in /opt/buildpack/tools/python/3.10.6/lib/python3.10/site-packages (22.2.1)\nCollecting pip\n Downloading pip-22.2.2-py3-none-any.whl (2.0 MB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.0/2.0 MB 21.0 MB/s eta 0:00:00\nInstalling collected packages: pip\n Attempting uninstall: pip\n Found existing installation: pip 22.2.1\n Uninstalling pip-22.2.1:\n Successfully uninstalled pip-22.2.1\nSuccessfully installed pip-22.2.2\nFiles removed: 2\nlinking tool python v3.10.6\nPython 3.10.6\npip 22.2.2 from /opt/buildpack/tools/python/3.10.6/lib/python3.10/site-packages/pip (python 3.10)\nInstalled v2 /usr/local/buildpack/tools/v2/python.sh in 20 seconds\nCollecting hashin\n Downloading hashin-0.17.0-py2.py3-none-any.whl (16 kB)\nCollecting pip-api\n Downloading pip_api-0.0.30-py3-none-any.whl (111 kB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 111.6/111.6 kB 5.9 MB/s eta 0:00:00\nCollecting packaging\n Downloading packaging-21.3-py3-none-any.whl (40 kB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 40.8/40.8 kB 4.7 MB/s eta 0:00:00\nCollecting pyparsing!=3.0.5,>=2.0.2\n Downloading pyparsing-3.0.9-py3-none-any.whl (98 kB)\n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 98.3/98.3 kB 11.1 MB/s eta 0:00:00\nRequirement already satisfied: pip in /opt/buildpack/tools/python/3.10.6/lib/python3.10/site-packages (from pip-api->hashin) (22.2.2)\nInstalling collected packages: pyparsing, pip-api, packaging, hashin\nSuccessfully installed hashin-0.17.0 packaging-21.3 pip-api-0.0.30 pyparsing-3.0.9\n",
"stderr": "Found file in cache: /tmp/containerbase/f79e18f004c2278a4c76a549ca9e8beed6ee66d17e187e79a230ba06836df191/python-3.10.6-focal-x86_64.tar.xz\n"
}
DEBUG: Returning updated third_party/requirements.txt(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: pip_requirements.updateArtifacts(third_party/requirements.in)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: No hashin commands to run - returning(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Updated 2 package files(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Updated 1 lock files(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"updatedArtifacts": [
"third_party/requirements.txt"
]
}
DEBUG: Getting comments for #2(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Found 0 comments(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: isBranchConflicted(main, renovate/datasets-1.x)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Using cached result false for isBranchConflicted(main, renovate/datasets-1.x)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: 2 file(s) to commit(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Preparing files for committing to branch renovate/datasets-1.x(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Setting git author name(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"gitAuthorName": "Renovate Bot"
}
DEBUG: Setting git author email(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"gitAuthorEmail": "bot@renovateapp.com"
}
DEBUG: git commit(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"deletedFiles": [],
"ignoredFiles": [],
"result": {
"author": null,
"branch": "renovate/datasets-1.x",
"commit": "210146775398f010baed48f2e6891d780406efa2",
"root": false,
"summary": {
"changes": 2,
"insertions": 4,
"deletions": 4
}
}
}
DEBUG: resetToCommit(5267848cc1c6b6dfb1d031f573b25d2ab36ebc82)(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Fetching branch renovate/datasets-1.x(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
INFO: Branch updated(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"commitSha": "78dae2cf45f995142afb48597777d62fa9677a1f"
}
DEBUG: Updating renovate/stability-days status check state to green(branch="renovate/datasets-1.x")
{
"baseBranch": "main"
}
DEBUG: Setting branch status(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"context": "renovate/stability-days",
"state": "green"
}
DEBUG: Branch status pending(branch="renovate/datasets-1.x")
{
"baseBranch": "main",
"commitSha": "78dae2cf45f995142afb48597777d62fa9677a1f"
}
DEBUG: getBranchPr(renovate/datasets-2.x)(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: findPr(renovate/datasets-2.x, undefined, open)(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: findPr(renovate/datasets-2.x, undefined, closed)(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: branchExists=false(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: dependencyDashboardCheck=undefined(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: recreateClosed is false(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: findPr(renovate/datasets-2.x, build: update dependency datasets to v2, !open)(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: prAlreadyExisted=false(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: Checking schedule(after 10pm every weekday,before 4am every weekday, America/Chihuahua)(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: Found timezone(branch="renovate/datasets-2.x")
{
"baseBranch": "main",
"timezone": "America/Chihuahua"
}
DEBUG: Adjusting now for timezone(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: Checking 2 schedule(s)(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: Checking schedule "after 10pm every weekday"(branch="renovate/datasets-2.x")
{
"baseBranch": "main",
"parsedSchedule": {
"schedules": [
{
"t_a": [
79200
],
"d": [
2,
3,
4,
5,
6
]
}
],
"exceptions": [],
"error": -1
}
}
DEBUG: Checking schedule "before 4am every weekday"(branch="renovate/datasets-2.x")
{
"baseBranch": "main",
"parsedSchedule": {
"schedules": [
{
"t_b": [
14400
],
"d": [
2,
3,
4,
5,
6
]
}
],
"exceptions": [],
"error": -1
}
}
DEBUG: Package not scheduled(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: Skipping branch creation as not within schedule(branch="renovate/datasets-2.x")
{
"baseBranch": "main"
}
DEBUG: getBranchPr(renovate/datasets-1.x)
DEBUG: findPr(renovate/datasets-1.x, undefined, open)
DEBUG: Found PR #2
DEBUG: Ensuring Dependency Dashboard
DEBUG: ensureIssue(Dependency Dashboard)
DEBUG: Patching issue
DEBUG: Issue updated
DEBUG: Removing any stale branches
DEBUG: config.repoIsOnboarded=true
DEBUG: Branch lists
{
"branchList": [
"renovate/datasets-1.x",
"renovate/datasets-2.x"
],
"renovateBranches": [
"renovate/datasets-1.x"
]
}
DEBUG: remainingBranches=
DEBUG: No branches to clean up
DEBUG: Cleaning up Renovate refs: refs/renovate/*
DEBUG: PackageFiles.clear() - Package files deleted
{
"baseBranches": [
"main"
]
}
DEBUG: Renovate repository PR statistics
{
"stats": {
"total": 1,
"open": 1,
"closed": 0,
"merged": 0
}
}
DEBUG: Repository result: done, status: onboarded, enabled: true, onboarded: true
DEBUG: Repository timing splits (milliseconds)
{
"splits": {
"init": 3599,
"extract": 494,
"lookup": 2772,
"onboarding": 0,
"update": 35845
},
"total": 45597
}
DEBUG: http statistics
{
"urls": {
"https://api.github.com/graphql (POST,200)": 2,
"https://api.github.com/repos/corypaik/renovate-loop/commits/78dae2cf45f995142afb48597777d62fa9677a1f/statuses (GET,200)": 2,
"https://api.github.com/repos/corypaik/renovate-loop/commits/renovate/datasets-1.x/status (GET,200)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/contents/renovate.json (GET,200)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/git/commits (POST,201)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/git/refs/heads/renovate/datasets-1.x (PATCH,200)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/git/trees (POST,201)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/issues/1 (GET,200)": 2,
"https://api.github.com/repos/corypaik/renovate-loop/issues/1 (PATCH,200)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/issues/2/comments (GET,200)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/pulls (GET,200)": 1,
"https://api.github.com/repos/corypaik/renovate-loop/statuses/78dae2cf45f995142afb48597777d62fa9677a1f (POST,201)": 1,
"https://api.github.com/repos/whitesource/merge-confidence/contents/beta.json (GET,200)": 1,
"https://pypi.org/pypi/certifi/json (GET,200)": 1
},
"hostStats": {
"api.github.com": {
"requestCount": 16,
"requestAvgMs": 269,
"queueAvgMs": 0
},
"pypi.org": {
"requestCount": 1,
"requestAvgMs": 989,
"queueAvgMs": 0
}
},
"totalRequests": 17
}
INFO: Repository finished
{
"durationMs": 45597
}
```
</details>
### Have you created a minimal reproduction repository?
I have linked to a minimal reproduction repository in the bug description | priority | renovate is stuck in an infinite loop overwriting a ci bot s commits how are you running renovate mend renovate hosted app on github com if you re self hosting renovate tell us what version of renovate you run no response if you re self hosting renovate select which platform you are using no response if you re self hosting renovate tell us what version of the platform you run no response was this something which used to work for you and then stopped it used to work and then stopped describe the bug renovate is stuck in an infinite loop overwriting a ci bot s fixup commits this started happening sometime in the past months the configuration has not been changed since so far i have only observed this with python packages but i do not have a similar procedure for other dependencies the process is something like this renovate updates package x and commits to the repository which triggers the first ci run the repository uses lock files with pip compile so before running builds and tests the ci agent runs a script to update the lockfiles this takes care of resolving dependency changes of x if any changes were made the ci bot commits them as a fixup commit git commit fixup head and pushes it to github the new commit triggers a second ci run this time the check should pass and move on the ci bot is registered as a gitignoredauthor but in the past renovate would only overwrite the changes if the main branch was updated i have rebasewhen as behind base branch currently every time the ci bot commits a change renovate seems to overwrite it the next time it runs this means the repository update process is effectively stuck in an infinite loop i m not sure exactly what triggers this behavior and it s been a while since it worked but it seems to have been around the same time renovate changed email names from bot renovateapp com to renovate users noreply github com i recall having to manually rebase the existing branches as they were marked as modified i ve provided debug logs from a minimal reproduction repository which is publicly available on the repository contains one example package that causes this issue all others have been removed or disabled there is also a public available for reference relevant debug logs logs debug no dangling containers to remove info repository started renovateversion debug using localdir mnt renovate gh corypaik renovate loop debug packagefiles clear package files deleted basebranches debug resetmemcache debug initrepo corypaik renovate loop debug corypaik renovate loop default branch main debug using app token for git init debug repository cache is restored from revision debug resetting npmrc debug detectsemanticcommits debug checkonboarding debug isonboarded debug checking cached config file name debug existing config file confirmed debug repository config filename renovate json config schema extends ignoremodulesandtests group monorepos group recommended workarounds all semanticcommits enabled semanticcommittype build semanticcommitscope prhourlylimit prconcurrentlimit dependencydashboard true labels target minor comp build ci action review timezone america chihuahua schedule after every weekday before every weekday basebranches main rebasewhen behind base branch pip requirements filematch requirements in prcreation not pending stabilitydays gitignoredauthors buildkite bot buildkite com pre commit enabled true packagerules description always pin direct python dependencies to ensure lockfile updates do not revert to a different version matchfiles third party requirements in rangestrategy pin description all python deps except datasets for minimal reproduction matchpackagepatterns matchupdatetypes major minor patch matchlanguages python excludepackagenames datasets enabled false debug repo is onboarded debug migrateandvalidate debug no config migration necessary debug massaged config config schema extends github whitesource merge confidence beta ignoremodulesandtests group monorepos group recommended workarounds all semanticcommits enabled semanticcommittype build semanticcommitscope prhourlylimit prconcurrentlimit dependencydashboard true labels target minor comp build ci action review timezone america chihuahua schedule after every weekday before every weekday basebranches main rebasewhen behind base branch pip requirements filematch requirements in prcreation not pending stabilitydays gitignoredauthors buildkite bot buildkite com pre commit enabled true packagerules description always pin direct python dependencies to ensure lockfile updates do not revert to a different version matchfiles third party requirements in rangestrategy pin description all python deps except datasets for minimal reproduction matchpackagepatterns matchupdatetypes major minor patch matchlanguages python excludepackagenames datasets enabled false debug migrated config config schema extends github whitesource merge confidence beta ignoremodulesandtests group monorepos group recommended workarounds all semanticcommits enabled semanticcommittype build semanticcommitscope prhourlylimit prconcurrentlimit dependencydashboard true labels target minor comp build ci action review timezone america chihuahua schedule after every weekday before every weekday basebranches main rebasewhen behind base branch pip requirements filematch requirements in prcreation not pending stabilitydays gitignoredauthors buildkite bot buildkite com pre commit enabled true packagerules description always pin direct python dependencies to ensure lockfile updates do not revert to a different version matchfiles third party requirements in rangestrategy pin description all python deps except datasets for minimal reproduction matchpackagepatterns matchupdatetypes major minor patch matchlanguages python excludepackagenames datasets enabled false debug setting hostrules from config debug found repo ignorepaths ignorepaths node modules bower components vendor examples tests test tests fixtures debug no vulnerability alerts found debug no vulnerability alerts found debug findissue dependency dashboard debug retrieving issuelist debug retrieved issues debug found issue debug basebranches basebranches main debug basebranch main basebranch main debug extract basebranch main debug found cached extract basebranch main basebranchsha debug deleted cached dep updates basebranch main info dependency extraction complete basebranch main stats managers bazel filecount depcount pip requirements filecount depcount total filecount depcount debug basebranch main basebranch main debug packagefiles add package file saved for branch basebranch main debug package releases lookups complete basebranch main debug branchifyupgrades basebranch main debug flattened updates found datasets datasets datasets datasets basebranch main debug returning branch es basebranch main debug config repoisonboarded true basebranch main debug packagefiles with updates basebranch main config bazel packagefile workspace deps datasource github tags deptype http archive depname rules python packagename bazelbuild rules python currentvalue refs managerdata def http archive n name rules python n n strip prefix rules python n url depindex warnings versioning semver skipreason invalid value updates pip requirements packagefile third party requirements in deps depname datasets currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate datasets x bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate datasets x packagefile third party requirements txt deps depname certifi currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate certifi x depname charset normalizer currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate charset normalizer x depname datasets currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate datasets x bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate datasets x depname dill currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype patch isrange true branchname renovate dill x depname filelock currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate filelock x depname fsspec currentvalue datasource pypi currentversion depindex warnings versioning sourceurl changelogurl issingleversion true fixedversion updates bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate fsspec x depname huggingface hub currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate huggingface hub x depname idna currentvalue datasource pypi currentversion depindex warnings versioning sourceurl fixedversion updates depname multiprocess currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype patch isrange true branchname renovate multiprocess x depname numpy currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate numpy x depname packaging currentvalue datasource pypi currentversion depindex warnings versioning sourceurl issingleversion true fixedversion updates bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate packaging x depname pandas currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype patch isrange true branchname renovate pandas x depname pyarrow currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage issingleversion true fixedversion updates bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate pyarrow x depname pyparsing currentvalue datasource pypi currentversion depindex warnings versioning issingleversion true fixedversion updates bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate pyparsing x depname python dateutil currentvalue datasource pypi currentversion depindex warnings versioning sourceurl fixedversion updates depname pytz currentvalue datasource pypi currentversion depindex warnings versioning homepage issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate pytz x depname pyyaml currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage issingleversion true fixedversion updates bucket major newversion newvalue releasetimestamp newmajor newminor updatetype major isrange true branchname renovate pyyaml x depname requests currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage changelogurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate requests x depname six currentvalue datasource pypi currentversion depindex warnings versioning sourceurl fixedversion updates depname tqdm currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage changelogurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate tqdm x depname typing extensions currentvalue datasource pypi currentversion depindex warnings versioning sourceurl changelogurl issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype minor isrange true branchname renovate typing extensions x depname currentvalue datasource pypi currentversion depindex warnings versioning sourceurl homepage issingleversion true fixedversion updates bucket non major newversion newvalue releasetimestamp newmajor newminor updatetype patch isrange true branchname renovate x depname xxhash currentvalue datasource pypi currentversion depindex warnings versioning sourceurl fixedversion updates debug processrepo debug processing branches renovate datasets x renovate datasets x debug calculating hourly prs remaining debug getprlist success pullstotal requeststotal apiquotaaffected true debug currenthourstart debug pr hourly limit remaining debug calculating prconcurrentlimit debug getbranchpr renovate datasets x debug findpr renovate datasets x undefined open debug found pr debug getbranchpr renovate datasets x debug findpr renovate datasets x undefined open debug findpr renovate datasets x undefined closed debug prs are currently open debug pr concurrent limit remaining debug calculated maximum prs remaining this run prsremaining debug pullrequests limit debug calculating hourly prs remaining debug currenthourstart debug pr hourly limit remaining debug calculating branchconcurrentlimit debug already existing branches found renovate datasets x debug branch concurrent limit remaining debug calculated maximum branches remaining this run branchesremaining debug branches limit debug branch fingerprint is unchanged updates check can be skipped branch renovate datasets x basebranch main debug getbranchpr renovate datasets x branch renovate datasets x basebranch main debug findpr renovate datasets x undefined open branch renovate datasets x basebranch main debug found pr branch renovate datasets x basebranch main debug branchexists true branch renovate datasets x basebranch main debug dependencydashboardcheck undefined branch renovate datasets x basebranch main debug pr rebase requested false branch renovate datasets x basebranch main debug checking if pr has been edited branch renovate datasets x basebranch main debug found existing branch pr branch renovate datasets x basebranch main debug checking schedule after every weekday before every weekday america chihuahua branch renovate datasets x basebranch main debug found timezone branch renovate datasets x basebranch main timezone america chihuahua debug adjusting now for timezone branch renovate datasets x basebranch main debug checking schedule s branch renovate datasets x basebranch main debug checking schedule after every weekday branch renovate datasets x basebranch main parsedschedule schedules t a d exceptions error debug checking schedule before every weekday branch renovate datasets x basebranch main parsedschedule schedules t b d exceptions error debug package not scheduled branch renovate datasets x basebranch main debug branch pr exists but is not scheduled will update if necessary branch renovate datasets x basebranch main debug branch already exists branch renovate datasets x basebranch main debug getbranchpr renovate datasets x branch renovate datasets x basebranch main debug findpr renovate datasets x undefined open branch renovate datasets x basebranch main debug found pr branch renovate datasets x basebranch main debug branch is behind base branch and needs rebasing branch renovate datasets x basebranch main debug branch is unmodified so can be rebased branch renovate datasets x basebranch main debug using reuseexistingbranch false branch renovate datasets x basebranch main debug setting current branch to main branch renovate datasets x basebranch main debug initializing git repository into mnt renovate gh corypaik renovate loop branch renovate datasets x basebranch main debug performing blobless clone branch renovate datasets x basebranch main debug git clone completed branch renovate datasets x basebranch main durationms debug latest repository commit branch renovate datasets x basebranch main latestcommit hash date message ci always run pip lockfile step refs head main origin main origin head body author name cory paik author email corypaik gmail com debug latest commit branch renovate datasets x basebranch main branchname main latestcommitdate debug manager getupdatedpackagefiles reuseexistingbranch false branch renovate datasets x basebranch main debug starting search at index packagefile third party requirements txt branch renovate datasets x basebranch main depname datasets debug found match at index packagefile third party requirements txt branch renovate datasets x basebranch main depname datasets debug contents updated packagefile third party requirements txt branch renovate datasets x basebranch main depname datasets debug starting search at index packagefile third party requirements in branch renovate datasets x basebranch main depname datasets debug found match at index packagefile third party requirements in branch renovate datasets x basebranch main depname datasets debug contents updated packagefile third party requirements in branch renovate datasets x basebranch main depname datasets debug pip requirements updateartifacts third party requirements txt branch renovate datasets x basebranch main debug setting buildpack cache dir to tmp containerbase branch renovate datasets x basebranch main debug using docker to execute branch renovate datasets x basebranch main image sidecar debug resolved stable matching version branch renovate datasets x basebranch main toolname python resolvedversion debug containerbasedir is separate from cachedir branch renovate datasets x basebranch main debug no tag or tagconstraint specified branch renovate datasets x basebranch main image docker io renovate sidecar debug fetching docker image docker io renovate sidecar branch renovate datasets x basebranch main debug finished fetching docker image docker io renovate sidecar branch renovate datasets x basebranch main debug executing command branch renovate datasets x basebranch main command docker run rm name renovate sidecar label renovate child v mnt renovate gh corypaik renovate loop mnt renovate gh corypaik renovate loop v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e pip cache dir e buildpack cache dir w mnt renovate gh corypaik renovate loop docker io renovate sidecar bash l c install tool python pip install user hashin hashin datasets r third party requirements txt debug exec completed branch renovate datasets x basebranch main cmd docker run rm name renovate sidecar label renovate child v mnt renovate gh corypaik renovate loop mnt renovate gh corypaik renovate loop v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e pip cache dir e buildpack cache dir w mnt renovate gh corypaik renovate loop docker io renovate sidecar bash l c install tool python pip install user hashin hashin datasets r third party requirements txt durationms stdout installing tool python nusing prebuild python nrequirement already satisfied pip in opt buildpack tools python lib site packages ncollecting pip n downloading pip none any whl mb n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ mb mb s eta ninstalling collected packages pip n attempting uninstall pip n found existing installation pip n uninstalling pip n successfully uninstalled pip nsuccessfully installed pip nfiles removed nlinking tool python npython npip from opt buildpack tools python lib site packages pip python ninstalled usr local buildpack tools python sh in seconds ncollecting hashin n downloading hashin none any whl kb ncollecting pip api n downloading pip api none any whl kb n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ kb mb s eta ncollecting packaging n downloading packaging none any whl kb n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ kb mb s eta ncollecting pyparsing n downloading pyparsing none any whl kb n ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ kb mb s eta nrequirement already satisfied pip in opt buildpack tools python lib site packages from pip api hashin ninstalling collected packages pyparsing pip api packaging hashin nsuccessfully installed hashin packaging pip api pyparsing n stderr found file in cache tmp containerbase python focal tar xz n debug returning updated third party requirements txt branch renovate datasets x basebranch main debug pip requirements updateartifacts third party requirements in branch renovate datasets x basebranch main debug no hashin commands to run returning branch renovate datasets x basebranch main debug updated package files branch renovate datasets x basebranch main debug updated lock files branch renovate datasets x basebranch main updatedartifacts third party requirements txt debug getting comments for branch renovate datasets x basebranch main debug found comments branch renovate datasets x basebranch main debug isbranchconflicted main renovate datasets x branch renovate datasets x basebranch main debug using cached result false for isbranchconflicted main renovate datasets x branch renovate datasets x basebranch main debug file s to commit branch renovate datasets x basebranch main debug preparing files for committing to branch renovate datasets x branch renovate datasets x basebranch main debug setting git author name branch renovate datasets x basebranch main gitauthorname renovate bot debug setting git author email branch renovate datasets x basebranch main gitauthoremail bot renovateapp com debug git commit branch renovate datasets x basebranch main deletedfiles ignoredfiles result author null branch renovate datasets x commit root false summary changes insertions deletions debug resettocommit branch renovate datasets x basebranch main debug fetching branch renovate datasets x branch renovate datasets x basebranch main info branch updated branch renovate datasets x basebranch main commitsha debug updating renovate stability days status check state to green branch renovate datasets x basebranch main debug setting branch status branch renovate datasets x basebranch main context renovate stability days state green debug branch status pending branch renovate datasets x basebranch main commitsha debug getbranchpr renovate datasets x branch renovate datasets x basebranch main debug findpr renovate datasets x undefined open branch renovate datasets x basebranch main debug findpr renovate datasets x undefined closed branch renovate datasets x basebranch main debug branchexists false branch renovate datasets x basebranch main debug dependencydashboardcheck undefined branch renovate datasets x basebranch main debug recreateclosed is false branch renovate datasets x basebranch main debug findpr renovate datasets x build update dependency datasets to open branch renovate datasets x basebranch main debug pralreadyexisted false branch renovate datasets x basebranch main debug checking schedule after every weekday before every weekday america chihuahua branch renovate datasets x basebranch main debug found timezone branch renovate datasets x basebranch main timezone america chihuahua debug adjusting now for timezone branch renovate datasets x basebranch main debug checking schedule s branch renovate datasets x basebranch main debug checking schedule after every weekday branch renovate datasets x basebranch main parsedschedule schedules t a d exceptions error debug checking schedule before every weekday branch renovate datasets x basebranch main parsedschedule schedules t b d exceptions error debug package not scheduled branch renovate datasets x basebranch main debug skipping branch creation as not within schedule branch renovate datasets x basebranch main debug getbranchpr renovate datasets x debug findpr renovate datasets x undefined open debug found pr debug ensuring dependency dashboard debug ensureissue dependency dashboard debug patching issue debug issue updated debug removing any stale branches debug config repoisonboarded true debug branch lists branchlist renovate datasets x renovate datasets x renovatebranches renovate datasets x debug remainingbranches debug no branches to clean up debug cleaning up renovate refs refs renovate debug packagefiles clear package files deleted basebranches main debug renovate repository pr statistics stats total open closed merged debug repository result done status onboarded enabled true onboarded true debug repository timing splits milliseconds splits init extract lookup onboarding update total debug http statistics urls post get get get post patch post get patch get get post get get hoststats api github com requestcount requestavgms queueavgms pypi org requestcount requestavgms queueavgms totalrequests info repository finished durationms have you created a minimal reproduction repository i have linked to a minimal reproduction repository in the bug description | 1 |
180,393 | 6,649,277,897 | IssuesEvent | 2017-09-28 12:41:23 | herbiehp/unicenta | https://api.github.com/repos/herbiehp/unicenta | closed | Modify Customer Form | 3-Medium Priority enhancement help wanted | Modify customer form to add additional attributes that are check boxes (like visible) and add columns to the customers table to store them.
To make it generic and usable by any business the names displayed in the GUI (and field names) should be something like CustAttr1, CustAttr2, etc with the ability to change the name displayed in the GUI with a Resources file.
A use case is for an organization that offers memberships to its customers.
Example
A country club that runs golf, curling and tennis operations.
It would be useful for them to track customer memberships and the types in uniCenta.
Some types could be:
- Golf Membership - Full
- Golf Membership - Weekdays
- Curling Membership - Full
- Curling Membership - 1 night per week
- Tennis Membership - Full
I am willing to take this on but don't know where to start on this yet so I may need some help.
Thx,
TekMason
| 1.0 | Modify Customer Form - Modify customer form to add additional attributes that are check boxes (like visible) and add columns to the customers table to store them.
To make it generic and usable by any business the names displayed in the GUI (and field names) should be something like CustAttr1, CustAttr2, etc with the ability to change the name displayed in the GUI with a Resources file.
A use case is for an organization that offers memberships to its customers.
Example
A country club that runs golf, curling and tennis operations.
It would be useful for them to track customer memberships and the types in uniCenta.
Some types could be:
- Golf Membership - Full
- Golf Membership - Weekdays
- Curling Membership - Full
- Curling Membership - 1 night per week
- Tennis Membership - Full
I am willing to take this on but don't know where to start on this yet so I may need some help.
Thx,
TekMason
| priority | modify customer form modify customer form to add additional attributes that are check boxes like visible and add columns to the customers table to store them to make it generic and usable by any business the names displayed in the gui and field names should be something like etc with the ability to change the name displayed in the gui with a resources file a use case is for an organization that offers memberships to its customers example a country club that runs golf curling and tennis operations it would be useful for them to track customer memberships and the types in unicenta some types could be golf membership full golf membership weekdays curling membership full curling membership night per week tennis membership full i am willing to take this on but don t know where to start on this yet so i may need some help thx tekmason | 1 |
559,360 | 16,557,057,288 | IssuesEvent | 2021-05-28 15:02:32 | guardicore/monkey | https://api.github.com/repos/guardicore/monkey | closed | Configure MongoDB on Monkey Island initialization | Complexity: Medium Enhancement Priority: High python | Monkey Island needs to write runtime artifacts to a writable location and assume that the source code directory is read-only. Currently, MongoDB is started by the `linux/run.sh`, `windows\run_mongodb.bat`, and `appimage/run_appimage.sh`. These scripts do not have access to the `data_dir` property in `server_config.json`, and giving them access would add even more redundant functionality to all of these scripts.
- [x] Add a `"mongodb"` section to the `server_config.json` with a `"start_mongodb"` boolean. (0) - @VakarisZ
```json
{
"log_level": "debug",
"data_dir": "$HOME/.monkey_island",
"environment": {
"server_config": "password",
"deployment": "development",
"user": "my_user",
"password_hash": "DEADBEEF"
}
"mongodb" {
"start_mongodb": true
}
}
```
- Make sure this section exists and is correct for the default `server_config.json`, the docker `server_config.json`, the windows installer `server_config.json`, and the AppImage `server_config.json`
- [x] If `"start_mongodb": true`, then monkey_island.py will start the mongodb process and set it's `--dbpath` argument to `data_dir/db`. (0d) - @VakarisZ
- [x] Refactor the data dir folder creation and permission setting into a separate function and re-use it to create db dir with relevant permissions as well (0d).
- [x] Refactor mongodb initialization to be decoupled from environment config (0d) - @VakarisZ
- [x] Remove the code that starts mongodb from `appimage/run_appimage.sh`. If this script no longer serves any purpose, remove it and modify the AppImage to run without it. (0d) - @VakarisZ
- [x] Remove `linux/run.sh`. Modify documentation with new instructions on how to start monkey island. (0d) - @VakarisZ
- [x] Remove code that creates db dir from deployment scripts, mongodb install scripts, docs, etc. (0) - @VakarisZ
- [x] Remove the code that starts mongodb from `monkey_island/windows/*`. If any of these scripts no longer serves a purpose, remove them.(0d) - @VakarisZ
- [x] Ensure the mongodb process is properly cleaned up when Monkey Island exits (0d)
- [x] Review PR and fix (0d) | 1.0 | Configure MongoDB on Monkey Island initialization - Monkey Island needs to write runtime artifacts to a writable location and assume that the source code directory is read-only. Currently, MongoDB is started by the `linux/run.sh`, `windows\run_mongodb.bat`, and `appimage/run_appimage.sh`. These scripts do not have access to the `data_dir` property in `server_config.json`, and giving them access would add even more redundant functionality to all of these scripts.
- [x] Add a `"mongodb"` section to the `server_config.json` with a `"start_mongodb"` boolean. (0) - @VakarisZ
```json
{
"log_level": "debug",
"data_dir": "$HOME/.monkey_island",
"environment": {
"server_config": "password",
"deployment": "development",
"user": "my_user",
"password_hash": "DEADBEEF"
}
"mongodb" {
"start_mongodb": true
}
}
```
- Make sure this section exists and is correct for the default `server_config.json`, the docker `server_config.json`, the windows installer `server_config.json`, and the AppImage `server_config.json`
- [x] If `"start_mongodb": true`, then monkey_island.py will start the mongodb process and set it's `--dbpath` argument to `data_dir/db`. (0d) - @VakarisZ
- [x] Refactor the data dir folder creation and permission setting into a separate function and re-use it to create db dir with relevant permissions as well (0d).
- [x] Refactor mongodb initialization to be decoupled from environment config (0d) - @VakarisZ
- [x] Remove the code that starts mongodb from `appimage/run_appimage.sh`. If this script no longer serves any purpose, remove it and modify the AppImage to run without it. (0d) - @VakarisZ
- [x] Remove `linux/run.sh`. Modify documentation with new instructions on how to start monkey island. (0d) - @VakarisZ
- [x] Remove code that creates db dir from deployment scripts, mongodb install scripts, docs, etc. (0) - @VakarisZ
- [x] Remove the code that starts mongodb from `monkey_island/windows/*`. If any of these scripts no longer serves a purpose, remove them.(0d) - @VakarisZ
- [x] Ensure the mongodb process is properly cleaned up when Monkey Island exits (0d)
- [x] Review PR and fix (0d) | priority | configure mongodb on monkey island initialization monkey island needs to write runtime artifacts to a writable location and assume that the source code directory is read only currently mongodb is started by the linux run sh windows run mongodb bat and appimage run appimage sh these scripts do not have access to the data dir property in server config json and giving them access would add even more redundant functionality to all of these scripts add a mongodb section to the server config json with a start mongodb boolean vakarisz json log level debug data dir home monkey island environment server config password deployment development user my user password hash deadbeef mongodb start mongodb true make sure this section exists and is correct for the default server config json the docker server config json the windows installer server config json and the appimage server config json if start mongodb true then monkey island py will start the mongodb process and set it s dbpath argument to data dir db vakarisz refactor the data dir folder creation and permission setting into a separate function and re use it to create db dir with relevant permissions as well refactor mongodb initialization to be decoupled from environment config vakarisz remove the code that starts mongodb from appimage run appimage sh if this script no longer serves any purpose remove it and modify the appimage to run without it vakarisz remove linux run sh modify documentation with new instructions on how to start monkey island vakarisz remove code that creates db dir from deployment scripts mongodb install scripts docs etc vakarisz remove the code that starts mongodb from monkey island windows if any of these scripts no longer serves a purpose remove them vakarisz ensure the mongodb process is properly cleaned up when monkey island exits review pr and fix | 1 |
247,239 | 7,915,576,371 | IssuesEvent | 2018-07-04 00:08:31 | facelessuser/pymdown-extensions | https://api.github.com/repos/facelessuser/pymdown-extensions | closed | SuperFences: Preserve Tabs and \r | Bug Priority - Medium Severity - Major | The preserve tabs feature is useful for preserving the tab character in code blocks, but it makes fences get processed before whitespace normalization. This means before `\r\n` is transformed to `\n`. On a Windows system, this can cause an issue as the content is scanned assuming normalization. We should strip trailing '\r' from lines to be safe. | 1.0 | SuperFences: Preserve Tabs and \r - The preserve tabs feature is useful for preserving the tab character in code blocks, but it makes fences get processed before whitespace normalization. This means before `\r\n` is transformed to `\n`. On a Windows system, this can cause an issue as the content is scanned assuming normalization. We should strip trailing '\r' from lines to be safe. | priority | superfences preserve tabs and r the preserve tabs feature is useful for preserving the tab character in code blocks but it makes fences get processed before whitespace normalization this means before r n is transformed to n on a windows system this can cause an issue as the content is scanned assuming normalization we should strip trailing r from lines to be safe | 1 |
476,723 | 13,749,106,737 | IssuesEvent | 2020-10-06 10:00:03 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | closed | User group invitation tab invited message not showing | bug hacktoberfest priority: medium | **Describe the bug**
When you go into the user profile groups tab and under that Invitation you will not show the invitation message. E.g https://learndash/members/mateo/groups/invites
**To Reproduce**
Steps to reproduce the behavior:
1. Invite member into your group with some invitation text.
2. Now login into the user you have invited.
3. Go to https://learndash/members/<member>/groups/invites
4. You will not show the invitation message. https://prnt.sc/unnnl7
**Expected behavior**
We should show the invitation message. https://prnt.sc/unno80
**Screenshots**
https://prnt.sc/unnnl7
**Support ticket links**
If applicable, add HelpScout link or ticket number where the issue was originally reported.
| 1.0 | User group invitation tab invited message not showing - **Describe the bug**
When you go into the user profile groups tab and under that Invitation you will not show the invitation message. E.g https://learndash/members/mateo/groups/invites
**To Reproduce**
Steps to reproduce the behavior:
1. Invite member into your group with some invitation text.
2. Now login into the user you have invited.
3. Go to https://learndash/members/<member>/groups/invites
4. You will not show the invitation message. https://prnt.sc/unnnl7
**Expected behavior**
We should show the invitation message. https://prnt.sc/unno80
**Screenshots**
https://prnt.sc/unnnl7
**Support ticket links**
If applicable, add HelpScout link or ticket number where the issue was originally reported.
| priority | user group invitation tab invited message not showing describe the bug when you go into the user profile groups tab and under that invitation you will not show the invitation message e g to reproduce steps to reproduce the behavior invite member into your group with some invitation text now login into the user you have invited go to you will not show the invitation message expected behavior we should show the invitation message screenshots support ticket links if applicable add helpscout link or ticket number where the issue was originally reported | 1 |
289,431 | 8,870,684,399 | IssuesEvent | 2019-01-11 10:13:44 | georchestra/georchestra | https://api.github.com/repos/georchestra/georchestra | closed | CAS - Relicates of config.jar in the JSP files | 2018 bug priority-medium | **Bug description**
We can find some traces of the shared config here:
https://github.com/georchestra/georchestra/blob/master/cas-server-webapp/src/main/webapp/WEB-INF/view/jsp/default/ui/includes/top.jsp#L38-L42
**geOrchestra version or branch**
since 18.06, also present on master currently
**Expected behavior**
We should get rid of these shared.* variables previously defined in the old config.jar process. Maybe the concerned variables should end in the default.properties.
| 1.0 | CAS - Relicates of config.jar in the JSP files - **Bug description**
We can find some traces of the shared config here:
https://github.com/georchestra/georchestra/blob/master/cas-server-webapp/src/main/webapp/WEB-INF/view/jsp/default/ui/includes/top.jsp#L38-L42
**geOrchestra version or branch**
since 18.06, also present on master currently
**Expected behavior**
We should get rid of these shared.* variables previously defined in the old config.jar process. Maybe the concerned variables should end in the default.properties.
| priority | cas relicates of config jar in the jsp files bug description we can find some traces of the shared config here georchestra version or branch since also present on master currently expected behavior we should get rid of these shared variables previously defined in the old config jar process maybe the concerned variables should end in the default properties | 1 |
100,528 | 4,097,848,842 | IssuesEvent | 2016-06-03 04:41:54 | Putaitu/mondai | https://api.github.com/repos/Putaitu/mondai | closed | Reloading resources doesn't work immediately | estimate:1h priority:medium type:bug version:0.1.3 | This appears to be a GitHub issue, that there is an extended output cache on the API. Is there anything that can be done to circumvent it? | 1.0 | Reloading resources doesn't work immediately - This appears to be a GitHub issue, that there is an extended output cache on the API. Is there anything that can be done to circumvent it? | priority | reloading resources doesn t work immediately this appears to be a github issue that there is an extended output cache on the api is there anything that can be done to circumvent it | 1 |
438,820 | 12,651,991,474 | IssuesEvent | 2020-06-17 02:09:13 | minio/minio | https://api.github.com/repos/minio/minio | closed | "Invalid argument" errors due to setrlimit calls on macOS | community priority: medium | Hello!
When using MinIO on macOS, we frequently see log lines that look like the following:
```
API: SYSTEM()
Time: <time>
Error: invalid argument
1: /Users/user/go/pkg/mod/github.com/minio/minio@v0.0.0-20180508161510-54cd29b51c38/cmd/gateway-main.go:166:cmd.StartGateway()
```
These are ignorable and everything seems to work, but still annoying. After a bunch of debugging, we tracked this down to how MinIO is trying to increase rlimits in https://github.com/minio/minio/blob/release/cmd/server-rlimit.go#L25-L55, and in https://github.com/minio/minio/tree/release/pkg/sys
There is an open Go bug about how the Go syscall for getting the max limit is lying. https://github.com/golang/go/issues/30401 This changed between Go releases on macOS, and we believe MinIO is hitting this bug.
Would it be useful to push up a code review to fix this? I think the solution is to extend github.com/minio/minio/pkg/sys to have special darwin-specific handling for not letting the max limit above 10240, as described in the Go issue.
## Expected Behavior
No inscrutable logs about invalid arguments while MinIO starts and tries to set limits.
## Current Behavior
MinIO logs "invalid argument" on macOS
## Possible Solution
Extend github.com/minio/minio/pkg/sys to specifically not allow the max limit above 10240, as described in the Go issue about this (Go's syscall behavior on Darwin changed): https://github.com/golang/go/issues/30401
## Steps to Reproduce (for bugs)
1. Run on macOS, look at logs
## Regression
This is a problem introduced by changing Go standard libraries.
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Version used (`minio version`): latest
* Environment name and version (e.g. nginx 1.9.1): Go 1.14
* Server type and version: Mac Mini
* Operating System and version (`uname -a`): macOS High Sierra
* Link to your project: https://www.storj.io :wave:
Thanks so much!
| 1.0 | "Invalid argument" errors due to setrlimit calls on macOS - Hello!
When using MinIO on macOS, we frequently see log lines that look like the following:
```
API: SYSTEM()
Time: <time>
Error: invalid argument
1: /Users/user/go/pkg/mod/github.com/minio/minio@v0.0.0-20180508161510-54cd29b51c38/cmd/gateway-main.go:166:cmd.StartGateway()
```
These are ignorable and everything seems to work, but still annoying. After a bunch of debugging, we tracked this down to how MinIO is trying to increase rlimits in https://github.com/minio/minio/blob/release/cmd/server-rlimit.go#L25-L55, and in https://github.com/minio/minio/tree/release/pkg/sys
There is an open Go bug about how the Go syscall for getting the max limit is lying. https://github.com/golang/go/issues/30401 This changed between Go releases on macOS, and we believe MinIO is hitting this bug.
Would it be useful to push up a code review to fix this? I think the solution is to extend github.com/minio/minio/pkg/sys to have special darwin-specific handling for not letting the max limit above 10240, as described in the Go issue.
## Expected Behavior
No inscrutable logs about invalid arguments while MinIO starts and tries to set limits.
## Current Behavior
MinIO logs "invalid argument" on macOS
## Possible Solution
Extend github.com/minio/minio/pkg/sys to specifically not allow the max limit above 10240, as described in the Go issue about this (Go's syscall behavior on Darwin changed): https://github.com/golang/go/issues/30401
## Steps to Reproduce (for bugs)
1. Run on macOS, look at logs
## Regression
This is a problem introduced by changing Go standard libraries.
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Version used (`minio version`): latest
* Environment name and version (e.g. nginx 1.9.1): Go 1.14
* Server type and version: Mac Mini
* Operating System and version (`uname -a`): macOS High Sierra
* Link to your project: https://www.storj.io :wave:
Thanks so much!
| priority | invalid argument errors due to setrlimit calls on macos hello when using minio on macos we frequently see log lines that look like the following api system time error invalid argument users user go pkg mod github com minio minio cmd gateway main go cmd startgateway these are ignorable and everything seems to work but still annoying after a bunch of debugging we tracked this down to how minio is trying to increase rlimits in and in there is an open go bug about how the go syscall for getting the max limit is lying this changed between go releases on macos and we believe minio is hitting this bug would it be useful to push up a code review to fix this i think the solution is to extend github com minio minio pkg sys to have special darwin specific handling for not letting the max limit above as described in the go issue expected behavior no inscrutable logs about invalid arguments while minio starts and tries to set limits current behavior minio logs invalid argument on macos possible solution extend github com minio minio pkg sys to specifically not allow the max limit above as described in the go issue about this go s syscall behavior on darwin changed steps to reproduce for bugs run on macos look at logs regression this is a problem introduced by changing go standard libraries your environment version used minio version latest environment name and version e g nginx go server type and version mac mini operating system and version uname a macos high sierra link to your project wave thanks so much | 1 |
533,130 | 15,577,446,931 | IssuesEvent | 2021-03-17 13:33:01 | Proof-Of-Humanity/proof-of-humanity-web | https://api.github.com/repos/Proof-Of-Humanity/proof-of-humanity-web | opened | Submit Profile Page - Reword Primary Document Button Text | priority: medium status: available type: enhancement :sparkles: | Instead of "primary document", we should say registry rules or something like that. | 1.0 | Submit Profile Page - Reword Primary Document Button Text - Instead of "primary document", we should say registry rules or something like that. | priority | submit profile page reword primary document button text instead of primary document we should say registry rules or something like that | 1 |
368,912 | 10,886,082,145 | IssuesEvent | 2019-11-18 11:46:39 | canonical-web-and-design/vanilla-framework | https://api.github.com/repos/canonical-web-and-design/vanilla-framework | closed | Contextual menu (left/center) doesn't grow to fit item's size | Priority: Medium |
Contextual menu pattern is defined to have a with between 10rem - 21rem and should adapt to contents (menu items) width.
This only works for default (right aligned) menu:
<img width="631" alt="screen shot 2018-11-19 at 17 10 16" src="https://user-images.githubusercontent.com/83575/48719886-cb5f6200-ec1e-11e8-8c66-6c8baacff434.png">
Left aligned or centred menus don't grow to fit size of the content and always stay at min-width of 10rem:
<img width="642" alt="screen shot 2018-11-19 at 17 13 27" src="https://user-images.githubusercontent.com/83575/48719898-d31f0680-ec1e-11e8-834f-6cb4bc06d187.png">
<img width="646" alt="screen shot 2018-11-19 at 17 10 36" src="https://user-images.githubusercontent.com/83575/48719900-d31f0680-ec1e-11e8-9692-8b5d706e6f83.png">
Found in snapcraft.io:
https://github.com/canonical-websites/snapcraft.io/issues/1332
**To Reproduce**
Steps to reproduce the behavior:
1. Use `p-contextual-menu--left` or `p-contextual-menu--center`
2. Have a menu item longer then 10 rem
3. See that menu doesn't grow to fit the item.
**Desktop:**
- OS: Mac OS
- Browser: Chrome, Firefox
- Version: latest
| 1.0 | Contextual menu (left/center) doesn't grow to fit item's size -
Contextual menu pattern is defined to have a with between 10rem - 21rem and should adapt to contents (menu items) width.
This only works for default (right aligned) menu:
<img width="631" alt="screen shot 2018-11-19 at 17 10 16" src="https://user-images.githubusercontent.com/83575/48719886-cb5f6200-ec1e-11e8-8c66-6c8baacff434.png">
Left aligned or centred menus don't grow to fit size of the content and always stay at min-width of 10rem:
<img width="642" alt="screen shot 2018-11-19 at 17 13 27" src="https://user-images.githubusercontent.com/83575/48719898-d31f0680-ec1e-11e8-834f-6cb4bc06d187.png">
<img width="646" alt="screen shot 2018-11-19 at 17 10 36" src="https://user-images.githubusercontent.com/83575/48719900-d31f0680-ec1e-11e8-9692-8b5d706e6f83.png">
Found in snapcraft.io:
https://github.com/canonical-websites/snapcraft.io/issues/1332
**To Reproduce**
Steps to reproduce the behavior:
1. Use `p-contextual-menu--left` or `p-contextual-menu--center`
2. Have a menu item longer then 10 rem
3. See that menu doesn't grow to fit the item.
**Desktop:**
- OS: Mac OS
- Browser: Chrome, Firefox
- Version: latest
| priority | contextual menu left center doesn t grow to fit item s size contextual menu pattern is defined to have a with between and should adapt to contents menu items width this only works for default right aligned menu img width alt screen shot at src left aligned or centred menus don t grow to fit size of the content and always stay at min width of img width alt screen shot at src img width alt screen shot at src found in snapcraft io to reproduce steps to reproduce the behavior use p contextual menu left or p contextual menu center have a menu item longer then rem see that menu doesn t grow to fit the item desktop os mac os browser chrome firefox version latest | 1 |
804,612 | 29,495,147,921 | IssuesEvent | 2023-06-02 16:17:54 | SolarWindss/Hearthstone.js | https://api.github.com/repos/SolarWindss/Hearthstone.js | closed | Move `validateCard` from interact to functions | priority: medium time: short improvement | This doesn't belong in interact
- [x] Move
- [x] Update src & tests | 1.0 | Move `validateCard` from interact to functions - This doesn't belong in interact
- [x] Move
- [x] Update src & tests | priority | move validatecard from interact to functions this doesn t belong in interact move update src tests | 1 |
99,940 | 4,074,675,549 | IssuesEvent | 2016-05-28 16:29:15 | BugBusterSWE/documentation | https://api.github.com/repos/BugBusterSWE/documentation | reopened | Aggiungere automazione APIDoc | priority:medium | *Documento in cui si trova il problema*:
Norme di Progetto
Activity #494
*Descrizione del problema*:
Aggiungere automazione APIDoc
Link task: [https://bugbusters.teamwork.com/tasks/6938623](https://bugbusters.teamwork.com/tasks/6938623) | 1.0 | Aggiungere automazione APIDoc - *Documento in cui si trova il problema*:
Norme di Progetto
Activity #494
*Descrizione del problema*:
Aggiungere automazione APIDoc
Link task: [https://bugbusters.teamwork.com/tasks/6938623](https://bugbusters.teamwork.com/tasks/6938623) | priority | aggiungere automazione apidoc documento in cui si trova il problema norme di progetto activity descrizione del problema aggiungere automazione apidoc link task | 1 |
481,788 | 13,891,889,024 | IssuesEvent | 2020-10-19 11:22:33 | sunpy/sunpy | https://api.github.com/repos/sunpy/sunpy | closed | Maps using the CD matrix are not correctly modified by resample | Bug(?) Close? Effort High Package Intermediate Priority Medium map | resample only changes the CDELT flags not CD if present.
| 1.0 | Maps using the CD matrix are not correctly modified by resample - resample only changes the CDELT flags not CD if present.
| priority | maps using the cd matrix are not correctly modified by resample resample only changes the cdelt flags not cd if present | 1 |
747,678 | 26,095,245,726 | IssuesEvent | 2022-12-26 18:18:08 | canaltin-byte/SWE573-SDP-Can | https://api.github.com/repos/canaltin-byte/SWE573-SDP-Can | closed | Home Page Name and surname | enhancement priority : Low Front-end Effort: Medium Home Page | User id should not be accessible. User Name and Surname should be there | 1.0 | Home Page Name and surname - User id should not be accessible. User Name and Surname should be there | priority | home page name and surname user id should not be accessible user name and surname should be there | 1 |
696,650 | 23,909,701,924 | IssuesEvent | 2022-09-09 06:54:44 | dmwm/CRABServer | https://api.github.com/repos/dmwm/CRABServer | opened | If extraJDL contains DESIRED_Sites, then submission fails after 25 tries | Priority: Medium | I submitted the task `220907_161045:dmapelli_crab_20220907_181041` that contained an `extraJDL` similar to
```python
config.Debug.extraJDL = [ '+DESIRED_Sites = "T1_DE_KIT,T2_US_Purdue,T3_UK_London_QMUL"' ]
```
and the submission failed with the error
```plaintext
Failure message from server: The CRAB server backend was not able to submit the jobs to the Grid schedulers. This could be a temporary glitch. Please try again later. If the error persists send an e-mail to cmstalk+computing-tools@dovecotmta.cern.ch. The submission was retried 25 times on 5 schedulers. These are the failures per Grid scheduler:
{'crab3@vocms059.cern.ch': ["Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cer
[... message truncated to the first 1000 chars ...]
```
Looking at the log `/data/container/TaskWorker/logs/tasks/dmapelli/220907_161045:dmapelli_crab_20220907_181041.log` (on `crab-preprod-tw01`) I noticed that the TW tried submitting the task from 18:11 to 18:39, and gave up after 25 tries on 5 different schedd.
I know that I made a mistake and that I should have used the `whiteList` instead, but maybe we can add a simple check that when we find `DESIRED_Sites` in the extraJDL we fail the task submission straightaway. | 1.0 | If extraJDL contains DESIRED_Sites, then submission fails after 25 tries - I submitted the task `220907_161045:dmapelli_crab_20220907_181041` that contained an `extraJDL` similar to
```python
config.Debug.extraJDL = [ '+DESIRED_Sites = "T1_DE_KIT,T2_US_Purdue,T3_UK_London_QMUL"' ]
```
and the submission failed with the error
```plaintext
Failure message from server: The CRAB server backend was not able to submit the jobs to the Grid schedulers. This could be a temporary glitch. Please try again later. If the error persists send an e-mail to cmstalk+computing-tools@dovecotmta.cern.ch. The submission was retried 25 times on 5 schedulers. These are the failures per Grid scheduler:
{'crab3@vocms059.cern.ch': ["Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cern.ch . Task: 220907_161045:dmapelli_crab_20220907_181041;\n'Error when performing action on the schedd.'", "Failed to submit task to: crab3@vocms059.cer
[... message truncated to the first 1000 chars ...]
```
Looking at the log `/data/container/TaskWorker/logs/tasks/dmapelli/220907_161045:dmapelli_crab_20220907_181041.log` (on `crab-preprod-tw01`) I noticed that the TW tried submitting the task from 18:11 to 18:39, and gave up after 25 tries on 5 different schedd.
I know that I made a mistake and that I should have used the `whiteList` instead, but maybe we can add a simple check that when we find `DESIRED_Sites` in the extraJDL we fail the task submission straightaway. | priority | if extrajdl contains desired sites then submission fails after tries i submitted the task dmapelli crab that contained an extrajdl similar to python config debug extrajdl and the submission failed with the error plaintext failure message from server the crab server backend was not able to submit the jobs to the grid schedulers this could be a temporary glitch please try again later if the error persists send an e mail to cmstalk computing tools dovecotmta cern ch the submission was retried times on schedulers these are the failures per grid scheduler cern ch failed to submit task to cern ch task dmapelli crab n error when performing action on the schedd failed to submit task to cern ch task dmapelli crab n error when performing action on the schedd failed to submit task to cern ch task dmapelli crab n error when performing action on the schedd failed to submit task to cern ch task dmapelli crab n error when performing action on the schedd failed to submit task to cer looking at the log data container taskworker logs tasks dmapelli dmapelli crab log on crab preprod i noticed that the tw tried submitting the task from to and gave up after tries on different schedd i know that i made a mistake and that i should have used the whitelist instead but maybe we can add a simple check that when we find desired sites in the extrajdl we fail the task submission straightaway | 1 |
556,151 | 16,476,186,850 | IssuesEvent | 2021-05-24 05:47:30 | Repair-DeskPOS/RepairDesk-BUGS-IMPROVEMENTS | https://api.github.com/repos/Repair-DeskPOS/RepairDesk-BUGS-IMPROVEMENTS | closed | iPAD - Deposit Feature | Cellular Surgeons Medium Priority enhancement iPAD POS Register | **Suggestion # 1**
I've created a product under iPhone 7 >> deposit / bench fee which is not refundable. Lets say customer comes to our store and using the iPAD I check in this customer >> tap on checkout and collect the deposit amount. This works fine
However when I open the device to find device issue and edit the ticket with new device issue, as soon I click on checkout it says invoice is already paid.
Just like web platform it would be nice to edit and existing ticket that has been completely paid.
**Suggestion # 2**
Also once I ring the item it would be cool to display paid amount along with balance on the right hand side.
**Brooks Johnson | Cellular Surgeon**
| 1.0 | iPAD - Deposit Feature - **Suggestion # 1**
I've created a product under iPhone 7 >> deposit / bench fee which is not refundable. Lets say customer comes to our store and using the iPAD I check in this customer >> tap on checkout and collect the deposit amount. This works fine
However when I open the device to find device issue and edit the ticket with new device issue, as soon I click on checkout it says invoice is already paid.
Just like web platform it would be nice to edit and existing ticket that has been completely paid.
**Suggestion # 2**
Also once I ring the item it would be cool to display paid amount along with balance on the right hand side.
**Brooks Johnson | Cellular Surgeon**
| priority | ipad deposit feature suggestion i ve created a product under iphone deposit bench fee which is not refundable lets say customer comes to our store and using the ipad i check in this customer tap on checkout and collect the deposit amount this works fine however when i open the device to find device issue and edit the ticket with new device issue as soon i click on checkout it says invoice is already paid just like web platform it would be nice to edit and existing ticket that has been completely paid suggestion also once i ring the item it would be cool to display paid amount along with balance on the right hand side brooks johnson cellular surgeon | 1 |
252,702 | 8,039,294,016 | IssuesEvent | 2018-07-30 17:55:36 | systers/communities | https://api.github.com/repos/systers/communities | closed | Set up the repository with basic angular files | Category: Coding Difficulty: MEDIUM Priority: HIGH Program: GSoC Type: Enhancement | ## Description
As a user,
I need set up the repository,
so that I can restart the project with angular framework
## Acceptance Criteria
- Basic working Angular App
- Basic files and modules established
### Update [Required]
- README
- Create multiple new files
## Definition of Done
- [ ] All of the required items are completed.
- [ ] Approval by 1 mentor.
## Estimation
1 hour
Can I be assigned this issue @Tharangi @divyanshu-rawat @Janiceilene @MeepyMay ? | 1.0 | Set up the repository with basic angular files - ## Description
As a user,
I need set up the repository,
so that I can restart the project with angular framework
## Acceptance Criteria
- Basic working Angular App
- Basic files and modules established
### Update [Required]
- README
- Create multiple new files
## Definition of Done
- [ ] All of the required items are completed.
- [ ] Approval by 1 mentor.
## Estimation
1 hour
Can I be assigned this issue @Tharangi @divyanshu-rawat @Janiceilene @MeepyMay ? | priority | set up the repository with basic angular files description as a user i need set up the repository so that i can restart the project with angular framework acceptance criteria basic working angular app basic files and modules established update readme create multiple new files definition of done all of the required items are completed approval by mentor estimation hour can i be assigned this issue tharangi divyanshu rawat janiceilene meepymay | 1 |
68,233 | 3,285,102,990 | IssuesEvent | 2015-10-28 19:07:44 | pantheon-systems/WordPress | https://api.github.com/repos/pantheon-systems/WordPress | closed | Allow for cache TTL of zero in cache plugin | priority:medium | In some instances such as development users dont want to have the minimum TTL of 600s. | 1.0 | Allow for cache TTL of zero in cache plugin - In some instances such as development users dont want to have the minimum TTL of 600s. | priority | allow for cache ttl of zero in cache plugin in some instances such as development users dont want to have the minimum ttl of | 1 |
725,678 | 24,971,367,564 | IssuesEvent | 2022-11-02 01:42:23 | aws-samples/aws-last-mile-delivery-hyperlocal | https://api.github.com/repos/aws-samples/aws-last-mile-delivery-hyperlocal | closed | Setup yarn dependency checks for all packages | enhancement dependencies priority:medium component:all effort:low | * first candidate would be [depcheck](https://github.com/depcheck/depcheck).
* to make sure that there are no missing dependencies in any packages | 1.0 | Setup yarn dependency checks for all packages - * first candidate would be [depcheck](https://github.com/depcheck/depcheck).
* to make sure that there are no missing dependencies in any packages | priority | setup yarn dependency checks for all packages first candidate would be to make sure that there are no missing dependencies in any packages | 1 |
94,280 | 3,923,997,027 | IssuesEvent | 2016-04-22 13:46:00 | EnvironmentAgency/pafs-user | https://api.github.com/repos/EnvironmentAgency/pafs-user | opened | CR:: User Requirements::Funding Sources:: Change the word 'expected' | Change Priority - 3 Medium Sprint 3 | **Issue:**
The word 'expected' caused participants to assume the funding types listed on this page, represent funding that is yet to be secured.

For example, if you have already secured some private funding. This was not seen as required to be entered here as it is not expected -you have already secured it.
**Recommendation**
To Be Discussed | 1.0 | CR:: User Requirements::Funding Sources:: Change the word 'expected' - **Issue:**
The word 'expected' caused participants to assume the funding types listed on this page, represent funding that is yet to be secured.

For example, if you have already secured some private funding. This was not seen as required to be entered here as it is not expected -you have already secured it.
**Recommendation**
To Be Discussed | priority | cr user requirements funding sources change the word expected issue the word expected caused participants to assume the funding types listed on this page represent funding that is yet to be secured for example if you have already secured some private funding this was not seen as required to be entered here as it is not expected you have already secured it recommendation to be discussed | 1 |
445,052 | 12,825,477,672 | IssuesEvent | 2020-07-06 15:01:28 | radcampaign/allaboveall2020 | https://api.github.com/repos/radcampaign/allaboveall2020 | closed | Tweak menu font size/icons | Priority: Medium | More QC from Allyson: she'd like the main menu to stand out a bit more. In comparing with the comps, I know the font size is dead on to what Kien provided. Possibly we could add more weight? How much bigger could the menu font go? I know there's a challenge to get the current items to fit.
Maybe the social and search icons could decrease in size a notch or two?
Or could we reduce the whitespace between the logo and the start of the black bar and menu?
Again, no rush, just logging.
| 1.0 | Tweak menu font size/icons - More QC from Allyson: she'd like the main menu to stand out a bit more. In comparing with the comps, I know the font size is dead on to what Kien provided. Possibly we could add more weight? How much bigger could the menu font go? I know there's a challenge to get the current items to fit.
Maybe the social and search icons could decrease in size a notch or two?
Or could we reduce the whitespace between the logo and the start of the black bar and menu?
Again, no rush, just logging.
| priority | tweak menu font size icons more qc from allyson she d like the main menu to stand out a bit more in comparing with the comps i know the font size is dead on to what kien provided possibly we could add more weight how much bigger could the menu font go i know there s a challenge to get the current items to fit maybe the social and search icons could decrease in size a notch or two or could we reduce the whitespace between the logo and the start of the black bar and menu again no rush just logging | 1 |
471,331 | 13,565,237,770 | IssuesEvent | 2020-09-18 11:21:56 | DXHeroes/dx-scanner | https://api.github.com/repos/DXHeroes/dx-scanner | opened | Create a language detector for Swift (~100 new lines of code) | Difficulty: Easy Priority: Medium Status: Available Type: Enhancement good first issue | We'd like to measure the quality of apps written in Swift.
There are "language detectors" to determine what language is on which path.
It's necessary to implement **SwiftLanguageDetector** the same way as is [JavaScriptLanguageDetector](https://github.com/DXHeroes/dx-scanner/blob/master/src/detectors/JavaScript/JavaScriptLanguageDetector.ts) implemented. It's around 100 new lines of code including tests.
More info in [our Contribution guide](https://github.com/DXHeroes/dx-scanner/blob/master/CONTRIBUTING.md). | 1.0 | Create a language detector for Swift (~100 new lines of code) - We'd like to measure the quality of apps written in Swift.
There are "language detectors" to determine what language is on which path.
It's necessary to implement **SwiftLanguageDetector** the same way as is [JavaScriptLanguageDetector](https://github.com/DXHeroes/dx-scanner/blob/master/src/detectors/JavaScript/JavaScriptLanguageDetector.ts) implemented. It's around 100 new lines of code including tests.
More info in [our Contribution guide](https://github.com/DXHeroes/dx-scanner/blob/master/CONTRIBUTING.md). | priority | create a language detector for swift new lines of code we d like to measure the quality of apps written in swift there are language detectors to determine what language is on which path it s necessary to implement swiftlanguagedetector the same way as is implemented it s around new lines of code including tests more info in | 1 |
650,780 | 21,416,982,859 | IssuesEvent | 2022-04-22 11:54:59 | sahar-avsh/SWE-599 | https://api.github.com/repos/sahar-avsh/SWE-599 | closed | Q&A - Creating a question UI | enhancement Show stopper Hard medium priority Q&A | For creating question:
- [x] A **form** shall open
- [x] User shall enter **title**
- [x] User shall be able to enter **description**
- [x] User shall be able to **attach** her/his **decks**
- [x] User shall be able to **attach** her/his **resources** | 1.0 | Q&A - Creating a question UI - For creating question:
- [x] A **form** shall open
- [x] User shall enter **title**
- [x] User shall be able to enter **description**
- [x] User shall be able to **attach** her/his **decks**
- [x] User shall be able to **attach** her/his **resources** | priority | q a creating a question ui for creating question a form shall open user shall enter title user shall be able to enter description user shall be able to attach her his decks user shall be able to attach her his resources | 1 |
224,225 | 7,467,858,395 | IssuesEvent | 2018-04-02 16:52:53 | enforcer574/smashclub | https://api.github.com/repos/enforcer574/smashclub | opened | Photo Slideshow on Home Page | Complexity: Medium Priority: 3 - Medium Type: User Request | Replace the static image on the "about" portlet of the home page with a timed slideshow. Admins can upload photos to a directory in the assets folder and they will be displayed in the rotation. A setting in the site_settings table controls the slideshow speed. | 1.0 | Photo Slideshow on Home Page - Replace the static image on the "about" portlet of the home page with a timed slideshow. Admins can upload photos to a directory in the assets folder and they will be displayed in the rotation. A setting in the site_settings table controls the slideshow speed. | priority | photo slideshow on home page replace the static image on the about portlet of the home page with a timed slideshow admins can upload photos to a directory in the assets folder and they will be displayed in the rotation a setting in the site settings table controls the slideshow speed | 1 |
406,501 | 11,894,049,716 | IssuesEvent | 2020-03-29 14:14:33 | robotframework/robotframework | https://api.github.com/repos/robotframework/robotframework | closed | Dynamic API: Add new `get_keyword_source` method | enhancement priority: medium rc 1 | This is needed to make it possible to add source information to Libdoc spec files (#3507) as well as to Libdoc's model objects (#3448). External tools like editors can then use this information to implement "go to definition" functionality.
This method needs to be able to return both the path to the source file and the line number, and we need to decide which format to use. Good alternatives include a tuple `(source, lineno)` as well as a string `source:lineno`. Not sure why, but I like the latter approach better. | 1.0 | Dynamic API: Add new `get_keyword_source` method - This is needed to make it possible to add source information to Libdoc spec files (#3507) as well as to Libdoc's model objects (#3448). External tools like editors can then use this information to implement "go to definition" functionality.
This method needs to be able to return both the path to the source file and the line number, and we need to decide which format to use. Good alternatives include a tuple `(source, lineno)` as well as a string `source:lineno`. Not sure why, but I like the latter approach better. | priority | dynamic api add new get keyword source method this is needed to make it possible to add source information to libdoc spec files as well as to libdoc s model objects external tools like editors can then use this information to implement go to definition functionality this method needs to be able to return both the path to the source file and the line number and we need to decide which format to use good alternatives include a tuple source lineno as well as a string source lineno not sure why but i like the latter approach better | 1 |
707,107 | 24,295,583,272 | IssuesEvent | 2022-09-29 09:46:52 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | Upgrade versions in .tool-versions | type:feature help wanted priority-3-medium new package manager status:in-progress | **What would you like Renovate to be able to do?**
When upgrading node, ruby or the like, it should also change the version in `.tool-versions`, used by [asdf](https://github.com/asdf-vm/asdf).
**Describe the solution you'd like**
See above.
**Describe alternatives you've considered**
Doing it manually.
**Additional context**
I'd imagine other version managers have similar files. Maybe renovate already supports them, but if not, it'd be great if it did as well.
| 1.0 | Upgrade versions in .tool-versions - **What would you like Renovate to be able to do?**
When upgrading node, ruby or the like, it should also change the version in `.tool-versions`, used by [asdf](https://github.com/asdf-vm/asdf).
**Describe the solution you'd like**
See above.
**Describe alternatives you've considered**
Doing it manually.
**Additional context**
I'd imagine other version managers have similar files. Maybe renovate already supports them, but if not, it'd be great if it did as well.
| priority | upgrade versions in tool versions what would you like renovate to be able to do when upgrading node ruby or the like it should also change the version in tool versions used by describe the solution you d like see above describe alternatives you ve considered doing it manually additional context i d imagine other version managers have similar files maybe renovate already supports them but if not it d be great if it did as well | 1 |
457,584 | 13,158,552,952 | IssuesEvent | 2020-08-10 14:30:06 | canonical-web-and-design/jaas-dashboard | https://api.github.com/repos/canonical-web-and-design/jaas-dashboard | opened | Leader information not shown in unit list | Model Details Priority: Medium | When a unit is in HA and a leader has been elected we should indicate as such.
I'd expect to see the leader status in the unit list but also in the unit details. @ziheliu214 can you add this to the unit list/details designs please.
related: https://discourse.juju.is/t/leadership-in-juju-operations-perspective/3406 | 1.0 | Leader information not shown in unit list - When a unit is in HA and a leader has been elected we should indicate as such.
I'd expect to see the leader status in the unit list but also in the unit details. @ziheliu214 can you add this to the unit list/details designs please.
related: https://discourse.juju.is/t/leadership-in-juju-operations-perspective/3406 | priority | leader information not shown in unit list when a unit is in ha and a leader has been elected we should indicate as such i d expect to see the leader status in the unit list but also in the unit details can you add this to the unit list details designs please related | 1 |
375,306 | 11,102,411,831 | IssuesEvent | 2019-12-16 23:58:12 | AugurProject/augur | https://api.github.com/repos/AugurProject/augur | closed | ZeroXTrade does not enforce ERC1155_PROXY_ID in AssetData | Priority: Medium V2 Audit | While the off-chain calls to ZeroXTrade (createZeroXOrderFor) will encode the ERC1155_PROXY_ID (`0xa7cb5fb7`) into the first four bytes of the order that is submitted to 0x.fillOrder, there is no check inside the actual trade function to ensure that is what is being passed along to the 0x Exchange. Alternate PROXY_IDs could behave differently and might even look in different places for actual asset being transferred, which could lead to unexpected fill amounts being returned, and potentially even unexpected trades occurring (although the tokenAddress being immediately after proxy_id is likely to limit severity). New proxies could be added in the future with unexpected functionality. | 1.0 | ZeroXTrade does not enforce ERC1155_PROXY_ID in AssetData - While the off-chain calls to ZeroXTrade (createZeroXOrderFor) will encode the ERC1155_PROXY_ID (`0xa7cb5fb7`) into the first four bytes of the order that is submitted to 0x.fillOrder, there is no check inside the actual trade function to ensure that is what is being passed along to the 0x Exchange. Alternate PROXY_IDs could behave differently and might even look in different places for actual asset being transferred, which could lead to unexpected fill amounts being returned, and potentially even unexpected trades occurring (although the tokenAddress being immediately after proxy_id is likely to limit severity). New proxies could be added in the future with unexpected functionality. | priority | zeroxtrade does not enforce proxy id in assetdata while the off chain calls to zeroxtrade createzeroxorderfor will encode the proxy id into the first four bytes of the order that is submitted to fillorder there is no check inside the actual trade function to ensure that is what is being passed along to the exchange alternate proxy ids could behave differently and might even look in different places for actual asset being transferred which could lead to unexpected fill amounts being returned and potentially even unexpected trades occurring although the tokenaddress being immediately after proxy id is likely to limit severity new proxies could be added in the future with unexpected functionality | 1 |
204,028 | 7,079,438,127 | IssuesEvent | 2018-01-10 09:34:02 | Automattic/liveblog | https://api.github.com/repos/Automattic/liveblog | closed | Remove reliance on GET params for single entry ajax requests | Priority::Medium enhancement | Scenario:
* A user loads a liveblog
* There are several key events shown in the key events widget
* Some of the key events have not loaded in the initial set of entries that are loaded by default
* The user clicks on one of the entries that isn't visible
* The plugin fires off an ajax request [appending an `index` GET param to the endpoint URL](https://github.com/Automattic/liveblog/blob/e50c8dad22e7519f42a0045d5b564d3e60d9c5e4/js/liveblog-lazyloader.js#L82-L85)
* The single entry is returned and displayed
See where the [ajax action is set up](https://github.com/Automattic/liveblog/blob/f82842828cad0c7b921459b10a6e534f9cbe5355/liveblog.php#L299) and the [request is processed](https://github.com/Automattic/liveblog/blob/f82842828cad0c7b921459b10a6e534f9cbe5355/liveblog.php#L519-L575).
We should refactor this to append an ID or some such to the endpoint URL, rather than a GET param, to avoid bypassing caching mechanisms like Batcache. | 1.0 | Remove reliance on GET params for single entry ajax requests - Scenario:
* A user loads a liveblog
* There are several key events shown in the key events widget
* Some of the key events have not loaded in the initial set of entries that are loaded by default
* The user clicks on one of the entries that isn't visible
* The plugin fires off an ajax request [appending an `index` GET param to the endpoint URL](https://github.com/Automattic/liveblog/blob/e50c8dad22e7519f42a0045d5b564d3e60d9c5e4/js/liveblog-lazyloader.js#L82-L85)
* The single entry is returned and displayed
See where the [ajax action is set up](https://github.com/Automattic/liveblog/blob/f82842828cad0c7b921459b10a6e534f9cbe5355/liveblog.php#L299) and the [request is processed](https://github.com/Automattic/liveblog/blob/f82842828cad0c7b921459b10a6e534f9cbe5355/liveblog.php#L519-L575).
We should refactor this to append an ID or some such to the endpoint URL, rather than a GET param, to avoid bypassing caching mechanisms like Batcache. | priority | remove reliance on get params for single entry ajax requests scenario a user loads a liveblog there are several key events shown in the key events widget some of the key events have not loaded in the initial set of entries that are loaded by default the user clicks on one of the entries that isn t visible the plugin fires off an ajax request the single entry is returned and displayed see where the and the we should refactor this to append an id or some such to the endpoint url rather than a get param to avoid bypassing caching mechanisms like batcache | 1 |
828,734 | 31,840,844,496 | IssuesEvent | 2023-09-14 16:12:04 | bcgov/foi-flow | https://api.github.com/repos/bcgov/foi-flow | closed | Single Source of Truth - Divisions | Task dev medium priority | Title of ticket:
#### Description
This task has been created to Remove existing `Program Area Divisions ` Tables and related dependencies like `PROGRAM AREAS ?? ` inside DOC Reviewer DB , in order to keep single source of truth of Divisions across FOI FLOW app. At present DOC REVIEWER Web App (only?) using this table(s)
#### Dependencies
Are there any dependencies?
#### DOD
- [x] Analyze the repercussions of removing Divisions, Program areas from Doc Reviewer DB - On initial analysis, its only used for Left panel on Doc Reviewer Web app, so Idea is to Create a new endpoint for Divisions by Program area from FOI FLOW
- [x] Once above Analyze DOD is over and correct, Develop an API endpoint or use/check existing APIs for Divisions
- [x] Update Left Panel Front End component to use FOI FLOW API endpoint for divisions.
- [x] Integration test
- [ ]
| 1.0 | Single Source of Truth - Divisions - Title of ticket:
#### Description
This task has been created to Remove existing `Program Area Divisions ` Tables and related dependencies like `PROGRAM AREAS ?? ` inside DOC Reviewer DB , in order to keep single source of truth of Divisions across FOI FLOW app. At present DOC REVIEWER Web App (only?) using this table(s)
#### Dependencies
Are there any dependencies?
#### DOD
- [x] Analyze the repercussions of removing Divisions, Program areas from Doc Reviewer DB - On initial analysis, its only used for Left panel on Doc Reviewer Web app, so Idea is to Create a new endpoint for Divisions by Program area from FOI FLOW
- [x] Once above Analyze DOD is over and correct, Develop an API endpoint or use/check existing APIs for Divisions
- [x] Update Left Panel Front End component to use FOI FLOW API endpoint for divisions.
- [x] Integration test
- [ ]
| priority | single source of truth divisions title of ticket description this task has been created to remove existing program area divisions tables and related dependencies like program areas inside doc reviewer db in order to keep single source of truth of divisions across foi flow app at present doc reviewer web app only using this table s dependencies are there any dependencies dod analyze the repercussions of removing divisions program areas from doc reviewer db on initial analysis its only used for left panel on doc reviewer web app so idea is to create a new endpoint for divisions by program area from foi flow once above analyze dod is over and correct develop an api endpoint or use check existing apis for divisions update left panel front end component to use foi flow api endpoint for divisions integration test | 1 |
706,036 | 24,258,019,046 | IssuesEvent | 2022-09-27 19:37:18 | gammapy/gammapy | https://api.github.com/repos/gammapy/gammapy | closed | Gammapy validation: HESS DL3 DR1 | effort-medium package-novice priority-high | As part of the Gammapy validation effort while preparing v1.0, we should script some of the H.E.S.S. data level 3, data release 1 analyses for validation.
Note that large parts of Gammapy were rewritten since Gammapy v0.12 (May 2019) which was used for https://arxiv.org/pdf/1910.08088.pdf requiring the validation exercise to be re-done with the latest version of Gammapy, and from now on moving forward we'll maintain the validation scripts in the Gammapy team to avoid regressions. One concrete case of a regression that we already noticed and fixed in the meantime is https://github.com/gammapy/gammapy/pull/2367 .
A task description and references are here:
- https://github.com/gammapy/gammapy-benchmarks
- https://github.com/gammapy/gammapy-benchmarks/blob/master/validation
- https://github.com/gammapy/gammapy-benchmarks/tree/master/validation/hess-dl3-dr1
We're looking for help! If you can contribute, please leave a comment here, or contact me on Slack.
For this, you don't have to be a Gammapy developer, what's required is some Python & Gammapy & IACT analysis experience, and at least ~ 2 full days to contribute in Nov 2019, although ~ 1 week in Nov & Dec 2019 is more realistic if you're willing to script the analysis for all targets.
| 1.0 | Gammapy validation: HESS DL3 DR1 - As part of the Gammapy validation effort while preparing v1.0, we should script some of the H.E.S.S. data level 3, data release 1 analyses for validation.
Note that large parts of Gammapy were rewritten since Gammapy v0.12 (May 2019) which was used for https://arxiv.org/pdf/1910.08088.pdf requiring the validation exercise to be re-done with the latest version of Gammapy, and from now on moving forward we'll maintain the validation scripts in the Gammapy team to avoid regressions. One concrete case of a regression that we already noticed and fixed in the meantime is https://github.com/gammapy/gammapy/pull/2367 .
A task description and references are here:
- https://github.com/gammapy/gammapy-benchmarks
- https://github.com/gammapy/gammapy-benchmarks/blob/master/validation
- https://github.com/gammapy/gammapy-benchmarks/tree/master/validation/hess-dl3-dr1
We're looking for help! If you can contribute, please leave a comment here, or contact me on Slack.
For this, you don't have to be a Gammapy developer, what's required is some Python & Gammapy & IACT analysis experience, and at least ~ 2 full days to contribute in Nov 2019, although ~ 1 week in Nov & Dec 2019 is more realistic if you're willing to script the analysis for all targets.
| priority | gammapy validation hess as part of the gammapy validation effort while preparing we should script some of the h e s s data level data release analyses for validation note that large parts of gammapy were rewritten since gammapy may which was used for requiring the validation exercise to be re done with the latest version of gammapy and from now on moving forward we ll maintain the validation scripts in the gammapy team to avoid regressions one concrete case of a regression that we already noticed and fixed in the meantime is a task description and references are here we re looking for help if you can contribute please leave a comment here or contact me on slack for this you don t have to be a gammapy developer what s required is some python gammapy iact analysis experience and at least full days to contribute in nov although week in nov dec is more realistic if you re willing to script the analysis for all targets | 1 |
236,074 | 7,745,930,033 | IssuesEvent | 2018-05-29 19:58:15 | strapi/strapi | https://api.github.com/repos/strapi/strapi | closed | Overwriting GraphQL schema | priority: medium type: bug 🐛 | <!--
⚠️ If you do not respect these two points, your issue will be closed.
- Don't forget the checkbox at the end of your issue.
- Respect this template.
-->
<!-- ⚠️ Before writing your issue make sure you are using:-->
<!-- Node 9.x.x -->
<!-- npm 5.x.x -->
<!-- The latest version of Strapi. -->
**Informations**
- **Node.js version**: 9.11.1
- **npm version**: 5.6.0
- **Strapi version**: 3.0.0-alpha.12.2
- **Database**: mongo
- **Operating system**: macOS
**What is the current behavior?**
I'm trying to overwrite the GraphQL query from
```graphql
type Query {
practice(id: String): Practice
type: {
Practice: false
},
resolver: {
Query: {
practice: {
description: 'Return the practice of the authenticated user',
resolver: 'Practice.findMine'
},
practices: false
}
}
}
```
to
```graphql
type Query {
practice: Practice
type: {
Practice: false
},
resolver: {
Query: {
practice: {
description: 'Return the practice of the authenticated user',
resolver: 'Practice.findMine'
},
practices: false
}
}
}
```
But running strapi will error in `Error: Resolver Query.practice must be object or function`
**Steps to reproduce the problem**
Try to overwrite the default `findOne` route of any model like the example above
**What is the expected behavior?**
It should overwrite the schema without a problem, even the generated code looks correct
```graphql
# generated schema file
type Query {
patient(id: String!): Patient
patients(sort: String, limit: Int, start: Int, where: JSON): [Patient]
practice: Practice # this looks correct, it doesn't have the id: String, argument
```
**Suggested solutions**
I don't know any, I'm trying to find a solution myself without success
---
<!-- ⚠️ Make sure to browse the opened and closed issues. -->
- [x] I'm sure that this feature hasn't already been referenced.
<!--
⬆️ Don't delete this checkbox from your issue and approve it.
Add an `x` -> [x] or click on the checkbox when your issue is created.
-->
| 1.0 | Overwriting GraphQL schema - <!--
⚠️ If you do not respect these two points, your issue will be closed.
- Don't forget the checkbox at the end of your issue.
- Respect this template.
-->
<!-- ⚠️ Before writing your issue make sure you are using:-->
<!-- Node 9.x.x -->
<!-- npm 5.x.x -->
<!-- The latest version of Strapi. -->
**Informations**
- **Node.js version**: 9.11.1
- **npm version**: 5.6.0
- **Strapi version**: 3.0.0-alpha.12.2
- **Database**: mongo
- **Operating system**: macOS
**What is the current behavior?**
I'm trying to overwrite the GraphQL query from
```graphql
type Query {
practice(id: String): Practice
type: {
Practice: false
},
resolver: {
Query: {
practice: {
description: 'Return the practice of the authenticated user',
resolver: 'Practice.findMine'
},
practices: false
}
}
}
```
to
```graphql
type Query {
practice: Practice
type: {
Practice: false
},
resolver: {
Query: {
practice: {
description: 'Return the practice of the authenticated user',
resolver: 'Practice.findMine'
},
practices: false
}
}
}
```
But running strapi will error in `Error: Resolver Query.practice must be object or function`
**Steps to reproduce the problem**
Try to overwrite the default `findOne` route of any model like the example above
**What is the expected behavior?**
It should overwrite the schema without a problem, even the generated code looks correct
```graphql
# generated schema file
type Query {
patient(id: String!): Patient
patients(sort: String, limit: Int, start: Int, where: JSON): [Patient]
practice: Practice # this looks correct, it doesn't have the id: String, argument
```
**Suggested solutions**
I don't know any, I'm trying to find a solution myself without success
---
<!-- ⚠️ Make sure to browse the opened and closed issues. -->
- [x] I'm sure that this feature hasn't already been referenced.
<!--
⬆️ Don't delete this checkbox from your issue and approve it.
Add an `x` -> [x] or click on the checkbox when your issue is created.
-->
| priority | overwriting graphql schema ⚠️ if you do not respect these two points your issue will be closed don t forget the checkbox at the end of your issue respect this template informations node js version npm version strapi version alpha database mongo operating system macos what is the current behavior i m trying to overwrite the graphql query from graphql type query practice id string practice type practice false resolver query practice description return the practice of the authenticated user resolver practice findmine practices false to graphql type query practice practice type practice false resolver query practice description return the practice of the authenticated user resolver practice findmine practices false but running strapi will error in error resolver query practice must be object or function steps to reproduce the problem try to overwrite the default findone route of any model like the example above what is the expected behavior it should overwrite the schema without a problem even the generated code looks correct graphql generated schema file type query patient id string patient patients sort string limit int start int where json practice practice this looks correct it doesn t have the id string argument suggested solutions i don t know any i m trying to find a solution myself without success i m sure that this feature hasn t already been referenced ⬆️ don t delete this checkbox from your issue and approve it add an x or click on the checkbox when your issue is created | 1 |
803,510 | 29,180,221,691 | IssuesEvent | 2023-05-19 11:14:37 | Vatsim-Scandinavia/controlcenter | https://api.github.com/repos/Vatsim-Scandinavia/controlcenter | closed | Visiting Controllers can't book positions | back-end bug priority: medium | We missed them out in the logic checks. Users who have an active visiting endorsement should also be able to book. | 1.0 | Visiting Controllers can't book positions - We missed them out in the logic checks. Users who have an active visiting endorsement should also be able to book. | priority | visiting controllers can t book positions we missed them out in the logic checks users who have an active visiting endorsement should also be able to book | 1 |
39,837 | 2,860,368,169 | IssuesEvent | 2015-06-03 15:32:09 | CUL-DigitalServices/grasshopper-ui | https://api.github.com/repos/CUL-DigitalServices/grasshopper-ui | opened | Switching from batch edit to calendar view binds event handlers again | Medium Priority To Fix | Instead of binding them all over again, binding them once when switching to the view initially should be enough. | 1.0 | Switching from batch edit to calendar view binds event handlers again - Instead of binding them all over again, binding them once when switching to the view initially should be enough. | priority | switching from batch edit to calendar view binds event handlers again instead of binding them all over again binding them once when switching to the view initially should be enough | 1 |
781,917 | 27,454,426,190 | IssuesEvent | 2023-03-02 20:03:42 | phetsims/scenery | https://api.github.com/repos/phetsims/scenery | closed | Moving to phet.{{X}} namespaces instead of multiple globals | dev:enhancement priority:3-medium status:ready-for-review dev:chip-away | During the namespace discussion, it was noted that there are a number of advantages to move to a single global variable (`phet`). The consistency will help with things like https://github.com/phetsims/scenery/issues/965, and it will override fewer names.
Those seem like good arguments to me (that somewhat override the verbosity concerns, since you can always create local variables if things are too verbose).
This seems like a few core changes (changing the available namespacing), and then a lot of refactoring to point everything to the new namespace (phetCore/axon/dot/kite/scenery at a minimum).
I'm curious if there are objections to this approach | 1.0 | Moving to phet.{{X}} namespaces instead of multiple globals - During the namespace discussion, it was noted that there are a number of advantages to move to a single global variable (`phet`). The consistency will help with things like https://github.com/phetsims/scenery/issues/965, and it will override fewer names.
Those seem like good arguments to me (that somewhat override the verbosity concerns, since you can always create local variables if things are too verbose).
This seems like a few core changes (changing the available namespacing), and then a lot of refactoring to point everything to the new namespace (phetCore/axon/dot/kite/scenery at a minimum).
I'm curious if there are objections to this approach | priority | moving to phet x namespaces instead of multiple globals during the namespace discussion it was noted that there are a number of advantages to move to a single global variable phet the consistency will help with things like and it will override fewer names those seem like good arguments to me that somewhat override the verbosity concerns since you can always create local variables if things are too verbose this seems like a few core changes changing the available namespacing and then a lot of refactoring to point everything to the new namespace phetcore axon dot kite scenery at a minimum i m curious if there are objections to this approach | 1 |
546,041 | 15,982,997,889 | IssuesEvent | 2021-04-18 07:10:25 | r-lib/styler | https://api.github.com/repos/r-lib/styler | closed | Indenting leading spaces is inconsistent? | Complexity: Medium Priority: High Status: Unassigned Type: Bug | Hi, I noticed that when there is a certain number of leading spaces, the line does not indent to the correct number. For example, with `indent_by = 2` (though this issue appears for more than just `2` spaces),
### 4 leading spaces (for example):
```R
function() {
print("hi")
}
```
does not indent to 2 spaces. However,
### 17 leading spaces (for example):
```R
function() {
print("hi")
}
```
does indent to 2. Can anyone reproduce this, or does anyone know why this is?
### System info
styler: 1.3.2
R: 4.0.4 | 1.0 | Indenting leading spaces is inconsistent? - Hi, I noticed that when there is a certain number of leading spaces, the line does not indent to the correct number. For example, with `indent_by = 2` (though this issue appears for more than just `2` spaces),
### 4 leading spaces (for example):
```R
function() {
print("hi")
}
```
does not indent to 2 spaces. However,
### 17 leading spaces (for example):
```R
function() {
print("hi")
}
```
does indent to 2. Can anyone reproduce this, or does anyone know why this is?
### System info
styler: 1.3.2
R: 4.0.4 | priority | indenting leading spaces is inconsistent hi i noticed that when there is a certain number of leading spaces the line does not indent to the correct number for example with indent by though this issue appears for more than just spaces leading spaces for example r function print hi does not indent to spaces however leading spaces for example r function print hi does indent to can anyone reproduce this or does anyone know why this is system info styler r | 1 |
116,944 | 4,709,096,787 | IssuesEvent | 2016-10-14 03:23:24 | honestbleeps/Reddit-Enhancement-Suite | https://api.github.com/repos/honestbleeps/Reddit-Enhancement-Suite | closed | Add a class to notifications with module id and notification ID | Difficulty-2_Medium Difficulty-1_Easy hacktoberfest Priority-4_Some Interest RE-Enhancement | This might be a duplicate
https://www.reddit.com/r/Enhancement/comments/4c9c5f/feature_request_add_a_class_or_id_to_the/ | 1.0 | Add a class to notifications with module id and notification ID - This might be a duplicate
https://www.reddit.com/r/Enhancement/comments/4c9c5f/feature_request_add_a_class_or_id_to_the/ | priority | add a class to notifications with module id and notification id this might be a duplicate | 1 |
809,125 | 30,175,401,407 | IssuesEvent | 2023-07-04 03:48:36 | Laserwolve-Games/DaggerQuest | https://api.github.com/repos/Laserwolve-Games/DaggerQuest | opened | Spawning Loot Occasionally Instead Rotates Other Loot | Bug Hard Difficulty Medium Priority | This is very hard to reproduce. Occasionally when spawning loot, no loot is a spawned, and a previously spawned piece of loot changes it's direction (animation). Need to find out if the loot that enemies drop also can cause this issue. | 1.0 | Spawning Loot Occasionally Instead Rotates Other Loot - This is very hard to reproduce. Occasionally when spawning loot, no loot is a spawned, and a previously spawned piece of loot changes it's direction (animation). Need to find out if the loot that enemies drop also can cause this issue. | priority | spawning loot occasionally instead rotates other loot this is very hard to reproduce occasionally when spawning loot no loot is a spawned and a previously spawned piece of loot changes it s direction animation need to find out if the loot that enemies drop also can cause this issue | 1 |
199,741 | 6,993,824,286 | IssuesEvent | 2017-12-15 13:04:19 | emory-libraries/ezpaarse-platforms | https://api.github.com/repos/emory-libraries/ezpaarse-platforms | closed | Policy Map | Additional Parser Medium Priority review | ### Example:star::star: :
www.policymap.com
### Priority:
Medium
### Subscriber (Library):
Woodruff
| 1.0 | Policy Map - ### Example:star::star: :
www.policymap.com
### Priority:
Medium
### Subscriber (Library):
Woodruff
| priority | policy map example star star priority medium subscriber library woodruff | 1 |
523,321 | 15,178,210,461 | IssuesEvent | 2021-02-14 14:33:19 | Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2 | https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2 | opened | Adapt Religious Clothes | :art: 3D graphics :triangular_ruler: :grey_exclamation: priority medium :question: suggestion :question: | <!--
DO NOT REMOVE PRE-EXISTING LINES
IF YOU WANT TO SUGGEST A FEW THINGS, OPEN A NEW ISSUE PER EVERY SUGGESTION
----------------------------------------------------------------------------------------------------------
-->
**Describe your suggestion in full detail below:**
We should adapt religious clothes making sure every religion will have them.
I started to write their triggers...
<details>
<summary>Click to expand</summary>

</details>
...but it turned out a too big project, so here I'm.
If the religion can't get any special clothes, it should be using Catholic clothes instead. | 1.0 | Adapt Religious Clothes - <!--
DO NOT REMOVE PRE-EXISTING LINES
IF YOU WANT TO SUGGEST A FEW THINGS, OPEN A NEW ISSUE PER EVERY SUGGESTION
----------------------------------------------------------------------------------------------------------
-->
**Describe your suggestion in full detail below:**
We should adapt religious clothes making sure every religion will have them.
I started to write their triggers...
<details>
<summary>Click to expand</summary>

</details>
...but it turned out a too big project, so here I'm.
If the religion can't get any special clothes, it should be using Catholic clothes instead. | priority | adapt religious clothes do not remove pre existing lines if you want to suggest a few things open a new issue per every suggestion describe your suggestion in full detail below we should adapt religious clothes making sure every religion will have them i started to write their triggers click to expand but it turned out a too big project so here i m if the religion can t get any special clothes it should be using catholic clothes instead | 1 |
30,501 | 2,723,981,498 | IssuesEvent | 2015-04-14 15:27:21 | CruxFramework/crux-widgets | https://api.github.com/repos/CruxFramework/crux-widgets | closed | FileProtocolIteratorFactory pass file relative names for its filter, when scanning for resources | bug CruxScannotation imported Milestone-3.0.1 Priority-Medium | _From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on October 08, 2010 11:49:50_
FileProtocolIteratorFactory pass file relative names for its filter, when scanning for resources
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=204_ | 1.0 | FileProtocolIteratorFactory pass file relative names for its filter, when scanning for resources - _From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on October 08, 2010 11:49:50_
FileProtocolIteratorFactory pass file relative names for its filter, when scanning for resources
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=204_ | priority | fileprotocoliteratorfactory pass file relative names for its filter when scanning for resources from on october fileprotocoliteratorfactory pass file relative names for its filter when scanning for resources original issue | 1 |
466,771 | 13,433,366,639 | IssuesEvent | 2020-09-07 09:43:44 | enso-org/ide | https://api.github.com/repos/enso-org/ide | closed | Lexer: Implement parsing of literals. | Category: IDE Change: Non-Breaking Difficulty: Core Contributor Priority: Medium Status: Invalid Type: Enhancement | ### Summary
Reimplement parsing of literals from scala in rust.
### Value
Literals are another primitive of our AST.
### Specification
Reimplement parsing of literals from scala in rust.
### Acceptance Criteria & Test
Implement and test the specified task above.
| 1.0 | Lexer: Implement parsing of literals. - ### Summary
Reimplement parsing of literals from scala in rust.
### Value
Literals are another primitive of our AST.
### Specification
Reimplement parsing of literals from scala in rust.
### Acceptance Criteria & Test
Implement and test the specified task above.
| priority | lexer implement parsing of literals summary reimplement parsing of literals from scala in rust value literals are another primitive of our ast specification reimplement parsing of literals from scala in rust acceptance criteria test implement and test the specified task above | 1 |
612,417 | 19,012,081,779 | IssuesEvent | 2021-11-23 10:24:27 | containrrr/watchtower | https://api.github.com/repos/containrrr/watchtower | closed | alibaba Cloud private warehouse cannot pull the image and automatically upgrade | Type: Bug Priority: Medium Status: Available | my docker-compose.yml
```yaml
watchtower:
container_name: watchtower
image: containrrr/watchtower
restart: always
environment:
- TZ=Asia/Shanghai
privileged: true
depends_on:
- mj-admin-service
- mj-admin-portal
volumes:
- /var/run/docker.sock:/var/run/docker.sock
- /root/.docker/config.json:/config.json
command: mj-admin-service --cleanup --interval 10
logging:
driver: "json-file"
options:
max-size: "1m"
max-file: "1"
```
the log shows that it is always ready, and there is no more log output:

| 1.0 | alibaba Cloud private warehouse cannot pull the image and automatically upgrade - my docker-compose.yml
```yaml
watchtower:
container_name: watchtower
image: containrrr/watchtower
restart: always
environment:
- TZ=Asia/Shanghai
privileged: true
depends_on:
- mj-admin-service
- mj-admin-portal
volumes:
- /var/run/docker.sock:/var/run/docker.sock
- /root/.docker/config.json:/config.json
command: mj-admin-service --cleanup --interval 10
logging:
driver: "json-file"
options:
max-size: "1m"
max-file: "1"
```
the log shows that it is always ready, and there is no more log output:

| priority | alibaba cloud private warehouse cannot pull the image and automatically upgrade my docker compose yml yaml watchtower container name watchtower image containrrr watchtower restart always environment tz asia shanghai privileged true depends on mj admin service mj admin portal volumes var run docker sock var run docker sock root docker config json config json command mj admin service cleanup interval logging driver json file options max size max file the log shows that it is always ready and there is no more log output: | 1 |
711,280 | 24,456,551,366 | IssuesEvent | 2022-10-07 07:20:09 | HiAvatar/backend | https://api.github.com/repos/HiAvatar/backend | closed | Project API 관련 예외처리 | Priority: Medium | ### Description
프로젝트와 관련된 예외 처리를 수행한다.
`@RestControllerAdvice`를 적용한 `ProjectControllerAdvice.class`에서 `@ExceptionHandler`를 사용하여 `ProjectController.class`의 API 로직으로부터 예외 처리 코드를 분리시키고, 필요한 예외 로직을 추가한다.
<br>
### Todo List
공통
- [ ] 존재하지 않는 프로젝트를 조회하는 요청이 왔을 때
- [ ] 특정 파일을 찾을 수 없을 때
음성 관련
- [ ] base64 -> file 변환에 실패했을 때
- [ ] file -> base64 변환에 실패했을 때
- [ ] 음성 파일 생성이 실패 했을 때
- [ ] 음성 파일 URL을 찾을 수 없을 때
영상 관련
- [ ] 영상 파일 생성이 실패 했을 때
<br>
### e.t.c (Optional)
(참조 및 링크 첨부하기)
| 1.0 | Project API 관련 예외처리 - ### Description
프로젝트와 관련된 예외 처리를 수행한다.
`@RestControllerAdvice`를 적용한 `ProjectControllerAdvice.class`에서 `@ExceptionHandler`를 사용하여 `ProjectController.class`의 API 로직으로부터 예외 처리 코드를 분리시키고, 필요한 예외 로직을 추가한다.
<br>
### Todo List
공통
- [ ] 존재하지 않는 프로젝트를 조회하는 요청이 왔을 때
- [ ] 특정 파일을 찾을 수 없을 때
음성 관련
- [ ] base64 -> file 변환에 실패했을 때
- [ ] file -> base64 변환에 실패했을 때
- [ ] 음성 파일 생성이 실패 했을 때
- [ ] 음성 파일 URL을 찾을 수 없을 때
영상 관련
- [ ] 영상 파일 생성이 실패 했을 때
<br>
### e.t.c (Optional)
(참조 및 링크 첨부하기)
| priority | project api 관련 예외처리 description 프로젝트와 관련된 예외 처리를 수행한다 restcontrolleradvice 를 적용한 projectcontrolleradvice class 에서 exceptionhandler 를 사용하여 projectcontroller class 의 api 로직으로부터 예외 처리 코드를 분리시키고 필요한 예외 로직을 추가한다 todo list 공통 존재하지 않는 프로젝트를 조회하는 요청이 왔을 때 특정 파일을 찾을 수 없을 때 음성 관련 file 변환에 실패했을 때 file 변환에 실패했을 때 음성 파일 생성이 실패 했을 때 음성 파일 url을 찾을 수 없을 때 영상 관련 영상 파일 생성이 실패 했을 때 e t c optional 참조 및 링크 첨부하기 | 1 |
752,736 | 26,323,118,746 | IssuesEvent | 2023-01-10 02:36:42 | LuanRT/YouTube.js | https://api.github.com/repos/LuanRT/YouTube.js | closed | V2.0.0 --- InnertubeError: Something went wrong at MusicDetailHeader! | bug good first issue Stale priority: medium | ### Steps to reproduce
I'm no sure why but calling `client.music.getAlbum();` with the id `"MPREb_Ml86XqtmqaJ"` crashes.
### Failure Logs
```shell
InnertubeError: Something went wrong at MusicDetailHeader!
This is a bug, please report it at https://github.com/LuanRT/YouTube.js/issues
at Function.formatError (/app/node_modules/youtubei.js/dist/src/parser/index.js:317:22)
at Function.parseItem (/app/node_modules/youtubei.js/dist/src/parser/index.js:268:22)
at Function.parse (/app/node_modules/youtubei.js/dist/src/parser/index.js:310:53)
at Function.parseResponse (/app/node_modules/youtubei.js/dist/src/parser/index.js:178:28)
at new Album (/app/node_modules/youtubei.js/dist/src/parser/ytmusic/Album.js:27:67)
at Music.<anonymous> (/app/node_modules/youtubei.js/dist/src/core/Music.js:121:20)
at Generator.next (<anonymous>)
at fulfilled (/app/node_modules/youtubei.js/dist/src/core/Music.js:5:58)
at processTicksAndRejections (node:internal/process/task_queues:96:5) {
info: {
stack: "TypeError: Cannot read properties of undefined (reading 'text')\n" +
' at new MusicDetailHeader (/app/node_modules/youtubei.js/dist/src/parser/classes/MusicDetailHeader.js:17:88)\n' +
' at Function.parseItem (/app/node_modules/youtubei.js/dist/src/parser/index.js:263:32)\n' +
' at Function.parse (/app/node_modules/youtubei.js/dist/src/parser/index.js:310:53)\n' +
' at Function.parseResponse (/app/node_modules/youtubei.js/dist/src/parser/index.js:178:28)\n' +
' at new Album (/app/node_modules/youtubei.js/dist/src/parser/ytmusic/Album.js:27:67)\n' +
' at Music.<anonymous> (/app/node_modules/youtubei.js/dist/src/core/Music.js:121:20)\n' +
' at Generator.next (<anonymous>)\n' +
' at fulfilled (/app/node_modules/youtubei.js/dist/src/core/Music.js:5:58)\n' +
' at processTicksAndRejections (node:internal/process/task_queues:96:5)'
},
date: 2022-10-06T17:47:24.863Z,
version: '2.0.0'
}
```
### Expected behavior
not crash
### Current behavior
crashes
### Version
Default
### Anything else?
_No response_
### Checklist
- [X] I am running the latest version.
- [X] I checked the documentation and found no answer.
- [X] I have searched the existing issues and made sure this is not a duplicate.
- [X] I have provided sufficient information. | 1.0 | V2.0.0 --- InnertubeError: Something went wrong at MusicDetailHeader! - ### Steps to reproduce
I'm no sure why but calling `client.music.getAlbum();` with the id `"MPREb_Ml86XqtmqaJ"` crashes.
### Failure Logs
```shell
InnertubeError: Something went wrong at MusicDetailHeader!
This is a bug, please report it at https://github.com/LuanRT/YouTube.js/issues
at Function.formatError (/app/node_modules/youtubei.js/dist/src/parser/index.js:317:22)
at Function.parseItem (/app/node_modules/youtubei.js/dist/src/parser/index.js:268:22)
at Function.parse (/app/node_modules/youtubei.js/dist/src/parser/index.js:310:53)
at Function.parseResponse (/app/node_modules/youtubei.js/dist/src/parser/index.js:178:28)
at new Album (/app/node_modules/youtubei.js/dist/src/parser/ytmusic/Album.js:27:67)
at Music.<anonymous> (/app/node_modules/youtubei.js/dist/src/core/Music.js:121:20)
at Generator.next (<anonymous>)
at fulfilled (/app/node_modules/youtubei.js/dist/src/core/Music.js:5:58)
at processTicksAndRejections (node:internal/process/task_queues:96:5) {
info: {
stack: "TypeError: Cannot read properties of undefined (reading 'text')\n" +
' at new MusicDetailHeader (/app/node_modules/youtubei.js/dist/src/parser/classes/MusicDetailHeader.js:17:88)\n' +
' at Function.parseItem (/app/node_modules/youtubei.js/dist/src/parser/index.js:263:32)\n' +
' at Function.parse (/app/node_modules/youtubei.js/dist/src/parser/index.js:310:53)\n' +
' at Function.parseResponse (/app/node_modules/youtubei.js/dist/src/parser/index.js:178:28)\n' +
' at new Album (/app/node_modules/youtubei.js/dist/src/parser/ytmusic/Album.js:27:67)\n' +
' at Music.<anonymous> (/app/node_modules/youtubei.js/dist/src/core/Music.js:121:20)\n' +
' at Generator.next (<anonymous>)\n' +
' at fulfilled (/app/node_modules/youtubei.js/dist/src/core/Music.js:5:58)\n' +
' at processTicksAndRejections (node:internal/process/task_queues:96:5)'
},
date: 2022-10-06T17:47:24.863Z,
version: '2.0.0'
}
```
### Expected behavior
not crash
### Current behavior
crashes
### Version
Default
### Anything else?
_No response_
### Checklist
- [X] I am running the latest version.
- [X] I checked the documentation and found no answer.
- [X] I have searched the existing issues and made sure this is not a duplicate.
- [X] I have provided sufficient information. | priority | innertubeerror something went wrong at musicdetailheader steps to reproduce i m no sure why but calling client music getalbum with the id mpreb crashes failure logs shell innertubeerror something went wrong at musicdetailheader this is a bug please report it at at function formaterror app node modules youtubei js dist src parser index js at function parseitem app node modules youtubei js dist src parser index js at function parse app node modules youtubei js dist src parser index js at function parseresponse app node modules youtubei js dist src parser index js at new album app node modules youtubei js dist src parser ytmusic album js at music app node modules youtubei js dist src core music js at generator next at fulfilled app node modules youtubei js dist src core music js at processticksandrejections node internal process task queues info stack typeerror cannot read properties of undefined reading text n at new musicdetailheader app node modules youtubei js dist src parser classes musicdetailheader js n at function parseitem app node modules youtubei js dist src parser index js n at function parse app node modules youtubei js dist src parser index js n at function parseresponse app node modules youtubei js dist src parser index js n at new album app node modules youtubei js dist src parser ytmusic album js n at music app node modules youtubei js dist src core music js n at generator next n at fulfilled app node modules youtubei js dist src core music js n at processticksandrejections node internal process task queues date version expected behavior not crash current behavior crashes version default anything else no response checklist i am running the latest version i checked the documentation and found no answer i have searched the existing issues and made sure this is not a duplicate i have provided sufficient information | 1 |
696,467 | 23,902,227,886 | IssuesEvent | 2022-09-08 19:56:34 | PrefectHQ/prefect | https://api.github.com/repos/PrefectHQ/prefect | closed | Display intended run date time on flow run page | enhancement from:slack ui v2 priority:medium | ### Prefect Version
2.x
### Describe the proposed behavior
Display/expose the intended flow run date time in the table on the flow run details page.
### Describe the current behavior
Currently only the Created and Updated date times are displayed on this table.
### Additional context
From community Slack:
When a Flow is running on a schedule and it runs late or is delayed for any reason, the date time recorded on the flow run instance in the Flow Run tab is the date time at which it was run not the actual intended time it was scheduled to run. Is there a way to find this information? I found a date time in the idempotency key in the flow instance information, however, i am not sure if this is its intended purpose | 1.0 | Display intended run date time on flow run page - ### Prefect Version
2.x
### Describe the proposed behavior
Display/expose the intended flow run date time in the table on the flow run details page.
### Describe the current behavior
Currently only the Created and Updated date times are displayed on this table.
### Additional context
From community Slack:
When a Flow is running on a schedule and it runs late or is delayed for any reason, the date time recorded on the flow run instance in the Flow Run tab is the date time at which it was run not the actual intended time it was scheduled to run. Is there a way to find this information? I found a date time in the idempotency key in the flow instance information, however, i am not sure if this is its intended purpose | priority | display intended run date time on flow run page prefect version x describe the proposed behavior display expose the intended flow run date time in the table on the flow run details page describe the current behavior currently only the created and updated date times are displayed on this table additional context from community slack when a flow is running on a schedule and it runs late or is delayed for any reason the date time recorded on the flow run instance in the flow run tab is the date time at which it was run not the actual intended time it was scheduled to run is there a way to find this information i found a date time in the idempotency key in the flow instance information however i am not sure if this is its intended purpose | 1 |
608,381 | 18,837,044,628 | IssuesEvent | 2021-11-11 03:05:10 | dhowe/AdNauseam | https://api.github.com/repos/dhowe/AdNauseam | closed | [3.11.5b1 ] Any changes in filter list activates "Activate the AdNauseam filter ?" warning | PRIORITY: Medium Bug | Whenever you disactivate/activate some filter in the filter list (for example, uBlock filters), after applying changes it displays: `Activate the AdNauseam filter ?` warning. When it should not.

| 1.0 | [3.11.5b1 ] Any changes in filter list activates "Activate the AdNauseam filter ?" warning - Whenever you disactivate/activate some filter in the filter list (for example, uBlock filters), after applying changes it displays: `Activate the AdNauseam filter ?` warning. When it should not.

| priority | any changes in filter list activates activate the adnauseam filter warning whenever you disactivate activate some filter in the filter list for example ublock filters after applying changes it displays activate the adnauseam filter warning when it should not | 1 |
217,944 | 7,329,325,442 | IssuesEvent | 2018-03-05 04:09:36 | tootsuite/mastodon | https://api.github.com/repos/tootsuite/mastodon | closed | Add badge for muted/blocked of user profile in last column | enhancement fit for beginners priority - medium ui | It would be great to add a badge (like the ```FOLLOWS YOU``` one) if the user is currently muted and/or blocked when displaying an user profile in the last column.
* * * *
- [x] I searched or browsed the repo’s other issues to ensure this is not a duplicate.
- [ ] This bug happens on a [tagged release](https://github.com/tootsuite/mastodon/releases) and not on `master` (If you're a user, don't worry about this).
| 1.0 | Add badge for muted/blocked of user profile in last column - It would be great to add a badge (like the ```FOLLOWS YOU``` one) if the user is currently muted and/or blocked when displaying an user profile in the last column.
* * * *
- [x] I searched or browsed the repo’s other issues to ensure this is not a duplicate.
- [ ] This bug happens on a [tagged release](https://github.com/tootsuite/mastodon/releases) and not on `master` (If you're a user, don't worry about this).
| priority | add badge for muted blocked of user profile in last column it would be great to add a badge like the follows you one if the user is currently muted and or blocked when displaying an user profile in the last column i searched or browsed the repo’s other issues to ensure this is not a duplicate this bug happens on a and not on master if you re a user don t worry about this | 1 |
617,735 | 19,403,360,892 | IssuesEvent | 2021-12-19 15:28:26 | ibissource/frank-flow | https://api.github.com/repos/ibissource/frank-flow | closed | Highlight the element in the canvas at which the user is editing in the Editor. | feature priority:medium work:medium | **Is your feature request related to a problem? Please describe.**
Clicking on an element shows the code. But clicking on the code doesn't show the element.
**Describe the solution you'd like**
Do the same but the other way around.
**Additional context**
Panning to that position would be even better, but may be confusing.
| 1.0 | Highlight the element in the canvas at which the user is editing in the Editor. - **Is your feature request related to a problem? Please describe.**
Clicking on an element shows the code. But clicking on the code doesn't show the element.
**Describe the solution you'd like**
Do the same but the other way around.
**Additional context**
Panning to that position would be even better, but may be confusing.
| priority | highlight the element in the canvas at which the user is editing in the editor is your feature request related to a problem please describe clicking on an element shows the code but clicking on the code doesn t show the element describe the solution you d like do the same but the other way around additional context panning to that position would be even better but may be confusing | 1 |
514,281 | 14,936,523,807 | IssuesEvent | 2021-01-25 13:30:15 | bounswe/bounswe2020group4 | https://api.github.com/repos/bounswe/bounswe2020group4 | opened | (BKND) Unit tests | Backend Coding Effort: Medium Priority: Medium | Unit tests will be added for email sending and report endpoints.
Deadline
25.01.2021 | 1.0 | (BKND) Unit tests - Unit tests will be added for email sending and report endpoints.
Deadline
25.01.2021 | priority | bknd unit tests unit tests will be added for email sending and report endpoints deadline | 1 |
806,033 | 29,797,987,007 | IssuesEvent | 2023-06-16 05:20:28 | SubnauticaModding/Nautilus | https://api.github.com/repos/SubnauticaModding/Nautilus | closed | Docs page on asynchronous prefab construction | Priority: Medium Type: Feature Request | **Describe the feature**
Does not need to be elaborate, but a beginner's guide with examples and tips and tricks would be nice. We should also warn beginners that asynchronous prefabs are visible in the world at the origin while being constructed, so you need to set them inactive. | 1.0 | Docs page on asynchronous prefab construction - **Describe the feature**
Does not need to be elaborate, but a beginner's guide with examples and tips and tricks would be nice. We should also warn beginners that asynchronous prefabs are visible in the world at the origin while being constructed, so you need to set them inactive. | priority | docs page on asynchronous prefab construction describe the feature does not need to be elaborate but a beginner s guide with examples and tips and tricks would be nice we should also warn beginners that asynchronous prefabs are visible in the world at the origin while being constructed so you need to set them inactive | 1 |
813,159 | 30,447,060,693 | IssuesEvent | 2023-07-15 20:17:23 | berkeli/My-Coursework-Planner | https://api.github.com/repos/berkeli/My-Coursework-Planner | opened | [TECH ED] What is an algorithm | 🐂 Size Medium 🏕 Priority Mandatory 📅 Week 2 🎯 Topic Communication 📅 Fundamentals 🎯 Topic Problem-Solving | ### Link to the coursework
https://www.youtube.com/watch?v=cDA3_5982h8
### Why are we doing this?
Programming is all about algorithms. But what is an algorithm?
Read [Introduction to Algorithms](https://github.com/thejsway/thejsway/blob/master/manuscript/intro02.md#introduction-to-algorithms) in the book the JS Way
Watch [The Exact Instructions Challenge😄](https://www.youtube.com/watch?v=cDA3_5982h8)
These show "algorithm recipes", and the video shows that you have to think carefully when building them. How does CYF Blocks help you build algorithms? Discuss this together in a Slack thread.
### Maximum time in hours
2
### How to submit
Discuss this together in a Slack thread | 1.0 | [TECH ED] What is an algorithm - ### Link to the coursework
https://www.youtube.com/watch?v=cDA3_5982h8
### Why are we doing this?
Programming is all about algorithms. But what is an algorithm?
Read [Introduction to Algorithms](https://github.com/thejsway/thejsway/blob/master/manuscript/intro02.md#introduction-to-algorithms) in the book the JS Way
Watch [The Exact Instructions Challenge😄](https://www.youtube.com/watch?v=cDA3_5982h8)
These show "algorithm recipes", and the video shows that you have to think carefully when building them. How does CYF Blocks help you build algorithms? Discuss this together in a Slack thread.
### Maximum time in hours
2
### How to submit
Discuss this together in a Slack thread | priority | what is an algorithm link to the coursework why are we doing this programming is all about algorithms but what is an algorithm read in the book the js way watch these show algorithm recipes and the video shows that you have to think carefully when building them how does cyf blocks help you build algorithms discuss this together in a slack thread maximum time in hours how to submit discuss this together in a slack thread | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.