Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
1,452
4,028,464,228
IssuesEvent
2016-05-18 06:29:33
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
opened
Possibility to improve parsing of treatments and conditions on a trial
enhancement Processors
See this trial: http://explorer.opentrials.net/trials/66906b54-0dee-4944-a621-c17f5e54fc5c There are patterns in the strings that could be detected. I understand that is this pattern does not apply across all records from this data source, it is difficult. Still, worth investigating. Screenshot: <img width="1280" alt="screenshot 2016-05-18 09 27 15" src="https://cloud.githubusercontent.com/assets/791678/15349358/e17cfbe4-1cda-11e6-8572-e3c40ff46b05.png">
1.0
Possibility to improve parsing of treatments and conditions on a trial - See this trial: http://explorer.opentrials.net/trials/66906b54-0dee-4944-a621-c17f5e54fc5c There are patterns in the strings that could be detected. I understand that is this pattern does not apply across all records from this data source, it is difficult. Still, worth investigating. Screenshot: <img width="1280" alt="screenshot 2016-05-18 09 27 15" src="https://cloud.githubusercontent.com/assets/791678/15349358/e17cfbe4-1cda-11e6-8572-e3c40ff46b05.png">
process
possibility to improve parsing of treatments and conditions on a trial see this trial there are patterns in the strings that could be detected i understand that is this pattern does not apply across all records from this data source it is difficult still worth investigating screenshot img width alt screenshot src
1
250,790
18,907,895,552
IssuesEvent
2021-11-16 11:01:53
KartoffelToby/ai_thermostat
https://api.github.com/repos/KartoffelToby/ai_thermostat
closed
Errors in HomeAssistant log - Can't find device model
documentation help wanted question
Hello, I have Eurotronic Spirit Zigbee (SPZB0001) but in homeassistant log I see the following errors: Logger: homeassistant Source: custom_components/ai_thermostat/climate.py:372 Integration: ai_thermostat (documentation, issues) First occurred: 14:45:41 (5 occurrences) Last logged: 14:46:12 Error doing job: Task exception was never retrieved Traceback (most recent call last): File "/config/custom_components/ai_thermostat/climate.py", line 372, in _async_tvr_changed if self.hass.states.get(self.heater_entity_id).attributes.get('device').get('model') is not None: AttributeError: 'NoneType' object has no attribute 'get' My config is: - platform: ai_thermostat name: ai_kitchen thermostat: climate.thermo_kitchen temperature_sensor: sensor.conditions_kitchen_temperature window_sensors: group.windows weather: weather.yandex_weather off_temperature: 2 "legacy" is enabled in specific settings of zigbee2mqtt. Thanks.
1.0
Errors in HomeAssistant log - Can't find device model - Hello, I have Eurotronic Spirit Zigbee (SPZB0001) but in homeassistant log I see the following errors: Logger: homeassistant Source: custom_components/ai_thermostat/climate.py:372 Integration: ai_thermostat (documentation, issues) First occurred: 14:45:41 (5 occurrences) Last logged: 14:46:12 Error doing job: Task exception was never retrieved Traceback (most recent call last): File "/config/custom_components/ai_thermostat/climate.py", line 372, in _async_tvr_changed if self.hass.states.get(self.heater_entity_id).attributes.get('device').get('model') is not None: AttributeError: 'NoneType' object has no attribute 'get' My config is: - platform: ai_thermostat name: ai_kitchen thermostat: climate.thermo_kitchen temperature_sensor: sensor.conditions_kitchen_temperature window_sensors: group.windows weather: weather.yandex_weather off_temperature: 2 "legacy" is enabled in specific settings of zigbee2mqtt. Thanks.
non_process
errors in homeassistant log can t find device model hello i have eurotronic spirit zigbee but in homeassistant log i see the following errors logger homeassistant source custom components ai thermostat climate py integration ai thermostat documentation issues first occurred occurrences last logged error doing job task exception was never retrieved traceback most recent call last file config custom components ai thermostat climate py line in async tvr changed if self hass states get self heater entity id attributes get device get model is not none attributeerror nonetype object has no attribute get my config is platform ai thermostat name ai kitchen thermostat climate thermo kitchen temperature sensor sensor conditions kitchen temperature window sensors group windows weather weather yandex weather off temperature legacy is enabled in specific settings of thanks
0
4,278
7,190,582,076
IssuesEvent
2018-02-02 17:45:52
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
Pull price from multiple locations and average per day/week
apps-ethPrice status-inprocess type-enhancement
Accountants need price per block. At 14 second resolution, a transaction may be more advantagiously priced for a token purchase, or perhaps less advantagiously priced. The idea is to price the transaction at the average of 5 min, 10 min, 1 hour, per day that best advantages the client. One could do this per block, per minute, per hour, per half-day, per day whatever, but one would have to do this for all transactions in the reporting period to avoid 'price shopping.' So here's the algo: Price a list of transactions per block (14 seconds), then do exact same pricing per minute, then per 15 minutes, then per hour, then per half-day, then per day. Do this across all days. Whichever is the most advantagious pricing mode, use that for the whole period (month, year, etc.)
1.0
Pull price from multiple locations and average per day/week - Accountants need price per block. At 14 second resolution, a transaction may be more advantagiously priced for a token purchase, or perhaps less advantagiously priced. The idea is to price the transaction at the average of 5 min, 10 min, 1 hour, per day that best advantages the client. One could do this per block, per minute, per hour, per half-day, per day whatever, but one would have to do this for all transactions in the reporting period to avoid 'price shopping.' So here's the algo: Price a list of transactions per block (14 seconds), then do exact same pricing per minute, then per 15 minutes, then per hour, then per half-day, then per day. Do this across all days. Whichever is the most advantagious pricing mode, use that for the whole period (month, year, etc.)
process
pull price from multiple locations and average per day week accountants need price per block at second resolution a transaction may be more advantagiously priced for a token purchase or perhaps less advantagiously priced the idea is to price the transaction at the average of min min hour per day that best advantages the client one could do this per block per minute per hour per half day per day whatever but one would have to do this for all transactions in the reporting period to avoid price shopping so here s the algo price a list of transactions per block seconds then do exact same pricing per minute then per minutes then per hour then per half day then per day do this across all days whichever is the most advantagious pricing mode use that for the whole period month year etc
1
281,102
21,315,373,507
IssuesEvent
2022-04-16 07:13:05
riakhaitan/pe
https://api.github.com/repos/riakhaitan/pe
opened
UG still refers to AddressBook
type.DocumentationBug severity.Low
![Screen Shot 2022-04-16 at 3.11.42 PM.png](https://raw.githubusercontent.com/riakhaitan/pe/main/files/6a0591d0-ebec-4679-abbe-312ef0dda302.png) All instances of addressbook have not been changed to MyGM. <!--session: 1650086337343-bc035de2-7316-4500-992a-e4b207d9b173--> <!--Version: Web v3.4.2-->
1.0
UG still refers to AddressBook - ![Screen Shot 2022-04-16 at 3.11.42 PM.png](https://raw.githubusercontent.com/riakhaitan/pe/main/files/6a0591d0-ebec-4679-abbe-312ef0dda302.png) All instances of addressbook have not been changed to MyGM. <!--session: 1650086337343-bc035de2-7316-4500-992a-e4b207d9b173--> <!--Version: Web v3.4.2-->
non_process
ug still refers to addressbook all instances of addressbook have not been changed to mygm
0
340,708
10,277,744,777
IssuesEvent
2019-08-25 08:03:06
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Qgis 3.4.2 crash when applied draw effects
Bug Crash/Data Corruption Feedback High Priority Symbology
Author Name: **Branko Bass** (Branko Bass) Original Redmine Issue: [20866](https://issues.qgis.org/issues/20866) Affected QGIS version: 3.4.2 Redmine category:symbology --- I tried to apply draw effects on vector layer of 1000 points. When "Outer Glow" added everything runs smoothly. But when tired to render all of them Qgis would crash. Tried multiple times same action. Thank you --- - [qgisCrash.mp4](https://issues.qgis.org/attachments/download/14016/qgisCrash.mp4) (Branko Bass)
1.0
Qgis 3.4.2 crash when applied draw effects - Author Name: **Branko Bass** (Branko Bass) Original Redmine Issue: [20866](https://issues.qgis.org/issues/20866) Affected QGIS version: 3.4.2 Redmine category:symbology --- I tried to apply draw effects on vector layer of 1000 points. When "Outer Glow" added everything runs smoothly. But when tired to render all of them Qgis would crash. Tried multiple times same action. Thank you --- - [qgisCrash.mp4](https://issues.qgis.org/attachments/download/14016/qgisCrash.mp4) (Branko Bass)
non_process
qgis crash when applied draw effects author name branko bass branko bass original redmine issue affected qgis version redmine category symbology i tried to apply draw effects on vector layer of points when outer glow added everything runs smoothly but when tired to render all of them qgis would crash tried multiple times same action thank you branko bass
0
90,480
3,822,261,122
IssuesEvent
2016-03-30 00:00:12
docker/docker
https://api.github.com/repos/docker/docker
closed
Add support for systemd cgroups
priority/P1
See opencontainers/runc#667. Cc @anusha-ragunathan who is working on it (just opening the issue for tracking).
1.0
Add support for systemd cgroups - See opencontainers/runc#667. Cc @anusha-ragunathan who is working on it (just opening the issue for tracking).
non_process
add support for systemd cgroups see opencontainers runc cc anusha ragunathan who is working on it just opening the issue for tracking
0
6,161
9,039,026,754
IssuesEvent
2019-02-10 01:29:22
material-components/material-components-ios
https://api.github.com/repos/material-components/material-components-ios
closed
[BottomNavigation] Graduate to Alpha
[BottomNavigation] type:Process
This was filed as an internal issue. If you are a Googler, please visit [b/119190664](http://b/119190664) for more details. <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/119190664](http://b/119190664) - Blocked by: https://github.com/material-components/material-components-ios/issues/5660
1.0
[BottomNavigation] Graduate to Alpha - This was filed as an internal issue. If you are a Googler, please visit [b/119190664](http://b/119190664) for more details. <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/119190664](http://b/119190664) - Blocked by: https://github.com/material-components/material-components-ios/issues/5660
process
graduate to alpha this was filed as an internal issue if you are a googler please visit for more details internal data associated internal bug blocked by
1
282,043
30,889,136,320
IssuesEvent
2023-08-04 02:17:36
madhans23/linux-4.1.15
https://api.github.com/repos/madhans23/linux-4.1.15
reopened
CVE-2019-15212 (Medium) detected in linux-stable-rtv4.1.33
Mend: dependency security vulnerability
## CVE-2019-15212 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.1.8. There is a double-free caused by a malicious USB device in the drivers/usb/misc/rio500.c driver. <p>Publish Date: 2019-08-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-15212>CVE-2019-15212</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15212">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15212</a></p> <p>Release Date: 2019-08-19</p> <p>Fix Resolution: v5.2-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-15212 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2019-15212 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.1.8. There is a double-free caused by a malicious USB device in the drivers/usb/misc/rio500.c driver. <p>Publish Date: 2019-08-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-15212>CVE-2019-15212</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15212">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15212</a></p> <p>Release Date: 2019-08-19</p> <p>Fix Resolution: v5.2-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in base branch master vulnerable source files drivers usb misc c drivers usb misc c drivers usb misc c vulnerability details an issue was discovered in the linux kernel before there is a double free caused by a malicious usb device in the drivers usb misc c driver publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
281,360
30,888,777,184
IssuesEvent
2023-08-04 01:48:54
nidhi7598/linux-4.1.15_CVE-2019-10220
https://api.github.com/repos/nidhi7598/linux-4.1.15_CVE-2019-10220
reopened
CVE-2018-16276 (High) detected in linuxlinux-4.4.302
Mend: dependency security vulnerability
## CVE-2018-16276 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2019-10220/commit/6a0d304d962ca933d73f507ce02157ef2791851c">6a0d304d962ca933d73f507ce02157ef2791851c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/yurex.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/yurex.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in yurex_read in drivers/usb/misc/yurex.c in the Linux kernel before 4.17.7. Local attackers could use user access read/writes with incorrect bounds checking in the yurex USB driver to crash the kernel or potentially escalate privileges. <p>Publish Date: 2018-08-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-16276>CVE-2018-16276</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-16276">https://nvd.nist.gov/vuln/detail/CVE-2018-16276</a></p> <p>Release Date: 2018-08-31</p> <p>Fix Resolution: 4.17.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-16276 (High) detected in linuxlinux-4.4.302 - ## CVE-2018-16276 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2019-10220/commit/6a0d304d962ca933d73f507ce02157ef2791851c">6a0d304d962ca933d73f507ce02157ef2791851c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/yurex.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/yurex.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in yurex_read in drivers/usb/misc/yurex.c in the Linux kernel before 4.17.7. Local attackers could use user access read/writes with incorrect bounds checking in the yurex USB driver to crash the kernel or potentially escalate privileges. <p>Publish Date: 2018-08-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-16276>CVE-2018-16276</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-16276">https://nvd.nist.gov/vuln/detail/CVE-2018-16276</a></p> <p>Release Date: 2018-08-31</p> <p>Fix Resolution: 4.17.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers usb misc yurex c drivers usb misc yurex c vulnerability details an issue was discovered in yurex read in drivers usb misc yurex c in the linux kernel before local attackers could use user access read writes with incorrect bounds checking in the yurex usb driver to crash the kernel or potentially escalate privileges publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
71,772
13,745,230,600
IssuesEvent
2020-10-06 02:15:48
new-zealand-research-information-system/nzris-data-specifications
https://api.github.com/repos/new-zealand-research-information-system/nzris-data-specifications
closed
Typo with IG01 description near Co-operative companies in Organisation Type codeset
codeset error
The description for the IG01 classification in the Organisation Type codeset contains the text “o-operative companies”, and this should be “Co-operative companies”.
1.0
Typo with IG01 description near Co-operative companies in Organisation Type codeset - The description for the IG01 classification in the Organisation Type codeset contains the text “o-operative companies”, and this should be “Co-operative companies”.
non_process
typo with description near co operative companies in organisation type codeset the description for the classification in the organisation type codeset contains the text “o operative companies” and this should be “co operative companies”
0
90,142
18,065,091,182
IssuesEvent
2021-09-20 18:08:35
zyantific/zydis
https://api.github.com/repos/zyantific/zydis
closed
Inconsistent segment override behavior for agen/agen_norel operands
C-bug C-enhancement A-decoder P-medium
I've been looking at `accepts_segment` evaluation logic and found a discrepancy in segment override handling. | Instruction(s) | operand_type | prefix_flags | accepts_segment | | --- | --- | --- | --- | | bndcl, bndcn, bndcu | agen | none | yes | | bndmk | agen_norel | none | no | | lea | agen | remove_segment | no | Issue with the database is that `remove_segment` is used only for `lea`. Issue in the generator is that currently `agen` and `agen_norel` are handled differently becasue `agen_norel` wasn't added here: ```delphi if (Definition.Operands.Items[I].OperandType in [optImplicitMem, optMEM, optMEMVSIBX, optMEMVSIBY, optMEMVSIBZ, optPTR, optAGEN, optMOFFS, optMIB]) then begin if (Definition.Operands.Items[I].IgnoreSegmentOverride) then begin Continue; end; Result := true; Break; end; ``` My suggestion to fix this is to modify `TZYDefinitionTableGenerator.AcceptsSegment`, so both `agen` and `agen_norel` are NOT present there. Then I would remove `remove_segment` from `lea` definition just to clean it up. This will ensure consistent, default LEA-like behavior for all `agen/agen_norel` operands without having to remember about adding extra stuff like `remove_segment` flag.
1.0
Inconsistent segment override behavior for agen/agen_norel operands - I've been looking at `accepts_segment` evaluation logic and found a discrepancy in segment override handling. | Instruction(s) | operand_type | prefix_flags | accepts_segment | | --- | --- | --- | --- | | bndcl, bndcn, bndcu | agen | none | yes | | bndmk | agen_norel | none | no | | lea | agen | remove_segment | no | Issue with the database is that `remove_segment` is used only for `lea`. Issue in the generator is that currently `agen` and `agen_norel` are handled differently becasue `agen_norel` wasn't added here: ```delphi if (Definition.Operands.Items[I].OperandType in [optImplicitMem, optMEM, optMEMVSIBX, optMEMVSIBY, optMEMVSIBZ, optPTR, optAGEN, optMOFFS, optMIB]) then begin if (Definition.Operands.Items[I].IgnoreSegmentOverride) then begin Continue; end; Result := true; Break; end; ``` My suggestion to fix this is to modify `TZYDefinitionTableGenerator.AcceptsSegment`, so both `agen` and `agen_norel` are NOT present there. Then I would remove `remove_segment` from `lea` definition just to clean it up. This will ensure consistent, default LEA-like behavior for all `agen/agen_norel` operands without having to remember about adding extra stuff like `remove_segment` flag.
non_process
inconsistent segment override behavior for agen agen norel operands i ve been looking at accepts segment evaluation logic and found a discrepancy in segment override handling instruction s operand type prefix flags accepts segment bndcl bndcn bndcu agen none yes bndmk agen norel none no lea agen remove segment no issue with the database is that remove segment is used only for lea issue in the generator is that currently agen and agen norel are handled differently becasue agen norel wasn t added here delphi if definition operands items operandtype in optimplicitmem optmem optmemvsibx optmemvsiby optmemvsibz optptr optagen optmoffs optmib then begin if definition operands items ignoresegmentoverride then begin continue end result true break end my suggestion to fix this is to modify tzydefinitiontablegenerator acceptssegment so both agen and agen norel are not present there then i would remove remove segment from lea definition just to clean it up this will ensure consistent default lea like behavior for all agen agen norel operands without having to remember about adding extra stuff like remove segment flag
0
31,445
14,967,685,337
IssuesEvent
2021-01-27 15:57:42
surge-synthesizer/surge
https://api.github.com/repos/surge-synthesizer/surge
opened
Optimize 11-band graphic EQ with SSE
Effects Feature Request Performance
We could have a 4x speedup by using SSE instead of 11 individual calls to BiquadFilter. So not a bad thing to do for 1.8.2!
True
Optimize 11-band graphic EQ with SSE - We could have a 4x speedup by using SSE instead of 11 individual calls to BiquadFilter. So not a bad thing to do for 1.8.2!
non_process
optimize band graphic eq with sse we could have a speedup by using sse instead of individual calls to biquadfilter so not a bad thing to do for
0
158,817
24,900,695,327
IssuesEvent
2022-10-28 20:31:55
Ingressive-for-Good/I4G-OPENSOURCE-FRONTEND-PROJECT-2022
https://api.github.com/repos/Ingressive-for-Good/I4G-OPENSOURCE-FRONTEND-PROJECT-2022
closed
Admin Design - Products Page (Tablet View)
documentation enhancement hacktoberfest-accepted hacktoberfest design
- Design the tablet view that enables the admin to be able to see the products uploaded for sale by the users pf the platform. - On this page, the admin can sort products based on name & category, view details of products, mark products as unavailable and permanently delete uploaded products - Design an empty state page - a page that indicates that no user have uploaded any product. - Design a filled state page - a page that shows the list of uploaded products by users of the platform. - There should be a more option button with three options: "view details" , "mark as unavailable" and "delete product" - Ensure you use the colors and typography on the style guide to ensure consistency.
1.0
Admin Design - Products Page (Tablet View) - - Design the tablet view that enables the admin to be able to see the products uploaded for sale by the users pf the platform. - On this page, the admin can sort products based on name & category, view details of products, mark products as unavailable and permanently delete uploaded products - Design an empty state page - a page that indicates that no user have uploaded any product. - Design a filled state page - a page that shows the list of uploaded products by users of the platform. - There should be a more option button with three options: "view details" , "mark as unavailable" and "delete product" - Ensure you use the colors and typography on the style guide to ensure consistency.
non_process
admin design products page tablet view design the tablet view that enables the admin to be able to see the products uploaded for sale by the users pf the platform on this page the admin can sort products based on name category view details of products mark products as unavailable and permanently delete uploaded products design an empty state page a page that indicates that no user have uploaded any product design a filled state page a page that shows the list of uploaded products by users of the platform there should be a more option button with three options view details mark as unavailable and delete product ensure you use the colors and typography on the style guide to ensure consistency
0
6,859
9,995,857,452
IssuesEvent
2019-07-11 21:23:35
AcademySoftwareFoundation/OpenCue
https://api.github.com/repos/AcademySoftwareFoundation/OpenCue
opened
Update CONTRIBUTING
process
**Describe the process** A clear and concise description of the action needing to be taken. I *think* the CLA section of CONTRIBUTING.md needs updating as the links look like they haven't been updated to reflect the migration to the ASWF?
1.0
Update CONTRIBUTING - **Describe the process** A clear and concise description of the action needing to be taken. I *think* the CLA section of CONTRIBUTING.md needs updating as the links look like they haven't been updated to reflect the migration to the ASWF?
process
update contributing describe the process a clear and concise description of the action needing to be taken i think the cla section of contributing md needs updating as the links look like they haven t been updated to reflect the migration to the aswf
1
18,799
24,700,981,215
IssuesEvent
2022-10-19 15:16:19
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
opened
docker image build broken on missing python
bug P1 process
### Description When the deployment action runs `RUN npm ci --only=production && npm cache clean --force --loglevel=error` there is now a consistent failure ``` #8 16.92 npm ERR! path /home/node/app/node_modules/@parcel/watcher #8 16.92 npm ERR! command failed #8 16.92 npm ERR! command sh -c -- node-gyp-build #8 16.92 npm ERR! gyp info it worked if it ends with ok #8 16.92 npm ERR! gyp info using node-gyp@9.3.0 #8 16.92 npm ERR! gyp info using node@16.18.0 | linux | arm64 #8 16.92 npm ERR! gyp ERR! find Python #8 16.92 npm ERR! gyp ERR! find Python Python is not set from command line or npm configuration #8 16.92 npm ERR! gyp ERR! find Python Python is not set from environment variable PYTHON #8 16.92 npm ERR! gyp ERR! find Python checking if "python3" can be used #8 16.92 npm ERR! gyp ERR! find Python - "python3" is not in PATH or produced an error #8 16.92 npm ERR! gyp ERR! find Python checking if "python" can be used #8 16.92 npm ERR! gyp ERR! find Python - "python" is not in PATH or produced an error #8 16.92 npm ERR! gyp ERR! find Python #8 16.92 npm ERR! gyp ERR! find Python ********************************************************** #8 16.93 npm ERR! gyp ERR! find Python You need to install the latest version of Python. #8 16.93 npm ERR! gyp ERR! find Python Node-gyp should be able to find and use Python. If not, #8 16.93 npm ERR! gyp ERR! find Python you can try one of the following options: #8 16.93 npm ERR! gyp ERR! find Python - Use the switch --python="/path/to/pythonexecutable" #8 16.93 npm ERR! gyp ERR! find Python (accepted by both node-gyp and npm) #8 16.93 npm ERR! gyp ERR! find Python - Set the environment variable PYTHON #8 16.93 npm ERR! gyp ERR! find Python - Set the npm configuration variable python: #8 16.93 npm ERR! gyp ERR! find Python npm config set python "/path/to/pythonexecutable" #8 16.93 npm ERR! gyp ERR! find Python For more information consult the documentation at: #8 16.93 npm ERR! gyp ERR! find Python https://github.com/nodejs/node-gyp#installation #8 16.93 npm ERR! gyp ERR! find Python ********************************************************** #8 16.93 npm ERR! gyp ERR! find Python #8 16.93 npm ERR! gyp ERR! configure error #8 16.93 npm ERR! gyp ERR! stack Error: Could not find any Python installation to use #8 16.93 npm ERR! gyp ERR! stack at PythonFinder.fail (/home/node/app/node_modules/node-gyp/lib/find-python.js:330:47) #8 16.93 npm ERR! gyp ERR! stack at PythonFinder.runChecks (/home/node/app/node_modules/node-gyp/lib/find-python.js:159:21) #8 16.93 npm ERR! gyp ERR! stack at PythonFinder.<anonymous> (/home/node/app/node_modules/node-gyp/lib/find-python.js:202:16) #8 16.93 npm ERR! gyp ERR! stack at PythonFinder.execFileCallback (/home/node/app/node_modules/node-gyp/lib/find-python.js:294:16) #8 16.93 npm ERR! gyp ERR! stack at exithandler (node:child_process:410:5) #8 16.93 npm ERR! gyp ERR! stack at ChildProcess.errorhandler (node:child_process:422:5) #8 16.93 npm ERR! gyp ERR! stack at ChildProcess.emit (node:events:513:28) #8 16.93 npm ERR! gyp ERR! stack at Process.ChildProcess._handle.onexit (node:internal/child_process:291:12) #8 16.93 npm ERR! gyp ERR! stack at onErrorNT (node:internal/child_process:485:16) #8 16.93 npm ERR! gyp ERR! stack at processTicksAndRejections (node:internal/process/task_queues:83:21) #8 16.93 npm ERR! gyp ERR! System Linux 5.10.124-linuxkit #8 16.93 npm ERR! gyp ERR! command "/usr/local/bin/node" "/home/node/app/node_modules/.bin/node-gyp" "rebuild" #8 16.93 npm ERR! gyp ERR! cwd /home/node/app/node_modules/@parcel/watcher #8 16.93 npm ERR! gyp ERR! node -v v16.18.0 #8 16.93 npm ERR! gyp ERR! node-gyp -v v9.3.0 #8 16.93 npm ERR! gyp ERR! not ok ``` ### Steps to reproduce build docker file with `docker build .` ### Additional context _No response_ ### Hedera network other ### Version v0.10.0 ### Operating system _No response_
1.0
docker image build broken on missing python - ### Description When the deployment action runs `RUN npm ci --only=production && npm cache clean --force --loglevel=error` there is now a consistent failure ``` #8 16.92 npm ERR! path /home/node/app/node_modules/@parcel/watcher #8 16.92 npm ERR! command failed #8 16.92 npm ERR! command sh -c -- node-gyp-build #8 16.92 npm ERR! gyp info it worked if it ends with ok #8 16.92 npm ERR! gyp info using node-gyp@9.3.0 #8 16.92 npm ERR! gyp info using node@16.18.0 | linux | arm64 #8 16.92 npm ERR! gyp ERR! find Python #8 16.92 npm ERR! gyp ERR! find Python Python is not set from command line or npm configuration #8 16.92 npm ERR! gyp ERR! find Python Python is not set from environment variable PYTHON #8 16.92 npm ERR! gyp ERR! find Python checking if "python3" can be used #8 16.92 npm ERR! gyp ERR! find Python - "python3" is not in PATH or produced an error #8 16.92 npm ERR! gyp ERR! find Python checking if "python" can be used #8 16.92 npm ERR! gyp ERR! find Python - "python" is not in PATH or produced an error #8 16.92 npm ERR! gyp ERR! find Python #8 16.92 npm ERR! gyp ERR! find Python ********************************************************** #8 16.93 npm ERR! gyp ERR! find Python You need to install the latest version of Python. #8 16.93 npm ERR! gyp ERR! find Python Node-gyp should be able to find and use Python. If not, #8 16.93 npm ERR! gyp ERR! find Python you can try one of the following options: #8 16.93 npm ERR! gyp ERR! find Python - Use the switch --python="/path/to/pythonexecutable" #8 16.93 npm ERR! gyp ERR! find Python (accepted by both node-gyp and npm) #8 16.93 npm ERR! gyp ERR! find Python - Set the environment variable PYTHON #8 16.93 npm ERR! gyp ERR! find Python - Set the npm configuration variable python: #8 16.93 npm ERR! gyp ERR! find Python npm config set python "/path/to/pythonexecutable" #8 16.93 npm ERR! gyp ERR! find Python For more information consult the documentation at: #8 16.93 npm ERR! gyp ERR! find Python https://github.com/nodejs/node-gyp#installation #8 16.93 npm ERR! gyp ERR! find Python ********************************************************** #8 16.93 npm ERR! gyp ERR! find Python #8 16.93 npm ERR! gyp ERR! configure error #8 16.93 npm ERR! gyp ERR! stack Error: Could not find any Python installation to use #8 16.93 npm ERR! gyp ERR! stack at PythonFinder.fail (/home/node/app/node_modules/node-gyp/lib/find-python.js:330:47) #8 16.93 npm ERR! gyp ERR! stack at PythonFinder.runChecks (/home/node/app/node_modules/node-gyp/lib/find-python.js:159:21) #8 16.93 npm ERR! gyp ERR! stack at PythonFinder.<anonymous> (/home/node/app/node_modules/node-gyp/lib/find-python.js:202:16) #8 16.93 npm ERR! gyp ERR! stack at PythonFinder.execFileCallback (/home/node/app/node_modules/node-gyp/lib/find-python.js:294:16) #8 16.93 npm ERR! gyp ERR! stack at exithandler (node:child_process:410:5) #8 16.93 npm ERR! gyp ERR! stack at ChildProcess.errorhandler (node:child_process:422:5) #8 16.93 npm ERR! gyp ERR! stack at ChildProcess.emit (node:events:513:28) #8 16.93 npm ERR! gyp ERR! stack at Process.ChildProcess._handle.onexit (node:internal/child_process:291:12) #8 16.93 npm ERR! gyp ERR! stack at onErrorNT (node:internal/child_process:485:16) #8 16.93 npm ERR! gyp ERR! stack at processTicksAndRejections (node:internal/process/task_queues:83:21) #8 16.93 npm ERR! gyp ERR! System Linux 5.10.124-linuxkit #8 16.93 npm ERR! gyp ERR! command "/usr/local/bin/node" "/home/node/app/node_modules/.bin/node-gyp" "rebuild" #8 16.93 npm ERR! gyp ERR! cwd /home/node/app/node_modules/@parcel/watcher #8 16.93 npm ERR! gyp ERR! node -v v16.18.0 #8 16.93 npm ERR! gyp ERR! node-gyp -v v9.3.0 #8 16.93 npm ERR! gyp ERR! not ok ``` ### Steps to reproduce build docker file with `docker build .` ### Additional context _No response_ ### Hedera network other ### Version v0.10.0 ### Operating system _No response_
process
docker image build broken on missing python description when the deployment action runs run npm ci only production npm cache clean force loglevel error there is now a consistent failure npm err path home node app node modules parcel watcher npm err command failed npm err command sh c node gyp build npm err gyp info it worked if it ends with ok npm err gyp info using node gyp npm err gyp info using node linux npm err gyp err find python npm err gyp err find python python is not set from command line or npm configuration npm err gyp err find python python is not set from environment variable python npm err gyp err find python checking if can be used npm err gyp err find python is not in path or produced an error npm err gyp err find python checking if python can be used npm err gyp err find python python is not in path or produced an error npm err gyp err find python npm err gyp err find python npm err gyp err find python you need to install the latest version of python npm err gyp err find python node gyp should be able to find and use python if not npm err gyp err find python you can try one of the following options npm err gyp err find python use the switch python path to pythonexecutable npm err gyp err find python accepted by both node gyp and npm npm err gyp err find python set the environment variable python npm err gyp err find python set the npm configuration variable python npm err gyp err find python npm config set python path to pythonexecutable npm err gyp err find python for more information consult the documentation at npm err gyp err find python npm err gyp err find python npm err gyp err find python npm err gyp err configure error npm err gyp err stack error could not find any python installation to use npm err gyp err stack at pythonfinder fail home node app node modules node gyp lib find python js npm err gyp err stack at pythonfinder runchecks home node app node modules node gyp lib find python js npm err gyp err stack at pythonfinder home node app node modules node gyp lib find python js npm err gyp err stack at pythonfinder execfilecallback home node app node modules node gyp lib find python js npm err gyp err stack at exithandler node child process npm err gyp err stack at childprocess errorhandler node child process npm err gyp err stack at childprocess emit node events npm err gyp err stack at process childprocess handle onexit node internal child process npm err gyp err stack at onerrornt node internal child process npm err gyp err stack at processticksandrejections node internal process task queues npm err gyp err system linux linuxkit npm err gyp err command usr local bin node home node app node modules bin node gyp rebuild npm err gyp err cwd home node app node modules parcel watcher npm err gyp err node v npm err gyp err node gyp v npm err gyp err not ok steps to reproduce build docker file with docker build additional context no response hedera network other version operating system no response
1
8,951
12,059,149,910
IssuesEvent
2020-04-15 18:45:06
googleapis/cloud-profiler-nodejs
https://api.github.com/repos/googleapis/cloud-profiler-nodejs
closed
CircleCI is getting slow.
api: cloudprofiler type: process
CircleCI builds seem to be taking longer. I was looking at the builds which our tests were queued behind and a saw this message: ``` This build is running under GoogleCloudPlatform's plan which provides 7 containers, plus 3 additional containers available for free and open source projects. Add Containers to run more builds concurrently. ``` Our builds were queued behind builds for GoogleCloudPlatform/google-cloud-java and GoogleCloudPlatform/google-cloud-python. What's the right thing to do here?
1.0
CircleCI is getting slow. - CircleCI builds seem to be taking longer. I was looking at the builds which our tests were queued behind and a saw this message: ``` This build is running under GoogleCloudPlatform's plan which provides 7 containers, plus 3 additional containers available for free and open source projects. Add Containers to run more builds concurrently. ``` Our builds were queued behind builds for GoogleCloudPlatform/google-cloud-java and GoogleCloudPlatform/google-cloud-python. What's the right thing to do here?
process
circleci is getting slow circleci builds seem to be taking longer i was looking at the builds which our tests were queued behind and a saw this message this build is running under googlecloudplatform s plan which provides containers plus additional containers available for free and open source projects add containers to run more builds concurrently our builds were queued behind builds for googlecloudplatform google cloud java and googlecloudplatform google cloud python what s the right thing to do here
1
143,631
19,190,473,727
IssuesEvent
2021-12-05 22:29:08
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Pod probes lead to blind SSRF from the node
kind/bug sig/network lifecycle/rotten sig/security triage/accepted
Note: I disclosed this private and was asked to open in public #### What happened: A pod probe can be configured to send readiness probes to arbitrary destinations, including localhost on the node. This is similar to CVE-2020-8555. Kubernetes Version: v0.20.0 (Kind) and v1.19.6 (GKE) were tested. I suspect its an issue on all versions Steps To Reproduce: ``` apiVersion: v1 kind: Pod metadata: name: test spec: containers: - image: docker.io/kennethreitz/httpbin name: httpbin readinessProbe: httpGet: host: 169.254.169.254 httpHeaders: - name: Metadata-Flavor value: Google path: /computeMetadata/v1/ port: 80 scheme: HTTP ``` This shows an example of accessing GCE metadata. Other possible targets are localhost endpoints, such as 127.0.0.1:2381/metrics (etcd metrics), or external targets (httpbin.org, for example). Note that because this is coming from the node, we can also access internal targets (pods) running on the same node, **bypassing NetworkPolicy restrictions**. In generally, the attack only allows sending GET requests, and does not allow viewing the data in the response. The status code of a response is, however, returned in pod events: "HTTP probe failed with statuscode: .... This could be used to extract some degree of private information about the node. Additionally, access to kubelet logs with verbosity >= 4 will log the full body, which would make this much work severe. Other possible risks are that these external calls are now coming from the k8s node which may have differing firewall/routing rules than a pod would. #### What you expected to happen: I don't know. Its certainly a tricky problem since its a part of the API and hard to change.
True
Pod probes lead to blind SSRF from the node - Note: I disclosed this private and was asked to open in public #### What happened: A pod probe can be configured to send readiness probes to arbitrary destinations, including localhost on the node. This is similar to CVE-2020-8555. Kubernetes Version: v0.20.0 (Kind) and v1.19.6 (GKE) were tested. I suspect its an issue on all versions Steps To Reproduce: ``` apiVersion: v1 kind: Pod metadata: name: test spec: containers: - image: docker.io/kennethreitz/httpbin name: httpbin readinessProbe: httpGet: host: 169.254.169.254 httpHeaders: - name: Metadata-Flavor value: Google path: /computeMetadata/v1/ port: 80 scheme: HTTP ``` This shows an example of accessing GCE metadata. Other possible targets are localhost endpoints, such as 127.0.0.1:2381/metrics (etcd metrics), or external targets (httpbin.org, for example). Note that because this is coming from the node, we can also access internal targets (pods) running on the same node, **bypassing NetworkPolicy restrictions**. In generally, the attack only allows sending GET requests, and does not allow viewing the data in the response. The status code of a response is, however, returned in pod events: "HTTP probe failed with statuscode: .... This could be used to extract some degree of private information about the node. Additionally, access to kubelet logs with verbosity >= 4 will log the full body, which would make this much work severe. Other possible risks are that these external calls are now coming from the k8s node which may have differing firewall/routing rules than a pod would. #### What you expected to happen: I don't know. Its certainly a tricky problem since its a part of the API and hard to change.
non_process
pod probes lead to blind ssrf from the node note i disclosed this private and was asked to open in public what happened a pod probe can be configured to send readiness probes to arbitrary destinations including localhost on the node this is similar to cve kubernetes version kind and gke were tested i suspect its an issue on all versions steps to reproduce apiversion kind pod metadata name test spec containers image docker io kennethreitz httpbin name httpbin readinessprobe httpget host httpheaders name metadata flavor value google path computemetadata port scheme http this shows an example of accessing gce metadata other possible targets are localhost endpoints such as metrics etcd metrics or external targets httpbin org for example note that because this is coming from the node we can also access internal targets pods running on the same node bypassing networkpolicy restrictions in generally the attack only allows sending get requests and does not allow viewing the data in the response the status code of a response is however returned in pod events http probe failed with statuscode this could be used to extract some degree of private information about the node additionally access to kubelet logs with verbosity will log the full body which would make this much work severe other possible risks are that these external calls are now coming from the node which may have differing firewall routing rules than a pod would what you expected to happen i don t know its certainly a tricky problem since its a part of the api and hard to change
0
409
2,851,438,547
IssuesEvent
2015-06-01 06:48:46
genomizer/genomizer-server
https://api.github.com/repos/genomizer/genomizer-server
closed
author when processing
Data Storage enhancement Processing
When sending a rawtoprofile request to the server, should the author be the logged in user or the author of the input file? Currently some clients use the author of the file and some use the logged in users name.
1.0
author when processing - When sending a rawtoprofile request to the server, should the author be the logged in user or the author of the input file? Currently some clients use the author of the file and some use the logged in users name.
process
author when processing when sending a rawtoprofile request to the server should the author be the logged in user or the author of the input file currently some clients use the author of the file and some use the logged in users name
1
57,385
11,745,348,204
IssuesEvent
2020-03-12 09:39:23
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
Fix generated TIMESTAMP precision in code generation
C: Code Generation E: All Editions P: High T: Defect
A lot of dialects do not produce correct `TIMESTAMP` precision in generated code. Either, they produce no precision at all (which jOOQ interprets as `0`), or they produce quite wrong numbers, such as in Informix: `4365`. This issue will list all the dialects that are affected. Fixing this is a prerequisite for #9933 `Optimistic locking does not work when recordTimestamp column has precision less than TIMESTAMP(3)`
1.0
Fix generated TIMESTAMP precision in code generation - A lot of dialects do not produce correct `TIMESTAMP` precision in generated code. Either, they produce no precision at all (which jOOQ interprets as `0`), or they produce quite wrong numbers, such as in Informix: `4365`. This issue will list all the dialects that are affected. Fixing this is a prerequisite for #9933 `Optimistic locking does not work when recordTimestamp column has precision less than TIMESTAMP(3)`
non_process
fix generated timestamp precision in code generation a lot of dialects do not produce correct timestamp precision in generated code either they produce no precision at all which jooq interprets as or they produce quite wrong numbers such as in informix this issue will list all the dialects that are affected fixing this is a prerequisite for optimistic locking does not work when recordtimestamp column has precision less than timestamp
0
21,685
30,178,542,682
IssuesEvent
2023-07-04 07:15:50
juspay/hyperswitch
https://api.github.com/repos/juspay/hyperswitch
closed
[FEATURE} Setting up process tracker to schedule email
A-process-tracker C-feature R-waiting-on-L1
### Feature description - Since expiry of the api_key would start right from its creation, while creating an api_key, insert a db entry into process_tracker table with new process_tracker workflow `ApiKeyExpiryWorkflow` and write logic for sending email by fetching merchant details. ### Possible implementation - When a merchant creates an API key with a future expiry date, a new process can be added to the `process_tracker` table. This process includes `tracking_data` that contains the `key_id`, `merchant_id`, `api_key_expiry`, and `expiry_reminder_days`. The api_key_expiry represents the date and time when the API key expires, and `expiry_reminder_days` is a vector specifying the number of days before the expiry when reminder emails should be sent (currently set as [7, 3, 1]). `schedule_time` will be set to 1st reminder date, and after the 1st email is sent, we update the `schedule_time` as `tracking_data.expiry_reminder_days[retry]` where `retry` initially set to 0, acts as a counter. This results in emails being scheduled at the respective number of days before API key expiry. So during 1st time, `schedule_time` would be 7 days prior to api_key expiry and during 2nd time, it would be 3 days prior and so on. Hence, for a given api_key, only one process should reside in `process_tracker` whose `schedule_time` changes accordingly. A new workflow `ApiKeyExpiryWorkflow` should be introduced on which `ProcessTrackerWorkflow` is implemented. The `execute_workflow()` function retrieves the merchant's email address and determines whether to send the email, schedule it for a later time, or finish the current process based on the retry count.
1.0
[FEATURE} Setting up process tracker to schedule email - ### Feature description - Since expiry of the api_key would start right from its creation, while creating an api_key, insert a db entry into process_tracker table with new process_tracker workflow `ApiKeyExpiryWorkflow` and write logic for sending email by fetching merchant details. ### Possible implementation - When a merchant creates an API key with a future expiry date, a new process can be added to the `process_tracker` table. This process includes `tracking_data` that contains the `key_id`, `merchant_id`, `api_key_expiry`, and `expiry_reminder_days`. The api_key_expiry represents the date and time when the API key expires, and `expiry_reminder_days` is a vector specifying the number of days before the expiry when reminder emails should be sent (currently set as [7, 3, 1]). `schedule_time` will be set to 1st reminder date, and after the 1st email is sent, we update the `schedule_time` as `tracking_data.expiry_reminder_days[retry]` where `retry` initially set to 0, acts as a counter. This results in emails being scheduled at the respective number of days before API key expiry. So during 1st time, `schedule_time` would be 7 days prior to api_key expiry and during 2nd time, it would be 3 days prior and so on. Hence, for a given api_key, only one process should reside in `process_tracker` whose `schedule_time` changes accordingly. A new workflow `ApiKeyExpiryWorkflow` should be introduced on which `ProcessTrackerWorkflow` is implemented. The `execute_workflow()` function retrieves the merchant's email address and determines whether to send the email, schedule it for a later time, or finish the current process based on the retry count.
process
feature setting up process tracker to schedule email feature description since expiry of the api key would start right from its creation while creating an api key insert a db entry into process tracker table with new process tracker workflow apikeyexpiryworkflow and write logic for sending email by fetching merchant details possible implementation when a merchant creates an api key with a future expiry date a new process can be added to the process tracker table this process includes tracking data that contains the key id merchant id api key expiry and expiry reminder days the api key expiry represents the date and time when the api key expires and expiry reminder days is a vector specifying the number of days before the expiry when reminder emails should be sent currently set as schedule time will be set to reminder date and after the email is sent we update the schedule time as tracking data expiry reminder days where retry initially set to acts as a counter this results in emails being scheduled at the respective number of days before api key expiry so during time schedule time would be days prior to api key expiry and during time it would be days prior and so on hence for a given api key only one process should reside in process tracker whose schedule time changes accordingly a new workflow apikeyexpiryworkflow should be introduced on which processtrackerworkflow is implemented the execute workflow function retrieves the merchant s email address and determines whether to send the email schedule it for a later time or finish the current process based on the retry count
1
22,121
30,658,517,297
IssuesEvent
2023-07-25 13:39:49
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] Mark selected join condition columns
.metabase-lib .Team/QueryProcessor :hammer_and_wrench:
For an existing join, the FE needs to highlight currently selected condition columns in column pickers. It's possible to implement this by calling `externalOp(condition)` and looking up for its columns, but it's not very robust (have to compare by name + table name) and isn't consistent with other MLv2 methods. It'd be nice if MLv2 could mark selected columns with `selected: true` in `joinConditionLHSColumns` and `joinConditionRHSColumns`
1.0
[MLv2] Mark selected join condition columns - For an existing join, the FE needs to highlight currently selected condition columns in column pickers. It's possible to implement this by calling `externalOp(condition)` and looking up for its columns, but it's not very robust (have to compare by name + table name) and isn't consistent with other MLv2 methods. It'd be nice if MLv2 could mark selected columns with `selected: true` in `joinConditionLHSColumns` and `joinConditionRHSColumns`
process
mark selected join condition columns for an existing join the fe needs to highlight currently selected condition columns in column pickers it s possible to implement this by calling externalop condition and looking up for its columns but it s not very robust have to compare by name table name and isn t consistent with other methods it d be nice if could mark selected columns with selected true in joinconditionlhscolumns and joinconditionrhscolumns
1
15,765
19,913,036,936
IssuesEvent
2022-01-25 19:13:57
input-output-hk/high-assurance-legacy
https://api.github.com/repos/input-output-hk/high-assurance-legacy
closed
Change the term “standard weak residual” to “normal weak residual”
language: isabelle topic: process calculus type: improvement
We believe that the term “normal weak residual” is more appropriate than the term “standard weak residual”. The term “normal weak residual” is being used already in @jeltsch’s [WFLP ’19 contribution][wflp-2019-contribution]. Our goal is to switch to this term in identifiers and documentation. [wflp-2019-contribution]: https://github.com/jeltsch/wflp-2019 "A Process Calculus for Formally Verifying Blockchain Consensus Protocols"
1.0
Change the term “standard weak residual” to “normal weak residual” - We believe that the term “normal weak residual” is more appropriate than the term “standard weak residual”. The term “normal weak residual” is being used already in @jeltsch’s [WFLP ’19 contribution][wflp-2019-contribution]. Our goal is to switch to this term in identifiers and documentation. [wflp-2019-contribution]: https://github.com/jeltsch/wflp-2019 "A Process Calculus for Formally Verifying Blockchain Consensus Protocols"
process
change the term “standard weak residual” to “normal weak residual” we believe that the term “normal weak residual” is more appropriate than the term “standard weak residual” the term “normal weak residual” is being used already in jeltsch’s our goal is to switch to this term in identifiers and documentation a process calculus for formally verifying blockchain consensus protocols
1
5,952
2,799,496,685
IssuesEvent
2015-05-13 01:21:39
ThoughtWorks-Xi-an-Scala-Camp/scalaz-monad-factory
https://api.github.com/repos/ThoughtWorks-Xi-an-Scala-Camp/scalaz-monad-factory
closed
Add test cases for List monads
test case
For now, there are only test cases for Option monads. We need test if the macro works with scalaz's List monads.
1.0
Add test cases for List monads - For now, there are only test cases for Option monads. We need test if the macro works with scalaz's List monads.
non_process
add test cases for list monads for now there are only test cases for option monads we need test if the macro works with scalaz s list monads
0
5,915
8,736,236,804
IssuesEvent
2018-12-11 18:58:30
ipfs/go-ipfs
https://api.github.com/repos/ipfs/go-ipfs
opened
Improving impromptu communication
process
## Goal Based on voting in #5819. We want to support easier impromptu communicate as a team. There are a collection of reasons why we think this would be helpful: a faster way to get unblocked, easy way to ask questions etc. A number of interesting solutions have already been mentioned like Teamspeak, Mumble or Discord. In this issue we're going to discuss (and decide on) something to try first. ## Summary *stebalien* - I wish there were some way to flag ourselves as busy/available. Going to IRC and just shouting into the ether hoping someone's listening can be a bit isolating. [also concerned about the distracting nature of lots of notifications] *eingenito* - I wish there were a more chatty place where just go-ipfs development was discussed. [interested in more conversation] *djdv* - I do think how we communicate on a regular basis, is an important thing to focus on, and would like to discuss this at some point somewhere. To me subtle changes in our communication patterns and tools could have a big impact. [mentions risk of tribal knowledge and how it might be mitigated by discipline] *magik2k* - IMO this is definitely worth trying at some point. [mentions experiences with Teamspeak, and also is concerned about tribal knowledge] *DonaldTsang* - [Discord is an option] *kevina* - [mentions risk of tribal knowledge in emphasizing voice/video communications]
1.0
Improving impromptu communication - ## Goal Based on voting in #5819. We want to support easier impromptu communicate as a team. There are a collection of reasons why we think this would be helpful: a faster way to get unblocked, easy way to ask questions etc. A number of interesting solutions have already been mentioned like Teamspeak, Mumble or Discord. In this issue we're going to discuss (and decide on) something to try first. ## Summary *stebalien* - I wish there were some way to flag ourselves as busy/available. Going to IRC and just shouting into the ether hoping someone's listening can be a bit isolating. [also concerned about the distracting nature of lots of notifications] *eingenito* - I wish there were a more chatty place where just go-ipfs development was discussed. [interested in more conversation] *djdv* - I do think how we communicate on a regular basis, is an important thing to focus on, and would like to discuss this at some point somewhere. To me subtle changes in our communication patterns and tools could have a big impact. [mentions risk of tribal knowledge and how it might be mitigated by discipline] *magik2k* - IMO this is definitely worth trying at some point. [mentions experiences with Teamspeak, and also is concerned about tribal knowledge] *DonaldTsang* - [Discord is an option] *kevina* - [mentions risk of tribal knowledge in emphasizing voice/video communications]
process
improving impromptu communication goal based on voting in we want to support easier impromptu communicate as a team there are a collection of reasons why we think this would be helpful a faster way to get unblocked easy way to ask questions etc a number of interesting solutions have already been mentioned like teamspeak mumble or discord in this issue we re going to discuss and decide on something to try first summary stebalien   i wish there were some way to flag ourselves as busy available going to irc and just shouting into the ether hoping someone s listening can be a bit isolating eingenito   i wish there were a more chatty place where just go ipfs development was discussed djdv   i do think how we communicate on a regular basis is an important thing to focus on and would like to discuss this at some point somewhere to me subtle changes in our communication patterns and tools could have a big impact   imo this is definitely worth trying at some point donaldtsang   kevina  
1
33,896
6,266,116,819
IssuesEvent
2017-07-16 23:13:36
MartinLoeper/KAMP-DSL
https://api.github.com/repos/MartinLoeper/KAMP-DSL
closed
Create an easy installer
documentation enhancement
As the name says: create an installer using software configurations and project sets. Update the wiki accordingly.
1.0
Create an easy installer - As the name says: create an installer using software configurations and project sets. Update the wiki accordingly.
non_process
create an easy installer as the name says create an installer using software configurations and project sets update the wiki accordingly
0
1,088
3,551,807,803
IssuesEvent
2016-01-21 06:49:53
Project60/org.project60.sepa
https://api.github.com/repos/Project60/org.project60.sepa
opened
contact_id not set in partial mandate
bug CiviCRM 4.4 payment processor
Under certain circumstances (not fully understood), the ``$this->getVar('_contactID');`` in [this line](https://github.com/Project60/org.project60.sepa/blob/1.2.1/CRM/Core/Payment/SDD.php#L157) doesn't yield the contact id, which can cause troubles in the further processing in the partial mandates. Observed with CiviSEPA 1.2.1 on CiviCRM 4.4.20 on a donation page for OOFF contributions.
1.0
contact_id not set in partial mandate - Under certain circumstances (not fully understood), the ``$this->getVar('_contactID');`` in [this line](https://github.com/Project60/org.project60.sepa/blob/1.2.1/CRM/Core/Payment/SDD.php#L157) doesn't yield the contact id, which can cause troubles in the further processing in the partial mandates. Observed with CiviSEPA 1.2.1 on CiviCRM 4.4.20 on a donation page for OOFF contributions.
process
contact id not set in partial mandate under certain circumstances not fully understood the this getvar contactid in doesn t yield the contact id which can cause troubles in the further processing in the partial mandates observed with civisepa on civicrm on a donation page for ooff contributions
1
257,915
19,535,343,616
IssuesEvent
2021-12-31 04:53:05
seanpm2001/SeansLifeArchive_Images_MotorWorld_CarFactory
https://api.github.com/repos/seanpm2001/SeansLifeArchive_Images_MotorWorld_CarFactory
opened
Contributor attribution thread
documentation enhancement good first issue
*** ### Seanpm2001/SeansLifeArchive_Images_MotorWorld_CarFactory/ Contributor attribution If you are a contributor on this project, you will be credited in this thread. @all-contributors will be helping out here. ***
1.0
Contributor attribution thread - *** ### Seanpm2001/SeansLifeArchive_Images_MotorWorld_CarFactory/ Contributor attribution If you are a contributor on this project, you will be credited in this thread. @all-contributors will be helping out here. ***
non_process
contributor attribution thread seanslifearchive images motorworld carfactory contributor attribution if you are a contributor on this project you will be credited in this thread all contributors will be helping out here
0
236,725
18,107,291,231
IssuesEvent
2021-09-22 20:41:02
hashicorp/terraform-provider-azuread
https://api.github.com/repos/hashicorp/terraform-provider-azuread
closed
azuread_directory_role_member argument documentation is not accurate
documentation feature/directory-roles
### Terraform Version ``` Terraform v1.0.7 on windows_amd64 + provider registry.terraform.io/hashicorp/azuread v2.3 ``` ### Additional Context The documentation on 2.3 resource azuread_directory_role_member should be updated: - Documentation lists the argument as `directory_role_object_id` but it seems to be just `role_object_id` - Documentation of `member_object_id` states that this only works for _Users_ and _Service Principals_. However; role assignable groups do also work perfectly. Nice :-) Maybe this could be added to the documentation.
1.0
azuread_directory_role_member argument documentation is not accurate - ### Terraform Version ``` Terraform v1.0.7 on windows_amd64 + provider registry.terraform.io/hashicorp/azuread v2.3 ``` ### Additional Context The documentation on 2.3 resource azuread_directory_role_member should be updated: - Documentation lists the argument as `directory_role_object_id` but it seems to be just `role_object_id` - Documentation of `member_object_id` states that this only works for _Users_ and _Service Principals_. However; role assignable groups do also work perfectly. Nice :-) Maybe this could be added to the documentation.
non_process
azuread directory role member argument documentation is not accurate terraform version terraform on windows provider registry terraform io hashicorp azuread additional context the documentation on resource azuread directory role member should be updated documentation lists the argument as directory role object id but it seems to be just role object id documentation of member object id states that this only works for users and service principals however role assignable groups do also work perfectly nice maybe this could be added to the documentation
0
63,730
26,501,421,835
IssuesEvent
2023-01-18 10:33:36
hashicorp/terraform-provider-azurerm
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
closed
Related to issue 12288: I have updated to the 2.68.0 azurerm provider and still getting the "MissingRequiredParameter" error
question service/postgresql
### Is there an existing issue for this? - [X] I have searched the existing issues ### Community Note <!--- Please keep this note for the community ---> * Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version 1.0.0 ### AzureRM Provider Version 2.68.0 ### Affected Resource(s)/Data Source(s) azurerm_postgresql_flexible_server ### Terraform Configuration Files ```hcl resource "azurerm_postgresql_flexible_server" "cpro-prod-east" { name = "aaaotcproprodeastus-paasdb" resource_group_name = data.azurerm_resource_group.cpro-east.name location = data.azurerm_resource_group.cpro-east.location version = "12" delegated_subnet_id = data.azurerm_subnet.cpro-east.id zone = "1" storage_mb = 32768 sku_name = "GP_Standard_D4s_v3" depends_on = [azurerm_private_dns_zone_virtual_network_link.cpro-east] } ``` ### Debug Output/Panic Output ```shell │ Error: creating Postgresql Flexible Server "aaaotcproprodeastus-paasdb" (Resource Group "aa-aot-cpro-prod-eastus-rg"): postgresqlflexibleservers.ServersClient#Create: Failure sending request: StatusCode=0 -- Original Error: Code="MissingRequiredParameter" Message="Parameter 'PrivateDnsZoneArmResourceId' must be specified. This parameter cannot be NULL or empty." │ │ with azurerm_postgresql_flexible_server.cpro-prod-east, │ on postgres.tf line 21, in resource "azurerm_postgresql_flexible_server" "cpro-prod-east": │ 21: resource "azurerm_postgresql_flexible_server" "cpro-prod-east" { ``` ### Expected Behaviour Postgres Flexible Server should be created with Private Connection to Subnet. ### Actual Behaviour Resource doesn't create and error displayes ### Steps to Reproduce terraform apply -target azurerm_postgresql_flexible_server.cpro-prod-east ### Important Factoids _No response_ ### References I researched https://github.com/hashicorp/terraform-provider-azurerm/issues/12053 and updated the Provider in my provider.tf file.
1.0
Related to issue 12288: I have updated to the 2.68.0 azurerm provider and still getting the "MissingRequiredParameter" error - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Community Note <!--- Please keep this note for the community ---> * Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version 1.0.0 ### AzureRM Provider Version 2.68.0 ### Affected Resource(s)/Data Source(s) azurerm_postgresql_flexible_server ### Terraform Configuration Files ```hcl resource "azurerm_postgresql_flexible_server" "cpro-prod-east" { name = "aaaotcproprodeastus-paasdb" resource_group_name = data.azurerm_resource_group.cpro-east.name location = data.azurerm_resource_group.cpro-east.location version = "12" delegated_subnet_id = data.azurerm_subnet.cpro-east.id zone = "1" storage_mb = 32768 sku_name = "GP_Standard_D4s_v3" depends_on = [azurerm_private_dns_zone_virtual_network_link.cpro-east] } ``` ### Debug Output/Panic Output ```shell │ Error: creating Postgresql Flexible Server "aaaotcproprodeastus-paasdb" (Resource Group "aa-aot-cpro-prod-eastus-rg"): postgresqlflexibleservers.ServersClient#Create: Failure sending request: StatusCode=0 -- Original Error: Code="MissingRequiredParameter" Message="Parameter 'PrivateDnsZoneArmResourceId' must be specified. This parameter cannot be NULL or empty." │ │ with azurerm_postgresql_flexible_server.cpro-prod-east, │ on postgres.tf line 21, in resource "azurerm_postgresql_flexible_server" "cpro-prod-east": │ 21: resource "azurerm_postgresql_flexible_server" "cpro-prod-east" { ``` ### Expected Behaviour Postgres Flexible Server should be created with Private Connection to Subnet. ### Actual Behaviour Resource doesn't create and error displayes ### Steps to Reproduce terraform apply -target azurerm_postgresql_flexible_server.cpro-prod-east ### Important Factoids _No response_ ### References I researched https://github.com/hashicorp/terraform-provider-azurerm/issues/12053 and updated the Provider in my provider.tf file.
non_process
related to issue i have updated to the azurerm provider and still getting the missingrequiredparameter error is there an existing issue for this i have searched the existing issues community note please vote on this issue by adding a thumbsup to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version azurerm provider version affected resource s data source s azurerm postgresql flexible server terraform configuration files hcl resource azurerm postgresql flexible server cpro prod east name aaaotcproprodeastus paasdb resource group name data azurerm resource group cpro east name location data azurerm resource group cpro east location version delegated subnet id data azurerm subnet cpro east id zone storage mb sku name gp standard depends on debug output panic output shell │ error creating postgresql flexible server aaaotcproprodeastus paasdb resource group aa aot cpro prod eastus rg postgresqlflexibleservers serversclient create failure sending request statuscode original error code missingrequiredparameter message parameter privatednszonearmresourceid must be specified this parameter cannot be null or empty │ │ with azurerm postgresql flexible server cpro prod east │ on postgres tf line in resource azurerm postgresql flexible server cpro prod east │ resource azurerm postgresql flexible server cpro prod east expected behaviour postgres flexible server should be created with private connection to subnet actual behaviour resource doesn t create and error displayes steps to reproduce terraform apply target azurerm postgresql flexible server cpro prod east important factoids no response references i researched and updated the provider in my provider tf file
0
13,507
3,151,273,385
IssuesEvent
2015-09-16 06:55:36
owncloud/ios
https://api.github.com/repos/owncloud/ios
closed
spacing in toolbar buttons
Design Enhancement
One think that always bothered me is that the icons on the bottom (Files, Upload, Share link,..) look a bit cramped. It would be cool to have a tiny bit of additional space at the bottom and top of the toolbar. What do you think? @jancborchardt
1.0
spacing in toolbar buttons - One think that always bothered me is that the icons on the bottom (Files, Upload, Share link,..) look a bit cramped. It would be cool to have a tiny bit of additional space at the bottom and top of the toolbar. What do you think? @jancborchardt
non_process
spacing in toolbar buttons one think that always bothered me is that the icons on the bottom files upload share link look a bit cramped it would be cool to have a tiny bit of additional space at the bottom and top of the toolbar what do you think jancborchardt
0
341,991
24,724,352,048
IssuesEvent
2022-10-20 13:06:04
ham-radio-software/lzhuf
https://api.github.com/repos/ham-radio-software/lzhuf
reopened
Add script to build RPM packages for x86_64
documentation enhancement
Need to use a GitHub action with a Fedora docker container to build RPM packages for lzhuf. Initial target will be el-8 using either Rocky 8 or AlmaLinux 8 in the chroot used for the build.
1.0
Add script to build RPM packages for x86_64 - Need to use a GitHub action with a Fedora docker container to build RPM packages for lzhuf. Initial target will be el-8 using either Rocky 8 or AlmaLinux 8 in the chroot used for the build.
non_process
add script to build rpm packages for need to use a github action with a fedora docker container to build rpm packages for lzhuf initial target will be el using either rocky or almalinux in the chroot used for the build
0
22,570
31,791,320,564
IssuesEvent
2023-09-13 03:45:02
medic/cht-core
https://api.github.com/repos/medic/cht-core
opened
Release 4.2.3
Type: Internal process
# Planning - Product Manager - [ ] Create an GH Milestone and add this issue to it. - [ ] Add all the issues to be worked on to the Milestone. - [ ] Ensure that all issues are labelled correctly, particularly ensure that "Regressions" are labelled with "Affects: <version>" labels. The "Affects" label is used in a link in the Known Issues section of the release notes of that version so it has to match exactly. To make sure the label is correct go to the [release notes](https://docs.communityhealthtoolkit.org/core/releases/#release-notes) and ensure the issue is listed. # Development - Release Engineer When development is ready to begin one of the engineers should be nominated as a Release Engineer. They will be responsible for making sure the following tasks are completed though not necessarily completing them. - [ ] Set the version number in `package.json` and `package-lock.json` and submit a PR to the release branch. The easiest way to do this is to use `npm --no-git-tag-version version patch`. - [ ] Ensure that issues from merged commits are closed and mapped to a milestone. - [ ] Write an update in the #product-team Slack channel summarising development and identifying any blockers (the [milestone-status](https://github.com/medic/support-scripts/tree/master/milestone-status) script can be used to get a breakdown of the issues). The Release Engineer is to update this every week until the version is released. # Releasing - Release Engineer Once all issues have been merged into `master` then the release process can start: - [ ] Build a beta named `<major>.<minor>.<patch>-beta.1` by pushing a lightweight git tag (e.g. `git tag <major>.<minor>.<patch>-beta.1`). - [ ] Once the CI completes successfully notify the team by writing a message in the #product-team Slack channel: ``` @product_team, I’ve just created the `<major>.<minor>.<patch>-beta.1` tag. Please let me know if there’s any final update we need to make. If all is good, then in 24h, I will start the release. Thanks! ``` - [ ] Add release notes to the [Core Framework Releases](https://docs.communityhealthtoolkit.org/core/releases/) page: - [ ] Create a new document for the release in the [releases folder](https://github.com/medic/cht-docs/tree/main/content/en/core/releases). - [ ] Ensure all issues are in the GH Milestone, that they're correctly labelled (in particular: they have the right Type, "UI/UX" if they change the UI, and "Breaking change" if appropriate), and have human readable descriptions. - [ ] Use [this script](https://github.com/medic/cht-core/blob/master/scripts/release-notes) to export the issues into our release note format. - [ ] Manually document any known migration steps and known issues. - [ ] Add a link to the new release page in the [Release Notes](https://docs.communityhealthtoolkit.org/core/releases/#release-notes) section. - [ ] Assign the PR to: - The Director of Technology or a developer - An SRE to review and confirm the documentation on upgrade instructions and breaking changes is sufficient - [ ] Create a release in GitHub from the release branch so it shows up under the [Releases tab](https://github.com/medic/cht-core/releases) with the naming convention `<major>.<minor>.<patch>`. This will create the git tag automatically. Ensure the release notes PR above is merged. Link to the release notes in the description of the release. - [ ] Confirm the release build completes successfully and the new release is available on the [market](https://staging.dev.medicmobile.org/builds_4/releases). Make sure that the document has new entry with `id: medic:medic:<major>.<minor>.<patch>` - [ ] Announce the release on the [CHT forum](https://forum.communityhealthtoolkit.org/), under the "Product - Releases" category using this template: ``` *Announcing the release of {{version}}* This release fixes {{number of bugs}}. Read the [release notes]({{url}}) for full details. ``` - [ ] Mark this issue "done" and close the Milestone.
1.0
Release 4.2.3 - # Planning - Product Manager - [ ] Create an GH Milestone and add this issue to it. - [ ] Add all the issues to be worked on to the Milestone. - [ ] Ensure that all issues are labelled correctly, particularly ensure that "Regressions" are labelled with "Affects: <version>" labels. The "Affects" label is used in a link in the Known Issues section of the release notes of that version so it has to match exactly. To make sure the label is correct go to the [release notes](https://docs.communityhealthtoolkit.org/core/releases/#release-notes) and ensure the issue is listed. # Development - Release Engineer When development is ready to begin one of the engineers should be nominated as a Release Engineer. They will be responsible for making sure the following tasks are completed though not necessarily completing them. - [ ] Set the version number in `package.json` and `package-lock.json` and submit a PR to the release branch. The easiest way to do this is to use `npm --no-git-tag-version version patch`. - [ ] Ensure that issues from merged commits are closed and mapped to a milestone. - [ ] Write an update in the #product-team Slack channel summarising development and identifying any blockers (the [milestone-status](https://github.com/medic/support-scripts/tree/master/milestone-status) script can be used to get a breakdown of the issues). The Release Engineer is to update this every week until the version is released. # Releasing - Release Engineer Once all issues have been merged into `master` then the release process can start: - [ ] Build a beta named `<major>.<minor>.<patch>-beta.1` by pushing a lightweight git tag (e.g. `git tag <major>.<minor>.<patch>-beta.1`). - [ ] Once the CI completes successfully notify the team by writing a message in the #product-team Slack channel: ``` @product_team, I’ve just created the `<major>.<minor>.<patch>-beta.1` tag. Please let me know if there’s any final update we need to make. If all is good, then in 24h, I will start the release. Thanks! ``` - [ ] Add release notes to the [Core Framework Releases](https://docs.communityhealthtoolkit.org/core/releases/) page: - [ ] Create a new document for the release in the [releases folder](https://github.com/medic/cht-docs/tree/main/content/en/core/releases). - [ ] Ensure all issues are in the GH Milestone, that they're correctly labelled (in particular: they have the right Type, "UI/UX" if they change the UI, and "Breaking change" if appropriate), and have human readable descriptions. - [ ] Use [this script](https://github.com/medic/cht-core/blob/master/scripts/release-notes) to export the issues into our release note format. - [ ] Manually document any known migration steps and known issues. - [ ] Add a link to the new release page in the [Release Notes](https://docs.communityhealthtoolkit.org/core/releases/#release-notes) section. - [ ] Assign the PR to: - The Director of Technology or a developer - An SRE to review and confirm the documentation on upgrade instructions and breaking changes is sufficient - [ ] Create a release in GitHub from the release branch so it shows up under the [Releases tab](https://github.com/medic/cht-core/releases) with the naming convention `<major>.<minor>.<patch>`. This will create the git tag automatically. Ensure the release notes PR above is merged. Link to the release notes in the description of the release. - [ ] Confirm the release build completes successfully and the new release is available on the [market](https://staging.dev.medicmobile.org/builds_4/releases). Make sure that the document has new entry with `id: medic:medic:<major>.<minor>.<patch>` - [ ] Announce the release on the [CHT forum](https://forum.communityhealthtoolkit.org/), under the "Product - Releases" category using this template: ``` *Announcing the release of {{version}}* This release fixes {{number of bugs}}. Read the [release notes]({{url}}) for full details. ``` - [ ] Mark this issue "done" and close the Milestone.
process
release planning product manager create an gh milestone and add this issue to it add all the issues to be worked on to the milestone ensure that all issues are labelled correctly particularly ensure that regressions are labelled with affects labels the affects label is used in a link in the known issues section of the release notes of that version so it has to match exactly to make sure the label is correct go to the and ensure the issue is listed development release engineer when development is ready to begin one of the engineers should be nominated as a release engineer they will be responsible for making sure the following tasks are completed though not necessarily completing them set the version number in package json and package lock json and submit a pr to the release branch the easiest way to do this is to use npm no git tag version version patch ensure that issues from merged commits are closed and mapped to a milestone write an update in the product team slack channel summarising development and identifying any blockers the script can be used to get a breakdown of the issues the release engineer is to update this every week until the version is released releasing release engineer once all issues have been merged into master then the release process can start build a beta named beta by pushing a lightweight git tag e g git tag beta once the ci completes successfully notify the team by writing a message in the product team slack channel product team i’ve just created the beta tag please let me know if there’s any final update we need to make if all is good then in i will start the release thanks add release notes to the page create a new document for the release in the ensure all issues are in the gh milestone that they re correctly labelled in particular they have the right type ui ux if they change the ui and breaking change if appropriate and have human readable descriptions use to export the issues into our release note format manually document any known migration steps and known issues add a link to the new release page in the section assign the pr to the director of technology or a developer an sre to review and confirm the documentation on upgrade instructions and breaking changes is sufficient create a release in github from the release branch so it shows up under the with the naming convention this will create the git tag automatically ensure the release notes pr above is merged link to the release notes in the description of the release confirm the release build completes successfully and the new release is available on the make sure that the document has new entry with id medic medic announce the release on the under the product releases category using this template announcing the release of version this release fixes number of bugs read the url for full details mark this issue done and close the milestone
1
258,128
27,563,857,841
IssuesEvent
2023-03-08 01:11:28
LynRodWS/alcor
https://api.github.com/repos/LynRodWS/alcor
opened
CVE-2018-17196 (High) detected in kafka-clients-2.0.1.jar
security vulnerability
## CVE-2018-17196 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kafka-clients-2.0.1.jar</b></p></summary> <p></p> <p>Library home page: <a href="http://kafka.apache.org">http://kafka.apache.org</a></p> <p>Path to dependency file: /services/vpc_manager/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/kafka/kafka-clients/2.0.1/kafka-clients-2.0.1.jar,/home/wss-scanner/.m2/repository/org/apache/kafka/kafka-clients/2.0.1/kafka-clients-2.0.1.jar</p> <p> Dependency Hierarchy: - common-0.1.0-SNAPSHOT.jar (Root Library) - :x: **kafka-clients-2.0.1.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Kafka versions between 0.11.0.0 and 2.1.0, it is possible to manually craft a Produce request which bypasses transaction/idempotent ACL validation. Only authenticated clients with Write permission on the respective topics are able to exploit this vulnerability. Users should upgrade to 2.1.1 or later where this vulnerability has been fixed. <p>Publish Date: 2019-07-11 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-17196>CVE-2018-17196</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-17196">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-17196</a></p> <p>Release Date: 2019-07-11</p> <p>Fix Resolution: org.apache.kafka:kafka-clients:2.1.1;org.apache.servicemix.bundles:org.apache.servicemix.bundles.kafka-clients:2.1.1_1</p> </p> </details> <p></p>
True
CVE-2018-17196 (High) detected in kafka-clients-2.0.1.jar - ## CVE-2018-17196 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kafka-clients-2.0.1.jar</b></p></summary> <p></p> <p>Library home page: <a href="http://kafka.apache.org">http://kafka.apache.org</a></p> <p>Path to dependency file: /services/vpc_manager/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/kafka/kafka-clients/2.0.1/kafka-clients-2.0.1.jar,/home/wss-scanner/.m2/repository/org/apache/kafka/kafka-clients/2.0.1/kafka-clients-2.0.1.jar</p> <p> Dependency Hierarchy: - common-0.1.0-SNAPSHOT.jar (Root Library) - :x: **kafka-clients-2.0.1.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Kafka versions between 0.11.0.0 and 2.1.0, it is possible to manually craft a Produce request which bypasses transaction/idempotent ACL validation. Only authenticated clients with Write permission on the respective topics are able to exploit this vulnerability. Users should upgrade to 2.1.1 or later where this vulnerability has been fixed. <p>Publish Date: 2019-07-11 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-17196>CVE-2018-17196</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-17196">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-17196</a></p> <p>Release Date: 2019-07-11</p> <p>Fix Resolution: org.apache.kafka:kafka-clients:2.1.1;org.apache.servicemix.bundles:org.apache.servicemix.bundles.kafka-clients:2.1.1_1</p> </p> </details> <p></p>
non_process
cve high detected in kafka clients jar cve high severity vulnerability vulnerable library kafka clients jar library home page a href path to dependency file services vpc manager pom xml path to vulnerable library home wss scanner repository org apache kafka kafka clients kafka clients jar home wss scanner repository org apache kafka kafka clients kafka clients jar dependency hierarchy common snapshot jar root library x kafka clients jar vulnerable library found in base branch master vulnerability details in apache kafka versions between and it is possible to manually craft a produce request which bypasses transaction idempotent acl validation only authenticated clients with write permission on the respective topics are able to exploit this vulnerability users should upgrade to or later where this vulnerability has been fixed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache kafka kafka clients org apache servicemix bundles org apache servicemix bundles kafka clients
0
714,718
24,571,741,545
IssuesEvent
2022-10-13 09:10:36
mozilla/addons-linter
https://api.github.com/repos/mozilla/addons-linter
opened
Consider simplifying the schema files
component: schema priority: p3 neverstale
Once https://github.com/mozilla/addons-linter/issues/4511 is done, we'll end up with a schema file per manifest version (so that should be one for mv2 and one for mv3). Let's see if we can change its structure to (1) remove useless information and (2) make it easier to consume for the linter.
1.0
Consider simplifying the schema files - Once https://github.com/mozilla/addons-linter/issues/4511 is done, we'll end up with a schema file per manifest version (so that should be one for mv2 and one for mv3). Let's see if we can change its structure to (1) remove useless information and (2) make it easier to consume for the linter.
non_process
consider simplifying the schema files once is done we ll end up with a schema file per manifest version so that should be one for and one for let s see if we can change its structure to remove useless information and make it easier to consume for the linter
0
7,267
10,421,338,005
IssuesEvent
2019-09-16 05:43:38
axa-group/Parsr
https://api.github.com/repos/axa-group/Parsr
closed
python3 command is just python sometimes
bug processing
**Summary** On Archlinux and some other OS (maybe Windows as well?), python 3.x is ran using the `python` command and python 2.x using the `python2` command. Here, we're assuming it's always `python3`: https://github.com/axa-group/Parsr/blob/dc3bec7ae62b0db12a075f31c26ba8951a63c2f7/server/src/modules/TableDetectionModule/TableDetectionModule.ts#L35 **Steps To Reproduce** Steps to reproduce the behavior: 1. Have your path with `python` pointing on Python 3.x 2. Remove `python3` from your path. **Expected behavior** Table extraction should run. **Actual behavior** A node error: `Error: spawnSync python3 ENOENT`. **Environment** - Reference commit/version: `dc3bec7ae62b0db12a075f31c26ba8951a63c2f7` - Other platform details: Python 3.x is just `python`, not `python3` - OS: Arch Linux, maybe others?
1.0
python3 command is just python sometimes - **Summary** On Archlinux and some other OS (maybe Windows as well?), python 3.x is ran using the `python` command and python 2.x using the `python2` command. Here, we're assuming it's always `python3`: https://github.com/axa-group/Parsr/blob/dc3bec7ae62b0db12a075f31c26ba8951a63c2f7/server/src/modules/TableDetectionModule/TableDetectionModule.ts#L35 **Steps To Reproduce** Steps to reproduce the behavior: 1. Have your path with `python` pointing on Python 3.x 2. Remove `python3` from your path. **Expected behavior** Table extraction should run. **Actual behavior** A node error: `Error: spawnSync python3 ENOENT`. **Environment** - Reference commit/version: `dc3bec7ae62b0db12a075f31c26ba8951a63c2f7` - Other platform details: Python 3.x is just `python`, not `python3` - OS: Arch Linux, maybe others?
process
command is just python sometimes summary on archlinux and some other os maybe windows as well python x is ran using the python command and python x using the command here we re assuming it s always steps to reproduce steps to reproduce the behavior have your path with python pointing on python x remove from your path expected behavior table extraction should run actual behavior a node error error spawnsync enoent environment reference commit version other platform details python x is just python not os arch linux maybe others
1
228,761
25,253,462,051
IssuesEvent
2022-11-15 16:14:48
vital-ws/splunk-test
https://api.github.com/repos/vital-ws/splunk-test
opened
lodash-3.0.0.tgz: 7 vulnerabilities (highest severity is: 9.1)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.0.0.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (lodash version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2019-10744](https://www.mend.io/vulnerability-database/CVE-2019-10744) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | lodash-3.0.0.tgz | Direct | 4.17.12 | &#9989; | | [CVE-2020-8203](https://www.mend.io/vulnerability-database/CVE-2020-8203) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.4 | lodash-3.0.0.tgz | Direct | 4.17.9 | &#9989; | | [CVE-2021-23337](https://www.mend.io/vulnerability-database/CVE-2021-23337) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.2 | lodash-3.0.0.tgz | Direct | 4.17.21 | &#9989; | | [CVE-2019-1010266](https://www.mend.io/vulnerability-database/CVE-2019-1010266) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | lodash-3.0.0.tgz | Direct | 4.17.11 | &#9989; | | [CVE-2018-3721](https://www.mend.io/vulnerability-database/CVE-2018-3721) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | lodash-3.0.0.tgz | Direct | 4.17.5 | &#9989; | | [CVE-2018-16487](https://www.mend.io/vulnerability-database/CVE-2018-16487) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.6 | lodash-3.0.0.tgz | Direct | 4.17.11 | &#9989; | | [CVE-2020-28500](https://www.mend.io/vulnerability-database/CVE-2020-28500) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | lodash-3.0.0.tgz | Direct | 4.17.21 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-10744</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Versions of lodash lower than 4.17.12 are vulnerable to Prototype Pollution. The function defaultsDeep could be tricked into adding or modifying properties of Object.prototype using a constructor payload. <p>Publish Date: 2019-07-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-10744>CVE-2019-10744</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-jf85-cpcp-j695">https://github.com/advisories/GHSA-jf85-cpcp-j695</a></p> <p>Release Date: 2019-07-26</p> <p>Fix Resolution: 4.17.12</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-8203</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-8203>CVE-2020-8203</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.4</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution: 4.17.9</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23337</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23337>CVE-2021-23337</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.2</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: 4.17.21</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-1010266</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11. <p>Publish Date: 2019-07-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-1010266>CVE-2019-1010266</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p> <p>Release Date: 2019-07-17</p> <p>Fix Resolution: 4.17.11</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-3721</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> lodash node module before 4.17.5 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via defaultsDeep, merge, and mergeWith functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-3721>CVE-2018-3721</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3721">https://nvd.nist.gov/vuln/detail/CVE-2018-3721</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 4.17.5</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-16487</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-16487>CVE-2018-16487</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.6</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution: 4.17.11</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-28500</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. Mend Note: After conducting further research, Mend has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-28500>CVE-2020-28500</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: 4.17.21</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
True
lodash-3.0.0.tgz: 7 vulnerabilities (highest severity is: 9.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.0.0.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (lodash version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2019-10744](https://www.mend.io/vulnerability-database/CVE-2019-10744) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | lodash-3.0.0.tgz | Direct | 4.17.12 | &#9989; | | [CVE-2020-8203](https://www.mend.io/vulnerability-database/CVE-2020-8203) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.4 | lodash-3.0.0.tgz | Direct | 4.17.9 | &#9989; | | [CVE-2021-23337](https://www.mend.io/vulnerability-database/CVE-2021-23337) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.2 | lodash-3.0.0.tgz | Direct | 4.17.21 | &#9989; | | [CVE-2019-1010266](https://www.mend.io/vulnerability-database/CVE-2019-1010266) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | lodash-3.0.0.tgz | Direct | 4.17.11 | &#9989; | | [CVE-2018-3721](https://www.mend.io/vulnerability-database/CVE-2018-3721) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | lodash-3.0.0.tgz | Direct | 4.17.5 | &#9989; | | [CVE-2018-16487](https://www.mend.io/vulnerability-database/CVE-2018-16487) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.6 | lodash-3.0.0.tgz | Direct | 4.17.11 | &#9989; | | [CVE-2020-28500](https://www.mend.io/vulnerability-database/CVE-2020-28500) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | lodash-3.0.0.tgz | Direct | 4.17.21 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-10744</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Versions of lodash lower than 4.17.12 are vulnerable to Prototype Pollution. The function defaultsDeep could be tricked into adding or modifying properties of Object.prototype using a constructor payload. <p>Publish Date: 2019-07-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-10744>CVE-2019-10744</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-jf85-cpcp-j695">https://github.com/advisories/GHSA-jf85-cpcp-j695</a></p> <p>Release Date: 2019-07-26</p> <p>Fix Resolution: 4.17.12</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-8203</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-8203>CVE-2020-8203</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.4</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution: 4.17.9</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23337</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23337>CVE-2021-23337</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.2</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: 4.17.21</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-1010266</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> lodash prior to 4.17.11 is affected by: CWE-400: Uncontrolled Resource Consumption. The impact is: Denial of service. The component is: Date handler. The attack vector is: Attacker provides very long strings, which the library attempts to match using a regular expression. The fixed version is: 4.17.11. <p>Publish Date: 2019-07-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-1010266>CVE-2019-1010266</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-1010266</a></p> <p>Release Date: 2019-07-17</p> <p>Fix Resolution: 4.17.11</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-3721</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> lodash node module before 4.17.5 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via defaultsDeep, merge, and mergeWith functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-3721>CVE-2018-3721</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3721">https://nvd.nist.gov/vuln/detail/CVE-2018-3721</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 4.17.5</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-16487</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-16487>CVE-2018-16487</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.6</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution: 4.17.11</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-28500</summary> ### Vulnerable Library - <b>lodash-3.0.0.tgz</b></p> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vital-ws/splunk-test/commit/08e7ae040cab7db134a00d8b763c00dd8bc070a7">08e7ae040cab7db134a00d8b763c00dd8bc070a7</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. Mend Note: After conducting further research, Mend has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-28500>CVE-2020-28500</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: 4.17.21</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
non_process
lodash tgz vulnerabilities highest severity is vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in lodash version remediation available high lodash tgz direct high lodash tgz direct high lodash tgz direct medium lodash tgz direct medium lodash tgz direct medium lodash tgz direct medium lodash tgz direct details cve vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details versions of lodash lower than are vulnerable to prototype pollution the function defaultsdeep could be tricked into adding or modifying properties of object prototype using a constructor payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue cve vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details prototype pollution attack when using zipobjectdeep in lodash before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue cve vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution rescue worker helmet automatic remediation is available for this issue cve vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details lodash prior to is affected by cwe uncontrolled resource consumption the impact is denial of service the component is date handler the attack vector is attacker provides very long strings which the library attempts to match using a regular expression the fixed version is publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue cve vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details lodash node module before suffers from a modification of assumed immutable data maid vulnerability via defaultsdeep merge and mergewith functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue cve vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue cve vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions mend note after conducting further research mend has determined that cve only affects environments with versions to of lodash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
0
973
3,423,281,745
IssuesEvent
2015-12-09 05:10:58
MaretEngineering/MROV
https://api.github.com/repos/MaretEngineering/MROV
closed
Why are these inverted x and y??? -Ethan
Processing question
@Eclip5e5 <- lol ```Java // Joysticks! joy1Vec[1] = (int) joy1.getX(); joy1Vec[0] = (int) joy1.getY(); joy2Vec[1] = (int) joy2.getX(); joy2Vec[0] = (int) joy2.getY(); ``` at the top it says this ```Java // Notes on the controller: // - Axis are inverted, so we used negative multipliers // - The "x" axis is the y-axis irl, but the controller's wierd like that ```
1.0
Why are these inverted x and y??? -Ethan - @Eclip5e5 <- lol ```Java // Joysticks! joy1Vec[1] = (int) joy1.getX(); joy1Vec[0] = (int) joy1.getY(); joy2Vec[1] = (int) joy2.getX(); joy2Vec[0] = (int) joy2.getY(); ``` at the top it says this ```Java // Notes on the controller: // - Axis are inverted, so we used negative multipliers // - The "x" axis is the y-axis irl, but the controller's wierd like that ```
process
why are these inverted x and y ethan lol java joysticks int getx int gety int getx int gety at the top it says this java notes on the controller axis are inverted so we used negative multipliers the x axis is the y axis irl but the controller s wierd like that
1
4,165
7,107,918,934
IssuesEvent
2018-01-16 21:45:53
18F/product-guide
https://api.github.com/repos/18F/product-guide
closed
SECTION UPDATE (Research) - Mindmaps
help wanted process change question
Does 18F use mindmaps during early project discovery? If this is a useful tool to us, we should add it to our user research deliverables perhaps. @porta-antiporta has created a mindmap that we could use as an example (not sure where it is documented, but more info about that [here](https://18f.slack.com/archives/product/p1453925867000077)). [slack convo for background](https://18f.slack.com/archives/product/p1453846341000024)
1.0
SECTION UPDATE (Research) - Mindmaps - Does 18F use mindmaps during early project discovery? If this is a useful tool to us, we should add it to our user research deliverables perhaps. @porta-antiporta has created a mindmap that we could use as an example (not sure where it is documented, but more info about that [here](https://18f.slack.com/archives/product/p1453925867000077)). [slack convo for background](https://18f.slack.com/archives/product/p1453846341000024)
process
section update research mindmaps does use mindmaps during early project discovery if this is a useful tool to us we should add it to our user research deliverables perhaps porta antiporta has created a mindmap that we could use as an example not sure where it is documented but more info about that
1
1,693
4,344,831,139
IssuesEvent
2016-07-29 09:56:53
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
reopened
Add contributed data to our database
3. In Development Processors
# Goal We had a contribution of ~100 CSRs and some other documents. We need to get them all, upload individually to our S3 instance, and create the related entry in the `documents` table in the API database. We should also document this process, as we'll be doing it more in the future. This is blocked on #254 # Analysis This documents are stored on s3 with link in `exporerdb`. Documents are in zip format. The process could be fully automated but the processor should includes specific to this contribution data like filename mapping etc. Initial implementation could be partially ad-hoc because we need to upload this documents ASAP. # Development - [x] create new `contrib_johnson` processor - [x] add `explorerdb` credentials to env vars and `config.py` - [x] add s3 credential to env vars and `config.py` - [x] in code download archive and extract it - [x] in code write files metadata to `database.documents` table - [x] create mapping between filenames in zip archive and metadata for `documents` table - [x] in code upload files from zip to s3 Estimated time: 2d
1.0
Add contributed data to our database - # Goal We had a contribution of ~100 CSRs and some other documents. We need to get them all, upload individually to our S3 instance, and create the related entry in the `documents` table in the API database. We should also document this process, as we'll be doing it more in the future. This is blocked on #254 # Analysis This documents are stored on s3 with link in `exporerdb`. Documents are in zip format. The process could be fully automated but the processor should includes specific to this contribution data like filename mapping etc. Initial implementation could be partially ad-hoc because we need to upload this documents ASAP. # Development - [x] create new `contrib_johnson` processor - [x] add `explorerdb` credentials to env vars and `config.py` - [x] add s3 credential to env vars and `config.py` - [x] in code download archive and extract it - [x] in code write files metadata to `database.documents` table - [x] create mapping between filenames in zip archive and metadata for `documents` table - [x] in code upload files from zip to s3 Estimated time: 2d
process
add contributed data to our database goal we had a contribution of csrs and some other documents we need to get them all upload individually to our instance and create the related entry in the documents table in the api database we should also document this process as we ll be doing it more in the future this is blocked on analysis this documents are stored on with link in exporerdb documents are in zip format the process could be fully automated but the processor should includes specific to this contribution data like filename mapping etc initial implementation could be partially ad hoc because we need to upload this documents asap development create new contrib johnson processor add explorerdb credentials to env vars and config py add credential to env vars and config py in code download archive and extract it in code write files metadata to database documents table create mapping between filenames in zip archive and metadata for documents table in code upload files from zip to estimated time
1
14,434
17,484,433,561
IssuesEvent
2021-08-09 09:07:10
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
Wrong JzCzhz blending results on CPU path
priority: high scope: image processing bug: pending
**Did you buy darktable from an application store ?** No. **Describe the bug/issue** Mask blending on CPU produces wrong results when JzCzhz blending is used. GPU path works correctly. E.g. selecting a hz range based on a colorful area of the image, that area isn't included in the mask. I found the root cause: `blendif_rgb_jzczhz.c` uses a "hacked" profile with modified `matrix_out` here: https://github.com/darktable-org/darktable/blob/40421d57d99c2b25f57898cd451b1bb357d10fa9/src/develop/blends/blendif_rgb_jzczhz.c#L150 However since https://github.com/darktable-org/darktable/commit/6ba7a7e47b131e5d30da4f78b226b822ffbcdec3 `matrix_out_transposed` is used instead. However that matrix is not initialized here: https://github.com/darktable-org/darktable/blob/40421d57d99c2b25f57898cd451b1bb357d10fa9/src/develop/blend.c#L219-L228 GPU code uses `matrix_out` directly, hence it works correctly. @ralfbrown, would you mind having a look at this? I can also post a simple fix, adding the initialization of `matrix_out_transposed` there but I feel you might come up with a cleaner fix. :)
1.0
Wrong JzCzhz blending results on CPU path - **Did you buy darktable from an application store ?** No. **Describe the bug/issue** Mask blending on CPU produces wrong results when JzCzhz blending is used. GPU path works correctly. E.g. selecting a hz range based on a colorful area of the image, that area isn't included in the mask. I found the root cause: `blendif_rgb_jzczhz.c` uses a "hacked" profile with modified `matrix_out` here: https://github.com/darktable-org/darktable/blob/40421d57d99c2b25f57898cd451b1bb357d10fa9/src/develop/blends/blendif_rgb_jzczhz.c#L150 However since https://github.com/darktable-org/darktable/commit/6ba7a7e47b131e5d30da4f78b226b822ffbcdec3 `matrix_out_transposed` is used instead. However that matrix is not initialized here: https://github.com/darktable-org/darktable/blob/40421d57d99c2b25f57898cd451b1bb357d10fa9/src/develop/blend.c#L219-L228 GPU code uses `matrix_out` directly, hence it works correctly. @ralfbrown, would you mind having a look at this? I can also post a simple fix, adding the initialization of `matrix_out_transposed` there but I feel you might come up with a cleaner fix. :)
process
wrong jzczhz blending results on cpu path did you buy darktable from an application store no describe the bug issue mask blending on cpu produces wrong results when jzczhz blending is used gpu path works correctly e g selecting a hz range based on a colorful area of the image that area isn t included in the mask i found the root cause blendif rgb jzczhz c uses a hacked profile with modified matrix out here however since matrix out transposed is used instead however that matrix is not initialized here gpu code uses matrix out directly hence it works correctly ralfbrown would you mind having a look at this i can also post a simple fix adding the initialization of matrix out transposed there but i feel you might come up with a cleaner fix
1
10,749
13,542,101,418
IssuesEvent
2020-09-16 16:49:05
nion-software/nionswift
https://api.github.com/repos/nion-software/nionswift
opened
Line profile and others performed on navigable data should optionally produce navigable data
f - computations f - processing type - enhancement
The line profile currently takes the indexed element and performs the line profile on just that element. Along with other mapping techniques, the user should have the option to produce a sequence or collection of 1D data as the result of the line profile operation. See #430 map 1d/2d operations to sequences/collections.
1.0
Line profile and others performed on navigable data should optionally produce navigable data - The line profile currently takes the indexed element and performs the line profile on just that element. Along with other mapping techniques, the user should have the option to produce a sequence or collection of 1D data as the result of the line profile operation. See #430 map 1d/2d operations to sequences/collections.
process
line profile and others performed on navigable data should optionally produce navigable data the line profile currently takes the indexed element and performs the line profile on just that element along with other mapping techniques the user should have the option to produce a sequence or collection of data as the result of the line profile operation see map operations to sequences collections
1
29,069
7,051,611,469
IssuesEvent
2018-01-03 12:37:51
bunq/sdk_java
https://api.github.com/repos/bunq/sdk_java
closed
Token request ideal is missing id attribute in response.
bug needs code regeneration pending review by andré
## Steps to reproduce: 1. Take a look at https://github.com/bunq/sdk_java/blob/dbc455fa50375a0d7771d00806628dc0d8c88ee7/src/main/java/com/bunq/sdk/model/generated/endpoint/TokenQrRequestIdeal.java#L32 ## What should happen: 1. There is an `id` field that holds the id of this request token. ## What happens: 1. The `id` field is missing ## Traceback [//]: # (If there is a traceback please share it in a quote! You can do this by pasting the traceback text, highlighting it and pressing the quote button.) ## SDK version and environment - Tested on [0.12.4](https://github.com/bunq/sdk_java/releases/tag/0.12.4) - [x] Sandbox - [x] Production ## Response id [//]: # (If this error has something to do with a request that fails, please provide the response id of the request.) - Response id: ## Extra info: [//]: # (Please provide any other relevant information here) - bunq/sdk_php#97
1.0
Token request ideal is missing id attribute in response. - ## Steps to reproduce: 1. Take a look at https://github.com/bunq/sdk_java/blob/dbc455fa50375a0d7771d00806628dc0d8c88ee7/src/main/java/com/bunq/sdk/model/generated/endpoint/TokenQrRequestIdeal.java#L32 ## What should happen: 1. There is an `id` field that holds the id of this request token. ## What happens: 1. The `id` field is missing ## Traceback [//]: # (If there is a traceback please share it in a quote! You can do this by pasting the traceback text, highlighting it and pressing the quote button.) ## SDK version and environment - Tested on [0.12.4](https://github.com/bunq/sdk_java/releases/tag/0.12.4) - [x] Sandbox - [x] Production ## Response id [//]: # (If this error has something to do with a request that fails, please provide the response id of the request.) - Response id: ## Extra info: [//]: # (Please provide any other relevant information here) - bunq/sdk_php#97
non_process
token request ideal is missing id attribute in response steps to reproduce take a look at what should happen there is an id field that holds the id of this request token what happens the id field is missing traceback if there is a traceback please share it in a quote you can do this by pasting the traceback text highlighting it and pressing the quote button sdk version and environment tested on sandbox production response id if this error has something to do with a request that fails please provide the response id of the request response id extra info please provide any other relevant information here bunq sdk php
0
20,216
26,806,412,319
IssuesEvent
2023-02-01 18:39:44
mmattDonk/AI-TTS-Donations
https://api.github.com/repos/mmattDonk/AI-TTS-Donations
closed
[SOL-10] testing client for viewers
💫 feature_request @solrock/processor @solrock/frontend frontend processor Medium priority
makes it so that it just polls an external api (one made by me) so that the viewer doesn't need any credentials. barebones ai-tts-donations with only the testing interface
2.0
[SOL-10] testing client for viewers - makes it so that it just polls an external api (one made by me) so that the viewer doesn't need any credentials. barebones ai-tts-donations with only the testing interface
process
testing client for viewers makes it so that it just polls an external api one made by me so that the viewer doesn t need any credentials barebones ai tts donations with only the testing interface
1
7,643
7,036,691,072
IssuesEvent
2017-12-28 10:16:38
kingalois/friedmanng
https://api.github.com/repos/kingalois/friedmanng
closed
install development environment (IDE) for java
Alois Florian infrastructure Jonas
examples are - eclipse - netbeans article whith ide's : https://blog.idrsolutions.com/2015/03/the-top-11-free-ide-for-java-coding-development-programming/
1.0
install development environment (IDE) for java - examples are - eclipse - netbeans article whith ide's : https://blog.idrsolutions.com/2015/03/the-top-11-free-ide-for-java-coding-development-programming/
non_process
install development environment ide for java examples are eclipse netbeans article whith ide s
0
175,530
6,551,787,880
IssuesEvent
2017-09-05 15:51:30
opencurrents/opencurrents
https://api.github.com/repos/opencurrents/opencurrents
opened
Email Confirmation of Volunteer: Message could be improved telling the coordinator what event this is.
priority low
![event confirmation email to admin](https://user-images.githubusercontent.com/26234440/30070124-1be3e4be-9228-11e7-8af5-35c3cfadf5c5.PNG)
1.0
Email Confirmation of Volunteer: Message could be improved telling the coordinator what event this is. - ![event confirmation email to admin](https://user-images.githubusercontent.com/26234440/30070124-1be3e4be-9228-11e7-8af5-35c3cfadf5c5.PNG)
non_process
email confirmation of volunteer message could be improved telling the coordinator what event this is
0
261
2,689,478,477
IssuesEvent
2015-03-31 10:30:36
cs2103jan2015-f10-3c/main
https://api.github.com/repos/cs2103jan2015-f10-3c/main
closed
search algorithm that support more than 1 word (also migrate from data processor to data storage
component.DataProcessor component.DataStorage priority.high type.task
@kevin-christian may need your help! :)
1.0
search algorithm that support more than 1 word (also migrate from data processor to data storage - @kevin-christian may need your help! :)
process
search algorithm that support more than word also migrate from data processor to data storage kevin christian may need your help
1
15,705
19,848,430,073
IssuesEvent
2022-01-21 09:33:08
ooi-data/CE09OSPM-WFP01-03-CTDPFK000-recovered_wfp-ctdpf_ckl_wfp_instrument_recovered
https://api.github.com/repos/ooi-data/CE09OSPM-WFP01-03-CTDPFK000-recovered_wfp-ctdpf_ckl_wfp_instrument_recovered
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:33:07.904915. ## Details Flow name: `CE09OSPM-WFP01-03-CTDPFK000-recovered_wfp-ctdpf_ckl_wfp_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:33:07.904915. ## Details Flow name: `CE09OSPM-WFP01-03-CTDPFK000-recovered_wfp-ctdpf_ckl_wfp_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered wfp ctdpf ckl wfp instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
4,321
7,227,792,645
IssuesEvent
2018-02-11 00:50:41
bartop/tpl
https://api.github.com/repos/bartop/tpl
closed
Add Visual Studio 2015 CI in AppVeyor
delayed new feature process upgrade waiting for merge
As far as I know the code should be compilable under VS2015 - and so CI should also cover this compiler/ide.
1.0
Add Visual Studio 2015 CI in AppVeyor - As far as I know the code should be compilable under VS2015 - and so CI should also cover this compiler/ide.
process
add visual studio ci in appveyor as far as i know the code should be compilable under and so ci should also cover this compiler ide
1
6,153
9,036,573,996
IssuesEvent
2019-02-09 01:11:38
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Format Date to month always shows 1 January
Bug Database/SQLServer Priority/P1 Query Processor
Hi all, i'm using Metabase 0.31.1 on a Data Warehouse built on SQL Server 2017. I have a table like this, in which i have on a row level (among other fields), day (DataCompleta), first day of month (PrimoGiornoMese) and first day of year (PrimoGiornoAnno): ![image](https://user-images.githubusercontent.com/36055023/49298489-d6877d00-f4bd-11e8-81ff-63973259bd84.png) What happens is when i choose to aggregate by the first field (DataCompleta) by month, to leverage Metabase datetime formatting (I'm working on this table as a test), every row of that column gets formatted as January ![image](https://user-images.githubusercontent.com/36055023/49298654-40078b80-f4be-11e8-904e-79a481378702.png) The SQL Code shows that the convert function is using `YYYY-MM-DD` instead of `YYYYMMDD` (my SQL Server is Italian, and it appears that if i convert to `YYYY-MM-DD` SQL Server is inverting days with months, while if i use `YYYYMMDD` is working correctly). I tried to change the server language to US English using: `EXEC sp_configure 'default language', 0 ; RECONFIGURE;` but still I can't make it work as I intend to. Has anyone experienced this? I forgot to mention that I modified the field metadata, to show YYYY/M/D in the admin page, but apparently it's not related. Thanks a lot, Andrea
1.0
Format Date to month always shows 1 January - Hi all, i'm using Metabase 0.31.1 on a Data Warehouse built on SQL Server 2017. I have a table like this, in which i have on a row level (among other fields), day (DataCompleta), first day of month (PrimoGiornoMese) and first day of year (PrimoGiornoAnno): ![image](https://user-images.githubusercontent.com/36055023/49298489-d6877d00-f4bd-11e8-81ff-63973259bd84.png) What happens is when i choose to aggregate by the first field (DataCompleta) by month, to leverage Metabase datetime formatting (I'm working on this table as a test), every row of that column gets formatted as January ![image](https://user-images.githubusercontent.com/36055023/49298654-40078b80-f4be-11e8-904e-79a481378702.png) The SQL Code shows that the convert function is using `YYYY-MM-DD` instead of `YYYYMMDD` (my SQL Server is Italian, and it appears that if i convert to `YYYY-MM-DD` SQL Server is inverting days with months, while if i use `YYYYMMDD` is working correctly). I tried to change the server language to US English using: `EXEC sp_configure 'default language', 0 ; RECONFIGURE;` but still I can't make it work as I intend to. Has anyone experienced this? I forgot to mention that I modified the field metadata, to show YYYY/M/D in the admin page, but apparently it's not related. Thanks a lot, Andrea
process
format date to month always shows january hi all i m using metabase on a data warehouse built on sql server i have a table like this in which i have on a row level among other fields day datacompleta first day of month primogiornomese and first day of year primogiornoanno what happens is when i choose to aggregate by the first field datacompleta by month to leverage metabase datetime formatting i m working on this table as a test every row of that column gets formatted as january the sql code shows that the convert function is using yyyy mm dd instead of yyyymmdd my sql server is italian and it appears that if i convert to yyyy mm dd sql server is inverting days with months while if i use yyyymmdd is working correctly i tried to change the server language to us english using exec sp configure default language reconfigure but still i can t make it work as i intend to has anyone experienced this i forgot to mention that i modified the field metadata to show yyyy m d in the admin page but apparently it s not related thanks a lot andrea
1
12,745
15,107,042,233
IssuesEvent
2021-02-08 14:59:33
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Prisma format and issues with implicit relations
process/candidate topic: prisma-format
When we validate the data model, we make certain implicit assumptions on how relations should work. We can demonstrate it with a simple data model: ```prisma datasource chinook { provider = "sqlserver" url = "jdbc:sqlserver://" } model Code { id String @id userId String? user User? } model User { id String @id name String? codes Code[] } ``` Here we can format the code with `cat dm.prisma | cargo run --bin prisma-fmt -- format`, which outputs the correct relation attributes: ```prisma datasource chinook { provider = "sqlserver" url = "jdbc:sqlserver://" } model Code { id String @id userId String? user User? @relation(fields: [userId], references: [id]) } model User { id String @id name String? codes Code[] } ``` Or we can remove the `userId` from our datamodel, and the formatter will add it back. We'll get into trouble though when we have two relations to the same model: ```prisma datasource chinook { provider = "sqlserver" url = "jdbc:sqlserver://" } model Code { id String @id user User? createdBy User? userId String? createdById String? } model User { id String @id name String? codes Code[] } ``` Running this through `prisma-fmt` will give the exact same model back. The relation columns are not added, but the data model passes the validation just to result in broken migrations later on.
1.0
Prisma format and issues with implicit relations - When we validate the data model, we make certain implicit assumptions on how relations should work. We can demonstrate it with a simple data model: ```prisma datasource chinook { provider = "sqlserver" url = "jdbc:sqlserver://" } model Code { id String @id userId String? user User? } model User { id String @id name String? codes Code[] } ``` Here we can format the code with `cat dm.prisma | cargo run --bin prisma-fmt -- format`, which outputs the correct relation attributes: ```prisma datasource chinook { provider = "sqlserver" url = "jdbc:sqlserver://" } model Code { id String @id userId String? user User? @relation(fields: [userId], references: [id]) } model User { id String @id name String? codes Code[] } ``` Or we can remove the `userId` from our datamodel, and the formatter will add it back. We'll get into trouble though when we have two relations to the same model: ```prisma datasource chinook { provider = "sqlserver" url = "jdbc:sqlserver://" } model Code { id String @id user User? createdBy User? userId String? createdById String? } model User { id String @id name String? codes Code[] } ``` Running this through `prisma-fmt` will give the exact same model back. The relation columns are not added, but the data model passes the validation just to result in broken migrations later on.
process
prisma format and issues with implicit relations when we validate the data model we make certain implicit assumptions on how relations should work we can demonstrate it with a simple data model prisma datasource chinook provider sqlserver url jdbc sqlserver model code id string id userid string user user model user id string id name string codes code here we can format the code with cat dm prisma cargo run bin prisma fmt format which outputs the correct relation attributes prisma datasource chinook provider sqlserver url jdbc sqlserver model code id string id userid string user user relation fields references model user id string id name string codes code or we can remove the userid from our datamodel and the formatter will add it back we ll get into trouble though when we have two relations to the same model prisma datasource chinook provider sqlserver url jdbc sqlserver model code id string id user user createdby user userid string createdbyid string model user id string id name string codes code running this through prisma fmt will give the exact same model back the relation columns are not added but the data model passes the validation just to result in broken migrations later on
1
1,323
3,874,105,092
IssuesEvent
2016-04-11 19:19:07
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
closed
harmonize use of -vert and -t flag
sct_extract_metric sct_process_segmentation sct_segment_graymatter
- [x] sct_segment_graymatter - [ ] sct_extract_metric --> add -vertfile? - [x] sct_process_segmentation see: https://github.com/neuropoly/spinalcordtoolbox/wiki/flags
1.0
harmonize use of -vert and -t flag - - [x] sct_segment_graymatter - [ ] sct_extract_metric --> add -vertfile? - [x] sct_process_segmentation see: https://github.com/neuropoly/spinalcordtoolbox/wiki/flags
process
harmonize use of vert and t flag sct segment graymatter sct extract metric add vertfile sct process segmentation see
1
59,361
17,023,099,340
IssuesEvent
2021-07-03 00:22:33
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Edit tab behaviour when not logged in
Component: website Priority: trivial Resolution: duplicate Type: defect
**[Submitted to the original trac issue database at 3.49pm, Monday, 27th February 2006]** On the viewer, when not logged in, the 'edit' tab is greyed out. Some nice easy minor tweaks to this: Make the font colour grey too to make it look more disabled. Make a mouse-over tool-tip saying "You must be logged-in to edit" (just a TITLE attribute in the HTML tag) If you the user still clicks it, send them over to the login page. (p.s. I tried to add this before, but I'm new to 'trac'. Sorry if I've added it twice)
1.0
Edit tab behaviour when not logged in - **[Submitted to the original trac issue database at 3.49pm, Monday, 27th February 2006]** On the viewer, when not logged in, the 'edit' tab is greyed out. Some nice easy minor tweaks to this: Make the font colour grey too to make it look more disabled. Make a mouse-over tool-tip saying "You must be logged-in to edit" (just a TITLE attribute in the HTML tag) If you the user still clicks it, send them over to the login page. (p.s. I tried to add this before, but I'm new to 'trac'. Sorry if I've added it twice)
non_process
edit tab behaviour when not logged in on the viewer when not logged in the edit tab is greyed out some nice easy minor tweaks to this make the font colour grey too to make it look more disabled make a mouse over tool tip saying you must be logged in to edit just a title attribute in the html tag if you the user still clicks it send them over to the login page p s i tried to add this before but i m new to trac sorry if i ve added it twice
0
68,124
3,284,260,677
IssuesEvent
2015-10-28 15:59:14
CenterForOpenScience/osf.io
https://api.github.com/repos/CenterForOpenScience/osf.io
closed
View-only link names display escaped HTML characters: & < >
2 - ready bug: production priority - medium
Reported on the Markup Safe 3 board by @jinluyuan # Steps 1. Go to the Sharing page of any of your projects 2. Click the Add button for View-only links 3. Name your link with the characters `& < >` ![](https://trello-attachments.s3.amazonaws.com/55c2395a385f787cf2c86df6/600x479/d4df4f875c63aef8f8f2bd2426ebdfad/Screen_Shot_2015-08-05_at_12.27.31_PM.png) # Expected behavior I expected that under the View-only Links header, `& < >` would display properly as a part of the name of the link. # Actual behavior ![](https://trello-attachments.s3.amazonaws.com/55c2395a385f787cf2c86df6/600x120/a33fb54038d968abf996256de821853f/Screen_Shot_2015-08-05_at_12.27.55_PM.png) ## Comments @abought "This is one of those (many) cases where knockout is faithfully rendering the data as it appears in the database."
1.0
View-only link names display escaped HTML characters: & < > - Reported on the Markup Safe 3 board by @jinluyuan # Steps 1. Go to the Sharing page of any of your projects 2. Click the Add button for View-only links 3. Name your link with the characters `& < >` ![](https://trello-attachments.s3.amazonaws.com/55c2395a385f787cf2c86df6/600x479/d4df4f875c63aef8f8f2bd2426ebdfad/Screen_Shot_2015-08-05_at_12.27.31_PM.png) # Expected behavior I expected that under the View-only Links header, `& < >` would display properly as a part of the name of the link. # Actual behavior ![](https://trello-attachments.s3.amazonaws.com/55c2395a385f787cf2c86df6/600x120/a33fb54038d968abf996256de821853f/Screen_Shot_2015-08-05_at_12.27.55_PM.png) ## Comments @abought "This is one of those (many) cases where knockout is faithfully rendering the data as it appears in the database."
non_process
view only link names display escaped html characters reported on the markup safe board by jinluyuan steps go to the sharing page of any of your projects click the add button for view only links name your link with the characters expected behavior i expected that under the view only links header would display properly as a part of the name of the link actual behavior comments abought this is one of those many cases where knockout is faithfully rendering the data as it appears in the database
0
370,646
10,935,039,564
IssuesEvent
2019-11-24 15:42:57
dave-ok/devc-teamwork-rest-api
https://api.github.com/repos/dave-ok/devc-teamwork-rest-api
closed
[chore] - Setup databases and models
blocker chore priority: high
### Description Create database migrations and models ### Acceptance Criteria - [x] 1. Create custom database migrations (with tests) for verbosity - [x] 2. Create data models with CRUD functionality and data validation - [x] 3. Run migrations to create local databases
1.0
[chore] - Setup databases and models - ### Description Create database migrations and models ### Acceptance Criteria - [x] 1. Create custom database migrations (with tests) for verbosity - [x] 2. Create data models with CRUD functionality and data validation - [x] 3. Run migrations to create local databases
non_process
setup databases and models description create database migrations and models acceptance criteria create custom database migrations with tests for verbosity create data models with crud functionality and data validation run migrations to create local databases
0
16,681
21,784,842,230
IssuesEvent
2022-05-14 01:31:06
nodejs/node
https://api.github.com/repos/nodejs/node
closed
`process`/`child_process`: be able to kill all descendent processes for a given process.
child_process feature request process stale
Hi 👋, Hope everyone is doing great. **Is your feature request related to a problem? Please describe.** I used `require("child_process").spawn` recently with the option `shell: true`. I observe in this case that two processes are spawned on for the shell itself and another one for the current command spawned by the shell. So when I use ChildProcess.kill(), it kills the shell process but not its child (descendent processes). (In my case the child process is an HTTP server, so a long process that doesn't exist itself") **Describe the solution you'd like** I thought about two different APIs: - First one (with the ChildProcess) ```js const { spawn } = require("child_process"); const p1 = spawn(httpServerCommand, { shell: true }); p1.killDeep(); ``` - Second one (with `process`) ```js const { spawn } = require("child_process"); const p1 = spawn(httpServerCommand, { shell: true }); process.killDeep(p1.pid); ``` In both cases `killDeep` will kill all the descendent processes. **Describe alternatives you've considered** I used this library: https://www.npmjs.com/package/tree-kill I'm sorry in advance if I miss something in the documentation (But I promise that I tried ahahah) If this feature request is acceptable, I could help and contribute (with a bit of guidance ^^) Thanks for reading me.
2.0
`process`/`child_process`: be able to kill all descendent processes for a given process. - Hi 👋, Hope everyone is doing great. **Is your feature request related to a problem? Please describe.** I used `require("child_process").spawn` recently with the option `shell: true`. I observe in this case that two processes are spawned on for the shell itself and another one for the current command spawned by the shell. So when I use ChildProcess.kill(), it kills the shell process but not its child (descendent processes). (In my case the child process is an HTTP server, so a long process that doesn't exist itself") **Describe the solution you'd like** I thought about two different APIs: - First one (with the ChildProcess) ```js const { spawn } = require("child_process"); const p1 = spawn(httpServerCommand, { shell: true }); p1.killDeep(); ``` - Second one (with `process`) ```js const { spawn } = require("child_process"); const p1 = spawn(httpServerCommand, { shell: true }); process.killDeep(p1.pid); ``` In both cases `killDeep` will kill all the descendent processes. **Describe alternatives you've considered** I used this library: https://www.npmjs.com/package/tree-kill I'm sorry in advance if I miss something in the documentation (But I promise that I tried ahahah) If this feature request is acceptable, I could help and contribute (with a bit of guidance ^^) Thanks for reading me.
process
process child process be able to kill all descendent processes for a given process hi 👋 hope everyone is doing great is your feature request related to a problem please describe i used require child process spawn recently with the option shell true i observe in this case that two processes are spawned on for the shell itself and another one for the current command spawned by the shell so when i use childprocess kill it kills the shell process but not its child descendent processes in my case the child process is an http server so a long process that doesn t exist itself describe the solution you d like i thought about two different apis first one with the childprocess js const spawn   require child process const spawn httpservercommand shell true killdeep second one with process js const spawn require child process const spawn httpservercommand shell true process killdeep pid in both cases killdeep will kill all the descendent processes describe alternatives you ve considered i used this library i m sorry in advance if i miss something in the documentation but i promise that i tried ahahah if this feature request is acceptable i could help and contribute with a bit of guidance thanks for reading me
1
13,064
15,395,205,690
IssuesEvent
2021-03-03 18:54:48
LLNL/axom
https://api.github.com/repos/LLNL/axom
closed
Convert uberenv to submodule
Reviewed Software process TPL
Now that `uberenv` has `axom`'s changes to support `vcpkg` (https://github.com/LLNL/uberenv/pull/40), we should convert our usage of uberenv from a copy of `uberenv.py` to using an `uberenv` submodule. This will make it easier to track changes and upgrade between our version and the main version of uberenv. See: * https://uberenv.readthedocs.io/en/latest/#uberenv-py * https://uberenv.readthedocs.io/en/latest/#project-configuration
1.0
Convert uberenv to submodule - Now that `uberenv` has `axom`'s changes to support `vcpkg` (https://github.com/LLNL/uberenv/pull/40), we should convert our usage of uberenv from a copy of `uberenv.py` to using an `uberenv` submodule. This will make it easier to track changes and upgrade between our version and the main version of uberenv. See: * https://uberenv.readthedocs.io/en/latest/#uberenv-py * https://uberenv.readthedocs.io/en/latest/#project-configuration
process
convert uberenv to submodule now that uberenv has axom s changes to support vcpkg we should convert our usage of uberenv from a copy of uberenv py to using an uberenv submodule this will make it easier to track changes and upgrade between our version and the main version of uberenv see
1
18,687
24,594,985,608
IssuesEvent
2022-10-14 07:31:27
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Studies > Study participant registry > Search bar placeholder text should be changed
Bug P2 Participant manager Process: Fixed Process: Tested dev
Actual: "Search by site ID or participant email" Expected: "Search by site name or participant email" Note: Functionality wise, search results are getting displayed if admin searches by site name and NOT site ID. Hence we have to specify name intead of ID in placeholder text ![Screenshot_3](https://user-images.githubusercontent.com/60386291/180806374-8f21c13e-5f71-447f-b30f-8495445b2898.png)
2.0
[PM] Studies > Study participant registry > Search bar placeholder text should be changed - Actual: "Search by site ID or participant email" Expected: "Search by site name or participant email" Note: Functionality wise, search results are getting displayed if admin searches by site name and NOT site ID. Hence we have to specify name intead of ID in placeholder text ![Screenshot_3](https://user-images.githubusercontent.com/60386291/180806374-8f21c13e-5f71-447f-b30f-8495445b2898.png)
process
studies study participant registry search bar placeholder text should be changed actual search by site id or participant email expected search by site name or participant email note functionality wise search results are getting displayed if admin searches by site name and not site id hence we have to specify name intead of id in placeholder text
1
206,366
23,380,022,099
IssuesEvent
2022-08-11 08:35:58
Mohammad-HGH/TODO-Django-React
https://api.github.com/repos/Mohammad-HGH/TODO-Django-React
opened
react-scripts-5.0.1.tgz: 2 vulnerabilities (highest severity is: 7.5)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>react-scripts-5.0.1.tgz</b></p></summary> <p></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/svgo/node_modules/nth-check/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Mohammad-HGH/TODO-Django-React/commit/3504794e0e5bd1ef8f008a1bdaaed932a60b6a50">3504794e0e5bd1ef8f008a1bdaaed932a60b6a50</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2021-33587](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33587) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | css-what-3.4.2.tgz | Transitive | N/A | &#10060; | | [CVE-2021-3803](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | nth-check-1.0.2.tgz | Transitive | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-33587</summary> ### Vulnerable Library - <b>css-what-3.4.2.tgz</b></p> <p>a CSS selector parser</p> <p>Library home page: <a href="https://registry.npmjs.org/css-what/-/css-what-3.4.2.tgz">https://registry.npmjs.org/css-what/-/css-what-3.4.2.tgz</a></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/svgo/node_modules/css-what/package.json</p> <p> Dependency Hierarchy: - react-scripts-5.0.1.tgz (Root Library) - webpack-5.5.0.tgz - plugin-svgo-5.5.0.tgz - svgo-1.3.2.tgz - css-select-2.1.0.tgz - :x: **css-what-3.4.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Mohammad-HGH/TODO-Django-React/commit/3504794e0e5bd1ef8f008a1bdaaed932a60b6a50">3504794e0e5bd1ef8f008a1bdaaed932a60b6a50</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The css-what package 4.0.0 through 5.0.0 for Node.js does not ensure that attribute parsing has Linear Time Complexity relative to the size of the input. <p>Publish Date: 2021-05-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33587>CVE-2021-33587</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33587">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33587</a></p> <p>Release Date: 2021-05-28</p> <p>Fix Resolution: css-what - 5.0.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-3803</summary> ### Vulnerable Library - <b>nth-check-1.0.2.tgz</b></p> <p>performant nth-check parser & compiler</p> <p>Library home page: <a href="https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz">https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz</a></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/svgo/node_modules/nth-check/package.json</p> <p> Dependency Hierarchy: - react-scripts-5.0.1.tgz (Root Library) - webpack-5.5.0.tgz - plugin-svgo-5.5.0.tgz - svgo-1.3.2.tgz - css-select-2.1.0.tgz - :x: **nth-check-1.0.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Mohammad-HGH/TODO-Django-React/commit/3504794e0e5bd1ef8f008a1bdaaed932a60b6a50">3504794e0e5bd1ef8f008a1bdaaed932a60b6a50</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> nth-check is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803>CVE-2021-3803</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: nth-check - v2.0.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
react-scripts-5.0.1.tgz: 2 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>react-scripts-5.0.1.tgz</b></p></summary> <p></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/svgo/node_modules/nth-check/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Mohammad-HGH/TODO-Django-React/commit/3504794e0e5bd1ef8f008a1bdaaed932a60b6a50">3504794e0e5bd1ef8f008a1bdaaed932a60b6a50</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2021-33587](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33587) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | css-what-3.4.2.tgz | Transitive | N/A | &#10060; | | [CVE-2021-3803](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | nth-check-1.0.2.tgz | Transitive | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-33587</summary> ### Vulnerable Library - <b>css-what-3.4.2.tgz</b></p> <p>a CSS selector parser</p> <p>Library home page: <a href="https://registry.npmjs.org/css-what/-/css-what-3.4.2.tgz">https://registry.npmjs.org/css-what/-/css-what-3.4.2.tgz</a></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/svgo/node_modules/css-what/package.json</p> <p> Dependency Hierarchy: - react-scripts-5.0.1.tgz (Root Library) - webpack-5.5.0.tgz - plugin-svgo-5.5.0.tgz - svgo-1.3.2.tgz - css-select-2.1.0.tgz - :x: **css-what-3.4.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Mohammad-HGH/TODO-Django-React/commit/3504794e0e5bd1ef8f008a1bdaaed932a60b6a50">3504794e0e5bd1ef8f008a1bdaaed932a60b6a50</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The css-what package 4.0.0 through 5.0.0 for Node.js does not ensure that attribute parsing has Linear Time Complexity relative to the size of the input. <p>Publish Date: 2021-05-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33587>CVE-2021-33587</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33587">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33587</a></p> <p>Release Date: 2021-05-28</p> <p>Fix Resolution: css-what - 5.0.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-3803</summary> ### Vulnerable Library - <b>nth-check-1.0.2.tgz</b></p> <p>performant nth-check parser & compiler</p> <p>Library home page: <a href="https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz">https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz</a></p> <p>Path to dependency file: /frontend/package.json</p> <p>Path to vulnerable library: /frontend/node_modules/svgo/node_modules/nth-check/package.json</p> <p> Dependency Hierarchy: - react-scripts-5.0.1.tgz (Root Library) - webpack-5.5.0.tgz - plugin-svgo-5.5.0.tgz - svgo-1.3.2.tgz - css-select-2.1.0.tgz - :x: **nth-check-1.0.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Mohammad-HGH/TODO-Django-React/commit/3504794e0e5bd1ef8f008a1bdaaed932a60b6a50">3504794e0e5bd1ef8f008a1bdaaed932a60b6a50</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> nth-check is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803>CVE-2021-3803</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: nth-check - v2.0.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_process
react scripts tgz vulnerabilities highest severity is vulnerable library react scripts tgz path to dependency file frontend package json path to vulnerable library frontend node modules svgo node modules nth check package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high css what tgz transitive n a high nth check tgz transitive n a details cve vulnerable library css what tgz a css selector parser library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules svgo node modules css what package json dependency hierarchy react scripts tgz root library webpack tgz plugin svgo tgz svgo tgz css select tgz x css what tgz vulnerable library found in head commit a href found in base branch main vulnerability details the css what package through for node js does not ensure that attribute parsing has linear time complexity relative to the size of the input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution css what step up your open source security game with mend cve vulnerable library nth check tgz performant nth check parser compiler library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules svgo node modules nth check package json dependency hierarchy react scripts tgz root library webpack tgz plugin svgo tgz svgo tgz css select tgz x nth check tgz vulnerable library found in head commit a href found in base branch main vulnerability details nth check is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution nth check step up your open source security game with mend
0
158,963
24,924,219,016
IssuesEvent
2022-10-31 05:16:25
SybooSyboo782/comproject-network
https://api.github.com/repos/SybooSyboo782/comproject-network
closed
Aggreagate 클래스 다이어그램
Design
**문제 제기** - 설계 **기능 요약** - Aggreagate 클래스 다이어그램 설계 **할일 목록** - [x] Snap-E 다이어그램 작성 - [x] Aggregate 클래스 다이어그램 작성 **비고**
1.0
Aggreagate 클래스 다이어그램 - **문제 제기** - 설계 **기능 요약** - Aggreagate 클래스 다이어그램 설계 **할일 목록** - [x] Snap-E 다이어그램 작성 - [x] Aggregate 클래스 다이어그램 작성 **비고**
non_process
aggreagate 클래스 다이어그램 문제 제기 설계 기능 요약 aggreagate 클래스 다이어그램 설계 할일 목록 snap e 다이어그램 작성 aggregate 클래스 다이어그램 작성 비고
0
14,269
17,225,449,056
IssuesEvent
2021-07-20 00:29:35
fluent/fluent-bit
https://api.github.com/repos/fluent/fluent-bit
closed
Join partial docker json-file lines
enhancement work-in-process
**Is your feature request related to a problem? Please describe.** <!--- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> Fluent-bit does not join log lines that are splitted as "partial" by docker daemon, when logging to json-file. https://github.com/moby/moby/blob/master/daemon/logger/jsonfilelog/jsonfilelog.go#L135 **Describe the solution you'd like** <!--- A clear and concise description of what you want to happen. --> Fluent-bit should join these lines, either in the parser or in a filter. **Describe alternatives you've considered** <!--- A clear and concise description of any alternative solutions or features you've considered. --> We currently have issues with this and don't see any good workarounds. **Additional context** <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> <!--- Add any other context or screenshots about the feature request here. --> Log events are sent to Splunk, splitted over multiple events and unparsed since they are not considered to be valid json when splitted.
1.0
Join partial docker json-file lines - **Is your feature request related to a problem? Please describe.** <!--- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> Fluent-bit does not join log lines that are splitted as "partial" by docker daemon, when logging to json-file. https://github.com/moby/moby/blob/master/daemon/logger/jsonfilelog/jsonfilelog.go#L135 **Describe the solution you'd like** <!--- A clear and concise description of what you want to happen. --> Fluent-bit should join these lines, either in the parser or in a filter. **Describe alternatives you've considered** <!--- A clear and concise description of any alternative solutions or features you've considered. --> We currently have issues with this and don't see any good workarounds. **Additional context** <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> <!--- Add any other context or screenshots about the feature request here. --> Log events are sent to Splunk, splitted over multiple events and unparsed since they are not considered to be valid json when splitted.
process
join partial docker json file lines is your feature request related to a problem please describe fluent bit does not join log lines that are splitted as partial by docker daemon when logging to json file describe the solution you d like fluent bit should join these lines either in the parser or in a filter describe alternatives you ve considered we currently have issues with this and don t see any good workarounds additional context log events are sent to splunk splitted over multiple events and unparsed since they are not considered to be valid json when splitted
1
328,265
9,991,851,896
IssuesEvent
2019-07-11 12:10:04
pombase/curation
https://api.github.com/repos/pombase/curation
closed
rad24 check
annotation priority high priority
GO:0046826 | negative regulation of protein export from nucleus | 3 -- | -- | -- regulates rgf1 | IMP | Muñoz S et al. (2014)
2.0
rad24 check - GO:0046826 | negative regulation of protein export from nucleus | 3 -- | -- | -- regulates rgf1 | IMP | Muñoz S et al. (2014)
non_process
check go negative regulation of protein export from nucleus regulates  imp muñoz s et al
0
14,338
17,367,805,938
IssuesEvent
2021-07-30 09:44:26
googleapis/python-spanner
https://api.github.com/repos/googleapis/python-spanner
closed
samples.samples.snippets_test: many tests failed
api: spanner flakybot: issue samples type: process
Many tests failed at the same time in this package. * I will close this issue when there are no more failures in this package _and_ there is at least one pass. * No new issues will be filed for this package until this issue is closed. * If there are already issues for individual test cases, I will close them when the corresponding test passes. You can close them earlier, if you prefer, and I won't reopen them while this issue is still open. Here are the tests that failed: * test_create_instance * test_create_database * test_create_database_with_encryption_config * test_insert_data * test_delete_data * test_query_data * test_add_column * test_read_data * test_update_data * test_read_stale_data * test_read_write_transaction * test_query_data_with_new_column * test_add_index (#403) * test_query_data_with_index (#404) * test_read_data_with_index (#405) * test_add_storing_index * test_read_data_with_storing_index * test_read_only_transaction * test_add_timestamp_column * test_update_data_with_timestamp * test_query_data_with_timestamp * test_create_table_with_timestamp * test_insert_data_with_timestamp * test_write_struct_data * test_query_with_struct * test_query_with_array_of_struct * test_query_struct_field * test_query_nested_struct_field * test_insert_data_with_dml * test_log_commit_stats * test_update_data_with_dml * test_delete_data_with_dml * test_update_data_with_dml_timestamp * test_dml_write_read_transaction * test_update_data_with_dml_struct * test_insert_with_dml * test_query_data_with_parameter * test_write_with_dml_transaction * test_create_table_with_datatypes * test_insert_datatypes_data * test_query_data_with_array * test_query_data_with_bool * test_query_data_with_bytes * test_query_data_with_date * test_query_data_with_float * test_query_data_with_int * test_query_data_with_string * test_add_numeric_column * test_update_data_with_numeric * test_query_data_with_numeric_parameter * test_query_data_with_timestamp_parameter * test_query_data_with_query_options * test_create_client_with_query_options (#406) ----- commit: 2487800e31842a44dcc37937c325e130c8c926b0 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/e2951c84-6fe7-446c-87e7-3b97256bee0b), [Sponge](http://sponge2/e2951c84-6fe7-446c-87e7-3b97256bee0b) status: failed
1.0
samples.samples.snippets_test: many tests failed - Many tests failed at the same time in this package. * I will close this issue when there are no more failures in this package _and_ there is at least one pass. * No new issues will be filed for this package until this issue is closed. * If there are already issues for individual test cases, I will close them when the corresponding test passes. You can close them earlier, if you prefer, and I won't reopen them while this issue is still open. Here are the tests that failed: * test_create_instance * test_create_database * test_create_database_with_encryption_config * test_insert_data * test_delete_data * test_query_data * test_add_column * test_read_data * test_update_data * test_read_stale_data * test_read_write_transaction * test_query_data_with_new_column * test_add_index (#403) * test_query_data_with_index (#404) * test_read_data_with_index (#405) * test_add_storing_index * test_read_data_with_storing_index * test_read_only_transaction * test_add_timestamp_column * test_update_data_with_timestamp * test_query_data_with_timestamp * test_create_table_with_timestamp * test_insert_data_with_timestamp * test_write_struct_data * test_query_with_struct * test_query_with_array_of_struct * test_query_struct_field * test_query_nested_struct_field * test_insert_data_with_dml * test_log_commit_stats * test_update_data_with_dml * test_delete_data_with_dml * test_update_data_with_dml_timestamp * test_dml_write_read_transaction * test_update_data_with_dml_struct * test_insert_with_dml * test_query_data_with_parameter * test_write_with_dml_transaction * test_create_table_with_datatypes * test_insert_datatypes_data * test_query_data_with_array * test_query_data_with_bool * test_query_data_with_bytes * test_query_data_with_date * test_query_data_with_float * test_query_data_with_int * test_query_data_with_string * test_add_numeric_column * test_update_data_with_numeric * test_query_data_with_numeric_parameter * test_query_data_with_timestamp_parameter * test_query_data_with_query_options * test_create_client_with_query_options (#406) ----- commit: 2487800e31842a44dcc37937c325e130c8c926b0 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/e2951c84-6fe7-446c-87e7-3b97256bee0b), [Sponge](http://sponge2/e2951c84-6fe7-446c-87e7-3b97256bee0b) status: failed
process
samples samples snippets test many tests failed many tests failed at the same time in this package i will close this issue when there are no more failures in this package and there is at least one pass no new issues will be filed for this package until this issue is closed if there are already issues for individual test cases i will close them when the corresponding test passes you can close them earlier if you prefer and i won t reopen them while this issue is still open here are the tests that failed test create instance test create database test create database with encryption config test insert data test delete data test query data test add column test read data test update data test read stale data test read write transaction test query data with new column test add index test query data with index test read data with index test add storing index test read data with storing index test read only transaction test add timestamp column test update data with timestamp test query data with timestamp test create table with timestamp test insert data with timestamp test write struct data test query with struct test query with array of struct test query struct field test query nested struct field test insert data with dml test log commit stats test update data with dml test delete data with dml test update data with dml timestamp test dml write read transaction test update data with dml struct test insert with dml test query data with parameter test write with dml transaction test create table with datatypes test insert datatypes data test query data with array test query data with bool test query data with bytes test query data with date test query data with float test query data with int test query data with string test add numeric column test update data with numeric test query data with numeric parameter test query data with timestamp parameter test query data with query options test create client with query options commit buildurl status failed
1
185,059
6,718,568,870
IssuesEvent
2017-10-15 14:26:27
Ekultek/Zeus-Scanner
https://api.github.com/repos/Ekultek/Zeus-Scanner
closed
WebDriverException: Message: Expected [object Undefined] undefined to be a string
browser issue bug priority: medium
Zeus version: `1.0.47.8f84` Firefox version: `Version: (55.0)` Error info: ```Traceback (most recent call last): File "/opt/Zeus-Scanner/var/google_search/search.py", line 242, in parse_search_results query_url = get_urls(query, url_to_search, verbose=verbose, user_agent=user_agent, proxy=proxy_string) File "/opt/Zeus-Scanner/var/google_search/search.py", line 141, in get_urls search.send_keys(query) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/webelement.py", line 350, in send_keys 'value': keys_to_typing(value)}) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/webelement.py", line 499, in _execute return self._parent.execute(command, params) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/webdriver.py", line 297, in execute self.error_handler.check_response(response) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/errorhandler.py", line 194, in check_response raise exception_class(message, screen, stacktrace) WebDriverException: Message: Expected [object Undefined] undefined to be a string ```` Running details: `Linux-2.6.32-042stab120.16-x86_64-with-Ubuntu-16.04-xenial` Commands used: `zeus.py -d inurl:php?id=` Log file info: ```2017-10-14 15:17:47,759;zeus-log;INFO;log file being saved to '/opt/Zeus-Scanner/log/zeus-log-6.log'... 2017-10-14 15:17:47,759;zeus-log;INFO;using default search engine... 2017-10-14 15:17:47,759;zeus-log;INFO;starting dork scan with query 'inurl:php?id='... 2017-10-14 15:17:47,759;zeus-log;INFO;attempting to gather query URL... 2017-10-14 15:17:47,876;zeus-log;INFO;firefox browser display will be hidden while it performs the query... 2017-10-14 15:17:47,876;zeus-log;WARNING;your web browser will be automated in order for Zeus to successfully bypass captchas and API calls. this is done in order to grab the URL from the search and parse the results. please give selenium time to finish it's task... 2017-10-14 15:17:51,182;zeus-log;INFO;browser will open shortly... 2017-10-14 15:17:51,686;zeus-log;INFO;searching 'http://google.com' using query 'inurl:php?id='... 2017-10-14 15:17:51,693;zeus-log;ERROR;search.pyc failed to gather the URL from search engine, caught exception 'Message: Expected [object Undefined] undefined to be a string' exception has been logged to current log file... Traceback (most recent call last): File "/opt/Zeus-Scanner/var/google_search/search.py", line 242, in parse_search_results query_url = get_urls(query, url_to_search, verbose=verbose, user_agent=user_agent, proxy=proxy_string) File "/opt/Zeus-Scanner/var/google_search/search.py", line 141, in get_urls search.send_keys(query) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/webelement.py", line 350, in send_keys 'value': keys_to_typing(value)}) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/webelement.py", line 499, in _execute return self._parent.execute(command, params) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/webdriver.py", line 297, in execute self.error_handler.check_response(response) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/errorhandler.py", line 194, in check_response raise exception_class(message, screen, stacktrace) WebDriverException: Message: Expected [object Undefined] undefined to be a string 2017-10-14 15:17:55,057;zeus-log;INFO;Zeus got an unexpected error and will automatically create an issue for this error, please wait... 2017-10-14 15:17:55,058;zeus-log;INFO;getting authorization... 2017-10-14 15:17:55,058;zeus-log;INFO;extracting traceback from log file... 2017-10-14 15:17:55,059;zeus-log;INFO;attempting to get firefox browser version... ```
1.0
WebDriverException: Message: Expected [object Undefined] undefined to be a string - Zeus version: `1.0.47.8f84` Firefox version: `Version: (55.0)` Error info: ```Traceback (most recent call last): File "/opt/Zeus-Scanner/var/google_search/search.py", line 242, in parse_search_results query_url = get_urls(query, url_to_search, verbose=verbose, user_agent=user_agent, proxy=proxy_string) File "/opt/Zeus-Scanner/var/google_search/search.py", line 141, in get_urls search.send_keys(query) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/webelement.py", line 350, in send_keys 'value': keys_to_typing(value)}) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/webelement.py", line 499, in _execute return self._parent.execute(command, params) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/webdriver.py", line 297, in execute self.error_handler.check_response(response) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/errorhandler.py", line 194, in check_response raise exception_class(message, screen, stacktrace) WebDriverException: Message: Expected [object Undefined] undefined to be a string ```` Running details: `Linux-2.6.32-042stab120.16-x86_64-with-Ubuntu-16.04-xenial` Commands used: `zeus.py -d inurl:php?id=` Log file info: ```2017-10-14 15:17:47,759;zeus-log;INFO;log file being saved to '/opt/Zeus-Scanner/log/zeus-log-6.log'... 2017-10-14 15:17:47,759;zeus-log;INFO;using default search engine... 2017-10-14 15:17:47,759;zeus-log;INFO;starting dork scan with query 'inurl:php?id='... 2017-10-14 15:17:47,759;zeus-log;INFO;attempting to gather query URL... 2017-10-14 15:17:47,876;zeus-log;INFO;firefox browser display will be hidden while it performs the query... 2017-10-14 15:17:47,876;zeus-log;WARNING;your web browser will be automated in order for Zeus to successfully bypass captchas and API calls. this is done in order to grab the URL from the search and parse the results. please give selenium time to finish it's task... 2017-10-14 15:17:51,182;zeus-log;INFO;browser will open shortly... 2017-10-14 15:17:51,686;zeus-log;INFO;searching 'http://google.com' using query 'inurl:php?id='... 2017-10-14 15:17:51,693;zeus-log;ERROR;search.pyc failed to gather the URL from search engine, caught exception 'Message: Expected [object Undefined] undefined to be a string' exception has been logged to current log file... Traceback (most recent call last): File "/opt/Zeus-Scanner/var/google_search/search.py", line 242, in parse_search_results query_url = get_urls(query, url_to_search, verbose=verbose, user_agent=user_agent, proxy=proxy_string) File "/opt/Zeus-Scanner/var/google_search/search.py", line 141, in get_urls search.send_keys(query) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/webelement.py", line 350, in send_keys 'value': keys_to_typing(value)}) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/webelement.py", line 499, in _execute return self._parent.execute(command, params) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/webdriver.py", line 297, in execute self.error_handler.check_response(response) File "/usr/local/lib/python2.7/dist-packages/selenium/webdriver/remote/errorhandler.py", line 194, in check_response raise exception_class(message, screen, stacktrace) WebDriverException: Message: Expected [object Undefined] undefined to be a string 2017-10-14 15:17:55,057;zeus-log;INFO;Zeus got an unexpected error and will automatically create an issue for this error, please wait... 2017-10-14 15:17:55,058;zeus-log;INFO;getting authorization... 2017-10-14 15:17:55,058;zeus-log;INFO;extracting traceback from log file... 2017-10-14 15:17:55,059;zeus-log;INFO;attempting to get firefox browser version... ```
non_process
webdriverexception message expected undefined to be a string zeus version firefox version version error info traceback most recent call last file opt zeus scanner var google search search py line in parse search results query url get urls query url to search verbose verbose user agent user agent proxy proxy string file opt zeus scanner var google search search py line in get urls search send keys query file usr local lib dist packages selenium webdriver remote webelement py line in send keys value keys to typing value file usr local lib dist packages selenium webdriver remote webelement py line in execute return self parent execute command params file usr local lib dist packages selenium webdriver remote webdriver py line in execute self error handler check response response file usr local lib dist packages selenium webdriver remote errorhandler py line in check response raise exception class message screen stacktrace webdriverexception message expected undefined to be a string running details linux with ubuntu xenial commands used zeus py d inurl php id log file info zeus log info log file being saved to opt zeus scanner log zeus log log zeus log info using default search engine zeus log info starting dork scan with query inurl php id zeus log info attempting to gather query url zeus log info firefox browser display will be hidden while it performs the query zeus log warning your web browser will be automated in order for zeus to successfully bypass captchas and api calls this is done in order to grab the url from the search and parse the results please give selenium time to finish it s task zeus log info browser will open shortly zeus log info searching using query inurl php id zeus log error  undefined to be a string exception has been logged to current log file traceback most recent call last file opt zeus scanner var google search search py line in parse search results query url get urls query url to search verbose verbose user agent user agent proxy proxy string file opt zeus scanner var google search search py line in get urls search send keys query file usr local lib dist packages selenium webdriver remote webelement py line in send keys value keys to typing value file usr local lib dist packages selenium webdriver remote webelement py line in execute return self parent execute command params file usr local lib dist packages selenium webdriver remote webdriver py line in execute self error handler check response response file usr local lib dist packages selenium webdriver remote errorhandler py line in check response raise exception class message screen stacktrace webdriverexception message expected undefined to be a string zeus log info  got an unexpected error and will automatically create an issue for this error please wait  zeus log info  authorization  zeus log info  traceback from log file  zeus log info  to get firefox browser version 
0
7,962
11,142,246,150
IssuesEvent
2019-12-22 08:01:46
linnovate/root
https://api.github.com/repos/linnovate/root
closed
Settings - Change participant status in Office not updated in attributed Folder
2.0.7 Process bug Settings critical
go to Settings create new Office (for example: new office6 - meiran) add participant create new Folder (for example: new folder23 - meiran) select the office you created (new office6 - meiran) go to the office you create and change the participant status go back to the folder you create the participant status not updated the office: ![image](https://user-images.githubusercontent.com/45143091/53695263-cf745180-3dc1-11e9-9a00-48d236c715a3.png) the attributed folder: ![image](https://user-images.githubusercontent.com/45143091/53695281-fc286900-3dc1-11e9-89aa-797a15cca921.png)
1.0
Settings - Change participant status in Office not updated in attributed Folder - go to Settings create new Office (for example: new office6 - meiran) add participant create new Folder (for example: new folder23 - meiran) select the office you created (new office6 - meiran) go to the office you create and change the participant status go back to the folder you create the participant status not updated the office: ![image](https://user-images.githubusercontent.com/45143091/53695263-cf745180-3dc1-11e9-9a00-48d236c715a3.png) the attributed folder: ![image](https://user-images.githubusercontent.com/45143091/53695281-fc286900-3dc1-11e9-89aa-797a15cca921.png)
process
settings change participant status in office not updated in attributed folder go to settings create new office for example new meiran add participant create new folder for example new meiran select the office you created new meiran go to the office you create and change the participant status go back to the folder you create the participant status not updated the office the attributed folder
1
165,329
26,146,242,896
IssuesEvent
2022-12-30 05:35:09
Already-12-O-clock/ABCDmoviE
https://api.github.com/repos/Already-12-O-clock/ABCDmoviE
closed
[issue] ticketingPage select 버튼 option 기능 구현
JavaScript design feat
## 🤔 무엇을 하실 건지 설명해주세요! - 빠른 예매 페이지 내 상영관 선택 article에서 select-option 기능을 li, button으로 구현 ## 👨‍💻 구현방법 및 예상 동작 - ## ⭐ 특이사항 -
1.0
[issue] ticketingPage select 버튼 option 기능 구현 - ## 🤔 무엇을 하실 건지 설명해주세요! - 빠른 예매 페이지 내 상영관 선택 article에서 select-option 기능을 li, button으로 구현 ## 👨‍💻 구현방법 및 예상 동작 - ## ⭐ 특이사항 -
non_process
ticketingpage select 버튼 option 기능 구현 🤔 무엇을 하실 건지 설명해주세요 빠른 예매 페이지 내 상영관 선택 article에서 select option 기능을 li button으로 구현 👨‍💻 구현방법 및 예상 동작 ⭐ 특이사항
0
829,821
31,900,133,774
IssuesEvent
2023-09-18 07:11:16
mobile-directing-system/mds-desktop
https://api.github.com/repos/mobile-directing-system/mds-desktop
closed
[BUG]: Intel creation - Address book entries
priority:high scope:implementation type:suggestion
### Bug Description Occurence: Assign an address book entry for an intel delivery (Intel creation page) Address book entry selection does not show the correct users that are available. Only address book entries are shown that are assigned to the current user. ### Expected Behavior Address book entries that do not have a user assigned should be shown in the list. ### Steps to Reproduce - ### Possible Solutions - ### Further Information -
1.0
[BUG]: Intel creation - Address book entries - ### Bug Description Occurence: Assign an address book entry for an intel delivery (Intel creation page) Address book entry selection does not show the correct users that are available. Only address book entries are shown that are assigned to the current user. ### Expected Behavior Address book entries that do not have a user assigned should be shown in the list. ### Steps to Reproduce - ### Possible Solutions - ### Further Information -
non_process
intel creation address book entries bug description occurence assign an address book entry for an intel delivery intel creation page address book entry selection does not show the correct users that are available only address book entries are shown that are assigned to the current user expected behavior address book entries that do not have a user assigned should be shown in the list steps to reproduce possible solutions further information
0
4,645
7,494,573,389
IssuesEvent
2018-04-07 11:29:22
pwittchen/ReactiveNetwork
https://api.github.com/repos/pwittchen/ReactiveNetwork
opened
Release 0.12.4
RxJava1.x RxJava2.x release process
**Initial release notes**: - updated project dependencies - PR #268, PR #269 - fixed Internet monitoring issues in Strict Mode on Android Oreo - issue #255 (not ready yet) **Things to do**: TBD.
1.0
Release 0.12.4 - **Initial release notes**: - updated project dependencies - PR #268, PR #269 - fixed Internet monitoring issues in Strict Mode on Android Oreo - issue #255 (not ready yet) **Things to do**: TBD.
process
release initial release notes updated project dependencies pr pr fixed internet monitoring issues in strict mode on android oreo issue not ready yet things to do tbd
1
72,641
8,757,788,501
IssuesEvent
2018-12-14 22:40:26
mozilla-lockbox/lockbox-addon
https://api.github.com/repos/mozilla-lockbox/lockbox-addon
opened
Consistant product branding between mobile apps & extension
needs-content needs-design
This is a placeholder issue as we need to go through and define all the changes to be made across the entire product suite. Once we explore what all those changes need to be, and open associated issues in the various repos, then we can close out this issue. This may or may not be related to the rebranding efforts happening (ie even if the product line stays Lockbox for the time being, there are some discrepancies that need to be addressed)
1.0
Consistant product branding between mobile apps & extension - This is a placeholder issue as we need to go through and define all the changes to be made across the entire product suite. Once we explore what all those changes need to be, and open associated issues in the various repos, then we can close out this issue. This may or may not be related to the rebranding efforts happening (ie even if the product line stays Lockbox for the time being, there are some discrepancies that need to be addressed)
non_process
consistant product branding between mobile apps extension this is a placeholder issue as we need to go through and define all the changes to be made across the entire product suite once we explore what all those changes need to be and open associated issues in the various repos then we can close out this issue this may or may not be related to the rebranding efforts happening ie even if the product line stays lockbox for the time being there are some discrepancies that need to be addressed
0
316
2,759,310,109
IssuesEvent
2015-04-28 02:17:37
iojs/io.js
https://api.github.com/repos/iojs/io.js
closed
spawn /bin/sh EMFILE
child_process
I am writing an auto-update for an io.js application, which restarts by spawning a new instance of the binary using: ```js var options = { cwd: Node.path.join(binary, '..'), detached: true, stdio: [ 'ignore', 'ignore', 'ignore' ] }; var child = Node.child.spawn(binary, [], options); child.on('error', function(error) { console.log(error); } ); child.on('close', function (code) { console.log('child process exited with code ' + code); } ); child.unref(); ``` When the binary is launched, it overwrites the pid file, the other process notices this and shuts itself down, and the newly launched binary continues starting up (there are a few more checks in place besides this). On Mac OS X Yosemite, I'm finding that after exactly 24 restarts, the 25th restart fails with "child process exited with code 1". Running an exec command instead of the spawn command at that point also gives "spawn /bin/sh EMFILE". Also, I have a dmg installer file, which when run copies itself to /Applications and then starts the binary from there and is then shutdown using the same process above. The information on open handles and ports for the newly run binary from /Applications shown in Activity Monitor shows that some of the handles of the original parent process are attributed to the child process. As a result, ejecting the dmg installer fails with a "still in use" error. Could it be that the max user processes or maxfiles limit is being reduced with each detached child process spawned so that eventually we hit the limit? Is there anyway around this or can I do the restart differently without creating a long hierarchy of processes?
1.0
spawn /bin/sh EMFILE - I am writing an auto-update for an io.js application, which restarts by spawning a new instance of the binary using: ```js var options = { cwd: Node.path.join(binary, '..'), detached: true, stdio: [ 'ignore', 'ignore', 'ignore' ] }; var child = Node.child.spawn(binary, [], options); child.on('error', function(error) { console.log(error); } ); child.on('close', function (code) { console.log('child process exited with code ' + code); } ); child.unref(); ``` When the binary is launched, it overwrites the pid file, the other process notices this and shuts itself down, and the newly launched binary continues starting up (there are a few more checks in place besides this). On Mac OS X Yosemite, I'm finding that after exactly 24 restarts, the 25th restart fails with "child process exited with code 1". Running an exec command instead of the spawn command at that point also gives "spawn /bin/sh EMFILE". Also, I have a dmg installer file, which when run copies itself to /Applications and then starts the binary from there and is then shutdown using the same process above. The information on open handles and ports for the newly run binary from /Applications shown in Activity Monitor shows that some of the handles of the original parent process are attributed to the child process. As a result, ejecting the dmg installer fails with a "still in use" error. Could it be that the max user processes or maxfiles limit is being reduced with each detached child process spawned so that eventually we hit the limit? Is there anyway around this or can I do the restart differently without creating a long hierarchy of processes?
process
spawn bin sh emfile i am writing an auto update for an io js application which restarts by spawning a new instance of the binary using js var options cwd node path join binary detached true stdio var child node child spawn binary options child on error function error console log error child on close function code console log child process exited with code code child unref when the binary is launched it overwrites the pid file the other process notices this and shuts itself down and the newly launched binary continues starting up there are a few more checks in place besides this on mac os x yosemite i m finding that after exactly restarts the restart fails with child process exited with code running an exec command instead of the spawn command at that point also gives spawn bin sh emfile also i have a dmg installer file which when run copies itself to applications and then starts the binary from there and is then shutdown using the same process above the information on open handles and ports for the newly run binary from applications shown in activity monitor shows that some of the handles of the original parent process are attributed to the child process as a result ejecting the dmg installer fails with a still in use error could it be that the max user processes or maxfiles limit is being reduced with each detached child process spawned so that eventually we hit the limit is there anyway around this or can i do the restart differently without creating a long hierarchy of processes
1
182,501
14,138,370,910
IssuesEvent
2020-11-10 08:19:38
unicode-org/icu4x
https://api.github.com/repos/unicode-org/icu4x
closed
Running tests with different feature sets / architectures
C-test-infra T-core
We should think about how we test different feature sets and architectures. By default, `cargo test` only tests your default architecture and the crate's default features. Examples of things we want to test: - `std` vs. `no_std` environment (by enabling or disabling the `std` feature) - Building for the wasm32_unknown_unknown architecture - Building on non-Linux, non-x86 architectures like Windows, macOS, ARM Note: https://github.com/rust-lang/cargo/issues/2911 is a feature request to allow integration tests to choose different feature sets.
1.0
Running tests with different feature sets / architectures - We should think about how we test different feature sets and architectures. By default, `cargo test` only tests your default architecture and the crate's default features. Examples of things we want to test: - `std` vs. `no_std` environment (by enabling or disabling the `std` feature) - Building for the wasm32_unknown_unknown architecture - Building on non-Linux, non-x86 architectures like Windows, macOS, ARM Note: https://github.com/rust-lang/cargo/issues/2911 is a feature request to allow integration tests to choose different feature sets.
non_process
running tests with different feature sets architectures we should think about how we test different feature sets and architectures by default cargo test only tests your default architecture and the crate s default features examples of things we want to test std vs no std environment by enabling or disabling the std feature building for the unknown unknown architecture building on non linux non architectures like windows macos arm note is a feature request to allow integration tests to choose different feature sets
0
7,933
11,114,953,612
IssuesEvent
2019-12-18 09:46:04
linnovate/root
https://api.github.com/repos/linnovate/root
closed
Sub - task and sub - project arrangement
2.0.7 Not Reproducible Process bug Projects Tasks bug
Does not work well, does not open properly, is not presented as a good, not intuitive hierarchy
1.0
Sub - task and sub - project arrangement - Does not work well, does not open properly, is not presented as a good, not intuitive hierarchy
process
sub task and sub project arrangement does not work well does not open properly is not presented as a good not intuitive hierarchy
1
11,582
14,444,657,020
IssuesEvent
2020-12-07 21:35:55
pacificclimate/quail
https://api.github.com/repos/pacificclimate/quail
closed
Frost Days
process
## Description This function takes a climdexInput object as input and computes the FD (frost days) climdex index: that is, the annual count of days where daily minimum temperature drops below 0 degrees Celsius. ## Function to wrap [`climdex.fd`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L720)
1.0
Frost Days - ## Description This function takes a climdexInput object as input and computes the FD (frost days) climdex index: that is, the annual count of days where daily minimum temperature drops below 0 degrees Celsius. ## Function to wrap [`climdex.fd`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L720)
process
frost days description this function takes a climdexinput object as input and computes the fd frost days climdex index that is the annual count of days where daily minimum temperature drops below degrees celsius function to wrap
1
22,495
31,468,927,755
IssuesEvent
2023-08-30 05:49:34
gradle/gradle
https://api.github.com/repos/gradle/gradle
closed
NPE in Gradle annotation processor handling
a:bug in:modular-java in:annotation-processing has:reproducer
### Current Behavior Gradle annotation processor generates NullPointException (NPE) when annotation is placed at the module level (very legitimate use case) which causes annotation based code generators such as (JStachio)[https://github.com/jstachio/jstachio] to crash. ### Expected Behavior well, not crashing with NPE would be a good start... ### Context (optional) @agentgt and myself discussed the issue in depth here: https://github.com/jstachio/jstachio/issues/223 and the faulty code seem to be this: https://github.com/jstachio/jstachio/issues/223#issuecomment-1685321532 This very much appears to be a poor null-check analysis prior to Gradle releases, which resulted in a few hours of wasted time to others. ### Steps to Reproduce https://github.com/hrstoyanov/gradleIssue https://github.com/jstachio/jstachio/issues/223#issuecomment-1685240694 This small sample project shows the issue with JStachio 1.3.1 and Gradle 8.3 in Java 20 environment. ### Gradle version 8.3, 8.2.1 ..possibly older too. ### Build scan URL (optional) _No response_ ### Your Environment (optional) Java 20
1.0
NPE in Gradle annotation processor handling - ### Current Behavior Gradle annotation processor generates NullPointException (NPE) when annotation is placed at the module level (very legitimate use case) which causes annotation based code generators such as (JStachio)[https://github.com/jstachio/jstachio] to crash. ### Expected Behavior well, not crashing with NPE would be a good start... ### Context (optional) @agentgt and myself discussed the issue in depth here: https://github.com/jstachio/jstachio/issues/223 and the faulty code seem to be this: https://github.com/jstachio/jstachio/issues/223#issuecomment-1685321532 This very much appears to be a poor null-check analysis prior to Gradle releases, which resulted in a few hours of wasted time to others. ### Steps to Reproduce https://github.com/hrstoyanov/gradleIssue https://github.com/jstachio/jstachio/issues/223#issuecomment-1685240694 This small sample project shows the issue with JStachio 1.3.1 and Gradle 8.3 in Java 20 environment. ### Gradle version 8.3, 8.2.1 ..possibly older too. ### Build scan URL (optional) _No response_ ### Your Environment (optional) Java 20
process
npe in gradle annotation processor handling current behavior gradle annotation processor generates nullpointexception npe when annotation is placed at the module level very legitimate use case which causes annotation based code generators such as jstachio to crash expected behavior well not crashing with npe would be a good start context optional agentgt and myself discussed the issue in depth here and the faulty code seem to be this this very much appears to be a poor null check analysis prior to gradle releases which resulted in a few hours of wasted time to others steps to reproduce this small sample project shows the issue with jstachio and gradle in java environment gradle version possibly older too build scan url optional no response your environment optional java
1
12,966
15,216,063,405
IssuesEvent
2021-02-17 15:06:02
flamewave000/dragonflagon-fvtt
https://api.github.com/repos/flamewave000/dragonflagon-fvtt
closed
Conflict Chat Commands vs system DSA5
incompatibility
Hi, since the Chat Enhancements use Library: Chat Commands as a dependency it is no longer compatible with the system DSA5. I messaged both of them, but don't know if it is fixable. So... just so you know. Kind regards Danii
True
Conflict Chat Commands vs system DSA5 - Hi, since the Chat Enhancements use Library: Chat Commands as a dependency it is no longer compatible with the system DSA5. I messaged both of them, but don't know if it is fixable. So... just so you know. Kind regards Danii
non_process
conflict chat commands vs system hi since the chat enhancements use library chat commands as a dependency it is no longer compatible with the system i messaged both of them but don t know if it is fixable so just so you know kind regards danii
0
77,114
26,777,837,620
IssuesEvent
2023-01-31 18:32:35
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
Update error state for character limits on input fields in the CMS
CMS Team 508/Accessibility 508-defect-1
## Description Within the CMS on the edit screens, some fields that have a character limit enabled have a red indicator when the editor goes over the limit. This red color is used both on the border of the field and in the character count below the field. ## Screenshot ![image](https://user-images.githubusercontent.com/106678594/182946157-fc4fc87c-9074-4bf3-8637-b4367a8089f0.png) ## Acceptance Criteria - [ ] Update the input border color to be `#e31c3d` (red-bright) - [ ] Update the red text under the input field to be `#B20D01` (red-dark) - [x] Confirm with Accessibility Lead
1.0
Update error state for character limits on input fields in the CMS - ## Description Within the CMS on the edit screens, some fields that have a character limit enabled have a red indicator when the editor goes over the limit. This red color is used both on the border of the field and in the character count below the field. ## Screenshot ![image](https://user-images.githubusercontent.com/106678594/182946157-fc4fc87c-9074-4bf3-8637-b4367a8089f0.png) ## Acceptance Criteria - [ ] Update the input border color to be `#e31c3d` (red-bright) - [ ] Update the red text under the input field to be `#B20D01` (red-dark) - [x] Confirm with Accessibility Lead
non_process
update error state for character limits on input fields in the cms description within the cms on the edit screens some fields that have a character limit enabled have a red indicator when the editor goes over the limit this red color is used both on the border of the field and in the character count below the field screenshot acceptance criteria update the input border color to be red bright update the red text under the input field to be red dark confirm with accessibility lead
0
8,205
11,401,593,691
IssuesEvent
2020-01-31 00:01:11
aodn/imos-toolbox
https://api.github.com/repos/aodn/imos-toolbox
opened
check DEPTH against NOMINAL_DEPTH interaction broken
Type:bug Unit:Processing Unit:gui
From mark snell: >In the “check DEPTH against NOMINAL_DEPTH” window I can’t select an area for processing. Normally I hold down the left mouse button and select an area but now when I do that it drags the whole image around.
1.0
check DEPTH against NOMINAL_DEPTH interaction broken - From mark snell: >In the “check DEPTH against NOMINAL_DEPTH” window I can’t select an area for processing. Normally I hold down the left mouse button and select an area but now when I do that it drags the whole image around.
process
check depth against nominal depth interaction broken from mark snell in the “check depth against nominal depth” window i can’t select an area for processing normally i hold down the left mouse button and select an area but now when i do that it drags the whole image around
1
14,939
18,388,713,765
IssuesEvent
2021-10-12 00:38:24
ncbo/bioportal-project
https://api.github.com/repos/ncbo/bioportal-project
opened
Diff is failing for AFO ontology
ontology processing problem
Received a report from end user that [AFO](https://bioportal.bioontology.org/ontologies/AFO) is displaying an "Error Diff" status on the summary page. Stack trace from parsing log: ``` I, [2021-10-11T10:28:55.430441 #23475] INFO -- : ["Java call [java -DentityExpansionLimit=1500000 -Xmx5120M -jar /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/bundler/gems/ontologies_linked_data-c4562ccb355a/bin/bubastis.jar -ontology1 /srv/ncbo/repository/AFO/19/merged-without-qudt.ttl -ontology2 /srv/ncbo/repository/AFO/20/merged-without-qudt-and-inferred.ttl -output /srv/ncbo/repository/AFO/20/bubastis_diff.xml -format xml]"] I, [2021-10-11T10:28:55.888639 #23475] INFO -- : ["Bubastis diff finished OK."] I, [2021-10-11T10:28:55.888695 #23475] INFO -- : ["Performing diff failed. Reason: null"] I, [2021-10-11T10:28:55.888753 #23475] INFO -- : ["ontology 1 is a file /srv/ncbo/repository/AFO/19/merged-without-qudt.ttl ontology 2 is a file /srv/ncbo/repository/AFO/20/merged-without-qudt-and-inferred.ttl trying load now"] E, [2021-10-11T10:28:55.889135 #23475] ERROR -- : ["Bubastis diff for http://data.bioontology.org/ontologies/AFO/submissions/20 failed - LinkedData::Diff::BubastisDiffException: Bubastis diff command exited with status=0. Output file /srv/ncbo/repository/AFO/20/bubastis_diff.xml cannot be found."] E, [2021-10-11T10:28:55.889293 #23475] ERROR -- : ["LinkedData::Diff::BubastisDiffException: Bubastis diff command exited with status=0. Output file /srv/ncbo/repository/AFO/20/bubastis_diff.xml cannot be found. /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/bundler/gems/ontologies_linked_data-c4562ccb355a/lib/ontologies_linked_data/diff/bubastis_diff.rb:100:in `call_bubastis_java_cmd' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/bundler/gems/ontologies_linked_data-c4562ccb355a/lib/ontologies_linked_data/diff/bubastis_diff.rb:114:in `diff' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/bundler/gems/ontologies_linked_data-c4562ccb355a/lib/ontologies_linked_data/models/ontology_submission.rb:340:in `diff' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/bundler/gems/ontologies_linked_data-c4562ccb355a/lib/ontologies_linked_data/models/ontology_submission.rb:1090:in `process_submission' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:177:in `process_submission' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:47:in `block in process_queue_submissions' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:41:in `each' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:41:in `process_queue_submissions' /srv/ncbo/ncbo_cron/bin/ncbo_cron:246:in `block (3 levels) in <main>' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:65:in `block (3 levels) in scheduled_locking_job' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:51:in `fork' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:51:in `block (2 levels) in scheduled_locking_job' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/gems/mlanett-redis-lock-0.2.7/lib/redis-lock.rb:43:in `lock' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/gems/mlanett-redis-lock-0.2.7/lib/redis-lock.rb:234:in `lock' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:50:in `block in scheduled_locking_job' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/gems/rufus-scheduler-2.0.24/lib/rufus/sc/jobs.rb:230:in `trigger_block' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/gems/rufus-scheduler-2.0.24/lib/rufus/sc/jobs.rb:204:in `block in trigger' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/gems/rufus-scheduler-2.0.24/lib/rufus/sc/scheduler.rb:430:in `block in trigger_job'"] ```
1.0
Diff is failing for AFO ontology - Received a report from end user that [AFO](https://bioportal.bioontology.org/ontologies/AFO) is displaying an "Error Diff" status on the summary page. Stack trace from parsing log: ``` I, [2021-10-11T10:28:55.430441 #23475] INFO -- : ["Java call [java -DentityExpansionLimit=1500000 -Xmx5120M -jar /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/bundler/gems/ontologies_linked_data-c4562ccb355a/bin/bubastis.jar -ontology1 /srv/ncbo/repository/AFO/19/merged-without-qudt.ttl -ontology2 /srv/ncbo/repository/AFO/20/merged-without-qudt-and-inferred.ttl -output /srv/ncbo/repository/AFO/20/bubastis_diff.xml -format xml]"] I, [2021-10-11T10:28:55.888639 #23475] INFO -- : ["Bubastis diff finished OK."] I, [2021-10-11T10:28:55.888695 #23475] INFO -- : ["Performing diff failed. Reason: null"] I, [2021-10-11T10:28:55.888753 #23475] INFO -- : ["ontology 1 is a file /srv/ncbo/repository/AFO/19/merged-without-qudt.ttl ontology 2 is a file /srv/ncbo/repository/AFO/20/merged-without-qudt-and-inferred.ttl trying load now"] E, [2021-10-11T10:28:55.889135 #23475] ERROR -- : ["Bubastis diff for http://data.bioontology.org/ontologies/AFO/submissions/20 failed - LinkedData::Diff::BubastisDiffException: Bubastis diff command exited with status=0. Output file /srv/ncbo/repository/AFO/20/bubastis_diff.xml cannot be found."] E, [2021-10-11T10:28:55.889293 #23475] ERROR -- : ["LinkedData::Diff::BubastisDiffException: Bubastis diff command exited with status=0. Output file /srv/ncbo/repository/AFO/20/bubastis_diff.xml cannot be found. /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/bundler/gems/ontologies_linked_data-c4562ccb355a/lib/ontologies_linked_data/diff/bubastis_diff.rb:100:in `call_bubastis_java_cmd' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/bundler/gems/ontologies_linked_data-c4562ccb355a/lib/ontologies_linked_data/diff/bubastis_diff.rb:114:in `diff' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/bundler/gems/ontologies_linked_data-c4562ccb355a/lib/ontologies_linked_data/models/ontology_submission.rb:340:in `diff' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/bundler/gems/ontologies_linked_data-c4562ccb355a/lib/ontologies_linked_data/models/ontology_submission.rb:1090:in `process_submission' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:177:in `process_submission' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:47:in `block in process_queue_submissions' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:41:in `each' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:41:in `process_queue_submissions' /srv/ncbo/ncbo_cron/bin/ncbo_cron:246:in `block (3 levels) in <main>' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:65:in `block (3 levels) in scheduled_locking_job' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:51:in `fork' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:51:in `block (2 levels) in scheduled_locking_job' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/gems/mlanett-redis-lock-0.2.7/lib/redis-lock.rb:43:in `lock' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/gems/mlanett-redis-lock-0.2.7/lib/redis-lock.rb:234:in `lock' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:50:in `block in scheduled_locking_job' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/gems/rufus-scheduler-2.0.24/lib/rufus/sc/jobs.rb:230:in `trigger_block' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/gems/rufus-scheduler-2.0.24/lib/rufus/sc/jobs.rb:204:in `block in trigger' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.6.0/gems/rufus-scheduler-2.0.24/lib/rufus/sc/scheduler.rb:430:in `block in trigger_job'"] ```
process
diff is failing for afo ontology received a report from end user that is displaying an error diff status on the summary page stack trace from parsing log i info i info i info i info ontology is a file srv ncbo repository afo merged without qudt ttl ontology is a file srv ncbo repository afo merged without qudt and inferred ttl trying load now e error e error linkeddata diff bubastisdiffexception bubastis diff command exited with status output file srv ncbo repository afo bubastis diff xml cannot be found srv ncbo ncbo cron vendor bundle ruby bundler gems ontologies linked data lib ontologies linked data diff bubastis diff rb in call bubastis java cmd srv ncbo ncbo cron vendor bundle ruby bundler gems ontologies linked data lib ontologies linked data diff bubastis diff rb in diff srv ncbo ncbo cron vendor bundle ruby bundler gems ontologies linked data lib ontologies linked data models ontology submission rb in diff srv ncbo ncbo cron vendor bundle ruby bundler gems ontologies linked data lib ontologies linked data models ontology submission rb in process submission srv ncbo ncbo cron lib ncbo cron ontology submission parser rb in process submission srv ncbo ncbo cron lib ncbo cron ontology submission parser rb in block in process queue submissions srv ncbo ncbo cron lib ncbo cron ontology submission parser rb in each srv ncbo ncbo cron lib ncbo cron ontology submission parser rb in process queue submissions srv ncbo ncbo cron bin ncbo cron in block levels in srv ncbo ncbo cron lib ncbo cron scheduler rb in block levels in scheduled locking job srv ncbo ncbo cron lib ncbo cron scheduler rb in fork srv ncbo ncbo cron lib ncbo cron scheduler rb in block levels in scheduled locking job srv ncbo ncbo cron vendor bundle ruby gems mlanett redis lock lib redis lock rb in lock srv ncbo ncbo cron vendor bundle ruby gems mlanett redis lock lib redis lock rb in lock srv ncbo ncbo cron lib ncbo cron scheduler rb in block in scheduled locking job srv ncbo ncbo cron vendor bundle ruby gems rufus scheduler lib rufus sc jobs rb in trigger block srv ncbo ncbo cron vendor bundle ruby gems rufus scheduler lib rufus sc jobs rb in block in trigger srv ncbo ncbo cron vendor bundle ruby gems rufus scheduler lib rufus sc scheduler rb in block in trigger job
1
9,628
12,576,525,545
IssuesEvent
2020-06-09 08:02:45
kubeflow/manifests
https://api.github.com/repos/kubeflow/manifests
closed
Pin 0.7.0 manifests in preparation for 0.7.1
area/kfctl kind/process lifecycle/stale priority/p1
We want to cut a v0.7.0 release. Before we do that, all of the KFDef *.0.7.0.yaml files https://github.com/kubeflow/manifests/tree/v0.7-branch/kfdef Should be updated to point to a specific commit of the kubeflow/manifests v0.7-branch for the kubeflow-manifests repository. This way cherry picking changes to the manifests on the branch won't break things. e.g https://github.com/kubeflow/manifests/blob/c0e81bedec9a4df8acf568cc5ccacc4bc05a3b38/kfdef/kfctl_gcp_iap.0.7.0.yaml#L399 Is pointing to the tip of the branch. We might want to create a tag v0.7.0 pointing at the current tip https://github.com/kubeflow/manifests/commit/c0e81bedec9a4df8acf568cc5ccacc4bc05a3b38
1.0
Pin 0.7.0 manifests in preparation for 0.7.1 - We want to cut a v0.7.0 release. Before we do that, all of the KFDef *.0.7.0.yaml files https://github.com/kubeflow/manifests/tree/v0.7-branch/kfdef Should be updated to point to a specific commit of the kubeflow/manifests v0.7-branch for the kubeflow-manifests repository. This way cherry picking changes to the manifests on the branch won't break things. e.g https://github.com/kubeflow/manifests/blob/c0e81bedec9a4df8acf568cc5ccacc4bc05a3b38/kfdef/kfctl_gcp_iap.0.7.0.yaml#L399 Is pointing to the tip of the branch. We might want to create a tag v0.7.0 pointing at the current tip https://github.com/kubeflow/manifests/commit/c0e81bedec9a4df8acf568cc5ccacc4bc05a3b38
process
pin manifests in preparation for we want to cut a release before we do that all of the kfdef yaml files should be updated to point to a specific commit of the kubeflow manifests branch for the kubeflow manifests repository this way cherry picking changes to the manifests on the branch won t break things e g is pointing to the tip of the branch we might want to create a tag pointing at the current tip
1
2,266
5,101,458,906
IssuesEvent
2017-01-04 15:34:24
hbz/lobid-resources
https://api.github.com/repos/hbz/lobid-resources
reopened
Remove unnecessary JSON objects or add additional information
processing
There are some properties that have an array of one or more JSON objects with only the `id` field as information. One such property is `exemplar`. As many title have a lot of items connected (see e.g. http://lobid.org/resources/HT000786420#!) the already long array gets even more bloated by the object parentheses. Properties where this is the case: - `exemplar` - `fulltextOnline` - `hasSupplement` - `hasVersion` - `isFormatOf` - `isPartOf` - `isPrimaryTopicOf` - `language` - `owner` - `sameAs` - `seeAlso` - `similar` - `source` - `tableOfContents` - `webPageArchived` - `description` There are two options for dealing with these: 1. Adding additional informato the object, e.g. `type` (for exemplars this would be `Item`) or `label` if available. 2. Removing the object and directly list the URIs in the array. Add least for `language` we should add the labels. But I doubt that we could add useful information to all of these...
1.0
Remove unnecessary JSON objects or add additional information - There are some properties that have an array of one or more JSON objects with only the `id` field as information. One such property is `exemplar`. As many title have a lot of items connected (see e.g. http://lobid.org/resources/HT000786420#!) the already long array gets even more bloated by the object parentheses. Properties where this is the case: - `exemplar` - `fulltextOnline` - `hasSupplement` - `hasVersion` - `isFormatOf` - `isPartOf` - `isPrimaryTopicOf` - `language` - `owner` - `sameAs` - `seeAlso` - `similar` - `source` - `tableOfContents` - `webPageArchived` - `description` There are two options for dealing with these: 1. Adding additional informato the object, e.g. `type` (for exemplars this would be `Item`) or `label` if available. 2. Removing the object and directly list the URIs in the array. Add least for `language` we should add the labels. But I doubt that we could add useful information to all of these...
process
remove unnecessary json objects or add additional information there are some properties that have an array of one or more json objects with only the id field as information one such property is exemplar as many title have a lot of items connected see e g the already long array gets even more bloated by the object parentheses properties where this is the case exemplar fulltextonline hassupplement hasversion isformatof ispartof isprimarytopicof language owner sameas seealso similar source tableofcontents webpagearchived description there are two options for dealing with these adding additional informato the object e g type for exemplars this would be item or label if available removing the object and directly list the uris in the array add least for language we should add the labels but i doubt that we could add useful information to all of these
1
382,918
26,524,326,248
IssuesEvent
2023-01-19 07:14:43
RWD-data-environment-in-Hospital/Documents
https://api.github.com/repos/RWD-data-environment-in-Hospital/Documents
closed
SOURCE テーブルにレコードを追加するために、SQL を実行すると『ERROR: 列"is_cache_enabled"内のNULL値はNOT NULL制約違反です』とエラーが表示される。
documentation question
## **対象のドキュメント:Atlasセットアップ手順** ■6.2 source テーブルへのレコード追加 SOURCE テーブルにレコードを追加するために、SQL を実行すると『ERROR: 列"is_cache_enabled"内のNULL値はNOT NULL制約違反です』とエラーが表示される。 ![画像1](https://user-images.githubusercontent.com/88526779/207009634-0a285895-34ba-4421-96be-eb9e26cf401d.png)
1.0
SOURCE テーブルにレコードを追加するために、SQL を実行すると『ERROR: 列"is_cache_enabled"内のNULL値はNOT NULL制約違反です』とエラーが表示される。 - ## **対象のドキュメント:Atlasセットアップ手順** ■6.2 source テーブルへのレコード追加 SOURCE テーブルにレコードを追加するために、SQL を実行すると『ERROR: 列"is_cache_enabled"内のNULL値はNOT NULL制約違反です』とエラーが表示される。 ![画像1](https://user-images.githubusercontent.com/88526779/207009634-0a285895-34ba-4421-96be-eb9e26cf401d.png)
non_process
source テーブルにレコードを追加するために、sql を実行すると『error 列 is cache enabled 内のnull値はnot null制約違反です』とエラーが表示される。 対象のドキュメント:atlasセットアップ手順 ■ .  source テーブルへのレコード追加 source テーブルにレコードを追加するために、sql を実行すると『error 列 is cache enabled 内のnull値はnot null制約違反です』とエラーが表示される。
0
21,940
30,446,799,030
IssuesEvent
2023-07-15 19:28:37
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
pyutils 0.0.1b15 has 2 GuardDog issues
guarddog typosquatting silent-process-execution
https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b15", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:205", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmp2_m2wwpy/pyutils" } }```
1.0
pyutils 0.0.1b15 has 2 GuardDog issues - https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b15", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:205", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmp2_m2wwpy/pyutils" } }```
process
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt python utils pytils silent process execution location pyutils exec utils py pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp pyutils
1
2,435
5,215,849,170
IssuesEvent
2017-01-26 07:54:54
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
[Chrome] Last header cell has incorrect width after resize
bug inprocess
Hello, We have found an issue with incorrect sizing for the last table header cell in Chrome. In BootstrapTable.js: _adjustHeaderWidth() { const isScroll = headerContainer.offsetWidth !== tbody.parentNode.offsetWidth; // Here Chrome returns true while FF returns false (FF is right) ... const scrollBarWidth = isScroll ? Util.getScrollBarWidth() : 0; } So Chrome rounds offsetWidth differently. Could you please implement cross-browser solution for rounding width. For reference: http://stackoverflow.com/questions/24104905/offsetwidth-not-same-in-different-browsers
1.0
[Chrome] Last header cell has incorrect width after resize - Hello, We have found an issue with incorrect sizing for the last table header cell in Chrome. In BootstrapTable.js: _adjustHeaderWidth() { const isScroll = headerContainer.offsetWidth !== tbody.parentNode.offsetWidth; // Here Chrome returns true while FF returns false (FF is right) ... const scrollBarWidth = isScroll ? Util.getScrollBarWidth() : 0; } So Chrome rounds offsetWidth differently. Could you please implement cross-browser solution for rounding width. For reference: http://stackoverflow.com/questions/24104905/offsetwidth-not-same-in-different-browsers
process
last header cell has incorrect width after resize hello we have found an issue with incorrect sizing for the last table header cell in chrome in bootstraptable js adjustheaderwidth const isscroll headercontainer offsetwidth tbody parentnode offsetwidth here chrome returns true while ff returns false ff is right const scrollbarwidth isscroll util getscrollbarwidth so chrome rounds offsetwidth differently could you please implement cross browser solution for rounding width for reference
1
18,010
24,025,809,406
IssuesEvent
2022-09-15 11:24:47
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
obsoletion/merge : DNA damage induced protein phosphorylation
PomBase cell cycle and DNA processes obsoletion quick fix
has 13 annotations (including 2 pombe) GO:0006975 DNA damage induced protein phosphorylation Is only used for kinases in the checkpoint pathway. We don't have other phosphorylation events with the inducing mechanism but this could be protein kinase part_of DNA damage checkpoint signalling No need to capture that the signalling is induced by DNA damage in a term. Do you agree?
1.0
obsoletion/merge : DNA damage induced protein phosphorylation - has 13 annotations (including 2 pombe) GO:0006975 DNA damage induced protein phosphorylation Is only used for kinases in the checkpoint pathway. We don't have other phosphorylation events with the inducing mechanism but this could be protein kinase part_of DNA damage checkpoint signalling No need to capture that the signalling is induced by DNA damage in a term. Do you agree?
process
obsoletion merge dna damage induced protein phosphorylation has annotations including pombe go dna damage induced protein phosphorylation is only used for kinases in the checkpoint pathway we don t have other phosphorylation events with the inducing mechanism but this could be protein kinase part of dna damage checkpoint signalling no need to capture that the signalling is induced by dna damage in a term do you agree
1
440,193
30,734,342,971
IssuesEvent
2023-07-28 06:14:02
Nathan13888/VisitorBadgeReloaded
https://api.github.com/repos/Nathan13888/VisitorBadgeReloaded
closed
[FEATURE] New Method for Counting “Uniquely”
documentation enhancement help wanted
**Is your feature request related to a problem? Please describe.** Currently only counts per page reload. **Describe the solution you'd like** * New method for counting *uniquely*. * Doesn't impact performance much * Default to counting uniquely?? **Describe alternatives you've considered** * Camo server ignores expiry headers **Additional context** @miguelnietoa https://github.com/jwenjian/visitor-badge/issues/8
1.0
[FEATURE] New Method for Counting “Uniquely” - **Is your feature request related to a problem? Please describe.** Currently only counts per page reload. **Describe the solution you'd like** * New method for counting *uniquely*. * Doesn't impact performance much * Default to counting uniquely?? **Describe alternatives you've considered** * Camo server ignores expiry headers **Additional context** @miguelnietoa https://github.com/jwenjian/visitor-badge/issues/8
non_process
new method for counting “uniquely” is your feature request related to a problem please describe currently only counts per page reload describe the solution you d like new method for counting uniquely doesn t impact performance much default to counting uniquely describe alternatives you ve considered camo server ignores expiry headers additional context miguelnietoa
0
14,231
17,150,850,018
IssuesEvent
2021-07-13 20:23:12
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
Create Windows common.ps1
kind/process priority/important-longterm
We need to make all the Windows tests more generic - we can do this the same way common.sh works.
1.0
Create Windows common.ps1 - We need to make all the Windows tests more generic - we can do this the same way common.sh works.
process
create windows common we need to make all the windows tests more generic we can do this the same way common sh works
1
1,261
2,911,093,763
IssuesEvent
2015-06-22 06:47:57
piwik/piwik
https://api.github.com/repos/piwik/piwik
opened
Row Evolution can use too much memory
Bug c: Performance
* [Example URL to reproduce](https://demo.piwik.org/index.php?date=2015-06-22&apiMethod=UserCountry.getCity&label=%40Mumbai%252C%2520Maharashtra%252C%2520India%2C%40Tokyo%252C%2520Tokyo%252C%2520Japan&disableLink=1&module=CoreHome&action=getMultiRowEvolutionPopover&colors=%7B%22backgroundColor%22%3A%22%23ffffff%22%2C%22lineColor%22%3A%22%23162c4a%22%2C%22minPointColor%22%3A%22%23ff7f7f%22%2C%22maxPointColor%22%3A%22%2375bf7c%22%2C%22lastPointColor%22%3A%22%2355aaff%22%7D&idSite=1&period=month) (requires login to demo) * Got: `<b>Fatal error</b>: Allowed memory size of 536870912 bytes exhausted (tried to allocate 88 bytes) in <b>/home/piwik-demo/www/demo.piwik.org/core/Metrics.php</b> on line <b>196</b><br />` (I since reverted memory limit to 256M so message now is * Expected: report loads fast and does not use more than 256M of memory
True
Row Evolution can use too much memory - * [Example URL to reproduce](https://demo.piwik.org/index.php?date=2015-06-22&apiMethod=UserCountry.getCity&label=%40Mumbai%252C%2520Maharashtra%252C%2520India%2C%40Tokyo%252C%2520Tokyo%252C%2520Japan&disableLink=1&module=CoreHome&action=getMultiRowEvolutionPopover&colors=%7B%22backgroundColor%22%3A%22%23ffffff%22%2C%22lineColor%22%3A%22%23162c4a%22%2C%22minPointColor%22%3A%22%23ff7f7f%22%2C%22maxPointColor%22%3A%22%2375bf7c%22%2C%22lastPointColor%22%3A%22%2355aaff%22%7D&idSite=1&period=month) (requires login to demo) * Got: `<b>Fatal error</b>: Allowed memory size of 536870912 bytes exhausted (tried to allocate 88 bytes) in <b>/home/piwik-demo/www/demo.piwik.org/core/Metrics.php</b> on line <b>196</b><br />` (I since reverted memory limit to 256M so message now is * Expected: report loads fast and does not use more than 256M of memory
non_process
row evolution can use too much memory requires login to demo got fatal error allowed memory size of bytes exhausted tried to allocate bytes in home piwik demo www demo piwik org core metrics php on line i since reverted memory limit to so message now is expected report loads fast and does not use more than of memory
0
812,702
30,348,747,354
IssuesEvent
2023-07-11 17:16:35
Memmy-App/memmy
https://api.github.com/repos/Memmy-App/memmy
closed
Default Sort menu in Settings out of order, does not allow New
bug medium priority in progress
**Describe the bug** The Default Sort menu under Settings>Content is out of order. The option to sort by New is placed where the Cancel menu option would normally be, with Cancel in place of New. This means you cannot select New as a default sort option. Also, clicking Cancel (in the menu position of New) crashes the app. **To Reproduce** Steps to reproduce the behavior: 1. Go to Settings > Content > Default Sort 2. The list shows New and Cancel out of order in the menu 4. Selecting New has the same result as if you selected Cancel. Selecting Cancel crashes the app. **Expected behavior** The “New” default sort option would be listed among the other default sort options. Selecting New will change the default sort to New. The “Cancel” button for the default sort menu will appear at the very bottom and will function as expected to close the default sort menu and return to Settings > Content. **Screenshots** <img width="375" alt="image" src="https://github.com/Memmy-App/memmy/assets/138980525/039529a8-130d-48ec-be30-f0af4cd788d3"> **Smartphone (please complete the following information):** - Device: iPhone SE - OS: iOS 16.5.1 - Version (Memmy): 0.1.4 (14) TestFlight
1.0
Default Sort menu in Settings out of order, does not allow New - **Describe the bug** The Default Sort menu under Settings>Content is out of order. The option to sort by New is placed where the Cancel menu option would normally be, with Cancel in place of New. This means you cannot select New as a default sort option. Also, clicking Cancel (in the menu position of New) crashes the app. **To Reproduce** Steps to reproduce the behavior: 1. Go to Settings > Content > Default Sort 2. The list shows New and Cancel out of order in the menu 4. Selecting New has the same result as if you selected Cancel. Selecting Cancel crashes the app. **Expected behavior** The “New” default sort option would be listed among the other default sort options. Selecting New will change the default sort to New. The “Cancel” button for the default sort menu will appear at the very bottom and will function as expected to close the default sort menu and return to Settings > Content. **Screenshots** <img width="375" alt="image" src="https://github.com/Memmy-App/memmy/assets/138980525/039529a8-130d-48ec-be30-f0af4cd788d3"> **Smartphone (please complete the following information):** - Device: iPhone SE - OS: iOS 16.5.1 - Version (Memmy): 0.1.4 (14) TestFlight
non_process
default sort menu in settings out of order does not allow new describe the bug the default sort menu under settings content is out of order the option to sort by new is placed where the cancel menu option would normally be with cancel in place of new this means you cannot select new as a default sort option also clicking cancel in the menu position of new crashes the app to reproduce steps to reproduce the behavior go to settings content default sort the list shows new and cancel out of order in the menu selecting new has the same result as if you selected cancel selecting cancel crashes the app expected behavior the “new” default sort option would be listed among the other default sort options selecting new will change the default sort to new the “cancel” button for the default sort menu will appear at the very bottom and will function as expected to close the default sort menu and return to settings content screenshots img width alt image src smartphone please complete the following information device iphone se os ios version memmy testflight
0
17,904
23,878,986,571
IssuesEvent
2022-09-07 22:17:04
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
reopened
Dependency Dashboard
type: process
This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/serverless-scheduler-proxy-google.golang.org-genproto-digest -->fix(deps): update google.golang.org/genproto digest to 2234105 - [ ] <!-- unschedule-branch=renovate/app-all-minor-patch -->chore(deps): update dependency typescript to ~4.8.0 - [ ] <!-- unschedule-branch=renovate/cli-all-minor-patch -->chore(deps): update dependency typescript to ~4.8.0 - [ ] <!-- unschedule-branch=renovate/common-container-all-minor-patch -->chore(deps): update dependency typescript to ~4.8.0 - [ ] <!-- unschedule-branch=renovate/policy-all-minor-patch -->chore(deps): update dependency typescript to ~4.8.0 - [ ] <!-- unschedule-branch=renovate/flakybot-all-minor-patch -->fix(deps): update module google.golang.org/api to v0.95.0 - [ ] <!-- unschedule-branch=renovate/gcf-utils-google-auth-library-8.x -->chore(deps): update dependency google-auth-library to v8 - [ ] <!-- unschedule-branch=renovate/policy-meow-10.x -->chore(deps): update dependency meow to v10 - [ ] <!-- unschedule-branch=renovate/secret-rotator-google-cloud-secret-manager-4.x -->fix(deps): update dependency @google-cloud/secret-manager to v4 - [ ] <!-- unschedule-branch=renovate/cron-utils-google-auth-library-8.x -->fix(deps): update dependency google-auth-library to v8 - [ ] <!-- unschedule-branch=renovate/secret-rotator-google-auth-library-8.x -->fix(deps): update dependency google-auth-library to v8 - [ ] <!-- unschedule-branch=renovate/gcf-utils-into-stream-7.x -->fix(deps): update dependency into-stream to v7 - [ ] <!-- unschedule-branch=renovate/gcf-utils-yargs-17.x -->fix(deps): update dependency yargs to v17 (`yargs`, `@types/yargs`) - [ ] <!-- unschedule-branch=renovate/lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/auto-approve-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/auto-label-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/blunderbuss-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/bot-config-utils-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/canary-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/cherry-pick-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/conventional-commit-lint-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/cron-utils-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/datastore-lock-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/do-not-merge-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/failurechecker-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/flakybot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/gcf-utils-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/generate-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/generated-files-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/git-file-utils-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/header-checker-lint-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/issue-utils-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/label-sync-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/label-utils-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/loadtest-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/merge-on-green-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/mono-repo-publish-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/object-selector-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/owl-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/app-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/cli-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/common-container-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/owlbot-bootstrapper-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/policy-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/release-brancher-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/release-please-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/release-trigger-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/repo-metadata-lint-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/secret-rotator-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/snippet-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/sync-repo-settings-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/trusted-contribution-lock-file-maintenance -->chore(deps): lock file maintenance ## Detected dependencies <details><summary>cloudbuild</summary> <blockquote> <details><summary>cloudbuild.yaml</summary> </details> <details><summary>packages/auto-approve/cloudbuild.yaml</summary> </details> <details><summary>packages/auto-label/cloudbuild.yaml</summary> </details> <details><summary>packages/bazel-bot/cloudbuild.yaml</summary> </details> <details><summary>packages/blunderbuss/cloudbuild.yaml</summary> </details> <details><summary>packages/canary-bot/cloudbuild.yaml</summary> </details> <details><summary>packages/cherry-pick-bot/cloudbuild.yaml</summary> </details> <details><summary>packages/conventional-commit-lint/cloudbuild.yaml</summary> </details> <details><summary>packages/do-not-merge/cloudbuild.yaml</summary> </details> <details><summary>packages/failurechecker/cloudbuild.yaml</summary> </details> <details><summary>packages/flakybot/cloudbuild.yaml</summary> </details> <details><summary>packages/generated-files-bot/cloudbuild.yaml</summary> </details> <details><summary>packages/header-checker-lint/cloudbuild.yaml</summary> </details> <details><summary>packages/label-sync/cloudbuild.yaml</summary> </details> <details><summary>packages/merge-on-green/cloudbuild.yaml</summary> </details> <details><summary>packages/owl-bot/cloudbuild.yaml</summary> </details> <details><summary>packages/policy/cloudbuild.yaml</summary> </details> <details><summary>packages/release-please/cloudbuild.yaml</summary> </details> <details><summary>packages/release-trigger/cloudbuild.yaml</summary> </details> <details><summary>packages/repo-metadata-lint/cloudbuild.yaml</summary> </details> <details><summary>packages/snippet-bot/cloudbuild.yaml</summary> </details> <details><summary>packages/sync-repo-settings/cloudbuild.yaml</summary> </details> <details><summary>packages/trusted-contribution/cloudbuild.yaml</summary> </details> <details><summary>serverless-scheduler-proxy/cloudbuild.yaml</summary> </details> <details><summary>trigger-sync/cloudbuild.yaml</summary> </details> </blockquote> </details> <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/ci.yaml</summary> - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/setup-node v3@2fddd8803e2f5c9604345a0b591c3020ee971a93` - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/setup-node v3@2fddd8803e2f5c9604345a0b591c3020ee971a93` - `actions/setup-go v3@268d8c0ca0432bb2cf416faae41297df9d262d7f` - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/setup-node v3@2fddd8803e2f5c9604345a0b591c3020ee971a93` </details> <details><summary>.github/workflows/deps.yaml</summary> - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/setup-node v3@2fddd8803e2f5c9604345a0b591c3020ee971a93` - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/setup-node v3@2fddd8803e2f5c9604345a0b591c3020ee971a93` - `googleapis/code-suggester v4@212b0804e41c4f89edfefba37e1c39cdb9856e36` </details> <details><summary>.github/workflows/links.yaml</summary> - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `lycheeverse/lychee-action v1.5.1@4a5af7cd2958a2282cefbd9c10f63bdb89982d76` - `peter-evans/create-issue-from-file v4.0.0@99b87c35610e986ad2034a7b0518a9b3ebea541b` </details> </blockquote> </details> <details><summary>gomod</summary> <blockquote> <details><summary>packages/flakybot/go.mod</summary> - `go 1.19` - `cloud.google.com/go/pubsub v1.25.1` - `github.com/google/go-cmp v0.5.8` - `google.golang.org/api v0.94.0` </details> <details><summary>serverless-scheduler-proxy/go.mod</summary> - `go 1.19` - `cloud.google.com/go/secretmanager v1.5.0` - `github.com/google/uuid v1.3.0` - `google.golang.org/genproto v0.0.0-20220829175752-36a9c930ecbf@36a9c930ecbf` </details> </blockquote> </details> <details><summary>npm</summary> <blockquote> <details><summary>package.json</summary> - `c8 ^7.1.0` - `gaxios ^5.0.0` - `smee-client ^1.1.0` - `@google-cloud/mono-repo-publish ^1.0.0` - `node >=12.0.0` </details> <details><summary>packages/auto-approve/package.json</summary> - `@google-cloud/secret-manager ^4.1.1` - `ajv ^8.11.0` - `dayjs ^1.11.5` - `gcf-utils ^14.0.1` - `@octokit/rest ^19.0.4` - `@octokit/webhooks ^10.1.5` - `@octokit/webhooks-definitions ^3.67.3` - `@probot/octokit-plugin-config ^1.1.6` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `js-yaml ^4.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `ts-node ^10.9.1` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/auto-label/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/label-utils ^3.0.0` - `@google-cloud/storage ^6.4.2` - `gcf-utils ^14.0.1` - `@octokit/types ^7.2.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `js-yaml ^4.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/blunderbuss/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/datastore-lock ^4.0.0` - `@google-automations/label-utils ^3.0.0` - `gcf-utils ^14.0.1` - `@octokit/types ^7.1.1` - `@octokit/webhooks ^10.1.5` - `@types/mocha ^9.1.1` - `@types/node ^18.7.13` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `js-yaml ^4.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/bot-config-utils/package.json</summary> - `ajv ^8.11.0` - `ajv-formats ^2.1.1` - `gcf-utils ^14.0.1` - `js-yaml ^4.1.0` - `@octokit/types ^7.2.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/canary-bot/package.json</summary> - `dayjs ^1.11.5` - `gcf-utils ^14.0.1` - `lru-cache ^7.14.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `typescript ~4.8.2` - `node >= 14.0.0` </details> <details><summary>packages/cherry-pick-bot/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@octokit/rest ^19.0.4` - `gcf-utils ^14.0.1` - `yargs ^17.5.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/conventional-commit-lint/package.json</summary> - `@commitlint/config-conventional ^17.1.0` - `@commitlint/lint ^17.1.0` - `@google-automations/bot-config-utils ^6.1.0` - `@octokit/openapi-types ^13.8.0` - `@octokit/webhooks-types ^6.3.6` - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `js-yaml ^4.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/cron-utils/package.json</summary> - `@google-cloud/scheduler ^3.0.3` - `@googleapis/run ^10.0.2` - `google-auth-library ^7.14.1` - `js-yaml ^4.1.0` - `yargs ^17.5.1` - `@octokit/rest ^19.0.4` - `@types/js-yaml ^4.0.5` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `gts ^4.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/datastore-lock/package.json</summary> - `@google-cloud/datastore ^7.0.0` - `gcf-utils ^14.0.1` - `uuid ^9.0.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `google-datastore-emulator ^6.0.0` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/do-not-merge/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/failurechecker/package.json</summary> - `@google-automations/issue-utils ^2.0.0` - `gcf-utils ^14.0.1` - `@octokit/types ^7.2.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/flakybot/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/datastore-lock ^4.0.0` - `@google-automations/label-utils ^3.0.0` - `@octokit/openapi-types ^13.8.0` - `@octokit/types ^7.2.0` - `gcf-utils ^14.0.1` - `probot ^12.2.8` - `xml-js ^1.6.11` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/gcf-utils/package.json</summary> - `@google-cloud/kms ^3.0.1` - `@google-cloud/secret-manager ^4.1.1` - `@google-cloud/storage ^6.4.1` - `@google-cloud/tasks ^3.0.2` - `google-gax ^3.3.1` - `@googleapis/run ^10.0.2` - `@octokit/auth-app ^4.0.5` - `@octokit/plugin-enterprise-compatibility 2.0.2` - `@octokit/rest ^19.0.4` - `@probot/octokit-plugin-config ^1.1.6` - `@types/bunyan ^1.8.8` - `@types/dotenv ^6.1.1` - `@types/end-of-stream ^1.4.1` - `@types/express ^4.17.13` - `@types/into-stream ^3.1.1` - `@types/ioredis ^4.28.10` - `@types/lru-cache ^7.10.9` - `@types/sonic-boom ^2.1.0` - `@types/uuid ^8.3.4` - `body-parser ^1.20.0` - `express ^4.18.1` - `gaxios ^5.0.1` - `get-stream ^6.0.1` - `into-stream ^6.0.0` - `octokit-auth-probot ^1.2.8` - `pino ^8.4.2` - `probot ^12.2.7` - `tmp ^0.2.1` - `uuid ^9.0.0` - `yargs ^16.0.0` - `@types/get-stream ^3.0.2` - `@types/mocha ^9.1.1` - `@types/node ^18.7.14` - `@types/pino ^7.0.4` - `@types/sinon ^10.0.13` - `@types/tmp ^0.2.3` - `@types/yargs ^16.0.0` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `google-auth-library ^7.14.1` - `gts ^4.0.0` - `mocha ^10.0.0` - `mocked-env ^1.3.5` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `sonic-boom ^3.2.0` - `stream-mock ^2.0.5` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/generate-bot/package.json</summary> - `enquirer ^2.3.6` - `handlebars ^4.7.7` - `@types/mocha ^9.1.1` - `@types/recursive-readdir ^2.2.1` - `@types/rimraf ^3.0.2` - `c8 ^7.12.0` - `gts ^4.0.0` - `mocha ^10.0.0` - `recursive-readdir ^2.2.2` - `rimraf ^3.0.2` - `typescript ~4.8.2` - `node >=12.18.2` </details> <details><summary>packages/generated-files-bot/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `gcf-utils ^14.0.1` - `js-yaml ^4.1.0` - `jsonpath ^1.1.1` - `minimatch ^5.1.0` - `@types/js-yaml ^4.0.5` - `@types/jsonpath ^0.2.0` - `@types/minimatch ^5.1.2` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/git-file-utils/package.json</summary> - `@octokit/rest 19.0.4` - `minimatch ^5.1.0` - `@types/chai ^4.3.3` - `@types/minimatch ^5.1.2` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `chai ^4.3.6` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `gts ^4.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/header-checker-lint/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `gcf-utils ^14.0.1` - `minimatch ^5.1.0` - `@types/minimatch ^5.1.2` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/issue-utils/package.json</summary> - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/label-sync/package.json</summary> - `@google-cloud/storage ^6.4.2` - `gaxios ^5.0.1` - `gcf-utils ^14.0.1` - `@probot/octokit-plugin-config ^1.1.6` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 12.18.2` </details> <details><summary>packages/label-utils/package.json</summary> - `@google-automations/datastore-lock ^4.0.0` - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/loadtest-bot/package.json</summary> - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `smee-client ^1.2.3` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/merge-on-green/package.json</summary> - `@google-automations/label-utils ^3.0.0` - `@google-cloud/datastore ^7.0.0` - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/mono-repo-publish/package.json</summary> - `@octokit/auth-app ^4.0.5` - `@octokit/rest ^19.0.4` - `yargs ^17.5.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `semistandard ^16.0.1` - `sinon ^14.0.0` - `typescript ^4.8.2` - `node >= 12.18.2` </details> <details><summary>packages/object-selector/package.json</summary> - `ajv ^8.11.0` - `easy-object-selector ^1.3.0` - `gcf-utils ^14.0.1` - `js-yaml ^4.1.0` - `yargs ^17.5.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `gts ^4.0.0` - `mocha ^10.0.0` - `mocked-env ^1.3.5` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/owl-bot/package.json</summary> - `@google-automations/datastore-lock ^4.0.0` - `@google-automations/label-utils ^3.0.0` - `@google-cloud/cloudbuild ^3.1.0` - `@octokit/rest ^19.0.4` - `@octokit/webhooks-types ^6.3.6` - `adm-zip ^0.5.9` - `ajv ^8.11.0` - `code-suggester ^4.1.0` - `firebase-admin ^11.0.1` - `follow-redirects ^1.15.1` - `fs-extra ^10.1.0` - `gaxios ^5.0.1` - `gcf-utils ^14.0.1` - `glob ^8.0.3` - `js-yaml ^4.1.0` - `jsonwebtoken ^8.5.1` - `minimatch ^5.1.0` - `probot ^12.2.5` - `tmp ^0.2.1` - `yargs ^17.5.1` - `@types/adm-zip ^0.5.0` - `@types/follow-redirects ^1.14.1` - `@types/fs-extra ^9.0.13` - `@types/glob ^8.0.0` - `@types/js-yaml ^4.0.5` - `@types/jsonwebtoken ^8.5.9` - `@types/minimatch ^5.1.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.13` - `@types/rimraf ^3.0.2` - `@types/sinon ^10.0.13` - `@types/tmp ^0.2.3` - `@types/yargs ^17.0.11` - `c8 ^7.12.0` - `gts ^3.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `rimraf ^3.0.2` - `sinon ^14.0.0` - `ts-node ^10.9.1` - `typescript ~4.8.2` - `uuid ^9.0.0` - `node >=14.0.0` </details> <details><summary>packages/owlbot-bootstrapper/app/package.json</summary> - `probot ^12.2.5` - `@types/mocha ^9.1.1` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^3.1.0` - `mocha ^10.0.0` - `typescript ~4.7.4` - `node >= 14` </details> <details><summary>packages/owlbot-bootstrapper/cli/package.json</summary> - `@google-cloud/cloudbuild ^3.1.0` - `yargs ^17.5.1` - `@types/yargs ^17.0.11` - `@types/jwt-encode ^1.0.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.9` - `@types/node-fetch ^2.6.2` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^3.1.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `typescript ~4.7.4` - `node >= 14` </details> <details><summary>packages/owlbot-bootstrapper/common-container/package.json</summary> - `@google-cloud/cloudbuild ^3.1.0` - `@google-cloud/secret-manager ^4.0.0` - `@octokit/auth-app ^4.0.5` - `@octokit/rest ^19.0.4` - `@types/yargs ^17.0.11` - `fs 0.0.1-security` - `gaxios ^5.0.1` - `gcf-utils ^14.0.1` - `jsonwebtoken ^8.5.1` - `node-fetch ^3.2.10` - `nodejs-fetch ^1.0.0` - `snap-shot-it ^7.9.6` - `uuidv4 ^6.2.13` - `yargs ^17.5.1` - `@types/jwt-encode ^1.0.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.9` - `@types/node-fetch ^2.6.2` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^3.1.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `typescript ~4.7.4` - `node >= 14` </details> <details><summary>packages/owlbot-bootstrapper/package.json</summary> - `@google-cloud/cloudbuild ^3.1.0` - `@google-cloud/secret-manager ^4.1.1` - `@octokit/auth-app ^4.0.5` - `@octokit/rest ^19.0.4` - `@types/yargs ^17.0.12` - `fs 0.0.1-security` - `gaxios ^5.0.1` - `gcf-utils ^14.0.1` - `jsonwebtoken ^8.5.1` - `node-fetch ^3.2.10` - `nodejs-fetch ^1.0.0` - `snap-shot-it ^7.9.6` - `uuidv4 ^6.2.13` - `yargs ^17.5.1` - `@types/jwt-encode ^1.0.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/node-fetch ^2.6.2` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/policy/package.json</summary> - `@google-cloud/bigquery ^6.0.2` - `code-suggester ^4.0.1` - `gaxios ^5.0.1` - `gcf-utils ^14.0.1` - `lru-cache ^7.14.0` - `uuid ^9.0.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.9` - `@types/sinon ^10.0.13` - `@types/uuid ^8.3.4` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^3.1.0` - `meow ^9.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `typescript ~4.7.4` - `node >= 14` </details> <details><summary>packages/release-brancher/package.json</summary> - `@octokit/rest ^19.0.4` - `code-suggester ^4.1.0` - `js-yaml ^4.1.0` - `yargs ^17.5.1` - `@types/js-yaml ^4.0.5` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/pino ^7.0.4` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 12.16.0` </details> <details><summary>packages/release-please/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/issue-utils ^2.0.0` - `@google-automations/label-utils ^3.0.0` - `@octokit/rest ^19.0.4` - `@octokit/webhooks ^10.1.5` - `gcf-utils ^14.0.1` - `release-please ^14.3.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/semver ^7.3.12` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `js-yaml ^4.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14.0.0` </details> <details><summary>packages/release-trigger/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/datastore-lock ^4.0.0` - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/repo-metadata-lint/package.json</summary> - `@google-cloud/storage ^6.4.2` - `@octokit/rest ^19.0.4` - `ajv ^8.11.0` - `gaxios ^5.0.1` - `gcf-utils ^14.0.1` - `jsonwebtoken ^8.5.1` - `yargs ^17.5.1` - `@octokit/types ^7.2.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/secret-rotator/package.json</summary> - `@google-cloud/secret-manager ^3.12.0` - `@googleapis/iam ^3.0.1` - `express ^4.18.1` - `gcf-utils ^14.0.1` - `google-auth-library ^7.14.1` - `@types/express ^4.17.13` - `@types/mocha ^9.1.1` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `gaxios ^5.0.1` - `google-gax ^3.4.0` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `typescript ^4.8.2` - `node >=14.0.0` </details> <details><summary>packages/snippet-bot/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/git-file-utils ^1.2.0` - `@google-automations/label-utils ^3.0.0` - `@google-cloud/storage ^6.4.2` - `follow-redirects ^1.15.1` - `gcf-utils ^14.0.1` - `lru-cache ^7.14.0` - `minimatch ^5.1.0` - `parse-diff ^0.9.0` - `tar ^6.1.11` - `tmp-promise ^3.0.3` - `@octokit/webhooks-types ^6.3.6` - `@types/follow-redirects ^1.14.1` - `@types/minimatch ^5.1.2` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/tar ^6.1.2` - `@types/tmp ^0.2.3` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/sync-repo-settings/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/issue-utils ^2.0.0` - `extend ^3.0.2` - `gcf-utils ^14.0.1` - `js-yaml ^4.1.0` - `yargs ^17.5.1` - `@types/extend ^3.0.1` - `@types/js-yaml ^4.0.5` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `mocked-env ^1.3.5` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `ts-node ^10.9.1` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/trusted-contribution/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-cloud/secret-manager ^4.1.1` - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `js-yaml ^4.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> </blockquote> </details> <details><summary>pip_requirements</summary> <blockquote> <details><summary>packages/release-trigger/requirements.txt</summary> - `gcp-releasetool ==1.8.7` </details> </blockquote> </details> --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/serverless-scheduler-proxy-google.golang.org-genproto-digest -->fix(deps): update google.golang.org/genproto digest to 2234105 - [ ] <!-- unschedule-branch=renovate/app-all-minor-patch -->chore(deps): update dependency typescript to ~4.8.0 - [ ] <!-- unschedule-branch=renovate/cli-all-minor-patch -->chore(deps): update dependency typescript to ~4.8.0 - [ ] <!-- unschedule-branch=renovate/common-container-all-minor-patch -->chore(deps): update dependency typescript to ~4.8.0 - [ ] <!-- unschedule-branch=renovate/policy-all-minor-patch -->chore(deps): update dependency typescript to ~4.8.0 - [ ] <!-- unschedule-branch=renovate/flakybot-all-minor-patch -->fix(deps): update module google.golang.org/api to v0.95.0 - [ ] <!-- unschedule-branch=renovate/gcf-utils-google-auth-library-8.x -->chore(deps): update dependency google-auth-library to v8 - [ ] <!-- unschedule-branch=renovate/policy-meow-10.x -->chore(deps): update dependency meow to v10 - [ ] <!-- unschedule-branch=renovate/secret-rotator-google-cloud-secret-manager-4.x -->fix(deps): update dependency @google-cloud/secret-manager to v4 - [ ] <!-- unschedule-branch=renovate/cron-utils-google-auth-library-8.x -->fix(deps): update dependency google-auth-library to v8 - [ ] <!-- unschedule-branch=renovate/secret-rotator-google-auth-library-8.x -->fix(deps): update dependency google-auth-library to v8 - [ ] <!-- unschedule-branch=renovate/gcf-utils-into-stream-7.x -->fix(deps): update dependency into-stream to v7 - [ ] <!-- unschedule-branch=renovate/gcf-utils-yargs-17.x -->fix(deps): update dependency yargs to v17 (`yargs`, `@types/yargs`) - [ ] <!-- unschedule-branch=renovate/lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/auto-approve-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/auto-label-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/blunderbuss-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/bot-config-utils-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/canary-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/cherry-pick-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/conventional-commit-lint-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/cron-utils-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/datastore-lock-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/do-not-merge-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/failurechecker-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/flakybot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/gcf-utils-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/generate-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/generated-files-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/git-file-utils-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/header-checker-lint-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/issue-utils-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/label-sync-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/label-utils-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/loadtest-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/merge-on-green-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/mono-repo-publish-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/object-selector-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/owl-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/app-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/cli-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/common-container-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/owlbot-bootstrapper-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/policy-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/release-brancher-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/release-please-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/release-trigger-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/repo-metadata-lint-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/secret-rotator-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/snippet-bot-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/sync-repo-settings-lock-file-maintenance -->chore(deps): lock file maintenance - [ ] <!-- unschedule-branch=renovate/trusted-contribution-lock-file-maintenance -->chore(deps): lock file maintenance ## Detected dependencies <details><summary>cloudbuild</summary> <blockquote> <details><summary>cloudbuild.yaml</summary> </details> <details><summary>packages/auto-approve/cloudbuild.yaml</summary> </details> <details><summary>packages/auto-label/cloudbuild.yaml</summary> </details> <details><summary>packages/bazel-bot/cloudbuild.yaml</summary> </details> <details><summary>packages/blunderbuss/cloudbuild.yaml</summary> </details> <details><summary>packages/canary-bot/cloudbuild.yaml</summary> </details> <details><summary>packages/cherry-pick-bot/cloudbuild.yaml</summary> </details> <details><summary>packages/conventional-commit-lint/cloudbuild.yaml</summary> </details> <details><summary>packages/do-not-merge/cloudbuild.yaml</summary> </details> <details><summary>packages/failurechecker/cloudbuild.yaml</summary> </details> <details><summary>packages/flakybot/cloudbuild.yaml</summary> </details> <details><summary>packages/generated-files-bot/cloudbuild.yaml</summary> </details> <details><summary>packages/header-checker-lint/cloudbuild.yaml</summary> </details> <details><summary>packages/label-sync/cloudbuild.yaml</summary> </details> <details><summary>packages/merge-on-green/cloudbuild.yaml</summary> </details> <details><summary>packages/owl-bot/cloudbuild.yaml</summary> </details> <details><summary>packages/policy/cloudbuild.yaml</summary> </details> <details><summary>packages/release-please/cloudbuild.yaml</summary> </details> <details><summary>packages/release-trigger/cloudbuild.yaml</summary> </details> <details><summary>packages/repo-metadata-lint/cloudbuild.yaml</summary> </details> <details><summary>packages/snippet-bot/cloudbuild.yaml</summary> </details> <details><summary>packages/sync-repo-settings/cloudbuild.yaml</summary> </details> <details><summary>packages/trusted-contribution/cloudbuild.yaml</summary> </details> <details><summary>serverless-scheduler-proxy/cloudbuild.yaml</summary> </details> <details><summary>trigger-sync/cloudbuild.yaml</summary> </details> </blockquote> </details> <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/ci.yaml</summary> - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/setup-node v3@2fddd8803e2f5c9604345a0b591c3020ee971a93` - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/setup-node v3@2fddd8803e2f5c9604345a0b591c3020ee971a93` - `actions/setup-go v3@268d8c0ca0432bb2cf416faae41297df9d262d7f` - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/setup-node v3@2fddd8803e2f5c9604345a0b591c3020ee971a93` </details> <details><summary>.github/workflows/deps.yaml</summary> - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/setup-node v3@2fddd8803e2f5c9604345a0b591c3020ee971a93` - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `actions/setup-node v3@2fddd8803e2f5c9604345a0b591c3020ee971a93` - `googleapis/code-suggester v4@212b0804e41c4f89edfefba37e1c39cdb9856e36` </details> <details><summary>.github/workflows/links.yaml</summary> - `actions/checkout v3@2541b1294d2704b0964813337f33b291d3f8596b` - `lycheeverse/lychee-action v1.5.1@4a5af7cd2958a2282cefbd9c10f63bdb89982d76` - `peter-evans/create-issue-from-file v4.0.0@99b87c35610e986ad2034a7b0518a9b3ebea541b` </details> </blockquote> </details> <details><summary>gomod</summary> <blockquote> <details><summary>packages/flakybot/go.mod</summary> - `go 1.19` - `cloud.google.com/go/pubsub v1.25.1` - `github.com/google/go-cmp v0.5.8` - `google.golang.org/api v0.94.0` </details> <details><summary>serverless-scheduler-proxy/go.mod</summary> - `go 1.19` - `cloud.google.com/go/secretmanager v1.5.0` - `github.com/google/uuid v1.3.0` - `google.golang.org/genproto v0.0.0-20220829175752-36a9c930ecbf@36a9c930ecbf` </details> </blockquote> </details> <details><summary>npm</summary> <blockquote> <details><summary>package.json</summary> - `c8 ^7.1.0` - `gaxios ^5.0.0` - `smee-client ^1.1.0` - `@google-cloud/mono-repo-publish ^1.0.0` - `node >=12.0.0` </details> <details><summary>packages/auto-approve/package.json</summary> - `@google-cloud/secret-manager ^4.1.1` - `ajv ^8.11.0` - `dayjs ^1.11.5` - `gcf-utils ^14.0.1` - `@octokit/rest ^19.0.4` - `@octokit/webhooks ^10.1.5` - `@octokit/webhooks-definitions ^3.67.3` - `@probot/octokit-plugin-config ^1.1.6` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `js-yaml ^4.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `ts-node ^10.9.1` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/auto-label/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/label-utils ^3.0.0` - `@google-cloud/storage ^6.4.2` - `gcf-utils ^14.0.1` - `@octokit/types ^7.2.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `js-yaml ^4.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/blunderbuss/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/datastore-lock ^4.0.0` - `@google-automations/label-utils ^3.0.0` - `gcf-utils ^14.0.1` - `@octokit/types ^7.1.1` - `@octokit/webhooks ^10.1.5` - `@types/mocha ^9.1.1` - `@types/node ^18.7.13` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `js-yaml ^4.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/bot-config-utils/package.json</summary> - `ajv ^8.11.0` - `ajv-formats ^2.1.1` - `gcf-utils ^14.0.1` - `js-yaml ^4.1.0` - `@octokit/types ^7.2.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/canary-bot/package.json</summary> - `dayjs ^1.11.5` - `gcf-utils ^14.0.1` - `lru-cache ^7.14.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `typescript ~4.8.2` - `node >= 14.0.0` </details> <details><summary>packages/cherry-pick-bot/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@octokit/rest ^19.0.4` - `gcf-utils ^14.0.1` - `yargs ^17.5.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/conventional-commit-lint/package.json</summary> - `@commitlint/config-conventional ^17.1.0` - `@commitlint/lint ^17.1.0` - `@google-automations/bot-config-utils ^6.1.0` - `@octokit/openapi-types ^13.8.0` - `@octokit/webhooks-types ^6.3.6` - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `js-yaml ^4.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/cron-utils/package.json</summary> - `@google-cloud/scheduler ^3.0.3` - `@googleapis/run ^10.0.2` - `google-auth-library ^7.14.1` - `js-yaml ^4.1.0` - `yargs ^17.5.1` - `@octokit/rest ^19.0.4` - `@types/js-yaml ^4.0.5` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `gts ^4.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/datastore-lock/package.json</summary> - `@google-cloud/datastore ^7.0.0` - `gcf-utils ^14.0.1` - `uuid ^9.0.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `google-datastore-emulator ^6.0.0` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/do-not-merge/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/failurechecker/package.json</summary> - `@google-automations/issue-utils ^2.0.0` - `gcf-utils ^14.0.1` - `@octokit/types ^7.2.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/flakybot/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/datastore-lock ^4.0.0` - `@google-automations/label-utils ^3.0.0` - `@octokit/openapi-types ^13.8.0` - `@octokit/types ^7.2.0` - `gcf-utils ^14.0.1` - `probot ^12.2.8` - `xml-js ^1.6.11` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/gcf-utils/package.json</summary> - `@google-cloud/kms ^3.0.1` - `@google-cloud/secret-manager ^4.1.1` - `@google-cloud/storage ^6.4.1` - `@google-cloud/tasks ^3.0.2` - `google-gax ^3.3.1` - `@googleapis/run ^10.0.2` - `@octokit/auth-app ^4.0.5` - `@octokit/plugin-enterprise-compatibility 2.0.2` - `@octokit/rest ^19.0.4` - `@probot/octokit-plugin-config ^1.1.6` - `@types/bunyan ^1.8.8` - `@types/dotenv ^6.1.1` - `@types/end-of-stream ^1.4.1` - `@types/express ^4.17.13` - `@types/into-stream ^3.1.1` - `@types/ioredis ^4.28.10` - `@types/lru-cache ^7.10.9` - `@types/sonic-boom ^2.1.0` - `@types/uuid ^8.3.4` - `body-parser ^1.20.0` - `express ^4.18.1` - `gaxios ^5.0.1` - `get-stream ^6.0.1` - `into-stream ^6.0.0` - `octokit-auth-probot ^1.2.8` - `pino ^8.4.2` - `probot ^12.2.7` - `tmp ^0.2.1` - `uuid ^9.0.0` - `yargs ^16.0.0` - `@types/get-stream ^3.0.2` - `@types/mocha ^9.1.1` - `@types/node ^18.7.14` - `@types/pino ^7.0.4` - `@types/sinon ^10.0.13` - `@types/tmp ^0.2.3` - `@types/yargs ^16.0.0` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `google-auth-library ^7.14.1` - `gts ^4.0.0` - `mocha ^10.0.0` - `mocked-env ^1.3.5` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `sonic-boom ^3.2.0` - `stream-mock ^2.0.5` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/generate-bot/package.json</summary> - `enquirer ^2.3.6` - `handlebars ^4.7.7` - `@types/mocha ^9.1.1` - `@types/recursive-readdir ^2.2.1` - `@types/rimraf ^3.0.2` - `c8 ^7.12.0` - `gts ^4.0.0` - `mocha ^10.0.0` - `recursive-readdir ^2.2.2` - `rimraf ^3.0.2` - `typescript ~4.8.2` - `node >=12.18.2` </details> <details><summary>packages/generated-files-bot/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `gcf-utils ^14.0.1` - `js-yaml ^4.1.0` - `jsonpath ^1.1.1` - `minimatch ^5.1.0` - `@types/js-yaml ^4.0.5` - `@types/jsonpath ^0.2.0` - `@types/minimatch ^5.1.2` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/git-file-utils/package.json</summary> - `@octokit/rest 19.0.4` - `minimatch ^5.1.0` - `@types/chai ^4.3.3` - `@types/minimatch ^5.1.2` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `chai ^4.3.6` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `gts ^4.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/header-checker-lint/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `gcf-utils ^14.0.1` - `minimatch ^5.1.0` - `@types/minimatch ^5.1.2` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/issue-utils/package.json</summary> - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/label-sync/package.json</summary> - `@google-cloud/storage ^6.4.2` - `gaxios ^5.0.1` - `gcf-utils ^14.0.1` - `@probot/octokit-plugin-config ^1.1.6` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 12.18.2` </details> <details><summary>packages/label-utils/package.json</summary> - `@google-automations/datastore-lock ^4.0.0` - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/loadtest-bot/package.json</summary> - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `smee-client ^1.2.3` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/merge-on-green/package.json</summary> - `@google-automations/label-utils ^3.0.0` - `@google-cloud/datastore ^7.0.0` - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/mono-repo-publish/package.json</summary> - `@octokit/auth-app ^4.0.5` - `@octokit/rest ^19.0.4` - `yargs ^17.5.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `semistandard ^16.0.1` - `sinon ^14.0.0` - `typescript ^4.8.2` - `node >= 12.18.2` </details> <details><summary>packages/object-selector/package.json</summary> - `ajv ^8.11.0` - `easy-object-selector ^1.3.0` - `gcf-utils ^14.0.1` - `js-yaml ^4.1.0` - `yargs ^17.5.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `dotenv ^16.0.2` - `gts ^4.0.0` - `mocha ^10.0.0` - `mocked-env ^1.3.5` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/owl-bot/package.json</summary> - `@google-automations/datastore-lock ^4.0.0` - `@google-automations/label-utils ^3.0.0` - `@google-cloud/cloudbuild ^3.1.0` - `@octokit/rest ^19.0.4` - `@octokit/webhooks-types ^6.3.6` - `adm-zip ^0.5.9` - `ajv ^8.11.0` - `code-suggester ^4.1.0` - `firebase-admin ^11.0.1` - `follow-redirects ^1.15.1` - `fs-extra ^10.1.0` - `gaxios ^5.0.1` - `gcf-utils ^14.0.1` - `glob ^8.0.3` - `js-yaml ^4.1.0` - `jsonwebtoken ^8.5.1` - `minimatch ^5.1.0` - `probot ^12.2.5` - `tmp ^0.2.1` - `yargs ^17.5.1` - `@types/adm-zip ^0.5.0` - `@types/follow-redirects ^1.14.1` - `@types/fs-extra ^9.0.13` - `@types/glob ^8.0.0` - `@types/js-yaml ^4.0.5` - `@types/jsonwebtoken ^8.5.9` - `@types/minimatch ^5.1.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.13` - `@types/rimraf ^3.0.2` - `@types/sinon ^10.0.13` - `@types/tmp ^0.2.3` - `@types/yargs ^17.0.11` - `c8 ^7.12.0` - `gts ^3.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `rimraf ^3.0.2` - `sinon ^14.0.0` - `ts-node ^10.9.1` - `typescript ~4.8.2` - `uuid ^9.0.0` - `node >=14.0.0` </details> <details><summary>packages/owlbot-bootstrapper/app/package.json</summary> - `probot ^12.2.5` - `@types/mocha ^9.1.1` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^3.1.0` - `mocha ^10.0.0` - `typescript ~4.7.4` - `node >= 14` </details> <details><summary>packages/owlbot-bootstrapper/cli/package.json</summary> - `@google-cloud/cloudbuild ^3.1.0` - `yargs ^17.5.1` - `@types/yargs ^17.0.11` - `@types/jwt-encode ^1.0.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.9` - `@types/node-fetch ^2.6.2` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^3.1.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `typescript ~4.7.4` - `node >= 14` </details> <details><summary>packages/owlbot-bootstrapper/common-container/package.json</summary> - `@google-cloud/cloudbuild ^3.1.0` - `@google-cloud/secret-manager ^4.0.0` - `@octokit/auth-app ^4.0.5` - `@octokit/rest ^19.0.4` - `@types/yargs ^17.0.11` - `fs 0.0.1-security` - `gaxios ^5.0.1` - `gcf-utils ^14.0.1` - `jsonwebtoken ^8.5.1` - `node-fetch ^3.2.10` - `nodejs-fetch ^1.0.0` - `snap-shot-it ^7.9.6` - `uuidv4 ^6.2.13` - `yargs ^17.5.1` - `@types/jwt-encode ^1.0.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.9` - `@types/node-fetch ^2.6.2` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^3.1.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `typescript ~4.7.4` - `node >= 14` </details> <details><summary>packages/owlbot-bootstrapper/package.json</summary> - `@google-cloud/cloudbuild ^3.1.0` - `@google-cloud/secret-manager ^4.1.1` - `@octokit/auth-app ^4.0.5` - `@octokit/rest ^19.0.4` - `@types/yargs ^17.0.12` - `fs 0.0.1-security` - `gaxios ^5.0.1` - `gcf-utils ^14.0.1` - `jsonwebtoken ^8.5.1` - `node-fetch ^3.2.10` - `nodejs-fetch ^1.0.0` - `snap-shot-it ^7.9.6` - `uuidv4 ^6.2.13` - `yargs ^17.5.1` - `@types/jwt-encode ^1.0.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/node-fetch ^2.6.2` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/policy/package.json</summary> - `@google-cloud/bigquery ^6.0.2` - `code-suggester ^4.0.1` - `gaxios ^5.0.1` - `gcf-utils ^14.0.1` - `lru-cache ^7.14.0` - `uuid ^9.0.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.9` - `@types/sinon ^10.0.13` - `@types/uuid ^8.3.4` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^3.1.0` - `meow ^9.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `typescript ~4.7.4` - `node >= 14` </details> <details><summary>packages/release-brancher/package.json</summary> - `@octokit/rest ^19.0.4` - `code-suggester ^4.1.0` - `js-yaml ^4.1.0` - `yargs ^17.5.1` - `@types/js-yaml ^4.0.5` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/pino ^7.0.4` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 12.16.0` </details> <details><summary>packages/release-please/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/issue-utils ^2.0.0` - `@google-automations/label-utils ^3.0.0` - `@octokit/rest ^19.0.4` - `@octokit/webhooks ^10.1.5` - `gcf-utils ^14.0.1` - `release-please ^14.3.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/semver ^7.3.12` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `js-yaml ^4.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14.0.0` </details> <details><summary>packages/release-trigger/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/datastore-lock ^4.0.0` - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/repo-metadata-lint/package.json</summary> - `@google-cloud/storage ^6.4.2` - `@octokit/rest ^19.0.4` - `ajv ^8.11.0` - `gaxios ^5.0.1` - `gcf-utils ^14.0.1` - `jsonwebtoken ^8.5.1` - `yargs ^17.5.1` - `@octokit/types ^7.2.0` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/secret-rotator/package.json</summary> - `@google-cloud/secret-manager ^3.12.0` - `@googleapis/iam ^3.0.1` - `express ^4.18.1` - `gcf-utils ^14.0.1` - `google-auth-library ^7.14.1` - `@types/express ^4.17.13` - `@types/mocha ^9.1.1` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `gaxios ^5.0.1` - `google-gax ^3.4.0` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `typescript ^4.8.2` - `node >=14.0.0` </details> <details><summary>packages/snippet-bot/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/git-file-utils ^1.2.0` - `@google-automations/label-utils ^3.0.0` - `@google-cloud/storage ^6.4.2` - `follow-redirects ^1.15.1` - `gcf-utils ^14.0.1` - `lru-cache ^7.14.0` - `minimatch ^5.1.0` - `parse-diff ^0.9.0` - `tar ^6.1.11` - `tmp-promise ^3.0.3` - `@octokit/webhooks-types ^6.3.6` - `@types/follow-redirects ^1.14.1` - `@types/minimatch ^5.1.2` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/tar ^6.1.2` - `@types/tmp ^0.2.3` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `smee-client ^1.2.3` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/sync-repo-settings/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-automations/issue-utils ^2.0.0` - `extend ^3.0.2` - `gcf-utils ^14.0.1` - `js-yaml ^4.1.0` - `yargs ^17.5.1` - `@types/extend ^3.0.1` - `@types/js-yaml ^4.0.5` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `@types/yargs ^17.0.12` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `mocked-env ^1.3.5` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `ts-node ^10.9.1` - `typescript ~4.8.2` - `node >= 14` </details> <details><summary>packages/trusted-contribution/package.json</summary> - `@google-automations/bot-config-utils ^6.1.0` - `@google-cloud/secret-manager ^4.1.1` - `gcf-utils ^14.0.1` - `@types/mocha ^9.1.1` - `@types/node ^18.7.15` - `@types/sinon ^10.0.13` - `c8 ^7.12.0` - `cross-env ^7.0.3` - `gts ^4.0.0` - `js-yaml ^4.1.0` - `lru-cache ^7.14.0` - `mocha ^10.0.0` - `nock ^13.2.9` - `sinon ^14.0.0` - `snap-shot-it ^7.9.6` - `typescript ~4.8.2` - `node >= 14` </details> </blockquote> </details> <details><summary>pip_requirements</summary> <blockquote> <details><summary>packages/release-trigger/requirements.txt</summary> - `gcp-releasetool ==1.8.7` </details> </blockquote> </details> --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue lists renovate updates and detected dependencies read the docs to learn more awaiting schedule these updates are awaiting their schedule click on a checkbox to get an update now fix deps update google golang org genproto digest to chore deps update dependency typescript to chore deps update dependency typescript to chore deps update dependency typescript to chore deps update dependency typescript to fix deps update module google golang org api to chore deps update dependency google auth library to chore deps update dependency meow to fix deps update dependency google cloud secret manager to fix deps update dependency google auth library to fix deps update dependency google auth library to fix deps update dependency into stream to fix deps update dependency yargs to yargs types yargs chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance chore deps lock file maintenance detected dependencies cloudbuild cloudbuild yaml packages auto approve cloudbuild yaml packages auto label cloudbuild yaml packages bazel bot cloudbuild yaml packages blunderbuss cloudbuild yaml packages canary bot cloudbuild yaml packages cherry pick bot cloudbuild yaml packages conventional commit lint cloudbuild yaml packages do not merge cloudbuild yaml packages failurechecker cloudbuild yaml packages flakybot cloudbuild yaml packages generated files bot cloudbuild yaml packages header checker lint cloudbuild yaml packages label sync cloudbuild yaml packages merge on green cloudbuild yaml packages owl bot cloudbuild yaml packages policy cloudbuild yaml packages release please cloudbuild yaml packages release trigger cloudbuild yaml packages repo metadata lint cloudbuild yaml packages snippet bot cloudbuild yaml packages sync repo settings cloudbuild yaml packages trusted contribution cloudbuild yaml serverless scheduler proxy cloudbuild yaml trigger sync cloudbuild yaml github actions github workflows ci yaml actions checkout actions setup node actions checkout actions setup node actions setup go actions checkout actions checkout actions checkout actions setup node github workflows deps yaml actions checkout actions setup node actions checkout actions setup node googleapis code suggester github workflows links yaml actions checkout lycheeverse lychee action peter evans create issue from file gomod packages flakybot go mod go cloud google com go pubsub github com google go cmp google golang org api serverless scheduler proxy go mod go cloud google com go secretmanager github com google uuid google golang org genproto npm package json gaxios smee client google cloud mono repo publish node packages auto approve package json google cloud secret manager ajv dayjs gcf utils octokit rest octokit webhooks octokit webhooks definitions probot octokit plugin config types mocha types node types sinon cross env gts js yaml lru cache mocha nock sinon smee client snap shot it ts node typescript node packages auto label package json google automations bot config utils google automations label utils google cloud storage gcf utils octokit types types mocha types node types sinon cross env gts js yaml lru cache mocha nock sinon smee client snap shot it typescript node packages blunderbuss package json google automations bot config utils google automations datastore lock google automations label utils gcf utils octokit types octokit webhooks types mocha types node types sinon cross env gts js yaml lru cache mocha nock sinon smee client snap shot it typescript node packages bot config utils package json ajv ajv formats gcf utils js yaml octokit types types mocha types node types sinon cross env dotenv gts lru cache mocha nock sinon snap shot it typescript node packages canary bot package json dayjs gcf utils lru cache types mocha types node types sinon cross env gts mocha nock sinon smee client typescript node packages cherry pick bot package json google automations bot config utils octokit rest gcf utils yargs types mocha types node types sinon types yargs cross env gts lru cache mocha nock sinon smee client snap shot it typescript node packages conventional commit lint package json commitlint config conventional commitlint lint google automations bot config utils octokit openapi types octokit webhooks types gcf utils types mocha types node types sinon cross env gts js yaml lru cache mocha nock sinon smee client snap shot it typescript node packages cron utils package json google cloud scheduler googleapis run google auth library js yaml yargs octokit rest types js yaml types mocha types node types sinon types yargs cross env dotenv gts mocha nock sinon snap shot it typescript node packages datastore lock package json google cloud datastore gcf utils uuid types mocha types node types sinon cross env dotenv google datastore emulator gts lru cache mocha nock sinon typescript node packages do not merge package json google automations bot config utils gcf utils types mocha types node types sinon cross env gts lru cache mocha nock sinon smee client snap shot it typescript node packages failurechecker package json google automations issue utils gcf utils octokit types types mocha types node types sinon cross env gts lru cache mocha nock sinon smee client snap shot it typescript node packages flakybot package json google automations bot config utils google automations datastore lock google automations label utils octokit openapi types octokit types gcf utils probot xml js types mocha types node types sinon cross env gts lru cache mocha nock sinon snap shot it typescript node packages gcf utils package json google cloud kms google cloud secret manager google cloud storage google cloud tasks google gax googleapis run octokit auth app octokit plugin enterprise compatibility octokit rest probot octokit plugin config types bunyan types dotenv types end of stream types express types into stream types ioredis types lru cache types sonic boom types uuid body parser express gaxios get stream into stream octokit auth probot pino probot tmp uuid yargs types get stream types mocha types node types pino types sinon types tmp types yargs cross env dotenv google auth library gts mocha mocked env nock sinon snap shot it sonic boom stream mock typescript node packages generate bot package json enquirer handlebars types mocha types recursive readdir types rimraf gts mocha recursive readdir rimraf typescript node packages generated files bot package json google automations bot config utils gcf utils js yaml jsonpath minimatch types js yaml types jsonpath types minimatch types mocha types node types sinon cross env gts lru cache mocha nock sinon smee client snap shot it typescript node packages git file utils package json octokit rest minimatch types chai types minimatch types mocha types node types sinon chai cross env dotenv gts mocha nock sinon snap shot it typescript node packages header checker lint package json google automations bot config utils gcf utils minimatch types minimatch types mocha types node types sinon cross env gts lru cache mocha nock sinon snap shot it typescript node packages issue utils package json gcf utils types mocha types node types sinon cross env dotenv gts lru cache mocha nock sinon snap shot it typescript node packages label sync package json google cloud storage gaxios gcf utils probot octokit plugin config types mocha types node types sinon cross env gts lru cache mocha nock sinon snap shot it typescript node packages label utils package json google automations datastore lock gcf utils types mocha types node types sinon cross env dotenv gts lru cache mocha nock sinon snap shot it typescript node packages loadtest bot package json gcf utils types mocha types node cross env gts lru cache mocha nock smee client typescript node packages merge on green package json google automations label utils google cloud datastore gcf utils types mocha types node types sinon cross env gts lru cache mocha nock sinon snap shot it typescript node packages mono repo publish package json octokit auth app octokit rest yargs types mocha types node types sinon types yargs cross env gts mocha nock semistandard sinon typescript node packages object selector package json ajv easy object selector gcf utils js yaml yargs types mocha types node types sinon types yargs cross env dotenv gts mocha mocked env nock sinon snap shot it typescript node packages owl bot package json google automations datastore lock google automations label utils google cloud cloudbuild octokit rest octokit webhooks types adm zip ajv code suggester firebase admin follow redirects fs extra gaxios gcf utils glob js yaml jsonwebtoken minimatch probot tmp yargs types adm zip types follow redirects types fs extra types glob types js yaml types jsonwebtoken types minimatch types mocha types node types rimraf types sinon types tmp types yargs gts lru cache mocha nock rimraf sinon ts node typescript uuid node packages owlbot bootstrapper app package json probot types mocha cross env gts mocha typescript node packages owlbot bootstrapper cli package json google cloud cloudbuild yargs types yargs types jwt encode types mocha types node types node fetch types sinon cross env gts mocha nock sinon smee client typescript node packages owlbot bootstrapper common container package json google cloud cloudbuild google cloud secret manager octokit auth app octokit rest types yargs fs security gaxios gcf utils jsonwebtoken node fetch nodejs fetch snap shot it yargs types jwt encode types mocha types node types node fetch types sinon cross env gts mocha nock sinon smee client typescript node packages owlbot bootstrapper package json google cloud cloudbuild google cloud secret manager octokit auth app octokit rest types yargs fs security gaxios gcf utils jsonwebtoken node fetch nodejs fetch snap shot it yargs types jwt encode types mocha types node types node fetch types sinon cross env gts lru cache mocha nock sinon smee client typescript node packages policy package json google cloud bigquery code suggester gaxios gcf utils lru cache uuid types mocha types node types sinon types uuid cross env gts meow mocha nock sinon smee client typescript node packages release brancher package json octokit rest code suggester js yaml yargs types js yaml types mocha types node types pino types sinon types yargs cross env gts mocha nock sinon snap shot it typescript node packages release please package json google automations bot config utils google automations issue utils google automations label utils octokit rest octokit webhooks gcf utils release please types mocha types node types semver types sinon cross env gts js yaml lru cache mocha nock sinon snap shot it typescript node packages release trigger package json google automations bot config utils google automations datastore lock gcf utils types mocha types node types sinon cross env gts lru cache mocha nock sinon smee client typescript node packages repo metadata lint package json google cloud storage octokit rest ajv gaxios gcf utils jsonwebtoken yargs octokit types types mocha types node types sinon types yargs cross env gts lru cache mocha nock sinon typescript node packages secret rotator package json google cloud secret manager googleapis iam express gcf utils google auth library types express types mocha types sinon gaxios google gax gts lru cache mocha nock sinon typescript node packages snippet bot package json google automations bot config utils google automations git file utils google automations label utils google cloud storage follow redirects gcf utils lru cache minimatch parse diff tar tmp promise octokit webhooks types types follow redirects types minimatch types mocha types node types sinon types tar types tmp cross env gts mocha nock sinon smee client snap shot it typescript node packages sync repo settings package json google automations bot config utils google automations issue utils extend gcf utils js yaml yargs types extend types js yaml types mocha types node types sinon types yargs cross env gts lru cache mocha mocked env nock sinon snap shot it ts node typescript node packages trusted contribution package json google automations bot config utils google cloud secret manager gcf utils types mocha types node types sinon cross env gts js yaml lru cache mocha nock sinon snap shot it typescript node pip requirements packages release trigger requirements txt gcp releasetool check this box to trigger a request for renovate to run again on this repository
1
2,563
5,316,086,321
IssuesEvent
2017-02-13 18:59:16
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
closed
|subtitles] [fr] Manifestation du 23 juin contre la loi El Khomri : « Hollande doit rétropédaler » - Mélenchon
Language: French Process: [6] Approved
# Video title Manifestation du 23 juin contre la loi El Khomri : « Hollande doit rétropédaler » - Mélenchon # URL https://www.youtube.com/watch?v=SEaI67zXEAI # Youtube subtitles language French # Duration 1:55 # Subtitles URL https://www.youtube.com/timedtext_editor?lang=fr&bl=vmp&ui=hd&v=SEaI67zXEAI&tab=captions&ref=player&action_mde_edit_form=1
1.0
|subtitles] [fr] Manifestation du 23 juin contre la loi El Khomri : « Hollande doit rétropédaler » - Mélenchon - # Video title Manifestation du 23 juin contre la loi El Khomri : « Hollande doit rétropédaler » - Mélenchon # URL https://www.youtube.com/watch?v=SEaI67zXEAI # Youtube subtitles language French # Duration 1:55 # Subtitles URL https://www.youtube.com/timedtext_editor?lang=fr&bl=vmp&ui=hd&v=SEaI67zXEAI&tab=captions&ref=player&action_mde_edit_form=1
process
subtitles manifestation du juin contre la loi el khomri « hollande doit rétropédaler » mélenchon video title manifestation du juin contre la loi el khomri « hollande doit rétropédaler » mélenchon url youtube subtitles language french duration subtitles url
1
91,168
18,358,240,627
IssuesEvent
2021-10-08 21:53:22
microsoft/vscode-jupyter
https://api.github.com/repos/microsoft/vscode-jupyter
closed
Auto install ipykernel doesn't work on Devcontainer/Codespaces
bug upstream-python verification-needed codespaces
1. Create a Codespace from https://github.com/kieferrm/vscode-remote-python 2. Switch active interpreter to /usr/bin/python3 3. Install pip with `sudo apt update && sudo apt install python3-pip` 4. Open interactive window or notebook and try to run cells 5. 🐛 ipykernel install fails with the following spew: <details> ``` vscode ➜ /workspaces/vscode-remote-python (main ✗) $ /usr/bin/python3 /home/vscode/.vscode-remote/extensions/ms-python.python-2021.6.944021595/pythonFiles/shell_exec.py /usr/bin/python3 -m pip install -U --force-reinstall ipykernel --user /tmp/tmp-663b0naQ0jTIxQ9.log Executing command in shell >> /usr/bin/python3 -m pip install -U --force-reinstall ipykernel --user Collecting ipykernel Using cached https://files.pythonhosted.org/packages/b6/4a/b9e7d67637549e1b75c2946e93e35740ea63a521d7d455343e6a8a8a7b91/ipykernel-6.0.1-py3-none-any.whl Collecting tornado>=4.2 (from ipykernel) Using cached https://files.pythonhosted.org/packages/bf/fa/2befee379094720b54065daa9c6117f3edb7d35f86cde0f50b3a28ecfadf/tornado-6.1-cp37-cp37m-manylinux1_x86_64.whl Collecting jupyter-client (from ipykernel) Using cached https://files.pythonhosted.org/packages/88/4e/50fcf8b38d9c08d5b4839c1650e595f6bfa4fc9b419e2b800db8f14ee532/jupyter_client-6.2.0-py3-none-any.whl Collecting ipython>=7.23.1 (from ipykernel) Using cached https://files.pythonhosted.org/packages/f7/b2/733ea4551a04866bbcfbbade4d9d2c82c829cf1cc6fac1ac5974b8c7f756/ipython-7.25.0-py3-none-any.whl Collecting debugpy>=1.0.0 (from ipykernel) Using cached https://files.pythonhosted.org/packages/84/89/bb91c5e3b676721d3a9c13489767b2ec83c7035464e68e8e2fe477ad4566/debugpy-1.3.0-cp37-cp37m-manylinux1_x86_64.whl Collecting importlib-metadata<4; python_version < "3.8.0" (from ipykernel) Using cached https://files.pythonhosted.org/packages/52/d0/bdb31463f2d9ca111e39b268518e9baa3542ef73ca449b711a7b4da69764/importlib_metadata-3.10.1-py3-none-any.whl Collecting traitlets>=4.1.0 (from ipykernel) Using cached https://files.pythonhosted.org/packages/f6/7d/3ecb0ebd0ce8dcdfa7bd47ab85c1d4a521e6770ef283d0824f5804994dfe/traitlets-5.0.5-py3-none-any.whl Collecting jupyter-core>=4.6.0 (from jupyter-client->ipykernel) Using cached https://files.pythonhosted.org/packages/53/40/5af36bffa0af3ac71d3a6fc6709de10e4f6ff7c01745b8bc4715372189c9/jupyter_core-4.7.1-py3-none-any.whl Collecting python-dateutil>=2.1 (from jupyter-client->ipykernel) Using cached https://files.pythonhosted.org/packages/d4/70/d60450c3dd48ef87586924207ae8907090de0b306af2bce5d134d78615cb/python_dateutil-2.8.1-py2.py3-none-any.whl Collecting nest-asyncio>=1.5 (from jupyter-client->ipykernel) Using cached https://files.pythonhosted.org/packages/52/e2/9b37da54e6e9094d2f558ae643d1954a0fa8215dfee4fa261f31c5439796/nest_asyncio-1.5.1-py3-none-any.whl Collecting pyzmq>=13 (from jupyter-client->ipykernel) Using cached https://files.pythonhosted.org/packages/2d/4b/256767441a6c4223fbc7c3eab1b3762d9d70b657899d844228a782f48af3/pyzmq-22.1.0-cp37-cp37m-manylinux1_x86_64.whl Collecting pickleshare (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/9a/41/220f49aaea88bc6fa6cba8d05ecf24676326156c23b991e80b3f2fc24c77/pickleshare-0.7.5-py2.py3-none-any.whl Collecting matplotlib-inline (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/7f/de/6c111d687335729cf8c156394c8d119b0dc3c34b6966ff2a2f7fe4aa79cf/matplotlib_inline-0.1.2-py3-none-any.whl Collecting setuptools>=18.5 (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/a2/e1/902fbc2f61ad6243cd3d57ffa195a9eb123021ec912ec5d811acf54a39f8/setuptools-57.1.0-py3-none-any.whl Collecting jedi>=0.16 (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/f9/36/7aa67ae2663025b49e8426ead0bad983fee1b73f472536e9790655da0277/jedi-0.18.0-py2.py3-none-any.whl Collecting pygments (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/a6/c9/be11fce9810793676017f79ffab3c6cb18575844a6c7b8d4ed92f95de604/Pygments-2.9.0-py3-none-any.whl Collecting pexpect>4.3; sys_platform != "win32" (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/39/7b/88dbb785881c28a102619d46423cb853b46dbccc70d3ac362d99773a78ce/pexpect-4.8.0-py2.py3-none-any.whl Collecting prompt-toolkit!=3.0.0,!=3.0.1,<3.1.0,>=2.0.0 (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/1e/34/1d9880ac1339ad4c6697b330e7a507584105613751318249d9e820faa25f/prompt_toolkit-3.0.19-py3-none-any.whl Collecting decorator (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/6a/36/b1b9bfdf28690ae01d9ca0aa5b0d07cb4448ac65fb91dc7e2d094e3d992f/decorator-5.0.9-py3-none-any.whl Collecting backcall (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/4c/1c/ff6546b6c12603d8dd1070aa3c3d273ad4c07f5771689a7b69a550e8c951/backcall-0.2.0-py2.py3-none-any.whl Collecting typing-extensions>=3.6.4; python_version < "3.8" (from importlib-metadata<4; python_version < "3.8.0"->ipykernel) Using cached https://files.pythonhosted.org/packages/2e/35/6c4fff5ab443b57116cb1aad46421fb719bed2825664e8fe77d66d99bcbc/typing_extensions-3.10.0.0-py3-none-any.whl Collecting zipp>=0.5 (from importlib-metadata<4; python_version < "3.8.0"->ipykernel) Using cached https://files.pythonhosted.org/packages/92/d9/89f433969fb8dc5b9cbdd4b4deb587720ec1aeb59a020cf15002b9593eef/zipp-3.5.0-py3-none-any.whl Collecting ipython-genutils (from traitlets>=4.1.0->ipykernel) Using cached https://files.pythonhosted.org/packages/fa/bc/9bd3b5c2b4774d5f33b2d544f1460be9df7df2fe42f352135381c347c69a/ipython_genutils-0.2.0-py2.py3-none-any.whl Collecting six>=1.5 (from python-dateutil>=2.1->jupyter-client->ipykernel) Using cached https://files.pythonhosted.org/packages/d9/5a/e7c31adbe875f2abbb91bd84cf2dc52d792b5a01506781dbcf25c91daf11/six-1.16.0-py2.py3-none-any.whl Collecting parso<0.9.0,>=0.8.0 (from jedi>=0.16->ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/a9/c4/d5476373088c120ffed82f34c74b266ccae31a68d665b837354d4d8dc8be/parso-0.8.2-py2.py3-none-any.whl Collecting ptyprocess>=0.5 (from pexpect>4.3; sys_platform != "win32"->ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/22/a6/858897256d0deac81a172289110f31629fc4cee19b6f01283303e18c8db3/ptyprocess-0.7.0-py2.py3-none-any.whl Collecting wcwidth (from prompt-toolkit!=3.0.0,!=3.0.1,<3.1.0,>=2.0.0->ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/59/7c/e39aca596badaf1b78e8f547c807b04dae603a433d3e7a7e04d67f2ef3e5/wcwidth-0.2.5-py2.py3-none-any.whl Installing collected packages: tornado, ipython-genutils, traitlets, jupyter-core, six, python-dateutil, nest-asyncio, pyzmq, jupyter-client, pickleshare, matplotlib-inline, setuptools, parso, jedi, pygments, ptyprocess, pexpect, wcwidth, prompt-toolkit, decorator, backcall, ipython, debugpy, typing-extensions, zipp, importlib-metadata, ipykernel Exception: Traceback (most recent call last): File "/usr/lib/python3/dist-packages/pip/_internal/cli/base_command.py", line 143, in main status = self.run(options, args) File "/usr/lib/python3/dist-packages/pip/_internal/commands/install.py", line 386, in run use_user_site=options.use_user_site, File "/usr/lib/python3/dist-packages/pip/_internal/req/__init__.py", line 49, in install_given_reqs **kwargs File "/usr/lib/python3/dist-packages/pip/_internal/req/req_install.py", line 760, in install use_user_site=use_user_site, pycompile=pycompile, File "/usr/lib/python3/dist-packages/pip/_internal/req/req_install.py", line 382, in move_wheel_files warn_script_location=warn_script_location, File "/usr/lib/python3/dist-packages/pip/_internal/wheel.py", line 215, in move_wheel_files prefix=prefix, File "/usr/lib/python3/dist-packages/pip/_internal/locations.py", line 165, in distutils_scheme i.finalize_options() File "/usr/lib/python3.7/distutils/command/install.py", line 274, in finalize_options raise DistutilsOptionError("can't combine user with prefix, " distutils.errors.DistutilsOptionError: can't combine user with prefix, exec_prefix/home, or install_(plat)base Traceback (most recent call last): File "/home/vscode/.vscode-remote/extensions/ms-python.python-2021.6.944021595/pythonFiles/shell_exec.py", line 26, in <module> subprocess.check_call(shell_args, stdout=sys.stdout, stderr=sys.stderr) File "/usr/lib/python3.7/subprocess.py", line 347, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['/usr/bin/python3', '-m', 'pip', 'install', '-U', '--force-reinstall', 'ipykernel', '--user']' returned non-zero exit status 2. ``` </details> 6. Run `/usr/bin/python3 -m pip install ipykernel` and installation completes successfully 7. 🐛 Retry running cells. Newly installed ipykernel is not detected 8. 🐛 ipykernel install is attempted twice, once for main kernel and once for backup kernel used for fast switching
1.0
Auto install ipykernel doesn't work on Devcontainer/Codespaces - 1. Create a Codespace from https://github.com/kieferrm/vscode-remote-python 2. Switch active interpreter to /usr/bin/python3 3. Install pip with `sudo apt update && sudo apt install python3-pip` 4. Open interactive window or notebook and try to run cells 5. 🐛 ipykernel install fails with the following spew: <details> ``` vscode ➜ /workspaces/vscode-remote-python (main ✗) $ /usr/bin/python3 /home/vscode/.vscode-remote/extensions/ms-python.python-2021.6.944021595/pythonFiles/shell_exec.py /usr/bin/python3 -m pip install -U --force-reinstall ipykernel --user /tmp/tmp-663b0naQ0jTIxQ9.log Executing command in shell >> /usr/bin/python3 -m pip install -U --force-reinstall ipykernel --user Collecting ipykernel Using cached https://files.pythonhosted.org/packages/b6/4a/b9e7d67637549e1b75c2946e93e35740ea63a521d7d455343e6a8a8a7b91/ipykernel-6.0.1-py3-none-any.whl Collecting tornado>=4.2 (from ipykernel) Using cached https://files.pythonhosted.org/packages/bf/fa/2befee379094720b54065daa9c6117f3edb7d35f86cde0f50b3a28ecfadf/tornado-6.1-cp37-cp37m-manylinux1_x86_64.whl Collecting jupyter-client (from ipykernel) Using cached https://files.pythonhosted.org/packages/88/4e/50fcf8b38d9c08d5b4839c1650e595f6bfa4fc9b419e2b800db8f14ee532/jupyter_client-6.2.0-py3-none-any.whl Collecting ipython>=7.23.1 (from ipykernel) Using cached https://files.pythonhosted.org/packages/f7/b2/733ea4551a04866bbcfbbade4d9d2c82c829cf1cc6fac1ac5974b8c7f756/ipython-7.25.0-py3-none-any.whl Collecting debugpy>=1.0.0 (from ipykernel) Using cached https://files.pythonhosted.org/packages/84/89/bb91c5e3b676721d3a9c13489767b2ec83c7035464e68e8e2fe477ad4566/debugpy-1.3.0-cp37-cp37m-manylinux1_x86_64.whl Collecting importlib-metadata<4; python_version < "3.8.0" (from ipykernel) Using cached https://files.pythonhosted.org/packages/52/d0/bdb31463f2d9ca111e39b268518e9baa3542ef73ca449b711a7b4da69764/importlib_metadata-3.10.1-py3-none-any.whl Collecting traitlets>=4.1.0 (from ipykernel) Using cached https://files.pythonhosted.org/packages/f6/7d/3ecb0ebd0ce8dcdfa7bd47ab85c1d4a521e6770ef283d0824f5804994dfe/traitlets-5.0.5-py3-none-any.whl Collecting jupyter-core>=4.6.0 (from jupyter-client->ipykernel) Using cached https://files.pythonhosted.org/packages/53/40/5af36bffa0af3ac71d3a6fc6709de10e4f6ff7c01745b8bc4715372189c9/jupyter_core-4.7.1-py3-none-any.whl Collecting python-dateutil>=2.1 (from jupyter-client->ipykernel) Using cached https://files.pythonhosted.org/packages/d4/70/d60450c3dd48ef87586924207ae8907090de0b306af2bce5d134d78615cb/python_dateutil-2.8.1-py2.py3-none-any.whl Collecting nest-asyncio>=1.5 (from jupyter-client->ipykernel) Using cached https://files.pythonhosted.org/packages/52/e2/9b37da54e6e9094d2f558ae643d1954a0fa8215dfee4fa261f31c5439796/nest_asyncio-1.5.1-py3-none-any.whl Collecting pyzmq>=13 (from jupyter-client->ipykernel) Using cached https://files.pythonhosted.org/packages/2d/4b/256767441a6c4223fbc7c3eab1b3762d9d70b657899d844228a782f48af3/pyzmq-22.1.0-cp37-cp37m-manylinux1_x86_64.whl Collecting pickleshare (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/9a/41/220f49aaea88bc6fa6cba8d05ecf24676326156c23b991e80b3f2fc24c77/pickleshare-0.7.5-py2.py3-none-any.whl Collecting matplotlib-inline (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/7f/de/6c111d687335729cf8c156394c8d119b0dc3c34b6966ff2a2f7fe4aa79cf/matplotlib_inline-0.1.2-py3-none-any.whl Collecting setuptools>=18.5 (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/a2/e1/902fbc2f61ad6243cd3d57ffa195a9eb123021ec912ec5d811acf54a39f8/setuptools-57.1.0-py3-none-any.whl Collecting jedi>=0.16 (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/f9/36/7aa67ae2663025b49e8426ead0bad983fee1b73f472536e9790655da0277/jedi-0.18.0-py2.py3-none-any.whl Collecting pygments (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/a6/c9/be11fce9810793676017f79ffab3c6cb18575844a6c7b8d4ed92f95de604/Pygments-2.9.0-py3-none-any.whl Collecting pexpect>4.3; sys_platform != "win32" (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/39/7b/88dbb785881c28a102619d46423cb853b46dbccc70d3ac362d99773a78ce/pexpect-4.8.0-py2.py3-none-any.whl Collecting prompt-toolkit!=3.0.0,!=3.0.1,<3.1.0,>=2.0.0 (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/1e/34/1d9880ac1339ad4c6697b330e7a507584105613751318249d9e820faa25f/prompt_toolkit-3.0.19-py3-none-any.whl Collecting decorator (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/6a/36/b1b9bfdf28690ae01d9ca0aa5b0d07cb4448ac65fb91dc7e2d094e3d992f/decorator-5.0.9-py3-none-any.whl Collecting backcall (from ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/4c/1c/ff6546b6c12603d8dd1070aa3c3d273ad4c07f5771689a7b69a550e8c951/backcall-0.2.0-py2.py3-none-any.whl Collecting typing-extensions>=3.6.4; python_version < "3.8" (from importlib-metadata<4; python_version < "3.8.0"->ipykernel) Using cached https://files.pythonhosted.org/packages/2e/35/6c4fff5ab443b57116cb1aad46421fb719bed2825664e8fe77d66d99bcbc/typing_extensions-3.10.0.0-py3-none-any.whl Collecting zipp>=0.5 (from importlib-metadata<4; python_version < "3.8.0"->ipykernel) Using cached https://files.pythonhosted.org/packages/92/d9/89f433969fb8dc5b9cbdd4b4deb587720ec1aeb59a020cf15002b9593eef/zipp-3.5.0-py3-none-any.whl Collecting ipython-genutils (from traitlets>=4.1.0->ipykernel) Using cached https://files.pythonhosted.org/packages/fa/bc/9bd3b5c2b4774d5f33b2d544f1460be9df7df2fe42f352135381c347c69a/ipython_genutils-0.2.0-py2.py3-none-any.whl Collecting six>=1.5 (from python-dateutil>=2.1->jupyter-client->ipykernel) Using cached https://files.pythonhosted.org/packages/d9/5a/e7c31adbe875f2abbb91bd84cf2dc52d792b5a01506781dbcf25c91daf11/six-1.16.0-py2.py3-none-any.whl Collecting parso<0.9.0,>=0.8.0 (from jedi>=0.16->ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/a9/c4/d5476373088c120ffed82f34c74b266ccae31a68d665b837354d4d8dc8be/parso-0.8.2-py2.py3-none-any.whl Collecting ptyprocess>=0.5 (from pexpect>4.3; sys_platform != "win32"->ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/22/a6/858897256d0deac81a172289110f31629fc4cee19b6f01283303e18c8db3/ptyprocess-0.7.0-py2.py3-none-any.whl Collecting wcwidth (from prompt-toolkit!=3.0.0,!=3.0.1,<3.1.0,>=2.0.0->ipython>=7.23.1->ipykernel) Using cached https://files.pythonhosted.org/packages/59/7c/e39aca596badaf1b78e8f547c807b04dae603a433d3e7a7e04d67f2ef3e5/wcwidth-0.2.5-py2.py3-none-any.whl Installing collected packages: tornado, ipython-genutils, traitlets, jupyter-core, six, python-dateutil, nest-asyncio, pyzmq, jupyter-client, pickleshare, matplotlib-inline, setuptools, parso, jedi, pygments, ptyprocess, pexpect, wcwidth, prompt-toolkit, decorator, backcall, ipython, debugpy, typing-extensions, zipp, importlib-metadata, ipykernel Exception: Traceback (most recent call last): File "/usr/lib/python3/dist-packages/pip/_internal/cli/base_command.py", line 143, in main status = self.run(options, args) File "/usr/lib/python3/dist-packages/pip/_internal/commands/install.py", line 386, in run use_user_site=options.use_user_site, File "/usr/lib/python3/dist-packages/pip/_internal/req/__init__.py", line 49, in install_given_reqs **kwargs File "/usr/lib/python3/dist-packages/pip/_internal/req/req_install.py", line 760, in install use_user_site=use_user_site, pycompile=pycompile, File "/usr/lib/python3/dist-packages/pip/_internal/req/req_install.py", line 382, in move_wheel_files warn_script_location=warn_script_location, File "/usr/lib/python3/dist-packages/pip/_internal/wheel.py", line 215, in move_wheel_files prefix=prefix, File "/usr/lib/python3/dist-packages/pip/_internal/locations.py", line 165, in distutils_scheme i.finalize_options() File "/usr/lib/python3.7/distutils/command/install.py", line 274, in finalize_options raise DistutilsOptionError("can't combine user with prefix, " distutils.errors.DistutilsOptionError: can't combine user with prefix, exec_prefix/home, or install_(plat)base Traceback (most recent call last): File "/home/vscode/.vscode-remote/extensions/ms-python.python-2021.6.944021595/pythonFiles/shell_exec.py", line 26, in <module> subprocess.check_call(shell_args, stdout=sys.stdout, stderr=sys.stderr) File "/usr/lib/python3.7/subprocess.py", line 347, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['/usr/bin/python3', '-m', 'pip', 'install', '-U', '--force-reinstall', 'ipykernel', '--user']' returned non-zero exit status 2. ``` </details> 6. Run `/usr/bin/python3 -m pip install ipykernel` and installation completes successfully 7. 🐛 Retry running cells. Newly installed ipykernel is not detected 8. 🐛 ipykernel install is attempted twice, once for main kernel and once for backup kernel used for fast switching
non_process
auto install ipykernel doesn t work on devcontainer codespaces create a codespace from switch active interpreter to usr bin install pip with sudo apt update sudo apt install pip open interactive window or notebook and try to run cells 🐛 ipykernel install fails with the following spew vscode ➜ workspaces vscode remote python main ✗ usr bin home vscode vscode remote extensions ms python python pythonfiles shell exec py usr bin m pip install u force reinstall ipykernel user tmp tmp log executing command in shell usr bin m pip install u force reinstall ipykernel user collecting ipykernel using cached collecting tornado from ipykernel using cached collecting jupyter client from ipykernel using cached collecting ipython from ipykernel using cached collecting debugpy from ipykernel using cached collecting importlib metadata python version from ipykernel using cached collecting traitlets from ipykernel using cached collecting jupyter core from jupyter client ipykernel using cached collecting python dateutil from jupyter client ipykernel using cached collecting nest asyncio from jupyter client ipykernel using cached collecting pyzmq from jupyter client ipykernel using cached collecting pickleshare from ipython ipykernel using cached collecting matplotlib inline from ipython ipykernel using cached collecting setuptools from ipython ipykernel using cached collecting jedi from ipython ipykernel using cached collecting pygments from ipython ipykernel using cached collecting pexpect sys platform from ipython ipykernel using cached collecting prompt toolkit from ipython ipykernel using cached collecting decorator from ipython ipykernel using cached collecting backcall from ipython ipykernel using cached collecting typing extensions python version ipykernel using cached collecting zipp from importlib metadata ipykernel using cached collecting ipython genutils from traitlets ipykernel using cached collecting six from python dateutil jupyter client ipykernel using cached collecting parso from jedi ipython ipykernel using cached collecting ptyprocess from pexpect sys platform ipython ipykernel using cached collecting wcwidth from prompt toolkit ipython ipykernel using cached installing collected packages tornado ipython genutils traitlets jupyter core six python dateutil nest asyncio pyzmq jupyter client pickleshare matplotlib inline setuptools parso jedi pygments ptyprocess pexpect wcwidth prompt toolkit decorator backcall ipython debugpy typing extensions zipp importlib metadata ipykernel exception traceback most recent call last file usr lib dist packages pip internal cli base command py line in main status self run options args file usr lib dist packages pip internal commands install py line in run use user site options use user site file usr lib dist packages pip internal req init py line in install given reqs kwargs file usr lib dist packages pip internal req req install py line in install use user site use user site pycompile pycompile file usr lib dist packages pip internal req req install py line in move wheel files warn script location warn script location file usr lib dist packages pip internal wheel py line in move wheel files prefix prefix file usr lib dist packages pip internal locations py line in distutils scheme i finalize options file usr lib distutils command install py line in finalize options raise distutilsoptionerror can t combine user with prefix distutils errors distutilsoptionerror can t combine user with prefix exec prefix home or install plat base traceback most recent call last file home vscode vscode remote extensions ms python python pythonfiles shell exec py line in subprocess check call shell args stdout sys stdout stderr sys stderr file usr lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status run usr bin m pip install ipykernel and installation completes successfully 🐛 retry running cells newly installed ipykernel is not detected 🐛 ipykernel install is attempted twice once for main kernel and once for backup kernel used for fast switching
0
86,249
8,030,077,503
IssuesEvent
2018-07-27 18:16:55
Microsoft/AzureStorageExplorer
https://api.github.com/repos/Microsoft/AzureStorageExplorer
closed
New Created container doesn't have all property unless refresh both in the global admin and global tenant
:white_check_mark: merged testing
Storage Explorer Version: 1.4.0 Platform/OS Version: win10 Steps to Reproduce: 1.Install the 1.4.0 version Storage Explorer ,Import the certificate and connect one storage account. 2.Create one new container. 3.Right Click the new container. Expected Experience: The new container has all properties. Actual Experience: You just can see some of them, but it will show all after refreshing and right click.
1.0
New Created container doesn't have all property unless refresh both in the global admin and global tenant - Storage Explorer Version: 1.4.0 Platform/OS Version: win10 Steps to Reproduce: 1.Install the 1.4.0 version Storage Explorer ,Import the certificate and connect one storage account. 2.Create one new container. 3.Right Click the new container. Expected Experience: The new container has all properties. Actual Experience: You just can see some of them, but it will show all after refreshing and right click.
non_process
new created container doesn t have all property unless refresh both in the global admin and global tenant storage explorer version platform os version steps to reproduce install the version storage explorer import the certificate and connect one storage account create one new container right click the new container expected experience the new container has all properties actual experience you just can see some of them but it will show all after refreshing and right click
0
284,408
21,417,498,861
IssuesEvent
2022-04-22 12:27:05
hmathieu-insat/PIR
https://api.github.com/repos/hmathieu-insat/PIR
closed
Make a diff of the Old code and the new code
documentation
- [ ] Get the new Conti code (posted yesterday) on the gitlab of Clement - [ ] Compare every file with the Conti code we had at the beginning Snippet of Powershell code that should probably do the trick: `$fso = Get-ChildItem -Recurse -path _Path-of-the-former-code_` `$fsoBU = Get-ChildItem -Recurse -path _Path-of-the-new-code_` `Compare-Object -ReferenceObject (Get-Content $fso) -DifferenceObject (Get-Content $fsoBU)`
1.0
Make a diff of the Old code and the new code - - [ ] Get the new Conti code (posted yesterday) on the gitlab of Clement - [ ] Compare every file with the Conti code we had at the beginning Snippet of Powershell code that should probably do the trick: `$fso = Get-ChildItem -Recurse -path _Path-of-the-former-code_` `$fsoBU = Get-ChildItem -Recurse -path _Path-of-the-new-code_` `Compare-Object -ReferenceObject (Get-Content $fso) -DifferenceObject (Get-Content $fsoBU)`
non_process
make a diff of the old code and the new code get the new conti code posted yesterday on the gitlab of clement compare every file with the conti code we had at the beginning snippet of powershell code that should probably do the trick fso get childitem recurse path path of the former code fsobu get childitem recurse path path of the new code compare object referenceobject get content fso differenceobject get content fsobu
0
19,690
26,042,490,602
IssuesEvent
2022-12-22 11:39:17
threefoldtech/www-mastodon
https://api.github.com/repos/threefoldtech/www-mastodon
closed
A randomly generated name was added before the configured mastodon instance name
process_wontfix
### Description A randomly generated name was added before the mastodon instance name that i entered. ![image](https://user-images.githubusercontent.com/101194226/204253194-a8fecdff-0567-43c3-89aa-a3f7c6e72861.png) as seen in the screenshot above the name i entered was `mastodontest123` and the domain was created with `md7906mastodontest123`
1.0
A randomly generated name was added before the configured mastodon instance name - ### Description A randomly generated name was added before the mastodon instance name that i entered. ![image](https://user-images.githubusercontent.com/101194226/204253194-a8fecdff-0567-43c3-89aa-a3f7c6e72861.png) as seen in the screenshot above the name i entered was `mastodontest123` and the domain was created with `md7906mastodontest123`
process
a randomly generated name was added before the configured mastodon instance name description a randomly generated name was added before the mastodon instance name that i entered as seen in the screenshot above the name i entered was and the domain was created with
1
16,118
20,380,326,807
IssuesEvent
2022-02-21 20:45:29
ethereumclassic/ECIPs
https://api.github.com/repos/ethereumclassic/ECIPs
closed
ETC Core Devs Call 22: Proposed Rejection of ECIP-1049
status:8 rejected status:1 draft status:0 wip meta:5 call meta:1 governance meta:3 process
Agenda: https://ethereumclassic.org/blog/2022-02-21-core-devs-call-22-ecip-1049-proposed-rejection Discuss fate of ECIP-1049 after three years. (ECIP-1000 clause). Focus: REJECT Keccak-256 Mining Algorithm Change due to a high-probability risk of Contentious Chain Split between GPU Miners on ETCHash and FPGA & ASIC Miners on Keccak-256. ECIP-1049 is in violation of Ethereum Classic founding documents and the ECIP process. At this point, the contentious proposal has negative externalities on the network and is a resource drain. Move to reject the proposal after three years of technical discussion. If the proposal is not rejected, begin plans for a chain split. Decision to be made: + Move to `Rejected` status, or; + Set a Block and Push Contentious Fork with a `Chain Split` ECIP-1000 Clause: "ECIPs should be changed from Draft or Last Call status, to Rejected, upon request by any person, if they have not made progress in three years. Such a ECIP may be changed to Draft status if the champion provides revisions that meaningfully address public criticism of the proposal, or to Last Call if it meets the criteria required as described in the previous paragraph." The reason for "Rejected" status under this clause is that the champion (Alex Tsankov) has not met this requirement during the three years: "the champion provides revisions that meaningfully address public criticism of the proposal". Rather, the champion (Alex Tsankov) has ignored much valid criticism and abandoned the proposal. https://ecips.ethereumclassic.org/ECIPs/ecip-1000 Follow Up from: https://github.com/ethereumclassic/ECIPs/issues/382 Formal Proposed Rejection: https://github.com/ethereumclassic/ECIPs/issues/394#issuecomment-1022909537 Documented Github Opposition: https://github.com/ethereumclassic/ECIPs/issues/394#issuecomment-828160552 If time permits: review ECIP-1094 and ECIP-1096 for activity. Newer proposals but appear to be abandoned by the authors as well. Should these be `Withdrawn`? Please review the issue thread to find the most up to date information. Related Discussions: ECIP-1049: Change the ETC Proof of Work Algorithm to Keccak256 #8 ECIP-1049: Change the ETC Proof of Work Algorithm to Keccak-256 #13 ETC Core Devs Call(s) 2020 Q3: Hardfork #333 ECIP-1095: Change ETC PoW to "vanilla" Sha-3 Discussion #342 ETC Core Devs Call 13 & 14 #362 ETC Core Devs Call 15 - ECIP-1049 Breakout Session Keccak-256 #382 SHA3 Precompile ethereum/EIPs#2951 Core Devs Call 15 Recording https://vimeo.com/464336957 Change the ETC Proof of Work Algorithm to Keccak-256 #394 Admin Clean Up on ecip-1049: https://github.com/ethereumclassic/ECIPs/pull/400 Core Devs Call 19 Recording https://www.youtube.com/watch?v=WySNxZbDEkQ Community Call 005 Recording https://youtu.be/HaDANZN-ZUU?t=1585 Community Call 010 Recording https://youtu.be/6DRZEaKkpb4?t=3411 Community Call 011 Recording https://www.youtube.com/watch?v=ad_grFagA5k Community Call 012 Recording https://youtu.be/GCBv1VCN2tE?t=3339 Community Call 013 Recording https://www.youtube.com/watch?v=HQ9IKu3PVkA ETC Core Devs Call 22: Proposed Rejection of ECIP-1049 #460 Pull Request- Rejected Status ecip-1049: https://github.com/ethereumclassic/ECIPs/issues/465 It should be noted in this new discussion thread, this ECIP appears to be contentious (as documented in all the previous threads/recordings) and has a high-probability of a chain split between the GPU Miners on ETCHash and the FPGA & ASIC miners on Keccak-256. ![Etchash-vs-Keccak256](https://user-images.githubusercontent.com/58588775/154829756-4ef2ca53-d926-4514-949d-feee686c4c2a.png) How to join: Topic: ETC Core Devs Call 22: Proposed Rejection of ECIP-1049 Time: February 21, 2022 Time: 17:00 UTC Meeting Link: https://ethereumclassic.org/discord Channel: community-calls @ethereumclassic/all-hands
1.0
ETC Core Devs Call 22: Proposed Rejection of ECIP-1049 - Agenda: https://ethereumclassic.org/blog/2022-02-21-core-devs-call-22-ecip-1049-proposed-rejection Discuss fate of ECIP-1049 after three years. (ECIP-1000 clause). Focus: REJECT Keccak-256 Mining Algorithm Change due to a high-probability risk of Contentious Chain Split between GPU Miners on ETCHash and FPGA & ASIC Miners on Keccak-256. ECIP-1049 is in violation of Ethereum Classic founding documents and the ECIP process. At this point, the contentious proposal has negative externalities on the network and is a resource drain. Move to reject the proposal after three years of technical discussion. If the proposal is not rejected, begin plans for a chain split. Decision to be made: + Move to `Rejected` status, or; + Set a Block and Push Contentious Fork with a `Chain Split` ECIP-1000 Clause: "ECIPs should be changed from Draft or Last Call status, to Rejected, upon request by any person, if they have not made progress in three years. Such a ECIP may be changed to Draft status if the champion provides revisions that meaningfully address public criticism of the proposal, or to Last Call if it meets the criteria required as described in the previous paragraph." The reason for "Rejected" status under this clause is that the champion (Alex Tsankov) has not met this requirement during the three years: "the champion provides revisions that meaningfully address public criticism of the proposal". Rather, the champion (Alex Tsankov) has ignored much valid criticism and abandoned the proposal. https://ecips.ethereumclassic.org/ECIPs/ecip-1000 Follow Up from: https://github.com/ethereumclassic/ECIPs/issues/382 Formal Proposed Rejection: https://github.com/ethereumclassic/ECIPs/issues/394#issuecomment-1022909537 Documented Github Opposition: https://github.com/ethereumclassic/ECIPs/issues/394#issuecomment-828160552 If time permits: review ECIP-1094 and ECIP-1096 for activity. Newer proposals but appear to be abandoned by the authors as well. Should these be `Withdrawn`? Please review the issue thread to find the most up to date information. Related Discussions: ECIP-1049: Change the ETC Proof of Work Algorithm to Keccak256 #8 ECIP-1049: Change the ETC Proof of Work Algorithm to Keccak-256 #13 ETC Core Devs Call(s) 2020 Q3: Hardfork #333 ECIP-1095: Change ETC PoW to "vanilla" Sha-3 Discussion #342 ETC Core Devs Call 13 & 14 #362 ETC Core Devs Call 15 - ECIP-1049 Breakout Session Keccak-256 #382 SHA3 Precompile ethereum/EIPs#2951 Core Devs Call 15 Recording https://vimeo.com/464336957 Change the ETC Proof of Work Algorithm to Keccak-256 #394 Admin Clean Up on ecip-1049: https://github.com/ethereumclassic/ECIPs/pull/400 Core Devs Call 19 Recording https://www.youtube.com/watch?v=WySNxZbDEkQ Community Call 005 Recording https://youtu.be/HaDANZN-ZUU?t=1585 Community Call 010 Recording https://youtu.be/6DRZEaKkpb4?t=3411 Community Call 011 Recording https://www.youtube.com/watch?v=ad_grFagA5k Community Call 012 Recording https://youtu.be/GCBv1VCN2tE?t=3339 Community Call 013 Recording https://www.youtube.com/watch?v=HQ9IKu3PVkA ETC Core Devs Call 22: Proposed Rejection of ECIP-1049 #460 Pull Request- Rejected Status ecip-1049: https://github.com/ethereumclassic/ECIPs/issues/465 It should be noted in this new discussion thread, this ECIP appears to be contentious (as documented in all the previous threads/recordings) and has a high-probability of a chain split between the GPU Miners on ETCHash and the FPGA & ASIC miners on Keccak-256. ![Etchash-vs-Keccak256](https://user-images.githubusercontent.com/58588775/154829756-4ef2ca53-d926-4514-949d-feee686c4c2a.png) How to join: Topic: ETC Core Devs Call 22: Proposed Rejection of ECIP-1049 Time: February 21, 2022 Time: 17:00 UTC Meeting Link: https://ethereumclassic.org/discord Channel: community-calls @ethereumclassic/all-hands
process
etc core devs call proposed rejection of ecip agenda discuss fate of ecip after three years ecip clause focus reject keccak mining algorithm change due to a high probability risk of contentious chain split between gpu miners on etchash and fpga asic miners on keccak ecip is in violation of ethereum classic founding documents and the ecip process at this point the contentious proposal has negative externalities on the network and is a resource drain move to reject the proposal after three years of technical discussion if the proposal is not rejected begin plans for a chain split decision to be made move to rejected status or set a block and push contentious fork with a chain split ecip clause ecips should be changed from draft or last call status to rejected upon request by any person if they have not made progress in three years such a ecip may be changed to draft status if the champion provides revisions that meaningfully address public criticism of the proposal or to last call if it meets the criteria required as described in the previous paragraph the reason for rejected status under this clause is that the champion alex tsankov has not met this requirement during the three years the champion provides revisions that meaningfully address public criticism of the proposal rather the champion alex tsankov has ignored much valid criticism and abandoned the proposal follow up from formal proposed rejection documented github opposition if time permits review ecip and ecip for activity newer proposals but appear to be abandoned by the authors as well should these be withdrawn please review the issue thread to find the most up to date information related discussions ecip change the etc proof of work algorithm to ecip change the etc proof of work algorithm to keccak etc core devs call s hardfork ecip change etc pow to vanilla sha discussion etc core devs call etc core devs call ecip breakout session keccak precompile ethereum eips core devs call recording change the etc proof of work algorithm to keccak admin clean up on ecip core devs call recording community call recording community call recording community call recording community call recording community call recording etc core devs call proposed rejection of ecip pull request rejected status ecip it should be noted in this new discussion thread this ecip appears to be contentious as documented in all the previous threads recordings and has a high probability of a chain split between the gpu miners on etchash and the fpga asic miners on keccak how to join topic etc core devs call proposed rejection of ecip time february time utc meeting link channel community calls ethereumclassic all hands
1
3,384
13,111,993,266
IssuesEvent
2020-08-05 00:46:15
short-d/short
https://api.github.com/repos/short-d/short
closed
[Refactor] Use ptr.String to pass test case strings directly into test cases
maintainability
**What is frustrating you?** Test case strings should be inserted directly into test cases if possible, but [due to a restriction in the Golang spec](https://golang.org/ref/spec#Address_operators), string literals cannot have its address taken directly. **Your solution** In #938, `ptr.String` method was introduced to facilitate getting address out of string literal in a clean way. Update all test cases that use string addresses to use this helper method instead.
True
[Refactor] Use ptr.String to pass test case strings directly into test cases - **What is frustrating you?** Test case strings should be inserted directly into test cases if possible, but [due to a restriction in the Golang spec](https://golang.org/ref/spec#Address_operators), string literals cannot have its address taken directly. **Your solution** In #938, `ptr.String` method was introduced to facilitate getting address out of string literal in a clean way. Update all test cases that use string addresses to use this helper method instead.
non_process
use ptr string to pass test case strings directly into test cases what is frustrating you test case strings should be inserted directly into test cases if possible but string literals cannot have its address taken directly your solution in ptr string method was introduced to facilitate getting address out of string literal in a clean way update all test cases that use string addresses to use this helper method instead
0
181,313
6,658,328,814
IssuesEvent
2017-09-30 17:54:13
Pterodactyl/Panel
https://api.github.com/repos/Pterodactyl/Panel
closed
Deleting two services fails with an error
bug priority: low
Hello, New pterodactyl implementation, clean, without any modifications ... when I'm trying to delete "Minecraft" and "Source Engine" services I get An error occurred while attempting to update this service. This error has been logged. ``` [2017-07-24 13:23:42] production.ERROR: PDOException: SQLSTATE[23000]: Integrity constraint violation: 1451 Cannot delete or update a parent row: a foreign key constraint fails (`dockermanager`.`service_options`, CONSTRAINT `service_options_config_from_foreign` FOREIGN KEY (`config_from`) REFERENCES `service_options` (`id`)) in /usr/share/nginx/html/vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOStatement.php:91 Stack trace: #0 /usr/share/nginx/html/vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOStatement.php(91): PDOStatement->execute(NULL) #1 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(474): Doctrine\DBAL\Driver\PDOStatement->execute() #2 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(640): Illuminate\Database\Connection->Illuminate\Database\{closure}('delete from `se...', Array) #3 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(607): Illuminate\Database\Connection->runQueryCallback('delete from `se...', Array, Object(Closure)) #4 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(477): Illuminate\Database\Connection->run('delete from `se...', Array, Object(Closure)) #5 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(428): Illuminate\Database\Connection->affectingStatement('delete from `se...', Array) #6 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php(2230): Illuminate\Database\Connection->delete('delete from `se...', Array) #7 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Builder.php(821): Illuminate\Database\Query\Builder->delete() #8 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php(778): Illuminate\Database\Eloquent\Builder->delete() #9 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php(747): Illuminate\Database\Eloquent\Model->performDeleteOnModel() #10 /usr/share/nginx/html/app/Repositories/OptionRepository.php(94): Illuminate\Database\Eloquent\Model->delete() #11 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Concerns/ManagesTransactions.php(29): Pterodactyl\Repositories\OptionRepository->Pterodactyl\Repositories\{closure}(Object(Illuminate\Database\MySqlConnection)) #12 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/DatabaseManager.php(324): Illuminate\Database\Connection->transaction(Object(Closure)) #13 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Support/Facades/Facade.php(221): Illuminate\Database\DatabaseManager->__call('transaction', Array) #14 /usr/share/nginx/html/app/Repositories/OptionRepository.php(95): Illuminate\Support\Facades\Facade::__callStatic('transaction', Array) #15 /usr/share/nginx/html/app/Repositories/ServiceRepository.php(129): Pterodactyl\Repositories\OptionRepository->delete(5) #16 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Concerns/ManagesTransactions.php(29): Pterodactyl\Repositories\ServiceRepository->Pterodactyl\Repositories\{closure}(Object(Illuminate\Database\MySqlConnection)) #17 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/DatabaseManager.php(324): Illuminate\Database\Connection->transaction(Object(Closure)) #18 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Support/Facades/Facade.php(221): Illuminate\Database\DatabaseManager->__call('transaction', Array) #19 /usr/share/nginx/html/app/Repositories/ServiceRepository.php(133): Illuminate\Support\Facades\Facade::__callStatic('transaction', Array) #20 /usr/share/nginx/html/app/Http/Controllers/Admin/ServiceController.php(136): Pterodactyl\Repositories\ServiceRepository->delete('2') #21 [internal function]: Pterodactyl\Http\Controllers\Admin\ServiceController->edit(Object(Illuminate\Http\Request), '2') #22 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Controller.php(55): call_user_func_array(Array, Array) #23 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/ControllerDispatcher.php(44): Illuminate\Routing\Controller->callAction('edit', Array) #24 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Route.php(203): Illuminate\Routing\ControllerDispatcher->dispatch(Object(Illuminate\Routing\Route), Object(Pterodactyl\Http\Controllers\Admin\ServiceController), 'edit') #25 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Route.php(160): Illuminate\Routing\Route->runController() #26 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(574): Illuminate\Routing\Route->run() #27 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(30): Illuminate\Routing\Router->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #28 /usr/share/nginx/html/app/Http/Middleware/AdminAuthenticate.php(71): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #29 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Pterodactyl\Http\Middleware\AdminAuthenticate->handle(Object(Illuminate\Http\Request), Object(Closure)) #30 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #31 /usr/share/nginx/html/app/Http/Middleware/LanguageMiddleware.php(54): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #32 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Pterodactyl\Http\Middleware\LanguageMiddleware->handle(Object(Illuminate\Http\Request), Object(Closure)) #33 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #34 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Middleware/SubstituteBindings.php(41): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #35 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Routing\Middleware\SubstituteBindings->handle(Object(Illuminate\Http\Request), Object(Closure)) #36 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #37 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Auth/Middleware/Authenticate.php(43): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #38 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Auth\Middleware\Authenticate->handle(Object(Illuminate\Http\Request), Object(Closure)) #39 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #40 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/VerifyCsrfToken.php(65): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #41 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\VerifyCsrfToken->handle(Object(Illuminate\Http\Request), Object(Closure)) #42 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #43 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/View/Middleware/ShareErrorsFromSession.php(49): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #44 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\View\Middleware\ShareErrorsFromSession->handle(Object(Illuminate\Http\Request), Object(Closure)) #45 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #46 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Session/Middleware/StartSession.php(64): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #47 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Session\Middleware\StartSession->handle(Object(Illuminate\Http\Request), Object(Closure)) #48 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #49 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/AddQueuedCookiesToResponse.php(37): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #50 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\AddQueuedCookiesToResponse->handle(Object(Illuminate\Http\Request), Object(Closure)) #51 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #52 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/EncryptCookies.php(59): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #53 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\EncryptCookies->handle(Object(Illuminate\Http\Request), Object(Closure)) #54 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #55 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(102): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #56 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(576): Illuminate\Pipeline\Pipeline->then(Object(Closure)) #57 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(535): Illuminate\Routing\Router->runRouteWithinStack(Object(Illuminate\Routing\Route), Object(Illuminate\Http\Request)) #58 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(513): Illuminate\Routing\Router->dispatchToRoute(Object(Illuminate\Http\Request)) #59 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(174): Illuminate\Routing\Router->dispatch(Object(Illuminate\Http\Request)) #60 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(30): Illuminate\Foundation\Http\Kernel->Illuminate\Foundation\Http\{closure}(Object(Illuminate\Http\Request)) #61 /usr/share/nginx/html/vendor/fideloper/proxy/src/TrustProxies.php(55): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #62 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Fideloper\Proxy\TrustProxies->handle(Object(Illuminate\Http\Request), Object(Closure)) #63 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #64 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/TransformsRequest.php(30): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #65 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\TransformsRequest->handle(Object(Illuminate\Http\Request), Object(Closure)) #66 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #67 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/AddQueuedCookiesToResponse.php(37): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #68 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\AddQueuedCookiesToResponse->handle(Object(Illuminate\Http\Request), Object(Closure)) #69 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #70 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/EncryptCookies.php(59): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #71 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\EncryptCookies->handle(Object(Illuminate\Http\Request), Object(Closure)) #72 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #73 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/CheckForMaintenanceMode.php(46): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #74 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\CheckForMaintenanceMode->handle(Object(Illuminate\Http\Request), Object(Closure)) #75 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #76 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(102): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #77 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(149): Illuminate\Pipeline\Pipeline->then(Object(Closure)) #78 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(116): Illuminate\Foundation\Http\Kernel->sendRequestThroughRouter(Object(Illuminate\Http\Request)) #79 /usr/share/nginx/html/public/index.php(52): Illuminate\Foundation\Http\Kernel->handle(Object(Illuminate\Http\Request)) #80 {main} Next Doctrine\DBAL\Driver\PDOException: SQLSTATE[23000]: Integrity constraint violation: 1451 Cannot delete or update a parent row: a foreign key constraint fails (`dockermanager`.`service_options`, CONSTRAINT `service_options_config_from_foreign` FOREIGN KEY (`config_from`) REFERENCES `service_options` (`id`)) in /usr/share/nginx/html/vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOStatement.php:93 Stack trace: #0 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(474): Doctrine\DBAL\Driver\PDOStatement->execute() #1 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(640): Illuminate\Database\Connection->Illuminate\Database\{closure}('delete from `se...', Array) #2 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(607): Illuminate\Database\Connection->runQueryCallback('delete from `se...', Array, Object(Closure)) #3 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(477): Illuminate\Database\Connection->run('delete from `se...', Array, Object(Closure)) #4 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(428): Illuminate\Database\Connection->affectingStatement('delete from `se...', Array) #5 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php(2230): Illuminate\Database\Connection->delete('delete from `se...', Array) #6 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Builder.php(821): Illuminate\Database\Query\Builder->delete() #7 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php(778): Illuminate\Database\Eloquent\Builder->delete() #8 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php(747): Illuminate\Database\Eloquent\Model->performDeleteOnModel() #9 /usr/share/nginx/html/app/Repositories/OptionRepository.php(94): Illuminate\Database\Eloquent\Model->delete() #10 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Concerns/ManagesTransactions.php(29): Pterodactyl\Repositories\OptionRepository->Pterodactyl\Repositories\{closure}(Object(Illuminate\Database\MySqlConnection)) #11 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/DatabaseManager.php(324): Illuminate\Database\Connection->transaction(Object(Closure)) #12 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Support/Facades/Facade.php(221): Illuminate\Database\DatabaseManager->__call('transaction', Array) #13 /usr/share/nginx/html/app/Repositories/OptionRepository.php(95): Illuminate\Support\Facades\Facade::__callStatic('transaction', Array) #14 /usr/share/nginx/html/app/Repositories/ServiceRepository.php(129): Pterodactyl\Repositories\OptionRepository->delete(5) #15 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Concerns/ManagesTransactions.php(29): Pterodactyl\Repositories\ServiceRepository->Pterodactyl\Repositories\{closure}(Object(Illuminate\Database\MySqlConnection)) #16 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/DatabaseManager.php(324): Illuminate\Database\Connection->transaction(Object(Closure)) #17 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Support/Facades/Facade.php(221): Illuminate\Database\DatabaseManager->__call('transaction', Array) #18 /usr/share/nginx/html/app/Repositories/ServiceRepository.php(133): Illuminate\Support\Facades\Facade::__callStatic('transaction', Array) #19 /usr/share/nginx/html/app/Http/Controllers/Admin/ServiceController.php(136): Pterodactyl\Repositories\ServiceRepository->delete('2') #20 [internal function]: Pterodactyl\Http\Controllers\Admin\ServiceController->edit(Object(Illuminate\Http\Request), '2') #21 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Controller.php(55): call_user_func_array(Array, Array) #22 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/ControllerDispatcher.php(44): Illuminate\Routing\Controller->callAction('edit', Array) #23 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Route.php(203): Illuminate\Routing\ControllerDispatcher->dispatch(Object(Illuminate\Routing\Route), Object(Pterodactyl\Http\Controllers\Admin\ServiceController), 'edit') #24 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Route.php(160): Illuminate\Routing\Route->runController() #25 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(574): Illuminate\Routing\Route->run() #26 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(30): Illuminate\Routing\Router->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #27 /usr/share/nginx/html/app/Http/Middleware/AdminAuthenticate.php(71): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #28 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Pterodactyl\Http\Middleware\AdminAuthenticate->handle(Object(Illuminate\Http\Request), Object(Closure)) #29 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #30 /usr/share/nginx/html/app/Http/Middleware/LanguageMiddleware.php(54): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #31 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Pterodactyl\Http\Middleware\LanguageMiddleware->handle(Object(Illuminate\Http\Request), Object(Closure)) #32 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #33 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Middleware/SubstituteBindings.php(41): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #34 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Routing\Middleware\SubstituteBindings->handle(Object(Illuminate\Http\Request), Object(Closure)) #35 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #36 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Auth/Middleware/Authenticate.php(43): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #37 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Auth\Middleware\Authenticate->handle(Object(Illuminate\Http\Request), Object(Closure)) #38 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #39 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/VerifyCsrfToken.php(65): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #40 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\VerifyCsrfToken->handle(Object(Illuminate\Http\Request), Object(Closure)) #41 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #42 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/View/Middleware/ShareErrorsFromSession.php(49): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #43 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\View\Middleware\ShareErrorsFromSession->handle(Object(Illuminate\Http\Request), Object(Closure)) #44 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #45 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Session/Middleware/StartSession.php(64): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #46 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Session\Middleware\StartSession->handle(Object(Illuminate\Http\Request), Object(Closure)) #47 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #48 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/AddQueuedCookiesToResponse.php(37): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #49 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\AddQueuedCookiesToResponse->handle(Object(Illuminate\Http\Request), Object(Closure)) #50 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #51 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/EncryptCookies.php(59): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #52 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\EncryptCookies->handle(Object(Illuminate\Http\Request), Object(Closure)) #53 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #54 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(102): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #55 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(576): Illuminate\Pipeline\Pipeline->then(Object(Closure)) #56 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(535): Illuminate\Routing\Router->runRouteWithinStack(Object(Illuminate\Routing\Route), Object(Illuminate\Http\Request)) #57 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(513): Illuminate\Routing\Router->dispatchToRoute(Object(Illuminate\Http\Request)) #58 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(174): Illuminate\Routing\Router->dispatch(Object(Illuminate\Http\Request)) #59 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(30): Illuminate\Foundation\Http\Kernel->Illuminate\Foundation\Http\{closure}(Object(Illuminate\Http\Request)) #60 /usr/share/nginx/html/vendor/fideloper/proxy/src/TrustProxies.php(55): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #61 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Fideloper\Proxy\TrustProxies->handle(Object(Illuminate\Http\Request), Object(Closure)) #62 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #63 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/TransformsRequest.php(30): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #64 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\TransformsRequest->handle(Object(Illuminate\Http\Request), Object(Closure)) #65 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #66 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/AddQueuedCookiesToResponse.php(37): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #67 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\AddQueuedCookiesToResponse->handle(Object(Illuminate\Http\Request), Object(Closure)) #68 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #69 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/EncryptCookies.php(59): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #70 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\EncryptCookies->handle(Object(Illuminate\Http\Request), Object(Closure)) #71 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #72 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/CheckForMaintenanceMode.php(46): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #73 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\CheckForMaintenanceMode->handle(Object(Illuminate\Http\Request), Object(Closure)) #74 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #75 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(102): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #76 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(149): Illuminate\Pipeline\Pipeline->then(Object(Closure)) #77 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(116): Illuminate\Foundation\Http\Kernel->sendRequestThroughRouter(Object(Illuminate\Http\Request)) #78 /usr/share/nginx/html/public/index.php(52): Illuminate\Foundation\Http\Kernel->handle(Object(Illuminate\Http\Request)) #79 {main} Next Illuminate\Database\QueryException: SQLSTATE[23000]: Integrity constraint violation: 1451 Cannot delete or update a parent row: a foreign key constraint fails (`dockermanager`.`service_options`, CONSTRAINT `service_options_config_from_foreign` FOREIGN KEY (`config_from`) REFERENCES `service_options` (`id`)) (SQL: delete from `service_options` where `id` = 5) in /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php:647 Stack trace: #0 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(607): Illuminate\Database\Connection->runQueryCallback('delete from `se...', Array, Object(Closure)) #1 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(477): Illuminate\Database\Connection->run('delete from `se...', Array, Object(Closure)) #2 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(428): Illuminate\Database\Connection->affectingStatement('delete from `se...', Array) #3 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php(2230): Illuminate\Database\Connection->delete('delete from `se...', Array) #4 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Builder.php(821): Illuminate\Database\Query\Builder->delete() #5 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php(778): Illuminate\Database\Eloquent\Builder->delete() #6 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php(747): Illuminate\Database\Eloquent\Model->performDeleteOnModel() #7 /usr/share/nginx/html/app/Repositories/OptionRepository.php(94): Illuminate\Database\Eloquent\Model->delete() #8 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Concerns/ManagesTransactions.php(29): Pterodactyl\Repositories\OptionRepository->Pterodactyl\Repositories\{closure}(Object(Illuminate\Database\MySqlConnection)) #9 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/DatabaseManager.php(324): Illuminate\Database\Connection->transaction(Object(Closure)) #10 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Support/Facades/Facade.php(221): Illuminate\Database\DatabaseManager->__call('transaction', Array) #11 /usr/share/nginx/html/app/Repositories/OptionRepository.php(95): Illuminate\Support\Facades\Facade::__callStatic('transaction', Array) #12 /usr/share/nginx/html/app/Repositories/ServiceRepository.php(129): Pterodactyl\Repositories\OptionRepository->delete(5) #13 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Concerns/ManagesTransactions.php(29): Pterodactyl\Repositories\ServiceRepository->Pterodactyl\Repositories\{closure}(Object(Illuminate\Database\MySqlConnection)) #14 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/DatabaseManager.php(324): Illuminate\Database\Connection->transaction(Object(Closure)) #15 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Support/Facades/Facade.php(221): Illuminate\Database\DatabaseManager->__call('transaction', Array) #16 /usr/share/nginx/html/app/Repositories/ServiceRepository.php(133): Illuminate\Support\Facades\Facade::__callStatic('transaction', Array) #17 /usr/share/nginx/html/app/Http/Controllers/Admin/ServiceController.php(136): Pterodactyl\Repositories\ServiceRepository->delete('2') #18 [internal function]: Pterodactyl\Http\Controllers\Admin\ServiceController->edit(Object(Illuminate\Http\Request), '2') #19 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Controller.php(55): call_user_func_array(Array, Array) #20 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/ControllerDispatcher.php(44): Illuminate\Routing\Controller->callAction('edit', Array) #21 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Route.php(203): Illuminate\Routing\ControllerDispatcher->dispatch(Object(Illuminate\Routing\Route), Object(Pterodactyl\Http\Controllers\Admin\ServiceController), 'edit') #22 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Route.php(160): Illuminate\Routing\Route->runController() #23 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(574): Illuminate\Routing\Route->run() #24 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(30): Illuminate\Routing\Router->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #25 /usr/share/nginx/html/app/Http/Middleware/AdminAuthenticate.php(71): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #26 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Pterodactyl\Http\Middleware\AdminAuthenticate->handle(Object(Illuminate\Http\Request), Object(Closure)) #27 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #28 /usr/share/nginx/html/app/Http/Middleware/LanguageMiddleware.php(54): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #29 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Pterodactyl\Http\Middleware\LanguageMiddleware->handle(Object(Illuminate\Http\Request), Object(Closure)) #30 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #31 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Middleware/SubstituteBindings.php(41): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #32 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Routing\Middleware\SubstituteBindings->handle(Object(Illuminate\Http\Request), Object(Closure)) #33 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #34 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Auth/Middleware/Authenticate.php(43): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #35 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Auth\Middleware\Authenticate->handle(Object(Illuminate\Http\Request), Object(Closure)) #36 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #37 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/VerifyCsrfToken.php(65): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #38 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\VerifyCsrfToken->handle(Object(Illuminate\Http\Request), Object(Closure)) #39 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #40 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/View/Middleware/ShareErrorsFromSession.php(49): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #41 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\View\Middleware\ShareErrorsFromSession->handle(Object(Illuminate\Http\Request), Object(Closure)) #42 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #43 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Session/Middleware/StartSession.php(64): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #44 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Session\Middleware\StartSession->handle(Object(Illuminate\Http\Request), Object(Closure)) #45 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #46 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/AddQueuedCookiesToResponse.php(37): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #47 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\AddQueuedCookiesToResponse->handle(Object(Illuminate\Http\Request), Object(Closure)) #48 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #49 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/EncryptCookies.php(59): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #50 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\EncryptCookies->handle(Object(Illuminate\Http\Request), Object(Closure)) #51 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #52 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(102): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #53 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(576): Illuminate\Pipeline\Pipeline->then(Object(Closure)) #54 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(535): Illuminate\Routing\Router->runRouteWithinStack(Object(Illuminate\Routing\Route), Object(Illuminate\Http\Request)) #55 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(513): Illuminate\Routing\Router->dispatchToRoute(Object(Illuminate\Http\Request)) #56 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(174): Illuminate\Routing\Router->dispatch(Object(Illuminate\Http\Request)) #57 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(30): Illuminate\Foundation\Http\Kernel->Illuminate\Foundation\Http\{closure}(Object(Illuminate\Http\Request)) #58 /usr/share/nginx/html/vendor/fideloper/proxy/src/TrustProxies.php(55): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #59 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Fideloper\Proxy\TrustProxies->handle(Object(Illuminate\Http\Request), Object(Closure)) #60 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #61 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/TransformsRequest.php(30): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #62 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\TransformsRequest->handle(Object(Illuminate\Http\Request), Object(Closure)) #63 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #64 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/AddQueuedCookiesToResponse.php(37): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #65 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\AddQueuedCookiesToResponse->handle(Object(Illuminate\Http\Request), Object(Closure)) #66 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #67 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/EncryptCookies.php(59): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #68 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\EncryptCookies->handle(Object(Illuminate\Http\Request), Object(Closure)) #69 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #70 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/CheckForMaintenanceMode.php(46): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #71 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\CheckForMaintenanceMode->handle(Object(Illuminate\Http\Request), Object(Closure)) #72 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #73 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(102): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #74 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(149): Illuminate\Pipeline\Pipeline->then(Object(Closure)) #75 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(116): Illuminate\Foundation\Http\Kernel->sendRequestThroughRouter(Object(Illuminate\Http\Request)) #76 /usr/share/nginx/html/public/index.php(52): Illuminate\Foundation\Http\Kernel->handle(Object(Illuminate\Http\Request)) #77 {main} ```
1.0
Deleting two services fails with an error - Hello, New pterodactyl implementation, clean, without any modifications ... when I'm trying to delete "Minecraft" and "Source Engine" services I get An error occurred while attempting to update this service. This error has been logged. ``` [2017-07-24 13:23:42] production.ERROR: PDOException: SQLSTATE[23000]: Integrity constraint violation: 1451 Cannot delete or update a parent row: a foreign key constraint fails (`dockermanager`.`service_options`, CONSTRAINT `service_options_config_from_foreign` FOREIGN KEY (`config_from`) REFERENCES `service_options` (`id`)) in /usr/share/nginx/html/vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOStatement.php:91 Stack trace: #0 /usr/share/nginx/html/vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOStatement.php(91): PDOStatement->execute(NULL) #1 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(474): Doctrine\DBAL\Driver\PDOStatement->execute() #2 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(640): Illuminate\Database\Connection->Illuminate\Database\{closure}('delete from `se...', Array) #3 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(607): Illuminate\Database\Connection->runQueryCallback('delete from `se...', Array, Object(Closure)) #4 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(477): Illuminate\Database\Connection->run('delete from `se...', Array, Object(Closure)) #5 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(428): Illuminate\Database\Connection->affectingStatement('delete from `se...', Array) #6 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php(2230): Illuminate\Database\Connection->delete('delete from `se...', Array) #7 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Builder.php(821): Illuminate\Database\Query\Builder->delete() #8 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php(778): Illuminate\Database\Eloquent\Builder->delete() #9 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php(747): Illuminate\Database\Eloquent\Model->performDeleteOnModel() #10 /usr/share/nginx/html/app/Repositories/OptionRepository.php(94): Illuminate\Database\Eloquent\Model->delete() #11 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Concerns/ManagesTransactions.php(29): Pterodactyl\Repositories\OptionRepository->Pterodactyl\Repositories\{closure}(Object(Illuminate\Database\MySqlConnection)) #12 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/DatabaseManager.php(324): Illuminate\Database\Connection->transaction(Object(Closure)) #13 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Support/Facades/Facade.php(221): Illuminate\Database\DatabaseManager->__call('transaction', Array) #14 /usr/share/nginx/html/app/Repositories/OptionRepository.php(95): Illuminate\Support\Facades\Facade::__callStatic('transaction', Array) #15 /usr/share/nginx/html/app/Repositories/ServiceRepository.php(129): Pterodactyl\Repositories\OptionRepository->delete(5) #16 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Concerns/ManagesTransactions.php(29): Pterodactyl\Repositories\ServiceRepository->Pterodactyl\Repositories\{closure}(Object(Illuminate\Database\MySqlConnection)) #17 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/DatabaseManager.php(324): Illuminate\Database\Connection->transaction(Object(Closure)) #18 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Support/Facades/Facade.php(221): Illuminate\Database\DatabaseManager->__call('transaction', Array) #19 /usr/share/nginx/html/app/Repositories/ServiceRepository.php(133): Illuminate\Support\Facades\Facade::__callStatic('transaction', Array) #20 /usr/share/nginx/html/app/Http/Controllers/Admin/ServiceController.php(136): Pterodactyl\Repositories\ServiceRepository->delete('2') #21 [internal function]: Pterodactyl\Http\Controllers\Admin\ServiceController->edit(Object(Illuminate\Http\Request), '2') #22 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Controller.php(55): call_user_func_array(Array, Array) #23 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/ControllerDispatcher.php(44): Illuminate\Routing\Controller->callAction('edit', Array) #24 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Route.php(203): Illuminate\Routing\ControllerDispatcher->dispatch(Object(Illuminate\Routing\Route), Object(Pterodactyl\Http\Controllers\Admin\ServiceController), 'edit') #25 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Route.php(160): Illuminate\Routing\Route->runController() #26 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(574): Illuminate\Routing\Route->run() #27 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(30): Illuminate\Routing\Router->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #28 /usr/share/nginx/html/app/Http/Middleware/AdminAuthenticate.php(71): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #29 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Pterodactyl\Http\Middleware\AdminAuthenticate->handle(Object(Illuminate\Http\Request), Object(Closure)) #30 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #31 /usr/share/nginx/html/app/Http/Middleware/LanguageMiddleware.php(54): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #32 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Pterodactyl\Http\Middleware\LanguageMiddleware->handle(Object(Illuminate\Http\Request), Object(Closure)) #33 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #34 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Middleware/SubstituteBindings.php(41): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #35 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Routing\Middleware\SubstituteBindings->handle(Object(Illuminate\Http\Request), Object(Closure)) #36 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #37 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Auth/Middleware/Authenticate.php(43): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #38 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Auth\Middleware\Authenticate->handle(Object(Illuminate\Http\Request), Object(Closure)) #39 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #40 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/VerifyCsrfToken.php(65): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #41 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\VerifyCsrfToken->handle(Object(Illuminate\Http\Request), Object(Closure)) #42 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #43 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/View/Middleware/ShareErrorsFromSession.php(49): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #44 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\View\Middleware\ShareErrorsFromSession->handle(Object(Illuminate\Http\Request), Object(Closure)) #45 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #46 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Session/Middleware/StartSession.php(64): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #47 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Session\Middleware\StartSession->handle(Object(Illuminate\Http\Request), Object(Closure)) #48 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #49 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/AddQueuedCookiesToResponse.php(37): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #50 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\AddQueuedCookiesToResponse->handle(Object(Illuminate\Http\Request), Object(Closure)) #51 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #52 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/EncryptCookies.php(59): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #53 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\EncryptCookies->handle(Object(Illuminate\Http\Request), Object(Closure)) #54 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #55 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(102): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #56 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(576): Illuminate\Pipeline\Pipeline->then(Object(Closure)) #57 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(535): Illuminate\Routing\Router->runRouteWithinStack(Object(Illuminate\Routing\Route), Object(Illuminate\Http\Request)) #58 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(513): Illuminate\Routing\Router->dispatchToRoute(Object(Illuminate\Http\Request)) #59 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(174): Illuminate\Routing\Router->dispatch(Object(Illuminate\Http\Request)) #60 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(30): Illuminate\Foundation\Http\Kernel->Illuminate\Foundation\Http\{closure}(Object(Illuminate\Http\Request)) #61 /usr/share/nginx/html/vendor/fideloper/proxy/src/TrustProxies.php(55): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #62 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Fideloper\Proxy\TrustProxies->handle(Object(Illuminate\Http\Request), Object(Closure)) #63 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #64 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/TransformsRequest.php(30): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #65 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\TransformsRequest->handle(Object(Illuminate\Http\Request), Object(Closure)) #66 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #67 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/AddQueuedCookiesToResponse.php(37): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #68 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\AddQueuedCookiesToResponse->handle(Object(Illuminate\Http\Request), Object(Closure)) #69 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #70 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/EncryptCookies.php(59): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #71 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\EncryptCookies->handle(Object(Illuminate\Http\Request), Object(Closure)) #72 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #73 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/CheckForMaintenanceMode.php(46): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #74 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\CheckForMaintenanceMode->handle(Object(Illuminate\Http\Request), Object(Closure)) #75 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #76 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(102): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #77 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(149): Illuminate\Pipeline\Pipeline->then(Object(Closure)) #78 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(116): Illuminate\Foundation\Http\Kernel->sendRequestThroughRouter(Object(Illuminate\Http\Request)) #79 /usr/share/nginx/html/public/index.php(52): Illuminate\Foundation\Http\Kernel->handle(Object(Illuminate\Http\Request)) #80 {main} Next Doctrine\DBAL\Driver\PDOException: SQLSTATE[23000]: Integrity constraint violation: 1451 Cannot delete or update a parent row: a foreign key constraint fails (`dockermanager`.`service_options`, CONSTRAINT `service_options_config_from_foreign` FOREIGN KEY (`config_from`) REFERENCES `service_options` (`id`)) in /usr/share/nginx/html/vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOStatement.php:93 Stack trace: #0 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(474): Doctrine\DBAL\Driver\PDOStatement->execute() #1 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(640): Illuminate\Database\Connection->Illuminate\Database\{closure}('delete from `se...', Array) #2 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(607): Illuminate\Database\Connection->runQueryCallback('delete from `se...', Array, Object(Closure)) #3 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(477): Illuminate\Database\Connection->run('delete from `se...', Array, Object(Closure)) #4 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(428): Illuminate\Database\Connection->affectingStatement('delete from `se...', Array) #5 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php(2230): Illuminate\Database\Connection->delete('delete from `se...', Array) #6 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Builder.php(821): Illuminate\Database\Query\Builder->delete() #7 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php(778): Illuminate\Database\Eloquent\Builder->delete() #8 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php(747): Illuminate\Database\Eloquent\Model->performDeleteOnModel() #9 /usr/share/nginx/html/app/Repositories/OptionRepository.php(94): Illuminate\Database\Eloquent\Model->delete() #10 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Concerns/ManagesTransactions.php(29): Pterodactyl\Repositories\OptionRepository->Pterodactyl\Repositories\{closure}(Object(Illuminate\Database\MySqlConnection)) #11 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/DatabaseManager.php(324): Illuminate\Database\Connection->transaction(Object(Closure)) #12 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Support/Facades/Facade.php(221): Illuminate\Database\DatabaseManager->__call('transaction', Array) #13 /usr/share/nginx/html/app/Repositories/OptionRepository.php(95): Illuminate\Support\Facades\Facade::__callStatic('transaction', Array) #14 /usr/share/nginx/html/app/Repositories/ServiceRepository.php(129): Pterodactyl\Repositories\OptionRepository->delete(5) #15 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Concerns/ManagesTransactions.php(29): Pterodactyl\Repositories\ServiceRepository->Pterodactyl\Repositories\{closure}(Object(Illuminate\Database\MySqlConnection)) #16 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/DatabaseManager.php(324): Illuminate\Database\Connection->transaction(Object(Closure)) #17 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Support/Facades/Facade.php(221): Illuminate\Database\DatabaseManager->__call('transaction', Array) #18 /usr/share/nginx/html/app/Repositories/ServiceRepository.php(133): Illuminate\Support\Facades\Facade::__callStatic('transaction', Array) #19 /usr/share/nginx/html/app/Http/Controllers/Admin/ServiceController.php(136): Pterodactyl\Repositories\ServiceRepository->delete('2') #20 [internal function]: Pterodactyl\Http\Controllers\Admin\ServiceController->edit(Object(Illuminate\Http\Request), '2') #21 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Controller.php(55): call_user_func_array(Array, Array) #22 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/ControllerDispatcher.php(44): Illuminate\Routing\Controller->callAction('edit', Array) #23 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Route.php(203): Illuminate\Routing\ControllerDispatcher->dispatch(Object(Illuminate\Routing\Route), Object(Pterodactyl\Http\Controllers\Admin\ServiceController), 'edit') #24 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Route.php(160): Illuminate\Routing\Route->runController() #25 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(574): Illuminate\Routing\Route->run() #26 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(30): Illuminate\Routing\Router->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #27 /usr/share/nginx/html/app/Http/Middleware/AdminAuthenticate.php(71): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #28 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Pterodactyl\Http\Middleware\AdminAuthenticate->handle(Object(Illuminate\Http\Request), Object(Closure)) #29 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #30 /usr/share/nginx/html/app/Http/Middleware/LanguageMiddleware.php(54): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #31 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Pterodactyl\Http\Middleware\LanguageMiddleware->handle(Object(Illuminate\Http\Request), Object(Closure)) #32 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #33 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Middleware/SubstituteBindings.php(41): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #34 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Routing\Middleware\SubstituteBindings->handle(Object(Illuminate\Http\Request), Object(Closure)) #35 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #36 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Auth/Middleware/Authenticate.php(43): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #37 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Auth\Middleware\Authenticate->handle(Object(Illuminate\Http\Request), Object(Closure)) #38 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #39 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/VerifyCsrfToken.php(65): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #40 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\VerifyCsrfToken->handle(Object(Illuminate\Http\Request), Object(Closure)) #41 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #42 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/View/Middleware/ShareErrorsFromSession.php(49): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #43 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\View\Middleware\ShareErrorsFromSession->handle(Object(Illuminate\Http\Request), Object(Closure)) #44 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #45 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Session/Middleware/StartSession.php(64): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #46 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Session\Middleware\StartSession->handle(Object(Illuminate\Http\Request), Object(Closure)) #47 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #48 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/AddQueuedCookiesToResponse.php(37): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #49 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\AddQueuedCookiesToResponse->handle(Object(Illuminate\Http\Request), Object(Closure)) #50 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #51 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/EncryptCookies.php(59): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #52 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\EncryptCookies->handle(Object(Illuminate\Http\Request), Object(Closure)) #53 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #54 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(102): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #55 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(576): Illuminate\Pipeline\Pipeline->then(Object(Closure)) #56 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(535): Illuminate\Routing\Router->runRouteWithinStack(Object(Illuminate\Routing\Route), Object(Illuminate\Http\Request)) #57 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(513): Illuminate\Routing\Router->dispatchToRoute(Object(Illuminate\Http\Request)) #58 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(174): Illuminate\Routing\Router->dispatch(Object(Illuminate\Http\Request)) #59 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(30): Illuminate\Foundation\Http\Kernel->Illuminate\Foundation\Http\{closure}(Object(Illuminate\Http\Request)) #60 /usr/share/nginx/html/vendor/fideloper/proxy/src/TrustProxies.php(55): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #61 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Fideloper\Proxy\TrustProxies->handle(Object(Illuminate\Http\Request), Object(Closure)) #62 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #63 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/TransformsRequest.php(30): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #64 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\TransformsRequest->handle(Object(Illuminate\Http\Request), Object(Closure)) #65 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #66 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/AddQueuedCookiesToResponse.php(37): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #67 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\AddQueuedCookiesToResponse->handle(Object(Illuminate\Http\Request), Object(Closure)) #68 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #69 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/EncryptCookies.php(59): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #70 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\EncryptCookies->handle(Object(Illuminate\Http\Request), Object(Closure)) #71 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #72 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/CheckForMaintenanceMode.php(46): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #73 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\CheckForMaintenanceMode->handle(Object(Illuminate\Http\Request), Object(Closure)) #74 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #75 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(102): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #76 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(149): Illuminate\Pipeline\Pipeline->then(Object(Closure)) #77 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(116): Illuminate\Foundation\Http\Kernel->sendRequestThroughRouter(Object(Illuminate\Http\Request)) #78 /usr/share/nginx/html/public/index.php(52): Illuminate\Foundation\Http\Kernel->handle(Object(Illuminate\Http\Request)) #79 {main} Next Illuminate\Database\QueryException: SQLSTATE[23000]: Integrity constraint violation: 1451 Cannot delete or update a parent row: a foreign key constraint fails (`dockermanager`.`service_options`, CONSTRAINT `service_options_config_from_foreign` FOREIGN KEY (`config_from`) REFERENCES `service_options` (`id`)) (SQL: delete from `service_options` where `id` = 5) in /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php:647 Stack trace: #0 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(607): Illuminate\Database\Connection->runQueryCallback('delete from `se...', Array, Object(Closure)) #1 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(477): Illuminate\Database\Connection->run('delete from `se...', Array, Object(Closure)) #2 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Connection.php(428): Illuminate\Database\Connection->affectingStatement('delete from `se...', Array) #3 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Query/Builder.php(2230): Illuminate\Database\Connection->delete('delete from `se...', Array) #4 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Builder.php(821): Illuminate\Database\Query\Builder->delete() #5 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php(778): Illuminate\Database\Eloquent\Builder->delete() #6 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Eloquent/Model.php(747): Illuminate\Database\Eloquent\Model->performDeleteOnModel() #7 /usr/share/nginx/html/app/Repositories/OptionRepository.php(94): Illuminate\Database\Eloquent\Model->delete() #8 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Concerns/ManagesTransactions.php(29): Pterodactyl\Repositories\OptionRepository->Pterodactyl\Repositories\{closure}(Object(Illuminate\Database\MySqlConnection)) #9 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/DatabaseManager.php(324): Illuminate\Database\Connection->transaction(Object(Closure)) #10 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Support/Facades/Facade.php(221): Illuminate\Database\DatabaseManager->__call('transaction', Array) #11 /usr/share/nginx/html/app/Repositories/OptionRepository.php(95): Illuminate\Support\Facades\Facade::__callStatic('transaction', Array) #12 /usr/share/nginx/html/app/Repositories/ServiceRepository.php(129): Pterodactyl\Repositories\OptionRepository->delete(5) #13 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/Concerns/ManagesTransactions.php(29): Pterodactyl\Repositories\ServiceRepository->Pterodactyl\Repositories\{closure}(Object(Illuminate\Database\MySqlConnection)) #14 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Database/DatabaseManager.php(324): Illuminate\Database\Connection->transaction(Object(Closure)) #15 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Support/Facades/Facade.php(221): Illuminate\Database\DatabaseManager->__call('transaction', Array) #16 /usr/share/nginx/html/app/Repositories/ServiceRepository.php(133): Illuminate\Support\Facades\Facade::__callStatic('transaction', Array) #17 /usr/share/nginx/html/app/Http/Controllers/Admin/ServiceController.php(136): Pterodactyl\Repositories\ServiceRepository->delete('2') #18 [internal function]: Pterodactyl\Http\Controllers\Admin\ServiceController->edit(Object(Illuminate\Http\Request), '2') #19 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Controller.php(55): call_user_func_array(Array, Array) #20 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/ControllerDispatcher.php(44): Illuminate\Routing\Controller->callAction('edit', Array) #21 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Route.php(203): Illuminate\Routing\ControllerDispatcher->dispatch(Object(Illuminate\Routing\Route), Object(Pterodactyl\Http\Controllers\Admin\ServiceController), 'edit') #22 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Route.php(160): Illuminate\Routing\Route->runController() #23 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(574): Illuminate\Routing\Route->run() #24 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(30): Illuminate\Routing\Router->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #25 /usr/share/nginx/html/app/Http/Middleware/AdminAuthenticate.php(71): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #26 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Pterodactyl\Http\Middleware\AdminAuthenticate->handle(Object(Illuminate\Http\Request), Object(Closure)) #27 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #28 /usr/share/nginx/html/app/Http/Middleware/LanguageMiddleware.php(54): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #29 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Pterodactyl\Http\Middleware\LanguageMiddleware->handle(Object(Illuminate\Http\Request), Object(Closure)) #30 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #31 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Middleware/SubstituteBindings.php(41): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #32 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Routing\Middleware\SubstituteBindings->handle(Object(Illuminate\Http\Request), Object(Closure)) #33 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #34 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Auth/Middleware/Authenticate.php(43): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #35 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Auth\Middleware\Authenticate->handle(Object(Illuminate\Http\Request), Object(Closure)) #36 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #37 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/VerifyCsrfToken.php(65): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #38 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\VerifyCsrfToken->handle(Object(Illuminate\Http\Request), Object(Closure)) #39 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #40 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/View/Middleware/ShareErrorsFromSession.php(49): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #41 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\View\Middleware\ShareErrorsFromSession->handle(Object(Illuminate\Http\Request), Object(Closure)) #42 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #43 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Session/Middleware/StartSession.php(64): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #44 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Session\Middleware\StartSession->handle(Object(Illuminate\Http\Request), Object(Closure)) #45 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #46 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/AddQueuedCookiesToResponse.php(37): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #47 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\AddQueuedCookiesToResponse->handle(Object(Illuminate\Http\Request), Object(Closure)) #48 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #49 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/EncryptCookies.php(59): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #50 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\EncryptCookies->handle(Object(Illuminate\Http\Request), Object(Closure)) #51 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #52 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(102): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #53 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(576): Illuminate\Pipeline\Pipeline->then(Object(Closure)) #54 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(535): Illuminate\Routing\Router->runRouteWithinStack(Object(Illuminate\Routing\Route), Object(Illuminate\Http\Request)) #55 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Router.php(513): Illuminate\Routing\Router->dispatchToRoute(Object(Illuminate\Http\Request)) #56 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(174): Illuminate\Routing\Router->dispatch(Object(Illuminate\Http\Request)) #57 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(30): Illuminate\Foundation\Http\Kernel->Illuminate\Foundation\Http\{closure}(Object(Illuminate\Http\Request)) #58 /usr/share/nginx/html/vendor/fideloper/proxy/src/TrustProxies.php(55): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #59 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Fideloper\Proxy\TrustProxies->handle(Object(Illuminate\Http\Request), Object(Closure)) #60 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #61 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/TransformsRequest.php(30): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #62 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\TransformsRequest->handle(Object(Illuminate\Http\Request), Object(Closure)) #63 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #64 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/AddQueuedCookiesToResponse.php(37): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #65 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\AddQueuedCookiesToResponse->handle(Object(Illuminate\Http\Request), Object(Closure)) #66 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #67 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Cookie/Middleware/EncryptCookies.php(59): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #68 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Cookie\Middleware\EncryptCookies->handle(Object(Illuminate\Http\Request), Object(Closure)) #69 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #70 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Middleware/CheckForMaintenanceMode.php(46): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #71 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(148): Illuminate\Foundation\Http\Middleware\CheckForMaintenanceMode->handle(Object(Illuminate\Http\Request), Object(Closure)) #72 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Routing/Pipeline.php(53): Illuminate\Pipeline\Pipeline->Illuminate\Pipeline\{closure}(Object(Illuminate\Http\Request)) #73 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Pipeline/Pipeline.php(102): Illuminate\Routing\Pipeline->Illuminate\Routing\{closure}(Object(Illuminate\Http\Request)) #74 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(149): Illuminate\Pipeline\Pipeline->then(Object(Closure)) #75 /usr/share/nginx/html/vendor/laravel/framework/src/Illuminate/Foundation/Http/Kernel.php(116): Illuminate\Foundation\Http\Kernel->sendRequestThroughRouter(Object(Illuminate\Http\Request)) #76 /usr/share/nginx/html/public/index.php(52): Illuminate\Foundation\Http\Kernel->handle(Object(Illuminate\Http\Request)) #77 {main} ```
non_process
deleting two services fails with an error hello new pterodactyl implementation clean without any modifications when i m trying to delete minecraft and source engine services i get an error occurred while attempting to update this service this error has been logged production error pdoexception sqlstate integrity constraint violation cannot delete or update a parent row a foreign key constraint fails dockermanager service options constraint service options config from foreign foreign key config from references service options id in usr share nginx html vendor doctrine dbal lib doctrine dbal driver pdostatement php stack trace usr share nginx html vendor doctrine dbal lib doctrine dbal driver pdostatement php pdostatement execute null usr share nginx html vendor laravel framework src illuminate database connection php doctrine dbal driver pdostatement execute usr share nginx html vendor laravel framework src illuminate database connection php illuminate database connection illuminate database closure delete from se array usr share nginx html vendor laravel framework src illuminate database connection php illuminate database connection runquerycallback delete from se array object closure usr share nginx html vendor laravel framework src illuminate database connection php illuminate database connection run delete from se array object closure usr share nginx html vendor laravel framework src illuminate database connection php illuminate database connection affectingstatement delete from se array usr share nginx html vendor laravel framework src illuminate database query builder php illuminate database connection delete delete from se array usr share nginx html vendor laravel framework src illuminate database eloquent builder php illuminate database query builder delete usr share nginx html vendor laravel framework src illuminate database eloquent model php illuminate database eloquent builder delete usr share nginx html vendor laravel framework src illuminate database eloquent model php illuminate database eloquent model performdeleteonmodel usr share nginx html app repositories optionrepository php illuminate database eloquent model delete usr share nginx html vendor laravel framework src illuminate database concerns managestransactions php pterodactyl repositories optionrepository pterodactyl repositories closure object illuminate database mysqlconnection usr share nginx html vendor laravel framework src illuminate database databasemanager php illuminate database connection transaction object closure usr share nginx html vendor laravel framework src illuminate support facades facade php illuminate database databasemanager call transaction array usr share nginx html app repositories optionrepository php illuminate support facades facade callstatic transaction array usr share nginx html app repositories servicerepository php pterodactyl repositories optionrepository delete usr share nginx html vendor laravel framework src illuminate database concerns managestransactions php pterodactyl repositories servicerepository pterodactyl repositories closure object illuminate database mysqlconnection usr share nginx html vendor laravel framework src illuminate database databasemanager php illuminate database connection transaction object closure usr share nginx html vendor laravel framework src illuminate support facades facade php illuminate database databasemanager call transaction array usr share nginx html app repositories servicerepository php illuminate support facades facade callstatic transaction array usr share nginx html app http controllers admin servicecontroller php pterodactyl repositories servicerepository delete pterodactyl http controllers admin servicecontroller edit object illuminate http request usr share nginx html vendor laravel framework src illuminate routing controller php call user func array array array usr share nginx html vendor laravel framework src illuminate routing controllerdispatcher php illuminate routing controller callaction edit array usr share nginx html vendor laravel framework src illuminate routing route php illuminate routing controllerdispatcher dispatch object illuminate routing route object pterodactyl http controllers admin servicecontroller edit usr share nginx html vendor laravel framework src illuminate routing route php illuminate routing route runcontroller usr share nginx html vendor laravel framework src illuminate routing router php illuminate routing route run usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate routing router illuminate routing closure object illuminate http request usr share nginx html app http middleware adminauthenticate php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php pterodactyl http middleware adminauthenticate handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html app http middleware languagemiddleware php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php pterodactyl http middleware languagemiddleware handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate routing middleware substitutebindings php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate routing middleware substitutebindings handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate auth middleware authenticate php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate auth middleware authenticate handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http middleware verifycsrftoken php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate foundation http middleware verifycsrftoken handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate view middleware shareerrorsfromsession php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate view middleware shareerrorsfromsession handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate session middleware startsession php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate session middleware startsession handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate cookie middleware addqueuedcookiestoresponse php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate cookie middleware addqueuedcookiestoresponse handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate cookie middleware encryptcookies php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate cookie middleware encryptcookies handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate routing router php illuminate pipeline pipeline then object closure usr share nginx html vendor laravel framework src illuminate routing router php illuminate routing router runroutewithinstack object illuminate routing route object illuminate http request usr share nginx html vendor laravel framework src illuminate routing router php illuminate routing router dispatchtoroute object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http kernel php illuminate routing router dispatch object illuminate http request usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate foundation http kernel illuminate foundation http closure object illuminate http request usr share nginx html vendor fideloper proxy src trustproxies php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php fideloper proxy trustproxies handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http middleware transformsrequest php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate foundation http middleware transformsrequest handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate cookie middleware addqueuedcookiestoresponse php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate cookie middleware addqueuedcookiestoresponse handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate cookie middleware encryptcookies php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate cookie middleware encryptcookies handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http middleware checkformaintenancemode php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate foundation http middleware checkformaintenancemode handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http kernel php illuminate pipeline pipeline then object closure usr share nginx html vendor laravel framework src illuminate foundation http kernel php illuminate foundation http kernel sendrequestthroughrouter object illuminate http request usr share nginx html public index php illuminate foundation http kernel handle object illuminate http request main next doctrine dbal driver pdoexception sqlstate integrity constraint violation cannot delete or update a parent row a foreign key constraint fails dockermanager service options constraint service options config from foreign foreign key config from references service options id in usr share nginx html vendor doctrine dbal lib doctrine dbal driver pdostatement php stack trace usr share nginx html vendor laravel framework src illuminate database connection php doctrine dbal driver pdostatement execute usr share nginx html vendor laravel framework src illuminate database connection php illuminate database connection illuminate database closure delete from se array usr share nginx html vendor laravel framework src illuminate database connection php illuminate database connection runquerycallback delete from se array object closure usr share nginx html vendor laravel framework src illuminate database connection php illuminate database connection run delete from se array object closure usr share nginx html vendor laravel framework src illuminate database connection php illuminate database connection affectingstatement delete from se array usr share nginx html vendor laravel framework src illuminate database query builder php illuminate database connection delete delete from se array usr share nginx html vendor laravel framework src illuminate database eloquent builder php illuminate database query builder delete usr share nginx html vendor laravel framework src illuminate database eloquent model php illuminate database eloquent builder delete usr share nginx html vendor laravel framework src illuminate database eloquent model php illuminate database eloquent model performdeleteonmodel usr share nginx html app repositories optionrepository php illuminate database eloquent model delete usr share nginx html vendor laravel framework src illuminate database concerns managestransactions php pterodactyl repositories optionrepository pterodactyl repositories closure object illuminate database mysqlconnection usr share nginx html vendor laravel framework src illuminate database databasemanager php illuminate database connection transaction object closure usr share nginx html vendor laravel framework src illuminate support facades facade php illuminate database databasemanager call transaction array usr share nginx html app repositories optionrepository php illuminate support facades facade callstatic transaction array usr share nginx html app repositories servicerepository php pterodactyl repositories optionrepository delete usr share nginx html vendor laravel framework src illuminate database concerns managestransactions php pterodactyl repositories servicerepository pterodactyl repositories closure object illuminate database mysqlconnection usr share nginx html vendor laravel framework src illuminate database databasemanager php illuminate database connection transaction object closure usr share nginx html vendor laravel framework src illuminate support facades facade php illuminate database databasemanager call transaction array usr share nginx html app repositories servicerepository php illuminate support facades facade callstatic transaction array usr share nginx html app http controllers admin servicecontroller php pterodactyl repositories servicerepository delete pterodactyl http controllers admin servicecontroller edit object illuminate http request usr share nginx html vendor laravel framework src illuminate routing controller php call user func array array array usr share nginx html vendor laravel framework src illuminate routing controllerdispatcher php illuminate routing controller callaction edit array usr share nginx html vendor laravel framework src illuminate routing route php illuminate routing controllerdispatcher dispatch object illuminate routing route object pterodactyl http controllers admin servicecontroller edit usr share nginx html vendor laravel framework src illuminate routing route php illuminate routing route runcontroller usr share nginx html vendor laravel framework src illuminate routing router php illuminate routing route run usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate routing router illuminate routing closure object illuminate http request usr share nginx html app http middleware adminauthenticate php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php pterodactyl http middleware adminauthenticate handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html app http middleware languagemiddleware php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php pterodactyl http middleware languagemiddleware handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate routing middleware substitutebindings php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate routing middleware substitutebindings handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate auth middleware authenticate php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate auth middleware authenticate handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http middleware verifycsrftoken php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate foundation http middleware verifycsrftoken handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate view middleware shareerrorsfromsession php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate view middleware shareerrorsfromsession handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate session middleware startsession php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate session middleware startsession handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate cookie middleware addqueuedcookiestoresponse php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate cookie middleware addqueuedcookiestoresponse handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate cookie middleware encryptcookies php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate cookie middleware encryptcookies handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate routing router php illuminate pipeline pipeline then object closure usr share nginx html vendor laravel framework src illuminate routing router php illuminate routing router runroutewithinstack object illuminate routing route object illuminate http request usr share nginx html vendor laravel framework src illuminate routing router php illuminate routing router dispatchtoroute object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http kernel php illuminate routing router dispatch object illuminate http request usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate foundation http kernel illuminate foundation http closure object illuminate http request usr share nginx html vendor fideloper proxy src trustproxies php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php fideloper proxy trustproxies handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http middleware transformsrequest php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate foundation http middleware transformsrequest handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate cookie middleware addqueuedcookiestoresponse php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate cookie middleware addqueuedcookiestoresponse handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate cookie middleware encryptcookies php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate cookie middleware encryptcookies handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http middleware checkformaintenancemode php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate foundation http middleware checkformaintenancemode handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http kernel php illuminate pipeline pipeline then object closure usr share nginx html vendor laravel framework src illuminate foundation http kernel php illuminate foundation http kernel sendrequestthroughrouter object illuminate http request usr share nginx html public index php illuminate foundation http kernel handle object illuminate http request main next illuminate database queryexception sqlstate integrity constraint violation cannot delete or update a parent row a foreign key constraint fails dockermanager service options constraint service options config from foreign foreign key config from references service options id sql delete from service options where id in usr share nginx html vendor laravel framework src illuminate database connection php stack trace usr share nginx html vendor laravel framework src illuminate database connection php illuminate database connection runquerycallback delete from se array object closure usr share nginx html vendor laravel framework src illuminate database connection php illuminate database connection run delete from se array object closure usr share nginx html vendor laravel framework src illuminate database connection php illuminate database connection affectingstatement delete from se array usr share nginx html vendor laravel framework src illuminate database query builder php illuminate database connection delete delete from se array usr share nginx html vendor laravel framework src illuminate database eloquent builder php illuminate database query builder delete usr share nginx html vendor laravel framework src illuminate database eloquent model php illuminate database eloquent builder delete usr share nginx html vendor laravel framework src illuminate database eloquent model php illuminate database eloquent model performdeleteonmodel usr share nginx html app repositories optionrepository php illuminate database eloquent model delete usr share nginx html vendor laravel framework src illuminate database concerns managestransactions php pterodactyl repositories optionrepository pterodactyl repositories closure object illuminate database mysqlconnection usr share nginx html vendor laravel framework src illuminate database databasemanager php illuminate database connection transaction object closure usr share nginx html vendor laravel framework src illuminate support facades facade php illuminate database databasemanager call transaction array usr share nginx html app repositories optionrepository php illuminate support facades facade callstatic transaction array usr share nginx html app repositories servicerepository php pterodactyl repositories optionrepository delete usr share nginx html vendor laravel framework src illuminate database concerns managestransactions php pterodactyl repositories servicerepository pterodactyl repositories closure object illuminate database mysqlconnection usr share nginx html vendor laravel framework src illuminate database databasemanager php illuminate database connection transaction object closure usr share nginx html vendor laravel framework src illuminate support facades facade php illuminate database databasemanager call transaction array usr share nginx html app repositories servicerepository php illuminate support facades facade callstatic transaction array usr share nginx html app http controllers admin servicecontroller php pterodactyl repositories servicerepository delete pterodactyl http controllers admin servicecontroller edit object illuminate http request usr share nginx html vendor laravel framework src illuminate routing controller php call user func array array array usr share nginx html vendor laravel framework src illuminate routing controllerdispatcher php illuminate routing controller callaction edit array usr share nginx html vendor laravel framework src illuminate routing route php illuminate routing controllerdispatcher dispatch object illuminate routing route object pterodactyl http controllers admin servicecontroller edit usr share nginx html vendor laravel framework src illuminate routing route php illuminate routing route runcontroller usr share nginx html vendor laravel framework src illuminate routing router php illuminate routing route run usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate routing router illuminate routing closure object illuminate http request usr share nginx html app http middleware adminauthenticate php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php pterodactyl http middleware adminauthenticate handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html app http middleware languagemiddleware php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php pterodactyl http middleware languagemiddleware handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate routing middleware substitutebindings php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate routing middleware substitutebindings handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate auth middleware authenticate php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate auth middleware authenticate handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http middleware verifycsrftoken php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate foundation http middleware verifycsrftoken handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate view middleware shareerrorsfromsession php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate view middleware shareerrorsfromsession handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate session middleware startsession php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate session middleware startsession handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate cookie middleware addqueuedcookiestoresponse php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate cookie middleware addqueuedcookiestoresponse handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate cookie middleware encryptcookies php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate cookie middleware encryptcookies handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate routing router php illuminate pipeline pipeline then object closure usr share nginx html vendor laravel framework src illuminate routing router php illuminate routing router runroutewithinstack object illuminate routing route object illuminate http request usr share nginx html vendor laravel framework src illuminate routing router php illuminate routing router dispatchtoroute object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http kernel php illuminate routing router dispatch object illuminate http request usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate foundation http kernel illuminate foundation http closure object illuminate http request usr share nginx html vendor fideloper proxy src trustproxies php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php fideloper proxy trustproxies handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http middleware transformsrequest php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate foundation http middleware transformsrequest handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate cookie middleware addqueuedcookiestoresponse php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate cookie middleware addqueuedcookiestoresponse handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate cookie middleware encryptcookies php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate cookie middleware encryptcookies handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http middleware checkformaintenancemode php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate foundation http middleware checkformaintenancemode handle object illuminate http request object closure usr share nginx html vendor laravel framework src illuminate routing pipeline php illuminate pipeline pipeline illuminate pipeline closure object illuminate http request usr share nginx html vendor laravel framework src illuminate pipeline pipeline php illuminate routing pipeline illuminate routing closure object illuminate http request usr share nginx html vendor laravel framework src illuminate foundation http kernel php illuminate pipeline pipeline then object closure usr share nginx html vendor laravel framework src illuminate foundation http kernel php illuminate foundation http kernel sendrequestthroughrouter object illuminate http request usr share nginx html public index php illuminate foundation http kernel handle object illuminate http request main
0
864
3,322,551,589
IssuesEvent
2015-11-09 15:00:36
luc-github/Repetier-Firmware-0.92
https://api.github.com/repos/luc-github/Repetier-Firmware-0.92
reopened
Simplify 3D stop randomly to print then resume after few seconds
bug Waiting to be processed
When this happen on serial monitor I see bunch of "wait" commands and on LCD the buffer decreasing, Can navigate in printer menu when this happen - so printer is not frozen and this does not trigger the watchdog After few seconds print resume, I did not observed any crash neiter print failure but, seems weird. This never happened with repetier host so far I have set same setting for communication in S3D than for repetier 202 cache / Hw flow control / baud rate 230400 not sure how to address this
1.0
Simplify 3D stop randomly to print then resume after few seconds - When this happen on serial monitor I see bunch of "wait" commands and on LCD the buffer decreasing, Can navigate in printer menu when this happen - so printer is not frozen and this does not trigger the watchdog After few seconds print resume, I did not observed any crash neiter print failure but, seems weird. This never happened with repetier host so far I have set same setting for communication in S3D than for repetier 202 cache / Hw flow control / baud rate 230400 not sure how to address this
process
simplify stop randomly to print then resume after few seconds when this happen on serial monitor i see bunch of wait commands and on lcd the buffer decreasing can navigate in printer menu when this happen so printer is not frozen and this does not trigger the watchdog after few seconds print resume i did not observed any crash neiter print failure but seems weird this never happened with repetier host so far i have set same setting for communication in than for repetier cache hw flow control baud rate not sure how to address this
1
524,999
15,227,214,155
IssuesEvent
2021-02-18 09:54:50
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
Holder of Key Validation support for JWT access tokens
Priority/Normal Type/New Feature
### Describe your problem(s) Client A and Client B have two applications in APIM. Client A obtains a token T from APIM. Client B sends a request to access an API using the token T. This will not be failed with MTLS ( Since B also a valid client to APIM, his transport certificate is in APIM trust-store). HOK Mechanism suggests binding the transport certificate to the token. ( Introspection should return the hash value of transport cert). Then per each API call, we should validate the received transport cert against the transport cert bound to the token. When JWT access tokens are used with WSO2 API Manager, we can do this validation easily if the certificate hash is included in the JWT token itself. ### Describe your solution The specification regarding this feature is [1]. According to [1] Mutual TLS token binding has been used as the HoK mechanism. Once the token is bounded with the Mutual TLS cert, the hash of the certificate will be present in the following locations Introspection Response (for opaque tokens) [2] { "active": true,"iss": "https://server.example.com", "sub": "ty.webb@example.com", "exp": 1493726400, "nbf": 1493722800, "cnf":{ "x5t#S256": "bwcK0esc3ACC3DB2Y5_lESsXE8o9ltc05O89jdN-dg2" } } As a JWT claim (for JWT tokens) [3] { "iss": "https://server.example.com", "sub": "ty.webb@example.com", "exp": 1493726400, "nbf": 1493722800, "cnf":{ "x5t#S256": "bwcK0esc3ACC3DB2Y5_lESsXE8o9ltc05O89jdN-dg2" } } [1] https://tools.ietf.org/html/rfc8705 [2] https://tools.ietf.org/html/rfc8705#section-3.2 [3] https://tools.ietf.org/html/rfc8705#section-3.1
1.0
Holder of Key Validation support for JWT access tokens - ### Describe your problem(s) Client A and Client B have two applications in APIM. Client A obtains a token T from APIM. Client B sends a request to access an API using the token T. This will not be failed with MTLS ( Since B also a valid client to APIM, his transport certificate is in APIM trust-store). HOK Mechanism suggests binding the transport certificate to the token. ( Introspection should return the hash value of transport cert). Then per each API call, we should validate the received transport cert against the transport cert bound to the token. When JWT access tokens are used with WSO2 API Manager, we can do this validation easily if the certificate hash is included in the JWT token itself. ### Describe your solution The specification regarding this feature is [1]. According to [1] Mutual TLS token binding has been used as the HoK mechanism. Once the token is bounded with the Mutual TLS cert, the hash of the certificate will be present in the following locations Introspection Response (for opaque tokens) [2] { "active": true,"iss": "https://server.example.com", "sub": "ty.webb@example.com", "exp": 1493726400, "nbf": 1493722800, "cnf":{ "x5t#S256": "bwcK0esc3ACC3DB2Y5_lESsXE8o9ltc05O89jdN-dg2" } } As a JWT claim (for JWT tokens) [3] { "iss": "https://server.example.com", "sub": "ty.webb@example.com", "exp": 1493726400, "nbf": 1493722800, "cnf":{ "x5t#S256": "bwcK0esc3ACC3DB2Y5_lESsXE8o9ltc05O89jdN-dg2" } } [1] https://tools.ietf.org/html/rfc8705 [2] https://tools.ietf.org/html/rfc8705#section-3.2 [3] https://tools.ietf.org/html/rfc8705#section-3.1
non_process
holder of key validation support for jwt access tokens describe your problem s client a and client b have two applications in apim client a obtains a token t from apim client b sends a request to access an api using the token t this will not be failed with mtls since b also a valid client to apim his transport certificate is in apim trust store hok mechanism suggests binding the transport certificate to the token introspection should return the hash value of transport cert then per each api call we should validate the received transport cert against the transport cert bound to the token when jwt access tokens are used with api manager we can do this validation easily if the certificate hash is included in the jwt token itself describe your solution the specification regarding this feature is according to mutual tls token binding has been used as the hok mechanism once the token is bounded with the mutual tls cert the hash of the certificate will be present in the following locations introspection response for opaque tokens active true iss sub ty webb example com exp nbf cnf as a jwt claim for jwt tokens iss sub ty webb example com exp nbf cnf
0
333,439
24,374,874,678
IssuesEvent
2022-10-03 23:21:47
SJSURoboticsTeam/urc-control-systems-2022
https://api.github.com/repos/SJSURoboticsTeam/urc-control-systems-2022
closed
Research GPS modules
documentation
Need a GPS that talks over UART Will need a GPS for competition purposes. Dev2 doesn't have any GPS drivers so we would need to create our own driver.
1.0
Research GPS modules - Need a GPS that talks over UART Will need a GPS for competition purposes. Dev2 doesn't have any GPS drivers so we would need to create our own driver.
non_process
research gps modules need a gps that talks over uart will need a gps for competition purposes doesn t have any gps drivers so we would need to create our own driver
0
14,945
18,425,736,371
IssuesEvent
2021-10-13 21:44:58
brucemiller/LaTeXML
https://api.github.com/repos/brucemiller/LaTeXML
closed
Smarter sizing in PDF to PNG image conversion
enhancement postprocessing
As a running example, take arXiv's [1804.11027](https://arxiv.org/pdf/1804.11027.pdf), in particular Figure 11 on page 8. The two images in that figure, titled `CUHK03` and `Market-1501` are PDF files in the original paper sources, and the latex for including them is: ```tex \begin{figure}[t] \centering \begin{tabular}{c} \includegraphics[height=5cm,width=7cm]{cmc-cuhk03}\\ \includegraphics[height=5cm,width=7cm]{cmc-market} \end{tabular} \caption{The CMC curves on CUHK03 and Market-1501 datasets.} \label{fig:cmc-cuhk03-market} \end{figure} ``` My main focus with this issue is how to avoid harming the image quality, when respecting the sizing arguments (`[height=5cm,width=7cm]`) and creating a new PNG file. Currently, these particular sizing parameters generate an image so small as to be completely illegible: ![image](https://user-images.githubusercontent.com/348975/135242272-a63ee4d9-690e-4eaf-b9e5-e3f8563ba0c6.png) Converting with the sizing arguments removed creates a legible (but, for arXiv, unconstrained to its maximum size) plot PNG: ![image](https://user-images.githubusercontent.com/348975/135242149-7adadb45-6b87-4575-9184-c423b4bb5cf0.png) As a brainstorming, half-baked, usability suggestion to conclude with: Could we rework the width and height parameters in particular as attributes on the image container, rather than as inherent properties of the PNG image? And err on the side of creating higher-quality / larger size PNGs than necessary.
1.0
Smarter sizing in PDF to PNG image conversion - As a running example, take arXiv's [1804.11027](https://arxiv.org/pdf/1804.11027.pdf), in particular Figure 11 on page 8. The two images in that figure, titled `CUHK03` and `Market-1501` are PDF files in the original paper sources, and the latex for including them is: ```tex \begin{figure}[t] \centering \begin{tabular}{c} \includegraphics[height=5cm,width=7cm]{cmc-cuhk03}\\ \includegraphics[height=5cm,width=7cm]{cmc-market} \end{tabular} \caption{The CMC curves on CUHK03 and Market-1501 datasets.} \label{fig:cmc-cuhk03-market} \end{figure} ``` My main focus with this issue is how to avoid harming the image quality, when respecting the sizing arguments (`[height=5cm,width=7cm]`) and creating a new PNG file. Currently, these particular sizing parameters generate an image so small as to be completely illegible: ![image](https://user-images.githubusercontent.com/348975/135242272-a63ee4d9-690e-4eaf-b9e5-e3f8563ba0c6.png) Converting with the sizing arguments removed creates a legible (but, for arXiv, unconstrained to its maximum size) plot PNG: ![image](https://user-images.githubusercontent.com/348975/135242149-7adadb45-6b87-4575-9184-c423b4bb5cf0.png) As a brainstorming, half-baked, usability suggestion to conclude with: Could we rework the width and height parameters in particular as attributes on the image container, rather than as inherent properties of the PNG image? And err on the side of creating higher-quality / larger size PNGs than necessary.
process
smarter sizing in pdf to png image conversion as a running example take arxiv s in particular figure on page the two images in that figure titled and market are pdf files in the original paper sources and the latex for including them is tex begin figure centering begin tabular c includegraphics cmc includegraphics cmc market end tabular caption the cmc curves on and market datasets label fig cmc market end figure my main focus with this issue is how to avoid harming the image quality when respecting the sizing arguments and creating a new png file currently these particular sizing parameters generate an image so small as to be completely illegible converting with the sizing arguments removed creates a legible but for arxiv unconstrained to its maximum size plot png as a brainstorming half baked usability suggestion to conclude with could we rework the width and height parameters in particular as attributes on the image container rather than as inherent properties of the png image and err on the side of creating higher quality larger size pngs than necessary
1